at v6.13-rc4 1666 lines 40 kB view raw
1// SPDX-License-Identifier: GPL-2.0 2#include <asm/bug.h> 3#include <linux/kernel.h> 4#include <linux/string.h> 5#include <linux/zalloc.h> 6#include <sys/time.h> 7#include <sys/resource.h> 8#include <sys/types.h> 9#include <sys/stat.h> 10#include <unistd.h> 11#include <errno.h> 12#include <fcntl.h> 13#include <stdlib.h> 14#ifdef HAVE_LIBBPF_SUPPORT 15#include <bpf/libbpf.h> 16#include "bpf-event.h" 17#include "bpf-utils.h" 18#endif 19#include "compress.h" 20#include "env.h" 21#include "namespaces.h" 22#include "path.h" 23#include "map.h" 24#include "symbol.h" 25#include "srcline.h" 26#include "dso.h" 27#include "dsos.h" 28#include "machine.h" 29#include "auxtrace.h" 30#include "util.h" /* O_CLOEXEC for older systems */ 31#include "debug.h" 32#include "string2.h" 33#include "vdso.h" 34#include "annotate-data.h" 35 36static const char * const debuglink_paths[] = { 37 "%.0s%s", 38 "%s/%s", 39 "%s/.debug/%s", 40 "/usr/lib/debug%s/%s" 41}; 42 43void dso__set_nsinfo(struct dso *dso, struct nsinfo *nsi) 44{ 45 nsinfo__put(RC_CHK_ACCESS(dso)->nsinfo); 46 RC_CHK_ACCESS(dso)->nsinfo = nsi; 47} 48 49char dso__symtab_origin(const struct dso *dso) 50{ 51 static const char origin[] = { 52 [DSO_BINARY_TYPE__KALLSYMS] = 'k', 53 [DSO_BINARY_TYPE__VMLINUX] = 'v', 54 [DSO_BINARY_TYPE__JAVA_JIT] = 'j', 55 [DSO_BINARY_TYPE__DEBUGLINK] = 'l', 56 [DSO_BINARY_TYPE__BUILD_ID_CACHE] = 'B', 57 [DSO_BINARY_TYPE__BUILD_ID_CACHE_DEBUGINFO] = 'D', 58 [DSO_BINARY_TYPE__FEDORA_DEBUGINFO] = 'f', 59 [DSO_BINARY_TYPE__UBUNTU_DEBUGINFO] = 'u', 60 [DSO_BINARY_TYPE__MIXEDUP_UBUNTU_DEBUGINFO] = 'x', 61 [DSO_BINARY_TYPE__OPENEMBEDDED_DEBUGINFO] = 'o', 62 [DSO_BINARY_TYPE__BUILDID_DEBUGINFO] = 'b', 63 [DSO_BINARY_TYPE__SYSTEM_PATH_DSO] = 'd', 64 [DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE] = 'K', 65 [DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP] = 'm', 66 [DSO_BINARY_TYPE__GUEST_KALLSYMS] = 'g', 67 [DSO_BINARY_TYPE__GUEST_KMODULE] = 'G', 68 [DSO_BINARY_TYPE__GUEST_KMODULE_COMP] = 'M', 69 [DSO_BINARY_TYPE__GUEST_VMLINUX] = 'V', 70 }; 71 72 if (dso == NULL || dso__symtab_type(dso) == DSO_BINARY_TYPE__NOT_FOUND) 73 return '!'; 74 return origin[dso__symtab_type(dso)]; 75} 76 77bool dso__is_object_file(const struct dso *dso) 78{ 79 switch (dso__binary_type(dso)) { 80 case DSO_BINARY_TYPE__KALLSYMS: 81 case DSO_BINARY_TYPE__GUEST_KALLSYMS: 82 case DSO_BINARY_TYPE__JAVA_JIT: 83 case DSO_BINARY_TYPE__BPF_PROG_INFO: 84 case DSO_BINARY_TYPE__BPF_IMAGE: 85 case DSO_BINARY_TYPE__OOL: 86 return false; 87 case DSO_BINARY_TYPE__VMLINUX: 88 case DSO_BINARY_TYPE__GUEST_VMLINUX: 89 case DSO_BINARY_TYPE__DEBUGLINK: 90 case DSO_BINARY_TYPE__BUILD_ID_CACHE: 91 case DSO_BINARY_TYPE__BUILD_ID_CACHE_DEBUGINFO: 92 case DSO_BINARY_TYPE__FEDORA_DEBUGINFO: 93 case DSO_BINARY_TYPE__UBUNTU_DEBUGINFO: 94 case DSO_BINARY_TYPE__MIXEDUP_UBUNTU_DEBUGINFO: 95 case DSO_BINARY_TYPE__BUILDID_DEBUGINFO: 96 case DSO_BINARY_TYPE__SYSTEM_PATH_DSO: 97 case DSO_BINARY_TYPE__GUEST_KMODULE: 98 case DSO_BINARY_TYPE__GUEST_KMODULE_COMP: 99 case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE: 100 case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP: 101 case DSO_BINARY_TYPE__KCORE: 102 case DSO_BINARY_TYPE__GUEST_KCORE: 103 case DSO_BINARY_TYPE__OPENEMBEDDED_DEBUGINFO: 104 case DSO_BINARY_TYPE__NOT_FOUND: 105 default: 106 return true; 107 } 108} 109 110int dso__read_binary_type_filename(const struct dso *dso, 111 enum dso_binary_type type, 112 char *root_dir, char *filename, size_t size) 113{ 114 char build_id_hex[SBUILD_ID_SIZE]; 115 int ret = 0; 116 size_t len; 117 118 switch (type) { 119 case DSO_BINARY_TYPE__DEBUGLINK: 120 { 121 const char *last_slash; 122 char dso_dir[PATH_MAX]; 123 char symfile[PATH_MAX]; 124 unsigned int i; 125 126 len = __symbol__join_symfs(filename, size, dso__long_name(dso)); 127 last_slash = filename + len; 128 while (last_slash != filename && *last_slash != '/') 129 last_slash--; 130 131 strncpy(dso_dir, filename, last_slash - filename); 132 dso_dir[last_slash-filename] = '\0'; 133 134 if (!is_regular_file(filename)) { 135 ret = -1; 136 break; 137 } 138 139 ret = filename__read_debuglink(filename, symfile, PATH_MAX); 140 if (ret) 141 break; 142 143 /* Check predefined locations where debug file might reside */ 144 ret = -1; 145 for (i = 0; i < ARRAY_SIZE(debuglink_paths); i++) { 146 snprintf(filename, size, 147 debuglink_paths[i], dso_dir, symfile); 148 if (is_regular_file(filename)) { 149 ret = 0; 150 break; 151 } 152 } 153 154 break; 155 } 156 case DSO_BINARY_TYPE__BUILD_ID_CACHE: 157 if (dso__build_id_filename(dso, filename, size, false) == NULL) 158 ret = -1; 159 break; 160 161 case DSO_BINARY_TYPE__BUILD_ID_CACHE_DEBUGINFO: 162 if (dso__build_id_filename(dso, filename, size, true) == NULL) 163 ret = -1; 164 break; 165 166 case DSO_BINARY_TYPE__FEDORA_DEBUGINFO: 167 len = __symbol__join_symfs(filename, size, "/usr/lib/debug"); 168 snprintf(filename + len, size - len, "%s.debug", dso__long_name(dso)); 169 break; 170 171 case DSO_BINARY_TYPE__UBUNTU_DEBUGINFO: 172 len = __symbol__join_symfs(filename, size, "/usr/lib/debug"); 173 snprintf(filename + len, size - len, "%s", dso__long_name(dso)); 174 break; 175 176 case DSO_BINARY_TYPE__MIXEDUP_UBUNTU_DEBUGINFO: 177 /* 178 * Ubuntu can mixup /usr/lib with /lib, putting debuginfo in 179 * /usr/lib/debug/lib when it is expected to be in 180 * /usr/lib/debug/usr/lib 181 */ 182 if (strlen(dso__long_name(dso)) < 9 || 183 strncmp(dso__long_name(dso), "/usr/lib/", 9)) { 184 ret = -1; 185 break; 186 } 187 len = __symbol__join_symfs(filename, size, "/usr/lib/debug"); 188 snprintf(filename + len, size - len, "%s", dso__long_name(dso) + 4); 189 break; 190 191 case DSO_BINARY_TYPE__OPENEMBEDDED_DEBUGINFO: 192 { 193 const char *last_slash; 194 size_t dir_size; 195 196 last_slash = dso__long_name(dso) + dso__long_name_len(dso); 197 while (last_slash != dso__long_name(dso) && *last_slash != '/') 198 last_slash--; 199 200 len = __symbol__join_symfs(filename, size, ""); 201 dir_size = last_slash - dso__long_name(dso) + 2; 202 if (dir_size > (size - len)) { 203 ret = -1; 204 break; 205 } 206 len += scnprintf(filename + len, dir_size, "%s", dso__long_name(dso)); 207 len += scnprintf(filename + len , size - len, ".debug%s", 208 last_slash); 209 break; 210 } 211 212 case DSO_BINARY_TYPE__BUILDID_DEBUGINFO: 213 if (!dso__has_build_id(dso)) { 214 ret = -1; 215 break; 216 } 217 218 build_id__sprintf(dso__bid_const(dso), build_id_hex); 219 len = __symbol__join_symfs(filename, size, "/usr/lib/debug/.build-id/"); 220 snprintf(filename + len, size - len, "%.2s/%s.debug", 221 build_id_hex, build_id_hex + 2); 222 break; 223 224 case DSO_BINARY_TYPE__VMLINUX: 225 case DSO_BINARY_TYPE__GUEST_VMLINUX: 226 case DSO_BINARY_TYPE__SYSTEM_PATH_DSO: 227 __symbol__join_symfs(filename, size, dso__long_name(dso)); 228 break; 229 230 case DSO_BINARY_TYPE__GUEST_KMODULE: 231 case DSO_BINARY_TYPE__GUEST_KMODULE_COMP: 232 path__join3(filename, size, symbol_conf.symfs, 233 root_dir, dso__long_name(dso)); 234 break; 235 236 case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE: 237 case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP: 238 __symbol__join_symfs(filename, size, dso__long_name(dso)); 239 break; 240 241 case DSO_BINARY_TYPE__KCORE: 242 case DSO_BINARY_TYPE__GUEST_KCORE: 243 snprintf(filename, size, "%s", dso__long_name(dso)); 244 break; 245 246 default: 247 case DSO_BINARY_TYPE__KALLSYMS: 248 case DSO_BINARY_TYPE__GUEST_KALLSYMS: 249 case DSO_BINARY_TYPE__JAVA_JIT: 250 case DSO_BINARY_TYPE__BPF_PROG_INFO: 251 case DSO_BINARY_TYPE__BPF_IMAGE: 252 case DSO_BINARY_TYPE__OOL: 253 case DSO_BINARY_TYPE__NOT_FOUND: 254 ret = -1; 255 break; 256 } 257 258 return ret; 259} 260 261enum { 262 COMP_ID__NONE = 0, 263}; 264 265static const struct { 266 const char *fmt; 267 int (*decompress)(const char *input, int output); 268 bool (*is_compressed)(const char *input); 269} compressions[] = { 270 [COMP_ID__NONE] = { .fmt = NULL, }, 271#ifdef HAVE_ZLIB_SUPPORT 272 { "gz", gzip_decompress_to_file, gzip_is_compressed }, 273#endif 274#ifdef HAVE_LZMA_SUPPORT 275 { "xz", lzma_decompress_to_file, lzma_is_compressed }, 276#endif 277 { NULL, NULL, NULL }, 278}; 279 280static int is_supported_compression(const char *ext) 281{ 282 unsigned i; 283 284 for (i = 1; compressions[i].fmt; i++) { 285 if (!strcmp(ext, compressions[i].fmt)) 286 return i; 287 } 288 return COMP_ID__NONE; 289} 290 291bool is_kernel_module(const char *pathname, int cpumode) 292{ 293 struct kmod_path m; 294 int mode = cpumode & PERF_RECORD_MISC_CPUMODE_MASK; 295 296 WARN_ONCE(mode != cpumode, 297 "Internal error: passing unmasked cpumode (%x) to is_kernel_module", 298 cpumode); 299 300 switch (mode) { 301 case PERF_RECORD_MISC_USER: 302 case PERF_RECORD_MISC_HYPERVISOR: 303 case PERF_RECORD_MISC_GUEST_USER: 304 return false; 305 /* Treat PERF_RECORD_MISC_CPUMODE_UNKNOWN as kernel */ 306 default: 307 if (kmod_path__parse(&m, pathname)) { 308 pr_err("Failed to check whether %s is a kernel module or not. Assume it is.", 309 pathname); 310 return true; 311 } 312 } 313 314 return m.kmod; 315} 316 317bool dso__needs_decompress(struct dso *dso) 318{ 319 return dso__symtab_type(dso) == DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP || 320 dso__symtab_type(dso) == DSO_BINARY_TYPE__GUEST_KMODULE_COMP; 321} 322 323int filename__decompress(const char *name, char *pathname, 324 size_t len, int comp, int *err) 325{ 326 char tmpbuf[] = KMOD_DECOMP_NAME; 327 int fd = -1; 328 329 /* 330 * We have proper compression id for DSO and yet the file 331 * behind the 'name' can still be plain uncompressed object. 332 * 333 * The reason is behind the logic we open the DSO object files, 334 * when we try all possible 'debug' objects until we find the 335 * data. So even if the DSO is represented by 'krava.xz' module, 336 * we can end up here opening ~/.debug/....23432432/debug' file 337 * which is not compressed. 338 * 339 * To keep this transparent, we detect this and return the file 340 * descriptor to the uncompressed file. 341 */ 342 if (!compressions[comp].is_compressed(name)) 343 return open(name, O_RDONLY); 344 345 fd = mkstemp(tmpbuf); 346 if (fd < 0) { 347 *err = errno; 348 return -1; 349 } 350 351 if (compressions[comp].decompress(name, fd)) { 352 *err = DSO_LOAD_ERRNO__DECOMPRESSION_FAILURE; 353 close(fd); 354 fd = -1; 355 } 356 357 if (!pathname || (fd < 0)) 358 unlink(tmpbuf); 359 360 if (pathname && (fd >= 0)) 361 strlcpy(pathname, tmpbuf, len); 362 363 return fd; 364} 365 366static int decompress_kmodule(struct dso *dso, const char *name, 367 char *pathname, size_t len) 368{ 369 if (!dso__needs_decompress(dso)) 370 return -1; 371 372 if (dso__comp(dso) == COMP_ID__NONE) 373 return -1; 374 375 return filename__decompress(name, pathname, len, dso__comp(dso), dso__load_errno(dso)); 376} 377 378int dso__decompress_kmodule_fd(struct dso *dso, const char *name) 379{ 380 return decompress_kmodule(dso, name, NULL, 0); 381} 382 383int dso__decompress_kmodule_path(struct dso *dso, const char *name, 384 char *pathname, size_t len) 385{ 386 int fd = decompress_kmodule(dso, name, pathname, len); 387 388 close(fd); 389 return fd >= 0 ? 0 : -1; 390} 391 392/* 393 * Parses kernel module specified in @path and updates 394 * @m argument like: 395 * 396 * @comp - true if @path contains supported compression suffix, 397 * false otherwise 398 * @kmod - true if @path contains '.ko' suffix in right position, 399 * false otherwise 400 * @name - if (@alloc_name && @kmod) is true, it contains strdup-ed base name 401 * of the kernel module without suffixes, otherwise strudup-ed 402 * base name of @path 403 * @ext - if (@alloc_ext && @comp) is true, it contains strdup-ed string 404 * the compression suffix 405 * 406 * Returns 0 if there's no strdup error, -ENOMEM otherwise. 407 */ 408int __kmod_path__parse(struct kmod_path *m, const char *path, 409 bool alloc_name) 410{ 411 const char *name = strrchr(path, '/'); 412 const char *ext = strrchr(path, '.'); 413 bool is_simple_name = false; 414 415 memset(m, 0x0, sizeof(*m)); 416 name = name ? name + 1 : path; 417 418 /* 419 * '.' is also a valid character for module name. For example: 420 * [aaa.bbb] is a valid module name. '[' should have higher 421 * priority than '.ko' suffix. 422 * 423 * The kernel names are from machine__mmap_name. Such 424 * name should belong to kernel itself, not kernel module. 425 */ 426 if (name[0] == '[') { 427 is_simple_name = true; 428 if ((strncmp(name, "[kernel.kallsyms]", 17) == 0) || 429 (strncmp(name, "[guest.kernel.kallsyms", 22) == 0) || 430 (strncmp(name, "[vdso]", 6) == 0) || 431 (strncmp(name, "[vdso32]", 8) == 0) || 432 (strncmp(name, "[vdsox32]", 9) == 0) || 433 (strncmp(name, "[vsyscall]", 10) == 0)) { 434 m->kmod = false; 435 436 } else 437 m->kmod = true; 438 } 439 440 /* No extension, just return name. */ 441 if ((ext == NULL) || is_simple_name) { 442 if (alloc_name) { 443 m->name = strdup(name); 444 return m->name ? 0 : -ENOMEM; 445 } 446 return 0; 447 } 448 449 m->comp = is_supported_compression(ext + 1); 450 if (m->comp > COMP_ID__NONE) 451 ext -= 3; 452 453 /* Check .ko extension only if there's enough name left. */ 454 if (ext > name) 455 m->kmod = !strncmp(ext, ".ko", 3); 456 457 if (alloc_name) { 458 if (m->kmod) { 459 if (asprintf(&m->name, "[%.*s]", (int) (ext - name), name) == -1) 460 return -ENOMEM; 461 } else { 462 if (asprintf(&m->name, "%s", name) == -1) 463 return -ENOMEM; 464 } 465 466 strreplace(m->name, '-', '_'); 467 } 468 469 return 0; 470} 471 472void dso__set_module_info(struct dso *dso, struct kmod_path *m, 473 struct machine *machine) 474{ 475 if (machine__is_host(machine)) 476 dso__set_symtab_type(dso, DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE); 477 else 478 dso__set_symtab_type(dso, DSO_BINARY_TYPE__GUEST_KMODULE); 479 480 /* _KMODULE_COMP should be next to _KMODULE */ 481 if (m->kmod && m->comp) { 482 dso__set_symtab_type(dso, dso__symtab_type(dso) + 1); 483 dso__set_comp(dso, m->comp); 484 } 485 486 dso__set_is_kmod(dso); 487 dso__set_short_name(dso, strdup(m->name), true); 488} 489 490/* 491 * Global list of open DSOs and the counter. 492 */ 493static LIST_HEAD(dso__data_open); 494static long dso__data_open_cnt; 495static pthread_mutex_t dso__data_open_lock = PTHREAD_MUTEX_INITIALIZER; 496 497static void dso__list_add(struct dso *dso) 498{ 499 list_add_tail(&dso__data(dso)->open_entry, &dso__data_open); 500#ifdef REFCNT_CHECKING 501 dso__data(dso)->dso = dso__get(dso); 502#endif 503 /* Assume the dso is part of dsos, hence the optional reference count above. */ 504 assert(dso__dsos(dso)); 505 dso__data_open_cnt++; 506} 507 508static void dso__list_del(struct dso *dso) 509{ 510 list_del_init(&dso__data(dso)->open_entry); 511#ifdef REFCNT_CHECKING 512 dso__put(dso__data(dso)->dso); 513#endif 514 WARN_ONCE(dso__data_open_cnt <= 0, 515 "DSO data fd counter out of bounds."); 516 dso__data_open_cnt--; 517} 518 519static void close_first_dso(void); 520 521static int do_open(char *name) 522{ 523 int fd; 524 char sbuf[STRERR_BUFSIZE]; 525 526 do { 527 fd = open(name, O_RDONLY|O_CLOEXEC); 528 if (fd >= 0) 529 return fd; 530 531 pr_debug("dso open failed: %s\n", 532 str_error_r(errno, sbuf, sizeof(sbuf))); 533 if (!dso__data_open_cnt || errno != EMFILE) 534 break; 535 536 close_first_dso(); 537 } while (1); 538 539 return -1; 540} 541 542char *dso__filename_with_chroot(const struct dso *dso, const char *filename) 543{ 544 return filename_with_chroot(nsinfo__pid(dso__nsinfo_const(dso)), filename); 545} 546 547static int __open_dso(struct dso *dso, struct machine *machine) 548{ 549 int fd = -EINVAL; 550 char *root_dir = (char *)""; 551 char *name = malloc(PATH_MAX); 552 bool decomp = false; 553 554 if (!name) 555 return -ENOMEM; 556 557 mutex_lock(dso__lock(dso)); 558 if (machine) 559 root_dir = machine->root_dir; 560 561 if (dso__read_binary_type_filename(dso, dso__binary_type(dso), 562 root_dir, name, PATH_MAX)) 563 goto out; 564 565 if (!is_regular_file(name)) { 566 char *new_name; 567 568 if (errno != ENOENT || dso__nsinfo(dso) == NULL) 569 goto out; 570 571 new_name = dso__filename_with_chroot(dso, name); 572 if (!new_name) 573 goto out; 574 575 free(name); 576 name = new_name; 577 } 578 579 if (dso__needs_decompress(dso)) { 580 char newpath[KMOD_DECOMP_LEN]; 581 size_t len = sizeof(newpath); 582 583 if (dso__decompress_kmodule_path(dso, name, newpath, len) < 0) { 584 fd = -(*dso__load_errno(dso)); 585 goto out; 586 } 587 588 decomp = true; 589 strcpy(name, newpath); 590 } 591 592 fd = do_open(name); 593 594 if (decomp) 595 unlink(name); 596 597out: 598 mutex_unlock(dso__lock(dso)); 599 free(name); 600 return fd; 601} 602 603static void check_data_close(void); 604 605/** 606 * dso_close - Open DSO data file 607 * @dso: dso object 608 * 609 * Open @dso's data file descriptor and updates 610 * list/count of open DSO objects. 611 */ 612static int open_dso(struct dso *dso, struct machine *machine) 613{ 614 int fd; 615 struct nscookie nsc; 616 617 if (dso__binary_type(dso) != DSO_BINARY_TYPE__BUILD_ID_CACHE) { 618 mutex_lock(dso__lock(dso)); 619 nsinfo__mountns_enter(dso__nsinfo(dso), &nsc); 620 mutex_unlock(dso__lock(dso)); 621 } 622 fd = __open_dso(dso, machine); 623 if (dso__binary_type(dso) != DSO_BINARY_TYPE__BUILD_ID_CACHE) 624 nsinfo__mountns_exit(&nsc); 625 626 if (fd >= 0) { 627 dso__list_add(dso); 628 /* 629 * Check if we crossed the allowed number 630 * of opened DSOs and close one if needed. 631 */ 632 check_data_close(); 633 } 634 635 return fd; 636} 637 638static void close_data_fd(struct dso *dso) 639{ 640 if (dso__data(dso)->fd >= 0) { 641 close(dso__data(dso)->fd); 642 dso__data(dso)->fd = -1; 643 dso__data(dso)->file_size = 0; 644 dso__list_del(dso); 645 } 646} 647 648/** 649 * dso_close - Close DSO data file 650 * @dso: dso object 651 * 652 * Close @dso's data file descriptor and updates 653 * list/count of open DSO objects. 654 */ 655static void close_dso(struct dso *dso) 656{ 657 close_data_fd(dso); 658} 659 660static void close_first_dso(void) 661{ 662 struct dso_data *dso_data; 663 struct dso *dso; 664 665 dso_data = list_first_entry(&dso__data_open, struct dso_data, open_entry); 666#ifdef REFCNT_CHECKING 667 dso = dso_data->dso; 668#else 669 dso = container_of(dso_data, struct dso, data); 670#endif 671 close_dso(dso); 672} 673 674static rlim_t get_fd_limit(void) 675{ 676 struct rlimit l; 677 rlim_t limit = 0; 678 679 /* Allow half of the current open fd limit. */ 680 if (getrlimit(RLIMIT_NOFILE, &l) == 0) { 681 if (l.rlim_cur == RLIM_INFINITY) 682 limit = l.rlim_cur; 683 else 684 limit = l.rlim_cur / 2; 685 } else { 686 pr_err("failed to get fd limit\n"); 687 limit = 1; 688 } 689 690 return limit; 691} 692 693static rlim_t fd_limit; 694 695/* 696 * Used only by tests/dso-data.c to reset the environment 697 * for tests. I dont expect we should change this during 698 * standard runtime. 699 */ 700void reset_fd_limit(void) 701{ 702 fd_limit = 0; 703} 704 705static bool may_cache_fd(void) 706{ 707 if (!fd_limit) 708 fd_limit = get_fd_limit(); 709 710 if (fd_limit == RLIM_INFINITY) 711 return true; 712 713 return fd_limit > (rlim_t) dso__data_open_cnt; 714} 715 716/* 717 * Check and close LRU dso if we crossed allowed limit 718 * for opened dso file descriptors. The limit is half 719 * of the RLIMIT_NOFILE files opened. 720*/ 721static void check_data_close(void) 722{ 723 bool cache_fd = may_cache_fd(); 724 725 if (!cache_fd) 726 close_first_dso(); 727} 728 729/** 730 * dso__data_close - Close DSO data file 731 * @dso: dso object 732 * 733 * External interface to close @dso's data file descriptor. 734 */ 735void dso__data_close(struct dso *dso) 736{ 737 pthread_mutex_lock(&dso__data_open_lock); 738 close_dso(dso); 739 pthread_mutex_unlock(&dso__data_open_lock); 740} 741 742static void try_to_open_dso(struct dso *dso, struct machine *machine) 743{ 744 enum dso_binary_type binary_type_data[] = { 745 DSO_BINARY_TYPE__BUILD_ID_CACHE, 746 DSO_BINARY_TYPE__SYSTEM_PATH_DSO, 747 DSO_BINARY_TYPE__NOT_FOUND, 748 }; 749 int i = 0; 750 struct dso_data *dso_data = dso__data(dso); 751 752 if (dso_data->fd >= 0) 753 return; 754 755 if (dso__binary_type(dso) != DSO_BINARY_TYPE__NOT_FOUND) { 756 dso_data->fd = open_dso(dso, machine); 757 goto out; 758 } 759 760 do { 761 dso__set_binary_type(dso, binary_type_data[i++]); 762 763 dso_data->fd = open_dso(dso, machine); 764 if (dso_data->fd >= 0) 765 goto out; 766 767 } while (dso__binary_type(dso) != DSO_BINARY_TYPE__NOT_FOUND); 768out: 769 if (dso_data->fd >= 0) 770 dso_data->status = DSO_DATA_STATUS_OK; 771 else 772 dso_data->status = DSO_DATA_STATUS_ERROR; 773} 774 775/** 776 * dso__data_get_fd - Get dso's data file descriptor 777 * @dso: dso object 778 * @machine: machine object 779 * 780 * External interface to find dso's file, open it and 781 * returns file descriptor. It should be paired with 782 * dso__data_put_fd() if it returns non-negative value. 783 */ 784int dso__data_get_fd(struct dso *dso, struct machine *machine) 785{ 786 if (dso__data(dso)->status == DSO_DATA_STATUS_ERROR) 787 return -1; 788 789 if (pthread_mutex_lock(&dso__data_open_lock) < 0) 790 return -1; 791 792 try_to_open_dso(dso, machine); 793 794 if (dso__data(dso)->fd < 0) 795 pthread_mutex_unlock(&dso__data_open_lock); 796 797 return dso__data(dso)->fd; 798} 799 800void dso__data_put_fd(struct dso *dso __maybe_unused) 801{ 802 pthread_mutex_unlock(&dso__data_open_lock); 803} 804 805bool dso__data_status_seen(struct dso *dso, enum dso_data_status_seen by) 806{ 807 u32 flag = 1 << by; 808 809 if (dso__data(dso)->status_seen & flag) 810 return true; 811 812 dso__data(dso)->status_seen |= flag; 813 814 return false; 815} 816 817#ifdef HAVE_LIBBPF_SUPPORT 818static ssize_t bpf_read(struct dso *dso, u64 offset, char *data) 819{ 820 struct bpf_prog_info_node *node; 821 ssize_t size = DSO__DATA_CACHE_SIZE; 822 struct dso_bpf_prog *dso_bpf_prog = dso__bpf_prog(dso); 823 u64 len; 824 u8 *buf; 825 826 node = perf_env__find_bpf_prog_info(dso_bpf_prog->env, dso_bpf_prog->id); 827 if (!node || !node->info_linear) { 828 dso__data(dso)->status = DSO_DATA_STATUS_ERROR; 829 return -1; 830 } 831 832 len = node->info_linear->info.jited_prog_len; 833 buf = (u8 *)(uintptr_t)node->info_linear->info.jited_prog_insns; 834 835 if (offset >= len) 836 return -1; 837 838 size = (ssize_t)min(len - offset, (u64)size); 839 memcpy(data, buf + offset, size); 840 return size; 841} 842 843static int bpf_size(struct dso *dso) 844{ 845 struct bpf_prog_info_node *node; 846 struct dso_bpf_prog *dso_bpf_prog = dso__bpf_prog(dso); 847 848 node = perf_env__find_bpf_prog_info(dso_bpf_prog->env, dso_bpf_prog->id); 849 if (!node || !node->info_linear) { 850 dso__data(dso)->status = DSO_DATA_STATUS_ERROR; 851 return -1; 852 } 853 854 dso__data(dso)->file_size = node->info_linear->info.jited_prog_len; 855 return 0; 856} 857#endif // HAVE_LIBBPF_SUPPORT 858 859static void 860dso_cache__free(struct dso *dso) 861{ 862 struct rb_root *root = &dso__data(dso)->cache; 863 struct rb_node *next = rb_first(root); 864 865 mutex_lock(dso__lock(dso)); 866 while (next) { 867 struct dso_cache *cache; 868 869 cache = rb_entry(next, struct dso_cache, rb_node); 870 next = rb_next(&cache->rb_node); 871 rb_erase(&cache->rb_node, root); 872 free(cache); 873 } 874 mutex_unlock(dso__lock(dso)); 875} 876 877static struct dso_cache *__dso_cache__find(struct dso *dso, u64 offset) 878{ 879 const struct rb_root *root = &dso__data(dso)->cache; 880 struct rb_node * const *p = &root->rb_node; 881 const struct rb_node *parent = NULL; 882 struct dso_cache *cache; 883 884 while (*p != NULL) { 885 u64 end; 886 887 parent = *p; 888 cache = rb_entry(parent, struct dso_cache, rb_node); 889 end = cache->offset + DSO__DATA_CACHE_SIZE; 890 891 if (offset < cache->offset) 892 p = &(*p)->rb_left; 893 else if (offset >= end) 894 p = &(*p)->rb_right; 895 else 896 return cache; 897 } 898 899 return NULL; 900} 901 902static struct dso_cache * 903dso_cache__insert(struct dso *dso, struct dso_cache *new) 904{ 905 struct rb_root *root = &dso__data(dso)->cache; 906 struct rb_node **p = &root->rb_node; 907 struct rb_node *parent = NULL; 908 struct dso_cache *cache; 909 u64 offset = new->offset; 910 911 mutex_lock(dso__lock(dso)); 912 while (*p != NULL) { 913 u64 end; 914 915 parent = *p; 916 cache = rb_entry(parent, struct dso_cache, rb_node); 917 end = cache->offset + DSO__DATA_CACHE_SIZE; 918 919 if (offset < cache->offset) 920 p = &(*p)->rb_left; 921 else if (offset >= end) 922 p = &(*p)->rb_right; 923 else 924 goto out; 925 } 926 927 rb_link_node(&new->rb_node, parent, p); 928 rb_insert_color(&new->rb_node, root); 929 930 cache = NULL; 931out: 932 mutex_unlock(dso__lock(dso)); 933 return cache; 934} 935 936static ssize_t dso_cache__memcpy(struct dso_cache *cache, u64 offset, u8 *data, 937 u64 size, bool out) 938{ 939 u64 cache_offset = offset - cache->offset; 940 u64 cache_size = min(cache->size - cache_offset, size); 941 942 if (out) 943 memcpy(data, cache->data + cache_offset, cache_size); 944 else 945 memcpy(cache->data + cache_offset, data, cache_size); 946 return cache_size; 947} 948 949static ssize_t file_read(struct dso *dso, struct machine *machine, 950 u64 offset, char *data) 951{ 952 ssize_t ret; 953 954 pthread_mutex_lock(&dso__data_open_lock); 955 956 /* 957 * dso__data(dso)->fd might be closed if other thread opened another 958 * file (dso) due to open file limit (RLIMIT_NOFILE). 959 */ 960 try_to_open_dso(dso, machine); 961 962 if (dso__data(dso)->fd < 0) { 963 dso__data(dso)->status = DSO_DATA_STATUS_ERROR; 964 ret = -errno; 965 goto out; 966 } 967 968 ret = pread(dso__data(dso)->fd, data, DSO__DATA_CACHE_SIZE, offset); 969out: 970 pthread_mutex_unlock(&dso__data_open_lock); 971 return ret; 972} 973 974static struct dso_cache *dso_cache__populate(struct dso *dso, 975 struct machine *machine, 976 u64 offset, ssize_t *ret) 977{ 978 u64 cache_offset = offset & DSO__DATA_CACHE_MASK; 979 struct dso_cache *cache; 980 struct dso_cache *old; 981 982 cache = zalloc(sizeof(*cache) + DSO__DATA_CACHE_SIZE); 983 if (!cache) { 984 *ret = -ENOMEM; 985 return NULL; 986 } 987#ifdef HAVE_LIBBPF_SUPPORT 988 if (dso__binary_type(dso) == DSO_BINARY_TYPE__BPF_PROG_INFO) 989 *ret = bpf_read(dso, cache_offset, cache->data); 990 else 991#endif 992 if (dso__binary_type(dso) == DSO_BINARY_TYPE__OOL) 993 *ret = DSO__DATA_CACHE_SIZE; 994 else 995 *ret = file_read(dso, machine, cache_offset, cache->data); 996 997 if (*ret <= 0) { 998 free(cache); 999 return NULL; 1000 } 1001 1002 cache->offset = cache_offset; 1003 cache->size = *ret; 1004 1005 old = dso_cache__insert(dso, cache); 1006 if (old) { 1007 /* we lose the race */ 1008 free(cache); 1009 cache = old; 1010 } 1011 1012 return cache; 1013} 1014 1015static struct dso_cache *dso_cache__find(struct dso *dso, 1016 struct machine *machine, 1017 u64 offset, 1018 ssize_t *ret) 1019{ 1020 struct dso_cache *cache = __dso_cache__find(dso, offset); 1021 1022 return cache ? cache : dso_cache__populate(dso, machine, offset, ret); 1023} 1024 1025static ssize_t dso_cache_io(struct dso *dso, struct machine *machine, 1026 u64 offset, u8 *data, ssize_t size, bool out) 1027{ 1028 struct dso_cache *cache; 1029 ssize_t ret = 0; 1030 1031 cache = dso_cache__find(dso, machine, offset, &ret); 1032 if (!cache) 1033 return ret; 1034 1035 return dso_cache__memcpy(cache, offset, data, size, out); 1036} 1037 1038/* 1039 * Reads and caches dso data DSO__DATA_CACHE_SIZE size chunks 1040 * in the rb_tree. Any read to already cached data is served 1041 * by cached data. Writes update the cache only, not the backing file. 1042 */ 1043static ssize_t cached_io(struct dso *dso, struct machine *machine, 1044 u64 offset, u8 *data, ssize_t size, bool out) 1045{ 1046 ssize_t r = 0; 1047 u8 *p = data; 1048 1049 do { 1050 ssize_t ret; 1051 1052 ret = dso_cache_io(dso, machine, offset, p, size, out); 1053 if (ret < 0) 1054 return ret; 1055 1056 /* Reached EOF, return what we have. */ 1057 if (!ret) 1058 break; 1059 1060 BUG_ON(ret > size); 1061 1062 r += ret; 1063 p += ret; 1064 offset += ret; 1065 size -= ret; 1066 1067 } while (size); 1068 1069 return r; 1070} 1071 1072static int file_size(struct dso *dso, struct machine *machine) 1073{ 1074 int ret = 0; 1075 struct stat st; 1076 char sbuf[STRERR_BUFSIZE]; 1077 1078 pthread_mutex_lock(&dso__data_open_lock); 1079 1080 /* 1081 * dso__data(dso)->fd might be closed if other thread opened another 1082 * file (dso) due to open file limit (RLIMIT_NOFILE). 1083 */ 1084 try_to_open_dso(dso, machine); 1085 1086 if (dso__data(dso)->fd < 0) { 1087 ret = -errno; 1088 dso__data(dso)->status = DSO_DATA_STATUS_ERROR; 1089 goto out; 1090 } 1091 1092 if (fstat(dso__data(dso)->fd, &st) < 0) { 1093 ret = -errno; 1094 pr_err("dso cache fstat failed: %s\n", 1095 str_error_r(errno, sbuf, sizeof(sbuf))); 1096 dso__data(dso)->status = DSO_DATA_STATUS_ERROR; 1097 goto out; 1098 } 1099 dso__data(dso)->file_size = st.st_size; 1100 1101out: 1102 pthread_mutex_unlock(&dso__data_open_lock); 1103 return ret; 1104} 1105 1106int dso__data_file_size(struct dso *dso, struct machine *machine) 1107{ 1108 if (dso__data(dso)->file_size) 1109 return 0; 1110 1111 if (dso__data(dso)->status == DSO_DATA_STATUS_ERROR) 1112 return -1; 1113#ifdef HAVE_LIBBPF_SUPPORT 1114 if (dso__binary_type(dso) == DSO_BINARY_TYPE__BPF_PROG_INFO) 1115 return bpf_size(dso); 1116#endif 1117 return file_size(dso, machine); 1118} 1119 1120/** 1121 * dso__data_size - Return dso data size 1122 * @dso: dso object 1123 * @machine: machine object 1124 * 1125 * Return: dso data size 1126 */ 1127off_t dso__data_size(struct dso *dso, struct machine *machine) 1128{ 1129 if (dso__data_file_size(dso, machine)) 1130 return -1; 1131 1132 /* For now just estimate dso data size is close to file size */ 1133 return dso__data(dso)->file_size; 1134} 1135 1136static ssize_t data_read_write_offset(struct dso *dso, struct machine *machine, 1137 u64 offset, u8 *data, ssize_t size, 1138 bool out) 1139{ 1140 if (dso__data_file_size(dso, machine)) 1141 return -1; 1142 1143 /* Check the offset sanity. */ 1144 if (offset > dso__data(dso)->file_size) 1145 return -1; 1146 1147 if (offset + size < offset) 1148 return -1; 1149 1150 return cached_io(dso, machine, offset, data, size, out); 1151} 1152 1153/** 1154 * dso__data_read_offset - Read data from dso file offset 1155 * @dso: dso object 1156 * @machine: machine object 1157 * @offset: file offset 1158 * @data: buffer to store data 1159 * @size: size of the @data buffer 1160 * 1161 * External interface to read data from dso file offset. Open 1162 * dso data file and use cached_read to get the data. 1163 */ 1164ssize_t dso__data_read_offset(struct dso *dso, struct machine *machine, 1165 u64 offset, u8 *data, ssize_t size) 1166{ 1167 if (dso__data(dso)->status == DSO_DATA_STATUS_ERROR) 1168 return -1; 1169 1170 return data_read_write_offset(dso, machine, offset, data, size, true); 1171} 1172 1173/** 1174 * dso__data_read_addr - Read data from dso address 1175 * @dso: dso object 1176 * @machine: machine object 1177 * @add: virtual memory address 1178 * @data: buffer to store data 1179 * @size: size of the @data buffer 1180 * 1181 * External interface to read data from dso address. 1182 */ 1183ssize_t dso__data_read_addr(struct dso *dso, struct map *map, 1184 struct machine *machine, u64 addr, 1185 u8 *data, ssize_t size) 1186{ 1187 u64 offset = map__map_ip(map, addr); 1188 1189 return dso__data_read_offset(dso, machine, offset, data, size); 1190} 1191 1192/** 1193 * dso__data_write_cache_offs - Write data to dso data cache at file offset 1194 * @dso: dso object 1195 * @machine: machine object 1196 * @offset: file offset 1197 * @data: buffer to write 1198 * @size: size of the @data buffer 1199 * 1200 * Write into the dso file data cache, but do not change the file itself. 1201 */ 1202ssize_t dso__data_write_cache_offs(struct dso *dso, struct machine *machine, 1203 u64 offset, const u8 *data_in, ssize_t size) 1204{ 1205 u8 *data = (u8 *)data_in; /* cast away const to use same fns for r/w */ 1206 1207 if (dso__data(dso)->status == DSO_DATA_STATUS_ERROR) 1208 return -1; 1209 1210 return data_read_write_offset(dso, machine, offset, data, size, false); 1211} 1212 1213/** 1214 * dso__data_write_cache_addr - Write data to dso data cache at dso address 1215 * @dso: dso object 1216 * @machine: machine object 1217 * @add: virtual memory address 1218 * @data: buffer to write 1219 * @size: size of the @data buffer 1220 * 1221 * External interface to write into the dso file data cache, but do not change 1222 * the file itself. 1223 */ 1224ssize_t dso__data_write_cache_addr(struct dso *dso, struct map *map, 1225 struct machine *machine, u64 addr, 1226 const u8 *data, ssize_t size) 1227{ 1228 u64 offset = map__map_ip(map, addr); 1229 1230 return dso__data_write_cache_offs(dso, machine, offset, data, size); 1231} 1232 1233struct map *dso__new_map(const char *name) 1234{ 1235 struct map *map = NULL; 1236 struct dso *dso = dso__new(name); 1237 1238 if (dso) { 1239 map = map__new2(0, dso); 1240 dso__put(dso); 1241 } 1242 1243 return map; 1244} 1245 1246struct dso *machine__findnew_kernel(struct machine *machine, const char *name, 1247 const char *short_name, int dso_type) 1248{ 1249 /* 1250 * The kernel dso could be created by build_id processing. 1251 */ 1252 struct dso *dso = machine__findnew_dso(machine, name); 1253 1254 /* 1255 * We need to run this in all cases, since during the build_id 1256 * processing we had no idea this was the kernel dso. 1257 */ 1258 if (dso != NULL) { 1259 dso__set_short_name(dso, short_name, false); 1260 dso__set_kernel(dso, dso_type); 1261 } 1262 1263 return dso; 1264} 1265 1266static void dso__set_long_name_id(struct dso *dso, const char *name, bool name_allocated) 1267{ 1268 struct dsos *dsos = dso__dsos(dso); 1269 1270 if (name == NULL) 1271 return; 1272 1273 if (dsos) { 1274 /* 1275 * Need to avoid re-sorting the dsos breaking by non-atomically 1276 * renaming the dso. 1277 */ 1278 down_write(&dsos->lock); 1279 } 1280 1281 if (dso__long_name_allocated(dso)) 1282 free((char *)dso__long_name(dso)); 1283 1284 RC_CHK_ACCESS(dso)->long_name = name; 1285 RC_CHK_ACCESS(dso)->long_name_len = strlen(name); 1286 dso__set_long_name_allocated(dso, name_allocated); 1287 1288 if (dsos) { 1289 dsos->sorted = false; 1290 up_write(&dsos->lock); 1291 } 1292} 1293 1294static int __dso_id__cmp(const struct dso_id *a, const struct dso_id *b) 1295{ 1296 if (a->maj > b->maj) return -1; 1297 if (a->maj < b->maj) return 1; 1298 1299 if (a->min > b->min) return -1; 1300 if (a->min < b->min) return 1; 1301 1302 if (a->ino > b->ino) return -1; 1303 if (a->ino < b->ino) return 1; 1304 1305 /* 1306 * Synthesized MMAP events have zero ino_generation, avoid comparing 1307 * them with MMAP events with actual ino_generation. 1308 * 1309 * I found it harmful because the mismatch resulted in a new 1310 * dso that did not have a build ID whereas the original dso did have a 1311 * build ID. The build ID was essential because the object was not found 1312 * otherwise. - Adrian 1313 */ 1314 if (a->ino_generation && b->ino_generation) { 1315 if (a->ino_generation > b->ino_generation) return -1; 1316 if (a->ino_generation < b->ino_generation) return 1; 1317 } 1318 1319 return 0; 1320} 1321 1322bool dso_id__empty(const struct dso_id *id) 1323{ 1324 if (!id) 1325 return true; 1326 1327 return !id->maj && !id->min && !id->ino && !id->ino_generation; 1328} 1329 1330void __dso__inject_id(struct dso *dso, const struct dso_id *id) 1331{ 1332 struct dsos *dsos = dso__dsos(dso); 1333 struct dso_id *dso_id = dso__id(dso); 1334 1335 /* dsos write lock held by caller. */ 1336 1337 dso_id->maj = id->maj; 1338 dso_id->min = id->min; 1339 dso_id->ino = id->ino; 1340 dso_id->ino_generation = id->ino_generation; 1341 1342 if (dsos) 1343 dsos->sorted = false; 1344} 1345 1346int dso_id__cmp(const struct dso_id *a, const struct dso_id *b) 1347{ 1348 /* 1349 * The second is always dso->id, so zeroes if not set, assume passing 1350 * NULL for a means a zeroed id 1351 */ 1352 if (dso_id__empty(a) || dso_id__empty(b)) 1353 return 0; 1354 1355 return __dso_id__cmp(a, b); 1356} 1357 1358int dso__cmp_id(struct dso *a, struct dso *b) 1359{ 1360 return __dso_id__cmp(dso__id(a), dso__id(b)); 1361} 1362 1363void dso__set_long_name(struct dso *dso, const char *name, bool name_allocated) 1364{ 1365 dso__set_long_name_id(dso, name, name_allocated); 1366} 1367 1368void dso__set_short_name(struct dso *dso, const char *name, bool name_allocated) 1369{ 1370 struct dsos *dsos = dso__dsos(dso); 1371 1372 if (name == NULL) 1373 return; 1374 1375 if (dsos) { 1376 /* 1377 * Need to avoid re-sorting the dsos breaking by non-atomically 1378 * renaming the dso. 1379 */ 1380 down_write(&dsos->lock); 1381 } 1382 if (dso__short_name_allocated(dso)) 1383 free((char *)dso__short_name(dso)); 1384 1385 RC_CHK_ACCESS(dso)->short_name = name; 1386 RC_CHK_ACCESS(dso)->short_name_len = strlen(name); 1387 dso__set_short_name_allocated(dso, name_allocated); 1388 1389 if (dsos) { 1390 dsos->sorted = false; 1391 up_write(&dsos->lock); 1392 } 1393} 1394 1395int dso__name_len(const struct dso *dso) 1396{ 1397 if (!dso) 1398 return strlen("[unknown]"); 1399 if (verbose > 0) 1400 return dso__long_name_len(dso); 1401 1402 return dso__short_name_len(dso); 1403} 1404 1405bool dso__loaded(const struct dso *dso) 1406{ 1407 return RC_CHK_ACCESS(dso)->loaded; 1408} 1409 1410bool dso__sorted_by_name(const struct dso *dso) 1411{ 1412 return RC_CHK_ACCESS(dso)->sorted_by_name; 1413} 1414 1415void dso__set_sorted_by_name(struct dso *dso) 1416{ 1417 RC_CHK_ACCESS(dso)->sorted_by_name = true; 1418} 1419 1420struct dso *dso__new_id(const char *name, const struct dso_id *id) 1421{ 1422 RC_STRUCT(dso) *dso = zalloc(sizeof(*dso) + strlen(name) + 1); 1423 struct dso *res; 1424 struct dso_data *data; 1425 1426 if (!dso) 1427 return NULL; 1428 1429 if (ADD_RC_CHK(res, dso)) { 1430 strcpy(dso->name, name); 1431 if (id) 1432 dso->id = *id; 1433 dso__set_long_name_id(res, dso->name, false); 1434 dso__set_short_name(res, dso->name, false); 1435 dso->symbols = RB_ROOT_CACHED; 1436 dso->symbol_names = NULL; 1437 dso->symbol_names_len = 0; 1438 dso->inlined_nodes = RB_ROOT_CACHED; 1439 dso->srclines = RB_ROOT_CACHED; 1440 dso->data_types = RB_ROOT; 1441 dso->global_vars = RB_ROOT; 1442 dso->data.fd = -1; 1443 dso->data.status = DSO_DATA_STATUS_UNKNOWN; 1444 dso->symtab_type = DSO_BINARY_TYPE__NOT_FOUND; 1445 dso->binary_type = DSO_BINARY_TYPE__NOT_FOUND; 1446 dso->is_64_bit = (sizeof(void *) == 8); 1447 dso->loaded = 0; 1448 dso->rel = 0; 1449 dso->sorted_by_name = 0; 1450 dso->has_build_id = 0; 1451 dso->has_srcline = 1; 1452 dso->a2l_fails = 1; 1453 dso->kernel = DSO_SPACE__USER; 1454 dso->is_kmod = 0; 1455 dso->needs_swap = DSO_SWAP__UNSET; 1456 dso->comp = COMP_ID__NONE; 1457 mutex_init(&dso->lock); 1458 refcount_set(&dso->refcnt, 1); 1459 data = &dso->data; 1460 data->cache = RB_ROOT; 1461 data->fd = -1; 1462 data->status = DSO_DATA_STATUS_UNKNOWN; 1463 INIT_LIST_HEAD(&data->open_entry); 1464#ifdef REFCNT_CHECKING 1465 data->dso = NULL; /* Set when on the open_entry list. */ 1466#endif 1467 } 1468 return res; 1469} 1470 1471struct dso *dso__new(const char *name) 1472{ 1473 return dso__new_id(name, NULL); 1474} 1475 1476void dso__delete(struct dso *dso) 1477{ 1478 if (dso__dsos(dso)) 1479 pr_err("DSO %s is still in rbtree when being deleted!\n", dso__long_name(dso)); 1480 1481 /* free inlines first, as they reference symbols */ 1482 inlines__tree_delete(&RC_CHK_ACCESS(dso)->inlined_nodes); 1483 srcline__tree_delete(&RC_CHK_ACCESS(dso)->srclines); 1484 symbols__delete(&RC_CHK_ACCESS(dso)->symbols); 1485 RC_CHK_ACCESS(dso)->symbol_names_len = 0; 1486 zfree(&RC_CHK_ACCESS(dso)->symbol_names); 1487 annotated_data_type__tree_delete(dso__data_types(dso)); 1488 global_var_type__tree_delete(dso__global_vars(dso)); 1489 1490 if (RC_CHK_ACCESS(dso)->short_name_allocated) { 1491 zfree((char **)&RC_CHK_ACCESS(dso)->short_name); 1492 RC_CHK_ACCESS(dso)->short_name_allocated = false; 1493 } 1494 1495 if (RC_CHK_ACCESS(dso)->long_name_allocated) { 1496 zfree((char **)&RC_CHK_ACCESS(dso)->long_name); 1497 RC_CHK_ACCESS(dso)->long_name_allocated = false; 1498 } 1499 1500 dso__data_close(dso); 1501 auxtrace_cache__free(RC_CHK_ACCESS(dso)->auxtrace_cache); 1502 dso_cache__free(dso); 1503 dso__free_a2l(dso); 1504 dso__free_symsrc_filename(dso); 1505 nsinfo__zput(RC_CHK_ACCESS(dso)->nsinfo); 1506 mutex_destroy(dso__lock(dso)); 1507 RC_CHK_FREE(dso); 1508} 1509 1510struct dso *dso__get(struct dso *dso) 1511{ 1512 struct dso *result; 1513 1514 if (RC_CHK_GET(result, dso)) 1515 refcount_inc(&RC_CHK_ACCESS(dso)->refcnt); 1516 1517 return result; 1518} 1519 1520void dso__put(struct dso *dso) 1521{ 1522 if (dso && refcount_dec_and_test(&RC_CHK_ACCESS(dso)->refcnt)) 1523 dso__delete(dso); 1524 else 1525 RC_CHK_PUT(dso); 1526} 1527 1528void dso__set_build_id(struct dso *dso, struct build_id *bid) 1529{ 1530 RC_CHK_ACCESS(dso)->bid = *bid; 1531 RC_CHK_ACCESS(dso)->has_build_id = 1; 1532} 1533 1534bool dso__build_id_equal(const struct dso *dso, struct build_id *bid) 1535{ 1536 const struct build_id *dso_bid = dso__bid_const(dso); 1537 1538 if (dso_bid->size > bid->size && dso_bid->size == BUILD_ID_SIZE) { 1539 /* 1540 * For the backward compatibility, it allows a build-id has 1541 * trailing zeros. 1542 */ 1543 return !memcmp(dso_bid->data, bid->data, bid->size) && 1544 !memchr_inv(&dso_bid->data[bid->size], 0, 1545 dso_bid->size - bid->size); 1546 } 1547 1548 return dso_bid->size == bid->size && 1549 memcmp(dso_bid->data, bid->data, dso_bid->size) == 0; 1550} 1551 1552void dso__read_running_kernel_build_id(struct dso *dso, struct machine *machine) 1553{ 1554 char path[PATH_MAX]; 1555 1556 if (machine__is_default_guest(machine)) 1557 return; 1558 sprintf(path, "%s/sys/kernel/notes", machine->root_dir); 1559 if (sysfs__read_build_id(path, dso__bid(dso)) == 0) 1560 dso__set_has_build_id(dso); 1561} 1562 1563int dso__kernel_module_get_build_id(struct dso *dso, 1564 const char *root_dir) 1565{ 1566 char filename[PATH_MAX]; 1567 /* 1568 * kernel module short names are of the form "[module]" and 1569 * we need just "module" here. 1570 */ 1571 const char *name = dso__short_name(dso) + 1; 1572 1573 snprintf(filename, sizeof(filename), 1574 "%s/sys/module/%.*s/notes/.note.gnu.build-id", 1575 root_dir, (int)strlen(name) - 1, name); 1576 1577 if (sysfs__read_build_id(filename, dso__bid(dso)) == 0) 1578 dso__set_has_build_id(dso); 1579 1580 return 0; 1581} 1582 1583static size_t dso__fprintf_buildid(struct dso *dso, FILE *fp) 1584{ 1585 char sbuild_id[SBUILD_ID_SIZE]; 1586 1587 build_id__sprintf(dso__bid(dso), sbuild_id); 1588 return fprintf(fp, "%s", sbuild_id); 1589} 1590 1591size_t dso__fprintf(struct dso *dso, FILE *fp) 1592{ 1593 struct rb_node *nd; 1594 size_t ret = fprintf(fp, "dso: %s (", dso__short_name(dso)); 1595 1596 if (dso__short_name(dso) != dso__long_name(dso)) 1597 ret += fprintf(fp, "%s, ", dso__long_name(dso)); 1598 ret += fprintf(fp, "%sloaded, ", dso__loaded(dso) ? "" : "NOT "); 1599 ret += dso__fprintf_buildid(dso, fp); 1600 ret += fprintf(fp, ")\n"); 1601 for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) { 1602 struct symbol *pos = rb_entry(nd, struct symbol, rb_node); 1603 ret += symbol__fprintf(pos, fp); 1604 } 1605 1606 return ret; 1607} 1608 1609enum dso_type dso__type(struct dso *dso, struct machine *machine) 1610{ 1611 int fd; 1612 enum dso_type type = DSO__TYPE_UNKNOWN; 1613 1614 fd = dso__data_get_fd(dso, machine); 1615 if (fd >= 0) { 1616 type = dso__type_fd(fd); 1617 dso__data_put_fd(dso); 1618 } 1619 1620 return type; 1621} 1622 1623int dso__strerror_load(struct dso *dso, char *buf, size_t buflen) 1624{ 1625 int idx, errnum = *dso__load_errno(dso); 1626 /* 1627 * This must have a same ordering as the enum dso_load_errno. 1628 */ 1629 static const char *dso_load__error_str[] = { 1630 "Internal tools/perf/ library error", 1631 "Invalid ELF file", 1632 "Can not read build id", 1633 "Mismatching build id", 1634 "Decompression failure", 1635 }; 1636 1637 BUG_ON(buflen == 0); 1638 1639 if (errnum >= 0) { 1640 const char *err = str_error_r(errnum, buf, buflen); 1641 1642 if (err != buf) 1643 scnprintf(buf, buflen, "%s", err); 1644 1645 return 0; 1646 } 1647 1648 if (errnum < __DSO_LOAD_ERRNO__START || errnum >= __DSO_LOAD_ERRNO__END) 1649 return -1; 1650 1651 idx = errnum - __DSO_LOAD_ERRNO__START; 1652 scnprintf(buf, buflen, "%s", dso_load__error_str[idx]); 1653 return 0; 1654} 1655 1656bool perf_pid_map_tid(const char *dso_name, int *tid) 1657{ 1658 return sscanf(dso_name, "/tmp/perf-%d.map", tid) == 1; 1659} 1660 1661bool is_perf_pid_map_name(const char *dso_name) 1662{ 1663 int tid; 1664 1665 return perf_pid_map_tid(dso_name, &tid); 1666}