Linux Audio

Check our new training course

Loading...
v6.13.7
   1// SPDX-License-Identifier: GPL-2.0
   2#include <asm/bug.h>
   3#include <linux/kernel.h>
   4#include <linux/string.h>
   5#include <linux/zalloc.h>
   6#include <sys/time.h>
   7#include <sys/resource.h>
   8#include <sys/types.h>
   9#include <sys/stat.h>
  10#include <unistd.h>
  11#include <errno.h>
  12#include <fcntl.h>
  13#include <stdlib.h>
  14#ifdef HAVE_LIBBPF_SUPPORT
  15#include <bpf/libbpf.h>
  16#include "bpf-event.h"
  17#include "bpf-utils.h"
  18#endif
  19#include "compress.h"
  20#include "env.h"
  21#include "namespaces.h"
  22#include "path.h"
  23#include "map.h"
  24#include "symbol.h"
  25#include "srcline.h"
  26#include "dso.h"
  27#include "dsos.h"
  28#include "machine.h"
  29#include "auxtrace.h"
  30#include "util.h" /* O_CLOEXEC for older systems */
  31#include "debug.h"
  32#include "string2.h"
  33#include "vdso.h"
  34#include "annotate-data.h"
  35
  36static const char * const debuglink_paths[] = {
  37	"%.0s%s",
  38	"%s/%s",
  39	"%s/.debug/%s",
  40	"/usr/lib/debug%s/%s"
  41};
  42
  43void dso__set_nsinfo(struct dso *dso, struct nsinfo *nsi)
  44{
  45	nsinfo__put(RC_CHK_ACCESS(dso)->nsinfo);
  46	RC_CHK_ACCESS(dso)->nsinfo = nsi;
  47}
  48
  49char dso__symtab_origin(const struct dso *dso)
  50{
  51	static const char origin[] = {
  52		[DSO_BINARY_TYPE__KALLSYMS]			= 'k',
  53		[DSO_BINARY_TYPE__VMLINUX]			= 'v',
  54		[DSO_BINARY_TYPE__JAVA_JIT]			= 'j',
  55		[DSO_BINARY_TYPE__DEBUGLINK]			= 'l',
  56		[DSO_BINARY_TYPE__BUILD_ID_CACHE]		= 'B',
  57		[DSO_BINARY_TYPE__BUILD_ID_CACHE_DEBUGINFO]	= 'D',
  58		[DSO_BINARY_TYPE__FEDORA_DEBUGINFO]		= 'f',
  59		[DSO_BINARY_TYPE__UBUNTU_DEBUGINFO]		= 'u',
  60		[DSO_BINARY_TYPE__MIXEDUP_UBUNTU_DEBUGINFO]	= 'x',
  61		[DSO_BINARY_TYPE__OPENEMBEDDED_DEBUGINFO]	= 'o',
  62		[DSO_BINARY_TYPE__BUILDID_DEBUGINFO]		= 'b',
  63		[DSO_BINARY_TYPE__SYSTEM_PATH_DSO]		= 'd',
  64		[DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE]		= 'K',
  65		[DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP]	= 'm',
  66		[DSO_BINARY_TYPE__GUEST_KALLSYMS]		= 'g',
  67		[DSO_BINARY_TYPE__GUEST_KMODULE]		= 'G',
  68		[DSO_BINARY_TYPE__GUEST_KMODULE_COMP]		= 'M',
  69		[DSO_BINARY_TYPE__GUEST_VMLINUX]		= 'V',
  70	};
  71
  72	if (dso == NULL || dso__symtab_type(dso) == DSO_BINARY_TYPE__NOT_FOUND)
  73		return '!';
  74	return origin[dso__symtab_type(dso)];
  75}
  76
  77bool dso__is_object_file(const struct dso *dso)
  78{
  79	switch (dso__binary_type(dso)) {
  80	case DSO_BINARY_TYPE__KALLSYMS:
  81	case DSO_BINARY_TYPE__GUEST_KALLSYMS:
  82	case DSO_BINARY_TYPE__JAVA_JIT:
  83	case DSO_BINARY_TYPE__BPF_PROG_INFO:
  84	case DSO_BINARY_TYPE__BPF_IMAGE:
  85	case DSO_BINARY_TYPE__OOL:
  86		return false;
  87	case DSO_BINARY_TYPE__VMLINUX:
  88	case DSO_BINARY_TYPE__GUEST_VMLINUX:
  89	case DSO_BINARY_TYPE__DEBUGLINK:
  90	case DSO_BINARY_TYPE__BUILD_ID_CACHE:
  91	case DSO_BINARY_TYPE__BUILD_ID_CACHE_DEBUGINFO:
  92	case DSO_BINARY_TYPE__FEDORA_DEBUGINFO:
  93	case DSO_BINARY_TYPE__UBUNTU_DEBUGINFO:
  94	case DSO_BINARY_TYPE__MIXEDUP_UBUNTU_DEBUGINFO:
  95	case DSO_BINARY_TYPE__BUILDID_DEBUGINFO:
  96	case DSO_BINARY_TYPE__SYSTEM_PATH_DSO:
  97	case DSO_BINARY_TYPE__GUEST_KMODULE:
  98	case DSO_BINARY_TYPE__GUEST_KMODULE_COMP:
  99	case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE:
 100	case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP:
 101	case DSO_BINARY_TYPE__KCORE:
 102	case DSO_BINARY_TYPE__GUEST_KCORE:
 103	case DSO_BINARY_TYPE__OPENEMBEDDED_DEBUGINFO:
 104	case DSO_BINARY_TYPE__NOT_FOUND:
 105	default:
 106		return true;
 107	}
 108}
 109
 110int dso__read_binary_type_filename(const struct dso *dso,
 111				   enum dso_binary_type type,
 112				   char *root_dir, char *filename, size_t size)
 113{
 114	char build_id_hex[SBUILD_ID_SIZE];
 115	int ret = 0;
 116	size_t len;
 117
 118	switch (type) {
 119	case DSO_BINARY_TYPE__DEBUGLINK:
 120	{
 121		const char *last_slash;
 122		char dso_dir[PATH_MAX];
 123		char symfile[PATH_MAX];
 124		unsigned int i;
 125
 126		len = __symbol__join_symfs(filename, size, dso__long_name(dso));
 127		last_slash = filename + len;
 128		while (last_slash != filename && *last_slash != '/')
 129			last_slash--;
 130
 131		strncpy(dso_dir, filename, last_slash - filename);
 132		dso_dir[last_slash-filename] = '\0';
 
 
 
 
 133
 134		if (!is_regular_file(filename)) {
 135			ret = -1;
 136			break;
 137		}
 138
 139		ret = filename__read_debuglink(filename, symfile, PATH_MAX);
 140		if (ret)
 141			break;
 142
 143		/* Check predefined locations where debug file might reside */
 144		ret = -1;
 145		for (i = 0; i < ARRAY_SIZE(debuglink_paths); i++) {
 146			snprintf(filename, size,
 147					debuglink_paths[i], dso_dir, symfile);
 148			if (is_regular_file(filename)) {
 149				ret = 0;
 150				break;
 151			}
 152		}
 153
 154		break;
 155	}
 156	case DSO_BINARY_TYPE__BUILD_ID_CACHE:
 157		if (dso__build_id_filename(dso, filename, size, false) == NULL)
 158			ret = -1;
 159		break;
 160
 161	case DSO_BINARY_TYPE__BUILD_ID_CACHE_DEBUGINFO:
 162		if (dso__build_id_filename(dso, filename, size, true) == NULL)
 163			ret = -1;
 164		break;
 165
 166	case DSO_BINARY_TYPE__FEDORA_DEBUGINFO:
 167		len = __symbol__join_symfs(filename, size, "/usr/lib/debug");
 168		snprintf(filename + len, size - len, "%s.debug", dso__long_name(dso));
 169		break;
 170
 171	case DSO_BINARY_TYPE__UBUNTU_DEBUGINFO:
 172		len = __symbol__join_symfs(filename, size, "/usr/lib/debug");
 173		snprintf(filename + len, size - len, "%s", dso__long_name(dso));
 174		break;
 175
 176	case DSO_BINARY_TYPE__MIXEDUP_UBUNTU_DEBUGINFO:
 177		/*
 178		 * Ubuntu can mixup /usr/lib with /lib, putting debuginfo in
 179		 * /usr/lib/debug/lib when it is expected to be in
 180		 * /usr/lib/debug/usr/lib
 181		 */
 182		if (strlen(dso__long_name(dso)) < 9 ||
 183		    strncmp(dso__long_name(dso), "/usr/lib/", 9)) {
 184			ret = -1;
 185			break;
 186		}
 187		len = __symbol__join_symfs(filename, size, "/usr/lib/debug");
 188		snprintf(filename + len, size - len, "%s", dso__long_name(dso) + 4);
 189		break;
 190
 191	case DSO_BINARY_TYPE__OPENEMBEDDED_DEBUGINFO:
 192	{
 193		const char *last_slash;
 194		size_t dir_size;
 195
 196		last_slash = dso__long_name(dso) + dso__long_name_len(dso);
 197		while (last_slash != dso__long_name(dso) && *last_slash != '/')
 198			last_slash--;
 199
 200		len = __symbol__join_symfs(filename, size, "");
 201		dir_size = last_slash - dso__long_name(dso) + 2;
 202		if (dir_size > (size - len)) {
 203			ret = -1;
 204			break;
 205		}
 206		len += scnprintf(filename + len, dir_size, "%s",  dso__long_name(dso));
 207		len += scnprintf(filename + len , size - len, ".debug%s",
 208								last_slash);
 209		break;
 210	}
 211
 212	case DSO_BINARY_TYPE__BUILDID_DEBUGINFO:
 213		if (!dso__has_build_id(dso)) {
 214			ret = -1;
 215			break;
 216		}
 217
 218		build_id__sprintf(dso__bid_const(dso), build_id_hex);
 
 
 219		len = __symbol__join_symfs(filename, size, "/usr/lib/debug/.build-id/");
 220		snprintf(filename + len, size - len, "%.2s/%s.debug",
 221			 build_id_hex, build_id_hex + 2);
 222		break;
 223
 224	case DSO_BINARY_TYPE__VMLINUX:
 225	case DSO_BINARY_TYPE__GUEST_VMLINUX:
 226	case DSO_BINARY_TYPE__SYSTEM_PATH_DSO:
 227		__symbol__join_symfs(filename, size, dso__long_name(dso));
 228		break;
 229
 230	case DSO_BINARY_TYPE__GUEST_KMODULE:
 231	case DSO_BINARY_TYPE__GUEST_KMODULE_COMP:
 232		path__join3(filename, size, symbol_conf.symfs,
 233			    root_dir, dso__long_name(dso));
 234		break;
 235
 236	case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE:
 237	case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP:
 238		__symbol__join_symfs(filename, size, dso__long_name(dso));
 239		break;
 240
 241	case DSO_BINARY_TYPE__KCORE:
 242	case DSO_BINARY_TYPE__GUEST_KCORE:
 243		snprintf(filename, size, "%s", dso__long_name(dso));
 244		break;
 245
 246	default:
 247	case DSO_BINARY_TYPE__KALLSYMS:
 248	case DSO_BINARY_TYPE__GUEST_KALLSYMS:
 249	case DSO_BINARY_TYPE__JAVA_JIT:
 250	case DSO_BINARY_TYPE__BPF_PROG_INFO:
 251	case DSO_BINARY_TYPE__BPF_IMAGE:
 252	case DSO_BINARY_TYPE__OOL:
 253	case DSO_BINARY_TYPE__NOT_FOUND:
 254		ret = -1;
 255		break;
 256	}
 257
 258	return ret;
 259}
 260
 261enum {
 262	COMP_ID__NONE = 0,
 263};
 264
 265static const struct {
 266	const char *fmt;
 267	int (*decompress)(const char *input, int output);
 268	bool (*is_compressed)(const char *input);
 269} compressions[] = {
 270	[COMP_ID__NONE] = { .fmt = NULL, },
 271#ifdef HAVE_ZLIB_SUPPORT
 272	{ "gz", gzip_decompress_to_file, gzip_is_compressed },
 273#endif
 274#ifdef HAVE_LZMA_SUPPORT
 275	{ "xz", lzma_decompress_to_file, lzma_is_compressed },
 276#endif
 277	{ NULL, NULL, NULL },
 278};
 279
 280static int is_supported_compression(const char *ext)
 281{
 282	unsigned i;
 283
 284	for (i = 1; compressions[i].fmt; i++) {
 285		if (!strcmp(ext, compressions[i].fmt))
 286			return i;
 287	}
 288	return COMP_ID__NONE;
 289}
 290
 291bool is_kernel_module(const char *pathname, int cpumode)
 292{
 293	struct kmod_path m;
 294	int mode = cpumode & PERF_RECORD_MISC_CPUMODE_MASK;
 295
 296	WARN_ONCE(mode != cpumode,
 297		  "Internal error: passing unmasked cpumode (%x) to is_kernel_module",
 298		  cpumode);
 299
 300	switch (mode) {
 301	case PERF_RECORD_MISC_USER:
 302	case PERF_RECORD_MISC_HYPERVISOR:
 303	case PERF_RECORD_MISC_GUEST_USER:
 304		return false;
 305	/* Treat PERF_RECORD_MISC_CPUMODE_UNKNOWN as kernel */
 306	default:
 307		if (kmod_path__parse(&m, pathname)) {
 308			pr_err("Failed to check whether %s is a kernel module or not. Assume it is.",
 309					pathname);
 310			return true;
 311		}
 312	}
 313
 314	return m.kmod;
 315}
 316
 317bool dso__needs_decompress(struct dso *dso)
 318{
 319	return dso__symtab_type(dso) == DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP ||
 320		dso__symtab_type(dso) == DSO_BINARY_TYPE__GUEST_KMODULE_COMP;
 321}
 322
 323int filename__decompress(const char *name, char *pathname,
 324			 size_t len, int comp, int *err)
 325{
 326	char tmpbuf[] = KMOD_DECOMP_NAME;
 327	int fd = -1;
 328
 329	/*
 330	 * We have proper compression id for DSO and yet the file
 331	 * behind the 'name' can still be plain uncompressed object.
 332	 *
 333	 * The reason is behind the logic we open the DSO object files,
 334	 * when we try all possible 'debug' objects until we find the
 335	 * data. So even if the DSO is represented by 'krava.xz' module,
 336	 * we can end up here opening ~/.debug/....23432432/debug' file
 337	 * which is not compressed.
 338	 *
 339	 * To keep this transparent, we detect this and return the file
 340	 * descriptor to the uncompressed file.
 341	 */
 342	if (!compressions[comp].is_compressed(name))
 343		return open(name, O_RDONLY);
 344
 345	fd = mkstemp(tmpbuf);
 346	if (fd < 0) {
 347		*err = errno;
 348		return -1;
 349	}
 350
 351	if (compressions[comp].decompress(name, fd)) {
 352		*err = DSO_LOAD_ERRNO__DECOMPRESSION_FAILURE;
 353		close(fd);
 354		fd = -1;
 355	}
 356
 357	if (!pathname || (fd < 0))
 358		unlink(tmpbuf);
 359
 360	if (pathname && (fd >= 0))
 361		strlcpy(pathname, tmpbuf, len);
 362
 363	return fd;
 364}
 365
 366static int decompress_kmodule(struct dso *dso, const char *name,
 367			      char *pathname, size_t len)
 368{
 369	if (!dso__needs_decompress(dso))
 370		return -1;
 371
 372	if (dso__comp(dso) == COMP_ID__NONE)
 373		return -1;
 374
 375	return filename__decompress(name, pathname, len, dso__comp(dso), dso__load_errno(dso));
 376}
 377
 378int dso__decompress_kmodule_fd(struct dso *dso, const char *name)
 379{
 380	return decompress_kmodule(dso, name, NULL, 0);
 381}
 382
 383int dso__decompress_kmodule_path(struct dso *dso, const char *name,
 384				 char *pathname, size_t len)
 385{
 386	int fd = decompress_kmodule(dso, name, pathname, len);
 387
 388	close(fd);
 389	return fd >= 0 ? 0 : -1;
 390}
 391
 392/*
 393 * Parses kernel module specified in @path and updates
 394 * @m argument like:
 395 *
 396 *    @comp - true if @path contains supported compression suffix,
 397 *            false otherwise
 398 *    @kmod - true if @path contains '.ko' suffix in right position,
 399 *            false otherwise
 400 *    @name - if (@alloc_name && @kmod) is true, it contains strdup-ed base name
 401 *            of the kernel module without suffixes, otherwise strudup-ed
 402 *            base name of @path
 403 *    @ext  - if (@alloc_ext && @comp) is true, it contains strdup-ed string
 404 *            the compression suffix
 405 *
 406 * Returns 0 if there's no strdup error, -ENOMEM otherwise.
 407 */
 408int __kmod_path__parse(struct kmod_path *m, const char *path,
 409		       bool alloc_name)
 410{
 411	const char *name = strrchr(path, '/');
 412	const char *ext  = strrchr(path, '.');
 413	bool is_simple_name = false;
 414
 415	memset(m, 0x0, sizeof(*m));
 416	name = name ? name + 1 : path;
 417
 418	/*
 419	 * '.' is also a valid character for module name. For example:
 420	 * [aaa.bbb] is a valid module name. '[' should have higher
 421	 * priority than '.ko' suffix.
 422	 *
 423	 * The kernel names are from machine__mmap_name. Such
 424	 * name should belong to kernel itself, not kernel module.
 425	 */
 426	if (name[0] == '[') {
 427		is_simple_name = true;
 428		if ((strncmp(name, "[kernel.kallsyms]", 17) == 0) ||
 429		    (strncmp(name, "[guest.kernel.kallsyms", 22) == 0) ||
 430		    (strncmp(name, "[vdso]", 6) == 0) ||
 431		    (strncmp(name, "[vdso32]", 8) == 0) ||
 432		    (strncmp(name, "[vdsox32]", 9) == 0) ||
 433		    (strncmp(name, "[vsyscall]", 10) == 0)) {
 434			m->kmod = false;
 435
 436		} else
 437			m->kmod = true;
 438	}
 439
 440	/* No extension, just return name. */
 441	if ((ext == NULL) || is_simple_name) {
 442		if (alloc_name) {
 443			m->name = strdup(name);
 444			return m->name ? 0 : -ENOMEM;
 445		}
 446		return 0;
 447	}
 448
 449	m->comp = is_supported_compression(ext + 1);
 450	if (m->comp > COMP_ID__NONE)
 451		ext -= 3;
 
 452
 453	/* Check .ko extension only if there's enough name left. */
 454	if (ext > name)
 455		m->kmod = !strncmp(ext, ".ko", 3);
 456
 457	if (alloc_name) {
 458		if (m->kmod) {
 459			if (asprintf(&m->name, "[%.*s]", (int) (ext - name), name) == -1)
 460				return -ENOMEM;
 461		} else {
 462			if (asprintf(&m->name, "%s", name) == -1)
 463				return -ENOMEM;
 464		}
 465
 466		strreplace(m->name, '-', '_');
 467	}
 468
 469	return 0;
 470}
 471
 472void dso__set_module_info(struct dso *dso, struct kmod_path *m,
 473			  struct machine *machine)
 474{
 475	if (machine__is_host(machine))
 476		dso__set_symtab_type(dso, DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE);
 477	else
 478		dso__set_symtab_type(dso, DSO_BINARY_TYPE__GUEST_KMODULE);
 479
 480	/* _KMODULE_COMP should be next to _KMODULE */
 481	if (m->kmod && m->comp) {
 482		dso__set_symtab_type(dso, dso__symtab_type(dso) + 1);
 483		dso__set_comp(dso, m->comp);
 484	}
 485
 486	dso__set_is_kmod(dso);
 487	dso__set_short_name(dso, strdup(m->name), true);
 488}
 489
 490/*
 491 * Global list of open DSOs and the counter.
 492 */
 493static LIST_HEAD(dso__data_open);
 494static long dso__data_open_cnt;
 495static pthread_mutex_t dso__data_open_lock = PTHREAD_MUTEX_INITIALIZER;
 496
 497static void dso__list_add(struct dso *dso)
 498{
 499	list_add_tail(&dso__data(dso)->open_entry, &dso__data_open);
 500#ifdef REFCNT_CHECKING
 501	dso__data(dso)->dso = dso__get(dso);
 502#endif
 503	/* Assume the dso is part of dsos, hence the optional reference count above. */
 504	assert(dso__dsos(dso));
 505	dso__data_open_cnt++;
 506}
 507
 508static void dso__list_del(struct dso *dso)
 509{
 510	list_del_init(&dso__data(dso)->open_entry);
 511#ifdef REFCNT_CHECKING
 512	dso__put(dso__data(dso)->dso);
 513#endif
 514	WARN_ONCE(dso__data_open_cnt <= 0,
 515		  "DSO data fd counter out of bounds.");
 516	dso__data_open_cnt--;
 517}
 518
 519static void close_first_dso(void);
 520
 521static int do_open(char *name)
 522{
 523	int fd;
 524	char sbuf[STRERR_BUFSIZE];
 525
 526	do {
 527		fd = open(name, O_RDONLY|O_CLOEXEC);
 528		if (fd >= 0)
 529			return fd;
 530
 531		pr_debug("dso open failed: %s\n",
 532			 str_error_r(errno, sbuf, sizeof(sbuf)));
 533		if (!dso__data_open_cnt || errno != EMFILE)
 534			break;
 535
 536		close_first_dso();
 537	} while (1);
 538
 539	return -1;
 540}
 541
 542char *dso__filename_with_chroot(const struct dso *dso, const char *filename)
 543{
 544	return filename_with_chroot(nsinfo__pid(dso__nsinfo_const(dso)), filename);
 545}
 546
 547static int __open_dso(struct dso *dso, struct machine *machine)
 548{
 549	int fd = -EINVAL;
 550	char *root_dir = (char *)"";
 551	char *name = malloc(PATH_MAX);
 552	bool decomp = false;
 553
 554	if (!name)
 555		return -ENOMEM;
 556
 557	mutex_lock(dso__lock(dso));
 558	if (machine)
 559		root_dir = machine->root_dir;
 560
 561	if (dso__read_binary_type_filename(dso, dso__binary_type(dso),
 562					    root_dir, name, PATH_MAX))
 563		goto out;
 564
 565	if (!is_regular_file(name)) {
 566		char *new_name;
 567
 568		if (errno != ENOENT || dso__nsinfo(dso) == NULL)
 569			goto out;
 570
 571		new_name = dso__filename_with_chroot(dso, name);
 572		if (!new_name)
 573			goto out;
 574
 575		free(name);
 576		name = new_name;
 577	}
 578
 579	if (dso__needs_decompress(dso)) {
 580		char newpath[KMOD_DECOMP_LEN];
 581		size_t len = sizeof(newpath);
 582
 583		if (dso__decompress_kmodule_path(dso, name, newpath, len) < 0) {
 584			fd = -(*dso__load_errno(dso));
 585			goto out;
 586		}
 587
 588		decomp = true;
 589		strcpy(name, newpath);
 590	}
 591
 592	fd = do_open(name);
 593
 594	if (decomp)
 595		unlink(name);
 596
 597out:
 598	mutex_unlock(dso__lock(dso));
 599	free(name);
 600	return fd;
 601}
 602
 603static void check_data_close(void);
 604
 605/**
 606 * dso_close - Open DSO data file
 607 * @dso: dso object
 608 *
 609 * Open @dso's data file descriptor and updates
 610 * list/count of open DSO objects.
 611 */
 612static int open_dso(struct dso *dso, struct machine *machine)
 613{
 614	int fd;
 615	struct nscookie nsc;
 616
 617	if (dso__binary_type(dso) != DSO_BINARY_TYPE__BUILD_ID_CACHE) {
 618		mutex_lock(dso__lock(dso));
 619		nsinfo__mountns_enter(dso__nsinfo(dso), &nsc);
 620		mutex_unlock(dso__lock(dso));
 621	}
 622	fd = __open_dso(dso, machine);
 623	if (dso__binary_type(dso) != DSO_BINARY_TYPE__BUILD_ID_CACHE)
 624		nsinfo__mountns_exit(&nsc);
 625
 626	if (fd >= 0) {
 627		dso__list_add(dso);
 628		/*
 629		 * Check if we crossed the allowed number
 630		 * of opened DSOs and close one if needed.
 631		 */
 632		check_data_close();
 633	}
 634
 635	return fd;
 636}
 637
 638static void close_data_fd(struct dso *dso)
 639{
 640	if (dso__data(dso)->fd >= 0) {
 641		close(dso__data(dso)->fd);
 642		dso__data(dso)->fd = -1;
 643		dso__data(dso)->file_size = 0;
 644		dso__list_del(dso);
 645	}
 646}
 647
 648/**
 649 * dso_close - Close DSO data file
 650 * @dso: dso object
 651 *
 652 * Close @dso's data file descriptor and updates
 653 * list/count of open DSO objects.
 654 */
 655static void close_dso(struct dso *dso)
 656{
 657	close_data_fd(dso);
 658}
 659
 660static void close_first_dso(void)
 661{
 662	struct dso_data *dso_data;
 663	struct dso *dso;
 664
 665	dso_data = list_first_entry(&dso__data_open, struct dso_data, open_entry);
 666#ifdef REFCNT_CHECKING
 667	dso = dso_data->dso;
 668#else
 669	dso = container_of(dso_data, struct dso, data);
 670#endif
 671	close_dso(dso);
 672}
 673
 674static rlim_t get_fd_limit(void)
 675{
 676	struct rlimit l;
 677	rlim_t limit = 0;
 678
 679	/* Allow half of the current open fd limit. */
 680	if (getrlimit(RLIMIT_NOFILE, &l) == 0) {
 681		if (l.rlim_cur == RLIM_INFINITY)
 682			limit = l.rlim_cur;
 683		else
 684			limit = l.rlim_cur / 2;
 685	} else {
 686		pr_err("failed to get fd limit\n");
 687		limit = 1;
 688	}
 689
 690	return limit;
 691}
 692
 693static rlim_t fd_limit;
 694
 695/*
 696 * Used only by tests/dso-data.c to reset the environment
 697 * for tests. I dont expect we should change this during
 698 * standard runtime.
 699 */
 700void reset_fd_limit(void)
 701{
 702	fd_limit = 0;
 703}
 704
 705static bool may_cache_fd(void)
 706{
 707	if (!fd_limit)
 708		fd_limit = get_fd_limit();
 
 
 709
 710	if (fd_limit == RLIM_INFINITY)
 711		return true;
 712
 713	return fd_limit > (rlim_t) dso__data_open_cnt;
 714}
 715
 716/*
 717 * Check and close LRU dso if we crossed allowed limit
 718 * for opened dso file descriptors. The limit is half
 719 * of the RLIMIT_NOFILE files opened.
 720*/
 721static void check_data_close(void)
 722{
 723	bool cache_fd = may_cache_fd();
 724
 725	if (!cache_fd)
 726		close_first_dso();
 727}
 728
 729/**
 730 * dso__data_close - Close DSO data file
 731 * @dso: dso object
 732 *
 733 * External interface to close @dso's data file descriptor.
 734 */
 735void dso__data_close(struct dso *dso)
 736{
 737	pthread_mutex_lock(&dso__data_open_lock);
 738	close_dso(dso);
 739	pthread_mutex_unlock(&dso__data_open_lock);
 740}
 741
 742static void try_to_open_dso(struct dso *dso, struct machine *machine)
 743{
 744	enum dso_binary_type binary_type_data[] = {
 745		DSO_BINARY_TYPE__BUILD_ID_CACHE,
 746		DSO_BINARY_TYPE__SYSTEM_PATH_DSO,
 747		DSO_BINARY_TYPE__NOT_FOUND,
 748	};
 749	int i = 0;
 750	struct dso_data *dso_data = dso__data(dso);
 751
 752	if (dso_data->fd >= 0)
 753		return;
 754
 755	if (dso__binary_type(dso) != DSO_BINARY_TYPE__NOT_FOUND) {
 756		dso_data->fd = open_dso(dso, machine);
 757		goto out;
 758	}
 759
 760	do {
 761		dso__set_binary_type(dso, binary_type_data[i++]);
 762
 763		dso_data->fd = open_dso(dso, machine);
 764		if (dso_data->fd >= 0)
 765			goto out;
 766
 767	} while (dso__binary_type(dso) != DSO_BINARY_TYPE__NOT_FOUND);
 768out:
 769	if (dso_data->fd >= 0)
 770		dso_data->status = DSO_DATA_STATUS_OK;
 771	else
 772		dso_data->status = DSO_DATA_STATUS_ERROR;
 773}
 774
 775/**
 776 * dso__data_get_fd - Get dso's data file descriptor
 777 * @dso: dso object
 778 * @machine: machine object
 779 *
 780 * External interface to find dso's file, open it and
 781 * returns file descriptor.  It should be paired with
 782 * dso__data_put_fd() if it returns non-negative value.
 783 */
 784int dso__data_get_fd(struct dso *dso, struct machine *machine)
 785{
 786	if (dso__data(dso)->status == DSO_DATA_STATUS_ERROR)
 787		return -1;
 788
 789	if (pthread_mutex_lock(&dso__data_open_lock) < 0)
 790		return -1;
 791
 792	try_to_open_dso(dso, machine);
 793
 794	if (dso__data(dso)->fd < 0)
 795		pthread_mutex_unlock(&dso__data_open_lock);
 796
 797	return dso__data(dso)->fd;
 798}
 799
 800void dso__data_put_fd(struct dso *dso __maybe_unused)
 801{
 802	pthread_mutex_unlock(&dso__data_open_lock);
 803}
 804
 805bool dso__data_status_seen(struct dso *dso, enum dso_data_status_seen by)
 806{
 807	u32 flag = 1 << by;
 808
 809	if (dso__data(dso)->status_seen & flag)
 810		return true;
 811
 812	dso__data(dso)->status_seen |= flag;
 813
 814	return false;
 815}
 816
 817#ifdef HAVE_LIBBPF_SUPPORT
 818static ssize_t bpf_read(struct dso *dso, u64 offset, char *data)
 819{
 820	struct bpf_prog_info_node *node;
 821	ssize_t size = DSO__DATA_CACHE_SIZE;
 822	struct dso_bpf_prog *dso_bpf_prog = dso__bpf_prog(dso);
 823	u64 len;
 824	u8 *buf;
 825
 826	node = perf_env__find_bpf_prog_info(dso_bpf_prog->env, dso_bpf_prog->id);
 827	if (!node || !node->info_linear) {
 828		dso__data(dso)->status = DSO_DATA_STATUS_ERROR;
 829		return -1;
 830	}
 831
 832	len = node->info_linear->info.jited_prog_len;
 833	buf = (u8 *)(uintptr_t)node->info_linear->info.jited_prog_insns;
 834
 835	if (offset >= len)
 836		return -1;
 837
 838	size = (ssize_t)min(len - offset, (u64)size);
 839	memcpy(data, buf + offset, size);
 840	return size;
 841}
 842
 843static int bpf_size(struct dso *dso)
 844{
 845	struct bpf_prog_info_node *node;
 846	struct dso_bpf_prog *dso_bpf_prog = dso__bpf_prog(dso);
 847
 848	node = perf_env__find_bpf_prog_info(dso_bpf_prog->env, dso_bpf_prog->id);
 849	if (!node || !node->info_linear) {
 850		dso__data(dso)->status = DSO_DATA_STATUS_ERROR;
 851		return -1;
 852	}
 853
 854	dso__data(dso)->file_size = node->info_linear->info.jited_prog_len;
 855	return 0;
 856}
 857#endif // HAVE_LIBBPF_SUPPORT
 858
 859static void
 860dso_cache__free(struct dso *dso)
 861{
 862	struct rb_root *root = &dso__data(dso)->cache;
 863	struct rb_node *next = rb_first(root);
 864
 865	mutex_lock(dso__lock(dso));
 866	while (next) {
 867		struct dso_cache *cache;
 868
 869		cache = rb_entry(next, struct dso_cache, rb_node);
 870		next = rb_next(&cache->rb_node);
 871		rb_erase(&cache->rb_node, root);
 872		free(cache);
 873	}
 874	mutex_unlock(dso__lock(dso));
 875}
 876
 877static struct dso_cache *__dso_cache__find(struct dso *dso, u64 offset)
 878{
 879	const struct rb_root *root = &dso__data(dso)->cache;
 880	struct rb_node * const *p = &root->rb_node;
 881	const struct rb_node *parent = NULL;
 882	struct dso_cache *cache;
 883
 884	while (*p != NULL) {
 885		u64 end;
 886
 887		parent = *p;
 888		cache = rb_entry(parent, struct dso_cache, rb_node);
 889		end = cache->offset + DSO__DATA_CACHE_SIZE;
 890
 891		if (offset < cache->offset)
 892			p = &(*p)->rb_left;
 893		else if (offset >= end)
 894			p = &(*p)->rb_right;
 895		else
 896			return cache;
 897	}
 898
 899	return NULL;
 900}
 901
 902static struct dso_cache *
 903dso_cache__insert(struct dso *dso, struct dso_cache *new)
 904{
 905	struct rb_root *root = &dso__data(dso)->cache;
 906	struct rb_node **p = &root->rb_node;
 907	struct rb_node *parent = NULL;
 908	struct dso_cache *cache;
 909	u64 offset = new->offset;
 910
 911	mutex_lock(dso__lock(dso));
 912	while (*p != NULL) {
 913		u64 end;
 914
 915		parent = *p;
 916		cache = rb_entry(parent, struct dso_cache, rb_node);
 917		end = cache->offset + DSO__DATA_CACHE_SIZE;
 918
 919		if (offset < cache->offset)
 920			p = &(*p)->rb_left;
 921		else if (offset >= end)
 922			p = &(*p)->rb_right;
 923		else
 924			goto out;
 925	}
 926
 927	rb_link_node(&new->rb_node, parent, p);
 928	rb_insert_color(&new->rb_node, root);
 929
 930	cache = NULL;
 931out:
 932	mutex_unlock(dso__lock(dso));
 933	return cache;
 934}
 935
 936static ssize_t dso_cache__memcpy(struct dso_cache *cache, u64 offset, u8 *data,
 937				 u64 size, bool out)
 
 938{
 939	u64 cache_offset = offset - cache->offset;
 940	u64 cache_size   = min(cache->size - cache_offset, size);
 941
 942	if (out)
 943		memcpy(data, cache->data + cache_offset, cache_size);
 944	else
 945		memcpy(cache->data + cache_offset, data, cache_size);
 946	return cache_size;
 947}
 948
 949static ssize_t file_read(struct dso *dso, struct machine *machine,
 950			 u64 offset, char *data)
 
 951{
 
 
 952	ssize_t ret;
 953
 954	pthread_mutex_lock(&dso__data_open_lock);
 
 955
 956	/*
 957	 * dso__data(dso)->fd might be closed if other thread opened another
 958	 * file (dso) due to open file limit (RLIMIT_NOFILE).
 959	 */
 960	try_to_open_dso(dso, machine);
 961
 962	if (dso__data(dso)->fd < 0) {
 963		dso__data(dso)->status = DSO_DATA_STATUS_ERROR;
 964		ret = -errno;
 965		goto out;
 966	}
 967
 968	ret = pread(dso__data(dso)->fd, data, DSO__DATA_CACHE_SIZE, offset);
 969out:
 970	pthread_mutex_unlock(&dso__data_open_lock);
 971	return ret;
 972}
 973
 974static struct dso_cache *dso_cache__populate(struct dso *dso,
 975					     struct machine *machine,
 976					     u64 offset, ssize_t *ret)
 977{
 978	u64 cache_offset = offset & DSO__DATA_CACHE_MASK;
 979	struct dso_cache *cache;
 980	struct dso_cache *old;
 981
 982	cache = zalloc(sizeof(*cache) + DSO__DATA_CACHE_SIZE);
 983	if (!cache) {
 984		*ret = -ENOMEM;
 985		return NULL;
 986	}
 987#ifdef HAVE_LIBBPF_SUPPORT
 988	if (dso__binary_type(dso) == DSO_BINARY_TYPE__BPF_PROG_INFO)
 989		*ret = bpf_read(dso, cache_offset, cache->data);
 990	else
 991#endif
 992	if (dso__binary_type(dso) == DSO_BINARY_TYPE__OOL)
 993		*ret = DSO__DATA_CACHE_SIZE;
 994	else
 995		*ret = file_read(dso, machine, cache_offset, cache->data);
 996
 997	if (*ret <= 0) {
 998		free(cache);
 999		return NULL;
1000	}
1001
1002	cache->offset = cache_offset;
1003	cache->size   = *ret;
 
1004
1005	old = dso_cache__insert(dso, cache);
1006	if (old) {
1007		/* we lose the race */
1008		free(cache);
1009		cache = old;
1010	}
1011
1012	return cache;
1013}
 
 
 
 
 
1014
1015static struct dso_cache *dso_cache__find(struct dso *dso,
1016					 struct machine *machine,
1017					 u64 offset,
1018					 ssize_t *ret)
1019{
1020	struct dso_cache *cache = __dso_cache__find(dso, offset);
1021
1022	return cache ? cache : dso_cache__populate(dso, machine, offset, ret);
 
 
 
1023}
1024
1025static ssize_t dso_cache_io(struct dso *dso, struct machine *machine,
1026			    u64 offset, u8 *data, ssize_t size, bool out)
1027{
1028	struct dso_cache *cache;
1029	ssize_t ret = 0;
1030
1031	cache = dso_cache__find(dso, machine, offset, &ret);
1032	if (!cache)
1033		return ret;
1034
1035	return dso_cache__memcpy(cache, offset, data, size, out);
 
 
 
 
1036}
1037
1038/*
1039 * Reads and caches dso data DSO__DATA_CACHE_SIZE size chunks
1040 * in the rb_tree. Any read to already cached data is served
1041 * by cached data. Writes update the cache only, not the backing file.
1042 */
1043static ssize_t cached_io(struct dso *dso, struct machine *machine,
1044			 u64 offset, u8 *data, ssize_t size, bool out)
1045{
1046	ssize_t r = 0;
1047	u8 *p = data;
1048
1049	do {
1050		ssize_t ret;
1051
1052		ret = dso_cache_io(dso, machine, offset, p, size, out);
1053		if (ret < 0)
1054			return ret;
1055
1056		/* Reached EOF, return what we have. */
1057		if (!ret)
1058			break;
1059
1060		BUG_ON(ret > size);
1061
1062		r      += ret;
1063		p      += ret;
1064		offset += ret;
1065		size   -= ret;
1066
1067	} while (size);
1068
1069	return r;
1070}
1071
1072static int file_size(struct dso *dso, struct machine *machine)
1073{
1074	int ret = 0;
1075	struct stat st;
1076	char sbuf[STRERR_BUFSIZE];
1077
 
 
 
 
 
 
1078	pthread_mutex_lock(&dso__data_open_lock);
1079
1080	/*
1081	 * dso__data(dso)->fd might be closed if other thread opened another
1082	 * file (dso) due to open file limit (RLIMIT_NOFILE).
1083	 */
1084	try_to_open_dso(dso, machine);
1085
1086	if (dso__data(dso)->fd < 0) {
1087		ret = -errno;
1088		dso__data(dso)->status = DSO_DATA_STATUS_ERROR;
1089		goto out;
1090	}
1091
1092	if (fstat(dso__data(dso)->fd, &st) < 0) {
1093		ret = -errno;
1094		pr_err("dso cache fstat failed: %s\n",
1095		       str_error_r(errno, sbuf, sizeof(sbuf)));
1096		dso__data(dso)->status = DSO_DATA_STATUS_ERROR;
1097		goto out;
1098	}
1099	dso__data(dso)->file_size = st.st_size;
1100
1101out:
1102	pthread_mutex_unlock(&dso__data_open_lock);
1103	return ret;
1104}
1105
1106int dso__data_file_size(struct dso *dso, struct machine *machine)
1107{
1108	if (dso__data(dso)->file_size)
1109		return 0;
1110
1111	if (dso__data(dso)->status == DSO_DATA_STATUS_ERROR)
1112		return -1;
1113#ifdef HAVE_LIBBPF_SUPPORT
1114	if (dso__binary_type(dso) == DSO_BINARY_TYPE__BPF_PROG_INFO)
1115		return bpf_size(dso);
1116#endif
1117	return file_size(dso, machine);
1118}
1119
1120/**
1121 * dso__data_size - Return dso data size
1122 * @dso: dso object
1123 * @machine: machine object
1124 *
1125 * Return: dso data size
1126 */
1127off_t dso__data_size(struct dso *dso, struct machine *machine)
1128{
1129	if (dso__data_file_size(dso, machine))
1130		return -1;
1131
1132	/* For now just estimate dso data size is close to file size */
1133	return dso__data(dso)->file_size;
1134}
1135
1136static ssize_t data_read_write_offset(struct dso *dso, struct machine *machine,
1137				      u64 offset, u8 *data, ssize_t size,
1138				      bool out)
1139{
1140	if (dso__data_file_size(dso, machine))
1141		return -1;
1142
1143	/* Check the offset sanity. */
1144	if (offset > dso__data(dso)->file_size)
1145		return -1;
1146
1147	if (offset + size < offset)
1148		return -1;
1149
1150	return cached_io(dso, machine, offset, data, size, out);
1151}
1152
1153/**
1154 * dso__data_read_offset - Read data from dso file offset
1155 * @dso: dso object
1156 * @machine: machine object
1157 * @offset: file offset
1158 * @data: buffer to store data
1159 * @size: size of the @data buffer
1160 *
1161 * External interface to read data from dso file offset. Open
1162 * dso data file and use cached_read to get the data.
1163 */
1164ssize_t dso__data_read_offset(struct dso *dso, struct machine *machine,
1165			      u64 offset, u8 *data, ssize_t size)
1166{
1167	if (dso__data(dso)->status == DSO_DATA_STATUS_ERROR)
1168		return -1;
1169
1170	return data_read_write_offset(dso, machine, offset, data, size, true);
1171}
1172
1173/**
1174 * dso__data_read_addr - Read data from dso address
1175 * @dso: dso object
1176 * @machine: machine object
1177 * @add: virtual memory address
1178 * @data: buffer to store data
1179 * @size: size of the @data buffer
1180 *
1181 * External interface to read data from dso address.
1182 */
1183ssize_t dso__data_read_addr(struct dso *dso, struct map *map,
1184			    struct machine *machine, u64 addr,
1185			    u8 *data, ssize_t size)
1186{
1187	u64 offset = map__map_ip(map, addr);
1188
1189	return dso__data_read_offset(dso, machine, offset, data, size);
1190}
1191
1192/**
1193 * dso__data_write_cache_offs - Write data to dso data cache at file offset
1194 * @dso: dso object
1195 * @machine: machine object
1196 * @offset: file offset
1197 * @data: buffer to write
1198 * @size: size of the @data buffer
1199 *
1200 * Write into the dso file data cache, but do not change the file itself.
1201 */
1202ssize_t dso__data_write_cache_offs(struct dso *dso, struct machine *machine,
1203				   u64 offset, const u8 *data_in, ssize_t size)
1204{
1205	u8 *data = (u8 *)data_in; /* cast away const to use same fns for r/w */
1206
1207	if (dso__data(dso)->status == DSO_DATA_STATUS_ERROR)
1208		return -1;
1209
1210	return data_read_write_offset(dso, machine, offset, data, size, false);
1211}
1212
1213/**
1214 * dso__data_write_cache_addr - Write data to dso data cache at dso address
1215 * @dso: dso object
1216 * @machine: machine object
1217 * @add: virtual memory address
1218 * @data: buffer to write
1219 * @size: size of the @data buffer
1220 *
1221 * External interface to write into the dso file data cache, but do not change
1222 * the file itself.
1223 */
1224ssize_t dso__data_write_cache_addr(struct dso *dso, struct map *map,
1225				   struct machine *machine, u64 addr,
1226				   const u8 *data, ssize_t size)
1227{
1228	u64 offset = map__map_ip(map, addr);
1229
1230	return dso__data_write_cache_offs(dso, machine, offset, data, size);
1231}
1232
1233struct map *dso__new_map(const char *name)
1234{
1235	struct map *map = NULL;
1236	struct dso *dso = dso__new(name);
1237
1238	if (dso) {
1239		map = map__new2(0, dso);
1240		dso__put(dso);
1241	}
1242
1243	return map;
1244}
1245
1246struct dso *machine__findnew_kernel(struct machine *machine, const char *name,
1247				    const char *short_name, int dso_type)
1248{
1249	/*
1250	 * The kernel dso could be created by build_id processing.
1251	 */
1252	struct dso *dso = machine__findnew_dso(machine, name);
1253
1254	/*
1255	 * We need to run this in all cases, since during the build_id
1256	 * processing we had no idea this was the kernel dso.
1257	 */
1258	if (dso != NULL) {
1259		dso__set_short_name(dso, short_name, false);
1260		dso__set_kernel(dso, dso_type);
1261	}
1262
1263	return dso;
1264}
1265
1266static void dso__set_long_name_id(struct dso *dso, const char *name, bool name_allocated)
 
 
 
 
 
 
1267{
1268	struct dsos *dsos = dso__dsos(dso);
 
1269
1270	if (name == NULL)
1271		return;
 
 
 
 
 
 
1272
1273	if (dsos) {
1274		/*
1275		 * Need to avoid re-sorting the dsos breaking by non-atomically
1276		 * renaming the dso.
1277		 */
1278		down_write(&dsos->lock);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1279	}
1280
1281	if (dso__long_name_allocated(dso))
1282		free((char *)dso__long_name(dso));
1283
1284	RC_CHK_ACCESS(dso)->long_name = name;
1285	RC_CHK_ACCESS(dso)->long_name_len = strlen(name);
1286	dso__set_long_name_allocated(dso, name_allocated);
1287
1288	if (dsos) {
1289		dsos->sorted = false;
1290		up_write(&dsos->lock);
1291	}
 
1292}
1293
1294static int __dso_id__cmp(const struct dso_id *a, const struct dso_id *b)
 
1295{
1296	if (a->maj > b->maj) return -1;
1297	if (a->maj < b->maj) return 1;
1298
1299	if (a->min > b->min) return -1;
1300	if (a->min < b->min) return 1;
1301
1302	if (a->ino > b->ino) return -1;
1303	if (a->ino < b->ino) return 1;
1304
1305	/*
1306	 * Synthesized MMAP events have zero ino_generation, avoid comparing
1307	 * them with MMAP events with actual ino_generation.
1308	 *
1309	 * I found it harmful because the mismatch resulted in a new
1310	 * dso that did not have a build ID whereas the original dso did have a
1311	 * build ID. The build ID was essential because the object was not found
1312	 * otherwise. - Adrian
1313	 */
1314	if (a->ino_generation && b->ino_generation) {
1315		if (a->ino_generation > b->ino_generation) return -1;
1316		if (a->ino_generation < b->ino_generation) return 1;
1317	}
1318
1319	return 0;
1320}
1321
1322bool dso_id__empty(const struct dso_id *id)
1323{
1324	if (!id)
1325		return true;
1326
1327	return !id->maj && !id->min && !id->ino && !id->ino_generation;
1328}
1329
1330void __dso__inject_id(struct dso *dso, const struct dso_id *id)
1331{
1332	struct dsos *dsos = dso__dsos(dso);
1333	struct dso_id *dso_id = dso__id(dso);
1334
1335	/* dsos write lock held by caller. */
 
 
 
 
 
 
 
 
1336
1337	dso_id->maj = id->maj;
1338	dso_id->min = id->min;
1339	dso_id->ino = id->ino;
1340	dso_id->ino_generation = id->ino_generation;
1341
1342	if (dsos)
1343		dsos->sorted = false;
1344}
1345
1346int dso_id__cmp(const struct dso_id *a, const struct dso_id *b)
1347{
1348	/*
1349	 * The second is always dso->id, so zeroes if not set, assume passing
1350	 * NULL for a means a zeroed id
1351	 */
1352	if (dso_id__empty(a) || dso_id__empty(b))
1353		return 0;
1354
1355	return __dso_id__cmp(a, b);
1356}
1357
1358int dso__cmp_id(struct dso *a, struct dso *b)
1359{
1360	return __dso_id__cmp(dso__id(a), dso__id(b));
1361}
1362
1363void dso__set_long_name(struct dso *dso, const char *name, bool name_allocated)
1364{
1365	dso__set_long_name_id(dso, name, name_allocated);
1366}
 
 
 
1367
1368void dso__set_short_name(struct dso *dso, const char *name, bool name_allocated)
1369{
1370	struct dsos *dsos = dso__dsos(dso);
1371
1372	if (name == NULL)
1373		return;
 
 
 
 
1374
1375	if (dsos) {
1376		/*
1377		 * Need to avoid re-sorting the dsos breaking by non-atomically
1378		 * renaming the dso.
1379		 */
1380		down_write(&dsos->lock);
1381	}
1382	if (dso__short_name_allocated(dso))
1383		free((char *)dso__short_name(dso));
1384
1385	RC_CHK_ACCESS(dso)->short_name		  = name;
1386	RC_CHK_ACCESS(dso)->short_name_len	  = strlen(name);
1387	dso__set_short_name_allocated(dso, name_allocated);
1388
1389	if (dsos) {
1390		dsos->sorted = false;
1391		up_write(&dsos->lock);
1392	}
1393}
1394
1395int dso__name_len(const struct dso *dso)
1396{
1397	if (!dso)
1398		return strlen("[unknown]");
1399	if (verbose > 0)
1400		return dso__long_name_len(dso);
1401
1402	return dso__short_name_len(dso);
1403}
1404
1405bool dso__loaded(const struct dso *dso)
1406{
1407	return RC_CHK_ACCESS(dso)->loaded;
1408}
1409
1410bool dso__sorted_by_name(const struct dso *dso)
1411{
1412	return RC_CHK_ACCESS(dso)->sorted_by_name;
1413}
1414
1415void dso__set_sorted_by_name(struct dso *dso)
1416{
1417	RC_CHK_ACCESS(dso)->sorted_by_name = true;
1418}
1419
1420struct dso *dso__new_id(const char *name, const struct dso_id *id)
1421{
1422	RC_STRUCT(dso) *dso = zalloc(sizeof(*dso) + strlen(name) + 1);
1423	struct dso *res;
1424	struct dso_data *data;
1425
1426	if (!dso)
1427		return NULL;
1428
1429	if (ADD_RC_CHK(res, dso)) {
 
1430		strcpy(dso->name, name);
1431		if (id)
1432			dso->id = *id;
1433		dso__set_long_name_id(res, dso->name, false);
1434		dso__set_short_name(res, dso->name, false);
1435		dso->symbols = RB_ROOT_CACHED;
1436		dso->symbol_names = NULL;
1437		dso->symbol_names_len = 0;
1438		dso->inlined_nodes = RB_ROOT_CACHED;
1439		dso->srclines = RB_ROOT_CACHED;
1440		dso->data_types = RB_ROOT;
1441		dso->global_vars = RB_ROOT;
1442		dso->data.fd = -1;
1443		dso->data.status = DSO_DATA_STATUS_UNKNOWN;
1444		dso->symtab_type = DSO_BINARY_TYPE__NOT_FOUND;
1445		dso->binary_type = DSO_BINARY_TYPE__NOT_FOUND;
1446		dso->is_64_bit = (sizeof(void *) == 8);
1447		dso->loaded = 0;
1448		dso->rel = 0;
1449		dso->sorted_by_name = 0;
1450		dso->has_build_id = 0;
1451		dso->has_srcline = 1;
1452		dso->a2l_fails = 1;
1453		dso->kernel = DSO_SPACE__USER;
1454		dso->is_kmod = 0;
1455		dso->needs_swap = DSO_SWAP__UNSET;
1456		dso->comp = COMP_ID__NONE;
1457		mutex_init(&dso->lock);
1458		refcount_set(&dso->refcnt, 1);
1459		data = &dso->data;
1460		data->cache = RB_ROOT;
1461		data->fd = -1;
1462		data->status = DSO_DATA_STATUS_UNKNOWN;
1463		INIT_LIST_HEAD(&data->open_entry);
1464#ifdef REFCNT_CHECKING
1465		data->dso = NULL; /* Set when on the open_entry list. */
1466#endif
1467	}
1468	return res;
1469}
1470
1471struct dso *dso__new(const char *name)
1472{
1473	return dso__new_id(name, NULL);
1474}
1475
1476void dso__delete(struct dso *dso)
1477{
1478	if (dso__dsos(dso))
1479		pr_err("DSO %s is still in rbtree when being deleted!\n", dso__long_name(dso));
1480
1481	/* free inlines first, as they reference symbols */
1482	inlines__tree_delete(&RC_CHK_ACCESS(dso)->inlined_nodes);
1483	srcline__tree_delete(&RC_CHK_ACCESS(dso)->srclines);
1484	symbols__delete(&RC_CHK_ACCESS(dso)->symbols);
1485	RC_CHK_ACCESS(dso)->symbol_names_len = 0;
1486	zfree(&RC_CHK_ACCESS(dso)->symbol_names);
1487	annotated_data_type__tree_delete(dso__data_types(dso));
1488	global_var_type__tree_delete(dso__global_vars(dso));
1489
1490	if (RC_CHK_ACCESS(dso)->short_name_allocated) {
1491		zfree((char **)&RC_CHK_ACCESS(dso)->short_name);
1492		RC_CHK_ACCESS(dso)->short_name_allocated = false;
1493	}
1494
1495	if (RC_CHK_ACCESS(dso)->long_name_allocated) {
1496		zfree((char **)&RC_CHK_ACCESS(dso)->long_name);
1497		RC_CHK_ACCESS(dso)->long_name_allocated = false;
1498	}
1499
1500	dso__data_close(dso);
1501	auxtrace_cache__free(RC_CHK_ACCESS(dso)->auxtrace_cache);
1502	dso_cache__free(dso);
1503	dso__free_a2l(dso);
1504	dso__free_symsrc_filename(dso);
1505	nsinfo__zput(RC_CHK_ACCESS(dso)->nsinfo);
1506	mutex_destroy(dso__lock(dso));
1507	RC_CHK_FREE(dso);
1508}
1509
1510struct dso *dso__get(struct dso *dso)
1511{
1512	struct dso *result;
1513
1514	if (RC_CHK_GET(result, dso))
1515		refcount_inc(&RC_CHK_ACCESS(dso)->refcnt);
1516
1517	return result;
1518}
1519
1520void dso__put(struct dso *dso)
1521{
1522	if (dso && refcount_dec_and_test(&RC_CHK_ACCESS(dso)->refcnt))
1523		dso__delete(dso);
1524	else
1525		RC_CHK_PUT(dso);
1526}
1527
1528void dso__set_build_id(struct dso *dso, struct build_id *bid)
1529{
1530	RC_CHK_ACCESS(dso)->bid = *bid;
1531	RC_CHK_ACCESS(dso)->has_build_id = 1;
1532}
1533
1534bool dso__build_id_equal(const struct dso *dso, struct build_id *bid)
1535{
1536	const struct build_id *dso_bid = dso__bid_const(dso);
1537
1538	if (dso_bid->size > bid->size && dso_bid->size == BUILD_ID_SIZE) {
1539		/*
1540		 * For the backward compatibility, it allows a build-id has
1541		 * trailing zeros.
1542		 */
1543		return !memcmp(dso_bid->data, bid->data, bid->size) &&
1544			!memchr_inv(&dso_bid->data[bid->size], 0,
1545				    dso_bid->size - bid->size);
1546	}
1547
1548	return dso_bid->size == bid->size &&
1549	       memcmp(dso_bid->data, bid->data, dso_bid->size) == 0;
1550}
1551
1552void dso__read_running_kernel_build_id(struct dso *dso, struct machine *machine)
1553{
1554	char path[PATH_MAX];
1555
1556	if (machine__is_default_guest(machine))
1557		return;
1558	sprintf(path, "%s/sys/kernel/notes", machine->root_dir);
1559	if (sysfs__read_build_id(path, dso__bid(dso)) == 0)
1560		dso__set_has_build_id(dso);
 
1561}
1562
1563int dso__kernel_module_get_build_id(struct dso *dso,
1564				    const char *root_dir)
1565{
1566	char filename[PATH_MAX];
1567	/*
1568	 * kernel module short names are of the form "[module]" and
1569	 * we need just "module" here.
1570	 */
1571	const char *name = dso__short_name(dso) + 1;
1572
1573	snprintf(filename, sizeof(filename),
1574		 "%s/sys/module/%.*s/notes/.note.gnu.build-id",
1575		 root_dir, (int)strlen(name) - 1, name);
1576
1577	if (sysfs__read_build_id(filename, dso__bid(dso)) == 0)
1578		dso__set_has_build_id(dso);
 
1579
1580	return 0;
1581}
1582
1583static size_t dso__fprintf_buildid(struct dso *dso, FILE *fp)
1584{
1585	char sbuild_id[SBUILD_ID_SIZE];
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1586
1587	build_id__sprintf(dso__bid(dso), sbuild_id);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1588	return fprintf(fp, "%s", sbuild_id);
1589}
1590
1591size_t dso__fprintf(struct dso *dso, FILE *fp)
1592{
1593	struct rb_node *nd;
1594	size_t ret = fprintf(fp, "dso: %s (", dso__short_name(dso));
1595
1596	if (dso__short_name(dso) != dso__long_name(dso))
1597		ret += fprintf(fp, "%s, ", dso__long_name(dso));
1598	ret += fprintf(fp, "%sloaded, ", dso__loaded(dso) ? "" : "NOT ");
 
1599	ret += dso__fprintf_buildid(dso, fp);
1600	ret += fprintf(fp, ")\n");
1601	for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) {
1602		struct symbol *pos = rb_entry(nd, struct symbol, rb_node);
1603		ret += symbol__fprintf(pos, fp);
1604	}
1605
1606	return ret;
1607}
1608
1609enum dso_type dso__type(struct dso *dso, struct machine *machine)
1610{
1611	int fd;
1612	enum dso_type type = DSO__TYPE_UNKNOWN;
1613
1614	fd = dso__data_get_fd(dso, machine);
1615	if (fd >= 0) {
1616		type = dso__type_fd(fd);
1617		dso__data_put_fd(dso);
1618	}
1619
1620	return type;
1621}
1622
1623int dso__strerror_load(struct dso *dso, char *buf, size_t buflen)
1624{
1625	int idx, errnum = *dso__load_errno(dso);
1626	/*
1627	 * This must have a same ordering as the enum dso_load_errno.
1628	 */
1629	static const char *dso_load__error_str[] = {
1630	"Internal tools/perf/ library error",
1631	"Invalid ELF file",
1632	"Can not read build id",
1633	"Mismatching build id",
1634	"Decompression failure",
1635	};
1636
1637	BUG_ON(buflen == 0);
1638
1639	if (errnum >= 0) {
1640		const char *err = str_error_r(errnum, buf, buflen);
1641
1642		if (err != buf)
1643			scnprintf(buf, buflen, "%s", err);
1644
1645		return 0;
1646	}
1647
1648	if (errnum <  __DSO_LOAD_ERRNO__START || errnum >= __DSO_LOAD_ERRNO__END)
1649		return -1;
1650
1651	idx = errnum - __DSO_LOAD_ERRNO__START;
1652	scnprintf(buf, buflen, "%s", dso_load__error_str[idx]);
1653	return 0;
1654}
1655
1656bool perf_pid_map_tid(const char *dso_name, int *tid)
1657{
1658	return sscanf(dso_name, "/tmp/perf-%d.map", tid) == 1;
1659}
1660
1661bool is_perf_pid_map_name(const char *dso_name)
1662{
1663	int tid;
1664
1665	return perf_pid_map_tid(dso_name, &tid);
1666}
v4.6
 
   1#include <asm/bug.h>
 
 
 
   2#include <sys/time.h>
   3#include <sys/resource.h>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
   4#include "symbol.h"
 
   5#include "dso.h"
 
   6#include "machine.h"
   7#include "auxtrace.h"
   8#include "util.h"
   9#include "debug.h"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  10
  11char dso__symtab_origin(const struct dso *dso)
  12{
  13	static const char origin[] = {
  14		[DSO_BINARY_TYPE__KALLSYMS]			= 'k',
  15		[DSO_BINARY_TYPE__VMLINUX]			= 'v',
  16		[DSO_BINARY_TYPE__JAVA_JIT]			= 'j',
  17		[DSO_BINARY_TYPE__DEBUGLINK]			= 'l',
  18		[DSO_BINARY_TYPE__BUILD_ID_CACHE]		= 'B',
 
  19		[DSO_BINARY_TYPE__FEDORA_DEBUGINFO]		= 'f',
  20		[DSO_BINARY_TYPE__UBUNTU_DEBUGINFO]		= 'u',
 
  21		[DSO_BINARY_TYPE__OPENEMBEDDED_DEBUGINFO]	= 'o',
  22		[DSO_BINARY_TYPE__BUILDID_DEBUGINFO]		= 'b',
  23		[DSO_BINARY_TYPE__SYSTEM_PATH_DSO]		= 'd',
  24		[DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE]		= 'K',
  25		[DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP]	= 'm',
  26		[DSO_BINARY_TYPE__GUEST_KALLSYMS]		= 'g',
  27		[DSO_BINARY_TYPE__GUEST_KMODULE]		= 'G',
  28		[DSO_BINARY_TYPE__GUEST_KMODULE_COMP]		= 'M',
  29		[DSO_BINARY_TYPE__GUEST_VMLINUX]		= 'V',
  30	};
  31
  32	if (dso == NULL || dso->symtab_type == DSO_BINARY_TYPE__NOT_FOUND)
  33		return '!';
  34	return origin[dso->symtab_type];
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  35}
  36
  37int dso__read_binary_type_filename(const struct dso *dso,
  38				   enum dso_binary_type type,
  39				   char *root_dir, char *filename, size_t size)
  40{
  41	char build_id_hex[BUILD_ID_SIZE * 2 + 1];
  42	int ret = 0;
  43	size_t len;
  44
  45	switch (type) {
  46	case DSO_BINARY_TYPE__DEBUGLINK: {
  47		char *debuglink;
 
 
 
 
 
 
 
 
 
  48
  49		len = __symbol__join_symfs(filename, size, dso->long_name);
  50		debuglink = filename + len;
  51		while (debuglink != filename && *debuglink != '/')
  52			debuglink--;
  53		if (*debuglink == '/')
  54			debuglink++;
  55
  56		ret = -1;
  57		if (!is_regular_file(filename))
 
 
 
 
 
  58			break;
  59
  60		ret = filename__read_debuglink(filename, debuglink,
  61					       size - (debuglink - filename));
 
 
 
 
 
 
 
  62		}
 
  63		break;
 
  64	case DSO_BINARY_TYPE__BUILD_ID_CACHE:
  65		/* skip the locally configured cache if a symfs is given */
  66		if (symbol_conf.symfs[0] ||
  67		    (dso__build_id_filename(dso, filename, size) == NULL))
 
 
 
  68			ret = -1;
  69		break;
  70
  71	case DSO_BINARY_TYPE__FEDORA_DEBUGINFO:
  72		len = __symbol__join_symfs(filename, size, "/usr/lib/debug");
  73		snprintf(filename + len, size - len, "%s.debug", dso->long_name);
  74		break;
  75
  76	case DSO_BINARY_TYPE__UBUNTU_DEBUGINFO:
  77		len = __symbol__join_symfs(filename, size, "/usr/lib/debug");
  78		snprintf(filename + len, size - len, "%s", dso->long_name);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  79		break;
  80
  81	case DSO_BINARY_TYPE__OPENEMBEDDED_DEBUGINFO:
  82	{
  83		const char *last_slash;
  84		size_t dir_size;
  85
  86		last_slash = dso->long_name + dso->long_name_len;
  87		while (last_slash != dso->long_name && *last_slash != '/')
  88			last_slash--;
  89
  90		len = __symbol__join_symfs(filename, size, "");
  91		dir_size = last_slash - dso->long_name + 2;
  92		if (dir_size > (size - len)) {
  93			ret = -1;
  94			break;
  95		}
  96		len += scnprintf(filename + len, dir_size, "%s",  dso->long_name);
  97		len += scnprintf(filename + len , size - len, ".debug%s",
  98								last_slash);
  99		break;
 100	}
 101
 102	case DSO_BINARY_TYPE__BUILDID_DEBUGINFO:
 103		if (!dso->has_build_id) {
 104			ret = -1;
 105			break;
 106		}
 107
 108		build_id__sprintf(dso->build_id,
 109				  sizeof(dso->build_id),
 110				  build_id_hex);
 111		len = __symbol__join_symfs(filename, size, "/usr/lib/debug/.build-id/");
 112		snprintf(filename + len, size - len, "%.2s/%s.debug",
 113			 build_id_hex, build_id_hex + 2);
 114		break;
 115
 116	case DSO_BINARY_TYPE__VMLINUX:
 117	case DSO_BINARY_TYPE__GUEST_VMLINUX:
 118	case DSO_BINARY_TYPE__SYSTEM_PATH_DSO:
 119		__symbol__join_symfs(filename, size, dso->long_name);
 120		break;
 121
 122	case DSO_BINARY_TYPE__GUEST_KMODULE:
 123	case DSO_BINARY_TYPE__GUEST_KMODULE_COMP:
 124		path__join3(filename, size, symbol_conf.symfs,
 125			    root_dir, dso->long_name);
 126		break;
 127
 128	case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE:
 129	case DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP:
 130		__symbol__join_symfs(filename, size, dso->long_name);
 131		break;
 132
 133	case DSO_BINARY_TYPE__KCORE:
 134	case DSO_BINARY_TYPE__GUEST_KCORE:
 135		snprintf(filename, size, "%s", dso->long_name);
 136		break;
 137
 138	default:
 139	case DSO_BINARY_TYPE__KALLSYMS:
 140	case DSO_BINARY_TYPE__GUEST_KALLSYMS:
 141	case DSO_BINARY_TYPE__JAVA_JIT:
 
 
 
 142	case DSO_BINARY_TYPE__NOT_FOUND:
 143		ret = -1;
 144		break;
 145	}
 146
 147	return ret;
 148}
 149
 
 
 
 
 150static const struct {
 151	const char *fmt;
 152	int (*decompress)(const char *input, int output);
 
 153} compressions[] = {
 
 154#ifdef HAVE_ZLIB_SUPPORT
 155	{ "gz", gzip_decompress_to_file },
 156#endif
 157#ifdef HAVE_LZMA_SUPPORT
 158	{ "xz", lzma_decompress_to_file },
 159#endif
 160	{ NULL, NULL },
 161};
 162
 163bool is_supported_compression(const char *ext)
 164{
 165	unsigned i;
 166
 167	for (i = 0; compressions[i].fmt; i++) {
 168		if (!strcmp(ext, compressions[i].fmt))
 169			return true;
 170	}
 171	return false;
 172}
 173
 174bool is_kernel_module(const char *pathname, int cpumode)
 175{
 176	struct kmod_path m;
 177	int mode = cpumode & PERF_RECORD_MISC_CPUMODE_MASK;
 178
 179	WARN_ONCE(mode != cpumode,
 180		  "Internal error: passing unmasked cpumode (%x) to is_kernel_module",
 181		  cpumode);
 182
 183	switch (mode) {
 184	case PERF_RECORD_MISC_USER:
 185	case PERF_RECORD_MISC_HYPERVISOR:
 186	case PERF_RECORD_MISC_GUEST_USER:
 187		return false;
 188	/* Treat PERF_RECORD_MISC_CPUMODE_UNKNOWN as kernel */
 189	default:
 190		if (kmod_path__parse(&m, pathname)) {
 191			pr_err("Failed to check whether %s is a kernel module or not. Assume it is.",
 192					pathname);
 193			return true;
 194		}
 195	}
 196
 197	return m.kmod;
 198}
 199
 200bool decompress_to_file(const char *ext, const char *filename, int output_fd)
 
 
 
 
 
 
 
 201{
 202	unsigned i;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 203
 204	for (i = 0; compressions[i].fmt; i++) {
 205		if (!strcmp(ext, compressions[i].fmt))
 206			return !compressions[i].decompress(filename,
 207							   output_fd);
 208	}
 209	return false;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 210}
 211
 212bool dso__needs_decompress(struct dso *dso)
 
 213{
 214	return dso->symtab_type == DSO_BINARY_TYPE__SYSTEM_PATH_KMODULE_COMP ||
 215		dso->symtab_type == DSO_BINARY_TYPE__GUEST_KMODULE_COMP;
 
 
 216}
 217
 218/*
 219 * Parses kernel module specified in @path and updates
 220 * @m argument like:
 221 *
 222 *    @comp - true if @path contains supported compression suffix,
 223 *            false otherwise
 224 *    @kmod - true if @path contains '.ko' suffix in right position,
 225 *            false otherwise
 226 *    @name - if (@alloc_name && @kmod) is true, it contains strdup-ed base name
 227 *            of the kernel module without suffixes, otherwise strudup-ed
 228 *            base name of @path
 229 *    @ext  - if (@alloc_ext && @comp) is true, it contains strdup-ed string
 230 *            the compression suffix
 231 *
 232 * Returns 0 if there's no strdup error, -ENOMEM otherwise.
 233 */
 234int __kmod_path__parse(struct kmod_path *m, const char *path,
 235		       bool alloc_name, bool alloc_ext)
 236{
 237	const char *name = strrchr(path, '/');
 238	const char *ext  = strrchr(path, '.');
 239	bool is_simple_name = false;
 240
 241	memset(m, 0x0, sizeof(*m));
 242	name = name ? name + 1 : path;
 243
 244	/*
 245	 * '.' is also a valid character for module name. For example:
 246	 * [aaa.bbb] is a valid module name. '[' should have higher
 247	 * priority than '.ko' suffix.
 248	 *
 249	 * The kernel names are from machine__mmap_name. Such
 250	 * name should belong to kernel itself, not kernel module.
 251	 */
 252	if (name[0] == '[') {
 253		is_simple_name = true;
 254		if ((strncmp(name, "[kernel.kallsyms]", 17) == 0) ||
 255		    (strncmp(name, "[guest.kernel.kallsyms", 22) == 0) ||
 256		    (strncmp(name, "[vdso]", 6) == 0) ||
 
 
 257		    (strncmp(name, "[vsyscall]", 10) == 0)) {
 258			m->kmod = false;
 259
 260		} else
 261			m->kmod = true;
 262	}
 263
 264	/* No extension, just return name. */
 265	if ((ext == NULL) || is_simple_name) {
 266		if (alloc_name) {
 267			m->name = strdup(name);
 268			return m->name ? 0 : -ENOMEM;
 269		}
 270		return 0;
 271	}
 272
 273	if (is_supported_compression(ext + 1)) {
 274		m->comp = true;
 275		ext -= 3;
 276	}
 277
 278	/* Check .ko extension only if there's enough name left. */
 279	if (ext > name)
 280		m->kmod = !strncmp(ext, ".ko", 3);
 281
 282	if (alloc_name) {
 283		if (m->kmod) {
 284			if (asprintf(&m->name, "[%.*s]", (int) (ext - name), name) == -1)
 285				return -ENOMEM;
 286		} else {
 287			if (asprintf(&m->name, "%s", name) == -1)
 288				return -ENOMEM;
 289		}
 290
 291		strxfrchar(m->name, '-', '_');
 292	}
 293
 294	if (alloc_ext && m->comp) {
 295		m->ext = strdup(ext + 4);
 296		if (!m->ext) {
 297			free((void *) m->name);
 298			return -ENOMEM;
 299		}
 
 
 
 
 
 
 
 
 
 300	}
 301
 302	return 0;
 
 303}
 304
 305/*
 306 * Global list of open DSOs and the counter.
 307 */
 308static LIST_HEAD(dso__data_open);
 309static long dso__data_open_cnt;
 310static pthread_mutex_t dso__data_open_lock = PTHREAD_MUTEX_INITIALIZER;
 311
 312static void dso__list_add(struct dso *dso)
 313{
 314	list_add_tail(&dso->data.open_entry, &dso__data_open);
 
 
 
 
 
 315	dso__data_open_cnt++;
 316}
 317
 318static void dso__list_del(struct dso *dso)
 319{
 320	list_del(&dso->data.open_entry);
 
 
 
 321	WARN_ONCE(dso__data_open_cnt <= 0,
 322		  "DSO data fd counter out of bounds.");
 323	dso__data_open_cnt--;
 324}
 325
 326static void close_first_dso(void);
 327
 328static int do_open(char *name)
 329{
 330	int fd;
 331	char sbuf[STRERR_BUFSIZE];
 332
 333	do {
 334		fd = open(name, O_RDONLY);
 335		if (fd >= 0)
 336			return fd;
 337
 338		pr_debug("dso open failed: %s\n",
 339			 strerror_r(errno, sbuf, sizeof(sbuf)));
 340		if (!dso__data_open_cnt || errno != EMFILE)
 341			break;
 342
 343		close_first_dso();
 344	} while (1);
 345
 346	return -1;
 347}
 348
 
 
 
 
 
 349static int __open_dso(struct dso *dso, struct machine *machine)
 350{
 351	int fd;
 352	char *root_dir = (char *)"";
 353	char *name = malloc(PATH_MAX);
 
 354
 355	if (!name)
 356		return -ENOMEM;
 357
 
 358	if (machine)
 359		root_dir = machine->root_dir;
 360
 361	if (dso__read_binary_type_filename(dso, dso->binary_type,
 362					    root_dir, name, PATH_MAX)) {
 
 
 
 
 
 
 
 
 
 
 
 
 363		free(name);
 364		return -EINVAL;
 
 
 
 
 
 
 
 
 
 
 
 
 
 365	}
 366
 367	fd = do_open(name);
 
 
 
 
 
 
 368	free(name);
 369	return fd;
 370}
 371
 372static void check_data_close(void);
 373
 374/**
 375 * dso_close - Open DSO data file
 376 * @dso: dso object
 377 *
 378 * Open @dso's data file descriptor and updates
 379 * list/count of open DSO objects.
 380 */
 381static int open_dso(struct dso *dso, struct machine *machine)
 382{
 383	int fd = __open_dso(dso, machine);
 
 
 
 
 
 
 
 
 
 
 384
 385	if (fd >= 0) {
 386		dso__list_add(dso);
 387		/*
 388		 * Check if we crossed the allowed number
 389		 * of opened DSOs and close one if needed.
 390		 */
 391		check_data_close();
 392	}
 393
 394	return fd;
 395}
 396
 397static void close_data_fd(struct dso *dso)
 398{
 399	if (dso->data.fd >= 0) {
 400		close(dso->data.fd);
 401		dso->data.fd = -1;
 402		dso->data.file_size = 0;
 403		dso__list_del(dso);
 404	}
 405}
 406
 407/**
 408 * dso_close - Close DSO data file
 409 * @dso: dso object
 410 *
 411 * Close @dso's data file descriptor and updates
 412 * list/count of open DSO objects.
 413 */
 414static void close_dso(struct dso *dso)
 415{
 416	close_data_fd(dso);
 417}
 418
 419static void close_first_dso(void)
 420{
 
 421	struct dso *dso;
 422
 423	dso = list_first_entry(&dso__data_open, struct dso, data.open_entry);
 
 
 
 
 
 424	close_dso(dso);
 425}
 426
 427static rlim_t get_fd_limit(void)
 428{
 429	struct rlimit l;
 430	rlim_t limit = 0;
 431
 432	/* Allow half of the current open fd limit. */
 433	if (getrlimit(RLIMIT_NOFILE, &l) == 0) {
 434		if (l.rlim_cur == RLIM_INFINITY)
 435			limit = l.rlim_cur;
 436		else
 437			limit = l.rlim_cur / 2;
 438	} else {
 439		pr_err("failed to get fd limit\n");
 440		limit = 1;
 441	}
 442
 443	return limit;
 444}
 445
 
 
 
 
 
 
 
 
 
 
 
 
 446static bool may_cache_fd(void)
 447{
 448	static rlim_t limit;
 449
 450	if (!limit)
 451		limit = get_fd_limit();
 452
 453	if (limit == RLIM_INFINITY)
 454		return true;
 455
 456	return limit > (rlim_t) dso__data_open_cnt;
 457}
 458
 459/*
 460 * Check and close LRU dso if we crossed allowed limit
 461 * for opened dso file descriptors. The limit is half
 462 * of the RLIMIT_NOFILE files opened.
 463*/
 464static void check_data_close(void)
 465{
 466	bool cache_fd = may_cache_fd();
 467
 468	if (!cache_fd)
 469		close_first_dso();
 470}
 471
 472/**
 473 * dso__data_close - Close DSO data file
 474 * @dso: dso object
 475 *
 476 * External interface to close @dso's data file descriptor.
 477 */
 478void dso__data_close(struct dso *dso)
 479{
 480	pthread_mutex_lock(&dso__data_open_lock);
 481	close_dso(dso);
 482	pthread_mutex_unlock(&dso__data_open_lock);
 483}
 484
 485static void try_to_open_dso(struct dso *dso, struct machine *machine)
 486{
 487	enum dso_binary_type binary_type_data[] = {
 488		DSO_BINARY_TYPE__BUILD_ID_CACHE,
 489		DSO_BINARY_TYPE__SYSTEM_PATH_DSO,
 490		DSO_BINARY_TYPE__NOT_FOUND,
 491	};
 492	int i = 0;
 
 493
 494	if (dso->data.fd >= 0)
 495		return;
 496
 497	if (dso->binary_type != DSO_BINARY_TYPE__NOT_FOUND) {
 498		dso->data.fd = open_dso(dso, machine);
 499		goto out;
 500	}
 501
 502	do {
 503		dso->binary_type = binary_type_data[i++];
 504
 505		dso->data.fd = open_dso(dso, machine);
 506		if (dso->data.fd >= 0)
 507			goto out;
 508
 509	} while (dso->binary_type != DSO_BINARY_TYPE__NOT_FOUND);
 510out:
 511	if (dso->data.fd >= 0)
 512		dso->data.status = DSO_DATA_STATUS_OK;
 513	else
 514		dso->data.status = DSO_DATA_STATUS_ERROR;
 515}
 516
 517/**
 518 * dso__data_get_fd - Get dso's data file descriptor
 519 * @dso: dso object
 520 * @machine: machine object
 521 *
 522 * External interface to find dso's file, open it and
 523 * returns file descriptor.  It should be paired with
 524 * dso__data_put_fd() if it returns non-negative value.
 525 */
 526int dso__data_get_fd(struct dso *dso, struct machine *machine)
 527{
 528	if (dso->data.status == DSO_DATA_STATUS_ERROR)
 529		return -1;
 530
 531	if (pthread_mutex_lock(&dso__data_open_lock) < 0)
 532		return -1;
 533
 534	try_to_open_dso(dso, machine);
 535
 536	if (dso->data.fd < 0)
 537		pthread_mutex_unlock(&dso__data_open_lock);
 538
 539	return dso->data.fd;
 540}
 541
 542void dso__data_put_fd(struct dso *dso __maybe_unused)
 543{
 544	pthread_mutex_unlock(&dso__data_open_lock);
 545}
 546
 547bool dso__data_status_seen(struct dso *dso, enum dso_data_status_seen by)
 548{
 549	u32 flag = 1 << by;
 550
 551	if (dso->data.status_seen & flag)
 552		return true;
 553
 554	dso->data.status_seen |= flag;
 555
 556	return false;
 557}
 558
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 559static void
 560dso_cache__free(struct dso *dso)
 561{
 562	struct rb_root *root = &dso->data.cache;
 563	struct rb_node *next = rb_first(root);
 564
 565	pthread_mutex_lock(&dso->lock);
 566	while (next) {
 567		struct dso_cache *cache;
 568
 569		cache = rb_entry(next, struct dso_cache, rb_node);
 570		next = rb_next(&cache->rb_node);
 571		rb_erase(&cache->rb_node, root);
 572		free(cache);
 573	}
 574	pthread_mutex_unlock(&dso->lock);
 575}
 576
 577static struct dso_cache *dso_cache__find(struct dso *dso, u64 offset)
 578{
 579	const struct rb_root *root = &dso->data.cache;
 580	struct rb_node * const *p = &root->rb_node;
 581	const struct rb_node *parent = NULL;
 582	struct dso_cache *cache;
 583
 584	while (*p != NULL) {
 585		u64 end;
 586
 587		parent = *p;
 588		cache = rb_entry(parent, struct dso_cache, rb_node);
 589		end = cache->offset + DSO__DATA_CACHE_SIZE;
 590
 591		if (offset < cache->offset)
 592			p = &(*p)->rb_left;
 593		else if (offset >= end)
 594			p = &(*p)->rb_right;
 595		else
 596			return cache;
 597	}
 598
 599	return NULL;
 600}
 601
 602static struct dso_cache *
 603dso_cache__insert(struct dso *dso, struct dso_cache *new)
 604{
 605	struct rb_root *root = &dso->data.cache;
 606	struct rb_node **p = &root->rb_node;
 607	struct rb_node *parent = NULL;
 608	struct dso_cache *cache;
 609	u64 offset = new->offset;
 610
 611	pthread_mutex_lock(&dso->lock);
 612	while (*p != NULL) {
 613		u64 end;
 614
 615		parent = *p;
 616		cache = rb_entry(parent, struct dso_cache, rb_node);
 617		end = cache->offset + DSO__DATA_CACHE_SIZE;
 618
 619		if (offset < cache->offset)
 620			p = &(*p)->rb_left;
 621		else if (offset >= end)
 622			p = &(*p)->rb_right;
 623		else
 624			goto out;
 625	}
 626
 627	rb_link_node(&new->rb_node, parent, p);
 628	rb_insert_color(&new->rb_node, root);
 629
 630	cache = NULL;
 631out:
 632	pthread_mutex_unlock(&dso->lock);
 633	return cache;
 634}
 635
 636static ssize_t
 637dso_cache__memcpy(struct dso_cache *cache, u64 offset,
 638		  u8 *data, u64 size)
 639{
 640	u64 cache_offset = offset - cache->offset;
 641	u64 cache_size   = min(cache->size - cache_offset, size);
 642
 643	memcpy(data, cache->data + cache_offset, cache_size);
 
 
 
 644	return cache_size;
 645}
 646
 647static ssize_t
 648dso_cache__read(struct dso *dso, struct machine *machine,
 649		u64 offset, u8 *data, ssize_t size)
 650{
 651	struct dso_cache *cache;
 652	struct dso_cache *old;
 653	ssize_t ret;
 654
 655	do {
 656		u64 cache_offset;
 657
 658		cache = zalloc(sizeof(*cache) + DSO__DATA_CACHE_SIZE);
 659		if (!cache)
 660			return -ENOMEM;
 
 
 661
 662		pthread_mutex_lock(&dso__data_open_lock);
 
 
 
 
 663
 664		/*
 665		 * dso->data.fd might be closed if other thread opened another
 666		 * file (dso) due to open file limit (RLIMIT_NOFILE).
 667		 */
 668		try_to_open_dso(dso, machine);
 669
 670		if (dso->data.fd < 0) {
 671			ret = -errno;
 672			dso->data.status = DSO_DATA_STATUS_ERROR;
 673			break;
 674		}
 
 
 675
 676		cache_offset = offset & DSO__DATA_CACHE_MASK;
 
 
 
 
 
 
 
 
 
 
 
 
 
 677
 678		ret = pread(dso->data.fd, cache->data, DSO__DATA_CACHE_SIZE, cache_offset);
 679		if (ret <= 0)
 680			break;
 
 681
 682		cache->offset = cache_offset;
 683		cache->size   = ret;
 684	} while (0);
 685
 686	pthread_mutex_unlock(&dso__data_open_lock);
 
 
 
 
 
 687
 688	if (ret > 0) {
 689		old = dso_cache__insert(dso, cache);
 690		if (old) {
 691			/* we lose the race */
 692			free(cache);
 693			cache = old;
 694		}
 695
 696		ret = dso_cache__memcpy(cache, offset, data, size);
 697	}
 
 
 
 
 698
 699	if (ret <= 0)
 700		free(cache);
 701
 702	return ret;
 703}
 704
 705static ssize_t dso_cache_read(struct dso *dso, struct machine *machine,
 706			      u64 offset, u8 *data, ssize_t size)
 707{
 708	struct dso_cache *cache;
 
 
 
 
 
 709
 710	cache = dso_cache__find(dso, offset);
 711	if (cache)
 712		return dso_cache__memcpy(cache, offset, data, size);
 713	else
 714		return dso_cache__read(dso, machine, offset, data, size);
 715}
 716
 717/*
 718 * Reads and caches dso data DSO__DATA_CACHE_SIZE size chunks
 719 * in the rb_tree. Any read to already cached data is served
 720 * by cached data.
 721 */
 722static ssize_t cached_read(struct dso *dso, struct machine *machine,
 723			   u64 offset, u8 *data, ssize_t size)
 724{
 725	ssize_t r = 0;
 726	u8 *p = data;
 727
 728	do {
 729		ssize_t ret;
 730
 731		ret = dso_cache_read(dso, machine, offset, p, size);
 732		if (ret < 0)
 733			return ret;
 734
 735		/* Reached EOF, return what we have. */
 736		if (!ret)
 737			break;
 738
 739		BUG_ON(ret > size);
 740
 741		r      += ret;
 742		p      += ret;
 743		offset += ret;
 744		size   -= ret;
 745
 746	} while (size);
 747
 748	return r;
 749}
 750
 751static int data_file_size(struct dso *dso, struct machine *machine)
 752{
 753	int ret = 0;
 754	struct stat st;
 755	char sbuf[STRERR_BUFSIZE];
 756
 757	if (dso->data.file_size)
 758		return 0;
 759
 760	if (dso->data.status == DSO_DATA_STATUS_ERROR)
 761		return -1;
 762
 763	pthread_mutex_lock(&dso__data_open_lock);
 764
 765	/*
 766	 * dso->data.fd might be closed if other thread opened another
 767	 * file (dso) due to open file limit (RLIMIT_NOFILE).
 768	 */
 769	try_to_open_dso(dso, machine);
 770
 771	if (dso->data.fd < 0) {
 772		ret = -errno;
 773		dso->data.status = DSO_DATA_STATUS_ERROR;
 774		goto out;
 775	}
 776
 777	if (fstat(dso->data.fd, &st) < 0) {
 778		ret = -errno;
 779		pr_err("dso cache fstat failed: %s\n",
 780		       strerror_r(errno, sbuf, sizeof(sbuf)));
 781		dso->data.status = DSO_DATA_STATUS_ERROR;
 782		goto out;
 783	}
 784	dso->data.file_size = st.st_size;
 785
 786out:
 787	pthread_mutex_unlock(&dso__data_open_lock);
 788	return ret;
 789}
 790
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 791/**
 792 * dso__data_size - Return dso data size
 793 * @dso: dso object
 794 * @machine: machine object
 795 *
 796 * Return: dso data size
 797 */
 798off_t dso__data_size(struct dso *dso, struct machine *machine)
 799{
 800	if (data_file_size(dso, machine))
 801		return -1;
 802
 803	/* For now just estimate dso data size is close to file size */
 804	return dso->data.file_size;
 805}
 806
 807static ssize_t data_read_offset(struct dso *dso, struct machine *machine,
 808				u64 offset, u8 *data, ssize_t size)
 
 809{
 810	if (data_file_size(dso, machine))
 811		return -1;
 812
 813	/* Check the offset sanity. */
 814	if (offset > dso->data.file_size)
 815		return -1;
 816
 817	if (offset + size < offset)
 818		return -1;
 819
 820	return cached_read(dso, machine, offset, data, size);
 821}
 822
 823/**
 824 * dso__data_read_offset - Read data from dso file offset
 825 * @dso: dso object
 826 * @machine: machine object
 827 * @offset: file offset
 828 * @data: buffer to store data
 829 * @size: size of the @data buffer
 830 *
 831 * External interface to read data from dso file offset. Open
 832 * dso data file and use cached_read to get the data.
 833 */
 834ssize_t dso__data_read_offset(struct dso *dso, struct machine *machine,
 835			      u64 offset, u8 *data, ssize_t size)
 836{
 837	if (dso->data.status == DSO_DATA_STATUS_ERROR)
 838		return -1;
 839
 840	return data_read_offset(dso, machine, offset, data, size);
 841}
 842
 843/**
 844 * dso__data_read_addr - Read data from dso address
 845 * @dso: dso object
 846 * @machine: machine object
 847 * @add: virtual memory address
 848 * @data: buffer to store data
 849 * @size: size of the @data buffer
 850 *
 851 * External interface to read data from dso address.
 852 */
 853ssize_t dso__data_read_addr(struct dso *dso, struct map *map,
 854			    struct machine *machine, u64 addr,
 855			    u8 *data, ssize_t size)
 856{
 857	u64 offset = map->map_ip(map, addr);
 
 858	return dso__data_read_offset(dso, machine, offset, data, size);
 859}
 860
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 861struct map *dso__new_map(const char *name)
 862{
 863	struct map *map = NULL;
 864	struct dso *dso = dso__new(name);
 865
 866	if (dso)
 867		map = map__new2(0, dso, MAP__FUNCTION);
 
 
 868
 869	return map;
 870}
 871
 872struct dso *machine__findnew_kernel(struct machine *machine, const char *name,
 873				    const char *short_name, int dso_type)
 874{
 875	/*
 876	 * The kernel dso could be created by build_id processing.
 877	 */
 878	struct dso *dso = machine__findnew_dso(machine, name);
 879
 880	/*
 881	 * We need to run this in all cases, since during the build_id
 882	 * processing we had no idea this was the kernel dso.
 883	 */
 884	if (dso != NULL) {
 885		dso__set_short_name(dso, short_name, false);
 886		dso->kernel = dso_type;
 887	}
 888
 889	return dso;
 890}
 891
 892/*
 893 * Find a matching entry and/or link current entry to RB tree.
 894 * Either one of the dso or name parameter must be non-NULL or the
 895 * function will not work.
 896 */
 897static struct dso *__dso__findlink_by_longname(struct rb_root *root,
 898					       struct dso *dso, const char *name)
 899{
 900	struct rb_node **p = &root->rb_node;
 901	struct rb_node  *parent = NULL;
 902
 903	if (!name)
 904		name = dso->long_name;
 905	/*
 906	 * Find node with the matching name
 907	 */
 908	while (*p) {
 909		struct dso *this = rb_entry(*p, struct dso, rb_node);
 910		int rc = strcmp(name, this->long_name);
 911
 912		parent = *p;
 913		if (rc == 0) {
 914			/*
 915			 * In case the new DSO is a duplicate of an existing
 916			 * one, print an one-time warning & put the new entry
 917			 * at the end of the list of duplicates.
 918			 */
 919			if (!dso || (dso == this))
 920				return this;	/* Find matching dso */
 921			/*
 922			 * The core kernel DSOs may have duplicated long name.
 923			 * In this case, the short name should be different.
 924			 * Comparing the short names to differentiate the DSOs.
 925			 */
 926			rc = strcmp(dso->short_name, this->short_name);
 927			if (rc == 0) {
 928				pr_err("Duplicated dso name: %s\n", name);
 929				return NULL;
 930			}
 931		}
 932		if (rc < 0)
 933			p = &parent->rb_left;
 934		else
 935			p = &parent->rb_right;
 936	}
 937	if (dso) {
 938		/* Add new node and rebalance tree */
 939		rb_link_node(&dso->rb_node, parent, p);
 940		rb_insert_color(&dso->rb_node, root);
 941		dso->root = root;
 
 
 
 
 
 
 942	}
 943	return NULL;
 944}
 945
 946static inline struct dso *__dso__find_by_longname(struct rb_root *root,
 947						  const char *name)
 948{
 949	return __dso__findlink_by_longname(root, NULL, name);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 950}
 951
 952void dso__set_long_name(struct dso *dso, const char *name, bool name_allocated)
 953{
 954	struct rb_root *root = dso->root;
 
 955
 956	if (name == NULL)
 957		return;
 958
 959	if (dso->long_name_allocated)
 960		free((char *)dso->long_name);
 
 
 961
 962	if (root) {
 963		rb_erase(&dso->rb_node, root);
 964		/*
 965		 * __dso__findlink_by_longname() isn't guaranteed to add it
 966		 * back, so a clean removal is required here.
 967		 */
 968		RB_CLEAR_NODE(&dso->rb_node);
 969		dso->root = NULL;
 970	}
 971
 972	dso->long_name		 = name;
 973	dso->long_name_len	 = strlen(name);
 974	dso->long_name_allocated = name_allocated;
 
 975
 976	if (root)
 977		__dso__findlink_by_longname(root, dso, NULL);
 978}
 979
 980void dso__set_short_name(struct dso *dso, const char *name, bool name_allocated)
 981{
 982	if (name == NULL)
 983		return;
 
 
 
 
 984
 985	if (dso->short_name_allocated)
 986		free((char *)dso->short_name);
 987
 988	dso->short_name		  = name;
 989	dso->short_name_len	  = strlen(name);
 990	dso->short_name_allocated = name_allocated;
 991}
 992
 993static void dso__set_basename(struct dso *dso)
 994{
 995       /*
 996        * basename() may modify path buffer, so we must pass
 997        * a copy.
 998        */
 999       char *base, *lname = strdup(dso->long_name);
1000
1001       if (!lname)
1002               return;
 
1003
1004       /*
1005        * basename() may return a pointer to internal
1006        * storage which is reused in subsequent calls
1007        * so copy the result.
1008        */
1009       base = strdup(basename(lname));
1010
1011       free(lname);
 
 
 
 
 
 
 
 
1012
1013       if (!base)
1014               return;
1015
1016       dso__set_short_name(dso, base, true);
 
 
 
 
1017}
1018
1019int dso__name_len(const struct dso *dso)
1020{
1021	if (!dso)
1022		return strlen("[unknown]");
1023	if (verbose)
1024		return dso->long_name_len;
1025
1026	return dso->short_name_len;
1027}
1028
1029bool dso__loaded(const struct dso *dso, enum map_type type)
1030{
1031	return dso->loaded & (1 << type);
1032}
1033
1034bool dso__sorted_by_name(const struct dso *dso, enum map_type type)
1035{
1036	return dso->sorted_by_name & (1 << type);
1037}
1038
1039void dso__set_sorted_by_name(struct dso *dso, enum map_type type)
1040{
1041	dso->sorted_by_name |= (1 << type);
1042}
1043
1044struct dso *dso__new(const char *name)
1045{
1046	struct dso *dso = calloc(1, sizeof(*dso) + strlen(name) + 1);
 
 
 
 
 
1047
1048	if (dso != NULL) {
1049		int i;
1050		strcpy(dso->name, name);
1051		dso__set_long_name(dso, dso->name, false);
1052		dso__set_short_name(dso, dso->name, false);
1053		for (i = 0; i < MAP__NR_TYPES; ++i)
1054			dso->symbols[i] = dso->symbol_names[i] = RB_ROOT;
1055		dso->data.cache = RB_ROOT;
 
 
 
 
 
 
1056		dso->data.fd = -1;
1057		dso->data.status = DSO_DATA_STATUS_UNKNOWN;
1058		dso->symtab_type = DSO_BINARY_TYPE__NOT_FOUND;
1059		dso->binary_type = DSO_BINARY_TYPE__NOT_FOUND;
1060		dso->is_64_bit = (sizeof(void *) == 8);
1061		dso->loaded = 0;
1062		dso->rel = 0;
1063		dso->sorted_by_name = 0;
1064		dso->has_build_id = 0;
1065		dso->has_srcline = 1;
1066		dso->a2l_fails = 1;
1067		dso->kernel = DSO_TYPE_USER;
 
1068		dso->needs_swap = DSO_SWAP__UNSET;
1069		RB_CLEAR_NODE(&dso->rb_node);
1070		dso->root = NULL;
1071		INIT_LIST_HEAD(&dso->node);
1072		INIT_LIST_HEAD(&dso->data.open_entry);
1073		pthread_mutex_init(&dso->lock, NULL);
1074		atomic_set(&dso->refcnt, 1);
 
 
 
 
 
1075	}
 
 
1076
1077	return dso;
 
 
1078}
1079
1080void dso__delete(struct dso *dso)
1081{
1082	int i;
 
1083
1084	if (!RB_EMPTY_NODE(&dso->rb_node))
1085		pr_err("DSO %s is still in rbtree when being deleted!\n",
1086		       dso->long_name);
1087	for (i = 0; i < MAP__NR_TYPES; ++i)
1088		symbols__delete(&dso->symbols[i]);
1089
1090	if (dso->short_name_allocated) {
1091		zfree((char **)&dso->short_name);
1092		dso->short_name_allocated = false;
1093	}
1094
1095	if (dso->long_name_allocated) {
1096		zfree((char **)&dso->long_name);
1097		dso->long_name_allocated = false;
 
 
 
1098	}
1099
1100	dso__data_close(dso);
1101	auxtrace_cache__free(dso->auxtrace_cache);
1102	dso_cache__free(dso);
1103	dso__free_a2l(dso);
1104	zfree(&dso->symsrc_filename);
1105	pthread_mutex_destroy(&dso->lock);
1106	free(dso);
 
1107}
1108
1109struct dso *dso__get(struct dso *dso)
1110{
1111	if (dso)
1112		atomic_inc(&dso->refcnt);
1113	return dso;
 
 
 
1114}
1115
1116void dso__put(struct dso *dso)
1117{
1118	if (dso && atomic_dec_and_test(&dso->refcnt))
1119		dso__delete(dso);
 
 
1120}
1121
1122void dso__set_build_id(struct dso *dso, void *build_id)
1123{
1124	memcpy(dso->build_id, build_id, sizeof(dso->build_id));
1125	dso->has_build_id = 1;
1126}
1127
1128bool dso__build_id_equal(const struct dso *dso, u8 *build_id)
1129{
1130	return memcmp(dso->build_id, build_id, sizeof(dso->build_id)) == 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
1131}
1132
1133void dso__read_running_kernel_build_id(struct dso *dso, struct machine *machine)
1134{
1135	char path[PATH_MAX];
1136
1137	if (machine__is_default_guest(machine))
1138		return;
1139	sprintf(path, "%s/sys/kernel/notes", machine->root_dir);
1140	if (sysfs__read_build_id(path, dso->build_id,
1141				 sizeof(dso->build_id)) == 0)
1142		dso->has_build_id = true;
1143}
1144
1145int dso__kernel_module_get_build_id(struct dso *dso,
1146				    const char *root_dir)
1147{
1148	char filename[PATH_MAX];
1149	/*
1150	 * kernel module short names are of the form "[module]" and
1151	 * we need just "module" here.
1152	 */
1153	const char *name = dso->short_name + 1;
1154
1155	snprintf(filename, sizeof(filename),
1156		 "%s/sys/module/%.*s/notes/.note.gnu.build-id",
1157		 root_dir, (int)strlen(name) - 1, name);
1158
1159	if (sysfs__read_build_id(filename, dso->build_id,
1160				 sizeof(dso->build_id)) == 0)
1161		dso->has_build_id = true;
1162
1163	return 0;
1164}
1165
1166bool __dsos__read_build_ids(struct list_head *head, bool with_hits)
1167{
1168	bool have_build_id = false;
1169	struct dso *pos;
1170
1171	list_for_each_entry(pos, head, node) {
1172		if (with_hits && !pos->hit)
1173			continue;
1174		if (pos->has_build_id) {
1175			have_build_id = true;
1176			continue;
1177		}
1178		if (filename__read_build_id(pos->long_name, pos->build_id,
1179					    sizeof(pos->build_id)) > 0) {
1180			have_build_id	  = true;
1181			pos->has_build_id = true;
1182		}
1183	}
1184
1185	return have_build_id;
1186}
1187
1188void __dsos__add(struct dsos *dsos, struct dso *dso)
1189{
1190	list_add_tail(&dso->node, &dsos->head);
1191	__dso__findlink_by_longname(&dsos->root, dso, NULL);
1192	/*
1193	 * It is now in the linked list, grab a reference, then garbage collect
1194	 * this when needing memory, by looking at LRU dso instances in the
1195	 * list with atomic_read(&dso->refcnt) == 1, i.e. no references
1196	 * anywhere besides the one for the list, do, under a lock for the
1197	 * list: remove it from the list, then a dso__put(), that probably will
1198	 * be the last and will then call dso__delete(), end of life.
1199	 *
1200	 * That, or at the end of the 'struct machine' lifetime, when all
1201	 * 'struct dso' instances will be removed from the list, in
1202	 * dsos__exit(), if they have no other reference from some other data
1203	 * structure.
1204	 *
1205	 * E.g.: after processing a 'perf.data' file and storing references
1206	 * to objects instantiated while processing events, we will have
1207	 * references to the 'thread', 'map', 'dso' structs all from 'struct
1208	 * hist_entry' instances, but we may not need anything not referenced,
1209	 * so we might as well call machines__exit()/machines__delete() and
1210	 * garbage collect it.
1211	 */
1212	dso__get(dso);
1213}
1214
1215void dsos__add(struct dsos *dsos, struct dso *dso)
1216{
1217	pthread_rwlock_wrlock(&dsos->lock);
1218	__dsos__add(dsos, dso);
1219	pthread_rwlock_unlock(&dsos->lock);
1220}
1221
1222struct dso *__dsos__find(struct dsos *dsos, const char *name, bool cmp_short)
1223{
1224	struct dso *pos;
1225
1226	if (cmp_short) {
1227		list_for_each_entry(pos, &dsos->head, node)
1228			if (strcmp(pos->short_name, name) == 0)
1229				return pos;
1230		return NULL;
1231	}
1232	return __dso__find_by_longname(&dsos->root, name);
1233}
1234
1235struct dso *dsos__find(struct dsos *dsos, const char *name, bool cmp_short)
1236{
1237	struct dso *dso;
1238	pthread_rwlock_rdlock(&dsos->lock);
1239	dso = __dsos__find(dsos, name, cmp_short);
1240	pthread_rwlock_unlock(&dsos->lock);
1241	return dso;
1242}
1243
1244struct dso *__dsos__addnew(struct dsos *dsos, const char *name)
1245{
1246	struct dso *dso = dso__new(name);
1247
1248	if (dso != NULL) {
1249		__dsos__add(dsos, dso);
1250		dso__set_basename(dso);
1251		/* Put dso here because __dsos_add already got it */
1252		dso__put(dso);
1253	}
1254	return dso;
1255}
1256
1257struct dso *__dsos__findnew(struct dsos *dsos, const char *name)
1258{
1259	struct dso *dso = __dsos__find(dsos, name, false);
1260
1261	return dso ? dso : __dsos__addnew(dsos, name);
1262}
1263
1264struct dso *dsos__findnew(struct dsos *dsos, const char *name)
1265{
1266	struct dso *dso;
1267	pthread_rwlock_wrlock(&dsos->lock);
1268	dso = dso__get(__dsos__findnew(dsos, name));
1269	pthread_rwlock_unlock(&dsos->lock);
1270	return dso;
1271}
1272
1273size_t __dsos__fprintf_buildid(struct list_head *head, FILE *fp,
1274			       bool (skip)(struct dso *dso, int parm), int parm)
1275{
1276	struct dso *pos;
1277	size_t ret = 0;
1278
1279	list_for_each_entry(pos, head, node) {
1280		if (skip && skip(pos, parm))
1281			continue;
1282		ret += dso__fprintf_buildid(pos, fp);
1283		ret += fprintf(fp, " %s\n", pos->long_name);
1284	}
1285	return ret;
1286}
1287
1288size_t __dsos__fprintf(struct list_head *head, FILE *fp)
1289{
1290	struct dso *pos;
1291	size_t ret = 0;
1292
1293	list_for_each_entry(pos, head, node) {
1294		int i;
1295		for (i = 0; i < MAP__NR_TYPES; ++i)
1296			ret += dso__fprintf(pos, i, fp);
1297	}
1298
1299	return ret;
1300}
1301
1302size_t dso__fprintf_buildid(struct dso *dso, FILE *fp)
1303{
1304	char sbuild_id[BUILD_ID_SIZE * 2 + 1];
1305
1306	build_id__sprintf(dso->build_id, sizeof(dso->build_id), sbuild_id);
1307	return fprintf(fp, "%s", sbuild_id);
1308}
1309
1310size_t dso__fprintf(struct dso *dso, enum map_type type, FILE *fp)
1311{
1312	struct rb_node *nd;
1313	size_t ret = fprintf(fp, "dso: %s (", dso->short_name);
1314
1315	if (dso->short_name != dso->long_name)
1316		ret += fprintf(fp, "%s, ", dso->long_name);
1317	ret += fprintf(fp, "%s, %sloaded, ", map_type__name[type],
1318		       dso__loaded(dso, type) ? "" : "NOT ");
1319	ret += dso__fprintf_buildid(dso, fp);
1320	ret += fprintf(fp, ")\n");
1321	for (nd = rb_first(&dso->symbols[type]); nd; nd = rb_next(nd)) {
1322		struct symbol *pos = rb_entry(nd, struct symbol, rb_node);
1323		ret += symbol__fprintf(pos, fp);
1324	}
1325
1326	return ret;
1327}
1328
1329enum dso_type dso__type(struct dso *dso, struct machine *machine)
1330{
1331	int fd;
1332	enum dso_type type = DSO__TYPE_UNKNOWN;
1333
1334	fd = dso__data_get_fd(dso, machine);
1335	if (fd >= 0) {
1336		type = dso__type_fd(fd);
1337		dso__data_put_fd(dso);
1338	}
1339
1340	return type;
1341}
1342
1343int dso__strerror_load(struct dso *dso, char *buf, size_t buflen)
1344{
1345	int idx, errnum = dso->load_errno;
1346	/*
1347	 * This must have a same ordering as the enum dso_load_errno.
1348	 */
1349	static const char *dso_load__error_str[] = {
1350	"Internal tools/perf/ library error",
1351	"Invalid ELF file",
1352	"Can not read build id",
1353	"Mismatching build id",
1354	"Decompression failure",
1355	};
1356
1357	BUG_ON(buflen == 0);
1358
1359	if (errnum >= 0) {
1360		const char *err = strerror_r(errnum, buf, buflen);
1361
1362		if (err != buf)
1363			scnprintf(buf, buflen, "%s", err);
1364
1365		return 0;
1366	}
1367
1368	if (errnum <  __DSO_LOAD_ERRNO__START || errnum >= __DSO_LOAD_ERRNO__END)
1369		return -1;
1370
1371	idx = errnum - __DSO_LOAD_ERRNO__START;
1372	scnprintf(buf, buflen, "%s", dso_load__error_str[idx]);
1373	return 0;
 
 
 
 
 
 
 
 
 
 
 
 
1374}