Linux Audio

Check our new training course

Buildroot integration, development and maintenance

Need a Buildroot system for your embedded project?
Loading...
v6.8
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (C) 2009-2011, Frederic Weisbecker <fweisbec@gmail.com>
   4 *
   5 * Handle the callchains from the stream in an ad-hoc radix tree and then
   6 * sort them in an rbtree.
   7 *
   8 * Using a radix for code path provides a fast retrieval and factorizes
   9 * memory use. Also that lets us use the paths in a hierarchical graph view.
  10 *
  11 */
  12
  13#include <inttypes.h>
  14#include <stdlib.h>
  15#include <stdio.h>
  16#include <stdbool.h>
  17#include <errno.h>
  18#include <math.h>
  19#include <linux/string.h>
  20#include <linux/zalloc.h>
  21
  22#include "asm/bug.h"
  23
  24#include "debug.h"
  25#include "dso.h"
  26#include "event.h"
  27#include "hist.h"
  28#include "sort.h"
  29#include "machine.h"
  30#include "map.h"
  31#include "callchain.h"
  32#include "branch.h"
  33#include "symbol.h"
  34#include "util.h"
  35#include "../perf.h"
  36
  37#define CALLCHAIN_PARAM_DEFAULT			\
  38	.mode		= CHAIN_GRAPH_ABS,	\
  39	.min_percent	= 0.5,			\
  40	.order		= ORDER_CALLEE,		\
  41	.key		= CCKEY_FUNCTION,	\
  42	.value		= CCVAL_PERCENT,	\
  43
  44struct callchain_param callchain_param = {
  45	CALLCHAIN_PARAM_DEFAULT
  46};
  47
  48/*
  49 * Are there any events usind DWARF callchains?
  50 *
  51 * I.e.
  52 *
  53 * -e cycles/call-graph=dwarf/
  54 */
  55bool dwarf_callchain_users;
  56
  57struct callchain_param callchain_param_default = {
  58	CALLCHAIN_PARAM_DEFAULT
  59};
  60
  61/* Used for thread-local struct callchain_cursor. */
  62static pthread_key_t callchain_cursor;
  63
  64int parse_callchain_record_opt(const char *arg, struct callchain_param *param)
  65{
  66	return parse_callchain_record(arg, param);
  67}
  68
  69static int parse_callchain_mode(const char *value)
  70{
  71	if (!strncmp(value, "graph", strlen(value))) {
  72		callchain_param.mode = CHAIN_GRAPH_ABS;
  73		return 0;
  74	}
  75	if (!strncmp(value, "flat", strlen(value))) {
  76		callchain_param.mode = CHAIN_FLAT;
  77		return 0;
  78	}
  79	if (!strncmp(value, "fractal", strlen(value))) {
  80		callchain_param.mode = CHAIN_GRAPH_REL;
  81		return 0;
  82	}
  83	if (!strncmp(value, "folded", strlen(value))) {
  84		callchain_param.mode = CHAIN_FOLDED;
  85		return 0;
  86	}
  87	return -1;
  88}
  89
  90static int parse_callchain_order(const char *value)
  91{
  92	if (!strncmp(value, "caller", strlen(value))) {
  93		callchain_param.order = ORDER_CALLER;
  94		callchain_param.order_set = true;
  95		return 0;
  96	}
  97	if (!strncmp(value, "callee", strlen(value))) {
  98		callchain_param.order = ORDER_CALLEE;
  99		callchain_param.order_set = true;
 100		return 0;
 101	}
 102	return -1;
 103}
 104
 105static int parse_callchain_sort_key(const char *value)
 106{
 107	if (!strncmp(value, "function", strlen(value))) {
 108		callchain_param.key = CCKEY_FUNCTION;
 109		return 0;
 110	}
 111	if (!strncmp(value, "address", strlen(value))) {
 112		callchain_param.key = CCKEY_ADDRESS;
 113		return 0;
 114	}
 115	if (!strncmp(value, "srcline", strlen(value))) {
 116		callchain_param.key = CCKEY_SRCLINE;
 117		return 0;
 118	}
 119	if (!strncmp(value, "branch", strlen(value))) {
 120		callchain_param.branch_callstack = 1;
 121		return 0;
 122	}
 123	return -1;
 124}
 125
 126static int parse_callchain_value(const char *value)
 127{
 128	if (!strncmp(value, "percent", strlen(value))) {
 129		callchain_param.value = CCVAL_PERCENT;
 130		return 0;
 131	}
 132	if (!strncmp(value, "period", strlen(value))) {
 133		callchain_param.value = CCVAL_PERIOD;
 134		return 0;
 135	}
 136	if (!strncmp(value, "count", strlen(value))) {
 137		callchain_param.value = CCVAL_COUNT;
 138		return 0;
 139	}
 140	return -1;
 141}
 142
 143static int get_stack_size(const char *str, unsigned long *_size)
 144{
 145	char *endptr;
 146	unsigned long size;
 147	unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));
 148
 149	size = strtoul(str, &endptr, 0);
 150
 151	do {
 152		if (*endptr)
 153			break;
 154
 155		size = round_up(size, sizeof(u64));
 156		if (!size || size > max_size)
 157			break;
 158
 159		*_size = size;
 160		return 0;
 161
 162	} while (0);
 163
 164	pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
 165	       max_size, str);
 166	return -1;
 167}
 168
 169static int
 170__parse_callchain_report_opt(const char *arg, bool allow_record_opt)
 171{
 172	char *tok;
 173	char *endptr, *saveptr = NULL;
 174	bool minpcnt_set = false;
 175	bool record_opt_set = false;
 176	bool try_stack_size = false;
 177
 178	callchain_param.enabled = true;
 179	symbol_conf.use_callchain = true;
 180
 181	if (!arg)
 182		return 0;
 183
 184	while ((tok = strtok_r((char *)arg, ",", &saveptr)) != NULL) {
 185		if (!strncmp(tok, "none", strlen(tok))) {
 186			callchain_param.mode = CHAIN_NONE;
 187			callchain_param.enabled = false;
 188			symbol_conf.use_callchain = false;
 189			return 0;
 190		}
 191
 192		if (!parse_callchain_mode(tok) ||
 193		    !parse_callchain_order(tok) ||
 194		    !parse_callchain_sort_key(tok) ||
 195		    !parse_callchain_value(tok)) {
 196			/* parsing ok - move on to the next */
 197			try_stack_size = false;
 198			goto next;
 199		} else if (allow_record_opt && !record_opt_set) {
 200			if (parse_callchain_record(tok, &callchain_param))
 201				goto try_numbers;
 202
 203			/* assume that number followed by 'dwarf' is stack size */
 204			if (callchain_param.record_mode == CALLCHAIN_DWARF)
 205				try_stack_size = true;
 206
 207			record_opt_set = true;
 208			goto next;
 209		}
 210
 211try_numbers:
 212		if (try_stack_size) {
 213			unsigned long size = 0;
 214
 215			if (get_stack_size(tok, &size) < 0)
 216				return -1;
 217			callchain_param.dump_size = size;
 218			try_stack_size = false;
 219		} else if (!minpcnt_set) {
 220			/* try to get the min percent */
 221			callchain_param.min_percent = strtod(tok, &endptr);
 222			if (tok == endptr)
 223				return -1;
 224			minpcnt_set = true;
 225		} else {
 226			/* try print limit at last */
 227			callchain_param.print_limit = strtoul(tok, &endptr, 0);
 228			if (tok == endptr)
 229				return -1;
 230		}
 231next:
 232		arg = NULL;
 233	}
 234
 235	if (callchain_register_param(&callchain_param) < 0) {
 236		pr_err("Can't register callchain params\n");
 237		return -1;
 238	}
 239	return 0;
 240}
 241
 242int parse_callchain_report_opt(const char *arg)
 243{
 244	return __parse_callchain_report_opt(arg, false);
 245}
 246
 247int parse_callchain_top_opt(const char *arg)
 248{
 249	return __parse_callchain_report_opt(arg, true);
 250}
 251
 252int parse_callchain_record(const char *arg, struct callchain_param *param)
 253{
 254	char *tok, *name, *saveptr = NULL;
 255	char *buf;
 256	int ret = -1;
 257
 258	/* We need buffer that we know we can write to. */
 259	buf = malloc(strlen(arg) + 1);
 260	if (!buf)
 261		return -ENOMEM;
 262
 263	strcpy(buf, arg);
 264
 265	tok = strtok_r((char *)buf, ",", &saveptr);
 266	name = tok ? : (char *)buf;
 267
 268	do {
 269		/* Framepointer style */
 270		if (!strncmp(name, "fp", sizeof("fp"))) {
 271			ret = 0;
 272			param->record_mode = CALLCHAIN_FP;
 273
 274			tok = strtok_r(NULL, ",", &saveptr);
 275			if (tok) {
 276				unsigned long size;
 277
 278				size = strtoul(tok, &name, 0);
 279				if (size < (unsigned) sysctl__max_stack())
 280					param->max_stack = size;
 281			}
 282			break;
 283
 284		/* Dwarf style */
 285		} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
 286			const unsigned long default_stack_dump_size = 8192;
 287
 288			ret = 0;
 289			param->record_mode = CALLCHAIN_DWARF;
 290			param->dump_size = default_stack_dump_size;
 291			dwarf_callchain_users = true;
 292
 293			tok = strtok_r(NULL, ",", &saveptr);
 294			if (tok) {
 295				unsigned long size = 0;
 296
 297				ret = get_stack_size(tok, &size);
 298				param->dump_size = size;
 299			}
 300		} else if (!strncmp(name, "lbr", sizeof("lbr"))) {
 301			if (!strtok_r(NULL, ",", &saveptr)) {
 302				param->record_mode = CALLCHAIN_LBR;
 303				ret = 0;
 304			} else
 305				pr_err("callchain: No more arguments "
 306					"needed for --call-graph lbr\n");
 307			break;
 308		} else {
 309			pr_err("callchain: Unknown --call-graph option "
 310			       "value: %s\n", arg);
 311			break;
 312		}
 313
 314	} while (0);
 315
 316	free(buf);
 317	return ret;
 318}
 319
 320int perf_callchain_config(const char *var, const char *value)
 321{
 322	char *endptr;
 323
 324	if (!strstarts(var, "call-graph."))
 325		return 0;
 326	var += sizeof("call-graph.") - 1;
 327
 328	if (!strcmp(var, "record-mode"))
 329		return parse_callchain_record_opt(value, &callchain_param);
 330	if (!strcmp(var, "dump-size")) {
 331		unsigned long size = 0;
 332		int ret;
 333
 334		ret = get_stack_size(value, &size);
 335		callchain_param.dump_size = size;
 336
 337		return ret;
 338	}
 339	if (!strcmp(var, "print-type")){
 340		int ret;
 341		ret = parse_callchain_mode(value);
 342		if (ret == -1)
 343			pr_err("Invalid callchain mode: %s\n", value);
 344		return ret;
 345	}
 346	if (!strcmp(var, "order")){
 347		int ret;
 348		ret = parse_callchain_order(value);
 349		if (ret == -1)
 350			pr_err("Invalid callchain order: %s\n", value);
 351		return ret;
 352	}
 353	if (!strcmp(var, "sort-key")){
 354		int ret;
 355		ret = parse_callchain_sort_key(value);
 356		if (ret == -1)
 357			pr_err("Invalid callchain sort key: %s\n", value);
 358		return ret;
 359	}
 360	if (!strcmp(var, "threshold")) {
 361		callchain_param.min_percent = strtod(value, &endptr);
 362		if (value == endptr) {
 363			pr_err("Invalid callchain threshold: %s\n", value);
 364			return -1;
 365		}
 366	}
 367	if (!strcmp(var, "print-limit")) {
 368		callchain_param.print_limit = strtod(value, &endptr);
 369		if (value == endptr) {
 370			pr_err("Invalid callchain print limit: %s\n", value);
 371			return -1;
 372		}
 373	}
 374
 375	return 0;
 376}
 377
 378static void
 379rb_insert_callchain(struct rb_root *root, struct callchain_node *chain,
 380		    enum chain_mode mode)
 381{
 382	struct rb_node **p = &root->rb_node;
 383	struct rb_node *parent = NULL;
 384	struct callchain_node *rnode;
 385	u64 chain_cumul = callchain_cumul_hits(chain);
 386
 387	while (*p) {
 388		u64 rnode_cumul;
 389
 390		parent = *p;
 391		rnode = rb_entry(parent, struct callchain_node, rb_node);
 392		rnode_cumul = callchain_cumul_hits(rnode);
 393
 394		switch (mode) {
 395		case CHAIN_FLAT:
 396		case CHAIN_FOLDED:
 397			if (rnode->hit < chain->hit)
 398				p = &(*p)->rb_left;
 399			else
 400				p = &(*p)->rb_right;
 401			break;
 402		case CHAIN_GRAPH_ABS: /* Falldown */
 403		case CHAIN_GRAPH_REL:
 404			if (rnode_cumul < chain_cumul)
 405				p = &(*p)->rb_left;
 406			else
 407				p = &(*p)->rb_right;
 408			break;
 409		case CHAIN_NONE:
 410		default:
 411			break;
 412		}
 413	}
 414
 415	rb_link_node(&chain->rb_node, parent, p);
 416	rb_insert_color(&chain->rb_node, root);
 417}
 418
 419static void
 420__sort_chain_flat(struct rb_root *rb_root, struct callchain_node *node,
 421		  u64 min_hit)
 422{
 423	struct rb_node *n;
 424	struct callchain_node *child;
 425
 426	n = rb_first(&node->rb_root_in);
 427	while (n) {
 428		child = rb_entry(n, struct callchain_node, rb_node_in);
 429		n = rb_next(n);
 430
 431		__sort_chain_flat(rb_root, child, min_hit);
 432	}
 433
 434	if (node->hit && node->hit >= min_hit)
 435		rb_insert_callchain(rb_root, node, CHAIN_FLAT);
 436}
 437
 438/*
 439 * Once we get every callchains from the stream, we can now
 440 * sort them by hit
 441 */
 442static void
 443sort_chain_flat(struct rb_root *rb_root, struct callchain_root *root,
 444		u64 min_hit, struct callchain_param *param __maybe_unused)
 445{
 446	*rb_root = RB_ROOT;
 447	__sort_chain_flat(rb_root, &root->node, min_hit);
 448}
 449
 450static void __sort_chain_graph_abs(struct callchain_node *node,
 451				   u64 min_hit)
 452{
 453	struct rb_node *n;
 454	struct callchain_node *child;
 455
 456	node->rb_root = RB_ROOT;
 457	n = rb_first(&node->rb_root_in);
 458
 459	while (n) {
 460		child = rb_entry(n, struct callchain_node, rb_node_in);
 461		n = rb_next(n);
 462
 463		__sort_chain_graph_abs(child, min_hit);
 464		if (callchain_cumul_hits(child) >= min_hit)
 465			rb_insert_callchain(&node->rb_root, child,
 466					    CHAIN_GRAPH_ABS);
 467	}
 468}
 469
 470static void
 471sort_chain_graph_abs(struct rb_root *rb_root, struct callchain_root *chain_root,
 472		     u64 min_hit, struct callchain_param *param __maybe_unused)
 473{
 474	__sort_chain_graph_abs(&chain_root->node, min_hit);
 475	rb_root->rb_node = chain_root->node.rb_root.rb_node;
 476}
 477
 478static void __sort_chain_graph_rel(struct callchain_node *node,
 479				   double min_percent)
 480{
 481	struct rb_node *n;
 482	struct callchain_node *child;
 483	u64 min_hit;
 484
 485	node->rb_root = RB_ROOT;
 486	min_hit = ceil(node->children_hit * min_percent);
 487
 488	n = rb_first(&node->rb_root_in);
 489	while (n) {
 490		child = rb_entry(n, struct callchain_node, rb_node_in);
 491		n = rb_next(n);
 492
 493		__sort_chain_graph_rel(child, min_percent);
 494		if (callchain_cumul_hits(child) >= min_hit)
 495			rb_insert_callchain(&node->rb_root, child,
 496					    CHAIN_GRAPH_REL);
 497	}
 498}
 499
 500static void
 501sort_chain_graph_rel(struct rb_root *rb_root, struct callchain_root *chain_root,
 502		     u64 min_hit __maybe_unused, struct callchain_param *param)
 503{
 504	__sort_chain_graph_rel(&chain_root->node, param->min_percent / 100.0);
 505	rb_root->rb_node = chain_root->node.rb_root.rb_node;
 506}
 507
 508int callchain_register_param(struct callchain_param *param)
 509{
 510	switch (param->mode) {
 511	case CHAIN_GRAPH_ABS:
 512		param->sort = sort_chain_graph_abs;
 513		break;
 514	case CHAIN_GRAPH_REL:
 515		param->sort = sort_chain_graph_rel;
 516		break;
 517	case CHAIN_FLAT:
 518	case CHAIN_FOLDED:
 519		param->sort = sort_chain_flat;
 520		break;
 521	case CHAIN_NONE:
 522	default:
 523		return -1;
 524	}
 525	return 0;
 526}
 527
 528/*
 529 * Create a child for a parent. If inherit_children, then the new child
 530 * will become the new parent of it's parent children
 531 */
 532static struct callchain_node *
 533create_child(struct callchain_node *parent, bool inherit_children)
 534{
 535	struct callchain_node *new;
 536
 537	new = zalloc(sizeof(*new));
 538	if (!new) {
 539		perror("not enough memory to create child for code path tree");
 540		return NULL;
 541	}
 542	new->parent = parent;
 543	INIT_LIST_HEAD(&new->val);
 544	INIT_LIST_HEAD(&new->parent_val);
 545
 546	if (inherit_children) {
 547		struct rb_node *n;
 548		struct callchain_node *child;
 549
 550		new->rb_root_in = parent->rb_root_in;
 551		parent->rb_root_in = RB_ROOT;
 552
 553		n = rb_first(&new->rb_root_in);
 554		while (n) {
 555			child = rb_entry(n, struct callchain_node, rb_node_in);
 556			child->parent = new;
 557			n = rb_next(n);
 558		}
 559
 560		/* make it the first child */
 561		rb_link_node(&new->rb_node_in, NULL, &parent->rb_root_in.rb_node);
 562		rb_insert_color(&new->rb_node_in, &parent->rb_root_in);
 563	}
 564
 565	return new;
 566}
 567
 568
 569/*
 570 * Fill the node with callchain values
 571 */
 572static int
 573fill_node(struct callchain_node *node, struct callchain_cursor *cursor)
 574{
 575	struct callchain_cursor_node *cursor_node;
 576
 577	node->val_nr = cursor->nr - cursor->pos;
 578	if (!node->val_nr)
 579		pr_warning("Warning: empty node in callchain tree\n");
 580
 581	cursor_node = callchain_cursor_current(cursor);
 582
 583	while (cursor_node) {
 584		struct callchain_list *call;
 585
 586		call = zalloc(sizeof(*call));
 587		if (!call) {
 588			perror("not enough memory for the code path tree");
 589			return -ENOMEM;
 590		}
 591		call->ip = cursor_node->ip;
 592		call->ms = cursor_node->ms;
 593		call->ms.map = map__get(call->ms.map);
 594		call->ms.maps = maps__get(call->ms.maps);
 595		call->srcline = cursor_node->srcline;
 596
 597		if (cursor_node->branch) {
 598			call->branch_count = 1;
 599
 600			if (cursor_node->branch_from) {
 601				/*
 602				 * branch_from is set with value somewhere else
 603				 * to imply it's "to" of a branch.
 604				 */
 605				if (!call->brtype_stat) {
 606					call->brtype_stat = zalloc(sizeof(*call->brtype_stat));
 607					if (!call->brtype_stat) {
 608						perror("not enough memory for the code path branch statistics");
 609						free(call->brtype_stat);
 610						return -ENOMEM;
 611					}
 612				}
 613				call->brtype_stat->branch_to = true;
 614
 615				if (cursor_node->branch_flags.predicted)
 616					call->predicted_count = 1;
 617
 618				if (cursor_node->branch_flags.abort)
 619					call->abort_count = 1;
 620
 621				branch_type_count(call->brtype_stat,
 622						  &cursor_node->branch_flags,
 623						  cursor_node->branch_from,
 624						  cursor_node->ip);
 625			} else {
 626				/*
 627				 * It's "from" of a branch
 628				 */
 629				if (call->brtype_stat && call->brtype_stat->branch_to)
 630					call->brtype_stat->branch_to = false;
 631				call->cycles_count =
 632					cursor_node->branch_flags.cycles;
 633				call->iter_count = cursor_node->nr_loop_iter;
 634				call->iter_cycles = cursor_node->iter_cycles;
 635			}
 636		}
 637
 638		list_add_tail(&call->list, &node->val);
 639
 640		callchain_cursor_advance(cursor);
 641		cursor_node = callchain_cursor_current(cursor);
 642	}
 643	return 0;
 644}
 645
 646static struct callchain_node *
 647add_child(struct callchain_node *parent,
 648	  struct callchain_cursor *cursor,
 649	  u64 period)
 650{
 651	struct callchain_node *new;
 652
 653	new = create_child(parent, false);
 654	if (new == NULL)
 655		return NULL;
 656
 657	if (fill_node(new, cursor) < 0) {
 658		struct callchain_list *call, *tmp;
 659
 660		list_for_each_entry_safe(call, tmp, &new->val, list) {
 661			list_del_init(&call->list);
 662			map_symbol__exit(&call->ms);
 663			zfree(&call->brtype_stat);
 664			free(call);
 665		}
 666		free(new);
 667		return NULL;
 668	}
 669
 670	new->children_hit = 0;
 671	new->hit = period;
 672	new->children_count = 0;
 673	new->count = 1;
 674	return new;
 675}
 676
 677enum match_result {
 678	MATCH_ERROR  = -1,
 679	MATCH_EQ,
 680	MATCH_LT,
 681	MATCH_GT,
 682};
 683
 684static enum match_result match_chain_strings(const char *left,
 685					     const char *right)
 686{
 687	enum match_result ret = MATCH_EQ;
 688	int cmp;
 689
 690	if (left && right)
 691		cmp = strcmp(left, right);
 692	else if (!left && right)
 693		cmp = 1;
 694	else if (left && !right)
 695		cmp = -1;
 696	else
 697		return MATCH_ERROR;
 698
 699	if (cmp != 0)
 700		ret = cmp < 0 ? MATCH_LT : MATCH_GT;
 701
 702	return ret;
 703}
 704
 705/*
 706 * We need to always use relative addresses because we're aggregating
 707 * callchains from multiple threads, i.e. different address spaces, so
 708 * comparing absolute addresses make no sense as a symbol in a DSO may end up
 709 * in a different address when used in a different binary or even the same
 710 * binary but with some sort of address randomization technique, thus we need
 711 * to compare just relative addresses. -acme
 712 */
 713static enum match_result match_chain_dso_addresses(struct map *left_map, u64 left_ip,
 714						   struct map *right_map, u64 right_ip)
 715{
 716	struct dso *left_dso = left_map ? map__dso(left_map) : NULL;
 717	struct dso *right_dso = right_map ? map__dso(right_map) : NULL;
 718
 719	if (left_dso != right_dso)
 720		return left_dso < right_dso ? MATCH_LT : MATCH_GT;
 721
 722	if (left_ip != right_ip)
 723 		return left_ip < right_ip ? MATCH_LT : MATCH_GT;
 724
 725	return MATCH_EQ;
 726}
 727
 728static enum match_result match_chain(struct callchain_cursor_node *node,
 729				     struct callchain_list *cnode)
 730{
 731	enum match_result match = MATCH_ERROR;
 732
 733	switch (callchain_param.key) {
 734	case CCKEY_SRCLINE:
 735		match = match_chain_strings(cnode->srcline, node->srcline);
 736		if (match != MATCH_ERROR)
 737			break;
 738		/* otherwise fall-back to symbol-based comparison below */
 739		fallthrough;
 740	case CCKEY_FUNCTION:
 741		if (node->ms.sym && cnode->ms.sym) {
 742			/*
 743			 * Compare inlined frames based on their symbol name
 744			 * because different inlined frames will have the same
 745			 * symbol start. Otherwise do a faster comparison based
 746			 * on the symbol start address.
 747			 */
 748			if (cnode->ms.sym->inlined || node->ms.sym->inlined) {
 749				match = match_chain_strings(cnode->ms.sym->name,
 750							    node->ms.sym->name);
 751				if (match != MATCH_ERROR)
 752					break;
 753			} else {
 754				match = match_chain_dso_addresses(cnode->ms.map, cnode->ms.sym->start,
 755								  node->ms.map, node->ms.sym->start);
 756				break;
 757			}
 758		}
 759		/* otherwise fall-back to IP-based comparison below */
 760		fallthrough;
 761	case CCKEY_ADDRESS:
 762	default:
 763		match = match_chain_dso_addresses(cnode->ms.map, cnode->ip, node->ms.map, node->ip);
 764		break;
 765	}
 766
 767	if (match == MATCH_EQ && node->branch) {
 768		cnode->branch_count++;
 769
 770		if (node->branch_from) {
 771			/*
 772			 * It's "to" of a branch
 773			 */
 774			if (!cnode->brtype_stat) {
 775				cnode->brtype_stat = zalloc(sizeof(*cnode->brtype_stat));
 776				if (!cnode->brtype_stat) {
 777					perror("not enough memory for the code path branch statistics");
 778					return MATCH_ERROR;
 779				}
 780			}
 781			cnode->brtype_stat->branch_to = true;
 782
 783			if (node->branch_flags.predicted)
 784				cnode->predicted_count++;
 785
 786			if (node->branch_flags.abort)
 787				cnode->abort_count++;
 788
 789			branch_type_count(cnode->brtype_stat,
 790					  &node->branch_flags,
 791					  node->branch_from,
 792					  node->ip);
 793		} else {
 794			/*
 795			 * It's "from" of a branch
 796			 */
 797			if (cnode->brtype_stat && cnode->brtype_stat->branch_to)
 798				cnode->brtype_stat->branch_to = false;
 799			cnode->cycles_count += node->branch_flags.cycles;
 800			cnode->iter_count += node->nr_loop_iter;
 801			cnode->iter_cycles += node->iter_cycles;
 802			cnode->from_count++;
 803		}
 804	}
 805
 806	return match;
 807}
 808
 809/*
 810 * Split the parent in two parts (a new child is created) and
 811 * give a part of its callchain to the created child.
 812 * Then create another child to host the given callchain of new branch
 813 */
 814static int
 815split_add_child(struct callchain_node *parent,
 816		struct callchain_cursor *cursor,
 817		struct callchain_list *to_split,
 818		u64 idx_parents, u64 idx_local, u64 period)
 819{
 820	struct callchain_node *new;
 821	struct list_head *old_tail;
 822	unsigned int idx_total = idx_parents + idx_local;
 823
 824	/* split */
 825	new = create_child(parent, true);
 826	if (new == NULL)
 827		return -1;
 828
 829	/* split the callchain and move a part to the new child */
 830	old_tail = parent->val.prev;
 831	list_del_range(&to_split->list, old_tail);
 832	new->val.next = &to_split->list;
 833	new->val.prev = old_tail;
 834	to_split->list.prev = &new->val;
 835	old_tail->next = &new->val;
 836
 837	/* split the hits */
 838	new->hit = parent->hit;
 839	new->children_hit = parent->children_hit;
 840	parent->children_hit = callchain_cumul_hits(new);
 841	new->val_nr = parent->val_nr - idx_local;
 842	parent->val_nr = idx_local;
 843	new->count = parent->count;
 844	new->children_count = parent->children_count;
 845	parent->children_count = callchain_cumul_counts(new);
 846
 847	/* create a new child for the new branch if any */
 848	if (idx_total < cursor->nr) {
 849		struct callchain_node *first;
 850		struct callchain_list *cnode;
 851		struct callchain_cursor_node *node;
 852		struct rb_node *p, **pp;
 853
 854		parent->hit = 0;
 855		parent->children_hit += period;
 856		parent->count = 0;
 857		parent->children_count += 1;
 858
 859		node = callchain_cursor_current(cursor);
 860		new = add_child(parent, cursor, period);
 861		if (new == NULL)
 862			return -1;
 863
 864		/*
 865		 * This is second child since we moved parent's children
 866		 * to new (first) child above.
 867		 */
 868		p = parent->rb_root_in.rb_node;
 869		first = rb_entry(p, struct callchain_node, rb_node_in);
 870		cnode = list_first_entry(&first->val, struct callchain_list,
 871					 list);
 872
 873		if (match_chain(node, cnode) == MATCH_LT)
 874			pp = &p->rb_left;
 875		else
 876			pp = &p->rb_right;
 877
 878		rb_link_node(&new->rb_node_in, p, pp);
 879		rb_insert_color(&new->rb_node_in, &parent->rb_root_in);
 880	} else {
 881		parent->hit = period;
 882		parent->count = 1;
 883	}
 884	return 0;
 885}
 886
 887static enum match_result
 888append_chain(struct callchain_node *root,
 889	     struct callchain_cursor *cursor,
 890	     u64 period);
 891
 892static int
 893append_chain_children(struct callchain_node *root,
 894		      struct callchain_cursor *cursor,
 895		      u64 period)
 896{
 897	struct callchain_node *rnode;
 898	struct callchain_cursor_node *node;
 899	struct rb_node **p = &root->rb_root_in.rb_node;
 900	struct rb_node *parent = NULL;
 901
 902	node = callchain_cursor_current(cursor);
 903	if (!node)
 904		return -1;
 905
 906	/* lookup in children */
 907	while (*p) {
 908		enum match_result ret;
 909
 910		parent = *p;
 911		rnode = rb_entry(parent, struct callchain_node, rb_node_in);
 912
 913		/* If at least first entry matches, rely to children */
 914		ret = append_chain(rnode, cursor, period);
 915		if (ret == MATCH_EQ)
 916			goto inc_children_hit;
 917		if (ret == MATCH_ERROR)
 918			return -1;
 919
 920		if (ret == MATCH_LT)
 921			p = &parent->rb_left;
 922		else
 923			p = &parent->rb_right;
 924	}
 925	/* nothing in children, add to the current node */
 926	rnode = add_child(root, cursor, period);
 927	if (rnode == NULL)
 928		return -1;
 929
 930	rb_link_node(&rnode->rb_node_in, parent, p);
 931	rb_insert_color(&rnode->rb_node_in, &root->rb_root_in);
 932
 933inc_children_hit:
 934	root->children_hit += period;
 935	root->children_count++;
 936	return 0;
 937}
 938
 939static enum match_result
 940append_chain(struct callchain_node *root,
 941	     struct callchain_cursor *cursor,
 942	     u64 period)
 943{
 944	struct callchain_list *cnode;
 945	u64 start = cursor->pos;
 946	bool found = false;
 947	u64 matches;
 948	enum match_result cmp = MATCH_ERROR;
 949
 950	/*
 951	 * Lookup in the current node
 952	 * If we have a symbol, then compare the start to match
 953	 * anywhere inside a function, unless function
 954	 * mode is disabled.
 955	 */
 956	list_for_each_entry(cnode, &root->val, list) {
 957		struct callchain_cursor_node *node;
 958
 959		node = callchain_cursor_current(cursor);
 960		if (!node)
 961			break;
 962
 963		cmp = match_chain(node, cnode);
 964		if (cmp != MATCH_EQ)
 965			break;
 966
 967		found = true;
 968
 969		callchain_cursor_advance(cursor);
 970	}
 971
 972	/* matches not, relay no the parent */
 973	if (!found) {
 974		WARN_ONCE(cmp == MATCH_ERROR, "Chain comparison error\n");
 975		return cmp;
 976	}
 977
 978	matches = cursor->pos - start;
 979
 980	/* we match only a part of the node. Split it and add the new chain */
 981	if (matches < root->val_nr) {
 982		if (split_add_child(root, cursor, cnode, start, matches,
 983				    period) < 0)
 984			return MATCH_ERROR;
 985
 986		return MATCH_EQ;
 987	}
 988
 989	/* we match 100% of the path, increment the hit */
 990	if (matches == root->val_nr && cursor->pos == cursor->nr) {
 991		root->hit += period;
 992		root->count++;
 993		return MATCH_EQ;
 994	}
 995
 996	/* We match the node and still have a part remaining */
 997	if (append_chain_children(root, cursor, period) < 0)
 998		return MATCH_ERROR;
 999
1000	return MATCH_EQ;
1001}
1002
1003int callchain_append(struct callchain_root *root,
1004		     struct callchain_cursor *cursor,
1005		     u64 period)
1006{
1007	if (cursor == NULL)
1008		return -1;
1009
1010	if (!cursor->nr)
1011		return 0;
1012
1013	callchain_cursor_commit(cursor);
1014
1015	if (append_chain_children(&root->node, cursor, period) < 0)
1016		return -1;
1017
1018	if (cursor->nr > root->max_depth)
1019		root->max_depth = cursor->nr;
1020
1021	return 0;
1022}
1023
1024static int
1025merge_chain_branch(struct callchain_cursor *cursor,
1026		   struct callchain_node *dst, struct callchain_node *src)
1027{
1028	struct callchain_cursor_node **old_last = cursor->last;
1029	struct callchain_node *child;
1030	struct callchain_list *list, *next_list;
1031	struct rb_node *n;
1032	int old_pos = cursor->nr;
1033	int err = 0;
1034
1035	list_for_each_entry_safe(list, next_list, &src->val, list) {
1036		struct map_symbol ms = {
1037			.maps = maps__get(list->ms.maps),
1038			.map = map__get(list->ms.map),
1039		};
1040		callchain_cursor_append(cursor, list->ip, &ms, false, NULL, 0, 0, 0, list->srcline);
1041		list_del_init(&list->list);
1042		map_symbol__exit(&ms);
1043		map_symbol__exit(&list->ms);
1044		zfree(&list->brtype_stat);
1045		free(list);
1046	}
1047
1048	if (src->hit) {
1049		callchain_cursor_commit(cursor);
1050		if (append_chain_children(dst, cursor, src->hit) < 0)
1051			return -1;
1052	}
1053
1054	n = rb_first(&src->rb_root_in);
1055	while (n) {
1056		child = container_of(n, struct callchain_node, rb_node_in);
1057		n = rb_next(n);
1058		rb_erase(&child->rb_node_in, &src->rb_root_in);
1059
1060		err = merge_chain_branch(cursor, dst, child);
1061		if (err)
1062			break;
1063
1064		free(child);
1065	}
1066
1067	cursor->nr = old_pos;
1068	cursor->last = old_last;
1069
1070	return err;
1071}
1072
1073int callchain_merge(struct callchain_cursor *cursor,
1074		    struct callchain_root *dst, struct callchain_root *src)
1075{
1076	return merge_chain_branch(cursor, &dst->node, &src->node);
1077}
1078
1079int callchain_cursor_append(struct callchain_cursor *cursor,
1080			    u64 ip, struct map_symbol *ms,
1081			    bool branch, struct branch_flags *flags,
1082			    int nr_loop_iter, u64 iter_cycles, u64 branch_from,
1083			    const char *srcline)
1084{
1085	struct callchain_cursor_node *node = *cursor->last;
1086
1087	if (!node) {
1088		node = calloc(1, sizeof(*node));
1089		if (!node)
1090			return -ENOMEM;
1091
1092		*cursor->last = node;
1093	}
1094
1095	node->ip = ip;
1096	map_symbol__exit(&node->ms);
1097	node->ms = *ms;
1098	node->ms.maps = maps__get(ms->maps);
1099	node->ms.map = map__get(ms->map);
1100	node->branch = branch;
1101	node->nr_loop_iter = nr_loop_iter;
1102	node->iter_cycles = iter_cycles;
1103	node->srcline = srcline;
1104
1105	if (flags)
1106		memcpy(&node->branch_flags, flags,
1107			sizeof(struct branch_flags));
1108
1109	node->branch_from = branch_from;
1110	cursor->nr++;
1111
1112	cursor->last = &node->next;
1113
1114	return 0;
1115}
1116
1117int sample__resolve_callchain(struct perf_sample *sample,
1118			      struct callchain_cursor *cursor, struct symbol **parent,
1119			      struct evsel *evsel, struct addr_location *al,
1120			      int max_stack)
1121{
1122	if (sample->callchain == NULL && !symbol_conf.show_branchflag_count)
1123		return 0;
1124
1125	if (symbol_conf.use_callchain || symbol_conf.cumulate_callchain ||
1126	    perf_hpp_list.parent || symbol_conf.show_branchflag_count) {
1127		return thread__resolve_callchain(al->thread, cursor, evsel, sample,
1128						 parent, al, max_stack);
1129	}
1130	return 0;
1131}
1132
1133int hist_entry__append_callchain(struct hist_entry *he, struct perf_sample *sample)
1134{
1135	if ((!symbol_conf.use_callchain || sample->callchain == NULL) &&
1136		!symbol_conf.show_branchflag_count)
1137		return 0;
1138	return callchain_append(he->callchain, get_tls_callchain_cursor(), sample->period);
1139}
1140
1141int fill_callchain_info(struct addr_location *al, struct callchain_cursor_node *node,
1142			bool hide_unresolved)
1143{
1144	struct machine *machine = maps__machine(node->ms.maps);
1145
1146	maps__put(al->maps);
1147	al->maps = maps__get(node->ms.maps);
1148	map__put(al->map);
1149	al->map = map__get(node->ms.map);
1150	al->sym = node->ms.sym;
1151	al->srcline = node->srcline;
1152	al->addr = node->ip;
1153
1154	if (al->sym == NULL) {
1155		if (hide_unresolved)
1156			return 0;
1157		if (al->map == NULL)
1158			goto out;
1159	}
1160	if (RC_CHK_EQUAL(al->maps, machine__kernel_maps(machine))) {
1161		if (machine__is_host(machine)) {
 
1162			al->cpumode = PERF_RECORD_MISC_KERNEL;
1163			al->level = 'k';
1164		} else {
1165			al->cpumode = PERF_RECORD_MISC_GUEST_KERNEL;
1166			al->level = 'g';
1167		}
1168	} else {
1169		if (machine__is_host(machine)) {
1170			al->cpumode = PERF_RECORD_MISC_USER;
1171			al->level = '.';
1172		} else if (perf_guest) {
1173			al->cpumode = PERF_RECORD_MISC_GUEST_USER;
1174			al->level = 'u';
1175		} else {
1176			al->cpumode = PERF_RECORD_MISC_HYPERVISOR;
1177			al->level = 'H';
1178		}
1179	}
1180
1181out:
1182	return 1;
1183}
1184
1185char *callchain_list__sym_name(struct callchain_list *cl,
1186			       char *bf, size_t bfsize, bool show_dso)
1187{
1188	bool show_addr = callchain_param.key == CCKEY_ADDRESS;
1189	bool show_srcline = show_addr || callchain_param.key == CCKEY_SRCLINE;
1190	int printed;
1191
1192	if (cl->ms.sym) {
1193		const char *inlined = cl->ms.sym->inlined ? " (inlined)" : "";
1194
1195		if (show_srcline && cl->srcline)
1196			printed = scnprintf(bf, bfsize, "%s %s%s",
1197					    cl->ms.sym->name, cl->srcline,
1198					    inlined);
1199		else
1200			printed = scnprintf(bf, bfsize, "%s%s",
1201					    cl->ms.sym->name, inlined);
1202	} else
1203		printed = scnprintf(bf, bfsize, "%#" PRIx64, cl->ip);
1204
1205	if (show_dso)
1206		scnprintf(bf + printed, bfsize - printed, " %s",
1207			  cl->ms.map ?
1208			  map__dso(cl->ms.map)->short_name :
1209			  "unknown");
1210
1211	return bf;
1212}
1213
1214char *callchain_node__scnprintf_value(struct callchain_node *node,
1215				      char *bf, size_t bfsize, u64 total)
1216{
1217	double percent = 0.0;
1218	u64 period = callchain_cumul_hits(node);
1219	unsigned count = callchain_cumul_counts(node);
1220
1221	if (callchain_param.mode == CHAIN_FOLDED) {
1222		period = node->hit;
1223		count = node->count;
1224	}
1225
1226	switch (callchain_param.value) {
1227	case CCVAL_PERIOD:
1228		scnprintf(bf, bfsize, "%"PRIu64, period);
1229		break;
1230	case CCVAL_COUNT:
1231		scnprintf(bf, bfsize, "%u", count);
1232		break;
1233	case CCVAL_PERCENT:
1234	default:
1235		if (total)
1236			percent = period * 100.0 / total;
1237		scnprintf(bf, bfsize, "%.2f%%", percent);
1238		break;
1239	}
1240	return bf;
1241}
1242
1243int callchain_node__fprintf_value(struct callchain_node *node,
1244				 FILE *fp, u64 total)
1245{
1246	double percent = 0.0;
1247	u64 period = callchain_cumul_hits(node);
1248	unsigned count = callchain_cumul_counts(node);
1249
1250	if (callchain_param.mode == CHAIN_FOLDED) {
1251		period = node->hit;
1252		count = node->count;
1253	}
1254
1255	switch (callchain_param.value) {
1256	case CCVAL_PERIOD:
1257		return fprintf(fp, "%"PRIu64, period);
1258	case CCVAL_COUNT:
1259		return fprintf(fp, "%u", count);
1260	case CCVAL_PERCENT:
1261	default:
1262		if (total)
1263			percent = period * 100.0 / total;
1264		return percent_color_fprintf(fp, "%.2f%%", percent);
1265	}
1266	return 0;
1267}
1268
1269static void callchain_counts_value(struct callchain_node *node,
1270				   u64 *branch_count, u64 *predicted_count,
1271				   u64 *abort_count, u64 *cycles_count)
1272{
1273	struct callchain_list *clist;
1274
1275	list_for_each_entry(clist, &node->val, list) {
1276		if (branch_count)
1277			*branch_count += clist->branch_count;
1278
1279		if (predicted_count)
1280			*predicted_count += clist->predicted_count;
1281
1282		if (abort_count)
1283			*abort_count += clist->abort_count;
1284
1285		if (cycles_count)
1286			*cycles_count += clist->cycles_count;
1287	}
1288}
1289
1290static int callchain_node_branch_counts_cumul(struct callchain_node *node,
1291					      u64 *branch_count,
1292					      u64 *predicted_count,
1293					      u64 *abort_count,
1294					      u64 *cycles_count)
1295{
1296	struct callchain_node *child;
1297	struct rb_node *n;
1298
1299	n = rb_first(&node->rb_root_in);
1300	while (n) {
1301		child = rb_entry(n, struct callchain_node, rb_node_in);
1302		n = rb_next(n);
1303
1304		callchain_node_branch_counts_cumul(child, branch_count,
1305						   predicted_count,
1306						   abort_count,
1307						   cycles_count);
1308
1309		callchain_counts_value(child, branch_count,
1310				       predicted_count, abort_count,
1311				       cycles_count);
1312	}
1313
1314	return 0;
1315}
1316
1317int callchain_branch_counts(struct callchain_root *root,
1318			    u64 *branch_count, u64 *predicted_count,
1319			    u64 *abort_count, u64 *cycles_count)
1320{
1321	if (branch_count)
1322		*branch_count = 0;
1323
1324	if (predicted_count)
1325		*predicted_count = 0;
1326
1327	if (abort_count)
1328		*abort_count = 0;
1329
1330	if (cycles_count)
1331		*cycles_count = 0;
1332
1333	return callchain_node_branch_counts_cumul(&root->node,
1334						  branch_count,
1335						  predicted_count,
1336						  abort_count,
1337						  cycles_count);
1338}
1339
1340static int count_pri64_printf(int idx, const char *str, u64 value, char *bf, int bfsize)
1341{
1342	return scnprintf(bf, bfsize, "%s%s:%" PRId64 "", (idx) ? " " : " (", str, value);
1343}
1344
1345static int count_float_printf(int idx, const char *str, float value,
1346			      char *bf, int bfsize, float threshold)
1347{
1348	if (threshold != 0.0 && value < threshold)
1349		return 0;
1350
1351	return scnprintf(bf, bfsize, "%s%s:%.1f%%", (idx) ? " " : " (", str, value);
1352}
1353
1354static int branch_to_str(char *bf, int bfsize,
1355			 u64 branch_count, u64 predicted_count,
1356			 u64 abort_count,
1357			 const struct branch_type_stat *brtype_stat)
1358{
1359	int printed, i = 0;
1360
1361	printed = branch_type_str(brtype_stat, bf, bfsize);
1362	if (printed)
1363		i++;
1364
1365	if (predicted_count < branch_count) {
1366		printed += count_float_printf(i++, "predicted",
1367				predicted_count * 100.0 / branch_count,
1368				bf + printed, bfsize - printed, 0.0);
1369	}
1370
1371	if (abort_count) {
1372		printed += count_float_printf(i++, "abort",
1373				abort_count * 100.0 / branch_count,
1374				bf + printed, bfsize - printed, 0.1);
1375	}
1376
1377	if (i)
1378		printed += scnprintf(bf + printed, bfsize - printed, ")");
1379
1380	return printed;
1381}
1382
1383static int branch_from_str(char *bf, int bfsize,
1384			   u64 branch_count,
1385			   u64 cycles_count, u64 iter_count,
1386			   u64 iter_cycles, u64 from_count)
1387{
1388	int printed = 0, i = 0;
1389	u64 cycles, v = 0;
1390
1391	cycles = cycles_count / branch_count;
1392	if (cycles) {
1393		printed += count_pri64_printf(i++, "cycles",
1394				cycles,
1395				bf + printed, bfsize - printed);
1396	}
1397
1398	if (iter_count && from_count) {
1399		v = iter_count / from_count;
1400		if (v) {
1401			printed += count_pri64_printf(i++, "iter",
1402					v, bf + printed, bfsize - printed);
1403
1404			printed += count_pri64_printf(i++, "avg_cycles",
1405					iter_cycles / iter_count,
1406					bf + printed, bfsize - printed);
1407		}
1408	}
1409
1410	if (i)
1411		printed += scnprintf(bf + printed, bfsize - printed, ")");
1412
1413	return printed;
1414}
1415
1416static int counts_str_build(char *bf, int bfsize,
1417			     u64 branch_count, u64 predicted_count,
1418			     u64 abort_count, u64 cycles_count,
1419			     u64 iter_count, u64 iter_cycles,
1420			     u64 from_count,
1421			     const struct branch_type_stat *brtype_stat)
1422{
1423	int printed;
1424
1425	if (branch_count == 0)
1426		return scnprintf(bf, bfsize, " (calltrace)");
1427
1428	if (brtype_stat->branch_to) {
1429		printed = branch_to_str(bf, bfsize, branch_count,
1430				predicted_count, abort_count, brtype_stat);
1431	} else {
1432		printed = branch_from_str(bf, bfsize, branch_count,
1433				cycles_count, iter_count, iter_cycles,
1434				from_count);
1435	}
1436
1437	if (!printed)
1438		bf[0] = 0;
1439
1440	return printed;
1441}
1442
1443static int callchain_counts_printf(FILE *fp, char *bf, int bfsize,
1444				   u64 branch_count, u64 predicted_count,
1445				   u64 abort_count, u64 cycles_count,
1446				   u64 iter_count, u64 iter_cycles,
1447				   u64 from_count,
1448				   const struct branch_type_stat *brtype_stat)
1449{
1450	char str[256];
1451
1452	counts_str_build(str, sizeof(str), branch_count,
1453			 predicted_count, abort_count, cycles_count,
1454			 iter_count, iter_cycles, from_count, brtype_stat);
1455
1456	if (fp)
1457		return fprintf(fp, "%s", str);
1458
1459	return scnprintf(bf, bfsize, "%s", str);
1460}
1461
1462int callchain_list_counts__printf_value(struct callchain_list *clist,
1463					FILE *fp, char *bf, int bfsize)
1464{
1465	static const struct branch_type_stat empty_brtype_stat = {};
1466	const struct branch_type_stat *brtype_stat;
1467	u64 branch_count, predicted_count;
1468	u64 abort_count, cycles_count;
1469	u64 iter_count, iter_cycles;
1470	u64 from_count;
1471
1472	brtype_stat = clist->brtype_stat ?: &empty_brtype_stat;
1473	branch_count = clist->branch_count;
1474	predicted_count = clist->predicted_count;
1475	abort_count = clist->abort_count;
1476	cycles_count = clist->cycles_count;
1477	iter_count = clist->iter_count;
1478	iter_cycles = clist->iter_cycles;
1479	from_count = clist->from_count;
1480
1481	return callchain_counts_printf(fp, bf, bfsize, branch_count,
1482				       predicted_count, abort_count,
1483				       cycles_count, iter_count, iter_cycles,
1484				       from_count, brtype_stat);
1485}
1486
1487static void free_callchain_node(struct callchain_node *node)
1488{
1489	struct callchain_list *list, *tmp;
1490	struct callchain_node *child;
1491	struct rb_node *n;
1492
1493	list_for_each_entry_safe(list, tmp, &node->parent_val, list) {
1494		list_del_init(&list->list);
1495		map_symbol__exit(&list->ms);
1496		zfree(&list->brtype_stat);
1497		free(list);
1498	}
1499
1500	list_for_each_entry_safe(list, tmp, &node->val, list) {
1501		list_del_init(&list->list);
1502		map_symbol__exit(&list->ms);
1503		zfree(&list->brtype_stat);
1504		free(list);
1505	}
1506
1507	n = rb_first(&node->rb_root_in);
1508	while (n) {
1509		child = container_of(n, struct callchain_node, rb_node_in);
1510		n = rb_next(n);
1511		rb_erase(&child->rb_node_in, &node->rb_root_in);
1512
1513		free_callchain_node(child);
1514		free(child);
1515	}
1516}
1517
1518void free_callchain(struct callchain_root *root)
1519{
1520	if (!symbol_conf.use_callchain)
1521		return;
1522
1523	free_callchain_node(&root->node);
1524}
1525
1526static u64 decay_callchain_node(struct callchain_node *node)
1527{
1528	struct callchain_node *child;
1529	struct rb_node *n;
1530	u64 child_hits = 0;
1531
1532	n = rb_first(&node->rb_root_in);
1533	while (n) {
1534		child = container_of(n, struct callchain_node, rb_node_in);
1535
1536		child_hits += decay_callchain_node(child);
1537		n = rb_next(n);
1538	}
1539
1540	node->hit = (node->hit * 7) / 8;
1541	node->children_hit = child_hits;
1542
1543	return node->hit;
1544}
1545
1546void decay_callchain(struct callchain_root *root)
1547{
1548	if (!symbol_conf.use_callchain)
1549		return;
1550
1551	decay_callchain_node(&root->node);
1552}
1553
1554int callchain_node__make_parent_list(struct callchain_node *node)
1555{
1556	struct callchain_node *parent = node->parent;
1557	struct callchain_list *chain, *new;
1558	LIST_HEAD(head);
1559
1560	while (parent) {
1561		list_for_each_entry_reverse(chain, &parent->val, list) {
1562			new = malloc(sizeof(*new));
1563			if (new == NULL)
1564				goto out;
1565			*new = *chain;
1566			new->has_children = false;
1567			new->ms.map = map__get(new->ms.map);
1568			list_add_tail(&new->list, &head);
1569		}
1570		parent = parent->parent;
1571	}
1572
1573	list_for_each_entry_safe_reverse(chain, new, &head, list)
1574		list_move_tail(&chain->list, &node->parent_val);
1575
1576	if (!list_empty(&node->parent_val)) {
1577		chain = list_first_entry(&node->parent_val, struct callchain_list, list);
1578		chain->has_children = rb_prev(&node->rb_node) || rb_next(&node->rb_node);
1579
1580		chain = list_first_entry(&node->val, struct callchain_list, list);
1581		chain->has_children = false;
1582	}
1583	return 0;
1584
1585out:
1586	list_for_each_entry_safe(chain, new, &head, list) {
1587		list_del_init(&chain->list);
1588		map_symbol__exit(&chain->ms);
1589		zfree(&chain->brtype_stat);
1590		free(chain);
1591	}
1592	return -ENOMEM;
1593}
1594
1595static void callchain_cursor__delete(void *vcursor)
1596{
1597	struct callchain_cursor *cursor = vcursor;
1598	struct callchain_cursor_node *node, *next;
1599
1600	callchain_cursor_reset(cursor);
1601	for (node = cursor->first; node != NULL; node = next) {
1602		next = node->next;
1603		free(node);
1604	}
1605	free(cursor);
1606}
1607
1608static void init_callchain_cursor_key(void)
1609{
1610	if (pthread_key_create(&callchain_cursor, callchain_cursor__delete)) {
1611		pr_err("callchain cursor creation failed");
1612		abort();
1613	}
1614}
1615
1616struct callchain_cursor *get_tls_callchain_cursor(void)
1617{
1618	static pthread_once_t once_control = PTHREAD_ONCE_INIT;
1619	struct callchain_cursor *cursor;
1620
1621	pthread_once(&once_control, init_callchain_cursor_key);
1622	cursor = pthread_getspecific(callchain_cursor);
1623	if (!cursor) {
1624		cursor = zalloc(sizeof(*cursor));
1625		if (!cursor)
1626			pr_debug3("%s: not enough memory\n", __func__);
1627		pthread_setspecific(callchain_cursor, cursor);
1628	}
1629	return cursor;
1630}
1631
1632int callchain_cursor__copy(struct callchain_cursor *dst,
1633			   struct callchain_cursor *src)
1634{
1635	int rc = 0;
1636
1637	callchain_cursor_reset(dst);
1638	callchain_cursor_commit(src);
1639
1640	while (true) {
1641		struct callchain_cursor_node *node;
1642
1643		node = callchain_cursor_current(src);
1644		if (node == NULL)
1645			break;
1646
1647		rc = callchain_cursor_append(dst, node->ip, &node->ms,
1648					     node->branch, &node->branch_flags,
1649					     node->nr_loop_iter,
1650					     node->iter_cycles,
1651					     node->branch_from, node->srcline);
1652		if (rc)
1653			break;
1654
1655		callchain_cursor_advance(src);
1656	}
1657
1658	return rc;
1659}
1660
1661/*
1662 * Initialize a cursor before adding entries inside, but keep
1663 * the previously allocated entries as a cache.
1664 */
1665void callchain_cursor_reset(struct callchain_cursor *cursor)
1666{
1667	struct callchain_cursor_node *node;
1668
1669	cursor->nr = 0;
1670	cursor->last = &cursor->first;
1671
1672	for (node = cursor->first; node != NULL; node = node->next)
1673		map_symbol__exit(&node->ms);
1674}
1675
1676void callchain_param_setup(u64 sample_type, const char *arch)
1677{
1678	if (symbol_conf.use_callchain || symbol_conf.cumulate_callchain) {
1679		if ((sample_type & PERF_SAMPLE_REGS_USER) &&
1680		    (sample_type & PERF_SAMPLE_STACK_USER)) {
1681			callchain_param.record_mode = CALLCHAIN_DWARF;
1682			dwarf_callchain_users = true;
1683		} else if (sample_type & PERF_SAMPLE_BRANCH_STACK)
1684			callchain_param.record_mode = CALLCHAIN_LBR;
1685		else
1686			callchain_param.record_mode = CALLCHAIN_FP;
1687	}
1688
1689	/*
1690	 * It's necessary to use libunwind to reliably determine the caller of
1691	 * a leaf function on aarch64, as otherwise we cannot know whether to
1692	 * start from the LR or FP.
1693	 *
1694	 * Always starting from the LR can result in duplicate or entirely
1695	 * erroneous entries. Always skipping the LR and starting from the FP
1696	 * can result in missing entries.
1697	 */
1698	if (callchain_param.record_mode == CALLCHAIN_FP && !strcmp(arch, "arm64"))
1699		dwarf_callchain_users = true;
1700}
1701
1702static bool chain_match(struct callchain_list *base_chain,
1703			struct callchain_list *pair_chain)
1704{
1705	enum match_result match;
1706
1707	match = match_chain_strings(base_chain->srcline,
1708				    pair_chain->srcline);
1709	if (match != MATCH_ERROR)
1710		return match == MATCH_EQ;
1711
1712	match = match_chain_dso_addresses(base_chain->ms.map,
1713					  base_chain->ip,
1714					  pair_chain->ms.map,
1715					  pair_chain->ip);
1716
1717	return match == MATCH_EQ;
1718}
1719
1720bool callchain_cnode_matched(struct callchain_node *base_cnode,
1721			     struct callchain_node *pair_cnode)
1722{
1723	struct callchain_list *base_chain, *pair_chain;
1724	bool match = false;
1725
1726	pair_chain = list_first_entry(&pair_cnode->val,
1727				      struct callchain_list,
1728				      list);
1729
1730	list_for_each_entry(base_chain, &base_cnode->val, list) {
1731		if (&pair_chain->list == &pair_cnode->val)
1732			return false;
1733
1734		if (!base_chain->srcline || !pair_chain->srcline) {
1735			pair_chain = list_next_entry(pair_chain, list);
1736			continue;
1737		}
1738
1739		match = chain_match(base_chain, pair_chain);
1740		if (!match)
1741			return false;
1742
1743		pair_chain = list_next_entry(pair_chain, list);
1744	}
1745
1746	/*
1747	 * Say chain1 is ABC, chain2 is ABCD, we consider they are
1748	 * not fully matched.
1749	 */
1750	if (pair_chain && (&pair_chain->list != &pair_cnode->val))
1751		return false;
1752
1753	return match;
1754}
1755
1756static u64 count_callchain_hits(struct hist_entry *he)
1757{
1758	struct rb_root *root = &he->sorted_chain;
1759	struct rb_node *rb_node = rb_first(root);
1760	struct callchain_node *node;
1761	u64 chain_hits = 0;
1762
1763	while (rb_node) {
1764		node = rb_entry(rb_node, struct callchain_node, rb_node);
1765		chain_hits += node->hit;
1766		rb_node = rb_next(rb_node);
1767	}
1768
1769	return chain_hits;
1770}
1771
1772u64 callchain_total_hits(struct hists *hists)
1773{
1774	struct rb_node *next = rb_first_cached(&hists->entries);
1775	u64 chain_hits = 0;
1776
1777	while (next) {
1778		struct hist_entry *he = rb_entry(next, struct hist_entry,
1779						 rb_node);
1780
1781		chain_hits += count_callchain_hits(he);
1782		next = rb_next(&he->rb_node);
1783	}
1784
1785	return chain_hits;
1786}
1787
1788s64 callchain_avg_cycles(struct callchain_node *cnode)
1789{
1790	struct callchain_list *chain;
1791	s64 cycles = 0;
1792
1793	list_for_each_entry(chain, &cnode->val, list) {
1794		if (chain->srcline && chain->branch_count)
1795			cycles += chain->cycles_count / chain->branch_count;
1796	}
1797
1798	return cycles;
1799}
v6.2
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (C) 2009-2011, Frederic Weisbecker <fweisbec@gmail.com>
   4 *
   5 * Handle the callchains from the stream in an ad-hoc radix tree and then
   6 * sort them in an rbtree.
   7 *
   8 * Using a radix for code path provides a fast retrieval and factorizes
   9 * memory use. Also that lets us use the paths in a hierarchical graph view.
  10 *
  11 */
  12
  13#include <inttypes.h>
  14#include <stdlib.h>
  15#include <stdio.h>
  16#include <stdbool.h>
  17#include <errno.h>
  18#include <math.h>
  19#include <linux/string.h>
  20#include <linux/zalloc.h>
  21
  22#include "asm/bug.h"
  23
  24#include "debug.h"
  25#include "dso.h"
  26#include "event.h"
  27#include "hist.h"
  28#include "sort.h"
  29#include "machine.h"
  30#include "map.h"
  31#include "callchain.h"
  32#include "branch.h"
  33#include "symbol.h"
  34#include "util.h"
  35#include "../perf.h"
  36
  37#define CALLCHAIN_PARAM_DEFAULT			\
  38	.mode		= CHAIN_GRAPH_ABS,	\
  39	.min_percent	= 0.5,			\
  40	.order		= ORDER_CALLEE,		\
  41	.key		= CCKEY_FUNCTION,	\
  42	.value		= CCVAL_PERCENT,	\
  43
  44struct callchain_param callchain_param = {
  45	CALLCHAIN_PARAM_DEFAULT
  46};
  47
  48/*
  49 * Are there any events usind DWARF callchains?
  50 *
  51 * I.e.
  52 *
  53 * -e cycles/call-graph=dwarf/
  54 */
  55bool dwarf_callchain_users;
  56
  57struct callchain_param callchain_param_default = {
  58	CALLCHAIN_PARAM_DEFAULT
  59};
  60
  61__thread struct callchain_cursor callchain_cursor;
 
  62
  63int parse_callchain_record_opt(const char *arg, struct callchain_param *param)
  64{
  65	return parse_callchain_record(arg, param);
  66}
  67
  68static int parse_callchain_mode(const char *value)
  69{
  70	if (!strncmp(value, "graph", strlen(value))) {
  71		callchain_param.mode = CHAIN_GRAPH_ABS;
  72		return 0;
  73	}
  74	if (!strncmp(value, "flat", strlen(value))) {
  75		callchain_param.mode = CHAIN_FLAT;
  76		return 0;
  77	}
  78	if (!strncmp(value, "fractal", strlen(value))) {
  79		callchain_param.mode = CHAIN_GRAPH_REL;
  80		return 0;
  81	}
  82	if (!strncmp(value, "folded", strlen(value))) {
  83		callchain_param.mode = CHAIN_FOLDED;
  84		return 0;
  85	}
  86	return -1;
  87}
  88
  89static int parse_callchain_order(const char *value)
  90{
  91	if (!strncmp(value, "caller", strlen(value))) {
  92		callchain_param.order = ORDER_CALLER;
  93		callchain_param.order_set = true;
  94		return 0;
  95	}
  96	if (!strncmp(value, "callee", strlen(value))) {
  97		callchain_param.order = ORDER_CALLEE;
  98		callchain_param.order_set = true;
  99		return 0;
 100	}
 101	return -1;
 102}
 103
 104static int parse_callchain_sort_key(const char *value)
 105{
 106	if (!strncmp(value, "function", strlen(value))) {
 107		callchain_param.key = CCKEY_FUNCTION;
 108		return 0;
 109	}
 110	if (!strncmp(value, "address", strlen(value))) {
 111		callchain_param.key = CCKEY_ADDRESS;
 112		return 0;
 113	}
 114	if (!strncmp(value, "srcline", strlen(value))) {
 115		callchain_param.key = CCKEY_SRCLINE;
 116		return 0;
 117	}
 118	if (!strncmp(value, "branch", strlen(value))) {
 119		callchain_param.branch_callstack = 1;
 120		return 0;
 121	}
 122	return -1;
 123}
 124
 125static int parse_callchain_value(const char *value)
 126{
 127	if (!strncmp(value, "percent", strlen(value))) {
 128		callchain_param.value = CCVAL_PERCENT;
 129		return 0;
 130	}
 131	if (!strncmp(value, "period", strlen(value))) {
 132		callchain_param.value = CCVAL_PERIOD;
 133		return 0;
 134	}
 135	if (!strncmp(value, "count", strlen(value))) {
 136		callchain_param.value = CCVAL_COUNT;
 137		return 0;
 138	}
 139	return -1;
 140}
 141
 142static int get_stack_size(const char *str, unsigned long *_size)
 143{
 144	char *endptr;
 145	unsigned long size;
 146	unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));
 147
 148	size = strtoul(str, &endptr, 0);
 149
 150	do {
 151		if (*endptr)
 152			break;
 153
 154		size = round_up(size, sizeof(u64));
 155		if (!size || size > max_size)
 156			break;
 157
 158		*_size = size;
 159		return 0;
 160
 161	} while (0);
 162
 163	pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
 164	       max_size, str);
 165	return -1;
 166}
 167
 168static int
 169__parse_callchain_report_opt(const char *arg, bool allow_record_opt)
 170{
 171	char *tok;
 172	char *endptr, *saveptr = NULL;
 173	bool minpcnt_set = false;
 174	bool record_opt_set = false;
 175	bool try_stack_size = false;
 176
 177	callchain_param.enabled = true;
 178	symbol_conf.use_callchain = true;
 179
 180	if (!arg)
 181		return 0;
 182
 183	while ((tok = strtok_r((char *)arg, ",", &saveptr)) != NULL) {
 184		if (!strncmp(tok, "none", strlen(tok))) {
 185			callchain_param.mode = CHAIN_NONE;
 186			callchain_param.enabled = false;
 187			symbol_conf.use_callchain = false;
 188			return 0;
 189		}
 190
 191		if (!parse_callchain_mode(tok) ||
 192		    !parse_callchain_order(tok) ||
 193		    !parse_callchain_sort_key(tok) ||
 194		    !parse_callchain_value(tok)) {
 195			/* parsing ok - move on to the next */
 196			try_stack_size = false;
 197			goto next;
 198		} else if (allow_record_opt && !record_opt_set) {
 199			if (parse_callchain_record(tok, &callchain_param))
 200				goto try_numbers;
 201
 202			/* assume that number followed by 'dwarf' is stack size */
 203			if (callchain_param.record_mode == CALLCHAIN_DWARF)
 204				try_stack_size = true;
 205
 206			record_opt_set = true;
 207			goto next;
 208		}
 209
 210try_numbers:
 211		if (try_stack_size) {
 212			unsigned long size = 0;
 213
 214			if (get_stack_size(tok, &size) < 0)
 215				return -1;
 216			callchain_param.dump_size = size;
 217			try_stack_size = false;
 218		} else if (!minpcnt_set) {
 219			/* try to get the min percent */
 220			callchain_param.min_percent = strtod(tok, &endptr);
 221			if (tok == endptr)
 222				return -1;
 223			minpcnt_set = true;
 224		} else {
 225			/* try print limit at last */
 226			callchain_param.print_limit = strtoul(tok, &endptr, 0);
 227			if (tok == endptr)
 228				return -1;
 229		}
 230next:
 231		arg = NULL;
 232	}
 233
 234	if (callchain_register_param(&callchain_param) < 0) {
 235		pr_err("Can't register callchain params\n");
 236		return -1;
 237	}
 238	return 0;
 239}
 240
 241int parse_callchain_report_opt(const char *arg)
 242{
 243	return __parse_callchain_report_opt(arg, false);
 244}
 245
 246int parse_callchain_top_opt(const char *arg)
 247{
 248	return __parse_callchain_report_opt(arg, true);
 249}
 250
 251int parse_callchain_record(const char *arg, struct callchain_param *param)
 252{
 253	char *tok, *name, *saveptr = NULL;
 254	char *buf;
 255	int ret = -1;
 256
 257	/* We need buffer that we know we can write to. */
 258	buf = malloc(strlen(arg) + 1);
 259	if (!buf)
 260		return -ENOMEM;
 261
 262	strcpy(buf, arg);
 263
 264	tok = strtok_r((char *)buf, ",", &saveptr);
 265	name = tok ? : (char *)buf;
 266
 267	do {
 268		/* Framepointer style */
 269		if (!strncmp(name, "fp", sizeof("fp"))) {
 270			ret = 0;
 271			param->record_mode = CALLCHAIN_FP;
 272
 273			tok = strtok_r(NULL, ",", &saveptr);
 274			if (tok) {
 275				unsigned long size;
 276
 277				size = strtoul(tok, &name, 0);
 278				if (size < (unsigned) sysctl__max_stack())
 279					param->max_stack = size;
 280			}
 281			break;
 282
 283		/* Dwarf style */
 284		} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
 285			const unsigned long default_stack_dump_size = 8192;
 286
 287			ret = 0;
 288			param->record_mode = CALLCHAIN_DWARF;
 289			param->dump_size = default_stack_dump_size;
 290			dwarf_callchain_users = true;
 291
 292			tok = strtok_r(NULL, ",", &saveptr);
 293			if (tok) {
 294				unsigned long size = 0;
 295
 296				ret = get_stack_size(tok, &size);
 297				param->dump_size = size;
 298			}
 299		} else if (!strncmp(name, "lbr", sizeof("lbr"))) {
 300			if (!strtok_r(NULL, ",", &saveptr)) {
 301				param->record_mode = CALLCHAIN_LBR;
 302				ret = 0;
 303			} else
 304				pr_err("callchain: No more arguments "
 305					"needed for --call-graph lbr\n");
 306			break;
 307		} else {
 308			pr_err("callchain: Unknown --call-graph option "
 309			       "value: %s\n", arg);
 310			break;
 311		}
 312
 313	} while (0);
 314
 315	free(buf);
 316	return ret;
 317}
 318
 319int perf_callchain_config(const char *var, const char *value)
 320{
 321	char *endptr;
 322
 323	if (!strstarts(var, "call-graph."))
 324		return 0;
 325	var += sizeof("call-graph.") - 1;
 326
 327	if (!strcmp(var, "record-mode"))
 328		return parse_callchain_record_opt(value, &callchain_param);
 329	if (!strcmp(var, "dump-size")) {
 330		unsigned long size = 0;
 331		int ret;
 332
 333		ret = get_stack_size(value, &size);
 334		callchain_param.dump_size = size;
 335
 336		return ret;
 337	}
 338	if (!strcmp(var, "print-type")){
 339		int ret;
 340		ret = parse_callchain_mode(value);
 341		if (ret == -1)
 342			pr_err("Invalid callchain mode: %s\n", value);
 343		return ret;
 344	}
 345	if (!strcmp(var, "order")){
 346		int ret;
 347		ret = parse_callchain_order(value);
 348		if (ret == -1)
 349			pr_err("Invalid callchain order: %s\n", value);
 350		return ret;
 351	}
 352	if (!strcmp(var, "sort-key")){
 353		int ret;
 354		ret = parse_callchain_sort_key(value);
 355		if (ret == -1)
 356			pr_err("Invalid callchain sort key: %s\n", value);
 357		return ret;
 358	}
 359	if (!strcmp(var, "threshold")) {
 360		callchain_param.min_percent = strtod(value, &endptr);
 361		if (value == endptr) {
 362			pr_err("Invalid callchain threshold: %s\n", value);
 363			return -1;
 364		}
 365	}
 366	if (!strcmp(var, "print-limit")) {
 367		callchain_param.print_limit = strtod(value, &endptr);
 368		if (value == endptr) {
 369			pr_err("Invalid callchain print limit: %s\n", value);
 370			return -1;
 371		}
 372	}
 373
 374	return 0;
 375}
 376
 377static void
 378rb_insert_callchain(struct rb_root *root, struct callchain_node *chain,
 379		    enum chain_mode mode)
 380{
 381	struct rb_node **p = &root->rb_node;
 382	struct rb_node *parent = NULL;
 383	struct callchain_node *rnode;
 384	u64 chain_cumul = callchain_cumul_hits(chain);
 385
 386	while (*p) {
 387		u64 rnode_cumul;
 388
 389		parent = *p;
 390		rnode = rb_entry(parent, struct callchain_node, rb_node);
 391		rnode_cumul = callchain_cumul_hits(rnode);
 392
 393		switch (mode) {
 394		case CHAIN_FLAT:
 395		case CHAIN_FOLDED:
 396			if (rnode->hit < chain->hit)
 397				p = &(*p)->rb_left;
 398			else
 399				p = &(*p)->rb_right;
 400			break;
 401		case CHAIN_GRAPH_ABS: /* Falldown */
 402		case CHAIN_GRAPH_REL:
 403			if (rnode_cumul < chain_cumul)
 404				p = &(*p)->rb_left;
 405			else
 406				p = &(*p)->rb_right;
 407			break;
 408		case CHAIN_NONE:
 409		default:
 410			break;
 411		}
 412	}
 413
 414	rb_link_node(&chain->rb_node, parent, p);
 415	rb_insert_color(&chain->rb_node, root);
 416}
 417
 418static void
 419__sort_chain_flat(struct rb_root *rb_root, struct callchain_node *node,
 420		  u64 min_hit)
 421{
 422	struct rb_node *n;
 423	struct callchain_node *child;
 424
 425	n = rb_first(&node->rb_root_in);
 426	while (n) {
 427		child = rb_entry(n, struct callchain_node, rb_node_in);
 428		n = rb_next(n);
 429
 430		__sort_chain_flat(rb_root, child, min_hit);
 431	}
 432
 433	if (node->hit && node->hit >= min_hit)
 434		rb_insert_callchain(rb_root, node, CHAIN_FLAT);
 435}
 436
 437/*
 438 * Once we get every callchains from the stream, we can now
 439 * sort them by hit
 440 */
 441static void
 442sort_chain_flat(struct rb_root *rb_root, struct callchain_root *root,
 443		u64 min_hit, struct callchain_param *param __maybe_unused)
 444{
 445	*rb_root = RB_ROOT;
 446	__sort_chain_flat(rb_root, &root->node, min_hit);
 447}
 448
 449static void __sort_chain_graph_abs(struct callchain_node *node,
 450				   u64 min_hit)
 451{
 452	struct rb_node *n;
 453	struct callchain_node *child;
 454
 455	node->rb_root = RB_ROOT;
 456	n = rb_first(&node->rb_root_in);
 457
 458	while (n) {
 459		child = rb_entry(n, struct callchain_node, rb_node_in);
 460		n = rb_next(n);
 461
 462		__sort_chain_graph_abs(child, min_hit);
 463		if (callchain_cumul_hits(child) >= min_hit)
 464			rb_insert_callchain(&node->rb_root, child,
 465					    CHAIN_GRAPH_ABS);
 466	}
 467}
 468
 469static void
 470sort_chain_graph_abs(struct rb_root *rb_root, struct callchain_root *chain_root,
 471		     u64 min_hit, struct callchain_param *param __maybe_unused)
 472{
 473	__sort_chain_graph_abs(&chain_root->node, min_hit);
 474	rb_root->rb_node = chain_root->node.rb_root.rb_node;
 475}
 476
 477static void __sort_chain_graph_rel(struct callchain_node *node,
 478				   double min_percent)
 479{
 480	struct rb_node *n;
 481	struct callchain_node *child;
 482	u64 min_hit;
 483
 484	node->rb_root = RB_ROOT;
 485	min_hit = ceil(node->children_hit * min_percent);
 486
 487	n = rb_first(&node->rb_root_in);
 488	while (n) {
 489		child = rb_entry(n, struct callchain_node, rb_node_in);
 490		n = rb_next(n);
 491
 492		__sort_chain_graph_rel(child, min_percent);
 493		if (callchain_cumul_hits(child) >= min_hit)
 494			rb_insert_callchain(&node->rb_root, child,
 495					    CHAIN_GRAPH_REL);
 496	}
 497}
 498
 499static void
 500sort_chain_graph_rel(struct rb_root *rb_root, struct callchain_root *chain_root,
 501		     u64 min_hit __maybe_unused, struct callchain_param *param)
 502{
 503	__sort_chain_graph_rel(&chain_root->node, param->min_percent / 100.0);
 504	rb_root->rb_node = chain_root->node.rb_root.rb_node;
 505}
 506
 507int callchain_register_param(struct callchain_param *param)
 508{
 509	switch (param->mode) {
 510	case CHAIN_GRAPH_ABS:
 511		param->sort = sort_chain_graph_abs;
 512		break;
 513	case CHAIN_GRAPH_REL:
 514		param->sort = sort_chain_graph_rel;
 515		break;
 516	case CHAIN_FLAT:
 517	case CHAIN_FOLDED:
 518		param->sort = sort_chain_flat;
 519		break;
 520	case CHAIN_NONE:
 521	default:
 522		return -1;
 523	}
 524	return 0;
 525}
 526
 527/*
 528 * Create a child for a parent. If inherit_children, then the new child
 529 * will become the new parent of it's parent children
 530 */
 531static struct callchain_node *
 532create_child(struct callchain_node *parent, bool inherit_children)
 533{
 534	struct callchain_node *new;
 535
 536	new = zalloc(sizeof(*new));
 537	if (!new) {
 538		perror("not enough memory to create child for code path tree");
 539		return NULL;
 540	}
 541	new->parent = parent;
 542	INIT_LIST_HEAD(&new->val);
 543	INIT_LIST_HEAD(&new->parent_val);
 544
 545	if (inherit_children) {
 546		struct rb_node *n;
 547		struct callchain_node *child;
 548
 549		new->rb_root_in = parent->rb_root_in;
 550		parent->rb_root_in = RB_ROOT;
 551
 552		n = rb_first(&new->rb_root_in);
 553		while (n) {
 554			child = rb_entry(n, struct callchain_node, rb_node_in);
 555			child->parent = new;
 556			n = rb_next(n);
 557		}
 558
 559		/* make it the first child */
 560		rb_link_node(&new->rb_node_in, NULL, &parent->rb_root_in.rb_node);
 561		rb_insert_color(&new->rb_node_in, &parent->rb_root_in);
 562	}
 563
 564	return new;
 565}
 566
 567
 568/*
 569 * Fill the node with callchain values
 570 */
 571static int
 572fill_node(struct callchain_node *node, struct callchain_cursor *cursor)
 573{
 574	struct callchain_cursor_node *cursor_node;
 575
 576	node->val_nr = cursor->nr - cursor->pos;
 577	if (!node->val_nr)
 578		pr_warning("Warning: empty node in callchain tree\n");
 579
 580	cursor_node = callchain_cursor_current(cursor);
 581
 582	while (cursor_node) {
 583		struct callchain_list *call;
 584
 585		call = zalloc(sizeof(*call));
 586		if (!call) {
 587			perror("not enough memory for the code path tree");
 588			return -1;
 589		}
 590		call->ip = cursor_node->ip;
 591		call->ms = cursor_node->ms;
 592		map__get(call->ms.map);
 
 593		call->srcline = cursor_node->srcline;
 594
 595		if (cursor_node->branch) {
 596			call->branch_count = 1;
 597
 598			if (cursor_node->branch_from) {
 599				/*
 600				 * branch_from is set with value somewhere else
 601				 * to imply it's "to" of a branch.
 602				 */
 603				call->brtype_stat.branch_to = true;
 
 
 
 
 
 
 
 
 604
 605				if (cursor_node->branch_flags.predicted)
 606					call->predicted_count = 1;
 607
 608				if (cursor_node->branch_flags.abort)
 609					call->abort_count = 1;
 610
 611				branch_type_count(&call->brtype_stat,
 612						  &cursor_node->branch_flags,
 613						  cursor_node->branch_from,
 614						  cursor_node->ip);
 615			} else {
 616				/*
 617				 * It's "from" of a branch
 618				 */
 619				call->brtype_stat.branch_to = false;
 
 620				call->cycles_count =
 621					cursor_node->branch_flags.cycles;
 622				call->iter_count = cursor_node->nr_loop_iter;
 623				call->iter_cycles = cursor_node->iter_cycles;
 624			}
 625		}
 626
 627		list_add_tail(&call->list, &node->val);
 628
 629		callchain_cursor_advance(cursor);
 630		cursor_node = callchain_cursor_current(cursor);
 631	}
 632	return 0;
 633}
 634
 635static struct callchain_node *
 636add_child(struct callchain_node *parent,
 637	  struct callchain_cursor *cursor,
 638	  u64 period)
 639{
 640	struct callchain_node *new;
 641
 642	new = create_child(parent, false);
 643	if (new == NULL)
 644		return NULL;
 645
 646	if (fill_node(new, cursor) < 0) {
 647		struct callchain_list *call, *tmp;
 648
 649		list_for_each_entry_safe(call, tmp, &new->val, list) {
 650			list_del_init(&call->list);
 651			map__zput(call->ms.map);
 
 652			free(call);
 653		}
 654		free(new);
 655		return NULL;
 656	}
 657
 658	new->children_hit = 0;
 659	new->hit = period;
 660	new->children_count = 0;
 661	new->count = 1;
 662	return new;
 663}
 664
 665enum match_result {
 666	MATCH_ERROR  = -1,
 667	MATCH_EQ,
 668	MATCH_LT,
 669	MATCH_GT,
 670};
 671
 672static enum match_result match_chain_strings(const char *left,
 673					     const char *right)
 674{
 675	enum match_result ret = MATCH_EQ;
 676	int cmp;
 677
 678	if (left && right)
 679		cmp = strcmp(left, right);
 680	else if (!left && right)
 681		cmp = 1;
 682	else if (left && !right)
 683		cmp = -1;
 684	else
 685		return MATCH_ERROR;
 686
 687	if (cmp != 0)
 688		ret = cmp < 0 ? MATCH_LT : MATCH_GT;
 689
 690	return ret;
 691}
 692
 693/*
 694 * We need to always use relative addresses because we're aggregating
 695 * callchains from multiple threads, i.e. different address spaces, so
 696 * comparing absolute addresses make no sense as a symbol in a DSO may end up
 697 * in a different address when used in a different binary or even the same
 698 * binary but with some sort of address randomization technique, thus we need
 699 * to compare just relative addresses. -acme
 700 */
 701static enum match_result match_chain_dso_addresses(struct map *left_map, u64 left_ip,
 702						   struct map *right_map, u64 right_ip)
 703{
 704	struct dso *left_dso = left_map ? left_map->dso : NULL;
 705	struct dso *right_dso = right_map ? right_map->dso : NULL;
 706
 707	if (left_dso != right_dso)
 708		return left_dso < right_dso ? MATCH_LT : MATCH_GT;
 709
 710	if (left_ip != right_ip)
 711 		return left_ip < right_ip ? MATCH_LT : MATCH_GT;
 712
 713	return MATCH_EQ;
 714}
 715
 716static enum match_result match_chain(struct callchain_cursor_node *node,
 717				     struct callchain_list *cnode)
 718{
 719	enum match_result match = MATCH_ERROR;
 720
 721	switch (callchain_param.key) {
 722	case CCKEY_SRCLINE:
 723		match = match_chain_strings(cnode->srcline, node->srcline);
 724		if (match != MATCH_ERROR)
 725			break;
 726		/* otherwise fall-back to symbol-based comparison below */
 727		__fallthrough;
 728	case CCKEY_FUNCTION:
 729		if (node->ms.sym && cnode->ms.sym) {
 730			/*
 731			 * Compare inlined frames based on their symbol name
 732			 * because different inlined frames will have the same
 733			 * symbol start. Otherwise do a faster comparison based
 734			 * on the symbol start address.
 735			 */
 736			if (cnode->ms.sym->inlined || node->ms.sym->inlined) {
 737				match = match_chain_strings(cnode->ms.sym->name,
 738							    node->ms.sym->name);
 739				if (match != MATCH_ERROR)
 740					break;
 741			} else {
 742				match = match_chain_dso_addresses(cnode->ms.map, cnode->ms.sym->start,
 743								  node->ms.map, node->ms.sym->start);
 744				break;
 745			}
 746		}
 747		/* otherwise fall-back to IP-based comparison below */
 748		__fallthrough;
 749	case CCKEY_ADDRESS:
 750	default:
 751		match = match_chain_dso_addresses(cnode->ms.map, cnode->ip, node->ms.map, node->ip);
 752		break;
 753	}
 754
 755	if (match == MATCH_EQ && node->branch) {
 756		cnode->branch_count++;
 757
 758		if (node->branch_from) {
 759			/*
 760			 * It's "to" of a branch
 761			 */
 762			cnode->brtype_stat.branch_to = true;
 
 
 
 
 
 
 
 763
 764			if (node->branch_flags.predicted)
 765				cnode->predicted_count++;
 766
 767			if (node->branch_flags.abort)
 768				cnode->abort_count++;
 769
 770			branch_type_count(&cnode->brtype_stat,
 771					  &node->branch_flags,
 772					  node->branch_from,
 773					  node->ip);
 774		} else {
 775			/*
 776			 * It's "from" of a branch
 777			 */
 778			cnode->brtype_stat.branch_to = false;
 
 779			cnode->cycles_count += node->branch_flags.cycles;
 780			cnode->iter_count += node->nr_loop_iter;
 781			cnode->iter_cycles += node->iter_cycles;
 782			cnode->from_count++;
 783		}
 784	}
 785
 786	return match;
 787}
 788
 789/*
 790 * Split the parent in two parts (a new child is created) and
 791 * give a part of its callchain to the created child.
 792 * Then create another child to host the given callchain of new branch
 793 */
 794static int
 795split_add_child(struct callchain_node *parent,
 796		struct callchain_cursor *cursor,
 797		struct callchain_list *to_split,
 798		u64 idx_parents, u64 idx_local, u64 period)
 799{
 800	struct callchain_node *new;
 801	struct list_head *old_tail;
 802	unsigned int idx_total = idx_parents + idx_local;
 803
 804	/* split */
 805	new = create_child(parent, true);
 806	if (new == NULL)
 807		return -1;
 808
 809	/* split the callchain and move a part to the new child */
 810	old_tail = parent->val.prev;
 811	list_del_range(&to_split->list, old_tail);
 812	new->val.next = &to_split->list;
 813	new->val.prev = old_tail;
 814	to_split->list.prev = &new->val;
 815	old_tail->next = &new->val;
 816
 817	/* split the hits */
 818	new->hit = parent->hit;
 819	new->children_hit = parent->children_hit;
 820	parent->children_hit = callchain_cumul_hits(new);
 821	new->val_nr = parent->val_nr - idx_local;
 822	parent->val_nr = idx_local;
 823	new->count = parent->count;
 824	new->children_count = parent->children_count;
 825	parent->children_count = callchain_cumul_counts(new);
 826
 827	/* create a new child for the new branch if any */
 828	if (idx_total < cursor->nr) {
 829		struct callchain_node *first;
 830		struct callchain_list *cnode;
 831		struct callchain_cursor_node *node;
 832		struct rb_node *p, **pp;
 833
 834		parent->hit = 0;
 835		parent->children_hit += period;
 836		parent->count = 0;
 837		parent->children_count += 1;
 838
 839		node = callchain_cursor_current(cursor);
 840		new = add_child(parent, cursor, period);
 841		if (new == NULL)
 842			return -1;
 843
 844		/*
 845		 * This is second child since we moved parent's children
 846		 * to new (first) child above.
 847		 */
 848		p = parent->rb_root_in.rb_node;
 849		first = rb_entry(p, struct callchain_node, rb_node_in);
 850		cnode = list_first_entry(&first->val, struct callchain_list,
 851					 list);
 852
 853		if (match_chain(node, cnode) == MATCH_LT)
 854			pp = &p->rb_left;
 855		else
 856			pp = &p->rb_right;
 857
 858		rb_link_node(&new->rb_node_in, p, pp);
 859		rb_insert_color(&new->rb_node_in, &parent->rb_root_in);
 860	} else {
 861		parent->hit = period;
 862		parent->count = 1;
 863	}
 864	return 0;
 865}
 866
 867static enum match_result
 868append_chain(struct callchain_node *root,
 869	     struct callchain_cursor *cursor,
 870	     u64 period);
 871
 872static int
 873append_chain_children(struct callchain_node *root,
 874		      struct callchain_cursor *cursor,
 875		      u64 period)
 876{
 877	struct callchain_node *rnode;
 878	struct callchain_cursor_node *node;
 879	struct rb_node **p = &root->rb_root_in.rb_node;
 880	struct rb_node *parent = NULL;
 881
 882	node = callchain_cursor_current(cursor);
 883	if (!node)
 884		return -1;
 885
 886	/* lookup in children */
 887	while (*p) {
 888		enum match_result ret;
 889
 890		parent = *p;
 891		rnode = rb_entry(parent, struct callchain_node, rb_node_in);
 892
 893		/* If at least first entry matches, rely to children */
 894		ret = append_chain(rnode, cursor, period);
 895		if (ret == MATCH_EQ)
 896			goto inc_children_hit;
 897		if (ret == MATCH_ERROR)
 898			return -1;
 899
 900		if (ret == MATCH_LT)
 901			p = &parent->rb_left;
 902		else
 903			p = &parent->rb_right;
 904	}
 905	/* nothing in children, add to the current node */
 906	rnode = add_child(root, cursor, period);
 907	if (rnode == NULL)
 908		return -1;
 909
 910	rb_link_node(&rnode->rb_node_in, parent, p);
 911	rb_insert_color(&rnode->rb_node_in, &root->rb_root_in);
 912
 913inc_children_hit:
 914	root->children_hit += period;
 915	root->children_count++;
 916	return 0;
 917}
 918
 919static enum match_result
 920append_chain(struct callchain_node *root,
 921	     struct callchain_cursor *cursor,
 922	     u64 period)
 923{
 924	struct callchain_list *cnode;
 925	u64 start = cursor->pos;
 926	bool found = false;
 927	u64 matches;
 928	enum match_result cmp = MATCH_ERROR;
 929
 930	/*
 931	 * Lookup in the current node
 932	 * If we have a symbol, then compare the start to match
 933	 * anywhere inside a function, unless function
 934	 * mode is disabled.
 935	 */
 936	list_for_each_entry(cnode, &root->val, list) {
 937		struct callchain_cursor_node *node;
 938
 939		node = callchain_cursor_current(cursor);
 940		if (!node)
 941			break;
 942
 943		cmp = match_chain(node, cnode);
 944		if (cmp != MATCH_EQ)
 945			break;
 946
 947		found = true;
 948
 949		callchain_cursor_advance(cursor);
 950	}
 951
 952	/* matches not, relay no the parent */
 953	if (!found) {
 954		WARN_ONCE(cmp == MATCH_ERROR, "Chain comparison error\n");
 955		return cmp;
 956	}
 957
 958	matches = cursor->pos - start;
 959
 960	/* we match only a part of the node. Split it and add the new chain */
 961	if (matches < root->val_nr) {
 962		if (split_add_child(root, cursor, cnode, start, matches,
 963				    period) < 0)
 964			return MATCH_ERROR;
 965
 966		return MATCH_EQ;
 967	}
 968
 969	/* we match 100% of the path, increment the hit */
 970	if (matches == root->val_nr && cursor->pos == cursor->nr) {
 971		root->hit += period;
 972		root->count++;
 973		return MATCH_EQ;
 974	}
 975
 976	/* We match the node and still have a part remaining */
 977	if (append_chain_children(root, cursor, period) < 0)
 978		return MATCH_ERROR;
 979
 980	return MATCH_EQ;
 981}
 982
 983int callchain_append(struct callchain_root *root,
 984		     struct callchain_cursor *cursor,
 985		     u64 period)
 986{
 
 
 
 987	if (!cursor->nr)
 988		return 0;
 989
 990	callchain_cursor_commit(cursor);
 991
 992	if (append_chain_children(&root->node, cursor, period) < 0)
 993		return -1;
 994
 995	if (cursor->nr > root->max_depth)
 996		root->max_depth = cursor->nr;
 997
 998	return 0;
 999}
1000
1001static int
1002merge_chain_branch(struct callchain_cursor *cursor,
1003		   struct callchain_node *dst, struct callchain_node *src)
1004{
1005	struct callchain_cursor_node **old_last = cursor->last;
1006	struct callchain_node *child;
1007	struct callchain_list *list, *next_list;
1008	struct rb_node *n;
1009	int old_pos = cursor->nr;
1010	int err = 0;
1011
1012	list_for_each_entry_safe(list, next_list, &src->val, list) {
1013		callchain_cursor_append(cursor, list->ip, &list->ms,
1014					false, NULL, 0, 0, 0, list->srcline);
 
 
 
1015		list_del_init(&list->list);
1016		map__zput(list->ms.map);
 
 
1017		free(list);
1018	}
1019
1020	if (src->hit) {
1021		callchain_cursor_commit(cursor);
1022		if (append_chain_children(dst, cursor, src->hit) < 0)
1023			return -1;
1024	}
1025
1026	n = rb_first(&src->rb_root_in);
1027	while (n) {
1028		child = container_of(n, struct callchain_node, rb_node_in);
1029		n = rb_next(n);
1030		rb_erase(&child->rb_node_in, &src->rb_root_in);
1031
1032		err = merge_chain_branch(cursor, dst, child);
1033		if (err)
1034			break;
1035
1036		free(child);
1037	}
1038
1039	cursor->nr = old_pos;
1040	cursor->last = old_last;
1041
1042	return err;
1043}
1044
1045int callchain_merge(struct callchain_cursor *cursor,
1046		    struct callchain_root *dst, struct callchain_root *src)
1047{
1048	return merge_chain_branch(cursor, &dst->node, &src->node);
1049}
1050
1051int callchain_cursor_append(struct callchain_cursor *cursor,
1052			    u64 ip, struct map_symbol *ms,
1053			    bool branch, struct branch_flags *flags,
1054			    int nr_loop_iter, u64 iter_cycles, u64 branch_from,
1055			    const char *srcline)
1056{
1057	struct callchain_cursor_node *node = *cursor->last;
1058
1059	if (!node) {
1060		node = calloc(1, sizeof(*node));
1061		if (!node)
1062			return -ENOMEM;
1063
1064		*cursor->last = node;
1065	}
1066
1067	node->ip = ip;
1068	map__zput(node->ms.map);
1069	node->ms = *ms;
1070	map__get(node->ms.map);
 
1071	node->branch = branch;
1072	node->nr_loop_iter = nr_loop_iter;
1073	node->iter_cycles = iter_cycles;
1074	node->srcline = srcline;
1075
1076	if (flags)
1077		memcpy(&node->branch_flags, flags,
1078			sizeof(struct branch_flags));
1079
1080	node->branch_from = branch_from;
1081	cursor->nr++;
1082
1083	cursor->last = &node->next;
1084
1085	return 0;
1086}
1087
1088int sample__resolve_callchain(struct perf_sample *sample,
1089			      struct callchain_cursor *cursor, struct symbol **parent,
1090			      struct evsel *evsel, struct addr_location *al,
1091			      int max_stack)
1092{
1093	if (sample->callchain == NULL && !symbol_conf.show_branchflag_count)
1094		return 0;
1095
1096	if (symbol_conf.use_callchain || symbol_conf.cumulate_callchain ||
1097	    perf_hpp_list.parent || symbol_conf.show_branchflag_count) {
1098		return thread__resolve_callchain(al->thread, cursor, evsel, sample,
1099						 parent, al, max_stack);
1100	}
1101	return 0;
1102}
1103
1104int hist_entry__append_callchain(struct hist_entry *he, struct perf_sample *sample)
1105{
1106	if ((!symbol_conf.use_callchain || sample->callchain == NULL) &&
1107		!symbol_conf.show_branchflag_count)
1108		return 0;
1109	return callchain_append(he->callchain, &callchain_cursor, sample->period);
1110}
1111
1112int fill_callchain_info(struct addr_location *al, struct callchain_cursor_node *node,
1113			bool hide_unresolved)
1114{
1115	al->maps = node->ms.maps;
1116	al->map = node->ms.map;
 
 
 
 
1117	al->sym = node->ms.sym;
1118	al->srcline = node->srcline;
1119	al->addr = node->ip;
1120
1121	if (al->sym == NULL) {
1122		if (hide_unresolved)
1123			return 0;
1124		if (al->map == NULL)
1125			goto out;
1126	}
1127
1128	if (al->maps == machine__kernel_maps(al->maps->machine)) {
1129		if (machine__is_host(al->maps->machine)) {
1130			al->cpumode = PERF_RECORD_MISC_KERNEL;
1131			al->level = 'k';
1132		} else {
1133			al->cpumode = PERF_RECORD_MISC_GUEST_KERNEL;
1134			al->level = 'g';
1135		}
1136	} else {
1137		if (machine__is_host(al->maps->machine)) {
1138			al->cpumode = PERF_RECORD_MISC_USER;
1139			al->level = '.';
1140		} else if (perf_guest) {
1141			al->cpumode = PERF_RECORD_MISC_GUEST_USER;
1142			al->level = 'u';
1143		} else {
1144			al->cpumode = PERF_RECORD_MISC_HYPERVISOR;
1145			al->level = 'H';
1146		}
1147	}
1148
1149out:
1150	return 1;
1151}
1152
1153char *callchain_list__sym_name(struct callchain_list *cl,
1154			       char *bf, size_t bfsize, bool show_dso)
1155{
1156	bool show_addr = callchain_param.key == CCKEY_ADDRESS;
1157	bool show_srcline = show_addr || callchain_param.key == CCKEY_SRCLINE;
1158	int printed;
1159
1160	if (cl->ms.sym) {
1161		const char *inlined = cl->ms.sym->inlined ? " (inlined)" : "";
1162
1163		if (show_srcline && cl->srcline)
1164			printed = scnprintf(bf, bfsize, "%s %s%s",
1165					    cl->ms.sym->name, cl->srcline,
1166					    inlined);
1167		else
1168			printed = scnprintf(bf, bfsize, "%s%s",
1169					    cl->ms.sym->name, inlined);
1170	} else
1171		printed = scnprintf(bf, bfsize, "%#" PRIx64, cl->ip);
1172
1173	if (show_dso)
1174		scnprintf(bf + printed, bfsize - printed, " %s",
1175			  cl->ms.map ?
1176			  cl->ms.map->dso->short_name :
1177			  "unknown");
1178
1179	return bf;
1180}
1181
1182char *callchain_node__scnprintf_value(struct callchain_node *node,
1183				      char *bf, size_t bfsize, u64 total)
1184{
1185	double percent = 0.0;
1186	u64 period = callchain_cumul_hits(node);
1187	unsigned count = callchain_cumul_counts(node);
1188
1189	if (callchain_param.mode == CHAIN_FOLDED) {
1190		period = node->hit;
1191		count = node->count;
1192	}
1193
1194	switch (callchain_param.value) {
1195	case CCVAL_PERIOD:
1196		scnprintf(bf, bfsize, "%"PRIu64, period);
1197		break;
1198	case CCVAL_COUNT:
1199		scnprintf(bf, bfsize, "%u", count);
1200		break;
1201	case CCVAL_PERCENT:
1202	default:
1203		if (total)
1204			percent = period * 100.0 / total;
1205		scnprintf(bf, bfsize, "%.2f%%", percent);
1206		break;
1207	}
1208	return bf;
1209}
1210
1211int callchain_node__fprintf_value(struct callchain_node *node,
1212				 FILE *fp, u64 total)
1213{
1214	double percent = 0.0;
1215	u64 period = callchain_cumul_hits(node);
1216	unsigned count = callchain_cumul_counts(node);
1217
1218	if (callchain_param.mode == CHAIN_FOLDED) {
1219		period = node->hit;
1220		count = node->count;
1221	}
1222
1223	switch (callchain_param.value) {
1224	case CCVAL_PERIOD:
1225		return fprintf(fp, "%"PRIu64, period);
1226	case CCVAL_COUNT:
1227		return fprintf(fp, "%u", count);
1228	case CCVAL_PERCENT:
1229	default:
1230		if (total)
1231			percent = period * 100.0 / total;
1232		return percent_color_fprintf(fp, "%.2f%%", percent);
1233	}
1234	return 0;
1235}
1236
1237static void callchain_counts_value(struct callchain_node *node,
1238				   u64 *branch_count, u64 *predicted_count,
1239				   u64 *abort_count, u64 *cycles_count)
1240{
1241	struct callchain_list *clist;
1242
1243	list_for_each_entry(clist, &node->val, list) {
1244		if (branch_count)
1245			*branch_count += clist->branch_count;
1246
1247		if (predicted_count)
1248			*predicted_count += clist->predicted_count;
1249
1250		if (abort_count)
1251			*abort_count += clist->abort_count;
1252
1253		if (cycles_count)
1254			*cycles_count += clist->cycles_count;
1255	}
1256}
1257
1258static int callchain_node_branch_counts_cumul(struct callchain_node *node,
1259					      u64 *branch_count,
1260					      u64 *predicted_count,
1261					      u64 *abort_count,
1262					      u64 *cycles_count)
1263{
1264	struct callchain_node *child;
1265	struct rb_node *n;
1266
1267	n = rb_first(&node->rb_root_in);
1268	while (n) {
1269		child = rb_entry(n, struct callchain_node, rb_node_in);
1270		n = rb_next(n);
1271
1272		callchain_node_branch_counts_cumul(child, branch_count,
1273						   predicted_count,
1274						   abort_count,
1275						   cycles_count);
1276
1277		callchain_counts_value(child, branch_count,
1278				       predicted_count, abort_count,
1279				       cycles_count);
1280	}
1281
1282	return 0;
1283}
1284
1285int callchain_branch_counts(struct callchain_root *root,
1286			    u64 *branch_count, u64 *predicted_count,
1287			    u64 *abort_count, u64 *cycles_count)
1288{
1289	if (branch_count)
1290		*branch_count = 0;
1291
1292	if (predicted_count)
1293		*predicted_count = 0;
1294
1295	if (abort_count)
1296		*abort_count = 0;
1297
1298	if (cycles_count)
1299		*cycles_count = 0;
1300
1301	return callchain_node_branch_counts_cumul(&root->node,
1302						  branch_count,
1303						  predicted_count,
1304						  abort_count,
1305						  cycles_count);
1306}
1307
1308static int count_pri64_printf(int idx, const char *str, u64 value, char *bf, int bfsize)
1309{
1310	return scnprintf(bf, bfsize, "%s%s:%" PRId64 "", (idx) ? " " : " (", str, value);
1311}
1312
1313static int count_float_printf(int idx, const char *str, float value,
1314			      char *bf, int bfsize, float threshold)
1315{
1316	if (threshold != 0.0 && value < threshold)
1317		return 0;
1318
1319	return scnprintf(bf, bfsize, "%s%s:%.1f%%", (idx) ? " " : " (", str, value);
1320}
1321
1322static int branch_to_str(char *bf, int bfsize,
1323			 u64 branch_count, u64 predicted_count,
1324			 u64 abort_count,
1325			 struct branch_type_stat *brtype_stat)
1326{
1327	int printed, i = 0;
1328
1329	printed = branch_type_str(brtype_stat, bf, bfsize);
1330	if (printed)
1331		i++;
1332
1333	if (predicted_count < branch_count) {
1334		printed += count_float_printf(i++, "predicted",
1335				predicted_count * 100.0 / branch_count,
1336				bf + printed, bfsize - printed, 0.0);
1337	}
1338
1339	if (abort_count) {
1340		printed += count_float_printf(i++, "abort",
1341				abort_count * 100.0 / branch_count,
1342				bf + printed, bfsize - printed, 0.1);
1343	}
1344
1345	if (i)
1346		printed += scnprintf(bf + printed, bfsize - printed, ")");
1347
1348	return printed;
1349}
1350
1351static int branch_from_str(char *bf, int bfsize,
1352			   u64 branch_count,
1353			   u64 cycles_count, u64 iter_count,
1354			   u64 iter_cycles, u64 from_count)
1355{
1356	int printed = 0, i = 0;
1357	u64 cycles, v = 0;
1358
1359	cycles = cycles_count / branch_count;
1360	if (cycles) {
1361		printed += count_pri64_printf(i++, "cycles",
1362				cycles,
1363				bf + printed, bfsize - printed);
1364	}
1365
1366	if (iter_count && from_count) {
1367		v = iter_count / from_count;
1368		if (v) {
1369			printed += count_pri64_printf(i++, "iter",
1370					v, bf + printed, bfsize - printed);
1371
1372			printed += count_pri64_printf(i++, "avg_cycles",
1373					iter_cycles / iter_count,
1374					bf + printed, bfsize - printed);
1375		}
1376	}
1377
1378	if (i)
1379		printed += scnprintf(bf + printed, bfsize - printed, ")");
1380
1381	return printed;
1382}
1383
1384static int counts_str_build(char *bf, int bfsize,
1385			     u64 branch_count, u64 predicted_count,
1386			     u64 abort_count, u64 cycles_count,
1387			     u64 iter_count, u64 iter_cycles,
1388			     u64 from_count,
1389			     struct branch_type_stat *brtype_stat)
1390{
1391	int printed;
1392
1393	if (branch_count == 0)
1394		return scnprintf(bf, bfsize, " (calltrace)");
1395
1396	if (brtype_stat->branch_to) {
1397		printed = branch_to_str(bf, bfsize, branch_count,
1398				predicted_count, abort_count, brtype_stat);
1399	} else {
1400		printed = branch_from_str(bf, bfsize, branch_count,
1401				cycles_count, iter_count, iter_cycles,
1402				from_count);
1403	}
1404
1405	if (!printed)
1406		bf[0] = 0;
1407
1408	return printed;
1409}
1410
1411static int callchain_counts_printf(FILE *fp, char *bf, int bfsize,
1412				   u64 branch_count, u64 predicted_count,
1413				   u64 abort_count, u64 cycles_count,
1414				   u64 iter_count, u64 iter_cycles,
1415				   u64 from_count,
1416				   struct branch_type_stat *brtype_stat)
1417{
1418	char str[256];
1419
1420	counts_str_build(str, sizeof(str), branch_count,
1421			 predicted_count, abort_count, cycles_count,
1422			 iter_count, iter_cycles, from_count, brtype_stat);
1423
1424	if (fp)
1425		return fprintf(fp, "%s", str);
1426
1427	return scnprintf(bf, bfsize, "%s", str);
1428}
1429
1430int callchain_list_counts__printf_value(struct callchain_list *clist,
1431					FILE *fp, char *bf, int bfsize)
1432{
 
 
1433	u64 branch_count, predicted_count;
1434	u64 abort_count, cycles_count;
1435	u64 iter_count, iter_cycles;
1436	u64 from_count;
1437
 
1438	branch_count = clist->branch_count;
1439	predicted_count = clist->predicted_count;
1440	abort_count = clist->abort_count;
1441	cycles_count = clist->cycles_count;
1442	iter_count = clist->iter_count;
1443	iter_cycles = clist->iter_cycles;
1444	from_count = clist->from_count;
1445
1446	return callchain_counts_printf(fp, bf, bfsize, branch_count,
1447				       predicted_count, abort_count,
1448				       cycles_count, iter_count, iter_cycles,
1449				       from_count, &clist->brtype_stat);
1450}
1451
1452static void free_callchain_node(struct callchain_node *node)
1453{
1454	struct callchain_list *list, *tmp;
1455	struct callchain_node *child;
1456	struct rb_node *n;
1457
1458	list_for_each_entry_safe(list, tmp, &node->parent_val, list) {
1459		list_del_init(&list->list);
1460		map__zput(list->ms.map);
 
1461		free(list);
1462	}
1463
1464	list_for_each_entry_safe(list, tmp, &node->val, list) {
1465		list_del_init(&list->list);
1466		map__zput(list->ms.map);
 
1467		free(list);
1468	}
1469
1470	n = rb_first(&node->rb_root_in);
1471	while (n) {
1472		child = container_of(n, struct callchain_node, rb_node_in);
1473		n = rb_next(n);
1474		rb_erase(&child->rb_node_in, &node->rb_root_in);
1475
1476		free_callchain_node(child);
1477		free(child);
1478	}
1479}
1480
1481void free_callchain(struct callchain_root *root)
1482{
1483	if (!symbol_conf.use_callchain)
1484		return;
1485
1486	free_callchain_node(&root->node);
1487}
1488
1489static u64 decay_callchain_node(struct callchain_node *node)
1490{
1491	struct callchain_node *child;
1492	struct rb_node *n;
1493	u64 child_hits = 0;
1494
1495	n = rb_first(&node->rb_root_in);
1496	while (n) {
1497		child = container_of(n, struct callchain_node, rb_node_in);
1498
1499		child_hits += decay_callchain_node(child);
1500		n = rb_next(n);
1501	}
1502
1503	node->hit = (node->hit * 7) / 8;
1504	node->children_hit = child_hits;
1505
1506	return node->hit;
1507}
1508
1509void decay_callchain(struct callchain_root *root)
1510{
1511	if (!symbol_conf.use_callchain)
1512		return;
1513
1514	decay_callchain_node(&root->node);
1515}
1516
1517int callchain_node__make_parent_list(struct callchain_node *node)
1518{
1519	struct callchain_node *parent = node->parent;
1520	struct callchain_list *chain, *new;
1521	LIST_HEAD(head);
1522
1523	while (parent) {
1524		list_for_each_entry_reverse(chain, &parent->val, list) {
1525			new = malloc(sizeof(*new));
1526			if (new == NULL)
1527				goto out;
1528			*new = *chain;
1529			new->has_children = false;
1530			map__get(new->ms.map);
1531			list_add_tail(&new->list, &head);
1532		}
1533		parent = parent->parent;
1534	}
1535
1536	list_for_each_entry_safe_reverse(chain, new, &head, list)
1537		list_move_tail(&chain->list, &node->parent_val);
1538
1539	if (!list_empty(&node->parent_val)) {
1540		chain = list_first_entry(&node->parent_val, struct callchain_list, list);
1541		chain->has_children = rb_prev(&node->rb_node) || rb_next(&node->rb_node);
1542
1543		chain = list_first_entry(&node->val, struct callchain_list, list);
1544		chain->has_children = false;
1545	}
1546	return 0;
1547
1548out:
1549	list_for_each_entry_safe(chain, new, &head, list) {
1550		list_del_init(&chain->list);
1551		map__zput(chain->ms.map);
 
1552		free(chain);
1553	}
1554	return -ENOMEM;
1555}
1556
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1557int callchain_cursor__copy(struct callchain_cursor *dst,
1558			   struct callchain_cursor *src)
1559{
1560	int rc = 0;
1561
1562	callchain_cursor_reset(dst);
1563	callchain_cursor_commit(src);
1564
1565	while (true) {
1566		struct callchain_cursor_node *node;
1567
1568		node = callchain_cursor_current(src);
1569		if (node == NULL)
1570			break;
1571
1572		rc = callchain_cursor_append(dst, node->ip, &node->ms,
1573					     node->branch, &node->branch_flags,
1574					     node->nr_loop_iter,
1575					     node->iter_cycles,
1576					     node->branch_from, node->srcline);
1577		if (rc)
1578			break;
1579
1580		callchain_cursor_advance(src);
1581	}
1582
1583	return rc;
1584}
1585
1586/*
1587 * Initialize a cursor before adding entries inside, but keep
1588 * the previously allocated entries as a cache.
1589 */
1590void callchain_cursor_reset(struct callchain_cursor *cursor)
1591{
1592	struct callchain_cursor_node *node;
1593
1594	cursor->nr = 0;
1595	cursor->last = &cursor->first;
1596
1597	for (node = cursor->first; node != NULL; node = node->next)
1598		map__zput(node->ms.map);
1599}
1600
1601void callchain_param_setup(u64 sample_type, const char *arch)
1602{
1603	if (symbol_conf.use_callchain || symbol_conf.cumulate_callchain) {
1604		if ((sample_type & PERF_SAMPLE_REGS_USER) &&
1605		    (sample_type & PERF_SAMPLE_STACK_USER)) {
1606			callchain_param.record_mode = CALLCHAIN_DWARF;
1607			dwarf_callchain_users = true;
1608		} else if (sample_type & PERF_SAMPLE_BRANCH_STACK)
1609			callchain_param.record_mode = CALLCHAIN_LBR;
1610		else
1611			callchain_param.record_mode = CALLCHAIN_FP;
1612	}
1613
1614	/*
1615	 * It's necessary to use libunwind to reliably determine the caller of
1616	 * a leaf function on aarch64, as otherwise we cannot know whether to
1617	 * start from the LR or FP.
1618	 *
1619	 * Always starting from the LR can result in duplicate or entirely
1620	 * erroneous entries. Always skipping the LR and starting from the FP
1621	 * can result in missing entries.
1622	 */
1623	if (callchain_param.record_mode == CALLCHAIN_FP && !strcmp(arch, "arm64"))
1624		dwarf_callchain_users = true;
1625}
1626
1627static bool chain_match(struct callchain_list *base_chain,
1628			struct callchain_list *pair_chain)
1629{
1630	enum match_result match;
1631
1632	match = match_chain_strings(base_chain->srcline,
1633				    pair_chain->srcline);
1634	if (match != MATCH_ERROR)
1635		return match == MATCH_EQ;
1636
1637	match = match_chain_dso_addresses(base_chain->ms.map,
1638					  base_chain->ip,
1639					  pair_chain->ms.map,
1640					  pair_chain->ip);
1641
1642	return match == MATCH_EQ;
1643}
1644
1645bool callchain_cnode_matched(struct callchain_node *base_cnode,
1646			     struct callchain_node *pair_cnode)
1647{
1648	struct callchain_list *base_chain, *pair_chain;
1649	bool match = false;
1650
1651	pair_chain = list_first_entry(&pair_cnode->val,
1652				      struct callchain_list,
1653				      list);
1654
1655	list_for_each_entry(base_chain, &base_cnode->val, list) {
1656		if (&pair_chain->list == &pair_cnode->val)
1657			return false;
1658
1659		if (!base_chain->srcline || !pair_chain->srcline) {
1660			pair_chain = list_next_entry(pair_chain, list);
1661			continue;
1662		}
1663
1664		match = chain_match(base_chain, pair_chain);
1665		if (!match)
1666			return false;
1667
1668		pair_chain = list_next_entry(pair_chain, list);
1669	}
1670
1671	/*
1672	 * Say chain1 is ABC, chain2 is ABCD, we consider they are
1673	 * not fully matched.
1674	 */
1675	if (pair_chain && (&pair_chain->list != &pair_cnode->val))
1676		return false;
1677
1678	return match;
1679}
1680
1681static u64 count_callchain_hits(struct hist_entry *he)
1682{
1683	struct rb_root *root = &he->sorted_chain;
1684	struct rb_node *rb_node = rb_first(root);
1685	struct callchain_node *node;
1686	u64 chain_hits = 0;
1687
1688	while (rb_node) {
1689		node = rb_entry(rb_node, struct callchain_node, rb_node);
1690		chain_hits += node->hit;
1691		rb_node = rb_next(rb_node);
1692	}
1693
1694	return chain_hits;
1695}
1696
1697u64 callchain_total_hits(struct hists *hists)
1698{
1699	struct rb_node *next = rb_first_cached(&hists->entries);
1700	u64 chain_hits = 0;
1701
1702	while (next) {
1703		struct hist_entry *he = rb_entry(next, struct hist_entry,
1704						 rb_node);
1705
1706		chain_hits += count_callchain_hits(he);
1707		next = rb_next(&he->rb_node);
1708	}
1709
1710	return chain_hits;
1711}
1712
1713s64 callchain_avg_cycles(struct callchain_node *cnode)
1714{
1715	struct callchain_list *chain;
1716	s64 cycles = 0;
1717
1718	list_for_each_entry(chain, &cnode->val, list) {
1719		if (chain->srcline && chain->branch_count)
1720			cycles += chain->cycles_count / chain->branch_count;
1721	}
1722
1723	return cycles;
1724}