xref: /linux/tools/perf/util/annotate.c (revision 001821b0e79716c4e17c71d8e053a23599a7a508)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) 2011, Red Hat Inc, Arnaldo Carvalho de Melo <acme@redhat.com>
4  *
5  * Parts came from builtin-annotate.c, see those files for further
6  * copyright notes.
7  */
8 
9 #include <errno.h>
10 #include <inttypes.h>
11 #include <libgen.h>
12 #include <stdlib.h>
13 #include "util.h" // hex_width()
14 #include "ui/ui.h"
15 #include "sort.h"
16 #include "build-id.h"
17 #include "color.h"
18 #include "config.h"
19 #include "disasm.h"
20 #include "dso.h"
21 #include "env.h"
22 #include "map.h"
23 #include "maps.h"
24 #include "symbol.h"
25 #include "srcline.h"
26 #include "units.h"
27 #include "debug.h"
28 #include "annotate.h"
29 #include "annotate-data.h"
30 #include "evsel.h"
31 #include "evlist.h"
32 #include "bpf-event.h"
33 #include "bpf-utils.h"
34 #include "block-range.h"
35 #include "string2.h"
36 #include "dwarf-regs.h"
37 #include "util/event.h"
38 #include "util/sharded_mutex.h"
39 #include "arch/common.h"
40 #include "namespaces.h"
41 #include "thread.h"
42 #include "hashmap.h"
43 #include <regex.h>
44 #include <linux/bitops.h>
45 #include <linux/kernel.h>
46 #include <linux/string.h>
47 #include <linux/zalloc.h>
48 #include <subcmd/parse-options.h>
49 #include <subcmd/run-command.h>
50 
51 /* FIXME: For the HE_COLORSET */
52 #include "ui/browser.h"
53 
54 /*
55  * FIXME: Using the same values as slang.h,
56  * but that header may not be available everywhere
57  */
58 #define LARROW_CHAR	((unsigned char)',')
59 #define RARROW_CHAR	((unsigned char)'+')
60 #define DARROW_CHAR	((unsigned char)'.')
61 #define UARROW_CHAR	((unsigned char)'-')
62 
63 #include <linux/ctype.h>
64 
65 /* global annotation options */
66 struct annotation_options annotate_opts;
67 
68 /* Data type collection debug statistics */
69 struct annotated_data_stat ann_data_stat;
70 LIST_HEAD(ann_insn_stat);
71 
72 /* Pseudo data types */
73 struct annotated_data_type stackop_type = {
74 	.self = {
75 		.type_name = (char *)"(stack operation)",
76 		.children = LIST_HEAD_INIT(stackop_type.self.children),
77 	},
78 };
79 
80 struct annotated_data_type canary_type = {
81 	.self = {
82 		.type_name = (char *)"(stack canary)",
83 		.children = LIST_HEAD_INIT(canary_type.self.children),
84 	},
85 };
86 
87 /* symbol histogram: key = offset << 16 | evsel->core.idx */
88 static size_t sym_hist_hash(long key, void *ctx __maybe_unused)
89 {
90 	return (key >> 16) + (key & 0xffff);
91 }
92 
93 static bool sym_hist_equal(long key1, long key2, void *ctx __maybe_unused)
94 {
95 	return key1 == key2;
96 }
97 
98 static struct annotated_source *annotated_source__new(void)
99 {
100 	struct annotated_source *src = zalloc(sizeof(*src));
101 
102 	if (src != NULL)
103 		INIT_LIST_HEAD(&src->source);
104 
105 	return src;
106 }
107 
108 static __maybe_unused void annotated_source__delete(struct annotated_source *src)
109 {
110 	struct hashmap_entry *cur;
111 	size_t bkt;
112 
113 	if (src == NULL)
114 		return;
115 
116 	if (src->samples) {
117 		hashmap__for_each_entry(src->samples, cur, bkt)
118 			zfree(&cur->pvalue);
119 		hashmap__free(src->samples);
120 	}
121 	zfree(&src->histograms);
122 	free(src);
123 }
124 
125 static int annotated_source__alloc_histograms(struct annotated_source *src,
126 					      int nr_hists)
127 {
128 	src->nr_histograms   = nr_hists;
129 	src->histograms	     = calloc(nr_hists, sizeof(*src->histograms));
130 
131 	if (src->histograms == NULL)
132 		return -1;
133 
134 	src->samples = hashmap__new(sym_hist_hash, sym_hist_equal, NULL);
135 	if (src->samples == NULL)
136 		zfree(&src->histograms);
137 
138 	return src->histograms ? 0 : -1;
139 }
140 
141 void symbol__annotate_zero_histograms(struct symbol *sym)
142 {
143 	struct annotation *notes = symbol__annotation(sym);
144 
145 	annotation__lock(notes);
146 	if (notes->src != NULL) {
147 		memset(notes->src->histograms, 0,
148 		       notes->src->nr_histograms * sizeof(*notes->src->histograms));
149 		hashmap__clear(notes->src->samples);
150 	}
151 	if (notes->branch && notes->branch->cycles_hist) {
152 		memset(notes->branch->cycles_hist, 0,
153 		       symbol__size(sym) * sizeof(struct cyc_hist));
154 	}
155 	annotation__unlock(notes);
156 }
157 
158 static int __symbol__account_cycles(struct cyc_hist *ch,
159 				    u64 start,
160 				    unsigned offset, unsigned cycles,
161 				    unsigned have_start)
162 {
163 	/*
164 	 * For now we can only account one basic block per
165 	 * final jump. But multiple could be overlapping.
166 	 * Always account the longest one. So when
167 	 * a shorter one has been already seen throw it away.
168 	 *
169 	 * We separately always account the full cycles.
170 	 */
171 	ch[offset].num_aggr++;
172 	ch[offset].cycles_aggr += cycles;
173 
174 	if (cycles > ch[offset].cycles_max)
175 		ch[offset].cycles_max = cycles;
176 
177 	if (ch[offset].cycles_min) {
178 		if (cycles && cycles < ch[offset].cycles_min)
179 			ch[offset].cycles_min = cycles;
180 	} else
181 		ch[offset].cycles_min = cycles;
182 
183 	if (!have_start && ch[offset].have_start)
184 		return 0;
185 	if (ch[offset].num) {
186 		if (have_start && (!ch[offset].have_start ||
187 				   ch[offset].start > start)) {
188 			ch[offset].have_start = 0;
189 			ch[offset].cycles = 0;
190 			ch[offset].num = 0;
191 			if (ch[offset].reset < 0xffff)
192 				ch[offset].reset++;
193 		} else if (have_start &&
194 			   ch[offset].start < start)
195 			return 0;
196 	}
197 
198 	if (ch[offset].num < NUM_SPARKS)
199 		ch[offset].cycles_spark[ch[offset].num] = cycles;
200 
201 	ch[offset].have_start = have_start;
202 	ch[offset].start = start;
203 	ch[offset].cycles += cycles;
204 	ch[offset].num++;
205 	return 0;
206 }
207 
208 static int __symbol__inc_addr_samples(struct map_symbol *ms,
209 				      struct annotated_source *src, int evidx, u64 addr,
210 				      struct perf_sample *sample)
211 {
212 	struct symbol *sym = ms->sym;
213 	long hash_key;
214 	u64 offset;
215 	struct sym_hist *h;
216 	struct sym_hist_entry *entry;
217 
218 	pr_debug3("%s: addr=%#" PRIx64 "\n", __func__, map__unmap_ip(ms->map, addr));
219 
220 	if ((addr < sym->start || addr >= sym->end) &&
221 	    (addr != sym->end || sym->start != sym->end)) {
222 		pr_debug("%s(%d): ERANGE! sym->name=%s, start=%#" PRIx64 ", addr=%#" PRIx64 ", end=%#" PRIx64 "\n",
223 		       __func__, __LINE__, sym->name, sym->start, addr, sym->end);
224 		return -ERANGE;
225 	}
226 
227 	offset = addr - sym->start;
228 	h = annotated_source__histogram(src, evidx);
229 	if (h == NULL) {
230 		pr_debug("%s(%d): ENOMEM! sym->name=%s, start=%#" PRIx64 ", addr=%#" PRIx64 ", end=%#" PRIx64 ", func: %d\n",
231 			 __func__, __LINE__, sym->name, sym->start, addr, sym->end, sym->type == STT_FUNC);
232 		return -ENOMEM;
233 	}
234 
235 	hash_key = offset << 16 | evidx;
236 	if (!hashmap__find(src->samples, hash_key, &entry)) {
237 		entry = zalloc(sizeof(*entry));
238 		if (entry == NULL)
239 			return -ENOMEM;
240 
241 		if (hashmap__add(src->samples, hash_key, entry) < 0)
242 			return -ENOMEM;
243 	}
244 
245 	h->nr_samples++;
246 	h->period += sample->period;
247 	entry->nr_samples++;
248 	entry->period += sample->period;
249 
250 	pr_debug3("%#" PRIx64 " %s: period++ [addr: %#" PRIx64 ", %#" PRIx64
251 		  ", evidx=%d] => nr_samples: %" PRIu64 ", period: %" PRIu64 "\n",
252 		  sym->start, sym->name, addr, addr - sym->start, evidx,
253 		  entry->nr_samples, entry->period);
254 	return 0;
255 }
256 
257 struct annotated_branch *annotation__get_branch(struct annotation *notes)
258 {
259 	if (notes == NULL)
260 		return NULL;
261 
262 	if (notes->branch == NULL)
263 		notes->branch = zalloc(sizeof(*notes->branch));
264 
265 	return notes->branch;
266 }
267 
268 static struct cyc_hist *symbol__cycles_hist(struct symbol *sym)
269 {
270 	struct annotation *notes = symbol__annotation(sym);
271 	struct annotated_branch *branch;
272 
273 	branch = annotation__get_branch(notes);
274 	if (branch == NULL)
275 		return NULL;
276 
277 	if (branch->cycles_hist == NULL) {
278 		const size_t size = symbol__size(sym);
279 
280 		branch->cycles_hist = calloc(size, sizeof(struct cyc_hist));
281 	}
282 
283 	return branch->cycles_hist;
284 }
285 
286 struct annotated_source *symbol__hists(struct symbol *sym, int nr_hists)
287 {
288 	struct annotation *notes = symbol__annotation(sym);
289 
290 	if (notes->src == NULL) {
291 		notes->src = annotated_source__new();
292 		if (notes->src == NULL)
293 			return NULL;
294 		goto alloc_histograms;
295 	}
296 
297 	if (notes->src->histograms == NULL) {
298 alloc_histograms:
299 		annotated_source__alloc_histograms(notes->src, nr_hists);
300 	}
301 
302 	return notes->src;
303 }
304 
305 static int symbol__inc_addr_samples(struct map_symbol *ms,
306 				    struct evsel *evsel, u64 addr,
307 				    struct perf_sample *sample)
308 {
309 	struct symbol *sym = ms->sym;
310 	struct annotated_source *src;
311 
312 	if (sym == NULL)
313 		return 0;
314 	src = symbol__hists(sym, evsel->evlist->core.nr_entries);
315 	return src ? __symbol__inc_addr_samples(ms, src, evsel->core.idx, addr, sample) : 0;
316 }
317 
318 static int symbol__account_cycles(u64 addr, u64 start,
319 				  struct symbol *sym, unsigned cycles)
320 {
321 	struct cyc_hist *cycles_hist;
322 	unsigned offset;
323 
324 	if (sym == NULL)
325 		return 0;
326 	cycles_hist = symbol__cycles_hist(sym);
327 	if (cycles_hist == NULL)
328 		return -ENOMEM;
329 	if (addr < sym->start || addr >= sym->end)
330 		return -ERANGE;
331 
332 	if (start) {
333 		if (start < sym->start || start >= sym->end)
334 			return -ERANGE;
335 		if (start >= addr)
336 			start = 0;
337 	}
338 	offset = addr - sym->start;
339 	return __symbol__account_cycles(cycles_hist,
340 					start ? start - sym->start : 0,
341 					offset, cycles,
342 					!!start);
343 }
344 
345 int addr_map_symbol__account_cycles(struct addr_map_symbol *ams,
346 				    struct addr_map_symbol *start,
347 				    unsigned cycles)
348 {
349 	u64 saddr = 0;
350 	int err;
351 
352 	if (!cycles)
353 		return 0;
354 
355 	/*
356 	 * Only set start when IPC can be computed. We can only
357 	 * compute it when the basic block is completely in a single
358 	 * function.
359 	 * Special case the case when the jump is elsewhere, but
360 	 * it starts on the function start.
361 	 */
362 	if (start &&
363 		(start->ms.sym == ams->ms.sym ||
364 		 (ams->ms.sym &&
365 		  start->addr == ams->ms.sym->start + map__start(ams->ms.map))))
366 		saddr = start->al_addr;
367 	if (saddr == 0)
368 		pr_debug2("BB with bad start: addr %"PRIx64" start %"PRIx64" sym %"PRIx64" saddr %"PRIx64"\n",
369 			ams->addr,
370 			start ? start->addr : 0,
371 			ams->ms.sym ? ams->ms.sym->start + map__start(ams->ms.map) : 0,
372 			saddr);
373 	err = symbol__account_cycles(ams->al_addr, saddr, ams->ms.sym, cycles);
374 	if (err)
375 		pr_debug2("account_cycles failed %d\n", err);
376 	return err;
377 }
378 
379 struct annotation_line *annotated_source__get_line(struct annotated_source *src,
380 						   s64 offset)
381 {
382 	struct annotation_line *al;
383 
384 	list_for_each_entry(al, &src->source, node) {
385 		if (al->offset == offset)
386 			return al;
387 	}
388 	return NULL;
389 }
390 
391 static unsigned annotation__count_insn(struct annotation *notes, u64 start, u64 end)
392 {
393 	struct annotation_line *al;
394 	unsigned n_insn = 0;
395 
396 	al = annotated_source__get_line(notes->src, start);
397 	if (al == NULL)
398 		return 0;
399 
400 	list_for_each_entry_from(al, &notes->src->source, node) {
401 		if (al->offset == -1)
402 			continue;
403 		if ((u64)al->offset > end)
404 			break;
405 		n_insn++;
406 	}
407 	return n_insn;
408 }
409 
410 static void annotated_branch__delete(struct annotated_branch *branch)
411 {
412 	if (branch) {
413 		zfree(&branch->cycles_hist);
414 		free(branch);
415 	}
416 }
417 
418 static void annotation__count_and_fill(struct annotation *notes, u64 start, u64 end, struct cyc_hist *ch)
419 {
420 	unsigned n_insn;
421 	unsigned int cover_insn = 0;
422 
423 	n_insn = annotation__count_insn(notes, start, end);
424 	if (n_insn && ch->num && ch->cycles) {
425 		struct annotation_line *al;
426 		struct annotated_branch *branch;
427 		float ipc = n_insn / ((double)ch->cycles / (double)ch->num);
428 
429 		/* Hide data when there are too many overlaps. */
430 		if (ch->reset >= 0x7fff)
431 			return;
432 
433 		al = annotated_source__get_line(notes->src, start);
434 		if (al == NULL)
435 			return;
436 
437 		list_for_each_entry_from(al, &notes->src->source, node) {
438 			if (al->offset == -1)
439 				continue;
440 			if ((u64)al->offset > end)
441 				break;
442 			if (al->cycles && al->cycles->ipc == 0.0) {
443 				al->cycles->ipc = ipc;
444 				cover_insn++;
445 			}
446 		}
447 
448 		branch = annotation__get_branch(notes);
449 		if (cover_insn && branch) {
450 			branch->hit_cycles += ch->cycles;
451 			branch->hit_insn += n_insn * ch->num;
452 			branch->cover_insn += cover_insn;
453 		}
454 	}
455 }
456 
457 static int annotation__compute_ipc(struct annotation *notes, size_t size)
458 {
459 	int err = 0;
460 	s64 offset;
461 
462 	if (!notes->branch || !notes->branch->cycles_hist)
463 		return 0;
464 
465 	notes->branch->total_insn = annotation__count_insn(notes, 0, size - 1);
466 	notes->branch->hit_cycles = 0;
467 	notes->branch->hit_insn = 0;
468 	notes->branch->cover_insn = 0;
469 
470 	annotation__lock(notes);
471 	for (offset = size - 1; offset >= 0; --offset) {
472 		struct cyc_hist *ch;
473 
474 		ch = &notes->branch->cycles_hist[offset];
475 		if (ch && ch->cycles) {
476 			struct annotation_line *al;
477 
478 			al = annotated_source__get_line(notes->src, offset);
479 			if (al && al->cycles == NULL) {
480 				al->cycles = zalloc(sizeof(*al->cycles));
481 				if (al->cycles == NULL) {
482 					err = ENOMEM;
483 					break;
484 				}
485 			}
486 			if (ch->have_start)
487 				annotation__count_and_fill(notes, ch->start, offset, ch);
488 			if (al && ch->num_aggr) {
489 				al->cycles->avg = ch->cycles_aggr / ch->num_aggr;
490 				al->cycles->max = ch->cycles_max;
491 				al->cycles->min = ch->cycles_min;
492 			}
493 		}
494 	}
495 
496 	if (err) {
497 		while (++offset < (s64)size) {
498 			struct cyc_hist *ch = &notes->branch->cycles_hist[offset];
499 
500 			if (ch && ch->cycles) {
501 				struct annotation_line *al;
502 
503 				al = annotated_source__get_line(notes->src, offset);
504 				if (al)
505 					zfree(&al->cycles);
506 			}
507 		}
508 	}
509 
510 	annotation__unlock(notes);
511 	return 0;
512 }
513 
514 int addr_map_symbol__inc_samples(struct addr_map_symbol *ams, struct perf_sample *sample,
515 				 struct evsel *evsel)
516 {
517 	return symbol__inc_addr_samples(&ams->ms, evsel, ams->al_addr, sample);
518 }
519 
520 int hist_entry__inc_addr_samples(struct hist_entry *he, struct perf_sample *sample,
521 				 struct evsel *evsel, u64 ip)
522 {
523 	return symbol__inc_addr_samples(&he->ms, evsel, ip, sample);
524 }
525 
526 
527 void annotation__exit(struct annotation *notes)
528 {
529 	annotated_source__delete(notes->src);
530 	annotated_branch__delete(notes->branch);
531 }
532 
533 static struct sharded_mutex *sharded_mutex;
534 
535 static void annotation__init_sharded_mutex(void)
536 {
537 	/* As many mutexes as there are CPUs. */
538 	sharded_mutex = sharded_mutex__new(cpu__max_present_cpu().cpu);
539 }
540 
541 static size_t annotation__hash(const struct annotation *notes)
542 {
543 	return (size_t)notes;
544 }
545 
546 static struct mutex *annotation__get_mutex(const struct annotation *notes)
547 {
548 	static pthread_once_t once = PTHREAD_ONCE_INIT;
549 
550 	pthread_once(&once, annotation__init_sharded_mutex);
551 	if (!sharded_mutex)
552 		return NULL;
553 
554 	return sharded_mutex__get_mutex(sharded_mutex, annotation__hash(notes));
555 }
556 
557 void annotation__lock(struct annotation *notes)
558 	NO_THREAD_SAFETY_ANALYSIS
559 {
560 	struct mutex *mutex = annotation__get_mutex(notes);
561 
562 	if (mutex)
563 		mutex_lock(mutex);
564 }
565 
566 void annotation__unlock(struct annotation *notes)
567 	NO_THREAD_SAFETY_ANALYSIS
568 {
569 	struct mutex *mutex = annotation__get_mutex(notes);
570 
571 	if (mutex)
572 		mutex_unlock(mutex);
573 }
574 
575 bool annotation__trylock(struct annotation *notes)
576 {
577 	struct mutex *mutex = annotation__get_mutex(notes);
578 
579 	if (!mutex)
580 		return false;
581 
582 	return mutex_trylock(mutex);
583 }
584 
585 void annotation_line__add(struct annotation_line *al, struct list_head *head)
586 {
587 	list_add_tail(&al->node, head);
588 }
589 
590 struct annotation_line *
591 annotation_line__next(struct annotation_line *pos, struct list_head *head)
592 {
593 	list_for_each_entry_continue(pos, head, node)
594 		if (pos->offset >= 0)
595 			return pos;
596 
597 	return NULL;
598 }
599 
600 static const char *annotate__address_color(struct block_range *br)
601 {
602 	double cov = block_range__coverage(br);
603 
604 	if (cov >= 0) {
605 		/* mark red for >75% coverage */
606 		if (cov > 0.75)
607 			return PERF_COLOR_RED;
608 
609 		/* mark dull for <1% coverage */
610 		if (cov < 0.01)
611 			return PERF_COLOR_NORMAL;
612 	}
613 
614 	return PERF_COLOR_MAGENTA;
615 }
616 
617 static const char *annotate__asm_color(struct block_range *br)
618 {
619 	double cov = block_range__coverage(br);
620 
621 	if (cov >= 0) {
622 		/* mark dull for <1% coverage */
623 		if (cov < 0.01)
624 			return PERF_COLOR_NORMAL;
625 	}
626 
627 	return PERF_COLOR_BLUE;
628 }
629 
630 static void annotate__branch_printf(struct block_range *br, u64 addr)
631 {
632 	bool emit_comment = true;
633 
634 	if (!br)
635 		return;
636 
637 #if 1
638 	if (br->is_target && br->start == addr) {
639 		struct block_range *branch = br;
640 		double p;
641 
642 		/*
643 		 * Find matching branch to our target.
644 		 */
645 		while (!branch->is_branch)
646 			branch = block_range__next(branch);
647 
648 		p = 100 *(double)br->entry / branch->coverage;
649 
650 		if (p > 0.1) {
651 			if (emit_comment) {
652 				emit_comment = false;
653 				printf("\t#");
654 			}
655 
656 			/*
657 			 * The percentage of coverage joined at this target in relation
658 			 * to the next branch.
659 			 */
660 			printf(" +%.2f%%", p);
661 		}
662 	}
663 #endif
664 	if (br->is_branch && br->end == addr) {
665 		double p = 100*(double)br->taken / br->coverage;
666 
667 		if (p > 0.1) {
668 			if (emit_comment) {
669 				emit_comment = false;
670 				printf("\t#");
671 			}
672 
673 			/*
674 			 * The percentage of coverage leaving at this branch, and
675 			 * its prediction ratio.
676 			 */
677 			printf(" -%.2f%% (p:%.2f%%)", p, 100*(double)br->pred  / br->taken);
678 		}
679 	}
680 }
681 
682 static int disasm_line__print(struct disasm_line *dl, u64 start, int addr_fmt_width)
683 {
684 	s64 offset = dl->al.offset;
685 	const u64 addr = start + offset;
686 	struct block_range *br;
687 
688 	br = block_range__find(addr);
689 	color_fprintf(stdout, annotate__address_color(br), "  %*" PRIx64 ":", addr_fmt_width, addr);
690 	color_fprintf(stdout, annotate__asm_color(br), "%s", dl->al.line);
691 	annotate__branch_printf(br, addr);
692 	return 0;
693 }
694 
695 static int
696 annotation_line__print(struct annotation_line *al, struct symbol *sym, u64 start,
697 		       struct evsel *evsel, u64 len, int min_pcnt, int printed,
698 		       int max_lines, struct annotation_line *queue, int addr_fmt_width,
699 		       int percent_type)
700 {
701 	struct disasm_line *dl = container_of(al, struct disasm_line, al);
702 	static const char *prev_line;
703 
704 	if (al->offset != -1) {
705 		double max_percent = 0.0;
706 		int i, nr_percent = 1;
707 		const char *color;
708 		struct annotation *notes = symbol__annotation(sym);
709 
710 		for (i = 0; i < al->data_nr; i++) {
711 			double percent;
712 
713 			percent = annotation_data__percent(&al->data[i],
714 							   percent_type);
715 
716 			if (percent > max_percent)
717 				max_percent = percent;
718 		}
719 
720 		if (al->data_nr > nr_percent)
721 			nr_percent = al->data_nr;
722 
723 		if (max_percent < min_pcnt)
724 			return -1;
725 
726 		if (max_lines && printed >= max_lines)
727 			return 1;
728 
729 		if (queue != NULL) {
730 			list_for_each_entry_from(queue, &notes->src->source, node) {
731 				if (queue == al)
732 					break;
733 				annotation_line__print(queue, sym, start, evsel, len,
734 						       0, 0, 1, NULL, addr_fmt_width,
735 						       percent_type);
736 			}
737 		}
738 
739 		color = get_percent_color(max_percent);
740 
741 		for (i = 0; i < nr_percent; i++) {
742 			struct annotation_data *data = &al->data[i];
743 			double percent;
744 
745 			percent = annotation_data__percent(data, percent_type);
746 			color = get_percent_color(percent);
747 
748 			if (symbol_conf.show_total_period)
749 				color_fprintf(stdout, color, " %11" PRIu64,
750 					      data->he.period);
751 			else if (symbol_conf.show_nr_samples)
752 				color_fprintf(stdout, color, " %7" PRIu64,
753 					      data->he.nr_samples);
754 			else
755 				color_fprintf(stdout, color, " %7.2f", percent);
756 		}
757 
758 		printf(" : ");
759 
760 		disasm_line__print(dl, start, addr_fmt_width);
761 
762 		/*
763 		 * Also color the filename and line if needed, with
764 		 * the same color than the percentage. Don't print it
765 		 * twice for close colored addr with the same filename:line
766 		 */
767 		if (al->path) {
768 			if (!prev_line || strcmp(prev_line, al->path)) {
769 				color_fprintf(stdout, color, " // %s", al->path);
770 				prev_line = al->path;
771 			}
772 		}
773 
774 		printf("\n");
775 	} else if (max_lines && printed >= max_lines)
776 		return 1;
777 	else {
778 		int width = symbol_conf.show_total_period ? 12 : 8;
779 
780 		if (queue)
781 			return -1;
782 
783 		if (evsel__is_group_event(evsel))
784 			width *= evsel->core.nr_members;
785 
786 		if (!*al->line)
787 			printf(" %*s:\n", width, " ");
788 		else
789 			printf(" %*s: %-*d %s\n", width, " ", addr_fmt_width, al->line_nr, al->line);
790 	}
791 
792 	return 0;
793 }
794 
795 static void calc_percent(struct annotation *notes,
796 			 struct evsel *evsel,
797 			 struct annotation_data *data,
798 			 s64 offset, s64 end)
799 {
800 	struct hists *hists = evsel__hists(evsel);
801 	int evidx = evsel->core.idx;
802 	struct sym_hist *sym_hist = annotation__histogram(notes, evidx);
803 	unsigned int hits = 0;
804 	u64 period = 0;
805 
806 	while (offset < end) {
807 		struct sym_hist_entry *entry;
808 
809 		entry = annotated_source__hist_entry(notes->src, evidx, offset);
810 		if (entry) {
811 			hits   += entry->nr_samples;
812 			period += entry->period;
813 		}
814 		++offset;
815 	}
816 
817 	if (sym_hist->nr_samples) {
818 		data->he.period     = period;
819 		data->he.nr_samples = hits;
820 		data->percent[PERCENT_HITS_LOCAL] = 100.0 * hits / sym_hist->nr_samples;
821 	}
822 
823 	if (hists->stats.nr_non_filtered_samples)
824 		data->percent[PERCENT_HITS_GLOBAL] = 100.0 * hits / hists->stats.nr_non_filtered_samples;
825 
826 	if (sym_hist->period)
827 		data->percent[PERCENT_PERIOD_LOCAL] = 100.0 * period / sym_hist->period;
828 
829 	if (hists->stats.total_period)
830 		data->percent[PERCENT_PERIOD_GLOBAL] = 100.0 * period / hists->stats.total_period;
831 }
832 
833 static void annotation__calc_percent(struct annotation *notes,
834 				     struct evsel *leader, s64 len)
835 {
836 	struct annotation_line *al, *next;
837 	struct evsel *evsel;
838 
839 	list_for_each_entry(al, &notes->src->source, node) {
840 		s64 end;
841 		int i = 0;
842 
843 		if (al->offset == -1)
844 			continue;
845 
846 		next = annotation_line__next(al, &notes->src->source);
847 		end  = next ? next->offset : len;
848 
849 		for_each_group_evsel(evsel, leader) {
850 			struct annotation_data *data;
851 
852 			BUG_ON(i >= al->data_nr);
853 
854 			data = &al->data[i++];
855 
856 			calc_percent(notes, evsel, data, al->offset, end);
857 		}
858 	}
859 }
860 
861 void symbol__calc_percent(struct symbol *sym, struct evsel *evsel)
862 {
863 	struct annotation *notes = symbol__annotation(sym);
864 
865 	annotation__calc_percent(notes, evsel, symbol__size(sym));
866 }
867 
868 static int evsel__get_arch(struct evsel *evsel, struct arch **parch)
869 {
870 	struct perf_env *env = evsel__env(evsel);
871 	const char *arch_name = perf_env__arch(env);
872 	struct arch *arch;
873 	int err;
874 
875 	if (!arch_name) {
876 		*parch = NULL;
877 		return errno;
878 	}
879 
880 	*parch = arch = arch__find(arch_name);
881 	if (arch == NULL) {
882 		pr_err("%s: unsupported arch %s\n", __func__, arch_name);
883 		return ENOTSUP;
884 	}
885 
886 	if (arch->init) {
887 		err = arch->init(arch, env ? env->cpuid : NULL);
888 		if (err) {
889 			pr_err("%s: failed to initialize %s arch priv area\n",
890 			       __func__, arch->name);
891 			return err;
892 		}
893 	}
894 	return 0;
895 }
896 
897 int symbol__annotate(struct map_symbol *ms, struct evsel *evsel,
898 		     struct arch **parch)
899 {
900 	struct symbol *sym = ms->sym;
901 	struct annotation *notes = symbol__annotation(sym);
902 	struct annotate_args args = {
903 		.evsel		= evsel,
904 		.options	= &annotate_opts,
905 	};
906 	struct arch *arch = NULL;
907 	int err;
908 
909 	err = evsel__get_arch(evsel, &arch);
910 	if (err < 0)
911 		return err;
912 
913 	if (parch)
914 		*parch = arch;
915 
916 	if (notes->src && !list_empty(&notes->src->source))
917 		return 0;
918 
919 	args.arch = arch;
920 	args.ms = *ms;
921 
922 	if (notes->src == NULL) {
923 		notes->src = annotated_source__new();
924 		if (notes->src == NULL)
925 			return -1;
926 	}
927 
928 	if (annotate_opts.full_addr)
929 		notes->src->start = map__objdump_2mem(ms->map, ms->sym->start);
930 	else
931 		notes->src->start = map__rip_2objdump(ms->map, ms->sym->start);
932 
933 	return symbol__disassemble(sym, &args);
934 }
935 
936 static void insert_source_line(struct rb_root *root, struct annotation_line *al)
937 {
938 	struct annotation_line *iter;
939 	struct rb_node **p = &root->rb_node;
940 	struct rb_node *parent = NULL;
941 	unsigned int percent_type = annotate_opts.percent_type;
942 	int i, ret;
943 
944 	while (*p != NULL) {
945 		parent = *p;
946 		iter = rb_entry(parent, struct annotation_line, rb_node);
947 
948 		ret = strcmp(iter->path, al->path);
949 		if (ret == 0) {
950 			for (i = 0; i < al->data_nr; i++) {
951 				iter->data[i].percent_sum += annotation_data__percent(&al->data[i],
952 										      percent_type);
953 			}
954 			return;
955 		}
956 
957 		if (ret < 0)
958 			p = &(*p)->rb_left;
959 		else
960 			p = &(*p)->rb_right;
961 	}
962 
963 	for (i = 0; i < al->data_nr; i++) {
964 		al->data[i].percent_sum = annotation_data__percent(&al->data[i],
965 								   percent_type);
966 	}
967 
968 	rb_link_node(&al->rb_node, parent, p);
969 	rb_insert_color(&al->rb_node, root);
970 }
971 
972 static int cmp_source_line(struct annotation_line *a, struct annotation_line *b)
973 {
974 	int i;
975 
976 	for (i = 0; i < a->data_nr; i++) {
977 		if (a->data[i].percent_sum == b->data[i].percent_sum)
978 			continue;
979 		return a->data[i].percent_sum > b->data[i].percent_sum;
980 	}
981 
982 	return 0;
983 }
984 
985 static void __resort_source_line(struct rb_root *root, struct annotation_line *al)
986 {
987 	struct annotation_line *iter;
988 	struct rb_node **p = &root->rb_node;
989 	struct rb_node *parent = NULL;
990 
991 	while (*p != NULL) {
992 		parent = *p;
993 		iter = rb_entry(parent, struct annotation_line, rb_node);
994 
995 		if (cmp_source_line(al, iter))
996 			p = &(*p)->rb_left;
997 		else
998 			p = &(*p)->rb_right;
999 	}
1000 
1001 	rb_link_node(&al->rb_node, parent, p);
1002 	rb_insert_color(&al->rb_node, root);
1003 }
1004 
1005 static void resort_source_line(struct rb_root *dest_root, struct rb_root *src_root)
1006 {
1007 	struct annotation_line *al;
1008 	struct rb_node *node;
1009 
1010 	node = rb_first(src_root);
1011 	while (node) {
1012 		struct rb_node *next;
1013 
1014 		al = rb_entry(node, struct annotation_line, rb_node);
1015 		next = rb_next(node);
1016 		rb_erase(node, src_root);
1017 
1018 		__resort_source_line(dest_root, al);
1019 		node = next;
1020 	}
1021 }
1022 
1023 static void print_summary(struct rb_root *root, const char *filename)
1024 {
1025 	struct annotation_line *al;
1026 	struct rb_node *node;
1027 
1028 	printf("\nSorted summary for file %s\n", filename);
1029 	printf("----------------------------------------------\n\n");
1030 
1031 	if (RB_EMPTY_ROOT(root)) {
1032 		printf(" Nothing higher than %1.1f%%\n", MIN_GREEN);
1033 		return;
1034 	}
1035 
1036 	node = rb_first(root);
1037 	while (node) {
1038 		double percent, percent_max = 0.0;
1039 		const char *color;
1040 		char *path;
1041 		int i;
1042 
1043 		al = rb_entry(node, struct annotation_line, rb_node);
1044 		for (i = 0; i < al->data_nr; i++) {
1045 			percent = al->data[i].percent_sum;
1046 			color = get_percent_color(percent);
1047 			color_fprintf(stdout, color, " %7.2f", percent);
1048 
1049 			if (percent > percent_max)
1050 				percent_max = percent;
1051 		}
1052 
1053 		path = al->path;
1054 		color = get_percent_color(percent_max);
1055 		color_fprintf(stdout, color, " %s\n", path);
1056 
1057 		node = rb_next(node);
1058 	}
1059 }
1060 
1061 static void symbol__annotate_hits(struct symbol *sym, struct evsel *evsel)
1062 {
1063 	int evidx = evsel->core.idx;
1064 	struct annotation *notes = symbol__annotation(sym);
1065 	struct sym_hist *h = annotation__histogram(notes, evidx);
1066 	u64 len = symbol__size(sym), offset;
1067 
1068 	for (offset = 0; offset < len; ++offset) {
1069 		struct sym_hist_entry *entry;
1070 
1071 		entry = annotated_source__hist_entry(notes->src, evidx, offset);
1072 		if (entry && entry->nr_samples != 0)
1073 			printf("%*" PRIx64 ": %" PRIu64 "\n", BITS_PER_LONG / 2,
1074 			       sym->start + offset, entry->nr_samples);
1075 	}
1076 	printf("%*s: %" PRIu64 "\n", BITS_PER_LONG / 2, "h->nr_samples", h->nr_samples);
1077 }
1078 
1079 static int annotated_source__addr_fmt_width(struct list_head *lines, u64 start)
1080 {
1081 	char bf[32];
1082 	struct annotation_line *line;
1083 
1084 	list_for_each_entry_reverse(line, lines, node) {
1085 		if (line->offset != -1)
1086 			return scnprintf(bf, sizeof(bf), "%" PRIx64, start + line->offset);
1087 	}
1088 
1089 	return 0;
1090 }
1091 
1092 int symbol__annotate_printf(struct map_symbol *ms, struct evsel *evsel)
1093 {
1094 	struct map *map = ms->map;
1095 	struct symbol *sym = ms->sym;
1096 	struct dso *dso = map__dso(map);
1097 	char *filename;
1098 	const char *d_filename;
1099 	const char *evsel_name = evsel__name(evsel);
1100 	struct annotation *notes = symbol__annotation(sym);
1101 	struct sym_hist *h = annotation__histogram(notes, evsel->core.idx);
1102 	struct annotation_line *pos, *queue = NULL;
1103 	struct annotation_options *opts = &annotate_opts;
1104 	u64 start = map__rip_2objdump(map, sym->start);
1105 	int printed = 2, queue_len = 0, addr_fmt_width;
1106 	int more = 0;
1107 	bool context = opts->context;
1108 	u64 len;
1109 	int width = symbol_conf.show_total_period ? 12 : 8;
1110 	int graph_dotted_len;
1111 	char buf[512];
1112 
1113 	filename = strdup(dso__long_name(dso));
1114 	if (!filename)
1115 		return -ENOMEM;
1116 
1117 	if (opts->full_path)
1118 		d_filename = filename;
1119 	else
1120 		d_filename = basename(filename);
1121 
1122 	len = symbol__size(sym);
1123 
1124 	if (evsel__is_group_event(evsel)) {
1125 		width *= evsel->core.nr_members;
1126 		evsel__group_desc(evsel, buf, sizeof(buf));
1127 		evsel_name = buf;
1128 	}
1129 
1130 	graph_dotted_len = printf(" %-*.*s|	Source code & Disassembly of %s for %s (%" PRIu64 " samples, "
1131 				  "percent: %s)\n",
1132 				  width, width, symbol_conf.show_total_period ? "Period" :
1133 				  symbol_conf.show_nr_samples ? "Samples" : "Percent",
1134 				  d_filename, evsel_name, h->nr_samples,
1135 				  percent_type_str(opts->percent_type));
1136 
1137 	printf("%-*.*s----\n",
1138 	       graph_dotted_len, graph_dotted_len, graph_dotted_line);
1139 
1140 	if (verbose > 0)
1141 		symbol__annotate_hits(sym, evsel);
1142 
1143 	addr_fmt_width = annotated_source__addr_fmt_width(&notes->src->source, start);
1144 
1145 	list_for_each_entry(pos, &notes->src->source, node) {
1146 		int err;
1147 
1148 		if (context && queue == NULL) {
1149 			queue = pos;
1150 			queue_len = 0;
1151 		}
1152 
1153 		err = annotation_line__print(pos, sym, start, evsel, len,
1154 					     opts->min_pcnt, printed, opts->max_lines,
1155 					     queue, addr_fmt_width, opts->percent_type);
1156 
1157 		switch (err) {
1158 		case 0:
1159 			++printed;
1160 			if (context) {
1161 				printed += queue_len;
1162 				queue = NULL;
1163 				queue_len = 0;
1164 			}
1165 			break;
1166 		case 1:
1167 			/* filtered by max_lines */
1168 			++more;
1169 			break;
1170 		case -1:
1171 		default:
1172 			/*
1173 			 * Filtered by min_pcnt or non IP lines when
1174 			 * context != 0
1175 			 */
1176 			if (!context)
1177 				break;
1178 			if (queue_len == context)
1179 				queue = list_entry(queue->node.next, typeof(*queue), node);
1180 			else
1181 				++queue_len;
1182 			break;
1183 		}
1184 	}
1185 
1186 	free(filename);
1187 
1188 	return more;
1189 }
1190 
1191 static void FILE__set_percent_color(void *fp __maybe_unused,
1192 				    double percent __maybe_unused,
1193 				    bool current __maybe_unused)
1194 {
1195 }
1196 
1197 static int FILE__set_jumps_percent_color(void *fp __maybe_unused,
1198 					 int nr __maybe_unused, bool current __maybe_unused)
1199 {
1200 	return 0;
1201 }
1202 
1203 static int FILE__set_color(void *fp __maybe_unused, int color __maybe_unused)
1204 {
1205 	return 0;
1206 }
1207 
1208 static void FILE__printf(void *fp, const char *fmt, ...)
1209 {
1210 	va_list args;
1211 
1212 	va_start(args, fmt);
1213 	vfprintf(fp, fmt, args);
1214 	va_end(args);
1215 }
1216 
1217 static void FILE__write_graph(void *fp, int graph)
1218 {
1219 	const char *s;
1220 	switch (graph) {
1221 
1222 	case DARROW_CHAR: s = "↓"; break;
1223 	case UARROW_CHAR: s = "↑"; break;
1224 	case LARROW_CHAR: s = "←"; break;
1225 	case RARROW_CHAR: s = "→"; break;
1226 	default:		s = "?"; break;
1227 	}
1228 
1229 	fputs(s, fp);
1230 }
1231 
1232 static int symbol__annotate_fprintf2(struct symbol *sym, FILE *fp)
1233 {
1234 	struct annotation *notes = symbol__annotation(sym);
1235 	struct annotation_write_ops wops = {
1236 		.first_line		 = true,
1237 		.obj			 = fp,
1238 		.set_color		 = FILE__set_color,
1239 		.set_percent_color	 = FILE__set_percent_color,
1240 		.set_jumps_percent_color = FILE__set_jumps_percent_color,
1241 		.printf			 = FILE__printf,
1242 		.write_graph		 = FILE__write_graph,
1243 	};
1244 	struct annotation_line *al;
1245 
1246 	list_for_each_entry(al, &notes->src->source, node) {
1247 		if (annotation_line__filter(al))
1248 			continue;
1249 		annotation_line__write(al, notes, &wops);
1250 		fputc('\n', fp);
1251 		wops.first_line = false;
1252 	}
1253 
1254 	return 0;
1255 }
1256 
1257 int map_symbol__annotation_dump(struct map_symbol *ms, struct evsel *evsel)
1258 {
1259 	const char *ev_name = evsel__name(evsel);
1260 	char buf[1024];
1261 	char *filename;
1262 	int err = -1;
1263 	FILE *fp;
1264 
1265 	if (asprintf(&filename, "%s.annotation", ms->sym->name) < 0)
1266 		return -1;
1267 
1268 	fp = fopen(filename, "w");
1269 	if (fp == NULL)
1270 		goto out_free_filename;
1271 
1272 	if (evsel__is_group_event(evsel)) {
1273 		evsel__group_desc(evsel, buf, sizeof(buf));
1274 		ev_name = buf;
1275 	}
1276 
1277 	fprintf(fp, "%s() %s\nEvent: %s\n\n",
1278 		ms->sym->name, dso__long_name(map__dso(ms->map)), ev_name);
1279 	symbol__annotate_fprintf2(ms->sym, fp);
1280 
1281 	fclose(fp);
1282 	err = 0;
1283 out_free_filename:
1284 	free(filename);
1285 	return err;
1286 }
1287 
1288 void symbol__annotate_zero_histogram(struct symbol *sym, int evidx)
1289 {
1290 	struct annotation *notes = symbol__annotation(sym);
1291 	struct sym_hist *h = annotation__histogram(notes, evidx);
1292 
1293 	memset(h, 0, sizeof(*notes->src->histograms) * notes->src->nr_histograms);
1294 }
1295 
1296 void symbol__annotate_decay_histogram(struct symbol *sym, int evidx)
1297 {
1298 	struct annotation *notes = symbol__annotation(sym);
1299 	struct sym_hist *h = annotation__histogram(notes, evidx);
1300 	struct annotation_line *al;
1301 
1302 	h->nr_samples = 0;
1303 	list_for_each_entry(al, &notes->src->source, node) {
1304 		struct sym_hist_entry *entry;
1305 
1306 		if (al->offset == -1)
1307 			continue;
1308 
1309 		entry = annotated_source__hist_entry(notes->src, evidx, al->offset);
1310 		if (entry == NULL)
1311 			continue;
1312 
1313 		entry->nr_samples = entry->nr_samples * 7 / 8;
1314 		h->nr_samples += entry->nr_samples;
1315 	}
1316 }
1317 
1318 void annotated_source__purge(struct annotated_source *as)
1319 {
1320 	struct annotation_line *al, *n;
1321 
1322 	list_for_each_entry_safe(al, n, &as->source, node) {
1323 		list_del_init(&al->node);
1324 		disasm_line__free(disasm_line(al));
1325 	}
1326 }
1327 
1328 static size_t disasm_line__fprintf(struct disasm_line *dl, FILE *fp)
1329 {
1330 	size_t printed;
1331 
1332 	if (dl->al.offset == -1)
1333 		return fprintf(fp, "%s\n", dl->al.line);
1334 
1335 	printed = fprintf(fp, "%#" PRIx64 " %s", dl->al.offset, dl->ins.name);
1336 
1337 	if (dl->ops.raw[0] != '\0') {
1338 		printed += fprintf(fp, "%.*s %s\n", 6 - (int)printed, " ",
1339 				   dl->ops.raw);
1340 	}
1341 
1342 	return printed + fprintf(fp, "\n");
1343 }
1344 
1345 size_t disasm__fprintf(struct list_head *head, FILE *fp)
1346 {
1347 	struct disasm_line *pos;
1348 	size_t printed = 0;
1349 
1350 	list_for_each_entry(pos, head, al.node)
1351 		printed += disasm_line__fprintf(pos, fp);
1352 
1353 	return printed;
1354 }
1355 
1356 bool disasm_line__is_valid_local_jump(struct disasm_line *dl, struct symbol *sym)
1357 {
1358 	if (!dl || !dl->ins.ops || !ins__is_jump(&dl->ins) ||
1359 	    !disasm_line__has_local_offset(dl) || dl->ops.target.offset < 0 ||
1360 	    dl->ops.target.offset >= (s64)symbol__size(sym))
1361 		return false;
1362 
1363 	return true;
1364 }
1365 
1366 static void
1367 annotation__mark_jump_targets(struct annotation *notes, struct symbol *sym)
1368 {
1369 	struct annotation_line *al;
1370 
1371 	/* PLT symbols contain external offsets */
1372 	if (strstr(sym->name, "@plt"))
1373 		return;
1374 
1375 	list_for_each_entry(al, &notes->src->source, node) {
1376 		struct disasm_line *dl;
1377 		struct annotation_line *target;
1378 
1379 		dl = disasm_line(al);
1380 
1381 		if (!disasm_line__is_valid_local_jump(dl, sym))
1382 			continue;
1383 
1384 		target = annotated_source__get_line(notes->src,
1385 						    dl->ops.target.offset);
1386 		/*
1387 		 * FIXME: Oops, no jump target? Buggy disassembler? Or do we
1388 		 * have to adjust to the previous offset?
1389 		 */
1390 		if (target == NULL)
1391 			continue;
1392 
1393 		if (++target->jump_sources > notes->src->max_jump_sources)
1394 			notes->src->max_jump_sources = target->jump_sources;
1395 	}
1396 }
1397 
1398 static void annotation__set_index(struct annotation *notes)
1399 {
1400 	struct annotation_line *al;
1401 	struct annotated_source *src = notes->src;
1402 
1403 	src->widths.max_line_len = 0;
1404 	src->nr_entries = 0;
1405 	src->nr_asm_entries = 0;
1406 
1407 	list_for_each_entry(al, &src->source, node) {
1408 		size_t line_len = strlen(al->line);
1409 
1410 		if (src->widths.max_line_len < line_len)
1411 			src->widths.max_line_len = line_len;
1412 		al->idx = src->nr_entries++;
1413 		if (al->offset != -1)
1414 			al->idx_asm = src->nr_asm_entries++;
1415 		else
1416 			al->idx_asm = -1;
1417 	}
1418 }
1419 
1420 static inline int width_jumps(int n)
1421 {
1422 	if (n >= 100)
1423 		return 5;
1424 	if (n / 10)
1425 		return 2;
1426 	return 1;
1427 }
1428 
1429 static int annotation__max_ins_name(struct annotation *notes)
1430 {
1431 	int max_name = 0, len;
1432 	struct annotation_line *al;
1433 
1434         list_for_each_entry(al, &notes->src->source, node) {
1435 		if (al->offset == -1)
1436 			continue;
1437 
1438 		len = strlen(disasm_line(al)->ins.name);
1439 		if (max_name < len)
1440 			max_name = len;
1441 	}
1442 
1443 	return max_name;
1444 }
1445 
1446 static void
1447 annotation__init_column_widths(struct annotation *notes, struct symbol *sym)
1448 {
1449 	notes->src->widths.addr = notes->src->widths.target =
1450 		notes->src->widths.min_addr = hex_width(symbol__size(sym));
1451 	notes->src->widths.max_addr = hex_width(sym->end);
1452 	notes->src->widths.jumps = width_jumps(notes->src->max_jump_sources);
1453 	notes->src->widths.max_ins_name = annotation__max_ins_name(notes);
1454 }
1455 
1456 void annotation__update_column_widths(struct annotation *notes)
1457 {
1458 	if (annotate_opts.use_offset)
1459 		notes->src->widths.target = notes->src->widths.min_addr;
1460 	else if (annotate_opts.full_addr)
1461 		notes->src->widths.target = BITS_PER_LONG / 4;
1462 	else
1463 		notes->src->widths.target = notes->src->widths.max_addr;
1464 
1465 	notes->src->widths.addr = notes->src->widths.target;
1466 
1467 	if (annotate_opts.show_nr_jumps)
1468 		notes->src->widths.addr += notes->src->widths.jumps + 1;
1469 }
1470 
1471 void annotation__toggle_full_addr(struct annotation *notes, struct map_symbol *ms)
1472 {
1473 	annotate_opts.full_addr = !annotate_opts.full_addr;
1474 
1475 	if (annotate_opts.full_addr)
1476 		notes->src->start = map__objdump_2mem(ms->map, ms->sym->start);
1477 	else
1478 		notes->src->start = map__rip_2objdump(ms->map, ms->sym->start);
1479 
1480 	annotation__update_column_widths(notes);
1481 }
1482 
1483 static void annotation__calc_lines(struct annotation *notes, struct map_symbol *ms,
1484 				   struct rb_root *root)
1485 {
1486 	struct annotation_line *al;
1487 	struct rb_root tmp_root = RB_ROOT;
1488 
1489 	list_for_each_entry(al, &notes->src->source, node) {
1490 		double percent_max = 0.0;
1491 		u64 addr;
1492 		int i;
1493 
1494 		for (i = 0; i < al->data_nr; i++) {
1495 			double percent;
1496 
1497 			percent = annotation_data__percent(&al->data[i],
1498 							   annotate_opts.percent_type);
1499 
1500 			if (percent > percent_max)
1501 				percent_max = percent;
1502 		}
1503 
1504 		if (percent_max <= 0.5)
1505 			continue;
1506 
1507 		addr = map__rip_2objdump(ms->map, ms->sym->start);
1508 		al->path = get_srcline(map__dso(ms->map), addr + al->offset, NULL,
1509 				       false, true, ms->sym->start + al->offset);
1510 		insert_source_line(&tmp_root, al);
1511 	}
1512 
1513 	resort_source_line(root, &tmp_root);
1514 }
1515 
1516 static void symbol__calc_lines(struct map_symbol *ms, struct rb_root *root)
1517 {
1518 	struct annotation *notes = symbol__annotation(ms->sym);
1519 
1520 	annotation__calc_lines(notes, ms, root);
1521 }
1522 
1523 int symbol__tty_annotate2(struct map_symbol *ms, struct evsel *evsel)
1524 {
1525 	struct dso *dso = map__dso(ms->map);
1526 	struct symbol *sym = ms->sym;
1527 	struct rb_root source_line = RB_ROOT;
1528 	struct hists *hists = evsel__hists(evsel);
1529 	char buf[1024];
1530 	int err;
1531 
1532 	err = symbol__annotate2(ms, evsel, NULL);
1533 	if (err) {
1534 		char msg[BUFSIZ];
1535 
1536 		dso__set_annotate_warned(dso);
1537 		symbol__strerror_disassemble(ms, err, msg, sizeof(msg));
1538 		ui__error("Couldn't annotate %s:\n%s", sym->name, msg);
1539 		return -1;
1540 	}
1541 
1542 	if (annotate_opts.print_lines) {
1543 		srcline_full_filename = annotate_opts.full_path;
1544 		symbol__calc_lines(ms, &source_line);
1545 		print_summary(&source_line, dso__long_name(dso));
1546 	}
1547 
1548 	hists__scnprintf_title(hists, buf, sizeof(buf));
1549 	fprintf(stdout, "%s, [percent: %s]\n%s() %s\n",
1550 		buf, percent_type_str(annotate_opts.percent_type), sym->name, dso__long_name(dso));
1551 	symbol__annotate_fprintf2(sym, stdout);
1552 
1553 	annotated_source__purge(symbol__annotation(sym)->src);
1554 
1555 	return 0;
1556 }
1557 
1558 int symbol__tty_annotate(struct map_symbol *ms, struct evsel *evsel)
1559 {
1560 	struct dso *dso = map__dso(ms->map);
1561 	struct symbol *sym = ms->sym;
1562 	struct rb_root source_line = RB_ROOT;
1563 	int err;
1564 
1565 	err = symbol__annotate(ms, evsel, NULL);
1566 	if (err) {
1567 		char msg[BUFSIZ];
1568 
1569 		dso__set_annotate_warned(dso);
1570 		symbol__strerror_disassemble(ms, err, msg, sizeof(msg));
1571 		ui__error("Couldn't annotate %s:\n%s", sym->name, msg);
1572 		return -1;
1573 	}
1574 
1575 	symbol__calc_percent(sym, evsel);
1576 
1577 	if (annotate_opts.print_lines) {
1578 		srcline_full_filename = annotate_opts.full_path;
1579 		symbol__calc_lines(ms, &source_line);
1580 		print_summary(&source_line, dso__long_name(dso));
1581 	}
1582 
1583 	symbol__annotate_printf(ms, evsel);
1584 
1585 	annotated_source__purge(symbol__annotation(sym)->src);
1586 
1587 	return 0;
1588 }
1589 
1590 bool ui__has_annotation(void)
1591 {
1592 	return use_browser == 1 && perf_hpp_list.sym;
1593 }
1594 
1595 
1596 static double annotation_line__max_percent(struct annotation_line *al,
1597 					   struct annotation *notes,
1598 					   unsigned int percent_type)
1599 {
1600 	double percent_max = 0.0;
1601 	int i;
1602 
1603 	for (i = 0; i < notes->src->nr_events; i++) {
1604 		double percent;
1605 
1606 		percent = annotation_data__percent(&al->data[i],
1607 						   percent_type);
1608 
1609 		if (percent > percent_max)
1610 			percent_max = percent;
1611 	}
1612 
1613 	return percent_max;
1614 }
1615 
1616 static void disasm_line__write(struct disasm_line *dl, struct annotation *notes,
1617 			       void *obj, char *bf, size_t size,
1618 			       void (*obj__printf)(void *obj, const char *fmt, ...),
1619 			       void (*obj__write_graph)(void *obj, int graph))
1620 {
1621 	if (dl->ins.ops && dl->ins.ops->scnprintf) {
1622 		if (ins__is_jump(&dl->ins)) {
1623 			bool fwd;
1624 
1625 			if (dl->ops.target.outside)
1626 				goto call_like;
1627 			fwd = dl->ops.target.offset > dl->al.offset;
1628 			obj__write_graph(obj, fwd ? DARROW_CHAR : UARROW_CHAR);
1629 			obj__printf(obj, " ");
1630 		} else if (ins__is_call(&dl->ins)) {
1631 call_like:
1632 			obj__write_graph(obj, RARROW_CHAR);
1633 			obj__printf(obj, " ");
1634 		} else if (ins__is_ret(&dl->ins)) {
1635 			obj__write_graph(obj, LARROW_CHAR);
1636 			obj__printf(obj, " ");
1637 		} else {
1638 			obj__printf(obj, "  ");
1639 		}
1640 	} else {
1641 		obj__printf(obj, "  ");
1642 	}
1643 
1644 	disasm_line__scnprintf(dl, bf, size, !annotate_opts.use_offset,
1645 			       notes->src->widths.max_ins_name);
1646 }
1647 
1648 static void ipc_coverage_string(char *bf, int size, struct annotation *notes)
1649 {
1650 	double ipc = 0.0, coverage = 0.0;
1651 	struct annotated_branch *branch = annotation__get_branch(notes);
1652 
1653 	if (branch && branch->hit_cycles)
1654 		ipc = branch->hit_insn / ((double)branch->hit_cycles);
1655 
1656 	if (branch && branch->total_insn) {
1657 		coverage = branch->cover_insn * 100.0 /
1658 			((double)branch->total_insn);
1659 	}
1660 
1661 	scnprintf(bf, size, "(Average IPC: %.2f, IPC Coverage: %.1f%%)",
1662 		  ipc, coverage);
1663 }
1664 
1665 static void __annotation_line__write(struct annotation_line *al, struct annotation *notes,
1666 				     bool first_line, bool current_entry, bool change_color, int width,
1667 				     void *obj, unsigned int percent_type,
1668 				     int  (*obj__set_color)(void *obj, int color),
1669 				     void (*obj__set_percent_color)(void *obj, double percent, bool current),
1670 				     int  (*obj__set_jumps_percent_color)(void *obj, int nr, bool current),
1671 				     void (*obj__printf)(void *obj, const char *fmt, ...),
1672 				     void (*obj__write_graph)(void *obj, int graph))
1673 
1674 {
1675 	double percent_max = annotation_line__max_percent(al, notes, percent_type);
1676 	int pcnt_width = annotation__pcnt_width(notes),
1677 	    cycles_width = annotation__cycles_width(notes);
1678 	bool show_title = false;
1679 	char bf[256];
1680 	int printed;
1681 
1682 	if (first_line && (al->offset == -1 || percent_max == 0.0)) {
1683 		if (notes->branch && al->cycles) {
1684 			if (al->cycles->ipc == 0.0 && al->cycles->avg == 0)
1685 				show_title = true;
1686 		} else
1687 			show_title = true;
1688 	}
1689 
1690 	if (al->offset != -1 && percent_max != 0.0) {
1691 		int i;
1692 
1693 		for (i = 0; i < notes->src->nr_events; i++) {
1694 			double percent;
1695 
1696 			percent = annotation_data__percent(&al->data[i], percent_type);
1697 
1698 			obj__set_percent_color(obj, percent, current_entry);
1699 			if (symbol_conf.show_total_period) {
1700 				obj__printf(obj, "%11" PRIu64 " ", al->data[i].he.period);
1701 			} else if (symbol_conf.show_nr_samples) {
1702 				obj__printf(obj, "%6" PRIu64 " ",
1703 						   al->data[i].he.nr_samples);
1704 			} else {
1705 				obj__printf(obj, "%6.2f ", percent);
1706 			}
1707 		}
1708 	} else {
1709 		obj__set_percent_color(obj, 0, current_entry);
1710 
1711 		if (!show_title)
1712 			obj__printf(obj, "%-*s", pcnt_width, " ");
1713 		else {
1714 			obj__printf(obj, "%-*s", pcnt_width,
1715 					   symbol_conf.show_total_period ? "Period" :
1716 					   symbol_conf.show_nr_samples ? "Samples" : "Percent");
1717 		}
1718 	}
1719 
1720 	if (notes->branch) {
1721 		if (al->cycles && al->cycles->ipc)
1722 			obj__printf(obj, "%*.2f ", ANNOTATION__IPC_WIDTH - 1, al->cycles->ipc);
1723 		else if (!show_title)
1724 			obj__printf(obj, "%*s", ANNOTATION__IPC_WIDTH, " ");
1725 		else
1726 			obj__printf(obj, "%*s ", ANNOTATION__IPC_WIDTH - 1, "IPC");
1727 
1728 		if (!annotate_opts.show_minmax_cycle) {
1729 			if (al->cycles && al->cycles->avg)
1730 				obj__printf(obj, "%*" PRIu64 " ",
1731 					   ANNOTATION__CYCLES_WIDTH - 1, al->cycles->avg);
1732 			else if (!show_title)
1733 				obj__printf(obj, "%*s",
1734 					    ANNOTATION__CYCLES_WIDTH, " ");
1735 			else
1736 				obj__printf(obj, "%*s ",
1737 					    ANNOTATION__CYCLES_WIDTH - 1,
1738 					    "Cycle");
1739 		} else {
1740 			if (al->cycles) {
1741 				char str[32];
1742 
1743 				scnprintf(str, sizeof(str),
1744 					"%" PRIu64 "(%" PRIu64 "/%" PRIu64 ")",
1745 					al->cycles->avg, al->cycles->min,
1746 					al->cycles->max);
1747 
1748 				obj__printf(obj, "%*s ",
1749 					    ANNOTATION__MINMAX_CYCLES_WIDTH - 1,
1750 					    str);
1751 			} else if (!show_title)
1752 				obj__printf(obj, "%*s",
1753 					    ANNOTATION__MINMAX_CYCLES_WIDTH,
1754 					    " ");
1755 			else
1756 				obj__printf(obj, "%*s ",
1757 					    ANNOTATION__MINMAX_CYCLES_WIDTH - 1,
1758 					    "Cycle(min/max)");
1759 		}
1760 
1761 		if (show_title && !*al->line) {
1762 			ipc_coverage_string(bf, sizeof(bf), notes);
1763 			obj__printf(obj, "%*s", ANNOTATION__AVG_IPC_WIDTH, bf);
1764 		}
1765 	}
1766 
1767 	obj__printf(obj, " ");
1768 
1769 	if (!*al->line)
1770 		obj__printf(obj, "%-*s", width - pcnt_width - cycles_width, " ");
1771 	else if (al->offset == -1) {
1772 		if (al->line_nr && annotate_opts.show_linenr)
1773 			printed = scnprintf(bf, sizeof(bf), "%-*d ",
1774 					    notes->src->widths.addr + 1, al->line_nr);
1775 		else
1776 			printed = scnprintf(bf, sizeof(bf), "%-*s  ",
1777 					    notes->src->widths.addr, " ");
1778 		obj__printf(obj, bf);
1779 		obj__printf(obj, "%-*s", width - printed - pcnt_width - cycles_width + 1, al->line);
1780 	} else {
1781 		u64 addr = al->offset;
1782 		int color = -1;
1783 
1784 		if (!annotate_opts.use_offset)
1785 			addr += notes->src->start;
1786 
1787 		if (!annotate_opts.use_offset) {
1788 			printed = scnprintf(bf, sizeof(bf), "%" PRIx64 ": ", addr);
1789 		} else {
1790 			if (al->jump_sources &&
1791 			    annotate_opts.offset_level >= ANNOTATION__OFFSET_JUMP_TARGETS) {
1792 				if (annotate_opts.show_nr_jumps) {
1793 					int prev;
1794 					printed = scnprintf(bf, sizeof(bf), "%*d ",
1795 							    notes->src->widths.jumps,
1796 							    al->jump_sources);
1797 					prev = obj__set_jumps_percent_color(obj, al->jump_sources,
1798 									    current_entry);
1799 					obj__printf(obj, bf);
1800 					obj__set_color(obj, prev);
1801 				}
1802 print_addr:
1803 				printed = scnprintf(bf, sizeof(bf), "%*" PRIx64 ": ",
1804 						    notes->src->widths.target, addr);
1805 			} else if (ins__is_call(&disasm_line(al)->ins) &&
1806 				   annotate_opts.offset_level >= ANNOTATION__OFFSET_CALL) {
1807 				goto print_addr;
1808 			} else if (annotate_opts.offset_level == ANNOTATION__MAX_OFFSET_LEVEL) {
1809 				goto print_addr;
1810 			} else {
1811 				printed = scnprintf(bf, sizeof(bf), "%-*s  ",
1812 						    notes->src->widths.addr, " ");
1813 			}
1814 		}
1815 
1816 		if (change_color)
1817 			color = obj__set_color(obj, HE_COLORSET_ADDR);
1818 		obj__printf(obj, bf);
1819 		if (change_color)
1820 			obj__set_color(obj, color);
1821 
1822 		disasm_line__write(disasm_line(al), notes, obj, bf, sizeof(bf), obj__printf, obj__write_graph);
1823 
1824 		obj__printf(obj, "%-*s", width - pcnt_width - cycles_width - 3 - printed, bf);
1825 	}
1826 
1827 }
1828 
1829 void annotation_line__write(struct annotation_line *al, struct annotation *notes,
1830 			    struct annotation_write_ops *wops)
1831 {
1832 	__annotation_line__write(al, notes, wops->first_line, wops->current_entry,
1833 				 wops->change_color, wops->width, wops->obj,
1834 				 annotate_opts.percent_type,
1835 				 wops->set_color, wops->set_percent_color,
1836 				 wops->set_jumps_percent_color, wops->printf,
1837 				 wops->write_graph);
1838 }
1839 
1840 int symbol__annotate2(struct map_symbol *ms, struct evsel *evsel,
1841 		      struct arch **parch)
1842 {
1843 	struct symbol *sym = ms->sym;
1844 	struct annotation *notes = symbol__annotation(sym);
1845 	size_t size = symbol__size(sym);
1846 	int nr_pcnt = 1, err;
1847 
1848 	if (evsel__is_group_event(evsel))
1849 		nr_pcnt = evsel->core.nr_members;
1850 
1851 	err = symbol__annotate(ms, evsel, parch);
1852 	if (err)
1853 		return err;
1854 
1855 	symbol__calc_percent(sym, evsel);
1856 
1857 	annotation__set_index(notes);
1858 	annotation__mark_jump_targets(notes, sym);
1859 
1860 	err = annotation__compute_ipc(notes, size);
1861 	if (err)
1862 		return err;
1863 
1864 	annotation__init_column_widths(notes, sym);
1865 	notes->src->nr_events = nr_pcnt;
1866 
1867 	annotation__update_column_widths(notes);
1868 	sym->annotate2 = 1;
1869 
1870 	return 0;
1871 }
1872 
1873 static int annotation__config(const char *var, const char *value, void *data)
1874 {
1875 	struct annotation_options *opt = data;
1876 
1877 	if (!strstarts(var, "annotate."))
1878 		return 0;
1879 
1880 	if (!strcmp(var, "annotate.offset_level")) {
1881 		perf_config_u8(&opt->offset_level, "offset_level", value);
1882 
1883 		if (opt->offset_level > ANNOTATION__MAX_OFFSET_LEVEL)
1884 			opt->offset_level = ANNOTATION__MAX_OFFSET_LEVEL;
1885 		else if (opt->offset_level < ANNOTATION__MIN_OFFSET_LEVEL)
1886 			opt->offset_level = ANNOTATION__MIN_OFFSET_LEVEL;
1887 	} else if (!strcmp(var, "annotate.hide_src_code")) {
1888 		opt->hide_src_code = perf_config_bool("hide_src_code", value);
1889 	} else if (!strcmp(var, "annotate.jump_arrows")) {
1890 		opt->jump_arrows = perf_config_bool("jump_arrows", value);
1891 	} else if (!strcmp(var, "annotate.show_linenr")) {
1892 		opt->show_linenr = perf_config_bool("show_linenr", value);
1893 	} else if (!strcmp(var, "annotate.show_nr_jumps")) {
1894 		opt->show_nr_jumps = perf_config_bool("show_nr_jumps", value);
1895 	} else if (!strcmp(var, "annotate.show_nr_samples")) {
1896 		symbol_conf.show_nr_samples = perf_config_bool("show_nr_samples",
1897 								value);
1898 	} else if (!strcmp(var, "annotate.show_total_period")) {
1899 		symbol_conf.show_total_period = perf_config_bool("show_total_period",
1900 								value);
1901 	} else if (!strcmp(var, "annotate.use_offset")) {
1902 		opt->use_offset = perf_config_bool("use_offset", value);
1903 	} else if (!strcmp(var, "annotate.disassembler_style")) {
1904 		opt->disassembler_style = strdup(value);
1905 		if (!opt->disassembler_style) {
1906 			pr_err("Not enough memory for annotate.disassembler_style\n");
1907 			return -1;
1908 		}
1909 	} else if (!strcmp(var, "annotate.objdump")) {
1910 		opt->objdump_path = strdup(value);
1911 		if (!opt->objdump_path) {
1912 			pr_err("Not enough memory for annotate.objdump\n");
1913 			return -1;
1914 		}
1915 	} else if (!strcmp(var, "annotate.addr2line")) {
1916 		symbol_conf.addr2line_path = strdup(value);
1917 		if (!symbol_conf.addr2line_path) {
1918 			pr_err("Not enough memory for annotate.addr2line\n");
1919 			return -1;
1920 		}
1921 	} else if (!strcmp(var, "annotate.demangle")) {
1922 		symbol_conf.demangle = perf_config_bool("demangle", value);
1923 	} else if (!strcmp(var, "annotate.demangle_kernel")) {
1924 		symbol_conf.demangle_kernel = perf_config_bool("demangle_kernel", value);
1925 	} else {
1926 		pr_debug("%s variable unknown, ignoring...", var);
1927 	}
1928 
1929 	return 0;
1930 }
1931 
1932 void annotation_options__init(void)
1933 {
1934 	struct annotation_options *opt = &annotate_opts;
1935 
1936 	memset(opt, 0, sizeof(*opt));
1937 
1938 	/* Default values. */
1939 	opt->use_offset = true;
1940 	opt->jump_arrows = true;
1941 	opt->annotate_src = true;
1942 	opt->offset_level = ANNOTATION__OFFSET_JUMP_TARGETS;
1943 	opt->percent_type = PERCENT_PERIOD_LOCAL;
1944 }
1945 
1946 void annotation_options__exit(void)
1947 {
1948 	zfree(&annotate_opts.disassembler_style);
1949 	zfree(&annotate_opts.objdump_path);
1950 }
1951 
1952 void annotation_config__init(void)
1953 {
1954 	perf_config(annotation__config, &annotate_opts);
1955 }
1956 
1957 static unsigned int parse_percent_type(char *str1, char *str2)
1958 {
1959 	unsigned int type = (unsigned int) -1;
1960 
1961 	if (!strcmp("period", str1)) {
1962 		if (!strcmp("local", str2))
1963 			type = PERCENT_PERIOD_LOCAL;
1964 		else if (!strcmp("global", str2))
1965 			type = PERCENT_PERIOD_GLOBAL;
1966 	}
1967 
1968 	if (!strcmp("hits", str1)) {
1969 		if (!strcmp("local", str2))
1970 			type = PERCENT_HITS_LOCAL;
1971 		else if (!strcmp("global", str2))
1972 			type = PERCENT_HITS_GLOBAL;
1973 	}
1974 
1975 	return type;
1976 }
1977 
1978 int annotate_parse_percent_type(const struct option *opt __maybe_unused, const char *_str,
1979 				int unset __maybe_unused)
1980 {
1981 	unsigned int type;
1982 	char *str1, *str2;
1983 	int err = -1;
1984 
1985 	str1 = strdup(_str);
1986 	if (!str1)
1987 		return -ENOMEM;
1988 
1989 	str2 = strchr(str1, '-');
1990 	if (!str2)
1991 		goto out;
1992 
1993 	*str2++ = 0;
1994 
1995 	type = parse_percent_type(str1, str2);
1996 	if (type == (unsigned int) -1)
1997 		type = parse_percent_type(str2, str1);
1998 	if (type != (unsigned int) -1) {
1999 		annotate_opts.percent_type = type;
2000 		err = 0;
2001 	}
2002 
2003 out:
2004 	free(str1);
2005 	return err;
2006 }
2007 
2008 int annotate_check_args(void)
2009 {
2010 	struct annotation_options *args = &annotate_opts;
2011 
2012 	if (args->prefix_strip && !args->prefix) {
2013 		pr_err("--prefix-strip requires --prefix\n");
2014 		return -1;
2015 	}
2016 	return 0;
2017 }
2018 
2019 /*
2020  * Get register number and access offset from the given instruction.
2021  * It assumes AT&T x86 asm format like OFFSET(REG).  Maybe it needs
2022  * to revisit the format when it handles different architecture.
2023  * Fills @reg and @offset when return 0.
2024  */
2025 static int extract_reg_offset(struct arch *arch, const char *str,
2026 			      struct annotated_op_loc *op_loc)
2027 {
2028 	char *p;
2029 	char *regname;
2030 
2031 	if (arch->objdump.register_char == 0)
2032 		return -1;
2033 
2034 	/*
2035 	 * It should start from offset, but it's possible to skip 0
2036 	 * in the asm.  So 0(%rax) should be same as (%rax).
2037 	 *
2038 	 * However, it also start with a segment select register like
2039 	 * %gs:0x18(%rbx).  In that case it should skip the part.
2040 	 */
2041 	if (*str == arch->objdump.register_char) {
2042 		if (arch__is(arch, "x86")) {
2043 			/* FIXME: Handle other segment registers */
2044 			if (!strncmp(str, "%gs:", 4))
2045 				op_loc->segment = INSN_SEG_X86_GS;
2046 		}
2047 
2048 		while (*str && !isdigit(*str) &&
2049 		       *str != arch->objdump.memory_ref_char)
2050 			str++;
2051 	}
2052 
2053 	op_loc->offset = strtol(str, &p, 0);
2054 
2055 	p = strchr(p, arch->objdump.register_char);
2056 	if (p == NULL)
2057 		return -1;
2058 
2059 	regname = strdup(p);
2060 	if (regname == NULL)
2061 		return -1;
2062 
2063 	op_loc->reg1 = get_dwarf_regnum(regname, 0);
2064 	free(regname);
2065 
2066 	/* Get the second register */
2067 	if (op_loc->multi_regs) {
2068 		p = strchr(p + 1, arch->objdump.register_char);
2069 		if (p == NULL)
2070 			return -1;
2071 
2072 		regname = strdup(p);
2073 		if (regname == NULL)
2074 			return -1;
2075 
2076 		op_loc->reg2 = get_dwarf_regnum(regname, 0);
2077 		free(regname);
2078 	}
2079 	return 0;
2080 }
2081 
2082 /**
2083  * annotate_get_insn_location - Get location of instruction
2084  * @arch: the architecture info
2085  * @dl: the target instruction
2086  * @loc: a buffer to save the data
2087  *
2088  * Get detailed location info (register and offset) in the instruction.
2089  * It needs both source and target operand and whether it accesses a
2090  * memory location.  The offset field is meaningful only when the
2091  * corresponding mem flag is set.  The reg2 field is meaningful only
2092  * when multi_regs flag is set.
2093  *
2094  * Some examples on x86:
2095  *
2096  *   mov  (%rax), %rcx   # src_reg1 = rax, src_mem = 1, src_offset = 0
2097  *                       # dst_reg1 = rcx, dst_mem = 0
2098  *
2099  *   mov  0x18, %r8      # src_reg1 = -1, src_mem = 0
2100  *                       # dst_reg1 = r8, dst_mem = 0
2101  *
2102  *   mov  %rsi, 8(%rbx,%rcx,4)  # src_reg1 = rsi, src_mem = 0, src_multi_regs = 0
2103  *                              # dst_reg1 = rbx, dst_reg2 = rcx, dst_mem = 1
2104  *                              # dst_multi_regs = 1, dst_offset = 8
2105  */
2106 int annotate_get_insn_location(struct arch *arch, struct disasm_line *dl,
2107 			       struct annotated_insn_loc *loc)
2108 {
2109 	struct ins_operands *ops;
2110 	struct annotated_op_loc *op_loc;
2111 	int i;
2112 
2113 	if (ins__is_lock(&dl->ins))
2114 		ops = dl->ops.locked.ops;
2115 	else
2116 		ops = &dl->ops;
2117 
2118 	if (ops == NULL)
2119 		return -1;
2120 
2121 	memset(loc, 0, sizeof(*loc));
2122 
2123 	for_each_insn_op_loc(loc, i, op_loc) {
2124 		const char *insn_str = ops->source.raw;
2125 		bool multi_regs = ops->source.multi_regs;
2126 
2127 		if (i == INSN_OP_TARGET) {
2128 			insn_str = ops->target.raw;
2129 			multi_regs = ops->target.multi_regs;
2130 		}
2131 
2132 		/* Invalidate the register by default */
2133 		op_loc->reg1 = -1;
2134 		op_loc->reg2 = -1;
2135 
2136 		if (insn_str == NULL)
2137 			continue;
2138 
2139 		if (strchr(insn_str, arch->objdump.memory_ref_char)) {
2140 			op_loc->mem_ref = true;
2141 			op_loc->multi_regs = multi_regs;
2142 			extract_reg_offset(arch, insn_str, op_loc);
2143 		} else {
2144 			char *s, *p = NULL;
2145 
2146 			if (arch__is(arch, "x86")) {
2147 				/* FIXME: Handle other segment registers */
2148 				if (!strncmp(insn_str, "%gs:", 4)) {
2149 					op_loc->segment = INSN_SEG_X86_GS;
2150 					op_loc->offset = strtol(insn_str + 4,
2151 								&p, 0);
2152 					if (p && p != insn_str + 4)
2153 						op_loc->imm = true;
2154 					continue;
2155 				}
2156 			}
2157 
2158 			s = strdup(insn_str);
2159 			if (s == NULL)
2160 				return -1;
2161 
2162 			if (*s == arch->objdump.register_char)
2163 				op_loc->reg1 = get_dwarf_regnum(s, 0);
2164 			else if (*s == arch->objdump.imm_char) {
2165 				op_loc->offset = strtol(s + 1, &p, 0);
2166 				if (p && p != s + 1)
2167 					op_loc->imm = true;
2168 			}
2169 			free(s);
2170 		}
2171 	}
2172 
2173 	return 0;
2174 }
2175 
2176 static struct disasm_line *find_disasm_line(struct symbol *sym, u64 ip,
2177 					    bool allow_update)
2178 {
2179 	struct disasm_line *dl;
2180 	struct annotation *notes;
2181 
2182 	notes = symbol__annotation(sym);
2183 
2184 	list_for_each_entry(dl, &notes->src->source, al.node) {
2185 		if (dl->al.offset == -1)
2186 			continue;
2187 
2188 		if (sym->start + dl->al.offset == ip) {
2189 			/*
2190 			 * llvm-objdump places "lock" in a separate line and
2191 			 * in that case, we want to get the next line.
2192 			 */
2193 			if (ins__is_lock(&dl->ins) &&
2194 			    *dl->ops.raw == '\0' && allow_update) {
2195 				ip++;
2196 				continue;
2197 			}
2198 			return dl;
2199 		}
2200 	}
2201 	return NULL;
2202 }
2203 
2204 static struct annotated_item_stat *annotate_data_stat(struct list_head *head,
2205 						      const char *name)
2206 {
2207 	struct annotated_item_stat *istat;
2208 
2209 	list_for_each_entry(istat, head, list) {
2210 		if (!strcmp(istat->name, name))
2211 			return istat;
2212 	}
2213 
2214 	istat = zalloc(sizeof(*istat));
2215 	if (istat == NULL)
2216 		return NULL;
2217 
2218 	istat->name = strdup(name);
2219 	if (istat->name == NULL) {
2220 		free(istat);
2221 		return NULL;
2222 	}
2223 
2224 	list_add_tail(&istat->list, head);
2225 	return istat;
2226 }
2227 
2228 static bool is_stack_operation(struct arch *arch, struct disasm_line *dl)
2229 {
2230 	if (arch__is(arch, "x86")) {
2231 		if (!strncmp(dl->ins.name, "push", 4) ||
2232 		    !strncmp(dl->ins.name, "pop", 3) ||
2233 		    !strncmp(dl->ins.name, "ret", 3))
2234 			return true;
2235 	}
2236 
2237 	return false;
2238 }
2239 
2240 static bool is_stack_canary(struct arch *arch, struct annotated_op_loc *loc)
2241 {
2242 	/* On x86_64, %gs:40 is used for stack canary */
2243 	if (arch__is(arch, "x86")) {
2244 		if (loc->segment == INSN_SEG_X86_GS && loc->imm &&
2245 		    loc->offset == 40)
2246 			return true;
2247 	}
2248 
2249 	return false;
2250 }
2251 
2252 static struct disasm_line *
2253 annotation__prev_asm_line(struct annotation *notes, struct disasm_line *curr)
2254 {
2255 	struct list_head *sources = &notes->src->source;
2256 	struct disasm_line *prev;
2257 
2258 	if (curr == list_first_entry(sources, struct disasm_line, al.node))
2259 		return NULL;
2260 
2261 	prev = list_prev_entry(curr, al.node);
2262 	while (prev->al.offset == -1 &&
2263 	       prev != list_first_entry(sources, struct disasm_line, al.node))
2264 		prev = list_prev_entry(prev, al.node);
2265 
2266 	if (prev->al.offset == -1)
2267 		return NULL;
2268 
2269 	return prev;
2270 }
2271 
2272 static struct disasm_line *
2273 annotation__next_asm_line(struct annotation *notes, struct disasm_line *curr)
2274 {
2275 	struct list_head *sources = &notes->src->source;
2276 	struct disasm_line *next;
2277 
2278 	if (curr == list_last_entry(sources, struct disasm_line, al.node))
2279 		return NULL;
2280 
2281 	next = list_next_entry(curr, al.node);
2282 	while (next->al.offset == -1 &&
2283 	       next != list_last_entry(sources, struct disasm_line, al.node))
2284 		next = list_next_entry(next, al.node);
2285 
2286 	if (next->al.offset == -1)
2287 		return NULL;
2288 
2289 	return next;
2290 }
2291 
2292 u64 annotate_calc_pcrel(struct map_symbol *ms, u64 ip, int offset,
2293 			struct disasm_line *dl)
2294 {
2295 	struct annotation *notes;
2296 	struct disasm_line *next;
2297 	u64 addr;
2298 
2299 	notes = symbol__annotation(ms->sym);
2300 	/*
2301 	 * PC-relative addressing starts from the next instruction address
2302 	 * But the IP is for the current instruction.  Since disasm_line
2303 	 * doesn't have the instruction size, calculate it using the next
2304 	 * disasm_line.  If it's the last one, we can use symbol's end
2305 	 * address directly.
2306 	 */
2307 	next = annotation__next_asm_line(notes, dl);
2308 	if (next == NULL)
2309 		addr = ms->sym->end + offset;
2310 	else
2311 		addr = ip + (next->al.offset - dl->al.offset) + offset;
2312 
2313 	return map__rip_2objdump(ms->map, addr);
2314 }
2315 
2316 /**
2317  * hist_entry__get_data_type - find data type for given hist entry
2318  * @he: hist entry
2319  *
2320  * This function first annotates the instruction at @he->ip and extracts
2321  * register and offset info from it.  Then it searches the DWARF debug
2322  * info to get a variable and type information using the address, register,
2323  * and offset.
2324  */
2325 struct annotated_data_type *hist_entry__get_data_type(struct hist_entry *he)
2326 {
2327 	struct map_symbol *ms = &he->ms;
2328 	struct evsel *evsel = hists_to_evsel(he->hists);
2329 	struct arch *arch;
2330 	struct disasm_line *dl;
2331 	struct annotated_insn_loc loc;
2332 	struct annotated_op_loc *op_loc;
2333 	struct annotated_data_type *mem_type;
2334 	struct annotated_item_stat *istat;
2335 	u64 ip = he->ip;
2336 	int i;
2337 
2338 	ann_data_stat.total++;
2339 
2340 	if (ms->map == NULL || ms->sym == NULL) {
2341 		ann_data_stat.no_sym++;
2342 		return NULL;
2343 	}
2344 
2345 	if (!symbol_conf.init_annotation) {
2346 		ann_data_stat.no_sym++;
2347 		return NULL;
2348 	}
2349 
2350 	/* Make sure it has the disasm of the function */
2351 	if (symbol__annotate(ms, evsel, &arch) < 0) {
2352 		ann_data_stat.no_insn++;
2353 		return NULL;
2354 	}
2355 
2356 	/*
2357 	 * Get a disasm to extract the location from the insn.
2358 	 * This is too slow...
2359 	 */
2360 	dl = find_disasm_line(ms->sym, ip, /*allow_update=*/true);
2361 	if (dl == NULL) {
2362 		ann_data_stat.no_insn++;
2363 		return NULL;
2364 	}
2365 
2366 retry:
2367 	istat = annotate_data_stat(&ann_insn_stat, dl->ins.name);
2368 	if (istat == NULL) {
2369 		ann_data_stat.no_insn++;
2370 		return NULL;
2371 	}
2372 
2373 	if (annotate_get_insn_location(arch, dl, &loc) < 0) {
2374 		ann_data_stat.no_insn_ops++;
2375 		istat->bad++;
2376 		return NULL;
2377 	}
2378 
2379 	if (is_stack_operation(arch, dl)) {
2380 		istat->good++;
2381 		he->mem_type_off = 0;
2382 		return &stackop_type;
2383 	}
2384 
2385 	for_each_insn_op_loc(&loc, i, op_loc) {
2386 		struct data_loc_info dloc = {
2387 			.arch = arch,
2388 			.thread = he->thread,
2389 			.ms = ms,
2390 			/* Recalculate IP for LOCK prefix or insn fusion */
2391 			.ip = ms->sym->start + dl->al.offset,
2392 			.cpumode = he->cpumode,
2393 			.op = op_loc,
2394 		};
2395 
2396 		if (!op_loc->mem_ref && op_loc->segment == INSN_SEG_NONE)
2397 			continue;
2398 
2399 		/* Recalculate IP because of LOCK prefix or insn fusion */
2400 		ip = ms->sym->start + dl->al.offset;
2401 
2402 		/* PC-relative addressing */
2403 		if (op_loc->reg1 == DWARF_REG_PC) {
2404 			dloc.var_addr = annotate_calc_pcrel(ms, dloc.ip,
2405 							    op_loc->offset, dl);
2406 		}
2407 
2408 		/* This CPU access in kernel - pretend PC-relative addressing */
2409 		if (dso__kernel(map__dso(ms->map)) && arch__is(arch, "x86") &&
2410 		    op_loc->segment == INSN_SEG_X86_GS && op_loc->imm) {
2411 			dloc.var_addr = op_loc->offset;
2412 			op_loc->reg1 = DWARF_REG_PC;
2413 		}
2414 
2415 		mem_type = find_data_type(&dloc);
2416 
2417 		if (mem_type == NULL && is_stack_canary(arch, op_loc)) {
2418 			istat->good++;
2419 			he->mem_type_off = 0;
2420 			return &canary_type;
2421 		}
2422 
2423 		if (mem_type)
2424 			istat->good++;
2425 		else
2426 			istat->bad++;
2427 
2428 		if (symbol_conf.annotate_data_sample) {
2429 			annotated_data_type__update_samples(mem_type, evsel,
2430 							    dloc.type_offset,
2431 							    he->stat.nr_events,
2432 							    he->stat.period);
2433 		}
2434 		he->mem_type_off = dloc.type_offset;
2435 		return mem_type;
2436 	}
2437 
2438 	/*
2439 	 * Some instructions can be fused and the actual memory access came
2440 	 * from the previous instruction.
2441 	 */
2442 	if (dl->al.offset > 0) {
2443 		struct annotation *notes;
2444 		struct disasm_line *prev_dl;
2445 
2446 		notes = symbol__annotation(ms->sym);
2447 		prev_dl = annotation__prev_asm_line(notes, dl);
2448 
2449 		if (prev_dl && ins__is_fused(arch, prev_dl->ins.name, dl->ins.name)) {
2450 			dl = prev_dl;
2451 			goto retry;
2452 		}
2453 	}
2454 
2455 	ann_data_stat.no_mem_ops++;
2456 	istat->bad++;
2457 	return NULL;
2458 }
2459 
2460 /* Basic block traversal (BFS) data structure */
2461 struct basic_block_data {
2462 	struct list_head queue;
2463 	struct list_head visited;
2464 };
2465 
2466 /*
2467  * During the traversal, it needs to know the parent block where the current
2468  * block block started from.  Note that single basic block can be parent of
2469  * two child basic blocks (in case of condition jump).
2470  */
2471 struct basic_block_link {
2472 	struct list_head node;
2473 	struct basic_block_link *parent;
2474 	struct annotated_basic_block *bb;
2475 };
2476 
2477 /* Check any of basic block in the list already has the offset */
2478 static bool basic_block_has_offset(struct list_head *head, s64 offset)
2479 {
2480 	struct basic_block_link *link;
2481 
2482 	list_for_each_entry(link, head, node) {
2483 		s64 begin_offset = link->bb->begin->al.offset;
2484 		s64 end_offset = link->bb->end->al.offset;
2485 
2486 		if (begin_offset <= offset && offset <= end_offset)
2487 			return true;
2488 	}
2489 	return false;
2490 }
2491 
2492 static bool is_new_basic_block(struct basic_block_data *bb_data,
2493 			       struct disasm_line *dl)
2494 {
2495 	s64 offset = dl->al.offset;
2496 
2497 	if (basic_block_has_offset(&bb_data->visited, offset))
2498 		return false;
2499 	if (basic_block_has_offset(&bb_data->queue, offset))
2500 		return false;
2501 	return true;
2502 }
2503 
2504 /* Add a basic block starting from dl and link it to the parent */
2505 static int add_basic_block(struct basic_block_data *bb_data,
2506 			   struct basic_block_link *parent,
2507 			   struct disasm_line *dl)
2508 {
2509 	struct annotated_basic_block *bb;
2510 	struct basic_block_link *link;
2511 
2512 	if (dl == NULL)
2513 		return -1;
2514 
2515 	if (!is_new_basic_block(bb_data, dl))
2516 		return 0;
2517 
2518 	bb = zalloc(sizeof(*bb));
2519 	if (bb == NULL)
2520 		return -1;
2521 
2522 	bb->begin = dl;
2523 	bb->end = dl;
2524 	INIT_LIST_HEAD(&bb->list);
2525 
2526 	link = malloc(sizeof(*link));
2527 	if (link == NULL) {
2528 		free(bb);
2529 		return -1;
2530 	}
2531 
2532 	link->bb = bb;
2533 	link->parent = parent;
2534 	list_add_tail(&link->node, &bb_data->queue);
2535 	return 0;
2536 }
2537 
2538 /* Returns true when it finds the target in the current basic block */
2539 static bool process_basic_block(struct basic_block_data *bb_data,
2540 				struct basic_block_link *link,
2541 				struct symbol *sym, u64 target)
2542 {
2543 	struct disasm_line *dl, *next_dl, *last_dl;
2544 	struct annotation *notes = symbol__annotation(sym);
2545 	bool found = false;
2546 
2547 	dl = link->bb->begin;
2548 	/* Check if it's already visited */
2549 	if (basic_block_has_offset(&bb_data->visited, dl->al.offset))
2550 		return false;
2551 
2552 	last_dl = list_last_entry(&notes->src->source,
2553 				  struct disasm_line, al.node);
2554 	if (last_dl->al.offset == -1)
2555 		last_dl = annotation__prev_asm_line(notes, last_dl);
2556 
2557 	if (last_dl == NULL)
2558 		return false;
2559 
2560 	list_for_each_entry_from(dl, &notes->src->source, al.node) {
2561 		/* Skip comment or debug info line */
2562 		if (dl->al.offset == -1)
2563 			continue;
2564 		/* Found the target instruction */
2565 		if (sym->start + dl->al.offset == target) {
2566 			found = true;
2567 			break;
2568 		}
2569 		/* End of the function, finish the block */
2570 		if (dl == last_dl)
2571 			break;
2572 		/* 'return' instruction finishes the block */
2573 		if (ins__is_ret(&dl->ins))
2574 			break;
2575 		/* normal instructions are part of the basic block */
2576 		if (!ins__is_jump(&dl->ins))
2577 			continue;
2578 		/* jump to a different function, tail call or return */
2579 		if (dl->ops.target.outside)
2580 			break;
2581 		/* jump instruction creates new basic block(s) */
2582 		next_dl = find_disasm_line(sym, sym->start + dl->ops.target.offset,
2583 					   /*allow_update=*/false);
2584 		if (next_dl)
2585 			add_basic_block(bb_data, link, next_dl);
2586 
2587 		/*
2588 		 * FIXME: determine conditional jumps properly.
2589 		 * Conditional jumps create another basic block with the
2590 		 * next disasm line.
2591 		 */
2592 		if (!strstr(dl->ins.name, "jmp")) {
2593 			next_dl = annotation__next_asm_line(notes, dl);
2594 			if (next_dl)
2595 				add_basic_block(bb_data, link, next_dl);
2596 		}
2597 		break;
2598 
2599 	}
2600 	link->bb->end = dl;
2601 	return found;
2602 }
2603 
2604 /*
2605  * It founds a target basic block, build a proper linked list of basic blocks
2606  * by following the link recursively.
2607  */
2608 static void link_found_basic_blocks(struct basic_block_link *link,
2609 				    struct list_head *head)
2610 {
2611 	while (link) {
2612 		struct basic_block_link *parent = link->parent;
2613 
2614 		list_move(&link->bb->list, head);
2615 		list_del(&link->node);
2616 		free(link);
2617 
2618 		link = parent;
2619 	}
2620 }
2621 
2622 static void delete_basic_blocks(struct basic_block_data *bb_data)
2623 {
2624 	struct basic_block_link *link, *tmp;
2625 
2626 	list_for_each_entry_safe(link, tmp, &bb_data->queue, node) {
2627 		list_del(&link->node);
2628 		zfree(&link->bb);
2629 		free(link);
2630 	}
2631 
2632 	list_for_each_entry_safe(link, tmp, &bb_data->visited, node) {
2633 		list_del(&link->node);
2634 		zfree(&link->bb);
2635 		free(link);
2636 	}
2637 }
2638 
2639 /**
2640  * annotate_get_basic_blocks - Get basic blocks for given address range
2641  * @sym: symbol to annotate
2642  * @src: source address
2643  * @dst: destination address
2644  * @head: list head to save basic blocks
2645  *
2646  * This function traverses disasm_lines from @src to @dst and save them in a
2647  * list of annotated_basic_block to @head.  It uses BFS to find the shortest
2648  * path between two.  The basic_block_link is to maintain parent links so
2649  * that it can build a list of blocks from the start.
2650  */
2651 int annotate_get_basic_blocks(struct symbol *sym, s64 src, s64 dst,
2652 			      struct list_head *head)
2653 {
2654 	struct basic_block_data bb_data = {
2655 		.queue = LIST_HEAD_INIT(bb_data.queue),
2656 		.visited = LIST_HEAD_INIT(bb_data.visited),
2657 	};
2658 	struct basic_block_link *link;
2659 	struct disasm_line *dl;
2660 	int ret = -1;
2661 
2662 	dl = find_disasm_line(sym, src, /*allow_update=*/false);
2663 	if (dl == NULL)
2664 		return -1;
2665 
2666 	if (add_basic_block(&bb_data, /*parent=*/NULL, dl) < 0)
2667 		return -1;
2668 
2669 	/* Find shortest path from src to dst using BFS */
2670 	while (!list_empty(&bb_data.queue)) {
2671 		link = list_first_entry(&bb_data.queue, struct basic_block_link, node);
2672 
2673 		if (process_basic_block(&bb_data, link, sym, dst)) {
2674 			link_found_basic_blocks(link, head);
2675 			ret = 0;
2676 			break;
2677 		}
2678 		list_move(&link->node, &bb_data.visited);
2679 	}
2680 	delete_basic_blocks(&bb_data);
2681 	return ret;
2682 }
2683