1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (C) 2011, Red Hat Inc, Arnaldo Carvalho de Melo <acme@redhat.com> 4 * 5 * Parts came from builtin-annotate.c, see those files for further 6 * copyright notes. 7 */ 8 9 #include <errno.h> 10 #include <inttypes.h> 11 #include <libgen.h> 12 #include <stdlib.h> 13 #include "util.h" // hex_width() 14 #include "ui/ui.h" 15 #include "sort.h" 16 #include "build-id.h" 17 #include "color.h" 18 #include "config.h" 19 #include "disasm.h" 20 #include "dso.h" 21 #include "env.h" 22 #include "map.h" 23 #include "maps.h" 24 #include "symbol.h" 25 #include "srcline.h" 26 #include "units.h" 27 #include "debug.h" 28 #include "debuginfo.h" 29 #include "annotate.h" 30 #include "annotate-data.h" 31 #include "evsel.h" 32 #include "evlist.h" 33 #include "bpf-event.h" 34 #include "bpf-utils.h" 35 #include "block-range.h" 36 #include "string2.h" 37 #include "dwarf-regs.h" 38 #include "util/event.h" 39 #include "util/sharded_mutex.h" 40 #include "arch/common.h" 41 #include "namespaces.h" 42 #include "thread.h" 43 #include "hashmap.h" 44 #include "strbuf.h" 45 #include <regex.h> 46 #include <linux/bitops.h> 47 #include <linux/kernel.h> 48 #include <linux/string.h> 49 #include <linux/zalloc.h> 50 #include <subcmd/parse-options.h> 51 #include <subcmd/run-command.h> 52 #include <math.h> 53 54 /* FIXME: For the HE_COLORSET */ 55 #include "ui/browser.h" 56 57 /* 58 * FIXME: Using the same values as slang.h, 59 * but that header may not be available everywhere 60 */ 61 #define LARROW_CHAR ((unsigned char)',') 62 #define RARROW_CHAR ((unsigned char)'+') 63 #define DARROW_CHAR ((unsigned char)'.') 64 #define UARROW_CHAR ((unsigned char)'-') 65 66 #include <linux/ctype.h> 67 68 /* global annotation options */ 69 struct annotation_options annotate_opts; 70 71 /* Data type collection debug statistics */ 72 struct annotated_data_stat ann_data_stat; 73 LIST_HEAD(ann_insn_stat); 74 75 /* Pseudo data types */ 76 struct annotated_data_type stackop_type = { 77 .self = { 78 .type_name = (char *)"(stack operation)", 79 .children = LIST_HEAD_INIT(stackop_type.self.children), 80 }, 81 }; 82 83 struct annotated_data_type canary_type = { 84 .self = { 85 .type_name = (char *)"(stack canary)", 86 .children = LIST_HEAD_INIT(canary_type.self.children), 87 }, 88 }; 89 90 /* symbol histogram: key = offset << 16 | evsel->core.idx */ 91 static size_t sym_hist_hash(long key, void *ctx __maybe_unused) 92 { 93 return (key >> 16) + (key & 0xffff); 94 } 95 96 static bool sym_hist_equal(long key1, long key2, void *ctx __maybe_unused) 97 { 98 return key1 == key2; 99 } 100 101 static struct annotated_source *annotated_source__new(void) 102 { 103 struct annotated_source *src = zalloc(sizeof(*src)); 104 105 if (src != NULL) 106 INIT_LIST_HEAD(&src->source); 107 108 return src; 109 } 110 111 static __maybe_unused void annotated_source__delete(struct annotated_source *src) 112 { 113 struct hashmap_entry *cur; 114 size_t bkt; 115 116 if (src == NULL) 117 return; 118 119 if (src->samples) { 120 hashmap__for_each_entry(src->samples, cur, bkt) 121 zfree(&cur->pvalue); 122 hashmap__free(src->samples); 123 } 124 zfree(&src->histograms); 125 free(src); 126 } 127 128 static int annotated_source__alloc_histograms(struct annotated_source *src, 129 int nr_hists) 130 { 131 src->nr_histograms = nr_hists; 132 src->histograms = calloc(nr_hists, sizeof(*src->histograms)); 133 134 if (src->histograms == NULL) 135 return -1; 136 137 src->samples = hashmap__new(sym_hist_hash, sym_hist_equal, NULL); 138 if (src->samples == NULL) 139 zfree(&src->histograms); 140 141 return src->histograms ? 0 : -1; 142 } 143 144 void symbol__annotate_zero_histograms(struct symbol *sym) 145 { 146 struct annotation *notes = symbol__annotation(sym); 147 148 annotation__lock(notes); 149 if (notes->src != NULL) { 150 memset(notes->src->histograms, 0, 151 notes->src->nr_histograms * sizeof(*notes->src->histograms)); 152 hashmap__clear(notes->src->samples); 153 } 154 if (notes->branch && notes->branch->cycles_hist) { 155 memset(notes->branch->cycles_hist, 0, 156 symbol__size(sym) * sizeof(struct cyc_hist)); 157 } 158 annotation__unlock(notes); 159 } 160 161 static int __symbol__account_cycles(struct cyc_hist *ch, 162 u64 start, 163 unsigned offset, unsigned cycles, 164 unsigned have_start) 165 { 166 /* 167 * For now we can only account one basic block per 168 * final jump. But multiple could be overlapping. 169 * Always account the longest one. So when 170 * a shorter one has been already seen throw it away. 171 * 172 * We separately always account the full cycles. 173 */ 174 ch[offset].num_aggr++; 175 ch[offset].cycles_aggr += cycles; 176 177 if (cycles > ch[offset].cycles_max) 178 ch[offset].cycles_max = cycles; 179 180 if (ch[offset].cycles_min) { 181 if (cycles && cycles < ch[offset].cycles_min) 182 ch[offset].cycles_min = cycles; 183 } else 184 ch[offset].cycles_min = cycles; 185 186 if (!have_start && ch[offset].have_start) 187 return 0; 188 if (ch[offset].num) { 189 if (have_start && (!ch[offset].have_start || 190 ch[offset].start > start)) { 191 ch[offset].have_start = 0; 192 ch[offset].cycles = 0; 193 ch[offset].num = 0; 194 if (ch[offset].reset < 0xffff) 195 ch[offset].reset++; 196 } else if (have_start && 197 ch[offset].start < start) 198 return 0; 199 } 200 201 if (ch[offset].num < NUM_SPARKS) 202 ch[offset].cycles_spark[ch[offset].num] = cycles; 203 204 ch[offset].have_start = have_start; 205 ch[offset].start = start; 206 ch[offset].cycles += cycles; 207 ch[offset].num++; 208 return 0; 209 } 210 211 static int __symbol__inc_addr_samples(struct map_symbol *ms, 212 struct annotated_source *src, struct evsel *evsel, u64 addr, 213 struct perf_sample *sample) 214 { 215 struct symbol *sym = ms->sym; 216 long hash_key; 217 u64 offset; 218 struct sym_hist *h; 219 struct sym_hist_entry *entry; 220 221 pr_debug3("%s: addr=%#" PRIx64 "\n", __func__, map__unmap_ip(ms->map, addr)); 222 223 if ((addr < sym->start || addr >= sym->end) && 224 (addr != sym->end || sym->start != sym->end)) { 225 pr_debug("%s(%d): ERANGE! sym->name=%s, start=%#" PRIx64 ", addr=%#" PRIx64 ", end=%#" PRIx64 "\n", 226 __func__, __LINE__, sym->name, sym->start, addr, sym->end); 227 return -ERANGE; 228 } 229 230 offset = addr - sym->start; 231 h = annotated_source__histogram(src, evsel); 232 if (h == NULL) { 233 pr_debug("%s(%d): ENOMEM! sym->name=%s, start=%#" PRIx64 ", addr=%#" PRIx64 ", end=%#" PRIx64 ", func: %d\n", 234 __func__, __LINE__, sym->name, sym->start, addr, sym->end, sym->type == STT_FUNC); 235 return -ENOMEM; 236 } 237 238 hash_key = offset << 16 | evsel->core.idx; 239 if (!hashmap__find(src->samples, hash_key, &entry)) { 240 entry = zalloc(sizeof(*entry)); 241 if (entry == NULL) 242 return -ENOMEM; 243 244 if (hashmap__add(src->samples, hash_key, entry) < 0) 245 return -ENOMEM; 246 } 247 248 h->nr_samples++; 249 h->period += sample->period; 250 entry->nr_samples++; 251 entry->period += sample->period; 252 253 pr_debug3("%#" PRIx64 " %s: period++ [addr: %#" PRIx64 ", %#" PRIx64 254 ", evidx=%d] => nr_samples: %" PRIu64 ", period: %" PRIu64 "\n", 255 sym->start, sym->name, addr, addr - sym->start, evsel->core.idx, 256 entry->nr_samples, entry->period); 257 return 0; 258 } 259 260 struct annotated_branch *annotation__get_branch(struct annotation *notes) 261 { 262 if (notes == NULL) 263 return NULL; 264 265 if (notes->branch == NULL) 266 notes->branch = zalloc(sizeof(*notes->branch)); 267 268 return notes->branch; 269 } 270 271 static struct annotated_branch *symbol__find_branch_hist(struct symbol *sym, 272 unsigned int br_cntr_nr) 273 { 274 struct annotation *notes = symbol__annotation(sym); 275 struct annotated_branch *branch; 276 const size_t size = symbol__size(sym); 277 278 branch = annotation__get_branch(notes); 279 if (branch == NULL) 280 return NULL; 281 282 if (branch->cycles_hist == NULL) { 283 branch->cycles_hist = calloc(size, sizeof(struct cyc_hist)); 284 if (!branch->cycles_hist) 285 return NULL; 286 } 287 288 if (br_cntr_nr && branch->br_cntr == NULL) { 289 branch->br_cntr = calloc(br_cntr_nr * size, sizeof(u64)); 290 if (!branch->br_cntr) 291 return NULL; 292 } 293 294 return branch; 295 } 296 297 struct annotated_source *symbol__hists(struct symbol *sym, int nr_hists) 298 { 299 struct annotation *notes = symbol__annotation(sym); 300 301 if (notes->src == NULL) { 302 notes->src = annotated_source__new(); 303 if (notes->src == NULL) 304 return NULL; 305 goto alloc_histograms; 306 } 307 308 if (notes->src->histograms == NULL) { 309 alloc_histograms: 310 annotated_source__alloc_histograms(notes->src, nr_hists); 311 } 312 313 return notes->src; 314 } 315 316 static int symbol__inc_addr_samples(struct map_symbol *ms, 317 struct evsel *evsel, u64 addr, 318 struct perf_sample *sample) 319 { 320 struct symbol *sym = ms->sym; 321 struct annotated_source *src; 322 323 if (sym == NULL) 324 return 0; 325 src = symbol__hists(sym, evsel->evlist->core.nr_entries); 326 return src ? __symbol__inc_addr_samples(ms, src, evsel, addr, sample) : 0; 327 } 328 329 static int symbol__account_br_cntr(struct annotated_branch *branch, 330 struct evsel *evsel, 331 unsigned offset, 332 u64 br_cntr) 333 { 334 unsigned int br_cntr_nr = evsel__leader(evsel)->br_cntr_nr; 335 unsigned int base = evsel__leader(evsel)->br_cntr_idx; 336 unsigned int off = offset * evsel->evlist->nr_br_cntr; 337 u64 *branch_br_cntr = branch->br_cntr; 338 unsigned int i, mask, width; 339 340 if (!br_cntr || !branch_br_cntr) 341 return 0; 342 343 perf_env__find_br_cntr_info(evsel__env(evsel), NULL, &width); 344 mask = (1L << width) - 1; 345 for (i = 0; i < br_cntr_nr; i++) { 346 u64 cntr = (br_cntr >> i * width) & mask; 347 348 branch_br_cntr[off + i + base] += cntr; 349 if (cntr == mask) 350 branch_br_cntr[off + i + base] |= ANNOTATION__BR_CNTR_SATURATED_FLAG; 351 } 352 353 return 0; 354 } 355 356 static int symbol__account_cycles(u64 addr, u64 start, struct symbol *sym, 357 unsigned cycles, struct evsel *evsel, 358 u64 br_cntr) 359 { 360 struct annotated_branch *branch; 361 unsigned offset; 362 int ret; 363 364 if (sym == NULL) 365 return 0; 366 branch = symbol__find_branch_hist(sym, evsel->evlist->nr_br_cntr); 367 if (!branch) 368 return -ENOMEM; 369 if (addr < sym->start || addr >= sym->end) 370 return -ERANGE; 371 372 if (start) { 373 if (start < sym->start || start >= sym->end) 374 return -ERANGE; 375 if (start >= addr) 376 start = 0; 377 } 378 offset = addr - sym->start; 379 ret = __symbol__account_cycles(branch->cycles_hist, 380 start ? start - sym->start : 0, 381 offset, cycles, 382 !!start); 383 384 if (ret) 385 return ret; 386 387 return symbol__account_br_cntr(branch, evsel, offset, br_cntr); 388 } 389 390 int addr_map_symbol__account_cycles(struct addr_map_symbol *ams, 391 struct addr_map_symbol *start, 392 unsigned cycles, 393 struct evsel *evsel, 394 u64 br_cntr) 395 { 396 u64 saddr = 0; 397 int err; 398 399 if (!cycles) 400 return 0; 401 402 /* 403 * Only set start when IPC can be computed. We can only 404 * compute it when the basic block is completely in a single 405 * function. 406 * Special case the case when the jump is elsewhere, but 407 * it starts on the function start. 408 */ 409 if (start && 410 (start->ms.sym == ams->ms.sym || 411 (ams->ms.sym && 412 start->addr == ams->ms.sym->start + map__start(ams->ms.map)))) 413 saddr = start->al_addr; 414 if (saddr == 0) 415 pr_debug2("BB with bad start: addr %"PRIx64" start %"PRIx64" sym %"PRIx64" saddr %"PRIx64"\n", 416 ams->addr, 417 start ? start->addr : 0, 418 ams->ms.sym ? ams->ms.sym->start + map__start(ams->ms.map) : 0, 419 saddr); 420 err = symbol__account_cycles(ams->al_addr, saddr, ams->ms.sym, cycles, evsel, br_cntr); 421 if (err) 422 pr_debug2("account_cycles failed %d\n", err); 423 return err; 424 } 425 426 struct annotation_line *annotated_source__get_line(struct annotated_source *src, 427 s64 offset) 428 { 429 struct annotation_line *al; 430 431 list_for_each_entry(al, &src->source, node) { 432 if (al->offset == offset) 433 return al; 434 } 435 return NULL; 436 } 437 438 static unsigned annotation__count_insn(struct annotation *notes, u64 start, u64 end) 439 { 440 struct annotation_line *al; 441 unsigned n_insn = 0; 442 443 al = annotated_source__get_line(notes->src, start); 444 if (al == NULL) 445 return 0; 446 447 list_for_each_entry_from(al, ¬es->src->source, node) { 448 if (al->offset == -1) 449 continue; 450 if ((u64)al->offset > end) 451 break; 452 n_insn++; 453 } 454 return n_insn; 455 } 456 457 static void annotated_branch__delete(struct annotated_branch *branch) 458 { 459 if (branch) { 460 zfree(&branch->cycles_hist); 461 free(branch->br_cntr); 462 free(branch); 463 } 464 } 465 466 static void annotation__count_and_fill(struct annotation *notes, u64 start, u64 end, struct cyc_hist *ch) 467 { 468 unsigned n_insn; 469 unsigned int cover_insn = 0; 470 471 n_insn = annotation__count_insn(notes, start, end); 472 if (n_insn && ch->num && ch->cycles) { 473 struct annotation_line *al; 474 struct annotated_branch *branch; 475 float ipc = n_insn / ((double)ch->cycles / (double)ch->num); 476 477 /* Hide data when there are too many overlaps. */ 478 if (ch->reset >= 0x7fff) 479 return; 480 481 al = annotated_source__get_line(notes->src, start); 482 if (al == NULL) 483 return; 484 485 list_for_each_entry_from(al, ¬es->src->source, node) { 486 if (al->offset == -1) 487 continue; 488 if ((u64)al->offset > end) 489 break; 490 if (al->cycles && al->cycles->ipc == 0.0) { 491 al->cycles->ipc = ipc; 492 cover_insn++; 493 } 494 } 495 496 branch = annotation__get_branch(notes); 497 if (cover_insn && branch) { 498 branch->hit_cycles += ch->cycles; 499 branch->hit_insn += n_insn * ch->num; 500 branch->cover_insn += cover_insn; 501 } 502 } 503 } 504 505 static int annotation__compute_ipc(struct annotation *notes, size_t size, 506 struct evsel *evsel) 507 { 508 unsigned int br_cntr_nr = evsel->evlist->nr_br_cntr; 509 int err = 0; 510 s64 offset; 511 512 if (!notes->branch || !notes->branch->cycles_hist) 513 return 0; 514 515 notes->branch->total_insn = annotation__count_insn(notes, 0, size - 1); 516 notes->branch->hit_cycles = 0; 517 notes->branch->hit_insn = 0; 518 notes->branch->cover_insn = 0; 519 520 annotation__lock(notes); 521 for (offset = size - 1; offset >= 0; --offset) { 522 struct cyc_hist *ch; 523 524 ch = ¬es->branch->cycles_hist[offset]; 525 if (ch && ch->cycles) { 526 struct annotation_line *al; 527 528 al = annotated_source__get_line(notes->src, offset); 529 if (al && al->cycles == NULL) { 530 al->cycles = zalloc(sizeof(*al->cycles)); 531 if (al->cycles == NULL) { 532 err = ENOMEM; 533 break; 534 } 535 } 536 if (ch->have_start) 537 annotation__count_and_fill(notes, ch->start, offset, ch); 538 if (al && ch->num_aggr) { 539 al->cycles->avg = ch->cycles_aggr / ch->num_aggr; 540 al->cycles->max = ch->cycles_max; 541 al->cycles->min = ch->cycles_min; 542 } 543 if (al && notes->branch->br_cntr) { 544 if (!al->br_cntr) { 545 al->br_cntr = calloc(br_cntr_nr, sizeof(u64)); 546 if (!al->br_cntr) { 547 err = ENOMEM; 548 break; 549 } 550 } 551 al->num_aggr = ch->num_aggr; 552 al->br_cntr_nr = br_cntr_nr; 553 al->evsel = evsel; 554 memcpy(al->br_cntr, ¬es->branch->br_cntr[offset * br_cntr_nr], 555 br_cntr_nr * sizeof(u64)); 556 } 557 } 558 } 559 560 if (err) { 561 while (++offset < (s64)size) { 562 struct cyc_hist *ch = ¬es->branch->cycles_hist[offset]; 563 564 if (ch && ch->cycles) { 565 struct annotation_line *al; 566 567 al = annotated_source__get_line(notes->src, offset); 568 if (al) { 569 zfree(&al->cycles); 570 zfree(&al->br_cntr); 571 } 572 } 573 } 574 } 575 576 annotation__unlock(notes); 577 return 0; 578 } 579 580 int addr_map_symbol__inc_samples(struct addr_map_symbol *ams, struct perf_sample *sample, 581 struct evsel *evsel) 582 { 583 return symbol__inc_addr_samples(&ams->ms, evsel, ams->al_addr, sample); 584 } 585 586 int hist_entry__inc_addr_samples(struct hist_entry *he, struct perf_sample *sample, 587 struct evsel *evsel, u64 ip) 588 { 589 return symbol__inc_addr_samples(&he->ms, evsel, ip, sample); 590 } 591 592 593 void annotation__exit(struct annotation *notes) 594 { 595 annotated_source__delete(notes->src); 596 annotated_branch__delete(notes->branch); 597 } 598 599 static struct sharded_mutex *sharded_mutex; 600 601 static void annotation__init_sharded_mutex(void) 602 { 603 /* As many mutexes as there are CPUs. */ 604 sharded_mutex = sharded_mutex__new(cpu__max_present_cpu().cpu); 605 } 606 607 static size_t annotation__hash(const struct annotation *notes) 608 { 609 return (size_t)notes; 610 } 611 612 static struct mutex *annotation__get_mutex(const struct annotation *notes) 613 { 614 static pthread_once_t once = PTHREAD_ONCE_INIT; 615 616 pthread_once(&once, annotation__init_sharded_mutex); 617 if (!sharded_mutex) 618 return NULL; 619 620 return sharded_mutex__get_mutex(sharded_mutex, annotation__hash(notes)); 621 } 622 623 void annotation__lock(struct annotation *notes) 624 NO_THREAD_SAFETY_ANALYSIS 625 { 626 struct mutex *mutex = annotation__get_mutex(notes); 627 628 if (mutex) 629 mutex_lock(mutex); 630 } 631 632 void annotation__unlock(struct annotation *notes) 633 NO_THREAD_SAFETY_ANALYSIS 634 { 635 struct mutex *mutex = annotation__get_mutex(notes); 636 637 if (mutex) 638 mutex_unlock(mutex); 639 } 640 641 bool annotation__trylock(struct annotation *notes) 642 { 643 struct mutex *mutex = annotation__get_mutex(notes); 644 645 if (!mutex) 646 return false; 647 648 return mutex_trylock(mutex); 649 } 650 651 void annotation_line__add(struct annotation_line *al, struct list_head *head) 652 { 653 list_add_tail(&al->node, head); 654 } 655 656 struct annotation_line * 657 annotation_line__next(struct annotation_line *pos, struct list_head *head) 658 { 659 list_for_each_entry_continue(pos, head, node) 660 if (pos->offset >= 0) 661 return pos; 662 663 return NULL; 664 } 665 666 static const char *annotate__address_color(struct block_range *br) 667 { 668 double cov = block_range__coverage(br); 669 670 if (cov >= 0) { 671 /* mark red for >75% coverage */ 672 if (cov > 0.75) 673 return PERF_COLOR_RED; 674 675 /* mark dull for <1% coverage */ 676 if (cov < 0.01) 677 return PERF_COLOR_NORMAL; 678 } 679 680 return PERF_COLOR_MAGENTA; 681 } 682 683 static const char *annotate__asm_color(struct block_range *br) 684 { 685 double cov = block_range__coverage(br); 686 687 if (cov >= 0) { 688 /* mark dull for <1% coverage */ 689 if (cov < 0.01) 690 return PERF_COLOR_NORMAL; 691 } 692 693 return PERF_COLOR_BLUE; 694 } 695 696 static void annotate__branch_printf(struct block_range *br, u64 addr) 697 { 698 bool emit_comment = true; 699 700 if (!br) 701 return; 702 703 #if 1 704 if (br->is_target && br->start == addr) { 705 struct block_range *branch = br; 706 double p; 707 708 /* 709 * Find matching branch to our target. 710 */ 711 while (!branch->is_branch) 712 branch = block_range__next(branch); 713 714 p = 100 *(double)br->entry / branch->coverage; 715 716 if (p > 0.1) { 717 if (emit_comment) { 718 emit_comment = false; 719 printf("\t#"); 720 } 721 722 /* 723 * The percentage of coverage joined at this target in relation 724 * to the next branch. 725 */ 726 printf(" +%.2f%%", p); 727 } 728 } 729 #endif 730 if (br->is_branch && br->end == addr) { 731 double p = 100*(double)br->taken / br->coverage; 732 733 if (p > 0.1) { 734 if (emit_comment) { 735 emit_comment = false; 736 printf("\t#"); 737 } 738 739 /* 740 * The percentage of coverage leaving at this branch, and 741 * its prediction ratio. 742 */ 743 printf(" -%.2f%% (p:%.2f%%)", p, 100*(double)br->pred / br->taken); 744 } 745 } 746 } 747 748 static int disasm_line__print(struct disasm_line *dl, u64 start, int addr_fmt_width) 749 { 750 s64 offset = dl->al.offset; 751 const u64 addr = start + offset; 752 struct block_range *br; 753 754 br = block_range__find(addr); 755 color_fprintf(stdout, annotate__address_color(br), " %*" PRIx64 ":", addr_fmt_width, addr); 756 color_fprintf(stdout, annotate__asm_color(br), "%s", dl->al.line); 757 annotate__branch_printf(br, addr); 758 return 0; 759 } 760 761 static int 762 annotation_line__print(struct annotation_line *al, struct symbol *sym, u64 start, 763 struct evsel *evsel, u64 len, int min_pcnt, int printed, 764 int max_lines, struct annotation_line *queue, int addr_fmt_width, 765 int percent_type) 766 { 767 struct disasm_line *dl = container_of(al, struct disasm_line, al); 768 struct annotation *notes = symbol__annotation(sym); 769 static const char *prev_line; 770 771 if (al->offset != -1) { 772 double max_percent = 0.0; 773 int i, nr_percent = 1; 774 const char *color; 775 776 for (i = 0; i < al->data_nr; i++) { 777 double percent; 778 779 percent = annotation_data__percent(&al->data[i], 780 percent_type); 781 782 if (percent > max_percent) 783 max_percent = percent; 784 } 785 786 if (al->data_nr > nr_percent) 787 nr_percent = al->data_nr; 788 789 if (max_percent < min_pcnt) 790 return -1; 791 792 if (max_lines && printed >= max_lines) 793 return 1; 794 795 if (queue != NULL) { 796 list_for_each_entry_from(queue, ¬es->src->source, node) { 797 if (queue == al) 798 break; 799 annotation_line__print(queue, sym, start, evsel, len, 800 0, 0, 1, NULL, addr_fmt_width, 801 percent_type); 802 } 803 } 804 805 color = get_percent_color(max_percent); 806 807 for (i = 0; i < nr_percent; i++) { 808 struct annotation_data *data = &al->data[i]; 809 double percent; 810 811 percent = annotation_data__percent(data, percent_type); 812 color = get_percent_color(percent); 813 814 if (symbol_conf.show_total_period) 815 color_fprintf(stdout, color, " %11" PRIu64, 816 data->he.period); 817 else if (symbol_conf.show_nr_samples) 818 color_fprintf(stdout, color, " %7" PRIu64, 819 data->he.nr_samples); 820 else 821 color_fprintf(stdout, color, " %7.2f", percent); 822 } 823 824 printf(" : "); 825 826 disasm_line__print(dl, start, addr_fmt_width); 827 828 /* 829 * Also color the filename and line if needed, with 830 * the same color than the percentage. Don't print it 831 * twice for close colored addr with the same filename:line 832 */ 833 if (al->path) { 834 if (!prev_line || strcmp(prev_line, al->path)) { 835 color_fprintf(stdout, color, " // %s", al->path); 836 prev_line = al->path; 837 } 838 } 839 840 printf("\n"); 841 } else if (max_lines && printed >= max_lines) 842 return 1; 843 else { 844 int width = annotation__pcnt_width(notes); 845 846 if (queue) 847 return -1; 848 849 if (!*al->line) 850 printf(" %*s:\n", width, " "); 851 else 852 printf(" %*s: %-*d %s\n", width, " ", addr_fmt_width, al->line_nr, al->line); 853 } 854 855 return 0; 856 } 857 858 static void calc_percent(struct annotation *notes, 859 struct evsel *evsel, 860 struct annotation_data *data, 861 s64 offset, s64 end) 862 { 863 struct hists *hists = evsel__hists(evsel); 864 struct sym_hist *sym_hist = annotation__histogram(notes, evsel); 865 unsigned int hits = 0; 866 u64 period = 0; 867 868 while (offset < end) { 869 struct sym_hist_entry *entry; 870 871 entry = annotated_source__hist_entry(notes->src, evsel, offset); 872 if (entry) { 873 hits += entry->nr_samples; 874 period += entry->period; 875 } 876 ++offset; 877 } 878 879 if (sym_hist->nr_samples) { 880 data->he.period = period; 881 data->he.nr_samples = hits; 882 data->percent[PERCENT_HITS_LOCAL] = 100.0 * hits / sym_hist->nr_samples; 883 } 884 885 if (hists->stats.nr_non_filtered_samples) 886 data->percent[PERCENT_HITS_GLOBAL] = 100.0 * hits / hists->stats.nr_non_filtered_samples; 887 888 if (sym_hist->period) 889 data->percent[PERCENT_PERIOD_LOCAL] = 100.0 * period / sym_hist->period; 890 891 if (hists->stats.total_period) 892 data->percent[PERCENT_PERIOD_GLOBAL] = 100.0 * period / hists->stats.total_period; 893 } 894 895 static void annotation__calc_percent(struct annotation *notes, 896 struct evsel *leader, s64 len) 897 { 898 struct annotation_line *al, *next; 899 struct evsel *evsel; 900 901 list_for_each_entry(al, ¬es->src->source, node) { 902 s64 end; 903 int i = 0; 904 905 if (al->offset == -1) 906 continue; 907 908 next = annotation_line__next(al, ¬es->src->source); 909 end = next ? next->offset : len; 910 911 for_each_group_evsel(evsel, leader) { 912 struct annotation_data *data; 913 914 BUG_ON(i >= al->data_nr); 915 916 if (symbol_conf.skip_empty && 917 evsel__hists(evsel)->stats.nr_samples == 0) 918 continue; 919 920 data = &al->data[i++]; 921 922 calc_percent(notes, evsel, data, al->offset, end); 923 } 924 } 925 } 926 927 void symbol__calc_percent(struct symbol *sym, struct evsel *evsel) 928 { 929 struct annotation *notes = symbol__annotation(sym); 930 931 annotation__calc_percent(notes, evsel, symbol__size(sym)); 932 } 933 934 static int evsel__get_arch(struct evsel *evsel, struct arch **parch) 935 { 936 struct perf_env *env = evsel__env(evsel); 937 const char *arch_name = perf_env__arch(env); 938 struct arch *arch; 939 int err; 940 941 if (!arch_name) { 942 *parch = NULL; 943 return errno; 944 } 945 946 *parch = arch = arch__find(arch_name); 947 if (arch == NULL) { 948 pr_err("%s: unsupported arch %s\n", __func__, arch_name); 949 return ENOTSUP; 950 } 951 952 if (arch->init) { 953 err = arch->init(arch, env ? env->cpuid : NULL); 954 if (err) { 955 pr_err("%s: failed to initialize %s arch priv area\n", 956 __func__, arch->name); 957 return err; 958 } 959 } 960 return 0; 961 } 962 963 int symbol__annotate(struct map_symbol *ms, struct evsel *evsel, 964 struct arch **parch) 965 { 966 struct symbol *sym = ms->sym; 967 struct annotation *notes = symbol__annotation(sym); 968 struct annotate_args args = { 969 .evsel = evsel, 970 .options = &annotate_opts, 971 }; 972 struct arch *arch = NULL; 973 int err, nr; 974 975 err = evsel__get_arch(evsel, &arch); 976 if (err < 0) 977 return err; 978 979 if (parch) 980 *parch = arch; 981 982 if (notes->src && !list_empty(¬es->src->source)) 983 return 0; 984 985 args.arch = arch; 986 args.ms = *ms; 987 988 if (notes->src == NULL) { 989 notes->src = annotated_source__new(); 990 if (notes->src == NULL) 991 return -1; 992 } 993 994 nr = 0; 995 if (evsel__is_group_event(evsel)) { 996 struct evsel *pos; 997 998 for_each_group_evsel(pos, evsel) { 999 if (symbol_conf.skip_empty && 1000 evsel__hists(pos)->stats.nr_samples == 0) 1001 continue; 1002 nr++; 1003 } 1004 } 1005 notes->src->nr_events = nr ? nr : 1; 1006 1007 if (annotate_opts.full_addr) 1008 notes->src->start = map__objdump_2mem(ms->map, ms->sym->start); 1009 else 1010 notes->src->start = map__rip_2objdump(ms->map, ms->sym->start); 1011 1012 return symbol__disassemble(sym, &args); 1013 } 1014 1015 static void insert_source_line(struct rb_root *root, struct annotation_line *al) 1016 { 1017 struct annotation_line *iter; 1018 struct rb_node **p = &root->rb_node; 1019 struct rb_node *parent = NULL; 1020 unsigned int percent_type = annotate_opts.percent_type; 1021 int i, ret; 1022 1023 while (*p != NULL) { 1024 parent = *p; 1025 iter = rb_entry(parent, struct annotation_line, rb_node); 1026 1027 ret = strcmp(iter->path, al->path); 1028 if (ret == 0) { 1029 for (i = 0; i < al->data_nr; i++) { 1030 iter->data[i].percent_sum += annotation_data__percent(&al->data[i], 1031 percent_type); 1032 } 1033 return; 1034 } 1035 1036 if (ret < 0) 1037 p = &(*p)->rb_left; 1038 else 1039 p = &(*p)->rb_right; 1040 } 1041 1042 for (i = 0; i < al->data_nr; i++) { 1043 al->data[i].percent_sum = annotation_data__percent(&al->data[i], 1044 percent_type); 1045 } 1046 1047 rb_link_node(&al->rb_node, parent, p); 1048 rb_insert_color(&al->rb_node, root); 1049 } 1050 1051 static int cmp_source_line(struct annotation_line *a, struct annotation_line *b) 1052 { 1053 int i; 1054 1055 for (i = 0; i < a->data_nr; i++) { 1056 if (a->data[i].percent_sum == b->data[i].percent_sum) 1057 continue; 1058 return a->data[i].percent_sum > b->data[i].percent_sum; 1059 } 1060 1061 return 0; 1062 } 1063 1064 static void __resort_source_line(struct rb_root *root, struct annotation_line *al) 1065 { 1066 struct annotation_line *iter; 1067 struct rb_node **p = &root->rb_node; 1068 struct rb_node *parent = NULL; 1069 1070 while (*p != NULL) { 1071 parent = *p; 1072 iter = rb_entry(parent, struct annotation_line, rb_node); 1073 1074 if (cmp_source_line(al, iter)) 1075 p = &(*p)->rb_left; 1076 else 1077 p = &(*p)->rb_right; 1078 } 1079 1080 rb_link_node(&al->rb_node, parent, p); 1081 rb_insert_color(&al->rb_node, root); 1082 } 1083 1084 static void resort_source_line(struct rb_root *dest_root, struct rb_root *src_root) 1085 { 1086 struct annotation_line *al; 1087 struct rb_node *node; 1088 1089 node = rb_first(src_root); 1090 while (node) { 1091 struct rb_node *next; 1092 1093 al = rb_entry(node, struct annotation_line, rb_node); 1094 next = rb_next(node); 1095 rb_erase(node, src_root); 1096 1097 __resort_source_line(dest_root, al); 1098 node = next; 1099 } 1100 } 1101 1102 static void print_summary(struct rb_root *root, const char *filename) 1103 { 1104 struct annotation_line *al; 1105 struct rb_node *node; 1106 1107 printf("\nSorted summary for file %s\n", filename); 1108 printf("----------------------------------------------\n\n"); 1109 1110 if (RB_EMPTY_ROOT(root)) { 1111 printf(" Nothing higher than %1.1f%%\n", MIN_GREEN); 1112 return; 1113 } 1114 1115 node = rb_first(root); 1116 while (node) { 1117 double percent, percent_max = 0.0; 1118 const char *color; 1119 char *path; 1120 int i; 1121 1122 al = rb_entry(node, struct annotation_line, rb_node); 1123 for (i = 0; i < al->data_nr; i++) { 1124 percent = al->data[i].percent_sum; 1125 color = get_percent_color(percent); 1126 color_fprintf(stdout, color, " %7.2f", percent); 1127 1128 if (percent > percent_max) 1129 percent_max = percent; 1130 } 1131 1132 path = al->path; 1133 color = get_percent_color(percent_max); 1134 color_fprintf(stdout, color, " %s\n", path); 1135 1136 node = rb_next(node); 1137 } 1138 } 1139 1140 static void symbol__annotate_hits(struct symbol *sym, struct evsel *evsel) 1141 { 1142 struct annotation *notes = symbol__annotation(sym); 1143 struct sym_hist *h = annotation__histogram(notes, evsel); 1144 u64 len = symbol__size(sym), offset; 1145 1146 for (offset = 0; offset < len; ++offset) { 1147 struct sym_hist_entry *entry; 1148 1149 entry = annotated_source__hist_entry(notes->src, evsel, offset); 1150 if (entry && entry->nr_samples != 0) 1151 printf("%*" PRIx64 ": %" PRIu64 "\n", BITS_PER_LONG / 2, 1152 sym->start + offset, entry->nr_samples); 1153 } 1154 printf("%*s: %" PRIu64 "\n", BITS_PER_LONG / 2, "h->nr_samples", h->nr_samples); 1155 } 1156 1157 static int annotated_source__addr_fmt_width(struct list_head *lines, u64 start) 1158 { 1159 char bf[32]; 1160 struct annotation_line *line; 1161 1162 list_for_each_entry_reverse(line, lines, node) { 1163 if (line->offset != -1) 1164 return scnprintf(bf, sizeof(bf), "%" PRIx64, start + line->offset); 1165 } 1166 1167 return 0; 1168 } 1169 1170 int symbol__annotate_printf(struct map_symbol *ms, struct evsel *evsel) 1171 { 1172 struct map *map = ms->map; 1173 struct symbol *sym = ms->sym; 1174 struct dso *dso = map__dso(map); 1175 char *filename; 1176 const char *d_filename; 1177 const char *evsel_name = evsel__name(evsel); 1178 struct annotation *notes = symbol__annotation(sym); 1179 struct sym_hist *h = annotation__histogram(notes, evsel); 1180 struct annotation_line *pos, *queue = NULL; 1181 struct annotation_options *opts = &annotate_opts; 1182 u64 start = map__rip_2objdump(map, sym->start); 1183 int printed = 2, queue_len = 0, addr_fmt_width; 1184 int more = 0; 1185 bool context = opts->context; 1186 u64 len; 1187 int width = annotation__pcnt_width(notes); 1188 int graph_dotted_len; 1189 char buf[512]; 1190 1191 filename = strdup(dso__long_name(dso)); 1192 if (!filename) 1193 return -ENOMEM; 1194 1195 if (opts->full_path) 1196 d_filename = filename; 1197 else 1198 d_filename = basename(filename); 1199 1200 len = symbol__size(sym); 1201 1202 if (evsel__is_group_event(evsel)) { 1203 evsel__group_desc(evsel, buf, sizeof(buf)); 1204 evsel_name = buf; 1205 } 1206 1207 graph_dotted_len = printf(" %-*.*s| Source code & Disassembly of %s for %s (%" PRIu64 " samples, " 1208 "percent: %s)\n", 1209 width, width, symbol_conf.show_total_period ? "Period" : 1210 symbol_conf.show_nr_samples ? "Samples" : "Percent", 1211 d_filename, evsel_name, h->nr_samples, 1212 percent_type_str(opts->percent_type)); 1213 1214 printf("%-*.*s----\n", 1215 graph_dotted_len, graph_dotted_len, graph_dotted_line); 1216 1217 if (verbose > 0) 1218 symbol__annotate_hits(sym, evsel); 1219 1220 addr_fmt_width = annotated_source__addr_fmt_width(¬es->src->source, start); 1221 1222 list_for_each_entry(pos, ¬es->src->source, node) { 1223 int err; 1224 1225 if (context && queue == NULL) { 1226 queue = pos; 1227 queue_len = 0; 1228 } 1229 1230 err = annotation_line__print(pos, sym, start, evsel, len, 1231 opts->min_pcnt, printed, opts->max_lines, 1232 queue, addr_fmt_width, opts->percent_type); 1233 1234 switch (err) { 1235 case 0: 1236 ++printed; 1237 if (context) { 1238 printed += queue_len; 1239 queue = NULL; 1240 queue_len = 0; 1241 } 1242 break; 1243 case 1: 1244 /* filtered by max_lines */ 1245 ++more; 1246 break; 1247 case -1: 1248 default: 1249 /* 1250 * Filtered by min_pcnt or non IP lines when 1251 * context != 0 1252 */ 1253 if (!context) 1254 break; 1255 if (queue_len == context) 1256 queue = list_entry(queue->node.next, typeof(*queue), node); 1257 else 1258 ++queue_len; 1259 break; 1260 } 1261 } 1262 1263 free(filename); 1264 1265 return more; 1266 } 1267 1268 static void FILE__set_percent_color(void *fp __maybe_unused, 1269 double percent __maybe_unused, 1270 bool current __maybe_unused) 1271 { 1272 } 1273 1274 static int FILE__set_jumps_percent_color(void *fp __maybe_unused, 1275 int nr __maybe_unused, bool current __maybe_unused) 1276 { 1277 return 0; 1278 } 1279 1280 static int FILE__set_color(void *fp __maybe_unused, int color __maybe_unused) 1281 { 1282 return 0; 1283 } 1284 1285 static void FILE__printf(void *fp, const char *fmt, ...) 1286 { 1287 va_list args; 1288 1289 va_start(args, fmt); 1290 vfprintf(fp, fmt, args); 1291 va_end(args); 1292 } 1293 1294 static void FILE__write_graph(void *fp, int graph) 1295 { 1296 const char *s; 1297 switch (graph) { 1298 1299 case DARROW_CHAR: s = "↓"; break; 1300 case UARROW_CHAR: s = "↑"; break; 1301 case LARROW_CHAR: s = "←"; break; 1302 case RARROW_CHAR: s = "→"; break; 1303 default: s = "?"; break; 1304 } 1305 1306 fputs(s, fp); 1307 } 1308 1309 static int symbol__annotate_fprintf2(struct symbol *sym, FILE *fp) 1310 { 1311 struct annotation *notes = symbol__annotation(sym); 1312 struct annotation_write_ops wops = { 1313 .first_line = true, 1314 .obj = fp, 1315 .set_color = FILE__set_color, 1316 .set_percent_color = FILE__set_percent_color, 1317 .set_jumps_percent_color = FILE__set_jumps_percent_color, 1318 .printf = FILE__printf, 1319 .write_graph = FILE__write_graph, 1320 }; 1321 struct annotation_line *al; 1322 1323 list_for_each_entry(al, ¬es->src->source, node) { 1324 if (annotation_line__filter(al)) 1325 continue; 1326 annotation_line__write(al, notes, &wops); 1327 fputc('\n', fp); 1328 wops.first_line = false; 1329 } 1330 1331 return 0; 1332 } 1333 1334 int map_symbol__annotation_dump(struct map_symbol *ms, struct evsel *evsel) 1335 { 1336 const char *ev_name = evsel__name(evsel); 1337 char buf[1024]; 1338 char *filename; 1339 int err = -1; 1340 FILE *fp; 1341 1342 if (asprintf(&filename, "%s.annotation", ms->sym->name) < 0) 1343 return -1; 1344 1345 fp = fopen(filename, "w"); 1346 if (fp == NULL) 1347 goto out_free_filename; 1348 1349 if (evsel__is_group_event(evsel)) { 1350 evsel__group_desc(evsel, buf, sizeof(buf)); 1351 ev_name = buf; 1352 } 1353 1354 fprintf(fp, "%s() %s\nEvent: %s\n\n", 1355 ms->sym->name, dso__long_name(map__dso(ms->map)), ev_name); 1356 symbol__annotate_fprintf2(ms->sym, fp); 1357 1358 fclose(fp); 1359 err = 0; 1360 out_free_filename: 1361 free(filename); 1362 return err; 1363 } 1364 1365 void symbol__annotate_zero_histogram(struct symbol *sym, struct evsel *evsel) 1366 { 1367 struct annotation *notes = symbol__annotation(sym); 1368 struct sym_hist *h = annotation__histogram(notes, evsel); 1369 1370 memset(h, 0, sizeof(*notes->src->histograms) * notes->src->nr_histograms); 1371 } 1372 1373 void symbol__annotate_decay_histogram(struct symbol *sym, struct evsel *evsel) 1374 { 1375 struct annotation *notes = symbol__annotation(sym); 1376 struct sym_hist *h = annotation__histogram(notes, evsel); 1377 struct annotation_line *al; 1378 1379 h->nr_samples = 0; 1380 list_for_each_entry(al, ¬es->src->source, node) { 1381 struct sym_hist_entry *entry; 1382 1383 if (al->offset == -1) 1384 continue; 1385 1386 entry = annotated_source__hist_entry(notes->src, evsel, al->offset); 1387 if (entry == NULL) 1388 continue; 1389 1390 entry->nr_samples = entry->nr_samples * 7 / 8; 1391 h->nr_samples += entry->nr_samples; 1392 } 1393 } 1394 1395 void annotated_source__purge(struct annotated_source *as) 1396 { 1397 struct annotation_line *al, *n; 1398 1399 list_for_each_entry_safe(al, n, &as->source, node) { 1400 list_del_init(&al->node); 1401 disasm_line__free(disasm_line(al)); 1402 } 1403 } 1404 1405 static size_t disasm_line__fprintf(struct disasm_line *dl, FILE *fp) 1406 { 1407 size_t printed; 1408 1409 if (dl->al.offset == -1) 1410 return fprintf(fp, "%s\n", dl->al.line); 1411 1412 printed = fprintf(fp, "%#" PRIx64 " %s", dl->al.offset, dl->ins.name); 1413 1414 if (dl->ops.raw[0] != '\0') { 1415 printed += fprintf(fp, "%.*s %s\n", 6 - (int)printed, " ", 1416 dl->ops.raw); 1417 } 1418 1419 return printed + fprintf(fp, "\n"); 1420 } 1421 1422 size_t disasm__fprintf(struct list_head *head, FILE *fp) 1423 { 1424 struct disasm_line *pos; 1425 size_t printed = 0; 1426 1427 list_for_each_entry(pos, head, al.node) 1428 printed += disasm_line__fprintf(pos, fp); 1429 1430 return printed; 1431 } 1432 1433 bool disasm_line__is_valid_local_jump(struct disasm_line *dl, struct symbol *sym) 1434 { 1435 if (!dl || !dl->ins.ops || !ins__is_jump(&dl->ins) || 1436 !disasm_line__has_local_offset(dl) || dl->ops.target.offset < 0 || 1437 dl->ops.target.offset >= (s64)symbol__size(sym)) 1438 return false; 1439 1440 return true; 1441 } 1442 1443 static void 1444 annotation__mark_jump_targets(struct annotation *notes, struct symbol *sym) 1445 { 1446 struct annotation_line *al; 1447 1448 /* PLT symbols contain external offsets */ 1449 if (strstr(sym->name, "@plt")) 1450 return; 1451 1452 list_for_each_entry(al, ¬es->src->source, node) { 1453 struct disasm_line *dl; 1454 struct annotation_line *target; 1455 1456 dl = disasm_line(al); 1457 1458 if (!disasm_line__is_valid_local_jump(dl, sym)) 1459 continue; 1460 1461 target = annotated_source__get_line(notes->src, 1462 dl->ops.target.offset); 1463 /* 1464 * FIXME: Oops, no jump target? Buggy disassembler? Or do we 1465 * have to adjust to the previous offset? 1466 */ 1467 if (target == NULL) 1468 continue; 1469 1470 if (++target->jump_sources > notes->src->max_jump_sources) 1471 notes->src->max_jump_sources = target->jump_sources; 1472 } 1473 } 1474 1475 static void annotation__set_index(struct annotation *notes) 1476 { 1477 struct annotation_line *al; 1478 struct annotated_source *src = notes->src; 1479 1480 src->widths.max_line_len = 0; 1481 src->nr_entries = 0; 1482 src->nr_asm_entries = 0; 1483 1484 list_for_each_entry(al, &src->source, node) { 1485 size_t line_len = strlen(al->line); 1486 1487 if (src->widths.max_line_len < line_len) 1488 src->widths.max_line_len = line_len; 1489 al->idx = src->nr_entries++; 1490 if (al->offset != -1) 1491 al->idx_asm = src->nr_asm_entries++; 1492 else 1493 al->idx_asm = -1; 1494 } 1495 } 1496 1497 static inline int width_jumps(int n) 1498 { 1499 if (n >= 100) 1500 return 5; 1501 if (n / 10) 1502 return 2; 1503 return 1; 1504 } 1505 1506 static int annotation__max_ins_name(struct annotation *notes) 1507 { 1508 int max_name = 0, len; 1509 struct annotation_line *al; 1510 1511 list_for_each_entry(al, ¬es->src->source, node) { 1512 if (al->offset == -1) 1513 continue; 1514 1515 len = strlen(disasm_line(al)->ins.name); 1516 if (max_name < len) 1517 max_name = len; 1518 } 1519 1520 return max_name; 1521 } 1522 1523 static void 1524 annotation__init_column_widths(struct annotation *notes, struct symbol *sym) 1525 { 1526 notes->src->widths.addr = notes->src->widths.target = 1527 notes->src->widths.min_addr = hex_width(symbol__size(sym)); 1528 notes->src->widths.max_addr = hex_width(sym->end); 1529 notes->src->widths.jumps = width_jumps(notes->src->max_jump_sources); 1530 notes->src->widths.max_ins_name = annotation__max_ins_name(notes); 1531 } 1532 1533 void annotation__update_column_widths(struct annotation *notes) 1534 { 1535 if (annotate_opts.use_offset) 1536 notes->src->widths.target = notes->src->widths.min_addr; 1537 else if (annotate_opts.full_addr) 1538 notes->src->widths.target = BITS_PER_LONG / 4; 1539 else 1540 notes->src->widths.target = notes->src->widths.max_addr; 1541 1542 notes->src->widths.addr = notes->src->widths.target; 1543 1544 if (annotate_opts.show_nr_jumps) 1545 notes->src->widths.addr += notes->src->widths.jumps + 1; 1546 } 1547 1548 void annotation__toggle_full_addr(struct annotation *notes, struct map_symbol *ms) 1549 { 1550 annotate_opts.full_addr = !annotate_opts.full_addr; 1551 1552 if (annotate_opts.full_addr) 1553 notes->src->start = map__objdump_2mem(ms->map, ms->sym->start); 1554 else 1555 notes->src->start = map__rip_2objdump(ms->map, ms->sym->start); 1556 1557 annotation__update_column_widths(notes); 1558 } 1559 1560 static void annotation__calc_lines(struct annotation *notes, struct map_symbol *ms, 1561 struct rb_root *root) 1562 { 1563 struct annotation_line *al; 1564 struct rb_root tmp_root = RB_ROOT; 1565 1566 list_for_each_entry(al, ¬es->src->source, node) { 1567 double percent_max = 0.0; 1568 u64 addr; 1569 int i; 1570 1571 for (i = 0; i < al->data_nr; i++) { 1572 double percent; 1573 1574 percent = annotation_data__percent(&al->data[i], 1575 annotate_opts.percent_type); 1576 1577 if (percent > percent_max) 1578 percent_max = percent; 1579 } 1580 1581 if (percent_max <= 0.5) 1582 continue; 1583 1584 addr = map__rip_2objdump(ms->map, ms->sym->start); 1585 al->path = get_srcline(map__dso(ms->map), addr + al->offset, NULL, 1586 false, true, ms->sym->start + al->offset); 1587 insert_source_line(&tmp_root, al); 1588 } 1589 1590 resort_source_line(root, &tmp_root); 1591 } 1592 1593 static void symbol__calc_lines(struct map_symbol *ms, struct rb_root *root) 1594 { 1595 struct annotation *notes = symbol__annotation(ms->sym); 1596 1597 annotation__calc_lines(notes, ms, root); 1598 } 1599 1600 int symbol__tty_annotate2(struct map_symbol *ms, struct evsel *evsel) 1601 { 1602 struct dso *dso = map__dso(ms->map); 1603 struct symbol *sym = ms->sym; 1604 struct rb_root source_line = RB_ROOT; 1605 struct hists *hists = evsel__hists(evsel); 1606 char buf[1024]; 1607 int err; 1608 1609 err = symbol__annotate2(ms, evsel, NULL); 1610 if (err) { 1611 char msg[BUFSIZ]; 1612 1613 dso__set_annotate_warned(dso); 1614 symbol__strerror_disassemble(ms, err, msg, sizeof(msg)); 1615 ui__error("Couldn't annotate %s:\n%s", sym->name, msg); 1616 return -1; 1617 } 1618 1619 if (annotate_opts.print_lines) { 1620 srcline_full_filename = annotate_opts.full_path; 1621 symbol__calc_lines(ms, &source_line); 1622 print_summary(&source_line, dso__long_name(dso)); 1623 } 1624 1625 hists__scnprintf_title(hists, buf, sizeof(buf)); 1626 fprintf(stdout, "%s, [percent: %s]\n%s() %s\n", 1627 buf, percent_type_str(annotate_opts.percent_type), sym->name, dso__long_name(dso)); 1628 symbol__annotate_fprintf2(sym, stdout); 1629 1630 annotated_source__purge(symbol__annotation(sym)->src); 1631 1632 return 0; 1633 } 1634 1635 int symbol__tty_annotate(struct map_symbol *ms, struct evsel *evsel) 1636 { 1637 struct dso *dso = map__dso(ms->map); 1638 struct symbol *sym = ms->sym; 1639 struct rb_root source_line = RB_ROOT; 1640 int err; 1641 1642 err = symbol__annotate(ms, evsel, NULL); 1643 if (err) { 1644 char msg[BUFSIZ]; 1645 1646 dso__set_annotate_warned(dso); 1647 symbol__strerror_disassemble(ms, err, msg, sizeof(msg)); 1648 ui__error("Couldn't annotate %s:\n%s", sym->name, msg); 1649 return -1; 1650 } 1651 1652 symbol__calc_percent(sym, evsel); 1653 1654 if (annotate_opts.print_lines) { 1655 srcline_full_filename = annotate_opts.full_path; 1656 symbol__calc_lines(ms, &source_line); 1657 print_summary(&source_line, dso__long_name(dso)); 1658 } 1659 1660 symbol__annotate_printf(ms, evsel); 1661 1662 annotated_source__purge(symbol__annotation(sym)->src); 1663 1664 return 0; 1665 } 1666 1667 bool ui__has_annotation(void) 1668 { 1669 return use_browser == 1 && perf_hpp_list.sym; 1670 } 1671 1672 1673 static double annotation_line__max_percent(struct annotation_line *al, 1674 unsigned int percent_type) 1675 { 1676 double percent_max = 0.0; 1677 int i; 1678 1679 for (i = 0; i < al->data_nr; i++) { 1680 double percent; 1681 1682 percent = annotation_data__percent(&al->data[i], 1683 percent_type); 1684 1685 if (percent > percent_max) 1686 percent_max = percent; 1687 } 1688 1689 return percent_max; 1690 } 1691 1692 static void disasm_line__write(struct disasm_line *dl, struct annotation *notes, 1693 void *obj, char *bf, size_t size, 1694 void (*obj__printf)(void *obj, const char *fmt, ...), 1695 void (*obj__write_graph)(void *obj, int graph)) 1696 { 1697 if (dl->ins.ops && dl->ins.ops->scnprintf) { 1698 if (ins__is_jump(&dl->ins)) { 1699 bool fwd; 1700 1701 if (dl->ops.target.outside) 1702 goto call_like; 1703 fwd = dl->ops.target.offset > dl->al.offset; 1704 obj__write_graph(obj, fwd ? DARROW_CHAR : UARROW_CHAR); 1705 obj__printf(obj, " "); 1706 } else if (ins__is_call(&dl->ins)) { 1707 call_like: 1708 obj__write_graph(obj, RARROW_CHAR); 1709 obj__printf(obj, " "); 1710 } else if (ins__is_ret(&dl->ins)) { 1711 obj__write_graph(obj, LARROW_CHAR); 1712 obj__printf(obj, " "); 1713 } else { 1714 obj__printf(obj, " "); 1715 } 1716 } else { 1717 obj__printf(obj, " "); 1718 } 1719 1720 disasm_line__scnprintf(dl, bf, size, !annotate_opts.use_offset, 1721 notes->src->widths.max_ins_name); 1722 } 1723 1724 static void ipc_coverage_string(char *bf, int size, struct annotation *notes) 1725 { 1726 double ipc = 0.0, coverage = 0.0; 1727 struct annotated_branch *branch = annotation__get_branch(notes); 1728 1729 if (branch && branch->hit_cycles) 1730 ipc = branch->hit_insn / ((double)branch->hit_cycles); 1731 1732 if (branch && branch->total_insn) { 1733 coverage = branch->cover_insn * 100.0 / 1734 ((double)branch->total_insn); 1735 } 1736 1737 scnprintf(bf, size, "(Average IPC: %.2f, IPC Coverage: %.1f%%)", 1738 ipc, coverage); 1739 } 1740 1741 int annotation_br_cntr_abbr_list(char **str, struct evsel *evsel, bool header) 1742 { 1743 struct evsel *pos; 1744 struct strbuf sb; 1745 1746 if (evsel->evlist->nr_br_cntr <= 0) 1747 return -ENOTSUP; 1748 1749 strbuf_init(&sb, /*hint=*/ 0); 1750 1751 if (header && strbuf_addf(&sb, "# Branch counter abbr list:\n")) 1752 goto err; 1753 1754 evlist__for_each_entry(evsel->evlist, pos) { 1755 if (!(pos->core.attr.branch_sample_type & PERF_SAMPLE_BRANCH_COUNTERS)) 1756 continue; 1757 if (header && strbuf_addf(&sb, "#")) 1758 goto err; 1759 1760 if (strbuf_addf(&sb, " %s = %s\n", pos->name, pos->abbr_name)) 1761 goto err; 1762 } 1763 1764 if (header && strbuf_addf(&sb, "#")) 1765 goto err; 1766 if (strbuf_addf(&sb, " '-' No event occurs\n")) 1767 goto err; 1768 1769 if (header && strbuf_addf(&sb, "#")) 1770 goto err; 1771 if (strbuf_addf(&sb, " '+' Event occurrences may be lost due to branch counter saturated\n")) 1772 goto err; 1773 1774 *str = strbuf_detach(&sb, NULL); 1775 1776 return 0; 1777 err: 1778 strbuf_release(&sb); 1779 return -ENOMEM; 1780 } 1781 1782 /* Assume the branch counter saturated at 3 */ 1783 #define ANNOTATION_BR_CNTR_SATURATION 3 1784 1785 int annotation_br_cntr_entry(char **str, int br_cntr_nr, 1786 u64 *br_cntr, int num_aggr, 1787 struct evsel *evsel) 1788 { 1789 struct evsel *pos = evsel ? evlist__first(evsel->evlist) : NULL; 1790 bool saturated = false; 1791 int i, j, avg, used; 1792 struct strbuf sb; 1793 1794 strbuf_init(&sb, /*hint=*/ 0); 1795 for (i = 0; i < br_cntr_nr; i++) { 1796 used = 0; 1797 avg = ceil((double)(br_cntr[i] & ~ANNOTATION__BR_CNTR_SATURATED_FLAG) / 1798 (double)num_aggr); 1799 1800 /* 1801 * A histogram with the abbr name is displayed by default. 1802 * With -v, the exact number of branch counter is displayed. 1803 */ 1804 if (verbose) { 1805 evlist__for_each_entry_from(evsel->evlist, pos) { 1806 if ((pos->core.attr.branch_sample_type & PERF_SAMPLE_BRANCH_COUNTERS) && 1807 (pos->br_cntr_idx == i)) 1808 break; 1809 } 1810 if (strbuf_addstr(&sb, pos->abbr_name)) 1811 goto err; 1812 1813 if (!br_cntr[i]) { 1814 if (strbuf_addstr(&sb, "=-")) 1815 goto err; 1816 } else { 1817 if (strbuf_addf(&sb, "=%d", avg)) 1818 goto err; 1819 } 1820 if (br_cntr[i] & ANNOTATION__BR_CNTR_SATURATED_FLAG) { 1821 if (strbuf_addch(&sb, '+')) 1822 goto err; 1823 } else { 1824 if (strbuf_addch(&sb, ' ')) 1825 goto err; 1826 } 1827 1828 if ((i < br_cntr_nr - 1) && strbuf_addch(&sb, ',')) 1829 goto err; 1830 continue; 1831 } 1832 1833 if (strbuf_addch(&sb, '|')) 1834 goto err; 1835 1836 if (!br_cntr[i]) { 1837 if (strbuf_addch(&sb, '-')) 1838 goto err; 1839 used++; 1840 } else { 1841 evlist__for_each_entry_from(evsel->evlist, pos) { 1842 if ((pos->core.attr.branch_sample_type & PERF_SAMPLE_BRANCH_COUNTERS) && 1843 (pos->br_cntr_idx == i)) 1844 break; 1845 } 1846 if (br_cntr[i] & ANNOTATION__BR_CNTR_SATURATED_FLAG) 1847 saturated = true; 1848 1849 for (j = 0; j < avg; j++, used++) { 1850 /* Print + if the number of logged events > 3 */ 1851 if (j >= ANNOTATION_BR_CNTR_SATURATION) { 1852 saturated = true; 1853 break; 1854 } 1855 if (strbuf_addstr(&sb, pos->abbr_name)) 1856 goto err; 1857 } 1858 1859 if (saturated) { 1860 if (strbuf_addch(&sb, '+')) 1861 goto err; 1862 used++; 1863 } 1864 pos = list_next_entry(pos, core.node); 1865 } 1866 1867 for (j = used; j < ANNOTATION_BR_CNTR_SATURATION + 1; j++) { 1868 if (strbuf_addch(&sb, ' ')) 1869 goto err; 1870 } 1871 } 1872 1873 if (!verbose && strbuf_addch(&sb, br_cntr_nr ? '|' : ' ')) 1874 goto err; 1875 1876 *str = strbuf_detach(&sb, NULL); 1877 1878 return 0; 1879 err: 1880 strbuf_release(&sb); 1881 return -ENOMEM; 1882 } 1883 1884 static void __annotation_line__write(struct annotation_line *al, struct annotation *notes, 1885 bool first_line, bool current_entry, bool change_color, int width, 1886 void *obj, unsigned int percent_type, 1887 int (*obj__set_color)(void *obj, int color), 1888 void (*obj__set_percent_color)(void *obj, double percent, bool current), 1889 int (*obj__set_jumps_percent_color)(void *obj, int nr, bool current), 1890 void (*obj__printf)(void *obj, const char *fmt, ...), 1891 void (*obj__write_graph)(void *obj, int graph)) 1892 1893 { 1894 double percent_max = annotation_line__max_percent(al, percent_type); 1895 int pcnt_width = annotation__pcnt_width(notes), 1896 cycles_width = annotation__cycles_width(notes); 1897 bool show_title = false; 1898 char bf[256]; 1899 int printed; 1900 1901 if (first_line && (al->offset == -1 || percent_max == 0.0)) { 1902 if (notes->branch && al->cycles) { 1903 if (al->cycles->ipc == 0.0 && al->cycles->avg == 0) 1904 show_title = true; 1905 } else 1906 show_title = true; 1907 } 1908 1909 if (al->offset != -1 && percent_max != 0.0) { 1910 int i; 1911 1912 for (i = 0; i < al->data_nr; i++) { 1913 double percent; 1914 1915 percent = annotation_data__percent(&al->data[i], percent_type); 1916 1917 obj__set_percent_color(obj, percent, current_entry); 1918 if (symbol_conf.show_total_period) { 1919 obj__printf(obj, "%11" PRIu64 " ", al->data[i].he.period); 1920 } else if (symbol_conf.show_nr_samples) { 1921 obj__printf(obj, "%7" PRIu64 " ", 1922 al->data[i].he.nr_samples); 1923 } else { 1924 obj__printf(obj, "%7.2f ", percent); 1925 } 1926 } 1927 } else { 1928 obj__set_percent_color(obj, 0, current_entry); 1929 1930 if (!show_title) 1931 obj__printf(obj, "%-*s", pcnt_width, " "); 1932 else { 1933 obj__printf(obj, "%-*s", pcnt_width, 1934 symbol_conf.show_total_period ? "Period" : 1935 symbol_conf.show_nr_samples ? "Samples" : "Percent"); 1936 } 1937 } 1938 1939 if (notes->branch) { 1940 if (al->cycles && al->cycles->ipc) 1941 obj__printf(obj, "%*.2f ", ANNOTATION__IPC_WIDTH - 1, al->cycles->ipc); 1942 else if (!show_title) 1943 obj__printf(obj, "%*s", ANNOTATION__IPC_WIDTH, " "); 1944 else 1945 obj__printf(obj, "%*s ", ANNOTATION__IPC_WIDTH - 1, "IPC"); 1946 1947 if (!annotate_opts.show_minmax_cycle) { 1948 if (al->cycles && al->cycles->avg) 1949 obj__printf(obj, "%*" PRIu64 " ", 1950 ANNOTATION__CYCLES_WIDTH - 1, al->cycles->avg); 1951 else if (!show_title) 1952 obj__printf(obj, "%*s", 1953 ANNOTATION__CYCLES_WIDTH, " "); 1954 else 1955 obj__printf(obj, "%*s ", 1956 ANNOTATION__CYCLES_WIDTH - 1, 1957 "Cycle"); 1958 } else { 1959 if (al->cycles) { 1960 char str[32]; 1961 1962 scnprintf(str, sizeof(str), 1963 "%" PRIu64 "(%" PRIu64 "/%" PRIu64 ")", 1964 al->cycles->avg, al->cycles->min, 1965 al->cycles->max); 1966 1967 obj__printf(obj, "%*s ", 1968 ANNOTATION__MINMAX_CYCLES_WIDTH - 1, 1969 str); 1970 } else if (!show_title) 1971 obj__printf(obj, "%*s", 1972 ANNOTATION__MINMAX_CYCLES_WIDTH, 1973 " "); 1974 else 1975 obj__printf(obj, "%*s ", 1976 ANNOTATION__MINMAX_CYCLES_WIDTH - 1, 1977 "Cycle(min/max)"); 1978 } 1979 1980 if (annotate_opts.show_br_cntr) { 1981 if (show_title) { 1982 obj__printf(obj, "%*s ", 1983 ANNOTATION__BR_CNTR_WIDTH, 1984 "Branch Counter"); 1985 } else { 1986 char *buf; 1987 1988 if (!annotation_br_cntr_entry(&buf, al->br_cntr_nr, al->br_cntr, 1989 al->num_aggr, al->evsel)) { 1990 obj__printf(obj, "%*s ", ANNOTATION__BR_CNTR_WIDTH, buf); 1991 free(buf); 1992 } 1993 } 1994 } 1995 1996 if (show_title && !*al->line) { 1997 ipc_coverage_string(bf, sizeof(bf), notes); 1998 obj__printf(obj, "%*s", ANNOTATION__AVG_IPC_WIDTH, bf); 1999 } 2000 } 2001 2002 obj__printf(obj, " "); 2003 2004 if (!*al->line) 2005 obj__printf(obj, "%-*s", width - pcnt_width - cycles_width, " "); 2006 else if (al->offset == -1) { 2007 if (al->line_nr && annotate_opts.show_linenr) 2008 printed = scnprintf(bf, sizeof(bf), "%-*d ", 2009 notes->src->widths.addr + 1, al->line_nr); 2010 else 2011 printed = scnprintf(bf, sizeof(bf), "%-*s ", 2012 notes->src->widths.addr, " "); 2013 obj__printf(obj, bf); 2014 obj__printf(obj, "%-*s", width - printed - pcnt_width - cycles_width + 1, al->line); 2015 } else { 2016 u64 addr = al->offset; 2017 int color = -1; 2018 2019 if (!annotate_opts.use_offset) 2020 addr += notes->src->start; 2021 2022 if (!annotate_opts.use_offset) { 2023 printed = scnprintf(bf, sizeof(bf), "%" PRIx64 ": ", addr); 2024 } else { 2025 if (al->jump_sources && 2026 annotate_opts.offset_level >= ANNOTATION__OFFSET_JUMP_TARGETS) { 2027 if (annotate_opts.show_nr_jumps) { 2028 int prev; 2029 printed = scnprintf(bf, sizeof(bf), "%*d ", 2030 notes->src->widths.jumps, 2031 al->jump_sources); 2032 prev = obj__set_jumps_percent_color(obj, al->jump_sources, 2033 current_entry); 2034 obj__printf(obj, bf); 2035 obj__set_color(obj, prev); 2036 } 2037 print_addr: 2038 printed = scnprintf(bf, sizeof(bf), "%*" PRIx64 ": ", 2039 notes->src->widths.target, addr); 2040 } else if (ins__is_call(&disasm_line(al)->ins) && 2041 annotate_opts.offset_level >= ANNOTATION__OFFSET_CALL) { 2042 goto print_addr; 2043 } else if (annotate_opts.offset_level == ANNOTATION__MAX_OFFSET_LEVEL) { 2044 goto print_addr; 2045 } else { 2046 printed = scnprintf(bf, sizeof(bf), "%-*s ", 2047 notes->src->widths.addr, " "); 2048 } 2049 } 2050 2051 if (change_color) 2052 color = obj__set_color(obj, HE_COLORSET_ADDR); 2053 obj__printf(obj, bf); 2054 if (change_color) 2055 obj__set_color(obj, color); 2056 2057 disasm_line__write(disasm_line(al), notes, obj, bf, sizeof(bf), obj__printf, obj__write_graph); 2058 2059 obj__printf(obj, "%-*s", width - pcnt_width - cycles_width - 3 - printed, bf); 2060 } 2061 2062 } 2063 2064 void annotation_line__write(struct annotation_line *al, struct annotation *notes, 2065 struct annotation_write_ops *wops) 2066 { 2067 __annotation_line__write(al, notes, wops->first_line, wops->current_entry, 2068 wops->change_color, wops->width, wops->obj, 2069 annotate_opts.percent_type, 2070 wops->set_color, wops->set_percent_color, 2071 wops->set_jumps_percent_color, wops->printf, 2072 wops->write_graph); 2073 } 2074 2075 int symbol__annotate2(struct map_symbol *ms, struct evsel *evsel, 2076 struct arch **parch) 2077 { 2078 struct symbol *sym = ms->sym; 2079 struct annotation *notes = symbol__annotation(sym); 2080 size_t size = symbol__size(sym); 2081 int err; 2082 2083 err = symbol__annotate(ms, evsel, parch); 2084 if (err) 2085 return err; 2086 2087 symbol__calc_percent(sym, evsel); 2088 2089 annotation__set_index(notes); 2090 annotation__mark_jump_targets(notes, sym); 2091 2092 err = annotation__compute_ipc(notes, size, evsel); 2093 if (err) 2094 return err; 2095 2096 annotation__init_column_widths(notes, sym); 2097 annotation__update_column_widths(notes); 2098 sym->annotate2 = 1; 2099 2100 return 0; 2101 } 2102 2103 static int annotation__config(const char *var, const char *value, void *data) 2104 { 2105 struct annotation_options *opt = data; 2106 2107 if (!strstarts(var, "annotate.")) 2108 return 0; 2109 2110 if (!strcmp(var, "annotate.offset_level")) { 2111 perf_config_u8(&opt->offset_level, "offset_level", value); 2112 2113 if (opt->offset_level > ANNOTATION__MAX_OFFSET_LEVEL) 2114 opt->offset_level = ANNOTATION__MAX_OFFSET_LEVEL; 2115 else if (opt->offset_level < ANNOTATION__MIN_OFFSET_LEVEL) 2116 opt->offset_level = ANNOTATION__MIN_OFFSET_LEVEL; 2117 } else if (!strcmp(var, "annotate.disassemblers")) { 2118 opt->disassemblers_str = strdup(value); 2119 if (!opt->disassemblers_str) { 2120 pr_err("Not enough memory for annotate.disassemblers\n"); 2121 return -1; 2122 } 2123 } else if (!strcmp(var, "annotate.hide_src_code")) { 2124 opt->hide_src_code = perf_config_bool("hide_src_code", value); 2125 } else if (!strcmp(var, "annotate.jump_arrows")) { 2126 opt->jump_arrows = perf_config_bool("jump_arrows", value); 2127 } else if (!strcmp(var, "annotate.show_linenr")) { 2128 opt->show_linenr = perf_config_bool("show_linenr", value); 2129 } else if (!strcmp(var, "annotate.show_nr_jumps")) { 2130 opt->show_nr_jumps = perf_config_bool("show_nr_jumps", value); 2131 } else if (!strcmp(var, "annotate.show_nr_samples")) { 2132 symbol_conf.show_nr_samples = perf_config_bool("show_nr_samples", 2133 value); 2134 } else if (!strcmp(var, "annotate.show_total_period")) { 2135 symbol_conf.show_total_period = perf_config_bool("show_total_period", 2136 value); 2137 } else if (!strcmp(var, "annotate.use_offset")) { 2138 opt->use_offset = perf_config_bool("use_offset", value); 2139 } else if (!strcmp(var, "annotate.disassembler_style")) { 2140 opt->disassembler_style = strdup(value); 2141 if (!opt->disassembler_style) { 2142 pr_err("Not enough memory for annotate.disassembler_style\n"); 2143 return -1; 2144 } 2145 } else if (!strcmp(var, "annotate.objdump")) { 2146 opt->objdump_path = strdup(value); 2147 if (!opt->objdump_path) { 2148 pr_err("Not enough memory for annotate.objdump\n"); 2149 return -1; 2150 } 2151 } else if (!strcmp(var, "annotate.addr2line")) { 2152 symbol_conf.addr2line_path = strdup(value); 2153 if (!symbol_conf.addr2line_path) { 2154 pr_err("Not enough memory for annotate.addr2line\n"); 2155 return -1; 2156 } 2157 } else if (!strcmp(var, "annotate.demangle")) { 2158 symbol_conf.demangle = perf_config_bool("demangle", value); 2159 } else if (!strcmp(var, "annotate.demangle_kernel")) { 2160 symbol_conf.demangle_kernel = perf_config_bool("demangle_kernel", value); 2161 } else { 2162 pr_debug("%s variable unknown, ignoring...", var); 2163 } 2164 2165 return 0; 2166 } 2167 2168 void annotation_options__init(void) 2169 { 2170 struct annotation_options *opt = &annotate_opts; 2171 2172 memset(opt, 0, sizeof(*opt)); 2173 2174 /* Default values. */ 2175 opt->use_offset = true; 2176 opt->jump_arrows = true; 2177 opt->annotate_src = true; 2178 opt->offset_level = ANNOTATION__OFFSET_JUMP_TARGETS; 2179 opt->percent_type = PERCENT_PERIOD_LOCAL; 2180 } 2181 2182 void annotation_options__exit(void) 2183 { 2184 zfree(&annotate_opts.disassembler_style); 2185 zfree(&annotate_opts.objdump_path); 2186 } 2187 2188 void annotation_config__init(void) 2189 { 2190 perf_config(annotation__config, &annotate_opts); 2191 } 2192 2193 static unsigned int parse_percent_type(char *str1, char *str2) 2194 { 2195 unsigned int type = (unsigned int) -1; 2196 2197 if (!strcmp("period", str1)) { 2198 if (!strcmp("local", str2)) 2199 type = PERCENT_PERIOD_LOCAL; 2200 else if (!strcmp("global", str2)) 2201 type = PERCENT_PERIOD_GLOBAL; 2202 } 2203 2204 if (!strcmp("hits", str1)) { 2205 if (!strcmp("local", str2)) 2206 type = PERCENT_HITS_LOCAL; 2207 else if (!strcmp("global", str2)) 2208 type = PERCENT_HITS_GLOBAL; 2209 } 2210 2211 return type; 2212 } 2213 2214 int annotate_parse_percent_type(const struct option *opt __maybe_unused, const char *_str, 2215 int unset __maybe_unused) 2216 { 2217 unsigned int type; 2218 char *str1, *str2; 2219 int err = -1; 2220 2221 str1 = strdup(_str); 2222 if (!str1) 2223 return -ENOMEM; 2224 2225 str2 = strchr(str1, '-'); 2226 if (!str2) 2227 goto out; 2228 2229 *str2++ = 0; 2230 2231 type = parse_percent_type(str1, str2); 2232 if (type == (unsigned int) -1) 2233 type = parse_percent_type(str2, str1); 2234 if (type != (unsigned int) -1) { 2235 annotate_opts.percent_type = type; 2236 err = 0; 2237 } 2238 2239 out: 2240 free(str1); 2241 return err; 2242 } 2243 2244 int annotate_check_args(void) 2245 { 2246 struct annotation_options *args = &annotate_opts; 2247 2248 if (args->prefix_strip && !args->prefix) { 2249 pr_err("--prefix-strip requires --prefix\n"); 2250 return -1; 2251 } 2252 return 0; 2253 } 2254 2255 /* 2256 * Get register number and access offset from the given instruction. 2257 * It assumes AT&T x86 asm format like OFFSET(REG). Maybe it needs 2258 * to revisit the format when it handles different architecture. 2259 * Fills @reg and @offset when return 0. 2260 */ 2261 static int extract_reg_offset(struct arch *arch, const char *str, 2262 struct annotated_op_loc *op_loc) 2263 { 2264 char *p; 2265 char *regname; 2266 2267 if (arch->objdump.register_char == 0) 2268 return -1; 2269 2270 /* 2271 * It should start from offset, but it's possible to skip 0 2272 * in the asm. So 0(%rax) should be same as (%rax). 2273 * 2274 * However, it also start with a segment select register like 2275 * %gs:0x18(%rbx). In that case it should skip the part. 2276 */ 2277 if (*str == arch->objdump.register_char) { 2278 if (arch__is(arch, "x86")) { 2279 /* FIXME: Handle other segment registers */ 2280 if (!strncmp(str, "%gs:", 4)) 2281 op_loc->segment = INSN_SEG_X86_GS; 2282 } 2283 2284 while (*str && !isdigit(*str) && 2285 *str != arch->objdump.memory_ref_char) 2286 str++; 2287 } 2288 2289 op_loc->offset = strtol(str, &p, 0); 2290 2291 p = strchr(p, arch->objdump.register_char); 2292 if (p == NULL) 2293 return -1; 2294 2295 regname = strdup(p); 2296 if (regname == NULL) 2297 return -1; 2298 2299 op_loc->reg1 = get_dwarf_regnum(regname, arch->e_machine, arch->e_flags); 2300 free(regname); 2301 2302 /* Get the second register */ 2303 if (op_loc->multi_regs) { 2304 p = strchr(p + 1, arch->objdump.register_char); 2305 if (p == NULL) 2306 return -1; 2307 2308 regname = strdup(p); 2309 if (regname == NULL) 2310 return -1; 2311 2312 op_loc->reg2 = get_dwarf_regnum(regname, arch->e_machine, arch->e_flags); 2313 free(regname); 2314 } 2315 return 0; 2316 } 2317 2318 /** 2319 * annotate_get_insn_location - Get location of instruction 2320 * @arch: the architecture info 2321 * @dl: the target instruction 2322 * @loc: a buffer to save the data 2323 * 2324 * Get detailed location info (register and offset) in the instruction. 2325 * It needs both source and target operand and whether it accesses a 2326 * memory location. The offset field is meaningful only when the 2327 * corresponding mem flag is set. The reg2 field is meaningful only 2328 * when multi_regs flag is set. 2329 * 2330 * Some examples on x86: 2331 * 2332 * mov (%rax), %rcx # src_reg1 = rax, src_mem = 1, src_offset = 0 2333 * # dst_reg1 = rcx, dst_mem = 0 2334 * 2335 * mov 0x18, %r8 # src_reg1 = -1, src_mem = 0 2336 * # dst_reg1 = r8, dst_mem = 0 2337 * 2338 * mov %rsi, 8(%rbx,%rcx,4) # src_reg1 = rsi, src_mem = 0, src_multi_regs = 0 2339 * # dst_reg1 = rbx, dst_reg2 = rcx, dst_mem = 1 2340 * # dst_multi_regs = 1, dst_offset = 8 2341 */ 2342 int annotate_get_insn_location(struct arch *arch, struct disasm_line *dl, 2343 struct annotated_insn_loc *loc) 2344 { 2345 struct ins_operands *ops; 2346 struct annotated_op_loc *op_loc; 2347 int i; 2348 2349 if (ins__is_lock(&dl->ins)) 2350 ops = dl->ops.locked.ops; 2351 else 2352 ops = &dl->ops; 2353 2354 if (ops == NULL) 2355 return -1; 2356 2357 memset(loc, 0, sizeof(*loc)); 2358 2359 for_each_insn_op_loc(loc, i, op_loc) { 2360 const char *insn_str = ops->source.raw; 2361 bool multi_regs = ops->source.multi_regs; 2362 bool mem_ref = ops->source.mem_ref; 2363 2364 if (i == INSN_OP_TARGET) { 2365 insn_str = ops->target.raw; 2366 multi_regs = ops->target.multi_regs; 2367 mem_ref = ops->target.mem_ref; 2368 } 2369 2370 /* Invalidate the register by default */ 2371 op_loc->reg1 = -1; 2372 op_loc->reg2 = -1; 2373 2374 if (insn_str == NULL) { 2375 if (!arch__is(arch, "powerpc")) 2376 continue; 2377 } 2378 2379 /* 2380 * For powerpc, call get_powerpc_regs function which extracts the 2381 * required fields for op_loc, ie reg1, reg2, offset from the 2382 * raw instruction. 2383 */ 2384 if (arch__is(arch, "powerpc")) { 2385 op_loc->mem_ref = mem_ref; 2386 op_loc->multi_regs = multi_regs; 2387 get_powerpc_regs(dl->raw.raw_insn, !i, op_loc); 2388 } else if (strchr(insn_str, arch->objdump.memory_ref_char)) { 2389 op_loc->mem_ref = true; 2390 op_loc->multi_regs = multi_regs; 2391 extract_reg_offset(arch, insn_str, op_loc); 2392 } else { 2393 char *s, *p = NULL; 2394 2395 if (arch__is(arch, "x86")) { 2396 /* FIXME: Handle other segment registers */ 2397 if (!strncmp(insn_str, "%gs:", 4)) { 2398 op_loc->segment = INSN_SEG_X86_GS; 2399 op_loc->offset = strtol(insn_str + 4, 2400 &p, 0); 2401 if (p && p != insn_str + 4) 2402 op_loc->imm = true; 2403 continue; 2404 } 2405 } 2406 2407 s = strdup(insn_str); 2408 if (s == NULL) 2409 return -1; 2410 2411 if (*s == arch->objdump.register_char) 2412 op_loc->reg1 = get_dwarf_regnum(s, arch->e_machine, arch->e_flags); 2413 else if (*s == arch->objdump.imm_char) { 2414 op_loc->offset = strtol(s + 1, &p, 0); 2415 if (p && p != s + 1) 2416 op_loc->imm = true; 2417 } 2418 free(s); 2419 } 2420 } 2421 2422 return 0; 2423 } 2424 2425 static struct disasm_line *find_disasm_line(struct symbol *sym, u64 ip, 2426 bool allow_update) 2427 { 2428 struct disasm_line *dl; 2429 struct annotation *notes; 2430 2431 notes = symbol__annotation(sym); 2432 2433 list_for_each_entry(dl, ¬es->src->source, al.node) { 2434 if (dl->al.offset == -1) 2435 continue; 2436 2437 if (sym->start + dl->al.offset == ip) { 2438 /* 2439 * llvm-objdump places "lock" in a separate line and 2440 * in that case, we want to get the next line. 2441 */ 2442 if (ins__is_lock(&dl->ins) && 2443 *dl->ops.raw == '\0' && allow_update) { 2444 ip++; 2445 continue; 2446 } 2447 return dl; 2448 } 2449 } 2450 return NULL; 2451 } 2452 2453 static struct annotated_item_stat *annotate_data_stat(struct list_head *head, 2454 const char *name) 2455 { 2456 struct annotated_item_stat *istat; 2457 2458 list_for_each_entry(istat, head, list) { 2459 if (!strcmp(istat->name, name)) 2460 return istat; 2461 } 2462 2463 istat = zalloc(sizeof(*istat)); 2464 if (istat == NULL) 2465 return NULL; 2466 2467 istat->name = strdup(name); 2468 if ((istat->name == NULL) || (!strlen(istat->name))) { 2469 free(istat); 2470 return NULL; 2471 } 2472 2473 list_add_tail(&istat->list, head); 2474 return istat; 2475 } 2476 2477 static bool is_stack_operation(struct arch *arch, struct disasm_line *dl) 2478 { 2479 if (arch__is(arch, "x86")) { 2480 if (!strncmp(dl->ins.name, "push", 4) || 2481 !strncmp(dl->ins.name, "pop", 3) || 2482 !strncmp(dl->ins.name, "call", 4) || 2483 !strncmp(dl->ins.name, "ret", 3)) 2484 return true; 2485 } 2486 2487 return false; 2488 } 2489 2490 static bool is_stack_canary(struct arch *arch, struct annotated_op_loc *loc) 2491 { 2492 /* On x86_64, %gs:40 is used for stack canary */ 2493 if (arch__is(arch, "x86")) { 2494 if (loc->segment == INSN_SEG_X86_GS && loc->imm && 2495 loc->offset == 40) 2496 return true; 2497 } 2498 2499 return false; 2500 } 2501 2502 static struct disasm_line * 2503 annotation__prev_asm_line(struct annotation *notes, struct disasm_line *curr) 2504 { 2505 struct list_head *sources = ¬es->src->source; 2506 struct disasm_line *prev; 2507 2508 if (curr == list_first_entry(sources, struct disasm_line, al.node)) 2509 return NULL; 2510 2511 prev = list_prev_entry(curr, al.node); 2512 while (prev->al.offset == -1 && 2513 prev != list_first_entry(sources, struct disasm_line, al.node)) 2514 prev = list_prev_entry(prev, al.node); 2515 2516 if (prev->al.offset == -1) 2517 return NULL; 2518 2519 return prev; 2520 } 2521 2522 static struct disasm_line * 2523 annotation__next_asm_line(struct annotation *notes, struct disasm_line *curr) 2524 { 2525 struct list_head *sources = ¬es->src->source; 2526 struct disasm_line *next; 2527 2528 if (curr == list_last_entry(sources, struct disasm_line, al.node)) 2529 return NULL; 2530 2531 next = list_next_entry(curr, al.node); 2532 while (next->al.offset == -1 && 2533 next != list_last_entry(sources, struct disasm_line, al.node)) 2534 next = list_next_entry(next, al.node); 2535 2536 if (next->al.offset == -1) 2537 return NULL; 2538 2539 return next; 2540 } 2541 2542 u64 annotate_calc_pcrel(struct map_symbol *ms, u64 ip, int offset, 2543 struct disasm_line *dl) 2544 { 2545 struct annotation *notes; 2546 struct disasm_line *next; 2547 u64 addr; 2548 2549 notes = symbol__annotation(ms->sym); 2550 /* 2551 * PC-relative addressing starts from the next instruction address 2552 * But the IP is for the current instruction. Since disasm_line 2553 * doesn't have the instruction size, calculate it using the next 2554 * disasm_line. If it's the last one, we can use symbol's end 2555 * address directly. 2556 */ 2557 next = annotation__next_asm_line(notes, dl); 2558 if (next == NULL) 2559 addr = ms->sym->end + offset; 2560 else 2561 addr = ip + (next->al.offset - dl->al.offset) + offset; 2562 2563 return map__rip_2objdump(ms->map, addr); 2564 } 2565 2566 static struct debuginfo_cache { 2567 struct dso *dso; 2568 struct debuginfo *dbg; 2569 } di_cache; 2570 2571 void debuginfo_cache__delete(void) 2572 { 2573 dso__put(di_cache.dso); 2574 di_cache.dso = NULL; 2575 2576 debuginfo__delete(di_cache.dbg); 2577 di_cache.dbg = NULL; 2578 } 2579 2580 /** 2581 * hist_entry__get_data_type - find data type for given hist entry 2582 * @he: hist entry 2583 * 2584 * This function first annotates the instruction at @he->ip and extracts 2585 * register and offset info from it. Then it searches the DWARF debug 2586 * info to get a variable and type information using the address, register, 2587 * and offset. 2588 */ 2589 struct annotated_data_type *hist_entry__get_data_type(struct hist_entry *he) 2590 { 2591 struct map_symbol *ms = &he->ms; 2592 struct evsel *evsel = hists_to_evsel(he->hists); 2593 struct arch *arch; 2594 struct disasm_line *dl; 2595 struct annotated_insn_loc loc; 2596 struct annotated_op_loc *op_loc; 2597 struct annotated_data_type *mem_type; 2598 struct annotated_item_stat *istat; 2599 u64 ip = he->ip; 2600 int i; 2601 2602 ann_data_stat.total++; 2603 2604 if (ms->map == NULL || ms->sym == NULL) { 2605 ann_data_stat.no_sym++; 2606 return NULL; 2607 } 2608 2609 if (!symbol_conf.init_annotation) { 2610 ann_data_stat.no_sym++; 2611 return NULL; 2612 } 2613 2614 /* 2615 * di_cache holds a pair of values, but code below assumes 2616 * di_cache.dso can be compared/updated and di_cache.dbg can be 2617 * read/updated independently from each other. That assumption only 2618 * holds in single threaded code. 2619 */ 2620 assert(perf_singlethreaded); 2621 2622 if (map__dso(ms->map) != di_cache.dso) { 2623 dso__put(di_cache.dso); 2624 di_cache.dso = dso__get(map__dso(ms->map)); 2625 2626 debuginfo__delete(di_cache.dbg); 2627 di_cache.dbg = debuginfo__new(dso__long_name(di_cache.dso)); 2628 } 2629 2630 if (di_cache.dbg == NULL) { 2631 ann_data_stat.no_dbginfo++; 2632 return NULL; 2633 } 2634 2635 /* Make sure it has the disasm of the function */ 2636 if (symbol__annotate(ms, evsel, &arch) < 0) { 2637 ann_data_stat.no_insn++; 2638 return NULL; 2639 } 2640 2641 /* 2642 * Get a disasm to extract the location from the insn. 2643 * This is too slow... 2644 */ 2645 dl = find_disasm_line(ms->sym, ip, /*allow_update=*/true); 2646 if (dl == NULL) { 2647 ann_data_stat.no_insn++; 2648 return NULL; 2649 } 2650 2651 retry: 2652 istat = annotate_data_stat(&ann_insn_stat, dl->ins.name); 2653 if (istat == NULL) { 2654 ann_data_stat.no_insn++; 2655 return NULL; 2656 } 2657 2658 if (annotate_get_insn_location(arch, dl, &loc) < 0) { 2659 ann_data_stat.no_insn_ops++; 2660 istat->bad++; 2661 return NULL; 2662 } 2663 2664 if (is_stack_operation(arch, dl)) { 2665 istat->good++; 2666 he->mem_type_off = 0; 2667 return &stackop_type; 2668 } 2669 2670 for_each_insn_op_loc(&loc, i, op_loc) { 2671 struct data_loc_info dloc = { 2672 .arch = arch, 2673 .thread = he->thread, 2674 .ms = ms, 2675 /* Recalculate IP for LOCK prefix or insn fusion */ 2676 .ip = ms->sym->start + dl->al.offset, 2677 .cpumode = he->cpumode, 2678 .op = op_loc, 2679 .di = di_cache.dbg, 2680 }; 2681 2682 if (!op_loc->mem_ref && op_loc->segment == INSN_SEG_NONE) 2683 continue; 2684 2685 /* Recalculate IP because of LOCK prefix or insn fusion */ 2686 ip = ms->sym->start + dl->al.offset; 2687 2688 /* PC-relative addressing */ 2689 if (op_loc->reg1 == DWARF_REG_PC) { 2690 dloc.var_addr = annotate_calc_pcrel(ms, dloc.ip, 2691 op_loc->offset, dl); 2692 } 2693 2694 /* This CPU access in kernel - pretend PC-relative addressing */ 2695 if (dso__kernel(map__dso(ms->map)) && arch__is(arch, "x86") && 2696 op_loc->segment == INSN_SEG_X86_GS && op_loc->imm) { 2697 dloc.var_addr = op_loc->offset; 2698 op_loc->reg1 = DWARF_REG_PC; 2699 } 2700 2701 mem_type = find_data_type(&dloc); 2702 2703 if (mem_type == NULL && is_stack_canary(arch, op_loc)) { 2704 istat->good++; 2705 he->mem_type_off = 0; 2706 return &canary_type; 2707 } 2708 2709 if (mem_type) 2710 istat->good++; 2711 else 2712 istat->bad++; 2713 2714 if (symbol_conf.annotate_data_sample) { 2715 annotated_data_type__update_samples(mem_type, evsel, 2716 dloc.type_offset, 2717 he->stat.nr_events, 2718 he->stat.period); 2719 } 2720 he->mem_type_off = dloc.type_offset; 2721 return mem_type; 2722 } 2723 2724 /* 2725 * Some instructions can be fused and the actual memory access came 2726 * from the previous instruction. 2727 */ 2728 if (dl->al.offset > 0) { 2729 struct annotation *notes; 2730 struct disasm_line *prev_dl; 2731 2732 notes = symbol__annotation(ms->sym); 2733 prev_dl = annotation__prev_asm_line(notes, dl); 2734 2735 if (prev_dl && ins__is_fused(arch, prev_dl->ins.name, dl->ins.name)) { 2736 dl = prev_dl; 2737 goto retry; 2738 } 2739 } 2740 2741 ann_data_stat.no_mem_ops++; 2742 istat->bad++; 2743 return NULL; 2744 } 2745 2746 /* Basic block traversal (BFS) data structure */ 2747 struct basic_block_data { 2748 struct list_head queue; 2749 struct list_head visited; 2750 }; 2751 2752 /* 2753 * During the traversal, it needs to know the parent block where the current 2754 * block block started from. Note that single basic block can be parent of 2755 * two child basic blocks (in case of condition jump). 2756 */ 2757 struct basic_block_link { 2758 struct list_head node; 2759 struct basic_block_link *parent; 2760 struct annotated_basic_block *bb; 2761 }; 2762 2763 /* Check any of basic block in the list already has the offset */ 2764 static bool basic_block_has_offset(struct list_head *head, s64 offset) 2765 { 2766 struct basic_block_link *link; 2767 2768 list_for_each_entry(link, head, node) { 2769 s64 begin_offset = link->bb->begin->al.offset; 2770 s64 end_offset = link->bb->end->al.offset; 2771 2772 if (begin_offset <= offset && offset <= end_offset) 2773 return true; 2774 } 2775 return false; 2776 } 2777 2778 static bool is_new_basic_block(struct basic_block_data *bb_data, 2779 struct disasm_line *dl) 2780 { 2781 s64 offset = dl->al.offset; 2782 2783 if (basic_block_has_offset(&bb_data->visited, offset)) 2784 return false; 2785 if (basic_block_has_offset(&bb_data->queue, offset)) 2786 return false; 2787 return true; 2788 } 2789 2790 /* Add a basic block starting from dl and link it to the parent */ 2791 static int add_basic_block(struct basic_block_data *bb_data, 2792 struct basic_block_link *parent, 2793 struct disasm_line *dl) 2794 { 2795 struct annotated_basic_block *bb; 2796 struct basic_block_link *link; 2797 2798 if (dl == NULL) 2799 return -1; 2800 2801 if (!is_new_basic_block(bb_data, dl)) 2802 return 0; 2803 2804 bb = zalloc(sizeof(*bb)); 2805 if (bb == NULL) 2806 return -1; 2807 2808 bb->begin = dl; 2809 bb->end = dl; 2810 INIT_LIST_HEAD(&bb->list); 2811 2812 link = malloc(sizeof(*link)); 2813 if (link == NULL) { 2814 free(bb); 2815 return -1; 2816 } 2817 2818 link->bb = bb; 2819 link->parent = parent; 2820 list_add_tail(&link->node, &bb_data->queue); 2821 return 0; 2822 } 2823 2824 /* Returns true when it finds the target in the current basic block */ 2825 static bool process_basic_block(struct basic_block_data *bb_data, 2826 struct basic_block_link *link, 2827 struct symbol *sym, u64 target) 2828 { 2829 struct disasm_line *dl, *next_dl, *last_dl; 2830 struct annotation *notes = symbol__annotation(sym); 2831 bool found = false; 2832 2833 dl = link->bb->begin; 2834 /* Check if it's already visited */ 2835 if (basic_block_has_offset(&bb_data->visited, dl->al.offset)) 2836 return false; 2837 2838 last_dl = list_last_entry(¬es->src->source, 2839 struct disasm_line, al.node); 2840 if (last_dl->al.offset == -1) 2841 last_dl = annotation__prev_asm_line(notes, last_dl); 2842 2843 if (last_dl == NULL) 2844 return false; 2845 2846 list_for_each_entry_from(dl, ¬es->src->source, al.node) { 2847 /* Skip comment or debug info line */ 2848 if (dl->al.offset == -1) 2849 continue; 2850 /* Found the target instruction */ 2851 if (sym->start + dl->al.offset == target) { 2852 found = true; 2853 break; 2854 } 2855 /* End of the function, finish the block */ 2856 if (dl == last_dl) 2857 break; 2858 /* 'return' instruction finishes the block */ 2859 if (ins__is_ret(&dl->ins)) 2860 break; 2861 /* normal instructions are part of the basic block */ 2862 if (!ins__is_jump(&dl->ins)) 2863 continue; 2864 /* jump to a different function, tail call or return */ 2865 if (dl->ops.target.outside) 2866 break; 2867 /* jump instruction creates new basic block(s) */ 2868 next_dl = find_disasm_line(sym, sym->start + dl->ops.target.offset, 2869 /*allow_update=*/false); 2870 if (next_dl) 2871 add_basic_block(bb_data, link, next_dl); 2872 2873 /* 2874 * FIXME: determine conditional jumps properly. 2875 * Conditional jumps create another basic block with the 2876 * next disasm line. 2877 */ 2878 if (!strstr(dl->ins.name, "jmp")) { 2879 next_dl = annotation__next_asm_line(notes, dl); 2880 if (next_dl) 2881 add_basic_block(bb_data, link, next_dl); 2882 } 2883 break; 2884 2885 } 2886 link->bb->end = dl; 2887 return found; 2888 } 2889 2890 /* 2891 * It founds a target basic block, build a proper linked list of basic blocks 2892 * by following the link recursively. 2893 */ 2894 static void link_found_basic_blocks(struct basic_block_link *link, 2895 struct list_head *head) 2896 { 2897 while (link) { 2898 struct basic_block_link *parent = link->parent; 2899 2900 list_move(&link->bb->list, head); 2901 list_del(&link->node); 2902 free(link); 2903 2904 link = parent; 2905 } 2906 } 2907 2908 static void delete_basic_blocks(struct basic_block_data *bb_data) 2909 { 2910 struct basic_block_link *link, *tmp; 2911 2912 list_for_each_entry_safe(link, tmp, &bb_data->queue, node) { 2913 list_del(&link->node); 2914 zfree(&link->bb); 2915 free(link); 2916 } 2917 2918 list_for_each_entry_safe(link, tmp, &bb_data->visited, node) { 2919 list_del(&link->node); 2920 zfree(&link->bb); 2921 free(link); 2922 } 2923 } 2924 2925 /** 2926 * annotate_get_basic_blocks - Get basic blocks for given address range 2927 * @sym: symbol to annotate 2928 * @src: source address 2929 * @dst: destination address 2930 * @head: list head to save basic blocks 2931 * 2932 * This function traverses disasm_lines from @src to @dst and save them in a 2933 * list of annotated_basic_block to @head. It uses BFS to find the shortest 2934 * path between two. The basic_block_link is to maintain parent links so 2935 * that it can build a list of blocks from the start. 2936 */ 2937 int annotate_get_basic_blocks(struct symbol *sym, s64 src, s64 dst, 2938 struct list_head *head) 2939 { 2940 struct basic_block_data bb_data = { 2941 .queue = LIST_HEAD_INIT(bb_data.queue), 2942 .visited = LIST_HEAD_INIT(bb_data.visited), 2943 }; 2944 struct basic_block_link *link; 2945 struct disasm_line *dl; 2946 int ret = -1; 2947 2948 dl = find_disasm_line(sym, src, /*allow_update=*/false); 2949 if (dl == NULL) 2950 return -1; 2951 2952 if (add_basic_block(&bb_data, /*parent=*/NULL, dl) < 0) 2953 return -1; 2954 2955 /* Find shortest path from src to dst using BFS */ 2956 while (!list_empty(&bb_data.queue)) { 2957 link = list_first_entry(&bb_data.queue, struct basic_block_link, node); 2958 2959 if (process_basic_block(&bb_data, link, sym, dst)) { 2960 link_found_basic_blocks(link, head); 2961 ret = 0; 2962 break; 2963 } 2964 list_move(&link->node, &bb_data.visited); 2965 } 2966 delete_basic_blocks(&bb_data); 2967 return ret; 2968 } 2969