1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (C) 2011, Red Hat Inc, Arnaldo Carvalho de Melo <acme@redhat.com> 4 * 5 * Parts came from builtin-annotate.c, see those files for further 6 * copyright notes. 7 */ 8 9 #include <errno.h> 10 #include <inttypes.h> 11 #include <libgen.h> 12 #include <stdlib.h> 13 #include "util.h" // hex_width() 14 #include "ui/ui.h" 15 #include "sort.h" 16 #include "build-id.h" 17 #include "color.h" 18 #include "config.h" 19 #include "disasm.h" 20 #include "dso.h" 21 #include "env.h" 22 #include "map.h" 23 #include "maps.h" 24 #include "symbol.h" 25 #include "srcline.h" 26 #include "units.h" 27 #include "debug.h" 28 #include "debuginfo.h" 29 #include "annotate.h" 30 #include "annotate-data.h" 31 #include "evsel.h" 32 #include "evlist.h" 33 #include "bpf-event.h" 34 #include "bpf-utils.h" 35 #include "block-range.h" 36 #include "string2.h" 37 #include "dwarf-regs.h" 38 #include "util/event.h" 39 #include "util/sharded_mutex.h" 40 #include "arch/common.h" 41 #include "namespaces.h" 42 #include "thread.h" 43 #include "hashmap.h" 44 #include "strbuf.h" 45 #include <regex.h> 46 #include <linux/bitops.h> 47 #include <linux/kernel.h> 48 #include <linux/string.h> 49 #include <linux/zalloc.h> 50 #include <subcmd/parse-options.h> 51 #include <subcmd/run-command.h> 52 #include <math.h> 53 54 /* FIXME: For the HE_COLORSET */ 55 #include "ui/browser.h" 56 57 /* 58 * FIXME: Using the same values as slang.h, 59 * but that header may not be available everywhere 60 */ 61 #define LARROW_CHAR ((unsigned char)',') 62 #define RARROW_CHAR ((unsigned char)'+') 63 #define DARROW_CHAR ((unsigned char)'.') 64 #define UARROW_CHAR ((unsigned char)'-') 65 66 #include <linux/ctype.h> 67 68 /* global annotation options */ 69 struct annotation_options annotate_opts; 70 71 /* Data type collection debug statistics */ 72 struct annotated_data_stat ann_data_stat; 73 LIST_HEAD(ann_insn_stat); 74 75 /* Pseudo data types */ 76 struct annotated_data_type stackop_type = { 77 .self = { 78 .type_name = (char *)"(stack operation)", 79 .children = LIST_HEAD_INIT(stackop_type.self.children), 80 }, 81 }; 82 83 struct annotated_data_type canary_type = { 84 .self = { 85 .type_name = (char *)"(stack canary)", 86 .children = LIST_HEAD_INIT(canary_type.self.children), 87 }, 88 }; 89 90 /* symbol histogram: key = offset << 16 | evsel->core.idx */ 91 static size_t sym_hist_hash(long key, void *ctx __maybe_unused) 92 { 93 return (key >> 16) + (key & 0xffff); 94 } 95 96 static bool sym_hist_equal(long key1, long key2, void *ctx __maybe_unused) 97 { 98 return key1 == key2; 99 } 100 101 static struct annotated_source *annotated_source__new(void) 102 { 103 struct annotated_source *src = zalloc(sizeof(*src)); 104 105 if (src != NULL) 106 INIT_LIST_HEAD(&src->source); 107 108 return src; 109 } 110 111 static __maybe_unused void annotated_source__delete(struct annotated_source *src) 112 { 113 struct hashmap_entry *cur; 114 size_t bkt; 115 116 if (src == NULL) 117 return; 118 119 if (src->samples) { 120 hashmap__for_each_entry(src->samples, cur, bkt) 121 zfree(&cur->pvalue); 122 hashmap__free(src->samples); 123 } 124 zfree(&src->histograms); 125 free(src); 126 } 127 128 static int annotated_source__alloc_histograms(struct annotated_source *src, 129 int nr_hists) 130 { 131 src->nr_histograms = nr_hists; 132 src->histograms = calloc(nr_hists, sizeof(*src->histograms)); 133 134 if (src->histograms == NULL) 135 return -1; 136 137 src->samples = hashmap__new(sym_hist_hash, sym_hist_equal, NULL); 138 if (src->samples == NULL) 139 zfree(&src->histograms); 140 141 return src->histograms ? 0 : -1; 142 } 143 144 void symbol__annotate_zero_histograms(struct symbol *sym) 145 { 146 struct annotation *notes = symbol__annotation(sym); 147 148 annotation__lock(notes); 149 if (notes->src != NULL) { 150 memset(notes->src->histograms, 0, 151 notes->src->nr_histograms * sizeof(*notes->src->histograms)); 152 hashmap__clear(notes->src->samples); 153 } 154 if (notes->branch && notes->branch->cycles_hist) { 155 memset(notes->branch->cycles_hist, 0, 156 symbol__size(sym) * sizeof(struct cyc_hist)); 157 } 158 annotation__unlock(notes); 159 } 160 161 static int __symbol__account_cycles(struct cyc_hist *ch, 162 u64 start, 163 unsigned offset, unsigned cycles, 164 unsigned have_start) 165 { 166 /* 167 * For now we can only account one basic block per 168 * final jump. But multiple could be overlapping. 169 * Always account the longest one. So when 170 * a shorter one has been already seen throw it away. 171 * 172 * We separately always account the full cycles. 173 */ 174 ch[offset].num_aggr++; 175 ch[offset].cycles_aggr += cycles; 176 177 if (cycles > ch[offset].cycles_max) 178 ch[offset].cycles_max = cycles; 179 180 if (ch[offset].cycles_min) { 181 if (cycles && cycles < ch[offset].cycles_min) 182 ch[offset].cycles_min = cycles; 183 } else 184 ch[offset].cycles_min = cycles; 185 186 if (!have_start && ch[offset].have_start) 187 return 0; 188 if (ch[offset].num) { 189 if (have_start && (!ch[offset].have_start || 190 ch[offset].start > start)) { 191 ch[offset].have_start = 0; 192 ch[offset].cycles = 0; 193 ch[offset].num = 0; 194 if (ch[offset].reset < 0xffff) 195 ch[offset].reset++; 196 } else if (have_start && 197 ch[offset].start < start) 198 return 0; 199 } 200 201 if (ch[offset].num < NUM_SPARKS) 202 ch[offset].cycles_spark[ch[offset].num] = cycles; 203 204 ch[offset].have_start = have_start; 205 ch[offset].start = start; 206 ch[offset].cycles += cycles; 207 ch[offset].num++; 208 return 0; 209 } 210 211 static int __symbol__inc_addr_samples(struct map_symbol *ms, 212 struct annotated_source *src, int evidx, u64 addr, 213 struct perf_sample *sample) 214 { 215 struct symbol *sym = ms->sym; 216 long hash_key; 217 u64 offset; 218 struct sym_hist *h; 219 struct sym_hist_entry *entry; 220 221 pr_debug3("%s: addr=%#" PRIx64 "\n", __func__, map__unmap_ip(ms->map, addr)); 222 223 if ((addr < sym->start || addr >= sym->end) && 224 (addr != sym->end || sym->start != sym->end)) { 225 pr_debug("%s(%d): ERANGE! sym->name=%s, start=%#" PRIx64 ", addr=%#" PRIx64 ", end=%#" PRIx64 "\n", 226 __func__, __LINE__, sym->name, sym->start, addr, sym->end); 227 return -ERANGE; 228 } 229 230 offset = addr - sym->start; 231 h = annotated_source__histogram(src, evidx); 232 if (h == NULL) { 233 pr_debug("%s(%d): ENOMEM! sym->name=%s, start=%#" PRIx64 ", addr=%#" PRIx64 ", end=%#" PRIx64 ", func: %d\n", 234 __func__, __LINE__, sym->name, sym->start, addr, sym->end, sym->type == STT_FUNC); 235 return -ENOMEM; 236 } 237 238 hash_key = offset << 16 | evidx; 239 if (!hashmap__find(src->samples, hash_key, &entry)) { 240 entry = zalloc(sizeof(*entry)); 241 if (entry == NULL) 242 return -ENOMEM; 243 244 if (hashmap__add(src->samples, hash_key, entry) < 0) 245 return -ENOMEM; 246 } 247 248 h->nr_samples++; 249 h->period += sample->period; 250 entry->nr_samples++; 251 entry->period += sample->period; 252 253 pr_debug3("%#" PRIx64 " %s: period++ [addr: %#" PRIx64 ", %#" PRIx64 254 ", evidx=%d] => nr_samples: %" PRIu64 ", period: %" PRIu64 "\n", 255 sym->start, sym->name, addr, addr - sym->start, evidx, 256 entry->nr_samples, entry->period); 257 return 0; 258 } 259 260 struct annotated_branch *annotation__get_branch(struct annotation *notes) 261 { 262 if (notes == NULL) 263 return NULL; 264 265 if (notes->branch == NULL) 266 notes->branch = zalloc(sizeof(*notes->branch)); 267 268 return notes->branch; 269 } 270 271 static struct annotated_branch *symbol__find_branch_hist(struct symbol *sym, 272 unsigned int br_cntr_nr) 273 { 274 struct annotation *notes = symbol__annotation(sym); 275 struct annotated_branch *branch; 276 const size_t size = symbol__size(sym); 277 278 branch = annotation__get_branch(notes); 279 if (branch == NULL) 280 return NULL; 281 282 if (branch->cycles_hist == NULL) { 283 branch->cycles_hist = calloc(size, sizeof(struct cyc_hist)); 284 if (!branch->cycles_hist) 285 return NULL; 286 } 287 288 if (br_cntr_nr && branch->br_cntr == NULL) { 289 branch->br_cntr = calloc(br_cntr_nr * size, sizeof(u64)); 290 if (!branch->br_cntr) 291 return NULL; 292 } 293 294 return branch; 295 } 296 297 struct annotated_source *symbol__hists(struct symbol *sym, int nr_hists) 298 { 299 struct annotation *notes = symbol__annotation(sym); 300 301 if (notes->src == NULL) { 302 notes->src = annotated_source__new(); 303 if (notes->src == NULL) 304 return NULL; 305 goto alloc_histograms; 306 } 307 308 if (notes->src->histograms == NULL) { 309 alloc_histograms: 310 annotated_source__alloc_histograms(notes->src, nr_hists); 311 } 312 313 return notes->src; 314 } 315 316 static int symbol__inc_addr_samples(struct map_symbol *ms, 317 struct evsel *evsel, u64 addr, 318 struct perf_sample *sample) 319 { 320 struct symbol *sym = ms->sym; 321 struct annotated_source *src; 322 323 if (sym == NULL) 324 return 0; 325 src = symbol__hists(sym, evsel->evlist->core.nr_entries); 326 return src ? __symbol__inc_addr_samples(ms, src, evsel->core.idx, addr, sample) : 0; 327 } 328 329 static int symbol__account_br_cntr(struct annotated_branch *branch, 330 struct evsel *evsel, 331 unsigned offset, 332 u64 br_cntr) 333 { 334 unsigned int br_cntr_nr = evsel__leader(evsel)->br_cntr_nr; 335 unsigned int base = evsel__leader(evsel)->br_cntr_idx; 336 unsigned int off = offset * evsel->evlist->nr_br_cntr; 337 u64 *branch_br_cntr = branch->br_cntr; 338 unsigned int i, mask, width; 339 340 if (!br_cntr || !branch_br_cntr) 341 return 0; 342 343 perf_env__find_br_cntr_info(evsel__env(evsel), NULL, &width); 344 mask = (1L << width) - 1; 345 for (i = 0; i < br_cntr_nr; i++) { 346 u64 cntr = (br_cntr >> i * width) & mask; 347 348 branch_br_cntr[off + i + base] += cntr; 349 if (cntr == mask) 350 branch_br_cntr[off + i + base] |= ANNOTATION__BR_CNTR_SATURATED_FLAG; 351 } 352 353 return 0; 354 } 355 356 static int symbol__account_cycles(u64 addr, u64 start, struct symbol *sym, 357 unsigned cycles, struct evsel *evsel, 358 u64 br_cntr) 359 { 360 struct annotated_branch *branch; 361 unsigned offset; 362 int ret; 363 364 if (sym == NULL) 365 return 0; 366 branch = symbol__find_branch_hist(sym, evsel->evlist->nr_br_cntr); 367 if (!branch) 368 return -ENOMEM; 369 if (addr < sym->start || addr >= sym->end) 370 return -ERANGE; 371 372 if (start) { 373 if (start < sym->start || start >= sym->end) 374 return -ERANGE; 375 if (start >= addr) 376 start = 0; 377 } 378 offset = addr - sym->start; 379 ret = __symbol__account_cycles(branch->cycles_hist, 380 start ? start - sym->start : 0, 381 offset, cycles, 382 !!start); 383 384 if (ret) 385 return ret; 386 387 return symbol__account_br_cntr(branch, evsel, offset, br_cntr); 388 } 389 390 int addr_map_symbol__account_cycles(struct addr_map_symbol *ams, 391 struct addr_map_symbol *start, 392 unsigned cycles, 393 struct evsel *evsel, 394 u64 br_cntr) 395 { 396 u64 saddr = 0; 397 int err; 398 399 if (!cycles) 400 return 0; 401 402 /* 403 * Only set start when IPC can be computed. We can only 404 * compute it when the basic block is completely in a single 405 * function. 406 * Special case the case when the jump is elsewhere, but 407 * it starts on the function start. 408 */ 409 if (start && 410 (start->ms.sym == ams->ms.sym || 411 (ams->ms.sym && 412 start->addr == ams->ms.sym->start + map__start(ams->ms.map)))) 413 saddr = start->al_addr; 414 if (saddr == 0) 415 pr_debug2("BB with bad start: addr %"PRIx64" start %"PRIx64" sym %"PRIx64" saddr %"PRIx64"\n", 416 ams->addr, 417 start ? start->addr : 0, 418 ams->ms.sym ? ams->ms.sym->start + map__start(ams->ms.map) : 0, 419 saddr); 420 err = symbol__account_cycles(ams->al_addr, saddr, ams->ms.sym, cycles, evsel, br_cntr); 421 if (err) 422 pr_debug2("account_cycles failed %d\n", err); 423 return err; 424 } 425 426 struct annotation_line *annotated_source__get_line(struct annotated_source *src, 427 s64 offset) 428 { 429 struct annotation_line *al; 430 431 list_for_each_entry(al, &src->source, node) { 432 if (al->offset == offset) 433 return al; 434 } 435 return NULL; 436 } 437 438 static unsigned annotation__count_insn(struct annotation *notes, u64 start, u64 end) 439 { 440 struct annotation_line *al; 441 unsigned n_insn = 0; 442 443 al = annotated_source__get_line(notes->src, start); 444 if (al == NULL) 445 return 0; 446 447 list_for_each_entry_from(al, ¬es->src->source, node) { 448 if (al->offset == -1) 449 continue; 450 if ((u64)al->offset > end) 451 break; 452 n_insn++; 453 } 454 return n_insn; 455 } 456 457 static void annotated_branch__delete(struct annotated_branch *branch) 458 { 459 if (branch) { 460 zfree(&branch->cycles_hist); 461 free(branch->br_cntr); 462 free(branch); 463 } 464 } 465 466 static void annotation__count_and_fill(struct annotation *notes, u64 start, u64 end, struct cyc_hist *ch) 467 { 468 unsigned n_insn; 469 unsigned int cover_insn = 0; 470 471 n_insn = annotation__count_insn(notes, start, end); 472 if (n_insn && ch->num && ch->cycles) { 473 struct annotation_line *al; 474 struct annotated_branch *branch; 475 float ipc = n_insn / ((double)ch->cycles / (double)ch->num); 476 477 /* Hide data when there are too many overlaps. */ 478 if (ch->reset >= 0x7fff) 479 return; 480 481 al = annotated_source__get_line(notes->src, start); 482 if (al == NULL) 483 return; 484 485 list_for_each_entry_from(al, ¬es->src->source, node) { 486 if (al->offset == -1) 487 continue; 488 if ((u64)al->offset > end) 489 break; 490 if (al->cycles && al->cycles->ipc == 0.0) { 491 al->cycles->ipc = ipc; 492 cover_insn++; 493 } 494 } 495 496 branch = annotation__get_branch(notes); 497 if (cover_insn && branch) { 498 branch->hit_cycles += ch->cycles; 499 branch->hit_insn += n_insn * ch->num; 500 branch->cover_insn += cover_insn; 501 } 502 } 503 } 504 505 static int annotation__compute_ipc(struct annotation *notes, size_t size, 506 struct evsel *evsel) 507 { 508 unsigned int br_cntr_nr = evsel->evlist->nr_br_cntr; 509 int err = 0; 510 s64 offset; 511 512 if (!notes->branch || !notes->branch->cycles_hist) 513 return 0; 514 515 notes->branch->total_insn = annotation__count_insn(notes, 0, size - 1); 516 notes->branch->hit_cycles = 0; 517 notes->branch->hit_insn = 0; 518 notes->branch->cover_insn = 0; 519 520 annotation__lock(notes); 521 for (offset = size - 1; offset >= 0; --offset) { 522 struct cyc_hist *ch; 523 524 ch = ¬es->branch->cycles_hist[offset]; 525 if (ch && ch->cycles) { 526 struct annotation_line *al; 527 528 al = annotated_source__get_line(notes->src, offset); 529 if (al && al->cycles == NULL) { 530 al->cycles = zalloc(sizeof(*al->cycles)); 531 if (al->cycles == NULL) { 532 err = ENOMEM; 533 break; 534 } 535 } 536 if (ch->have_start) 537 annotation__count_and_fill(notes, ch->start, offset, ch); 538 if (al && ch->num_aggr) { 539 al->cycles->avg = ch->cycles_aggr / ch->num_aggr; 540 al->cycles->max = ch->cycles_max; 541 al->cycles->min = ch->cycles_min; 542 } 543 if (al && notes->branch->br_cntr) { 544 if (!al->br_cntr) { 545 al->br_cntr = calloc(br_cntr_nr, sizeof(u64)); 546 if (!al->br_cntr) { 547 err = ENOMEM; 548 break; 549 } 550 } 551 al->num_aggr = ch->num_aggr; 552 al->br_cntr_nr = br_cntr_nr; 553 al->evsel = evsel; 554 memcpy(al->br_cntr, ¬es->branch->br_cntr[offset * br_cntr_nr], 555 br_cntr_nr * sizeof(u64)); 556 } 557 } 558 } 559 560 if (err) { 561 while (++offset < (s64)size) { 562 struct cyc_hist *ch = ¬es->branch->cycles_hist[offset]; 563 564 if (ch && ch->cycles) { 565 struct annotation_line *al; 566 567 al = annotated_source__get_line(notes->src, offset); 568 if (al) { 569 zfree(&al->cycles); 570 zfree(&al->br_cntr); 571 } 572 } 573 } 574 } 575 576 annotation__unlock(notes); 577 return 0; 578 } 579 580 int addr_map_symbol__inc_samples(struct addr_map_symbol *ams, struct perf_sample *sample, 581 struct evsel *evsel) 582 { 583 return symbol__inc_addr_samples(&ams->ms, evsel, ams->al_addr, sample); 584 } 585 586 int hist_entry__inc_addr_samples(struct hist_entry *he, struct perf_sample *sample, 587 struct evsel *evsel, u64 ip) 588 { 589 return symbol__inc_addr_samples(&he->ms, evsel, ip, sample); 590 } 591 592 593 void annotation__exit(struct annotation *notes) 594 { 595 annotated_source__delete(notes->src); 596 annotated_branch__delete(notes->branch); 597 } 598 599 static struct sharded_mutex *sharded_mutex; 600 601 static void annotation__init_sharded_mutex(void) 602 { 603 /* As many mutexes as there are CPUs. */ 604 sharded_mutex = sharded_mutex__new(cpu__max_present_cpu().cpu); 605 } 606 607 static size_t annotation__hash(const struct annotation *notes) 608 { 609 return (size_t)notes; 610 } 611 612 static struct mutex *annotation__get_mutex(const struct annotation *notes) 613 { 614 static pthread_once_t once = PTHREAD_ONCE_INIT; 615 616 pthread_once(&once, annotation__init_sharded_mutex); 617 if (!sharded_mutex) 618 return NULL; 619 620 return sharded_mutex__get_mutex(sharded_mutex, annotation__hash(notes)); 621 } 622 623 void annotation__lock(struct annotation *notes) 624 NO_THREAD_SAFETY_ANALYSIS 625 { 626 struct mutex *mutex = annotation__get_mutex(notes); 627 628 if (mutex) 629 mutex_lock(mutex); 630 } 631 632 void annotation__unlock(struct annotation *notes) 633 NO_THREAD_SAFETY_ANALYSIS 634 { 635 struct mutex *mutex = annotation__get_mutex(notes); 636 637 if (mutex) 638 mutex_unlock(mutex); 639 } 640 641 bool annotation__trylock(struct annotation *notes) 642 { 643 struct mutex *mutex = annotation__get_mutex(notes); 644 645 if (!mutex) 646 return false; 647 648 return mutex_trylock(mutex); 649 } 650 651 void annotation_line__add(struct annotation_line *al, struct list_head *head) 652 { 653 list_add_tail(&al->node, head); 654 } 655 656 struct annotation_line * 657 annotation_line__next(struct annotation_line *pos, struct list_head *head) 658 { 659 list_for_each_entry_continue(pos, head, node) 660 if (pos->offset >= 0) 661 return pos; 662 663 return NULL; 664 } 665 666 static const char *annotate__address_color(struct block_range *br) 667 { 668 double cov = block_range__coverage(br); 669 670 if (cov >= 0) { 671 /* mark red for >75% coverage */ 672 if (cov > 0.75) 673 return PERF_COLOR_RED; 674 675 /* mark dull for <1% coverage */ 676 if (cov < 0.01) 677 return PERF_COLOR_NORMAL; 678 } 679 680 return PERF_COLOR_MAGENTA; 681 } 682 683 static const char *annotate__asm_color(struct block_range *br) 684 { 685 double cov = block_range__coverage(br); 686 687 if (cov >= 0) { 688 /* mark dull for <1% coverage */ 689 if (cov < 0.01) 690 return PERF_COLOR_NORMAL; 691 } 692 693 return PERF_COLOR_BLUE; 694 } 695 696 static void annotate__branch_printf(struct block_range *br, u64 addr) 697 { 698 bool emit_comment = true; 699 700 if (!br) 701 return; 702 703 #if 1 704 if (br->is_target && br->start == addr) { 705 struct block_range *branch = br; 706 double p; 707 708 /* 709 * Find matching branch to our target. 710 */ 711 while (!branch->is_branch) 712 branch = block_range__next(branch); 713 714 p = 100 *(double)br->entry / branch->coverage; 715 716 if (p > 0.1) { 717 if (emit_comment) { 718 emit_comment = false; 719 printf("\t#"); 720 } 721 722 /* 723 * The percentage of coverage joined at this target in relation 724 * to the next branch. 725 */ 726 printf(" +%.2f%%", p); 727 } 728 } 729 #endif 730 if (br->is_branch && br->end == addr) { 731 double p = 100*(double)br->taken / br->coverage; 732 733 if (p > 0.1) { 734 if (emit_comment) { 735 emit_comment = false; 736 printf("\t#"); 737 } 738 739 /* 740 * The percentage of coverage leaving at this branch, and 741 * its prediction ratio. 742 */ 743 printf(" -%.2f%% (p:%.2f%%)", p, 100*(double)br->pred / br->taken); 744 } 745 } 746 } 747 748 static int disasm_line__print(struct disasm_line *dl, u64 start, int addr_fmt_width) 749 { 750 s64 offset = dl->al.offset; 751 const u64 addr = start + offset; 752 struct block_range *br; 753 754 br = block_range__find(addr); 755 color_fprintf(stdout, annotate__address_color(br), " %*" PRIx64 ":", addr_fmt_width, addr); 756 color_fprintf(stdout, annotate__asm_color(br), "%s", dl->al.line); 757 annotate__branch_printf(br, addr); 758 return 0; 759 } 760 761 static int 762 annotation_line__print(struct annotation_line *al, struct symbol *sym, u64 start, 763 struct evsel *evsel, u64 len, int min_pcnt, int printed, 764 int max_lines, struct annotation_line *queue, int addr_fmt_width, 765 int percent_type) 766 { 767 struct disasm_line *dl = container_of(al, struct disasm_line, al); 768 struct annotation *notes = symbol__annotation(sym); 769 static const char *prev_line; 770 771 if (al->offset != -1) { 772 double max_percent = 0.0; 773 int i, nr_percent = 1; 774 const char *color; 775 776 for (i = 0; i < al->data_nr; i++) { 777 double percent; 778 779 percent = annotation_data__percent(&al->data[i], 780 percent_type); 781 782 if (percent > max_percent) 783 max_percent = percent; 784 } 785 786 if (al->data_nr > nr_percent) 787 nr_percent = al->data_nr; 788 789 if (max_percent < min_pcnt) 790 return -1; 791 792 if (max_lines && printed >= max_lines) 793 return 1; 794 795 if (queue != NULL) { 796 list_for_each_entry_from(queue, ¬es->src->source, node) { 797 if (queue == al) 798 break; 799 annotation_line__print(queue, sym, start, evsel, len, 800 0, 0, 1, NULL, addr_fmt_width, 801 percent_type); 802 } 803 } 804 805 color = get_percent_color(max_percent); 806 807 for (i = 0; i < nr_percent; i++) { 808 struct annotation_data *data = &al->data[i]; 809 double percent; 810 811 percent = annotation_data__percent(data, percent_type); 812 color = get_percent_color(percent); 813 814 if (symbol_conf.show_total_period) 815 color_fprintf(stdout, color, " %11" PRIu64, 816 data->he.period); 817 else if (symbol_conf.show_nr_samples) 818 color_fprintf(stdout, color, " %7" PRIu64, 819 data->he.nr_samples); 820 else 821 color_fprintf(stdout, color, " %7.2f", percent); 822 } 823 824 printf(" : "); 825 826 disasm_line__print(dl, start, addr_fmt_width); 827 828 /* 829 * Also color the filename and line if needed, with 830 * the same color than the percentage. Don't print it 831 * twice for close colored addr with the same filename:line 832 */ 833 if (al->path) { 834 if (!prev_line || strcmp(prev_line, al->path)) { 835 color_fprintf(stdout, color, " // %s", al->path); 836 prev_line = al->path; 837 } 838 } 839 840 printf("\n"); 841 } else if (max_lines && printed >= max_lines) 842 return 1; 843 else { 844 int width = annotation__pcnt_width(notes); 845 846 if (queue) 847 return -1; 848 849 if (!*al->line) 850 printf(" %*s:\n", width, " "); 851 else 852 printf(" %*s: %-*d %s\n", width, " ", addr_fmt_width, al->line_nr, al->line); 853 } 854 855 return 0; 856 } 857 858 static void calc_percent(struct annotation *notes, 859 struct evsel *evsel, 860 struct annotation_data *data, 861 s64 offset, s64 end) 862 { 863 struct hists *hists = evsel__hists(evsel); 864 int evidx = evsel->core.idx; 865 struct sym_hist *sym_hist = annotation__histogram(notes, evidx); 866 unsigned int hits = 0; 867 u64 period = 0; 868 869 while (offset < end) { 870 struct sym_hist_entry *entry; 871 872 entry = annotated_source__hist_entry(notes->src, evidx, offset); 873 if (entry) { 874 hits += entry->nr_samples; 875 period += entry->period; 876 } 877 ++offset; 878 } 879 880 if (sym_hist->nr_samples) { 881 data->he.period = period; 882 data->he.nr_samples = hits; 883 data->percent[PERCENT_HITS_LOCAL] = 100.0 * hits / sym_hist->nr_samples; 884 } 885 886 if (hists->stats.nr_non_filtered_samples) 887 data->percent[PERCENT_HITS_GLOBAL] = 100.0 * hits / hists->stats.nr_non_filtered_samples; 888 889 if (sym_hist->period) 890 data->percent[PERCENT_PERIOD_LOCAL] = 100.0 * period / sym_hist->period; 891 892 if (hists->stats.total_period) 893 data->percent[PERCENT_PERIOD_GLOBAL] = 100.0 * period / hists->stats.total_period; 894 } 895 896 static void annotation__calc_percent(struct annotation *notes, 897 struct evsel *leader, s64 len) 898 { 899 struct annotation_line *al, *next; 900 struct evsel *evsel; 901 902 list_for_each_entry(al, ¬es->src->source, node) { 903 s64 end; 904 int i = 0; 905 906 if (al->offset == -1) 907 continue; 908 909 next = annotation_line__next(al, ¬es->src->source); 910 end = next ? next->offset : len; 911 912 for_each_group_evsel(evsel, leader) { 913 struct annotation_data *data; 914 915 BUG_ON(i >= al->data_nr); 916 917 if (symbol_conf.skip_empty && 918 evsel__hists(evsel)->stats.nr_samples == 0) 919 continue; 920 921 data = &al->data[i++]; 922 923 calc_percent(notes, evsel, data, al->offset, end); 924 } 925 } 926 } 927 928 void symbol__calc_percent(struct symbol *sym, struct evsel *evsel) 929 { 930 struct annotation *notes = symbol__annotation(sym); 931 932 annotation__calc_percent(notes, evsel, symbol__size(sym)); 933 } 934 935 static int evsel__get_arch(struct evsel *evsel, struct arch **parch) 936 { 937 struct perf_env *env = evsel__env(evsel); 938 const char *arch_name = perf_env__arch(env); 939 struct arch *arch; 940 int err; 941 942 if (!arch_name) { 943 *parch = NULL; 944 return errno; 945 } 946 947 *parch = arch = arch__find(arch_name); 948 if (arch == NULL) { 949 pr_err("%s: unsupported arch %s\n", __func__, arch_name); 950 return ENOTSUP; 951 } 952 953 if (arch->init) { 954 err = arch->init(arch, env ? env->cpuid : NULL); 955 if (err) { 956 pr_err("%s: failed to initialize %s arch priv area\n", 957 __func__, arch->name); 958 return err; 959 } 960 } 961 return 0; 962 } 963 964 int symbol__annotate(struct map_symbol *ms, struct evsel *evsel, 965 struct arch **parch) 966 { 967 struct symbol *sym = ms->sym; 968 struct annotation *notes = symbol__annotation(sym); 969 struct annotate_args args = { 970 .evsel = evsel, 971 .options = &annotate_opts, 972 }; 973 struct arch *arch = NULL; 974 int err, nr; 975 976 err = evsel__get_arch(evsel, &arch); 977 if (err < 0) 978 return err; 979 980 if (parch) 981 *parch = arch; 982 983 if (notes->src && !list_empty(¬es->src->source)) 984 return 0; 985 986 args.arch = arch; 987 args.ms = *ms; 988 989 if (notes->src == NULL) { 990 notes->src = annotated_source__new(); 991 if (notes->src == NULL) 992 return -1; 993 } 994 995 nr = 0; 996 if (evsel__is_group_event(evsel)) { 997 struct evsel *pos; 998 999 for_each_group_evsel(pos, evsel) { 1000 if (symbol_conf.skip_empty && 1001 evsel__hists(pos)->stats.nr_samples == 0) 1002 continue; 1003 nr++; 1004 } 1005 } 1006 notes->src->nr_events = nr ? nr : 1; 1007 1008 if (annotate_opts.full_addr) 1009 notes->src->start = map__objdump_2mem(ms->map, ms->sym->start); 1010 else 1011 notes->src->start = map__rip_2objdump(ms->map, ms->sym->start); 1012 1013 return symbol__disassemble(sym, &args); 1014 } 1015 1016 static void insert_source_line(struct rb_root *root, struct annotation_line *al) 1017 { 1018 struct annotation_line *iter; 1019 struct rb_node **p = &root->rb_node; 1020 struct rb_node *parent = NULL; 1021 unsigned int percent_type = annotate_opts.percent_type; 1022 int i, ret; 1023 1024 while (*p != NULL) { 1025 parent = *p; 1026 iter = rb_entry(parent, struct annotation_line, rb_node); 1027 1028 ret = strcmp(iter->path, al->path); 1029 if (ret == 0) { 1030 for (i = 0; i < al->data_nr; i++) { 1031 iter->data[i].percent_sum += annotation_data__percent(&al->data[i], 1032 percent_type); 1033 } 1034 return; 1035 } 1036 1037 if (ret < 0) 1038 p = &(*p)->rb_left; 1039 else 1040 p = &(*p)->rb_right; 1041 } 1042 1043 for (i = 0; i < al->data_nr; i++) { 1044 al->data[i].percent_sum = annotation_data__percent(&al->data[i], 1045 percent_type); 1046 } 1047 1048 rb_link_node(&al->rb_node, parent, p); 1049 rb_insert_color(&al->rb_node, root); 1050 } 1051 1052 static int cmp_source_line(struct annotation_line *a, struct annotation_line *b) 1053 { 1054 int i; 1055 1056 for (i = 0; i < a->data_nr; i++) { 1057 if (a->data[i].percent_sum == b->data[i].percent_sum) 1058 continue; 1059 return a->data[i].percent_sum > b->data[i].percent_sum; 1060 } 1061 1062 return 0; 1063 } 1064 1065 static void __resort_source_line(struct rb_root *root, struct annotation_line *al) 1066 { 1067 struct annotation_line *iter; 1068 struct rb_node **p = &root->rb_node; 1069 struct rb_node *parent = NULL; 1070 1071 while (*p != NULL) { 1072 parent = *p; 1073 iter = rb_entry(parent, struct annotation_line, rb_node); 1074 1075 if (cmp_source_line(al, iter)) 1076 p = &(*p)->rb_left; 1077 else 1078 p = &(*p)->rb_right; 1079 } 1080 1081 rb_link_node(&al->rb_node, parent, p); 1082 rb_insert_color(&al->rb_node, root); 1083 } 1084 1085 static void resort_source_line(struct rb_root *dest_root, struct rb_root *src_root) 1086 { 1087 struct annotation_line *al; 1088 struct rb_node *node; 1089 1090 node = rb_first(src_root); 1091 while (node) { 1092 struct rb_node *next; 1093 1094 al = rb_entry(node, struct annotation_line, rb_node); 1095 next = rb_next(node); 1096 rb_erase(node, src_root); 1097 1098 __resort_source_line(dest_root, al); 1099 node = next; 1100 } 1101 } 1102 1103 static void print_summary(struct rb_root *root, const char *filename) 1104 { 1105 struct annotation_line *al; 1106 struct rb_node *node; 1107 1108 printf("\nSorted summary for file %s\n", filename); 1109 printf("----------------------------------------------\n\n"); 1110 1111 if (RB_EMPTY_ROOT(root)) { 1112 printf(" Nothing higher than %1.1f%%\n", MIN_GREEN); 1113 return; 1114 } 1115 1116 node = rb_first(root); 1117 while (node) { 1118 double percent, percent_max = 0.0; 1119 const char *color; 1120 char *path; 1121 int i; 1122 1123 al = rb_entry(node, struct annotation_line, rb_node); 1124 for (i = 0; i < al->data_nr; i++) { 1125 percent = al->data[i].percent_sum; 1126 color = get_percent_color(percent); 1127 color_fprintf(stdout, color, " %7.2f", percent); 1128 1129 if (percent > percent_max) 1130 percent_max = percent; 1131 } 1132 1133 path = al->path; 1134 color = get_percent_color(percent_max); 1135 color_fprintf(stdout, color, " %s\n", path); 1136 1137 node = rb_next(node); 1138 } 1139 } 1140 1141 static void symbol__annotate_hits(struct symbol *sym, struct evsel *evsel) 1142 { 1143 int evidx = evsel->core.idx; 1144 struct annotation *notes = symbol__annotation(sym); 1145 struct sym_hist *h = annotation__histogram(notes, evidx); 1146 u64 len = symbol__size(sym), offset; 1147 1148 for (offset = 0; offset < len; ++offset) { 1149 struct sym_hist_entry *entry; 1150 1151 entry = annotated_source__hist_entry(notes->src, evidx, offset); 1152 if (entry && entry->nr_samples != 0) 1153 printf("%*" PRIx64 ": %" PRIu64 "\n", BITS_PER_LONG / 2, 1154 sym->start + offset, entry->nr_samples); 1155 } 1156 printf("%*s: %" PRIu64 "\n", BITS_PER_LONG / 2, "h->nr_samples", h->nr_samples); 1157 } 1158 1159 static int annotated_source__addr_fmt_width(struct list_head *lines, u64 start) 1160 { 1161 char bf[32]; 1162 struct annotation_line *line; 1163 1164 list_for_each_entry_reverse(line, lines, node) { 1165 if (line->offset != -1) 1166 return scnprintf(bf, sizeof(bf), "%" PRIx64, start + line->offset); 1167 } 1168 1169 return 0; 1170 } 1171 1172 int symbol__annotate_printf(struct map_symbol *ms, struct evsel *evsel) 1173 { 1174 struct map *map = ms->map; 1175 struct symbol *sym = ms->sym; 1176 struct dso *dso = map__dso(map); 1177 char *filename; 1178 const char *d_filename; 1179 const char *evsel_name = evsel__name(evsel); 1180 struct annotation *notes = symbol__annotation(sym); 1181 struct sym_hist *h = annotation__histogram(notes, evsel->core.idx); 1182 struct annotation_line *pos, *queue = NULL; 1183 struct annotation_options *opts = &annotate_opts; 1184 u64 start = map__rip_2objdump(map, sym->start); 1185 int printed = 2, queue_len = 0, addr_fmt_width; 1186 int more = 0; 1187 bool context = opts->context; 1188 u64 len; 1189 int width = annotation__pcnt_width(notes); 1190 int graph_dotted_len; 1191 char buf[512]; 1192 1193 filename = strdup(dso__long_name(dso)); 1194 if (!filename) 1195 return -ENOMEM; 1196 1197 if (opts->full_path) 1198 d_filename = filename; 1199 else 1200 d_filename = basename(filename); 1201 1202 len = symbol__size(sym); 1203 1204 if (evsel__is_group_event(evsel)) { 1205 evsel__group_desc(evsel, buf, sizeof(buf)); 1206 evsel_name = buf; 1207 } 1208 1209 graph_dotted_len = printf(" %-*.*s| Source code & Disassembly of %s for %s (%" PRIu64 " samples, " 1210 "percent: %s)\n", 1211 width, width, symbol_conf.show_total_period ? "Period" : 1212 symbol_conf.show_nr_samples ? "Samples" : "Percent", 1213 d_filename, evsel_name, h->nr_samples, 1214 percent_type_str(opts->percent_type)); 1215 1216 printf("%-*.*s----\n", 1217 graph_dotted_len, graph_dotted_len, graph_dotted_line); 1218 1219 if (verbose > 0) 1220 symbol__annotate_hits(sym, evsel); 1221 1222 addr_fmt_width = annotated_source__addr_fmt_width(¬es->src->source, start); 1223 1224 list_for_each_entry(pos, ¬es->src->source, node) { 1225 int err; 1226 1227 if (context && queue == NULL) { 1228 queue = pos; 1229 queue_len = 0; 1230 } 1231 1232 err = annotation_line__print(pos, sym, start, evsel, len, 1233 opts->min_pcnt, printed, opts->max_lines, 1234 queue, addr_fmt_width, opts->percent_type); 1235 1236 switch (err) { 1237 case 0: 1238 ++printed; 1239 if (context) { 1240 printed += queue_len; 1241 queue = NULL; 1242 queue_len = 0; 1243 } 1244 break; 1245 case 1: 1246 /* filtered by max_lines */ 1247 ++more; 1248 break; 1249 case -1: 1250 default: 1251 /* 1252 * Filtered by min_pcnt or non IP lines when 1253 * context != 0 1254 */ 1255 if (!context) 1256 break; 1257 if (queue_len == context) 1258 queue = list_entry(queue->node.next, typeof(*queue), node); 1259 else 1260 ++queue_len; 1261 break; 1262 } 1263 } 1264 1265 free(filename); 1266 1267 return more; 1268 } 1269 1270 static void FILE__set_percent_color(void *fp __maybe_unused, 1271 double percent __maybe_unused, 1272 bool current __maybe_unused) 1273 { 1274 } 1275 1276 static int FILE__set_jumps_percent_color(void *fp __maybe_unused, 1277 int nr __maybe_unused, bool current __maybe_unused) 1278 { 1279 return 0; 1280 } 1281 1282 static int FILE__set_color(void *fp __maybe_unused, int color __maybe_unused) 1283 { 1284 return 0; 1285 } 1286 1287 static void FILE__printf(void *fp, const char *fmt, ...) 1288 { 1289 va_list args; 1290 1291 va_start(args, fmt); 1292 vfprintf(fp, fmt, args); 1293 va_end(args); 1294 } 1295 1296 static void FILE__write_graph(void *fp, int graph) 1297 { 1298 const char *s; 1299 switch (graph) { 1300 1301 case DARROW_CHAR: s = "↓"; break; 1302 case UARROW_CHAR: s = "↑"; break; 1303 case LARROW_CHAR: s = "←"; break; 1304 case RARROW_CHAR: s = "→"; break; 1305 default: s = "?"; break; 1306 } 1307 1308 fputs(s, fp); 1309 } 1310 1311 static int symbol__annotate_fprintf2(struct symbol *sym, FILE *fp) 1312 { 1313 struct annotation *notes = symbol__annotation(sym); 1314 struct annotation_write_ops wops = { 1315 .first_line = true, 1316 .obj = fp, 1317 .set_color = FILE__set_color, 1318 .set_percent_color = FILE__set_percent_color, 1319 .set_jumps_percent_color = FILE__set_jumps_percent_color, 1320 .printf = FILE__printf, 1321 .write_graph = FILE__write_graph, 1322 }; 1323 struct annotation_line *al; 1324 1325 list_for_each_entry(al, ¬es->src->source, node) { 1326 if (annotation_line__filter(al)) 1327 continue; 1328 annotation_line__write(al, notes, &wops); 1329 fputc('\n', fp); 1330 wops.first_line = false; 1331 } 1332 1333 return 0; 1334 } 1335 1336 int map_symbol__annotation_dump(struct map_symbol *ms, struct evsel *evsel) 1337 { 1338 const char *ev_name = evsel__name(evsel); 1339 char buf[1024]; 1340 char *filename; 1341 int err = -1; 1342 FILE *fp; 1343 1344 if (asprintf(&filename, "%s.annotation", ms->sym->name) < 0) 1345 return -1; 1346 1347 fp = fopen(filename, "w"); 1348 if (fp == NULL) 1349 goto out_free_filename; 1350 1351 if (evsel__is_group_event(evsel)) { 1352 evsel__group_desc(evsel, buf, sizeof(buf)); 1353 ev_name = buf; 1354 } 1355 1356 fprintf(fp, "%s() %s\nEvent: %s\n\n", 1357 ms->sym->name, dso__long_name(map__dso(ms->map)), ev_name); 1358 symbol__annotate_fprintf2(ms->sym, fp); 1359 1360 fclose(fp); 1361 err = 0; 1362 out_free_filename: 1363 free(filename); 1364 return err; 1365 } 1366 1367 void symbol__annotate_zero_histogram(struct symbol *sym, int evidx) 1368 { 1369 struct annotation *notes = symbol__annotation(sym); 1370 struct sym_hist *h = annotation__histogram(notes, evidx); 1371 1372 memset(h, 0, sizeof(*notes->src->histograms) * notes->src->nr_histograms); 1373 } 1374 1375 void symbol__annotate_decay_histogram(struct symbol *sym, int evidx) 1376 { 1377 struct annotation *notes = symbol__annotation(sym); 1378 struct sym_hist *h = annotation__histogram(notes, evidx); 1379 struct annotation_line *al; 1380 1381 h->nr_samples = 0; 1382 list_for_each_entry(al, ¬es->src->source, node) { 1383 struct sym_hist_entry *entry; 1384 1385 if (al->offset == -1) 1386 continue; 1387 1388 entry = annotated_source__hist_entry(notes->src, evidx, al->offset); 1389 if (entry == NULL) 1390 continue; 1391 1392 entry->nr_samples = entry->nr_samples * 7 / 8; 1393 h->nr_samples += entry->nr_samples; 1394 } 1395 } 1396 1397 void annotated_source__purge(struct annotated_source *as) 1398 { 1399 struct annotation_line *al, *n; 1400 1401 list_for_each_entry_safe(al, n, &as->source, node) { 1402 list_del_init(&al->node); 1403 disasm_line__free(disasm_line(al)); 1404 } 1405 } 1406 1407 static size_t disasm_line__fprintf(struct disasm_line *dl, FILE *fp) 1408 { 1409 size_t printed; 1410 1411 if (dl->al.offset == -1) 1412 return fprintf(fp, "%s\n", dl->al.line); 1413 1414 printed = fprintf(fp, "%#" PRIx64 " %s", dl->al.offset, dl->ins.name); 1415 1416 if (dl->ops.raw[0] != '\0') { 1417 printed += fprintf(fp, "%.*s %s\n", 6 - (int)printed, " ", 1418 dl->ops.raw); 1419 } 1420 1421 return printed + fprintf(fp, "\n"); 1422 } 1423 1424 size_t disasm__fprintf(struct list_head *head, FILE *fp) 1425 { 1426 struct disasm_line *pos; 1427 size_t printed = 0; 1428 1429 list_for_each_entry(pos, head, al.node) 1430 printed += disasm_line__fprintf(pos, fp); 1431 1432 return printed; 1433 } 1434 1435 bool disasm_line__is_valid_local_jump(struct disasm_line *dl, struct symbol *sym) 1436 { 1437 if (!dl || !dl->ins.ops || !ins__is_jump(&dl->ins) || 1438 !disasm_line__has_local_offset(dl) || dl->ops.target.offset < 0 || 1439 dl->ops.target.offset >= (s64)symbol__size(sym)) 1440 return false; 1441 1442 return true; 1443 } 1444 1445 static void 1446 annotation__mark_jump_targets(struct annotation *notes, struct symbol *sym) 1447 { 1448 struct annotation_line *al; 1449 1450 /* PLT symbols contain external offsets */ 1451 if (strstr(sym->name, "@plt")) 1452 return; 1453 1454 list_for_each_entry(al, ¬es->src->source, node) { 1455 struct disasm_line *dl; 1456 struct annotation_line *target; 1457 1458 dl = disasm_line(al); 1459 1460 if (!disasm_line__is_valid_local_jump(dl, sym)) 1461 continue; 1462 1463 target = annotated_source__get_line(notes->src, 1464 dl->ops.target.offset); 1465 /* 1466 * FIXME: Oops, no jump target? Buggy disassembler? Or do we 1467 * have to adjust to the previous offset? 1468 */ 1469 if (target == NULL) 1470 continue; 1471 1472 if (++target->jump_sources > notes->src->max_jump_sources) 1473 notes->src->max_jump_sources = target->jump_sources; 1474 } 1475 } 1476 1477 static void annotation__set_index(struct annotation *notes) 1478 { 1479 struct annotation_line *al; 1480 struct annotated_source *src = notes->src; 1481 1482 src->widths.max_line_len = 0; 1483 src->nr_entries = 0; 1484 src->nr_asm_entries = 0; 1485 1486 list_for_each_entry(al, &src->source, node) { 1487 size_t line_len = strlen(al->line); 1488 1489 if (src->widths.max_line_len < line_len) 1490 src->widths.max_line_len = line_len; 1491 al->idx = src->nr_entries++; 1492 if (al->offset != -1) 1493 al->idx_asm = src->nr_asm_entries++; 1494 else 1495 al->idx_asm = -1; 1496 } 1497 } 1498 1499 static inline int width_jumps(int n) 1500 { 1501 if (n >= 100) 1502 return 5; 1503 if (n / 10) 1504 return 2; 1505 return 1; 1506 } 1507 1508 static int annotation__max_ins_name(struct annotation *notes) 1509 { 1510 int max_name = 0, len; 1511 struct annotation_line *al; 1512 1513 list_for_each_entry(al, ¬es->src->source, node) { 1514 if (al->offset == -1) 1515 continue; 1516 1517 len = strlen(disasm_line(al)->ins.name); 1518 if (max_name < len) 1519 max_name = len; 1520 } 1521 1522 return max_name; 1523 } 1524 1525 static void 1526 annotation__init_column_widths(struct annotation *notes, struct symbol *sym) 1527 { 1528 notes->src->widths.addr = notes->src->widths.target = 1529 notes->src->widths.min_addr = hex_width(symbol__size(sym)); 1530 notes->src->widths.max_addr = hex_width(sym->end); 1531 notes->src->widths.jumps = width_jumps(notes->src->max_jump_sources); 1532 notes->src->widths.max_ins_name = annotation__max_ins_name(notes); 1533 } 1534 1535 void annotation__update_column_widths(struct annotation *notes) 1536 { 1537 if (annotate_opts.use_offset) 1538 notes->src->widths.target = notes->src->widths.min_addr; 1539 else if (annotate_opts.full_addr) 1540 notes->src->widths.target = BITS_PER_LONG / 4; 1541 else 1542 notes->src->widths.target = notes->src->widths.max_addr; 1543 1544 notes->src->widths.addr = notes->src->widths.target; 1545 1546 if (annotate_opts.show_nr_jumps) 1547 notes->src->widths.addr += notes->src->widths.jumps + 1; 1548 } 1549 1550 void annotation__toggle_full_addr(struct annotation *notes, struct map_symbol *ms) 1551 { 1552 annotate_opts.full_addr = !annotate_opts.full_addr; 1553 1554 if (annotate_opts.full_addr) 1555 notes->src->start = map__objdump_2mem(ms->map, ms->sym->start); 1556 else 1557 notes->src->start = map__rip_2objdump(ms->map, ms->sym->start); 1558 1559 annotation__update_column_widths(notes); 1560 } 1561 1562 static void annotation__calc_lines(struct annotation *notes, struct map_symbol *ms, 1563 struct rb_root *root) 1564 { 1565 struct annotation_line *al; 1566 struct rb_root tmp_root = RB_ROOT; 1567 1568 list_for_each_entry(al, ¬es->src->source, node) { 1569 double percent_max = 0.0; 1570 u64 addr; 1571 int i; 1572 1573 for (i = 0; i < al->data_nr; i++) { 1574 double percent; 1575 1576 percent = annotation_data__percent(&al->data[i], 1577 annotate_opts.percent_type); 1578 1579 if (percent > percent_max) 1580 percent_max = percent; 1581 } 1582 1583 if (percent_max <= 0.5) 1584 continue; 1585 1586 addr = map__rip_2objdump(ms->map, ms->sym->start); 1587 al->path = get_srcline(map__dso(ms->map), addr + al->offset, NULL, 1588 false, true, ms->sym->start + al->offset); 1589 insert_source_line(&tmp_root, al); 1590 } 1591 1592 resort_source_line(root, &tmp_root); 1593 } 1594 1595 static void symbol__calc_lines(struct map_symbol *ms, struct rb_root *root) 1596 { 1597 struct annotation *notes = symbol__annotation(ms->sym); 1598 1599 annotation__calc_lines(notes, ms, root); 1600 } 1601 1602 int symbol__tty_annotate2(struct map_symbol *ms, struct evsel *evsel) 1603 { 1604 struct dso *dso = map__dso(ms->map); 1605 struct symbol *sym = ms->sym; 1606 struct rb_root source_line = RB_ROOT; 1607 struct hists *hists = evsel__hists(evsel); 1608 char buf[1024]; 1609 int err; 1610 1611 err = symbol__annotate2(ms, evsel, NULL); 1612 if (err) { 1613 char msg[BUFSIZ]; 1614 1615 dso__set_annotate_warned(dso); 1616 symbol__strerror_disassemble(ms, err, msg, sizeof(msg)); 1617 ui__error("Couldn't annotate %s:\n%s", sym->name, msg); 1618 return -1; 1619 } 1620 1621 if (annotate_opts.print_lines) { 1622 srcline_full_filename = annotate_opts.full_path; 1623 symbol__calc_lines(ms, &source_line); 1624 print_summary(&source_line, dso__long_name(dso)); 1625 } 1626 1627 hists__scnprintf_title(hists, buf, sizeof(buf)); 1628 fprintf(stdout, "%s, [percent: %s]\n%s() %s\n", 1629 buf, percent_type_str(annotate_opts.percent_type), sym->name, dso__long_name(dso)); 1630 symbol__annotate_fprintf2(sym, stdout); 1631 1632 annotated_source__purge(symbol__annotation(sym)->src); 1633 1634 return 0; 1635 } 1636 1637 int symbol__tty_annotate(struct map_symbol *ms, struct evsel *evsel) 1638 { 1639 struct dso *dso = map__dso(ms->map); 1640 struct symbol *sym = ms->sym; 1641 struct rb_root source_line = RB_ROOT; 1642 int err; 1643 1644 err = symbol__annotate(ms, evsel, NULL); 1645 if (err) { 1646 char msg[BUFSIZ]; 1647 1648 dso__set_annotate_warned(dso); 1649 symbol__strerror_disassemble(ms, err, msg, sizeof(msg)); 1650 ui__error("Couldn't annotate %s:\n%s", sym->name, msg); 1651 return -1; 1652 } 1653 1654 symbol__calc_percent(sym, evsel); 1655 1656 if (annotate_opts.print_lines) { 1657 srcline_full_filename = annotate_opts.full_path; 1658 symbol__calc_lines(ms, &source_line); 1659 print_summary(&source_line, dso__long_name(dso)); 1660 } 1661 1662 symbol__annotate_printf(ms, evsel); 1663 1664 annotated_source__purge(symbol__annotation(sym)->src); 1665 1666 return 0; 1667 } 1668 1669 bool ui__has_annotation(void) 1670 { 1671 return use_browser == 1 && perf_hpp_list.sym; 1672 } 1673 1674 1675 static double annotation_line__max_percent(struct annotation_line *al, 1676 unsigned int percent_type) 1677 { 1678 double percent_max = 0.0; 1679 int i; 1680 1681 for (i = 0; i < al->data_nr; i++) { 1682 double percent; 1683 1684 percent = annotation_data__percent(&al->data[i], 1685 percent_type); 1686 1687 if (percent > percent_max) 1688 percent_max = percent; 1689 } 1690 1691 return percent_max; 1692 } 1693 1694 static void disasm_line__write(struct disasm_line *dl, struct annotation *notes, 1695 void *obj, char *bf, size_t size, 1696 void (*obj__printf)(void *obj, const char *fmt, ...), 1697 void (*obj__write_graph)(void *obj, int graph)) 1698 { 1699 if (dl->ins.ops && dl->ins.ops->scnprintf) { 1700 if (ins__is_jump(&dl->ins)) { 1701 bool fwd; 1702 1703 if (dl->ops.target.outside) 1704 goto call_like; 1705 fwd = dl->ops.target.offset > dl->al.offset; 1706 obj__write_graph(obj, fwd ? DARROW_CHAR : UARROW_CHAR); 1707 obj__printf(obj, " "); 1708 } else if (ins__is_call(&dl->ins)) { 1709 call_like: 1710 obj__write_graph(obj, RARROW_CHAR); 1711 obj__printf(obj, " "); 1712 } else if (ins__is_ret(&dl->ins)) { 1713 obj__write_graph(obj, LARROW_CHAR); 1714 obj__printf(obj, " "); 1715 } else { 1716 obj__printf(obj, " "); 1717 } 1718 } else { 1719 obj__printf(obj, " "); 1720 } 1721 1722 disasm_line__scnprintf(dl, bf, size, !annotate_opts.use_offset, 1723 notes->src->widths.max_ins_name); 1724 } 1725 1726 static void ipc_coverage_string(char *bf, int size, struct annotation *notes) 1727 { 1728 double ipc = 0.0, coverage = 0.0; 1729 struct annotated_branch *branch = annotation__get_branch(notes); 1730 1731 if (branch && branch->hit_cycles) 1732 ipc = branch->hit_insn / ((double)branch->hit_cycles); 1733 1734 if (branch && branch->total_insn) { 1735 coverage = branch->cover_insn * 100.0 / 1736 ((double)branch->total_insn); 1737 } 1738 1739 scnprintf(bf, size, "(Average IPC: %.2f, IPC Coverage: %.1f%%)", 1740 ipc, coverage); 1741 } 1742 1743 int annotation_br_cntr_abbr_list(char **str, struct evsel *evsel, bool header) 1744 { 1745 struct evsel *pos; 1746 struct strbuf sb; 1747 1748 if (evsel->evlist->nr_br_cntr <= 0) 1749 return -ENOTSUP; 1750 1751 strbuf_init(&sb, /*hint=*/ 0); 1752 1753 if (header && strbuf_addf(&sb, "# Branch counter abbr list:\n")) 1754 goto err; 1755 1756 evlist__for_each_entry(evsel->evlist, pos) { 1757 if (!(pos->core.attr.branch_sample_type & PERF_SAMPLE_BRANCH_COUNTERS)) 1758 continue; 1759 if (header && strbuf_addf(&sb, "#")) 1760 goto err; 1761 1762 if (strbuf_addf(&sb, " %s = %s\n", pos->name, pos->abbr_name)) 1763 goto err; 1764 } 1765 1766 if (header && strbuf_addf(&sb, "#")) 1767 goto err; 1768 if (strbuf_addf(&sb, " '-' No event occurs\n")) 1769 goto err; 1770 1771 if (header && strbuf_addf(&sb, "#")) 1772 goto err; 1773 if (strbuf_addf(&sb, " '+' Event occurrences may be lost due to branch counter saturated\n")) 1774 goto err; 1775 1776 *str = strbuf_detach(&sb, NULL); 1777 1778 return 0; 1779 err: 1780 strbuf_release(&sb); 1781 return -ENOMEM; 1782 } 1783 1784 /* Assume the branch counter saturated at 3 */ 1785 #define ANNOTATION_BR_CNTR_SATURATION 3 1786 1787 int annotation_br_cntr_entry(char **str, int br_cntr_nr, 1788 u64 *br_cntr, int num_aggr, 1789 struct evsel *evsel) 1790 { 1791 struct evsel *pos = evsel ? evlist__first(evsel->evlist) : NULL; 1792 bool saturated = false; 1793 int i, j, avg, used; 1794 struct strbuf sb; 1795 1796 strbuf_init(&sb, /*hint=*/ 0); 1797 for (i = 0; i < br_cntr_nr; i++) { 1798 used = 0; 1799 avg = ceil((double)(br_cntr[i] & ~ANNOTATION__BR_CNTR_SATURATED_FLAG) / 1800 (double)num_aggr); 1801 1802 /* 1803 * A histogram with the abbr name is displayed by default. 1804 * With -v, the exact number of branch counter is displayed. 1805 */ 1806 if (verbose) { 1807 evlist__for_each_entry_from(evsel->evlist, pos) { 1808 if ((pos->core.attr.branch_sample_type & PERF_SAMPLE_BRANCH_COUNTERS) && 1809 (pos->br_cntr_idx == i)) 1810 break; 1811 } 1812 if (strbuf_addstr(&sb, pos->abbr_name)) 1813 goto err; 1814 1815 if (!br_cntr[i]) { 1816 if (strbuf_addstr(&sb, "=-")) 1817 goto err; 1818 } else { 1819 if (strbuf_addf(&sb, "=%d", avg)) 1820 goto err; 1821 } 1822 if (br_cntr[i] & ANNOTATION__BR_CNTR_SATURATED_FLAG) { 1823 if (strbuf_addch(&sb, '+')) 1824 goto err; 1825 } else { 1826 if (strbuf_addch(&sb, ' ')) 1827 goto err; 1828 } 1829 1830 if ((i < br_cntr_nr - 1) && strbuf_addch(&sb, ',')) 1831 goto err; 1832 continue; 1833 } 1834 1835 if (strbuf_addch(&sb, '|')) 1836 goto err; 1837 1838 if (!br_cntr[i]) { 1839 if (strbuf_addch(&sb, '-')) 1840 goto err; 1841 used++; 1842 } else { 1843 evlist__for_each_entry_from(evsel->evlist, pos) { 1844 if ((pos->core.attr.branch_sample_type & PERF_SAMPLE_BRANCH_COUNTERS) && 1845 (pos->br_cntr_idx == i)) 1846 break; 1847 } 1848 if (br_cntr[i] & ANNOTATION__BR_CNTR_SATURATED_FLAG) 1849 saturated = true; 1850 1851 for (j = 0; j < avg; j++, used++) { 1852 /* Print + if the number of logged events > 3 */ 1853 if (j >= ANNOTATION_BR_CNTR_SATURATION) { 1854 saturated = true; 1855 break; 1856 } 1857 if (strbuf_addstr(&sb, pos->abbr_name)) 1858 goto err; 1859 } 1860 1861 if (saturated) { 1862 if (strbuf_addch(&sb, '+')) 1863 goto err; 1864 used++; 1865 } 1866 pos = list_next_entry(pos, core.node); 1867 } 1868 1869 for (j = used; j < ANNOTATION_BR_CNTR_SATURATION + 1; j++) { 1870 if (strbuf_addch(&sb, ' ')) 1871 goto err; 1872 } 1873 } 1874 1875 if (!verbose && strbuf_addch(&sb, br_cntr_nr ? '|' : ' ')) 1876 goto err; 1877 1878 *str = strbuf_detach(&sb, NULL); 1879 1880 return 0; 1881 err: 1882 strbuf_release(&sb); 1883 return -ENOMEM; 1884 } 1885 1886 static void __annotation_line__write(struct annotation_line *al, struct annotation *notes, 1887 bool first_line, bool current_entry, bool change_color, int width, 1888 void *obj, unsigned int percent_type, 1889 int (*obj__set_color)(void *obj, int color), 1890 void (*obj__set_percent_color)(void *obj, double percent, bool current), 1891 int (*obj__set_jumps_percent_color)(void *obj, int nr, bool current), 1892 void (*obj__printf)(void *obj, const char *fmt, ...), 1893 void (*obj__write_graph)(void *obj, int graph)) 1894 1895 { 1896 double percent_max = annotation_line__max_percent(al, percent_type); 1897 int pcnt_width = annotation__pcnt_width(notes), 1898 cycles_width = annotation__cycles_width(notes); 1899 bool show_title = false; 1900 char bf[256]; 1901 int printed; 1902 1903 if (first_line && (al->offset == -1 || percent_max == 0.0)) { 1904 if (notes->branch && al->cycles) { 1905 if (al->cycles->ipc == 0.0 && al->cycles->avg == 0) 1906 show_title = true; 1907 } else 1908 show_title = true; 1909 } 1910 1911 if (al->offset != -1 && percent_max != 0.0) { 1912 int i; 1913 1914 for (i = 0; i < al->data_nr; i++) { 1915 double percent; 1916 1917 percent = annotation_data__percent(&al->data[i], percent_type); 1918 1919 obj__set_percent_color(obj, percent, current_entry); 1920 if (symbol_conf.show_total_period) { 1921 obj__printf(obj, "%11" PRIu64 " ", al->data[i].he.period); 1922 } else if (symbol_conf.show_nr_samples) { 1923 obj__printf(obj, "%7" PRIu64 " ", 1924 al->data[i].he.nr_samples); 1925 } else { 1926 obj__printf(obj, "%7.2f ", percent); 1927 } 1928 } 1929 } else { 1930 obj__set_percent_color(obj, 0, current_entry); 1931 1932 if (!show_title) 1933 obj__printf(obj, "%-*s", pcnt_width, " "); 1934 else { 1935 obj__printf(obj, "%-*s", pcnt_width, 1936 symbol_conf.show_total_period ? "Period" : 1937 symbol_conf.show_nr_samples ? "Samples" : "Percent"); 1938 } 1939 } 1940 1941 if (notes->branch) { 1942 if (al->cycles && al->cycles->ipc) 1943 obj__printf(obj, "%*.2f ", ANNOTATION__IPC_WIDTH - 1, al->cycles->ipc); 1944 else if (!show_title) 1945 obj__printf(obj, "%*s", ANNOTATION__IPC_WIDTH, " "); 1946 else 1947 obj__printf(obj, "%*s ", ANNOTATION__IPC_WIDTH - 1, "IPC"); 1948 1949 if (!annotate_opts.show_minmax_cycle) { 1950 if (al->cycles && al->cycles->avg) 1951 obj__printf(obj, "%*" PRIu64 " ", 1952 ANNOTATION__CYCLES_WIDTH - 1, al->cycles->avg); 1953 else if (!show_title) 1954 obj__printf(obj, "%*s", 1955 ANNOTATION__CYCLES_WIDTH, " "); 1956 else 1957 obj__printf(obj, "%*s ", 1958 ANNOTATION__CYCLES_WIDTH - 1, 1959 "Cycle"); 1960 } else { 1961 if (al->cycles) { 1962 char str[32]; 1963 1964 scnprintf(str, sizeof(str), 1965 "%" PRIu64 "(%" PRIu64 "/%" PRIu64 ")", 1966 al->cycles->avg, al->cycles->min, 1967 al->cycles->max); 1968 1969 obj__printf(obj, "%*s ", 1970 ANNOTATION__MINMAX_CYCLES_WIDTH - 1, 1971 str); 1972 } else if (!show_title) 1973 obj__printf(obj, "%*s", 1974 ANNOTATION__MINMAX_CYCLES_WIDTH, 1975 " "); 1976 else 1977 obj__printf(obj, "%*s ", 1978 ANNOTATION__MINMAX_CYCLES_WIDTH - 1, 1979 "Cycle(min/max)"); 1980 } 1981 1982 if (annotate_opts.show_br_cntr) { 1983 if (show_title) { 1984 obj__printf(obj, "%*s ", 1985 ANNOTATION__BR_CNTR_WIDTH, 1986 "Branch Counter"); 1987 } else { 1988 char *buf; 1989 1990 if (!annotation_br_cntr_entry(&buf, al->br_cntr_nr, al->br_cntr, 1991 al->num_aggr, al->evsel)) { 1992 obj__printf(obj, "%*s ", ANNOTATION__BR_CNTR_WIDTH, buf); 1993 free(buf); 1994 } 1995 } 1996 } 1997 1998 if (show_title && !*al->line) { 1999 ipc_coverage_string(bf, sizeof(bf), notes); 2000 obj__printf(obj, "%*s", ANNOTATION__AVG_IPC_WIDTH, bf); 2001 } 2002 } 2003 2004 obj__printf(obj, " "); 2005 2006 if (!*al->line) 2007 obj__printf(obj, "%-*s", width - pcnt_width - cycles_width, " "); 2008 else if (al->offset == -1) { 2009 if (al->line_nr && annotate_opts.show_linenr) 2010 printed = scnprintf(bf, sizeof(bf), "%-*d ", 2011 notes->src->widths.addr + 1, al->line_nr); 2012 else 2013 printed = scnprintf(bf, sizeof(bf), "%-*s ", 2014 notes->src->widths.addr, " "); 2015 obj__printf(obj, bf); 2016 obj__printf(obj, "%-*s", width - printed - pcnt_width - cycles_width + 1, al->line); 2017 } else { 2018 u64 addr = al->offset; 2019 int color = -1; 2020 2021 if (!annotate_opts.use_offset) 2022 addr += notes->src->start; 2023 2024 if (!annotate_opts.use_offset) { 2025 printed = scnprintf(bf, sizeof(bf), "%" PRIx64 ": ", addr); 2026 } else { 2027 if (al->jump_sources && 2028 annotate_opts.offset_level >= ANNOTATION__OFFSET_JUMP_TARGETS) { 2029 if (annotate_opts.show_nr_jumps) { 2030 int prev; 2031 printed = scnprintf(bf, sizeof(bf), "%*d ", 2032 notes->src->widths.jumps, 2033 al->jump_sources); 2034 prev = obj__set_jumps_percent_color(obj, al->jump_sources, 2035 current_entry); 2036 obj__printf(obj, bf); 2037 obj__set_color(obj, prev); 2038 } 2039 print_addr: 2040 printed = scnprintf(bf, sizeof(bf), "%*" PRIx64 ": ", 2041 notes->src->widths.target, addr); 2042 } else if (ins__is_call(&disasm_line(al)->ins) && 2043 annotate_opts.offset_level >= ANNOTATION__OFFSET_CALL) { 2044 goto print_addr; 2045 } else if (annotate_opts.offset_level == ANNOTATION__MAX_OFFSET_LEVEL) { 2046 goto print_addr; 2047 } else { 2048 printed = scnprintf(bf, sizeof(bf), "%-*s ", 2049 notes->src->widths.addr, " "); 2050 } 2051 } 2052 2053 if (change_color) 2054 color = obj__set_color(obj, HE_COLORSET_ADDR); 2055 obj__printf(obj, bf); 2056 if (change_color) 2057 obj__set_color(obj, color); 2058 2059 disasm_line__write(disasm_line(al), notes, obj, bf, sizeof(bf), obj__printf, obj__write_graph); 2060 2061 obj__printf(obj, "%-*s", width - pcnt_width - cycles_width - 3 - printed, bf); 2062 } 2063 2064 } 2065 2066 void annotation_line__write(struct annotation_line *al, struct annotation *notes, 2067 struct annotation_write_ops *wops) 2068 { 2069 __annotation_line__write(al, notes, wops->first_line, wops->current_entry, 2070 wops->change_color, wops->width, wops->obj, 2071 annotate_opts.percent_type, 2072 wops->set_color, wops->set_percent_color, 2073 wops->set_jumps_percent_color, wops->printf, 2074 wops->write_graph); 2075 } 2076 2077 int symbol__annotate2(struct map_symbol *ms, struct evsel *evsel, 2078 struct arch **parch) 2079 { 2080 struct symbol *sym = ms->sym; 2081 struct annotation *notes = symbol__annotation(sym); 2082 size_t size = symbol__size(sym); 2083 int err; 2084 2085 err = symbol__annotate(ms, evsel, parch); 2086 if (err) 2087 return err; 2088 2089 symbol__calc_percent(sym, evsel); 2090 2091 annotation__set_index(notes); 2092 annotation__mark_jump_targets(notes, sym); 2093 2094 err = annotation__compute_ipc(notes, size, evsel); 2095 if (err) 2096 return err; 2097 2098 annotation__init_column_widths(notes, sym); 2099 annotation__update_column_widths(notes); 2100 sym->annotate2 = 1; 2101 2102 return 0; 2103 } 2104 2105 static int annotation__config(const char *var, const char *value, void *data) 2106 { 2107 struct annotation_options *opt = data; 2108 2109 if (!strstarts(var, "annotate.")) 2110 return 0; 2111 2112 if (!strcmp(var, "annotate.offset_level")) { 2113 perf_config_u8(&opt->offset_level, "offset_level", value); 2114 2115 if (opt->offset_level > ANNOTATION__MAX_OFFSET_LEVEL) 2116 opt->offset_level = ANNOTATION__MAX_OFFSET_LEVEL; 2117 else if (opt->offset_level < ANNOTATION__MIN_OFFSET_LEVEL) 2118 opt->offset_level = ANNOTATION__MIN_OFFSET_LEVEL; 2119 } else if (!strcmp(var, "annotate.disassemblers")) { 2120 opt->disassemblers_str = strdup(value); 2121 if (!opt->disassemblers_str) { 2122 pr_err("Not enough memory for annotate.disassemblers\n"); 2123 return -1; 2124 } 2125 } else if (!strcmp(var, "annotate.hide_src_code")) { 2126 opt->hide_src_code = perf_config_bool("hide_src_code", value); 2127 } else if (!strcmp(var, "annotate.jump_arrows")) { 2128 opt->jump_arrows = perf_config_bool("jump_arrows", value); 2129 } else if (!strcmp(var, "annotate.show_linenr")) { 2130 opt->show_linenr = perf_config_bool("show_linenr", value); 2131 } else if (!strcmp(var, "annotate.show_nr_jumps")) { 2132 opt->show_nr_jumps = perf_config_bool("show_nr_jumps", value); 2133 } else if (!strcmp(var, "annotate.show_nr_samples")) { 2134 symbol_conf.show_nr_samples = perf_config_bool("show_nr_samples", 2135 value); 2136 } else if (!strcmp(var, "annotate.show_total_period")) { 2137 symbol_conf.show_total_period = perf_config_bool("show_total_period", 2138 value); 2139 } else if (!strcmp(var, "annotate.use_offset")) { 2140 opt->use_offset = perf_config_bool("use_offset", value); 2141 } else if (!strcmp(var, "annotate.disassembler_style")) { 2142 opt->disassembler_style = strdup(value); 2143 if (!opt->disassembler_style) { 2144 pr_err("Not enough memory for annotate.disassembler_style\n"); 2145 return -1; 2146 } 2147 } else if (!strcmp(var, "annotate.objdump")) { 2148 opt->objdump_path = strdup(value); 2149 if (!opt->objdump_path) { 2150 pr_err("Not enough memory for annotate.objdump\n"); 2151 return -1; 2152 } 2153 } else if (!strcmp(var, "annotate.addr2line")) { 2154 symbol_conf.addr2line_path = strdup(value); 2155 if (!symbol_conf.addr2line_path) { 2156 pr_err("Not enough memory for annotate.addr2line\n"); 2157 return -1; 2158 } 2159 } else if (!strcmp(var, "annotate.demangle")) { 2160 symbol_conf.demangle = perf_config_bool("demangle", value); 2161 } else if (!strcmp(var, "annotate.demangle_kernel")) { 2162 symbol_conf.demangle_kernel = perf_config_bool("demangle_kernel", value); 2163 } else { 2164 pr_debug("%s variable unknown, ignoring...", var); 2165 } 2166 2167 return 0; 2168 } 2169 2170 void annotation_options__init(void) 2171 { 2172 struct annotation_options *opt = &annotate_opts; 2173 2174 memset(opt, 0, sizeof(*opt)); 2175 2176 /* Default values. */ 2177 opt->use_offset = true; 2178 opt->jump_arrows = true; 2179 opt->annotate_src = true; 2180 opt->offset_level = ANNOTATION__OFFSET_JUMP_TARGETS; 2181 opt->percent_type = PERCENT_PERIOD_LOCAL; 2182 } 2183 2184 void annotation_options__exit(void) 2185 { 2186 zfree(&annotate_opts.disassembler_style); 2187 zfree(&annotate_opts.objdump_path); 2188 } 2189 2190 void annotation_config__init(void) 2191 { 2192 perf_config(annotation__config, &annotate_opts); 2193 } 2194 2195 static unsigned int parse_percent_type(char *str1, char *str2) 2196 { 2197 unsigned int type = (unsigned int) -1; 2198 2199 if (!strcmp("period", str1)) { 2200 if (!strcmp("local", str2)) 2201 type = PERCENT_PERIOD_LOCAL; 2202 else if (!strcmp("global", str2)) 2203 type = PERCENT_PERIOD_GLOBAL; 2204 } 2205 2206 if (!strcmp("hits", str1)) { 2207 if (!strcmp("local", str2)) 2208 type = PERCENT_HITS_LOCAL; 2209 else if (!strcmp("global", str2)) 2210 type = PERCENT_HITS_GLOBAL; 2211 } 2212 2213 return type; 2214 } 2215 2216 int annotate_parse_percent_type(const struct option *opt __maybe_unused, const char *_str, 2217 int unset __maybe_unused) 2218 { 2219 unsigned int type; 2220 char *str1, *str2; 2221 int err = -1; 2222 2223 str1 = strdup(_str); 2224 if (!str1) 2225 return -ENOMEM; 2226 2227 str2 = strchr(str1, '-'); 2228 if (!str2) 2229 goto out; 2230 2231 *str2++ = 0; 2232 2233 type = parse_percent_type(str1, str2); 2234 if (type == (unsigned int) -1) 2235 type = parse_percent_type(str2, str1); 2236 if (type != (unsigned int) -1) { 2237 annotate_opts.percent_type = type; 2238 err = 0; 2239 } 2240 2241 out: 2242 free(str1); 2243 return err; 2244 } 2245 2246 int annotate_check_args(void) 2247 { 2248 struct annotation_options *args = &annotate_opts; 2249 2250 if (args->prefix_strip && !args->prefix) { 2251 pr_err("--prefix-strip requires --prefix\n"); 2252 return -1; 2253 } 2254 return 0; 2255 } 2256 2257 /* 2258 * Get register number and access offset from the given instruction. 2259 * It assumes AT&T x86 asm format like OFFSET(REG). Maybe it needs 2260 * to revisit the format when it handles different architecture. 2261 * Fills @reg and @offset when return 0. 2262 */ 2263 static int extract_reg_offset(struct arch *arch, const char *str, 2264 struct annotated_op_loc *op_loc) 2265 { 2266 char *p; 2267 char *regname; 2268 2269 if (arch->objdump.register_char == 0) 2270 return -1; 2271 2272 /* 2273 * It should start from offset, but it's possible to skip 0 2274 * in the asm. So 0(%rax) should be same as (%rax). 2275 * 2276 * However, it also start with a segment select register like 2277 * %gs:0x18(%rbx). In that case it should skip the part. 2278 */ 2279 if (*str == arch->objdump.register_char) { 2280 if (arch__is(arch, "x86")) { 2281 /* FIXME: Handle other segment registers */ 2282 if (!strncmp(str, "%gs:", 4)) 2283 op_loc->segment = INSN_SEG_X86_GS; 2284 } 2285 2286 while (*str && !isdigit(*str) && 2287 *str != arch->objdump.memory_ref_char) 2288 str++; 2289 } 2290 2291 op_loc->offset = strtol(str, &p, 0); 2292 2293 p = strchr(p, arch->objdump.register_char); 2294 if (p == NULL) 2295 return -1; 2296 2297 regname = strdup(p); 2298 if (regname == NULL) 2299 return -1; 2300 2301 op_loc->reg1 = get_dwarf_regnum(regname, arch->e_machine, arch->e_flags); 2302 free(regname); 2303 2304 /* Get the second register */ 2305 if (op_loc->multi_regs) { 2306 p = strchr(p + 1, arch->objdump.register_char); 2307 if (p == NULL) 2308 return -1; 2309 2310 regname = strdup(p); 2311 if (regname == NULL) 2312 return -1; 2313 2314 op_loc->reg2 = get_dwarf_regnum(regname, arch->e_machine, arch->e_flags); 2315 free(regname); 2316 } 2317 return 0; 2318 } 2319 2320 /** 2321 * annotate_get_insn_location - Get location of instruction 2322 * @arch: the architecture info 2323 * @dl: the target instruction 2324 * @loc: a buffer to save the data 2325 * 2326 * Get detailed location info (register and offset) in the instruction. 2327 * It needs both source and target operand and whether it accesses a 2328 * memory location. The offset field is meaningful only when the 2329 * corresponding mem flag is set. The reg2 field is meaningful only 2330 * when multi_regs flag is set. 2331 * 2332 * Some examples on x86: 2333 * 2334 * mov (%rax), %rcx # src_reg1 = rax, src_mem = 1, src_offset = 0 2335 * # dst_reg1 = rcx, dst_mem = 0 2336 * 2337 * mov 0x18, %r8 # src_reg1 = -1, src_mem = 0 2338 * # dst_reg1 = r8, dst_mem = 0 2339 * 2340 * mov %rsi, 8(%rbx,%rcx,4) # src_reg1 = rsi, src_mem = 0, src_multi_regs = 0 2341 * # dst_reg1 = rbx, dst_reg2 = rcx, dst_mem = 1 2342 * # dst_multi_regs = 1, dst_offset = 8 2343 */ 2344 int annotate_get_insn_location(struct arch *arch, struct disasm_line *dl, 2345 struct annotated_insn_loc *loc) 2346 { 2347 struct ins_operands *ops; 2348 struct annotated_op_loc *op_loc; 2349 int i; 2350 2351 if (ins__is_lock(&dl->ins)) 2352 ops = dl->ops.locked.ops; 2353 else 2354 ops = &dl->ops; 2355 2356 if (ops == NULL) 2357 return -1; 2358 2359 memset(loc, 0, sizeof(*loc)); 2360 2361 for_each_insn_op_loc(loc, i, op_loc) { 2362 const char *insn_str = ops->source.raw; 2363 bool multi_regs = ops->source.multi_regs; 2364 bool mem_ref = ops->source.mem_ref; 2365 2366 if (i == INSN_OP_TARGET) { 2367 insn_str = ops->target.raw; 2368 multi_regs = ops->target.multi_regs; 2369 mem_ref = ops->target.mem_ref; 2370 } 2371 2372 /* Invalidate the register by default */ 2373 op_loc->reg1 = -1; 2374 op_loc->reg2 = -1; 2375 2376 if (insn_str == NULL) { 2377 if (!arch__is(arch, "powerpc")) 2378 continue; 2379 } 2380 2381 /* 2382 * For powerpc, call get_powerpc_regs function which extracts the 2383 * required fields for op_loc, ie reg1, reg2, offset from the 2384 * raw instruction. 2385 */ 2386 if (arch__is(arch, "powerpc")) { 2387 op_loc->mem_ref = mem_ref; 2388 op_loc->multi_regs = multi_regs; 2389 get_powerpc_regs(dl->raw.raw_insn, !i, op_loc); 2390 } else if (strchr(insn_str, arch->objdump.memory_ref_char)) { 2391 op_loc->mem_ref = true; 2392 op_loc->multi_regs = multi_regs; 2393 extract_reg_offset(arch, insn_str, op_loc); 2394 } else { 2395 char *s, *p = NULL; 2396 2397 if (arch__is(arch, "x86")) { 2398 /* FIXME: Handle other segment registers */ 2399 if (!strncmp(insn_str, "%gs:", 4)) { 2400 op_loc->segment = INSN_SEG_X86_GS; 2401 op_loc->offset = strtol(insn_str + 4, 2402 &p, 0); 2403 if (p && p != insn_str + 4) 2404 op_loc->imm = true; 2405 continue; 2406 } 2407 } 2408 2409 s = strdup(insn_str); 2410 if (s == NULL) 2411 return -1; 2412 2413 if (*s == arch->objdump.register_char) 2414 op_loc->reg1 = get_dwarf_regnum(s, arch->e_machine, arch->e_flags); 2415 else if (*s == arch->objdump.imm_char) { 2416 op_loc->offset = strtol(s + 1, &p, 0); 2417 if (p && p != s + 1) 2418 op_loc->imm = true; 2419 } 2420 free(s); 2421 } 2422 } 2423 2424 return 0; 2425 } 2426 2427 static struct disasm_line *find_disasm_line(struct symbol *sym, u64 ip, 2428 bool allow_update) 2429 { 2430 struct disasm_line *dl; 2431 struct annotation *notes; 2432 2433 notes = symbol__annotation(sym); 2434 2435 list_for_each_entry(dl, ¬es->src->source, al.node) { 2436 if (dl->al.offset == -1) 2437 continue; 2438 2439 if (sym->start + dl->al.offset == ip) { 2440 /* 2441 * llvm-objdump places "lock" in a separate line and 2442 * in that case, we want to get the next line. 2443 */ 2444 if (ins__is_lock(&dl->ins) && 2445 *dl->ops.raw == '\0' && allow_update) { 2446 ip++; 2447 continue; 2448 } 2449 return dl; 2450 } 2451 } 2452 return NULL; 2453 } 2454 2455 static struct annotated_item_stat *annotate_data_stat(struct list_head *head, 2456 const char *name) 2457 { 2458 struct annotated_item_stat *istat; 2459 2460 list_for_each_entry(istat, head, list) { 2461 if (!strcmp(istat->name, name)) 2462 return istat; 2463 } 2464 2465 istat = zalloc(sizeof(*istat)); 2466 if (istat == NULL) 2467 return NULL; 2468 2469 istat->name = strdup(name); 2470 if ((istat->name == NULL) || (!strlen(istat->name))) { 2471 free(istat); 2472 return NULL; 2473 } 2474 2475 list_add_tail(&istat->list, head); 2476 return istat; 2477 } 2478 2479 static bool is_stack_operation(struct arch *arch, struct disasm_line *dl) 2480 { 2481 if (arch__is(arch, "x86")) { 2482 if (!strncmp(dl->ins.name, "push", 4) || 2483 !strncmp(dl->ins.name, "pop", 3) || 2484 !strncmp(dl->ins.name, "call", 4) || 2485 !strncmp(dl->ins.name, "ret", 3)) 2486 return true; 2487 } 2488 2489 return false; 2490 } 2491 2492 static bool is_stack_canary(struct arch *arch, struct annotated_op_loc *loc) 2493 { 2494 /* On x86_64, %gs:40 is used for stack canary */ 2495 if (arch__is(arch, "x86")) { 2496 if (loc->segment == INSN_SEG_X86_GS && loc->imm && 2497 loc->offset == 40) 2498 return true; 2499 } 2500 2501 return false; 2502 } 2503 2504 static struct disasm_line * 2505 annotation__prev_asm_line(struct annotation *notes, struct disasm_line *curr) 2506 { 2507 struct list_head *sources = ¬es->src->source; 2508 struct disasm_line *prev; 2509 2510 if (curr == list_first_entry(sources, struct disasm_line, al.node)) 2511 return NULL; 2512 2513 prev = list_prev_entry(curr, al.node); 2514 while (prev->al.offset == -1 && 2515 prev != list_first_entry(sources, struct disasm_line, al.node)) 2516 prev = list_prev_entry(prev, al.node); 2517 2518 if (prev->al.offset == -1) 2519 return NULL; 2520 2521 return prev; 2522 } 2523 2524 static struct disasm_line * 2525 annotation__next_asm_line(struct annotation *notes, struct disasm_line *curr) 2526 { 2527 struct list_head *sources = ¬es->src->source; 2528 struct disasm_line *next; 2529 2530 if (curr == list_last_entry(sources, struct disasm_line, al.node)) 2531 return NULL; 2532 2533 next = list_next_entry(curr, al.node); 2534 while (next->al.offset == -1 && 2535 next != list_last_entry(sources, struct disasm_line, al.node)) 2536 next = list_next_entry(next, al.node); 2537 2538 if (next->al.offset == -1) 2539 return NULL; 2540 2541 return next; 2542 } 2543 2544 u64 annotate_calc_pcrel(struct map_symbol *ms, u64 ip, int offset, 2545 struct disasm_line *dl) 2546 { 2547 struct annotation *notes; 2548 struct disasm_line *next; 2549 u64 addr; 2550 2551 notes = symbol__annotation(ms->sym); 2552 /* 2553 * PC-relative addressing starts from the next instruction address 2554 * But the IP is for the current instruction. Since disasm_line 2555 * doesn't have the instruction size, calculate it using the next 2556 * disasm_line. If it's the last one, we can use symbol's end 2557 * address directly. 2558 */ 2559 next = annotation__next_asm_line(notes, dl); 2560 if (next == NULL) 2561 addr = ms->sym->end + offset; 2562 else 2563 addr = ip + (next->al.offset - dl->al.offset) + offset; 2564 2565 return map__rip_2objdump(ms->map, addr); 2566 } 2567 2568 static struct debuginfo_cache { 2569 struct dso *dso; 2570 struct debuginfo *dbg; 2571 } di_cache; 2572 2573 void debuginfo_cache__delete(void) 2574 { 2575 dso__put(di_cache.dso); 2576 di_cache.dso = NULL; 2577 2578 debuginfo__delete(di_cache.dbg); 2579 di_cache.dbg = NULL; 2580 } 2581 2582 /** 2583 * hist_entry__get_data_type - find data type for given hist entry 2584 * @he: hist entry 2585 * 2586 * This function first annotates the instruction at @he->ip and extracts 2587 * register and offset info from it. Then it searches the DWARF debug 2588 * info to get a variable and type information using the address, register, 2589 * and offset. 2590 */ 2591 struct annotated_data_type *hist_entry__get_data_type(struct hist_entry *he) 2592 { 2593 struct map_symbol *ms = &he->ms; 2594 struct evsel *evsel = hists_to_evsel(he->hists); 2595 struct arch *arch; 2596 struct disasm_line *dl; 2597 struct annotated_insn_loc loc; 2598 struct annotated_op_loc *op_loc; 2599 struct annotated_data_type *mem_type; 2600 struct annotated_item_stat *istat; 2601 u64 ip = he->ip; 2602 int i; 2603 2604 ann_data_stat.total++; 2605 2606 if (ms->map == NULL || ms->sym == NULL) { 2607 ann_data_stat.no_sym++; 2608 return NULL; 2609 } 2610 2611 if (!symbol_conf.init_annotation) { 2612 ann_data_stat.no_sym++; 2613 return NULL; 2614 } 2615 2616 /* 2617 * di_cache holds a pair of values, but code below assumes 2618 * di_cache.dso can be compared/updated and di_cache.dbg can be 2619 * read/updated independently from each other. That assumption only 2620 * holds in single threaded code. 2621 */ 2622 assert(perf_singlethreaded); 2623 2624 if (map__dso(ms->map) != di_cache.dso) { 2625 dso__put(di_cache.dso); 2626 di_cache.dso = dso__get(map__dso(ms->map)); 2627 2628 debuginfo__delete(di_cache.dbg); 2629 di_cache.dbg = debuginfo__new(dso__long_name(di_cache.dso)); 2630 } 2631 2632 if (di_cache.dbg == NULL) { 2633 ann_data_stat.no_dbginfo++; 2634 return NULL; 2635 } 2636 2637 /* Make sure it has the disasm of the function */ 2638 if (symbol__annotate(ms, evsel, &arch) < 0) { 2639 ann_data_stat.no_insn++; 2640 return NULL; 2641 } 2642 2643 /* 2644 * Get a disasm to extract the location from the insn. 2645 * This is too slow... 2646 */ 2647 dl = find_disasm_line(ms->sym, ip, /*allow_update=*/true); 2648 if (dl == NULL) { 2649 ann_data_stat.no_insn++; 2650 return NULL; 2651 } 2652 2653 retry: 2654 istat = annotate_data_stat(&ann_insn_stat, dl->ins.name); 2655 if (istat == NULL) { 2656 ann_data_stat.no_insn++; 2657 return NULL; 2658 } 2659 2660 if (annotate_get_insn_location(arch, dl, &loc) < 0) { 2661 ann_data_stat.no_insn_ops++; 2662 istat->bad++; 2663 return NULL; 2664 } 2665 2666 if (is_stack_operation(arch, dl)) { 2667 istat->good++; 2668 he->mem_type_off = 0; 2669 return &stackop_type; 2670 } 2671 2672 for_each_insn_op_loc(&loc, i, op_loc) { 2673 struct data_loc_info dloc = { 2674 .arch = arch, 2675 .thread = he->thread, 2676 .ms = ms, 2677 /* Recalculate IP for LOCK prefix or insn fusion */ 2678 .ip = ms->sym->start + dl->al.offset, 2679 .cpumode = he->cpumode, 2680 .op = op_loc, 2681 .di = di_cache.dbg, 2682 }; 2683 2684 if (!op_loc->mem_ref && op_loc->segment == INSN_SEG_NONE) 2685 continue; 2686 2687 /* Recalculate IP because of LOCK prefix or insn fusion */ 2688 ip = ms->sym->start + dl->al.offset; 2689 2690 /* PC-relative addressing */ 2691 if (op_loc->reg1 == DWARF_REG_PC) { 2692 dloc.var_addr = annotate_calc_pcrel(ms, dloc.ip, 2693 op_loc->offset, dl); 2694 } 2695 2696 /* This CPU access in kernel - pretend PC-relative addressing */ 2697 if (dso__kernel(map__dso(ms->map)) && arch__is(arch, "x86") && 2698 op_loc->segment == INSN_SEG_X86_GS && op_loc->imm) { 2699 dloc.var_addr = op_loc->offset; 2700 op_loc->reg1 = DWARF_REG_PC; 2701 } 2702 2703 mem_type = find_data_type(&dloc); 2704 2705 if (mem_type == NULL && is_stack_canary(arch, op_loc)) { 2706 istat->good++; 2707 he->mem_type_off = 0; 2708 return &canary_type; 2709 } 2710 2711 if (mem_type) 2712 istat->good++; 2713 else 2714 istat->bad++; 2715 2716 if (symbol_conf.annotate_data_sample) { 2717 annotated_data_type__update_samples(mem_type, evsel, 2718 dloc.type_offset, 2719 he->stat.nr_events, 2720 he->stat.period); 2721 } 2722 he->mem_type_off = dloc.type_offset; 2723 return mem_type; 2724 } 2725 2726 /* 2727 * Some instructions can be fused and the actual memory access came 2728 * from the previous instruction. 2729 */ 2730 if (dl->al.offset > 0) { 2731 struct annotation *notes; 2732 struct disasm_line *prev_dl; 2733 2734 notes = symbol__annotation(ms->sym); 2735 prev_dl = annotation__prev_asm_line(notes, dl); 2736 2737 if (prev_dl && ins__is_fused(arch, prev_dl->ins.name, dl->ins.name)) { 2738 dl = prev_dl; 2739 goto retry; 2740 } 2741 } 2742 2743 ann_data_stat.no_mem_ops++; 2744 istat->bad++; 2745 return NULL; 2746 } 2747 2748 /* Basic block traversal (BFS) data structure */ 2749 struct basic_block_data { 2750 struct list_head queue; 2751 struct list_head visited; 2752 }; 2753 2754 /* 2755 * During the traversal, it needs to know the parent block where the current 2756 * block block started from. Note that single basic block can be parent of 2757 * two child basic blocks (in case of condition jump). 2758 */ 2759 struct basic_block_link { 2760 struct list_head node; 2761 struct basic_block_link *parent; 2762 struct annotated_basic_block *bb; 2763 }; 2764 2765 /* Check any of basic block in the list already has the offset */ 2766 static bool basic_block_has_offset(struct list_head *head, s64 offset) 2767 { 2768 struct basic_block_link *link; 2769 2770 list_for_each_entry(link, head, node) { 2771 s64 begin_offset = link->bb->begin->al.offset; 2772 s64 end_offset = link->bb->end->al.offset; 2773 2774 if (begin_offset <= offset && offset <= end_offset) 2775 return true; 2776 } 2777 return false; 2778 } 2779 2780 static bool is_new_basic_block(struct basic_block_data *bb_data, 2781 struct disasm_line *dl) 2782 { 2783 s64 offset = dl->al.offset; 2784 2785 if (basic_block_has_offset(&bb_data->visited, offset)) 2786 return false; 2787 if (basic_block_has_offset(&bb_data->queue, offset)) 2788 return false; 2789 return true; 2790 } 2791 2792 /* Add a basic block starting from dl and link it to the parent */ 2793 static int add_basic_block(struct basic_block_data *bb_data, 2794 struct basic_block_link *parent, 2795 struct disasm_line *dl) 2796 { 2797 struct annotated_basic_block *bb; 2798 struct basic_block_link *link; 2799 2800 if (dl == NULL) 2801 return -1; 2802 2803 if (!is_new_basic_block(bb_data, dl)) 2804 return 0; 2805 2806 bb = zalloc(sizeof(*bb)); 2807 if (bb == NULL) 2808 return -1; 2809 2810 bb->begin = dl; 2811 bb->end = dl; 2812 INIT_LIST_HEAD(&bb->list); 2813 2814 link = malloc(sizeof(*link)); 2815 if (link == NULL) { 2816 free(bb); 2817 return -1; 2818 } 2819 2820 link->bb = bb; 2821 link->parent = parent; 2822 list_add_tail(&link->node, &bb_data->queue); 2823 return 0; 2824 } 2825 2826 /* Returns true when it finds the target in the current basic block */ 2827 static bool process_basic_block(struct basic_block_data *bb_data, 2828 struct basic_block_link *link, 2829 struct symbol *sym, u64 target) 2830 { 2831 struct disasm_line *dl, *next_dl, *last_dl; 2832 struct annotation *notes = symbol__annotation(sym); 2833 bool found = false; 2834 2835 dl = link->bb->begin; 2836 /* Check if it's already visited */ 2837 if (basic_block_has_offset(&bb_data->visited, dl->al.offset)) 2838 return false; 2839 2840 last_dl = list_last_entry(¬es->src->source, 2841 struct disasm_line, al.node); 2842 if (last_dl->al.offset == -1) 2843 last_dl = annotation__prev_asm_line(notes, last_dl); 2844 2845 if (last_dl == NULL) 2846 return false; 2847 2848 list_for_each_entry_from(dl, ¬es->src->source, al.node) { 2849 /* Skip comment or debug info line */ 2850 if (dl->al.offset == -1) 2851 continue; 2852 /* Found the target instruction */ 2853 if (sym->start + dl->al.offset == target) { 2854 found = true; 2855 break; 2856 } 2857 /* End of the function, finish the block */ 2858 if (dl == last_dl) 2859 break; 2860 /* 'return' instruction finishes the block */ 2861 if (ins__is_ret(&dl->ins)) 2862 break; 2863 /* normal instructions are part of the basic block */ 2864 if (!ins__is_jump(&dl->ins)) 2865 continue; 2866 /* jump to a different function, tail call or return */ 2867 if (dl->ops.target.outside) 2868 break; 2869 /* jump instruction creates new basic block(s) */ 2870 next_dl = find_disasm_line(sym, sym->start + dl->ops.target.offset, 2871 /*allow_update=*/false); 2872 if (next_dl) 2873 add_basic_block(bb_data, link, next_dl); 2874 2875 /* 2876 * FIXME: determine conditional jumps properly. 2877 * Conditional jumps create another basic block with the 2878 * next disasm line. 2879 */ 2880 if (!strstr(dl->ins.name, "jmp")) { 2881 next_dl = annotation__next_asm_line(notes, dl); 2882 if (next_dl) 2883 add_basic_block(bb_data, link, next_dl); 2884 } 2885 break; 2886 2887 } 2888 link->bb->end = dl; 2889 return found; 2890 } 2891 2892 /* 2893 * It founds a target basic block, build a proper linked list of basic blocks 2894 * by following the link recursively. 2895 */ 2896 static void link_found_basic_blocks(struct basic_block_link *link, 2897 struct list_head *head) 2898 { 2899 while (link) { 2900 struct basic_block_link *parent = link->parent; 2901 2902 list_move(&link->bb->list, head); 2903 list_del(&link->node); 2904 free(link); 2905 2906 link = parent; 2907 } 2908 } 2909 2910 static void delete_basic_blocks(struct basic_block_data *bb_data) 2911 { 2912 struct basic_block_link *link, *tmp; 2913 2914 list_for_each_entry_safe(link, tmp, &bb_data->queue, node) { 2915 list_del(&link->node); 2916 zfree(&link->bb); 2917 free(link); 2918 } 2919 2920 list_for_each_entry_safe(link, tmp, &bb_data->visited, node) { 2921 list_del(&link->node); 2922 zfree(&link->bb); 2923 free(link); 2924 } 2925 } 2926 2927 /** 2928 * annotate_get_basic_blocks - Get basic blocks for given address range 2929 * @sym: symbol to annotate 2930 * @src: source address 2931 * @dst: destination address 2932 * @head: list head to save basic blocks 2933 * 2934 * This function traverses disasm_lines from @src to @dst and save them in a 2935 * list of annotated_basic_block to @head. It uses BFS to find the shortest 2936 * path between two. The basic_block_link is to maintain parent links so 2937 * that it can build a list of blocks from the start. 2938 */ 2939 int annotate_get_basic_blocks(struct symbol *sym, s64 src, s64 dst, 2940 struct list_head *head) 2941 { 2942 struct basic_block_data bb_data = { 2943 .queue = LIST_HEAD_INIT(bb_data.queue), 2944 .visited = LIST_HEAD_INIT(bb_data.visited), 2945 }; 2946 struct basic_block_link *link; 2947 struct disasm_line *dl; 2948 int ret = -1; 2949 2950 dl = find_disasm_line(sym, src, /*allow_update=*/false); 2951 if (dl == NULL) 2952 return -1; 2953 2954 if (add_basic_block(&bb_data, /*parent=*/NULL, dl) < 0) 2955 return -1; 2956 2957 /* Find shortest path from src to dst using BFS */ 2958 while (!list_empty(&bb_data.queue)) { 2959 link = list_first_entry(&bb_data.queue, struct basic_block_link, node); 2960 2961 if (process_basic_block(&bb_data, link, sym, dst)) { 2962 link_found_basic_blocks(link, head); 2963 ret = 0; 2964 break; 2965 } 2966 list_move(&link->node, &bb_data.visited); 2967 } 2968 delete_basic_blocks(&bb_data); 2969 return ret; 2970 } 2971