| /linux/tools/perf/util/ |
| H A D | counts.c | 12 struct perf_counts *counts = zalloc(sizeof(*counts)); in perf_counts__new() local 14 if (counts) { in perf_counts__new() 19 free(counts); in perf_counts__new() 23 counts->values = values; in perf_counts__new() 27 xyarray__delete(counts->values); in perf_counts__new() 28 free(counts); in perf_counts__new() 32 counts->loaded = values; in perf_counts__new() 35 return counts; in perf_counts__new() 38 void perf_counts__delete(struct perf_counts *counts) in perf_counts__delete() argument 40 if (counts) { in perf_counts__delete() [all …]
|
| H A D | counts.h | 20 perf_counts(struct perf_counts *counts, int cpu_map_idx, int thread) in perf_counts() argument 22 return xyarray__entry(counts->values, cpu_map_idx, thread); in perf_counts() 26 perf_counts__is_loaded(struct perf_counts *counts, int cpu_map_idx, int thread) in perf_counts__is_loaded() argument 28 return *((bool *) xyarray__entry(counts->loaded, cpu_map_idx, thread)); in perf_counts__is_loaded() 32 perf_counts__set_loaded(struct perf_counts *counts, int cpu_map_idx, int thread, bool loaded) in perf_counts__set_loaded() argument 34 *((bool *) xyarray__entry(counts->loaded, cpu_map_idx, thread)) = loaded; in perf_counts__set_loaded() 38 void perf_counts__delete(struct perf_counts *counts); 39 void perf_counts__reset(struct perf_counts *counts);
|
| H A D | stat.c | 158 struct perf_counts *counts; in evsel__alloc_prev_raw_counts() local 160 counts = perf_counts__new(cpu_map_nr, nthreads); in evsel__alloc_prev_raw_counts() 161 if (counts) in evsel__alloc_prev_raw_counts() 162 evsel->prev_raw_counts = counts; in evsel__alloc_prev_raw_counts() 164 return counts ? 0 : -ENOMEM; in evsel__alloc_prev_raw_counts() 253 *perf_counts(evsel->counts, idx, thread) = in evsel__copy_prev_raw_counts() 275 *ps->aggr[0].counts.values = avg_stats(&ps->res_stats); in evsel__copy_res_stats() 371 if (evsel->err || evsel->counts->scaled == -1) in evsel__count_has_error() 407 struct perf_counts_values *aggr_counts = &ps->aggr[thread].counts; in process_counter_values() 443 ps_aggr->counts.val = 0; in process_counter_values() [all …]
|
| H A D | bpf_counter.c | 294 struct perf_counts_values *counts; in bpf_program_profiler__read() local 303 counts = perf_counts(evsel->counts, idx, 0); in bpf_program_profiler__read() 304 counts->val = 0; in bpf_program_profiler__read() 305 counts->ena = 0; in bpf_program_profiler__read() 306 counts->run = 0; in bpf_program_profiler__read() 325 counts = perf_counts(evsel->counts, idx, 0); in bpf_program_profiler__read() 326 counts->val += values[bpf_cpu].counter; in bpf_program_profiler__read() 327 counts->ena += values[bpf_cpu].enabled; in bpf_program_profiler__read() 328 counts->run += values[bpf_cpu].running; in bpf_program_profiler__read() 696 struct perf_counts_values *counts; in bperf__read() local [all …]
|
| H A D | branch.c | 27 st->counts[flags->type]++; in branch_type_count() 118 total += st->counts[i]; in branch_type_stat_display() 152 if (st->counts[i] > 0) in branch_type_stat_display() 156 (double)st->counts[i] / (double)total); in branch_type_stat_display() 180 total += st->counts[i]; in branch_type_str() 198 if (st->counts[i] > 0) in branch_type_str()
|
| H A D | bpf_counter_cgroup.c | 272 struct perf_counts_values *counts; in bperf_cgrp__read() local 300 counts = perf_counts(evsel->counts, i, 0); in bperf_cgrp__read() 301 counts->val = values[cpu.cpu].counter; in bperf_cgrp__read() 302 counts->ena = values[cpu.cpu].enabled; in bperf_cgrp__read() 303 counts->run = values[cpu.cpu].running; in bperf_cgrp__read()
|
| /linux/tools/lib/perf/tests/ |
| H A D | test-evsel.c | 40 struct perf_counts_values counts = { .val = 0 }; in test_stat_cpu() local 42 perf_evsel__read(evsel, idx, 0, &counts); in test_stat_cpu() 43 __T("failed to read value for evsel", counts.val != 0); in test_stat_cpu() 55 struct perf_counts_values counts = { .val = 0 }; in test_stat_thread() local 75 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread() 76 __T("failed to read value for evsel", counts.val != 0); in test_stat_thread() 87 struct perf_counts_values counts = { .val = 0 }; in test_stat_thread_enable() local 108 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread_enable() 109 __T("failed to read value for evsel", counts.val == 0); in test_stat_thread_enable() 114 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread_enable() [all …]
|
| H A D | test-evlist.c | 78 struct perf_counts_values counts = { .val = 0 }; in test_stat_cpu() local 80 perf_evsel__read(evsel, idx, 0, &counts); in test_stat_cpu() 81 __T("failed to read value for evsel", counts.val != 0); in test_stat_cpu() 94 struct perf_counts_values counts = { .val = 0 }; in test_stat_thread() local 136 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread() 137 __T("failed to read value for evsel", counts.val != 0); in test_stat_thread() 149 struct perf_counts_values counts = { .val = 0 }; in test_stat_thread_enable() local 193 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread_enable() 194 __T("failed to read value for evsel", counts.val == 0); in test_stat_thread_enable() 200 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread_enable() [all …]
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | sock_iter_batch.c | 25 static int insert(__u64 cookie, struct sock_count counts[], int counts_len) in insert() argument 31 if (!counts[i].cookie) { in insert() 33 } else if (counts[i].cookie == cookie) { in insert() 41 counts[insert].cookie = cookie; in insert() 42 counts[insert].count++; in insert() 44 return counts[insert].count; in insert() 47 static int read_n(int iter_fd, int n, struct sock_count counts[], in read_n() argument 58 ASSERT_GE(insert(out.cookie, counts, counts_len), 0, "insert"); in read_n() 77 static bool was_seen(int fd, struct sock_count counts[], int counts_len) in was_seen() argument 83 if (cookie == counts[i].cookie) in was_seen() [all …]
|
| /linux/arch/x86/kernel/cpu/resctrl/ |
| H A D | pseudo_lock.c | 310 struct residency_counts *counts) in measure_residency_fn() argument 418 counts->miss_before = miss_before; in measure_residency_fn() 419 counts->hits_before = hits_before; in measure_residency_fn() 420 counts->miss_after = miss_after; in measure_residency_fn() 421 counts->hits_after = hits_after; in measure_residency_fn() 428 struct residency_counts counts = {0}; in resctrl_arch_measure_l2_residency() local 450 measure_residency_fn(&perf_miss_attr, &perf_hit_attr, plr, &counts); in resctrl_arch_measure_l2_residency() 455 trace_pseudo_lock_l2(counts.hits_after - counts.hits_before, in resctrl_arch_measure_l2_residency() 456 counts.miss_after - counts.miss_before); in resctrl_arch_measure_l2_residency() 466 struct residency_counts counts = {0}; in resctrl_arch_measure_l3_residency() local [all …]
|
| /linux/Documentation/admin-guide/perf/ |
| H A D | fujitsu_uncore_pmu.rst | 32 This event counts MAC cycles at MAC frequency. 34 This event counts the number of read requests to MAC. 36 This event counts the number of read requests including retry to MAC. 38 This event counts the number of responses to read requests to MAC. 40 This event counts the number of read requests including retry with PFTGT 43 This event counts the number of read requests including retry without PFTGT 46 This event counts the number of responses to read requests which hit the 49 This event counts the number of responses to read requests which miss the 52 This event counts outstanding read requests issued by DDR memory controller 55 This event counts the number of write requests to MAC (including zero write, [all …]
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | msi.c | 230 struct msi_counts *counts = data; in count_non_bridge_devices() 240 counts->num_devices++; in count_spare_msis() 247 struct msi_counts *counts = data; in count_spare_msis() 251 if (dn == counts->requestor) in count_spare_msis() 252 req = counts->request; in count_spare_msis() 266 if (req < counts->quota) in count_spare_msis() 267 counts->spare += counts->quota - req; in count_spare_msis() 268 else if (req > counts->quota) 269 counts 225 struct msi_counts *counts = data; count_non_bridge_devices() local 242 struct msi_counts *counts = data; count_spare_msis() local 272 struct msi_counts counts; msi_quota_for_device() local [all...] |
| /linux/samples/bpf/ |
| H A D | sampleip_user.c | 82 struct ipcount counts[MAX_IPS]; variable 102 counts[i].ip = next_key; in print_ip_map() 103 counts[i++].count = value; in print_ip_map() 109 qsort(counts, max, sizeof(struct ipcount), count_cmp); in print_ip_map() 111 if (counts[i].ip > _text_addr) { in print_ip_map() 112 sym = ksym_search(counts[i].ip); in print_ip_map() 118 printf("0x%-17llx %-32s %u\n", counts[i].ip, sym->name, in print_ip_map() 119 counts[i].count); in print_ip_map() 121 printf("0x%-17llx %-32s %u\n", counts[i].ip, "(user)", in print_ip_map() 122 counts[i].count); in print_ip_map()
|
| H A D | offwaketime.bpf.c | 32 } counts SEC(".maps"); 94 val = bpf_map_lookup_elem(&counts, &key); in update_counts() 96 bpf_map_update_elem(&counts, &key, &zero, BPF_NOEXIST); in update_counts() 97 val = bpf_map_lookup_elem(&counts, &key); in update_counts()
|
| H A D | trace_event_kern.c | 25 } counts SEC(".maps"); 71 val = bpf_map_lookup_elem(&counts, &key); in bpf_prog1() 75 bpf_map_update_elem(&counts, &key, &one, BPF_NOEXIST); in bpf_prog1()
|
| /linux/drivers/media/v4l2-core/ |
| H A D | v4l2-vp9.c | 1674 const struct v4l2_vp9_frame_symbol_counts *counts, in _adapt_coeff() argument 1683 *counts->eob[i][j][k][l][m][1], in _adapt_coeff() 1684 *counts->eob[i][j][k][l][m][0] - *counts->eob[i][j][k][l][m][1], in _adapt_coeff() 1687 adapt_probs_variant_a_coef(p, *counts->coeff[i][j][k][l][m], uf); in _adapt_coeff() 1694 const struct v4l2_vp9_frame_symbol_counts *counts, in _adapt_coef_probs() argument 1702 _adapt_coeff(i, j, k, probs, counts, uf); in _adapt_coef_probs() 1706 struct v4l2_vp9_frame_symbol_counts *counts, in v4l2_vp9_adapt_coef_probs() argument 1711 _adapt_coef_probs(probs, counts, 112); in v4l2_vp9_adapt_coef_probs() 1714 _adapt_coef_probs(probs, counts, 128); in v4l2_vp9_adapt_coef_probs() 1716 _adapt_coef_probs(probs, counts, 112); in v4l2_vp9_adapt_coef_probs() [all …]
|
| /linux/drivers/hv/ |
| H A D | hv_proc.c | 22 int *counts; in hv_call_deposit_pages() local 43 counts = kcalloc(HV_DEPOSIT_MAX, sizeof(int), GFP_KERNEL); in hv_call_deposit_pages() 44 if (!counts) { in hv_call_deposit_pages() 69 counts[i] = 1 << order; in hv_call_deposit_pages() 70 num_pages -= counts[i]; in hv_call_deposit_pages() 84 for (j = 0; j < counts[i]; ++j, ++page_count) in hv_call_deposit_pages() 102 for (j = 0; j < counts[i]; ++j) in hv_call_deposit_pages() 108 kfree(counts); in hv_call_deposit_pages()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_btf_nokv.c | 21 struct ipv_counts *counts; in test_long_fname_2() local 24 counts = bpf_map_lookup_elem(&btf_map, &key); in test_long_fname_2() 25 if (!counts) in test_long_fname_2() 28 counts->v6++; in test_long_fname_2()
|
| H A D | test_btf_newkv.c | 22 struct ipv_counts *counts; in test_long_fname_2() local 25 counts = bpf_map_lookup_elem(&btf_map, &key); in test_long_fname_2() 26 if (!counts) in test_long_fname_2() 29 counts->v6++; in test_long_fname_2()
|
| /linux/tools/lib/perf/Documentation/examples/ |
| H A D | counting.c | 24 struct perf_counts_values counts; in main() local 73 perf_evsel__read(evsel, 0, 0, &counts); in main() 75 counts.val, counts.ena, counts.run); in main()
|
| /linux/drivers/scsi/ |
| H A D | ch.c | 117 u_int counts[CH_TYPES]; member 235 ch->counts[i]) in ch_elem_to_typecode() 323 ch->counts[CHET_MT] = in ch_readconfig() 327 ch->counts[CHET_ST] = in ch_readconfig() 331 ch->counts[CHET_IE] = in ch_readconfig() 335 ch->counts[CHET_DT] = in ch_readconfig() 339 ch->counts[CHET_MT]); in ch_readconfig() 342 ch->counts[CHET_ST]); in ch_readconfig() 345 ch->counts[CHET_IE]); in ch_readconfig() 348 ch->counts[CHET_D in ch_readconfig() [all...] |
| /linux/tools/testing/kunit/ |
| H A D | kunit_parser.py | 35 counts : TestCounts - counts of the test statuses and errors of 46 self.counts = TestCounts() 51 f'{self.subtests}, {self.log}, {self.counts})') 59 self.counts.errors += 1 78 Tracks the counts of statuses of all test cases and any errors within 102 def add_subtest_counts(self, counts: TestCounts) -> None: 104 Adds the counts of another TestCounts object to the current 105 TestCounts object. Used to add the counts of a subtest to the 109 counts [all...] |
| H A D | kunit_tool_test.py | 147 self.assertEqual(result.counts.errors, 0) 154 self.assertEqual(result.counts.errors, 0) 161 self.assertEqual(result.counts.errors, 0) 168 self.assertEqual(result.counts.errors, 0) 175 self.assertEqual(result.counts.failed, 2) 188 self.assertEqual(result.counts.errors, 1) 198 self.assertEqual(result.counts, kunit_parser.TestCounts(passed=10, errors=0)) 208 self.assertEqual(result.counts.errors, 1) 219 self.assertEqual(result.counts, kunit_parser.TestCounts(passed=1, errors=1)) 231 self.assertEqual(result.counts [all...] |
| /linux/drivers/gpu/drm/amd/display/dc/link/accessories/ |
| H A D | link_dp_trace.c | 64 link->dp_trace.detect_lt_trace.counts.fail = fail_count; in dp_trace_lt_fail_count_update() 66 link->dp_trace.commit_lt_trace.counts.fail = fail_count; in dp_trace_lt_fail_count_update() 73 link->dp_trace.detect_lt_trace.counts.total++; in dp_trace_lt_total_count_increment() 75 link->dp_trace.commit_lt_trace.counts.total++; in dp_trace_lt_total_count_increment() 137 return &link->dp_trace.detect_lt_trace.counts; in dp_trace_get_lt_counts() 139 return &link->dp_trace.commit_lt_trace.counts; in dp_trace_get_lt_counts()
|
| /linux/tools/lib/perf/Documentation/ |
| H A D | libperf-counting.txt | 37 * reads and displays event counts 171 Now we need to get the counts from events, following code iterates through the 172 events list and read counts: 177 82 perf_evsel__read(evsel, 0, 0, &counts); 179 84 counts.val, counts.ena, counts.run);
|