| /linux/tools/testing/selftests/powerpc/pmu/ebb/ |
| H A D | lost_exception_test.c | 49 orig_period = max_period = sample_period = 400; in test_body() 51 mtspr(SPRN_PMC4, pmc_sample_period(sample_period)); in test_body() 66 if (sample_period >= (orig_period + 200)) in test_body() 67 sample_period = orig_period; in test_body() 69 sample_period++; in test_body() 71 if (sample_period > max_period) in test_body() 72 max_period = sample_period; in test_body()
|
| H A D | multi_counter_test.c | 58 mtspr(SPRN_PMC1, pmc_sample_period(sample_period)); in multi_counter() 59 mtspr(SPRN_PMC2, pmc_sample_period(sample_period)); in multi_counter() 60 mtspr(SPRN_PMC3, pmc_sample_period(sample_period)); in multi_counter() 61 mtspr(SPRN_PMC4, pmc_sample_period(sample_period)); in multi_counter() 62 mtspr(SPRN_PMC5, pmc_sample_period(sample_period)); in multi_counter() 63 mtspr(SPRN_PMC6, pmc_sample_period(sample_period)); in multi_counter()
|
| H A D | ebb.c | 31 u64 sample_period = 0x40000000ull; variable 71 bool ebb_check_count(int pmc, u64 sample_period, int fudge) in ebb_check_count() argument 77 lower = ebb_state.stats.ebb_count * (sample_period - fudge); in ebb_check_count() 85 upper = ebb_state.stats.ebb_count * (sample_period + fudge); in ebb_check_count() 119 found += count_pmc(i, sample_period); in standard_ebb_callee() 252 int count_pmc(int pmc, uint32_t sample_period) in count_pmc() argument 258 start_value = pmc_sample_period(sample_period); in count_pmc() 381 mtspr(SPRN_PMC1, pmc_sample_period(sample_period)); in ebb_child()
|
| H A D | ebb.h | 46 bool ebb_check_count(int pmc, u64 sample_period, int fudge); 58 int count_pmc(int pmc, uint32_t sample_period); 69 extern u64 sample_period;
|
| H A D | back_to_back_ebbs_test.c | 41 count_pmc(1, sample_period); in ebb_callee() 81 sample_period = 5; in back_to_back_ebbs() 84 mtspr(SPRN_PMC1, pmc_sample_period(sample_period)); in back_to_back_ebbs()
|
| H A D | instruction_count_test.c | 43 count_pmc(4, sample_period); in do_count_loop() 101 count_pmc(4, sample_period); in pmc4_ebb_callee() 125 sample_period = COUNTER_OVERFLOW; in instruction_count()
|
| H A D | no_handler_test.c | 38 sample_period = 1000; in no_handler_test() 39 mtspr(SPRN_PMC1, pmc_sample_period(sample_period)); in no_handler_test()
|
| H A D | cycles_test.c | 35 mtspr(SPRN_PMC1, pmc_sample_period(sample_period)); in cycles() 50 FAIL_IF(!ebb_check_count(1, sample_period, 100)); in cycles()
|
| H A D | pmae_handling_test.c | 42 count_pmc(1, sample_period); in syscall_ebb_callee() 77 mtspr(SPRN_PMC1, pmc_sample_period(sample_period)); in test_body()
|
| H A D | pmc56_overflow_test.c | 30 count_pmc(2, sample_period); in ebb_callee() 69 mtspr(SPRN_PMC2, pmc_sample_period(sample_period)); in pmc56_overflow()
|
| H A D | cycles_with_freeze_test.c | 48 count_pmc(1, sample_period); in ebb_callee() 74 mtspr(SPRN_PMC1, pmc_sample_period(sample_period)); in cycles_with_freeze()
|
| /linux/tools/testing/selftests/bpf/verifier/ |
| H A D | perf_event_sample_period.c | 7 offsetof(struct bpf_perf_event_data, sample_period)), 10 offsetof(struct bpf_perf_event_data, sample_period) + 7), 23 offsetof(struct bpf_perf_event_data, sample_period)), 26 offsetof(struct bpf_perf_event_data, sample_period) + 6), 39 offsetof(struct bpf_perf_event_data, sample_period)), 42 offsetof(struct bpf_perf_event_data, sample_period) + 4), 54 offsetof(struct bpf_perf_event_data, sample_period)),
|
| /linux/samples/bpf/ |
| H A D | tracex6_user.c | 104 .sample_period = SAMPLE_PERIOD, in test_bpf_perf_event() 113 .sample_period = SAMPLE_PERIOD, in test_bpf_perf_event() 122 .sample_period = SAMPLE_PERIOD, in test_bpf_perf_event() 132 .sample_period = SAMPLE_PERIOD, in test_bpf_perf_event() 144 .sample_period = SAMPLE_PERIOD, in test_bpf_perf_event() 156 .sample_period = 0, in test_bpf_perf_event()
|
| H A D | trace_event_kern.c | 50 if (ctx->sample_period < 10000) in bpf_prog1() 57 bpf_trace_printk(fmt, sizeof(fmt), cpu, ctx->sample_period, in bpf_prog1()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | verifier_unpriv_perf.c | 28 -(__s32) offsetof(struct bpf_perf_event_data, sample_period) - 8), in fill_of_different_pointers_ldx() 29 __imm_const(sample_period, in fill_of_different_pointers_ldx() 30 offsetof(struct bpf_perf_event_data, sample_period)) in fill_of_different_pointers_ldx()
|
| /linux/drivers/thermal/intel/int340x_thermal/ |
| H A D | acpi_thermal_rel.h | 49 u64 sample_period; member 64 u64 sample_period; member 110 u64 sample_period; member 121 u64 sample_period; member
|
| /linux/tools/perf/arch/arm64/util/ |
| H A D | arm-spe.c | 240 static __u64 sample_period; in arm_spe_pmu__sample_period() local 242 if (sample_period) in arm_spe_pmu__sample_period() 243 return sample_period; in arm_spe_pmu__sample_period() 250 &sample_period) != 1) { in arm_spe_pmu__sample_period() 252 sample_period = 4096; in arm_spe_pmu__sample_period() 254 return sample_period; in arm_spe_pmu__sample_period() 262 evsel->core.attr.sample_period = arm_spe_pmu__sample_period(evsel->pmu); in arm_spe_setup_evsel() 376 tracking_evsel->core.attr.sample_period = 1; in arm_spe_setup_tracking_event() 692 attr->sample_period = arm_spe_pmu__sample_period(arm_spe_pmu); in arm_spe_pmu_default_config()
|
| /linux/arch/powerpc/perf/ |
| H A D | vpa-dtl.c | 299 period = max_t(u64, NSEC_PER_MSEC, event->hw.sample_period); in vpa_dtl_hrtimer_handle() 310 period = max_t(u64, NSEC_PER_MSEC, hwc->sample_period); in vpa_dtl_start_hrtimer() 421 event->attr.sample_period = NSEC_PER_SEC / freq; in vpa_dtl_event_init() 422 hwc->sample_period = event->attr.sample_period; in vpa_dtl_event_init() 423 local64_set(&hwc->period_left, hwc->sample_period); in vpa_dtl_event_init() 424 hwc->last_period = hwc->sample_period; in vpa_dtl_event_init()
|
| /linux/drivers/media/rc/ |
| H A D | ene_ir.c | 30 static int sample_period; variable 103 if (sample_period != ENE_DEFAULT_SAMPLE_PERIOD) in ene_hw_detect() 392 if (sample_period == ENE_DEFAULT_SAMPLE_PERIOD) in ene_rx_setup() 397 (sample_period + sample_period_adjust) | in ene_rx_setup() 443 dev->rdev->rx_resolution = sample_period; in ene_rx_setup() 450 dev->rdev->min_timeout = 127 * sample_period; in ene_rx_setup() 635 dev->tx_sample = DIV_ROUND_CLOSEST(sample, sample_period); in ene_tx_sample() 644 dbg("TX: sample %8d (%s)", raw_tx * sample_period, in ene_tx_sample() 783 hw_sample = hw_value * sample_period; in ene_isr() 1025 if (sample_period < 5 || sample_period > 0x7F) in ene_probe() [all …]
|
| H A D | ite-cir.h | 126 #define ITE_BITS_TO_US(bits, sample_period) \ argument 127 ((u32)((bits) * ITE_BAUDRATE_DIVISOR * (sample_period) / 1000))
|
| /linux/kernel/ |
| H A D | watchdog_perf.c | 127 wd_attr->sample_period = hw_nmi_get_sample_period(watchdog_thresh); in hardlockup_detector_event_create() 134 wd_attr->sample_period = hw_nmi_get_sample_period(watchdog_thresh); in hardlockup_detector_event_create() 203 if (event->attr.sample_period == period) in hardlockup_detector_perf_adjust_period()
|
| H A D | watchdog.c | 369 static u64 __read_mostly sample_period; variable 472 u16 sample_period_16 = get_16bit_precision(sample_period); in update_cpustat() 499 u64 sample_period_msecond = sample_period; in print_cpustat() 654 sample_period = get_softlockup_thresh() * ((u64)NSEC_PER_SEC / NUM_SAMPLE_PERIODS); in set_sample_period() 655 watchdog_update_hrtimer_threshold(sample_period); in set_sample_period() 804 hrtimer_forward_now(hrtimer, ns_to_ktime(sample_period)); in watchdog_timer_fn() 906 hrtimer_start(hrtimer, ns_to_ktime(sample_period), in watchdog_enable()
|
| /linux/arch/x86/events/amd/ |
| H A D | ibs.c | 105 s64 period = hwc->sample_period; in perf_event_set_period() 316 if (hwc->sample_period) { in perf_ibs_init() 322 hwc->sample_period = perf_ibs->min_period; in perf_ibs_init() 325 hwc->sample_period &= ~0x0FULL; in perf_ibs_init() 326 if (hwc->sample_period < perf_ibs->min_period) in perf_ibs_init() 344 event->attr.sample_period = period; in perf_ibs_init() 345 hwc->sample_period = period; in perf_ibs_init() 347 if (hwc->sample_period < perf_ibs->min_period) in perf_ibs_init() 366 hwc->last_period = hwc->sample_period; in perf_ibs_init() 367 local64_set(&hwc->period_left, hwc->sample_period); in perf_ibs_init() [all …]
|
| /linux/arch/riscv/kvm/ |
| H A D | vcpu_pmu.c | 41 u64 sample_period; in kvm_pmu_get_sample_period() local 44 sample_period = counter_val_mask; in kvm_pmu_get_sample_period() 46 sample_period = (-pmc->counter_val) & counter_val_mask; in kvm_pmu_get_sample_period() 48 return sample_period; in kvm_pmu_get_sample_period() 308 perf_event->attr.sample_period = period; in kvm_riscv_pmu_overflow() 309 perf_event->hw.sample_period = period; in kvm_riscv_pmu_overflow() 334 attr->sample_period = kvm_pmu_get_sample_period(pmc); in kvm_pmu_create_perf_event()
|
| /linux/drivers/perf/ |
| H A D | riscv_pmu.c | 206 s64 period = hwc->sample_period; in riscv_pmu_event_set_period() 345 hwc->sample_period = cmask >> 1; in riscv_pmu_event_init() 346 hwc->last_period = hwc->sample_period; in riscv_pmu_event_init() 347 local64_set(&hwc->period_left, hwc->sample_period); in riscv_pmu_event_init()
|