/linux/arch/sh/kernel/ |
H A D | perf_event.c | 161 u64 prev_raw_count, new_raw_count; in sh_perf_event_update() local 178 prev_raw_count = local64_read(&hwc->prev_count); in sh_perf_event_update() 181 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sh_perf_event_update() 182 new_raw_count) != prev_raw_count) in sh_perf_event_update() 193 delta = (new_raw_count << shift) - (prev_raw_count << shift); in sh_perf_event_update()
|
/linux/drivers/perf/hisilicon/ |
H A D | hisi_uncore_pmu.c | 282 u64 delta, prev_raw_count, new_raw_count; in hisi_uncore_pmu_event_update() local 287 prev_raw_count = local64_read(&hwc->prev_count); in hisi_uncore_pmu_event_update() 288 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in hisi_uncore_pmu_event_update() 289 new_raw_count) != prev_raw_count); in hisi_uncore_pmu_event_update() 293 delta = (new_raw_count - prev_raw_count) & in hisi_uncore_pmu_event_update() 312 u64 prev_raw_count = local64_read(&hwc->prev_count); in hisi_uncore_pmu_start() local 314 hisi_pmu->ops->write_counter(hisi_pmu, hwc, prev_raw_count); in hisi_uncore_pmu_start()
|
/linux/arch/xtensa/kernel/ |
H A D | perf_event.c | 148 uint64_t prev_raw_count, new_raw_count; in xtensa_perf_event_update() local 152 prev_raw_count = local64_read(&hwc->prev_count); in xtensa_perf_event_update() 154 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in xtensa_perf_event_update() 155 new_raw_count) != prev_raw_count); in xtensa_perf_event_update() 157 delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK; in xtensa_perf_event_update()
|
/linux/drivers/perf/ |
H A D | riscv_pmu.c | 166 u64 prev_raw_count, new_raw_count; in riscv_pmu_event_update() local 176 prev_raw_count = local64_read(&hwc->prev_count); in riscv_pmu_event_update() 178 oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, in riscv_pmu_event_update() 180 } while (oldval != prev_raw_count); in riscv_pmu_event_update() 182 delta = (new_raw_count - prev_raw_count) & cmask; in riscv_pmu_event_update()
|
H A D | starfive_starlink_pmu.c | 237 u64 prev_raw_count, new_raw_count; in starlink_pmu_update() local 242 prev_raw_count = local64_read(&hwc->prev_count); in starlink_pmu_update() 250 oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, in starlink_pmu_update() 252 } while (oldval != prev_raw_count); in starlink_pmu_update() 254 delta = (new_raw_count - prev_raw_count) & STARLINK_PMU_COUNTER_MASK; in starlink_pmu_update()
|
H A D | arm_pmu.c | 246 u64 delta, prev_raw_count, new_raw_count; in armpmu_event_update() local 250 prev_raw_count = local64_read(&hwc->prev_count); in armpmu_event_update() 253 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in armpmu_event_update() 254 new_raw_count) != prev_raw_count) in armpmu_event_update() 257 delta = (new_raw_count - prev_raw_count) & max_period; in armpmu_event_update()
|
H A D | arm-cci.c | 973 u64 delta, prev_raw_count, new_raw_count; in cci5xx_pmu_write_counters() 976 prev_raw_count = local64_read(&hwc->prev_count); in cci5xx_pmu_write_counters() 978 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, 979 new_raw_count) != prev_raw_count); 981 delta = (new_raw_count - prev_raw_count) & CCI_PMU_CNTR_MASK; in pmu_event_update() 983 u64 delta, prev_raw_count, new_raw_count; pmu_event_update() local
|
H A D | xgene_pmu.c | 981 u64 delta, prev_raw_count, new_raw_count; in xgene_perf_event_set_period() 984 prev_raw_count = local64_read(&hw->prev_count); 987 if (local64_cmpxchg(&hw->prev_count, prev_raw_count, in xgene_perf_event_update() 988 new_raw_count) != prev_raw_count) in xgene_perf_event_update() 991 delta = (new_raw_count - prev_raw_count) & pmu_dev->max_period; in xgene_perf_event_update() 1016 u64 prev_raw_count = local64_read(&hw->prev_count); in xgene_perf_start() 1019 prev_raw_count); in xgene_perf_start() 990 u64 delta, prev_raw_count, new_raw_count; xgene_perf_event_update() local 1025 u64 prev_raw_count = local64_read(&hw->prev_count); xgene_perf_start() local
|
H A D | alibaba_uncore_drw_pmu.c | 586 unsigned long prev_raw_count = in ali_drw_pmu_start() local 588 writel(prev_raw_count, in ali_drw_pmu_start()
|
/linux/arch/arm/mach-imx/ |
H A D | mmdc.c | 311 u64 delta, prev_raw_count, new_raw_count; in mmdc_pmu_event_update() local 314 prev_raw_count = local64_read(&hwc->prev_count); in mmdc_pmu_event_update() 317 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in mmdc_pmu_event_update() 318 new_raw_count) != prev_raw_count); in mmdc_pmu_event_update() 320 delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF; in mmdc_pmu_event_update()
|
/linux/arch/alpha/kernel/ |
H A D | perf_event.c | 310 long prev_raw_count, new_raw_count; in alpha_perf_event_update() local 314 prev_raw_count = local64_read(&hwc->prev_count); in alpha_perf_event_update() 317 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in alpha_perf_event_update() 318 new_raw_count) != prev_raw_count) in alpha_perf_event_update() 321 delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf; in alpha_perf_event_update()
|
/linux/arch/loongarch/kernel/ |
H A D | perf_event.c | 340 u64 prev_raw_count, new_raw_count; in loongarch_pmu_event_update() local 343 prev_raw_count = local64_read(&hwc->prev_count); in loongarch_pmu_event_update() 346 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in loongarch_pmu_event_update() 347 new_raw_count) != prev_raw_count) in loongarch_pmu_event_update() 350 delta = new_raw_count - prev_raw_count; in loongarch_pmu_event_update()
|
/linux/arch/x86/events/amd/ |
H A D | ibs.c | 148 u64 prev_raw_count; in perf_event_try_update() local 158 prev_raw_count = local64_read(&hwc->prev_count); in perf_event_try_update() 160 &prev_raw_count, new_raw_count)) in perf_event_try_update() 171 delta = (new_raw_count << shift) - (prev_raw_count << shift); in perf_event_try_update()
|
/linux/arch/mips/kernel/ |
H A D | perf_event_mipsxx.c | 453 u64 prev_raw_count, new_raw_count; in mipspmu_event_update() local 457 prev_raw_count = local64_read(&hwc->prev_count); in mipspmu_event_update() 460 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in mipspmu_event_update() 461 new_raw_count) != prev_raw_count) in mipspmu_event_update() 464 delta = new_raw_count - prev_raw_count; in mipspmu_event_update()
|
/linux/arch/sparc/kernel/ |
H A D | perf_event.c | 867 u64 prev_raw_count, new_raw_count; in sparc_perf_event_update() local 871 prev_raw_count = local64_read(&hwc->prev_count); in sparc_perf_event_update() 874 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sparc_perf_event_update() 875 new_raw_count) != prev_raw_count) in sparc_perf_event_update() 878 delta = (new_raw_count << shift) - (prev_raw_count << shift); in sparc_perf_event_update()
|
/linux/arch/arc/kernel/ |
H A D | perf_event.c | 283 u64 prev_raw_count = local64_read(&hwc->prev_count); in arc_perf_event_update() local 285 s64 delta = new_raw_count - prev_raw_count; in arc_perf_event_update()
|
/linux/drivers/powercap/ |
H A D | intel_rapl_common.c | 1660 u64 prev_raw_count, new_raw_count; in rapl_event_update() local 1667 prev_raw_count = local64_read(&hwc->prev_count); in rapl_event_update() 1671 &prev_raw_count, new_raw_count)); in rapl_event_update() 1679 delta = new_raw_count - prev_raw_count; in rapl_event_update()
|
/linux/arch/csky/kernel/ |
H A D | perf_event.c | 922 uint64_t prev_raw_count = local64_read(&hwc->prev_count); in csky_perf_event_update() local 929 int64_t delta = new_raw_count - prev_raw_count; in csky_perf_event_update()
|
/linux/arch/x86/events/intel/ |
H A D | ds.c | 2100 u64 prev_raw_count, new_raw_count; in intel_pmu_save_and_restart_reload() local 2110 prev_raw_count = local64_read(&hwc->prev_count); in intel_pmu_save_and_restart_reload() 2142 old = ((s64)(prev_raw_count << shift) >> shift); in intel_pmu_save_and_restart_reload()
|