| /linux/drivers/pps/generators/ |
| H A D | pps_gen_tio.c | 47 u32 prev_count; member 86 tio->prev_count = 0; in pps_tio_disable() 144 if (tio->prev_count && tio->prev_count == event_count) in hrtimer_callback() 146 tio->prev_count = event_count; in hrtimer_callback()
|
| /linux/arch/arm/mm/ |
| H A D | cache-l2x0-pmu.c | 120 u64 prev_count, new_count, mask; in l2x0_pmu_event_read() local 123 prev_count = local64_read(&hw->prev_count); in l2x0_pmu_event_read() 125 } while (local64_xchg(&hw->prev_count, new_count) != prev_count); in l2x0_pmu_event_read() 128 local64_add((new_count - prev_count) & mask, &event->count); in l2x0_pmu_event_read() 147 local64_set(&hw->prev_count, 0); in l2x0_pmu_event_configure()
|
| /linux/drivers/perf/ |
| H A D | arm_dmc620_pmu.c | 327 u64 delta, prev_count, new_count; in dmc620_pmu_event_update() local 331 prev_count = local64_read(&hwc->prev_count); in dmc620_pmu_event_update() 333 } while (local64_cmpxchg(&hwc->prev_count, in dmc620_pmu_event_update() 334 prev_count, new_count) != prev_count); in dmc620_pmu_event_update() 335 delta = (new_count - prev_count) & DMC620_CNT_MAX_PERIOD; in dmc620_pmu_event_update() 343 local64_set(&event->hw.prev_count, DMC620_CNT_INIT); in dmc620_pmu_event_set_period()
|
| H A D | riscv_pmu.c | 174 prev_raw_count = local64_read(&hwc->prev_count); in riscv_pmu_event_update() 176 oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, in riscv_pmu_event_update() 233 local64_set(&hwc->prev_count, (u64)-left); in riscv_pmu_event_set_period() 252 init_val = local64_read(&hwc->prev_count) & max_period; in riscv_pmu_start()
|
| H A D | alibaba_uncore_drw_pmu.c | 303 prev = local64_read(&hwc->prev_count); in ali_drw_pmu_event_update() 305 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in ali_drw_pmu_event_update() 328 local64_set(&event->hw.prev_count, pre_val); in ali_drw_pmu_event_set_period() 587 local64_read(&event->hw.prev_count); in ali_drw_pmu_start()
|
| H A D | fujitsu_uncore_pmu.c | 76 local64_set(&event->hw.prev_count, 0); in fujitsu_uncore_counter_start() 109 prev = local64_read(&event->hw.prev_count); in fujitsu_uncore_counter_update() 111 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in fujitsu_uncore_counter_update()
|
| H A D | arm-ccn.c | 863 u64 prev_count, new_count, mask; in arm_ccn_pmu_event_update() local 866 prev_count = local64_read(&hw->prev_count); in arm_ccn_pmu_event_update() 868 } while (local64_xchg(&hw->prev_count, new_count) != prev_count); in arm_ccn_pmu_event_update() 872 local64_add((new_count - prev_count) & mask, &event->count); in arm_ccn_pmu_event_update() 913 local64_set(&event->hw.prev_count, in arm_ccn_pmu_event_start()
|
| H A D | arm_smmuv3_pmu.c | 262 prev = local64_read(&hwc->prev_count); in smmu_pmu_event_update() 264 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in smmu_pmu_event_update() 299 local64_set(&hwc->prev_count, new); in smmu_pmu_set_period() 500 local64_set(&hwc->prev_count, 0); in smmu_pmu_event_add()
|
| H A D | arm_pmu.c | 237 local64_set(&hwc->prev_count, (u64)-left); in armpmu_event_set_period() 254 prev_raw_count = local64_read(&hwc->prev_count); in armpmu_event_update() 257 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in armpmu_event_update()
|
| H A D | dwc_pcie_pmu.c | 371 prev = local64_read(&hwc->prev_count); in dwc_pcie_pmu_event_update() 374 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in dwc_pcie_pmu_event_update() 479 local64_set(&hwc->prev_count, 0); in dwc_pcie_pmu_event_start()
|
| /linux/fs/gfs2/ |
| H A D | file.c | 781 size_t *prev_count, in should_fault_in_pages() argument 798 if (*prev_count != count) { in should_fault_in_pages() 806 *prev_count = count; in should_fault_in_pages() 816 size_t prev_count = 0, window_size = 0; in gfs2_file_direct_read() local 857 if (should_fault_in_pages(to, iocb, &prev_count, &window_size)) { in gfs2_file_direct_read() 880 size_t prev_count = 0, window_size = 0; in gfs2_file_direct_write() local 926 enough_retries = prev_count == iov_iter_count(from) && in gfs2_file_direct_write() 928 if (should_fault_in_pages(from, iocb, &prev_count, &window_size)) { in gfs2_file_direct_write() 953 size_t prev_count = 0, window_size = 0; in gfs2_file_read_iter() local 996 if (should_fault_in_pages(to, iocb, &prev_count, &window_size)) { in gfs2_file_read_iter() [all …]
|
| /linux/drivers/iommu/intel/ |
| H A D | perfmon.c | 305 u64 prev_count, new_count, delta; in iommu_pmu_event_update() local 309 prev_count = local64_read(&hwc->prev_count); in iommu_pmu_event_update() 311 if (local64_xchg(&hwc->prev_count, new_count) != prev_count) in iommu_pmu_event_update() 318 delta = (new_count << shift) - (prev_count << shift); in iommu_pmu_event_update() 344 local64_set((&hwc->prev_count), count); in iommu_pmu_start()
|
| /linux/kernel/ |
| H A D | softirq.c | 612 int prev_count; in handle_softirqs() local 617 prev_count = preempt_count(); in handle_softirqs() 624 if (unlikely(prev_count != preempt_count())) { in handle_softirqs() 627 prev_count, preempt_count()); in handle_softirqs() 628 preempt_count_set(prev_count); in handle_softirqs()
|
| /linux/drivers/perf/hisilicon/ |
| H A D | hisi_uncore_pmu.c | 309 local64_set(&hwc->prev_count, val); in hisi_uncore_pmu_set_event_period() 324 prev_raw_count = local64_read(&hwc->prev_count); in hisi_uncore_pmu_event_update() 325 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in hisi_uncore_pmu_event_update() 349 u64 prev_raw_count = local64_read(&hwc->prev_count); in hisi_uncore_pmu_start()
|
| H A D | hisi_pcie_pmu.c | 452 prev_cnt = local64_read(&hwc->prev_count); in hisi_pcie_pmu_event_update() 454 } while (local64_cmpxchg(&hwc->prev_count, prev_cnt, in hisi_pcie_pmu_event_update() 475 local64_set(&hwc->prev_count, HISI_PCIE_INIT_VAL); in hisi_pcie_pmu_set_period() 488 local64_set(&hwc->prev_count, cnt); in hisi_pcie_pmu_set_period() 550 prev_cnt = local64_read(&hwc->prev_count); in hisi_pcie_pmu_start()
|
| /linux/arch/arm/mach-imx/ |
| H A D | mmdc.c | 314 prev_raw_count = local64_read(&hwc->prev_count); in mmdc_pmu_event_update() 317 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in mmdc_pmu_event_update() 342 local64_set(&hwc->prev_count, 0); in mmdc_pmu_event_start() 377 local64_set(&hwc->prev_count, mmdc_pmu_read_counter(pmu_mmdc, cfg)); in mmdc_pmu_event_add()
|
| /linux/arch/x86/events/amd/ |
| H A D | uncore.c | 146 prev = local64_read(&hwc->prev_count); in amd_uncore_read() 157 local64_set(&hwc->prev_count, new); in amd_uncore_read() 173 wrmsrq(hwc->event_base, (u64)local64_read(&hwc->prev_count)); in amd_uncore_start() 948 wrmsrq(hwc->event_base, (u64)local64_read(&hwc->prev_count)); in amd_uncore_umc_start() 963 prev = local64_read(&hwc->prev_count); in amd_uncore_umc_read() 979 local64_set(&hwc->prev_count, 0); in amd_uncore_umc_read() 981 local64_set(&hwc->prev_count, new); in amd_uncore_umc_read()
|
| /linux/arch/sh/kernel/ |
| H A D | perf_event.c | 178 prev_raw_count = local64_read(&hwc->prev_count); in sh_perf_event_update() 181 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sh_perf_event_update()
|
| /linux/drivers/misc/ |
| H A D | ntsync.c | 411 __u32 prev_count; in ntsync_sem_release() local 424 prev_count = sem->u.sem.count; in ntsync_sem_release() 434 if (!ret && put_user(prev_count, user_args)) in ntsync_sem_release() 461 __u32 prev_count; in ntsync_mutex_unlock() local 475 prev_count = mutex->u.mutex.count; in ntsync_mutex_unlock() 485 if (!ret && put_user(prev_count, &user_args->count)) in ntsync_mutex_unlock()
|
| /linux/drivers/dma/idxd/ |
| H A D | perfmon.c | 217 prev_raw_count = local64_read(&hwc->prev_count); in perfmon_pmu_event_update() 220 } while (!local64_try_cmpxchg(&hwc->prev_count, in perfmon_pmu_event_update() 330 local64_set(&event->hw.prev_count, cntrdata); in perfmon_pmu_event_start()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_pmu.c | 321 prev = local64_read(&hwc->prev_count); in xe_pmu_event_update() 324 } while (!local64_try_cmpxchg(&hwc->prev_count, &prev, new)); in xe_pmu_event_update() 356 local64_set(&event->hw.prev_count, __xe_pmu_event_read(event)); in xe_pmu_enable()
|
| /linux/arch/x86/events/intel/ |
| H A D | uncore.c | 277 u64 prev_count, new_count, delta; in uncore_perf_event_update() local 289 prev_count = local64_read(&event->hw.prev_count); in uncore_perf_event_update() 291 if (local64_xchg(&event->hw.prev_count, new_count) != prev_count) in uncore_perf_event_update() 294 delta = (new_count << shift) - (prev_count << shift); in uncore_perf_event_update() 519 local64_set(&event->hw.prev_count, in uncore_pmu_event_start() 534 local64_set(&event->hw.prev_count, uncore_read_counter(box, event)); in uncore_pmu_event_start()
|
| /linux/arch/loongarch/kernel/ |
| H A D | perf_event.c | 326 local64_set(&hwc->prev_count, loongarch_pmu.overflow - left); in loongarch_pmu_event_set_period() 343 prev_raw_count = local64_read(&hwc->prev_count); in loongarch_pmu_event_update() 346 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in loongarch_pmu_event_update()
|
| /linux/arch/x86/events/ |
| H A D | rapl.c | 223 prev_raw_count = local64_read(&hwc->prev_count); in rapl_event_update() 226 } while (!local64_try_cmpxchg(&hwc->prev_count, in rapl_event_update() 291 local64_set(&event->hw.prev_count, rapl_read_counter(event)); in __rapl_pmu_event_start()
|
| /linux/arch/s390/kernel/ |
| H A D | perf_pai.c | 426 prev = local64_read(&event->hw.prev_count); in pai_read() 428 local64_set(&event->hw.prev_count, new); in pai_read() 449 local64_set(&event->hw.prev_count, sum); in pai_start()
|