| /linux/fs/xfs/ |
| H A D | xfs_stats.c | 15 val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx)); in counter_val() 79 xs_xstrat_bytes += per_cpu_ptr(stats, i)->s.xs_xstrat_bytes; in xfs_stats_format() 80 xs_write_bytes += per_cpu_ptr(stats, i)->s.xs_write_bytes; in xfs_stats_format() 81 xs_read_bytes += per_cpu_ptr(stats, i)->s.xs_read_bytes; in xfs_stats_format() 82 xs_defer_relog += per_cpu_ptr(stats, i)->s.xs_defer_relog; in xfs_stats_format() 83 xs_gc_bytes += per_cpu_ptr(stats, i)->s.xs_gc_bytes; in xfs_stats_format() 112 xs_inodes_active = per_cpu_ptr(stats, c)->s.xs_inodes_active; in xfs_stats_clearall() 113 xs_inodes_meta = per_cpu_ptr(stats, c)->s.xs_inodes_meta; in xfs_stats_clearall() 114 memset(per_cpu_ptr(stats, c), 0, sizeof(*stats)); in xfs_stats_clearall() 115 per_cpu_ptr(stats, c)->s.xs_inodes_active = xs_inodes_active; in xfs_stats_clearall() [all …]
|
| H A D | xfs_stats.h | 177 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++; \ 178 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \ 183 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--; \ 184 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \ 189 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \ 190 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \ 195 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++; \ 196 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \ 201 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]; \ 202 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]; \ [all …]
|
| /linux/kernel/ |
| H A D | relay.c | 202 *per_cpu_ptr(chan->buf, buf->cpu) = NULL; in relay_destroy_buf() 332 if (chan->is_global && (buf = *per_cpu_ptr(chan->buf, 0))) { in relay_reset() 339 if ((buf = *per_cpu_ptr(chan->buf, i))) in relay_reset() 386 return *per_cpu_ptr(chan->buf, 0); in relay_open_buf() 410 *per_cpu_ptr(chan->buf, 0) = buf; in relay_open_buf() 444 if (*per_cpu_ptr(chan->buf, cpu)) in relay_prepare_cpu() 452 *per_cpu_ptr(chan->buf, cpu) = buf; in relay_prepare_cpu() 520 *per_cpu_ptr(chan->buf, i) = buf; in relay_open() 529 if ((buf = *per_cpu_ptr(chan->buf, i))) in relay_open() 632 buf = *per_cpu_ptr(chan->buf, cpu); in relay_subbufs_consumed() [all …]
|
| H A D | smpboot.c | 168 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in __smpboot_create_thread() 193 *per_cpu_ptr(ht->store, cpu) = tsk; in __smpboot_create_thread() 226 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_unpark_thread() 245 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_park_thread() 268 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_destroy_threads() 272 *per_cpu_ptr(ht->store, cpu) = NULL; in smpboot_destroy_threads()
|
| H A D | cpu.c | 173 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in cpuhp_invoke_callback() 311 atomic_t *st = per_cpu_ptr(&cpuhp_state.ap_sync_state, cpu); in cpuhp_wait_for_sync_state() 362 atomic_t *st = per_cpu_ptr(&cpuhp_state.ap_sync_state, cpu); in cpuhp_bp_sync_dead() 404 atomic_t *st = per_cpu_ptr(&cpuhp_state.ap_sync_state, cpu); in cpuhp_can_boot_ap() 800 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in bringup_wait_for_ap_online() 833 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in cpuhp_bringup_ap() 865 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in bringup_cpu() 1135 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in cpuhp_invoke_ap_callback() 1184 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in cpuhp_kick_ap_work() 1215 st = per_cpu_ptr( in cpuhp_init_state() [all...] |
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | rtrs-clt-stats.c | 27 s = per_cpu_ptr(stats->pcpu_stats, con->cpu); in rtrs_clt_update_wc_stats() 47 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_from_cnt_to_str() 66 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_to_cnt_to_str() 90 r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma; in rtrs_clt_stats_rdma_to_str() 119 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_rdma_stats() 135 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_cpu_migr_stats()
|
| /linux/kernel/irq/ |
| H A D | matrix.c | 72 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_alloc_matrix() 144 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu() 165 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu_managed() 221 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_reserve_managed() 262 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_remove_managed() 305 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc_managed() 400 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc() 427 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_free() 510 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_debug_show()
|
| /linux/arch/x86/events/amd/ |
| H A D | uncore.c | 166 struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu); in amd_uncore_start() 184 struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu); in amd_uncore_stop() 205 struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu); in amd_uncore_add() 252 struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu); in amd_uncore_del() 280 ctx = *per_cpu_ptr(pmu->ctx, event->cpu); in amd_uncore_event_init() 454 union amd_uncore_info *info = per_cpu_ptr(uncore->info, cpu); in amd_uncore_ctx_cid() 461 union amd_uncore_info *info = per_cpu_ptr(uncore->info, cpu); in amd_uncore_ctx_gid() 468 union amd_uncore_info *info = per_cpu_ptr(uncore->info, cpu); in amd_uncore_ctx_num_pmcs() 483 ctx = *per_cpu_ptr(pmu->ctx, cpu); in amd_uncore_ctx_free() 495 *per_cpu_ptr(pmu->ctx, cpu) = NULL; in amd_uncore_ctx_free() [all …]
|
| /linux/kernel/sched/ |
| H A D | cpuacct.c | 99 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_read() 100 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_read() 139 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_write() 140 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_write() 272 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_stats_show() 280 cputime.sum_exec_runtime += *per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_stats_show() 344 *per_cpu_ptr(ca->cpuusage, cpu) += cputime; in cpuacct_charge()
|
| H A D | topology.c | 926 sibling = *per_cpu_ptr(sdd->sd, i); in build_balance_mask() 987 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in init_overlap_sched_group() 1047 sibling = *per_cpu_ptr(sdd->sd, i); in build_overlap_sched_groups() 1197 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group() 1205 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group() 1206 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group() 1582 WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd); in claim_allocations() 1583 *per_cpu_ptr(sdd->sd, cpu) = NULL; in claim_allocations() 1585 if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) in claim_allocations() 1586 *per_cpu_ptr(sdd->sds, cpu) = NULL; in claim_allocations() [all …]
|
| /linux/include/linux/ |
| H A D | context_tracking_state.h | 96 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_rcu_watching_cpu() 103 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_rcu_watching_cpu_acquire() 115 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_nesting_cpu() 127 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_nmi_nesting_cpu()
|
| H A D | part_stat.h | 29 (per_cpu_ptr((part)->bd_stats, (cpu))->field) 39 res += per_cpu_ptr((part)->bd_stats, _cpu)->field; \ 48 memset(per_cpu_ptr(part->bd_stats, i), value, in part_stat_set_all()
|
| /linux/arch/s390/kernel/ |
| H A D | smp.c | 297 pcpu_sigp_retry(per_cpu_ptr(&pcpu_devices, cpu), SIGP_RESTART, 0); in pcpu_start_fn() 404 if (pcpu_running(per_cpu_ptr(&pcpu_devices, cpu))) in arch_vcpu_is_preempted() 437 struct pcpu *pcpu = per_cpu_ptr(&pcpu_devices, cpu); in smp_emergency_stop() 446 if (pcpu_stopped(per_cpu_ptr(&pcpu_devices, cpu))) in smp_emergency_stop() 477 pcpu = per_cpu_ptr(&pcpu_devices, cpu); in smp_send_stop() 518 pcpu_ec_call(per_cpu_ptr(&pcpu_devices, cpu), ec_call_function_single); in arch_send_call_function_ipi_mask() 523 pcpu_ec_call(per_cpu_ptr(&pcpu_devices, cpu), ec_call_function_single); in arch_send_call_function_single_ipi() 533 pcpu_ec_call(per_cpu_ptr(&pcpu_devices, cpu), ec_schedule); in arch_smp_send_reschedule() 551 pcpu = per_cpu_ptr(&pcpu_devices, cpu); in smp_store_status() 724 pcpu = per_cpu_ptr(&pcpu_devices, cpu); in smp_add_core() [all …]
|
| H A D | wti.c | 104 struct wti_state *st = per_cpu_ptr(&wti_state, cpu); in wti_pending() 133 st = per_cpu_ptr(&wti_state, cpu); in wti_show() 144 struct wti_state *st = per_cpu_ptr(&wti_state, cpu); in wti_thread_fn() 179 st = per_cpu_ptr(&wti_state, cpu); in wti_init()
|
| /linux/kernel/bpf/ |
| H A D | percpu_freelist.c | 15 struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_init() 58 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_push_nmi() 89 head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_push() 107 head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_populate()
|
| /linux/drivers/irqchip/ |
| H A D | irq-riscv-imsic-state.c | 186 tlocal = per_cpu_ptr(imsic->global.local, tvec->cpu); in __imsic_local_sync() 190 mlocal = per_cpu_ptr(imsic->global.local, mvec->cpu); in __imsic_local_sync() 195 mlocal = per_cpu_ptr(imsic->global.local, mvec->cpu); in __imsic_local_sync() 296 lpriv = per_cpu_ptr(imsic->lpriv, vec->cpu); in imsic_vector_mask() 318 lpriv = per_cpu_ptr(imsic->lpriv, vec->cpu); in imsic_vector_unmask() 342 lpriv = per_cpu_ptr(imsic->lpriv, vec->cpu); in imsic_vector_force_move_cleanup() 387 old_lpriv = per_cpu_ptr(imsic->lpriv, old_vec->cpu); in imsic_vector_move() 391 new_lpriv = per_cpu_ptr(imsic->lpriv, new_vec->cpu); in imsic_vector_move() 412 lpriv = per_cpu_ptr(imsic->lpriv, vec->cpu); in imsic_vector_debug_show() 451 lpriv = per_cpu_ptr(imsic->lpriv, cpu); in imsic_vector_alloc() [all …]
|
| H A D | irq-loongarch-avec.c | 85 struct pending_list *plist = per_cpu_ptr(&pending_list, cpu); in pending_list_init() 95 plist = per_cpu_ptr(&pending_list, adata->prev_cpu); in avecintc_sync() 125 per_cpu_ptr(irq_map, adata->cpu)[adata->vec] = irq_data_to_desc(data); in avecintc_set_affinity() 152 struct pending_list *plist = per_cpu_ptr(&pending_list, cpu); in avecintc_cpu_offline() 269 per_cpu_ptr(irq_map, adata->cpu)[adata->vec] = irq_data_to_desc(irqd); in avecintc_alloc_vector()
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | aperfmperf.c | 388 per_cpu_ptr(arch_cpu_scale, cpu)->capacity = SCHED_CAPACITY_SCALE; in arch_enable_hybrid_capacity_scale() 389 per_cpu_ptr(arch_cpu_scale, cpu)->freq_ratio = arch_max_freq_ratio; in arch_enable_hybrid_capacity_scale() 419 WRITE_ONCE(per_cpu_ptr(arch_cpu_scale, cpu)->capacity, in arch_set_cpu_capacity() 421 WRITE_ONCE(per_cpu_ptr(arch_cpu_scale, cpu)->freq_ratio, in arch_set_cpu_capacity() 431 return READ_ONCE(per_cpu_ptr(arch_cpu_scale, cpu)->capacity); in arch_scale_cpu_capacity() 508 struct aperfmperf *s = per_cpu_ptr(&cpu_samples, cpu); in arch_freq_get_on_cpu()
|
| /linux/fs/squashfs/ |
| H A D | decompressor_multi_percpu.c | 39 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create() 53 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create() 70 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_destroy()
|
| /linux/drivers/platform/x86/amd/hfi/ |
| H A D | hfi.c | 184 info = per_cpu_ptr(&amd_hfi_cpuinfo, cpu_index); in amd_hfi_fill_metadata() 188 info = per_cpu_ptr(&amd_hfi_cpuinfo, cpu_index); in amd_hfi_fill_metadata() 232 hfi_cpuinfo = per_cpu_ptr(&amd_hfi_cpuinfo, idx); in amd_hfi_alloc_class_data() 278 struct amd_hfi_cpuinfo *hfi_info = per_cpu_ptr(&amd_hfi_cpuinfo, cpu); in amd_hfi_online() 333 struct amd_hfi_cpuinfo *hfi_cpuinfo = per_cpu_ptr(&amd_hfi_cpuinfo, cpu); in update_hfi_ipcc_scores() 408 struct amd_hfi_cpuinfo *hfi_cpuinfo = per_cpu_ptr(&amd_hfi_cpuinfo, cpu); in class_capabilities_show()
|
| /linux/drivers/clocksource/ |
| H A D | timer-mp-csky.c | 78 struct timer_of *to = per_cpu_ptr(&csky_to, cpu); in csky_mptimer_starting_cpu() 145 to = per_cpu_ptr(&csky_to, cpu); in csky_mptimer_init() 168 to = per_cpu_ptr(&csky_to, cpu_rollback); in csky_mptimer_init()
|
| /linux/tools/testing/shared/linux/ |
| H A D | percpu.h | 10 #define per_cpu_ptr(ptr, cpu) ({ (void)(cpu); (ptr); }) macro 11 #define per_cpu(var, cpu) (*per_cpu_ptr(&(var), cpu))
|
| /linux/drivers/hv/ |
| H A D | hv.c | 182 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc() 193 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc() 250 per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_free() 279 per_cpu_ptr(hv_context.cpu_context, cpu); in hv_hyp_synic_enable_regs() 350 = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_para_synic_enable_regs() 407 per_cpu_ptr(hv_context.cpu_context, cpu); in hv_hyp_synic_disable_regs()
|
| /linux/arch/x86/kernel/ |
| H A D | kgdb.c | 208 bp = *per_cpu_ptr(breakinfo[breakno].pev, cpu); in kgdb_correct_hw_break() 237 pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); in hw_break_reserve_slot() 249 pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); in hw_break_reserve_slot() 264 pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); in hw_break_release_slot() 304 bp = *per_cpu_ptr(breakinfo[i].pev, cpu); in kgdb_remove_all_hw_break() 397 bp = *per_cpu_ptr(breakinfo[i].pev, cpu); in kgdb_disable_hw_debug() 666 pevent = per_cpu_ptr(breakinfo[i].pev, cpu); in kgdb_arch_late()
|
| /linux/arch/x86/kernel/acpi/ |
| H A D | cstate.c | 183 percpu_entry = per_cpu_ptr(cpu_cstate_entry, cpu); in acpi_processor_ffh_cstate_probe() 214 percpu_entry = per_cpu_ptr(cpu_cstate_entry, cpu); in acpi_processor_ffh_play_dead() 224 percpu_entry = per_cpu_ptr(cpu_cstate_entry, cpu); in acpi_processor_ffh_cstate_enter()
|