| /linux/arch/powerpc/include/asm/ |
| H A D | idle.h | 14 *this_cpu_ptr(&idle_entry_purr_snap) = mfspr(SPRN_PURR); in snapshot_purr_idle_entry() 19 *this_cpu_ptr(&idle_entry_spurr_snap) = mfspr(SPRN_SPURR); in snapshot_spurr_idle_entry() 25 u64 in_purr = *this_cpu_ptr(&idle_entry_purr_snap); in update_idle_purr_accounting() 34 u64 *idle_spurr_cycles_ptr = this_cpu_ptr(&idle_spurr_cycles); in update_idle_spurr_accounting() 35 u64 in_spurr = *this_cpu_ptr(&idle_entry_spurr_snap); in update_idle_spurr_accounting() 89 return *this_cpu_ptr(&idle_spurr_cycles); in read_this_idle_spurr()
|
| /linux/arch/x86/kernel/cpu/mce/ |
| H A D | intel.c | 166 machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); in intel_threshold_interrupt() 178 unsigned long *owned = (void *)this_cpu_ptr(&mce_banks_owned); in cmci_skip_bank() 192 __clear_bit(bank, this_cpu_ptr(mce_poll_banks)); in cmci_skip_bank() 232 struct mca_storm_desc *storm = this_cpu_ptr(&storm_desc); in cmci_claim_bank() 240 WARN_ON(!test_bit(bank, this_cpu_ptr(mce_poll_banks))); in cmci_claim_bank() 246 set_bit(bank, (void *)this_cpu_ptr(&mce_banks_owned)); in cmci_claim_bank() 253 __clear_bit(bank, this_cpu_ptr(mce_poll_banks)); in cmci_claim_bank() 316 machine_check_poll(0, this_cpu_ptr(&mce_banks_owned)); in cmci_recheck() 325 if (!test_bit(bank, this_cpu_ptr(mce_banks_owned))) in __cmci_disable_bank() 330 __clear_bit(bank, this_cpu_ptr(mce_banks_owned)); in __cmci_disable_bank() [all …]
|
| H A D | amd.c | 279 struct mce_amd_cpu_data *data = this_cpu_ptr(&mce_amd_data); in smca_configure() 280 u8 *bank_counts = this_cpu_ptr(smca_bank_counts); in smca_configure() 332 this_cpu_ptr(mce_banks_array)[bank].lsb_in_status = !!(low & BIT(8)); in smca_configure() 335 this_cpu_ptr(smca_banks)[bank].paddrv = 1; in smca_configure() 352 this_cpu_ptr(smca_banks)[bank].hwid = s_hwid; in smca_configure() 353 this_cpu_ptr(smca_banks)[bank].id = low; in smca_configure() 354 this_cpu_ptr(smca_banks)[bank].sysfs_id = bank_counts[s_hwid->bank_type]++; in smca_configure() 587 __set_bit(bank, this_cpu_ptr(&mce_amd_data)->thr_intr_banks); in prepare_threshold_block() 679 struct mce_bank *mce_banks = this_cpu_ptr(mce_banks_array); in amd_apply_cpu_quirks() 707 struct mce_amd_cpu_data *data = this_cpu_ptr(&mce_amd_data); in smca_enable_interrupt_vectors() [all …]
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | this_cpu_ops.rst | 118 y = this_cpu_ptr(&x) 121 ``this_cpu_ptr`` 避免了通用 ``get_cpu``/``put_cpu`` 序列所需的多个步骤。没有可用 148 &x和p是每CPU变量的偏移量。 ``this_cpu_ptr()`` 使用每CPU变量的偏移量,这让它看起来 181 ``this_cpu_ptr()``:: 185 pp = this_cpu_ptr(&p); 225 &this_cpu_ptr(pp)->n 对比 this_cpu_ptr(&pp->n) 259 struct data *p = this_cpu_ptr(&datap); 261 但是,将通过this_cpu_ptr计算的指针传递给其他cpu是不寻常的,应该避免。
|
| /linux/drivers/hv/ |
| H A D | mshv_root_hv_call.c | 60 input_page = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_withdraw_memory() 103 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_create_partition() 104 output = *this_cpu_ptr(hyperv_pcpu_output_arg); in hv_call_create_partition() 220 input_page = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_do_map_gpa_hcall() 322 input_page = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_unmap_gpa_pages() 359 input_page = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_get_gpa_access_states() 360 output_page = *this_cpu_ptr(hyperv_pcpu_output_arg); in hv_call_get_gpa_access_states() 395 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_assert_virtual_interrupt() 450 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_get_vp_state() 451 output = *this_cpu_ptr(hyperv_pcpu_output_arg); in hv_call_get_vp_state() [all …]
|
| H A D | mshv_common.c | 43 input_page = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_get_vp_registers() 44 output_page = *this_cpu_ptr(hyperv_pcpu_output_arg); in hv_call_get_vp_registers() 87 input_page = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_set_vp_registers() 126 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_get_partition_property() 127 output = *this_cpu_ptr(hyperv_pcpu_output_arg); in hv_call_get_partition_property() 173 in = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_initialize_sleep_states() 231 in = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_machine_power_off()
|
| H A D | hv_proc.c | 77 input_page = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_deposit_pages() 173 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_add_logical_proc() 175 output = *this_cpu_ptr(hyperv_pcpu_output_arg); in hv_call_add_logical_proc() 217 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_create_vp() 251 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_call_notify_all_processors_started() 273 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_lp_exists() 274 output = *this_cpu_ptr(hyperv_pcpu_output_arg); in hv_lp_exists()
|
| H A D | hv_common.c | 272 output = *this_cpu_ptr(hyperv_pcpu_input_arg); in hv_get_partition_id() 293 input = *this_cpu_ptr(hyperv_pcpu_input_arg); in get_vtl() 294 output = *this_cpu_ptr(hyperv_pcpu_output_arg); in get_vtl() 485 inputarg = (void **)this_cpu_ptr(hyperv_pcpu_input_arg); in hv_common_cpu_init() 497 outputarg = (void **)this_cpu_ptr(hyperv_pcpu_output_arg); in hv_common_cpu_init() 537 synic_eventring_tail = (u8 **)this_cpu_ptr(hv_synic_eventring_tail); in hv_common_cpu_init() 564 synic_eventring_tail = this_cpu_ptr(hv_synic_eventring_tail); in hv_common_cpu_die()
|
| /linux/arch/xtensa/kernel/ |
| H A D | hw_breakpoint.c | 178 i = alloc_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_install_hw_breakpoint() 185 i = alloc_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_install_hw_breakpoint() 216 i = free_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_uninstall_hw_breakpoint() 224 i = free_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_uninstall_hw_breakpoint() 269 struct perf_event *bp = this_cpu_ptr(wp_on_reg)[i]; in restore_dbreak() 281 struct perf_event **bp = this_cpu_ptr(bp_on_reg); in check_hw_breakpoint() 290 struct perf_event **bp = this_cpu_ptr(wp_on_reg); in check_hw_breakpoint()
|
| /linux/arch/x86/events/intel/ |
| H A D | bts.c | 235 struct bts_ctx *bts = this_cpu_ptr(bts_ctx); in __bts_event_start() 263 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_start() 264 struct bts_ctx *bts = this_cpu_ptr(bts_ctx); in bts_event_start() 294 struct bts_ctx *bts = this_cpu_ptr(bts_ctx); in __bts_event_stop() 308 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_stop() 309 struct bts_ctx *bts = this_cpu_ptr(bts_ctx); in bts_event_stop() 348 bts = this_cpu_ptr(bts_ctx); in intel_bts_enable_local() 372 bts = this_cpu_ptr(bts_ctx); in intel_bts_disable_local() 463 struct debug_store *ds = this_cpu_ptr(&cpu_hw_events)->ds; in intel_bts_interrupt() 473 bts = this_cpu_ptr(bts_ctx); in intel_bts_interrupt() [all …]
|
| /linux/kernel/time/ |
| H A D | tick-sched.c | 424 irq_work_queue(this_cpu_ptr(&nohz_full_kick_work)); in tick_nohz_full_kick() 615 ts = this_cpu_ptr(&tick_cpu_sched); in __tick_nohz_task_switch() 718 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_tick_stopped() 1250 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_stop_tick() 1286 tick_nohz_retain_tick(this_cpu_ptr(&tick_cpu_sched)); in tick_nohz_idle_retain_tick() 1302 ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_enter() 1333 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_irq_exit() 1348 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_got_tick() 1386 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_get_sleep_length() 1456 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_restart_tick() [all …]
|
| H A D | tick-common.c | 264 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_install_replacement() 377 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_broadcast_oneshot_control() 422 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_shutdown() 443 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_suspend_local() 457 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_resume_local()
|
| /linux/net/core/ |
| H A D | dst_cache.c | 75 dst = dst_cache_per_cpu_get(dst_cache, this_cpu_ptr(dst_cache->cache)); in dst_cache_get() 90 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_get_ip4() 112 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_set_ip4() 130 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_set_ip6() 149 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_get_ip6()
|
| /linux/kernel/irq/ |
| H A D | ipi-mux.c | 31 struct ipi_mux_cpu *icpu = this_cpu_ptr(ipi_mux_pcpu); in ipi_mux_mask() 38 struct ipi_mux_cpu *icpu = this_cpu_ptr(ipi_mux_pcpu); in ipi_mux_unmask() 56 struct ipi_mux_cpu *icpu = this_cpu_ptr(ipi_mux_pcpu); in ipi_mux_send_mask() 122 struct ipi_mux_cpu *icpu = this_cpu_ptr(ipi_mux_pcpu); in ipi_mux_process()
|
| /linux/arch/arm64/kvm/ |
| H A D | vmid.c | 135 atomic64_set(this_cpu_ptr(&active_vmids), VMID_ACTIVE_INVALID); in kvm_arm_vmid_clear_active() 155 old_active_vmid = atomic64_read(this_cpu_ptr(&active_vmids)); in kvm_arm_vmid_update() 157 0 != atomic64_cmpxchg_relaxed(this_cpu_ptr(&active_vmids), in kvm_arm_vmid_update() 168 atomic64_set(this_cpu_ptr(&active_vmids), vmid); in kvm_arm_vmid_update()
|
| /linux/arch/arm64/kernel/ |
| H A D | hw_breakpoint.c | 238 slots = this_cpu_ptr(bp_on_reg); in hw_breakpoint_control() 245 slots = this_cpu_ptr(wp_on_reg); in hw_breakpoint_control() 590 slots = this_cpu_ptr(bp_on_reg); in toggle_bp_registers() 594 slots = this_cpu_ptr(wp_on_reg); in toggle_bp_registers() 631 slots = this_cpu_ptr(bp_on_reg); in do_breakpoint() 681 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in do_breakpoint() 763 slots = this_cpu_ptr(wp_on_reg); in do_watchpoint() 829 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in do_watchpoint() 850 int *kernel_step = this_cpu_ptr(&stepping_kernel_bp); in try_step_suspended_breakpoints() 950 for (slots = this_cpu_ptr(bp_on_reg), i = 0; i < core_num_brps; ++i) { in hw_breakpoint_reset() [all …]
|
| /linux/include/linux/ |
| H A D | local_lock_internal.h | 180 #define __local_lock_is_locked(lock) READ_ONCE(this_cpu_ptr(lock)->acquired) 314 (rt_mutex_owner(&this_cpu_ptr(__lock)->lock) == current) 328 return this_cpu_ptr(base); in __this_cpu_local_lock() 334 return this_cpu_ptr(base); in __this_cpu_local_lock() 338 #define __this_cpu_local_lock(base) this_cpu_ptr(base)
|
| /linux/arch/x86/xen/ |
| H A D | multicalls.c | 68 return this_cpu_ptr(mc_debug_data_ptr); in get_mc_debug() 144 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_flush() 221 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in __xen_mc_entry() 256 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_extend_args() 286 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_callback()
|
| /linux/arch/sh/kernel/ |
| H A D | kprobes.c | 97 struct kprobe *saved = this_cpu_ptr(&saved_next_opcode); in arch_remove_kprobe() 106 saved = this_cpu_ptr(&saved_next_opcode2); in arch_remove_kprobe() 148 op1 = this_cpu_ptr(&saved_next_opcode); in prepare_singlestep() 149 op2 = this_cpu_ptr(&saved_next_opcode2); in prepare_singlestep() 321 p = this_cpu_ptr(&saved_next_opcode); in post_kprobe_handler() 333 p = this_cpu_ptr(&saved_next_opcode2); in post_kprobe_handler()
|
| H A D | perf_event.c | 201 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_stop() 219 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_start() 236 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_del() 246 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_add()
|
| /linux/mm/kasan/ |
| H A D | quarantine.c | 209 q = this_cpu_ptr(&cpu_quarantine); in kasan_quarantine_put() 320 sq = this_cpu_ptr(&shrink_qlist); in __per_cpu_remove_cache() 330 q = this_cpu_ptr(&cpu_quarantine); in per_cpu_remove_cache() 385 this_cpu_ptr(&cpu_quarantine)->offline = false; in kasan_cpu_online() 393 q = this_cpu_ptr(&cpu_quarantine); in kasan_cpu_offline()
|
| /linux/lib/ |
| H A D | irq_poll.c | 37 list_add_tail(&iop->list, this_cpu_ptr(&blk_cpu_iopoll)); in irq_poll_sched() 80 struct list_head *list = this_cpu_ptr(&blk_cpu_iopoll); in irq_poll_softirq() 199 this_cpu_ptr(&blk_cpu_iopoll)); in irq_poll_cpu_dead()
|
| /linux/kernel/ |
| H A D | user-return-notifier.c | 18 hlist_add_head(&urn->link, this_cpu_ptr(&return_notifier_list)); in user_return_notifier_register() 29 if (hlist_empty(this_cpu_ptr(&return_notifier_list))) in user_return_notifier_unregister()
|
| /linux/arch/powerpc/platforms/powernv/ |
| H A D | opal-tracepoints.c | 54 depth = this_cpu_ptr(&opal_trace_depth); in __trace_opal_entry() 75 depth = this_cpu_ptr(&opal_trace_depth); in __trace_opal_exit()
|
| /linux/arch/arm64/include/asm/ |
| H A D | hardirq.h | 36 ___ctx = this_cpu_ptr(&nmi_contexts); \ 69 ___ctx = this_cpu_ptr(&nmi_contexts); \
|