| /linux/arch/mips/kernel/ |
| H A D | mips-r2-to-r6-emul.c | 2243 (unsigned long)__this_cpu_read(mipsr2emustats.movs), in mipsr2_emul_show() 2244 (unsigned long)__this_cpu_read(mipsr2bdemustats.movs)); in mipsr2_emul_show() 2246 (unsigned long)__this_cpu_read(mipsr2emustats.hilo), in mipsr2_emul_show() 2247 (unsigned long)__this_cpu_read(mipsr2bdemustats.hilo)); in mipsr2_emul_show() 2249 (unsigned long)__this_cpu_read(mipsr2emustats.muls), in mipsr2_emul_show() 2250 (unsigned long)__this_cpu_read(mipsr2bdemustats.muls)); in mipsr2_emul_show() 2252 (unsigned long)__this_cpu_read(mipsr2emustats.divs), in mipsr2_emul_show() 2253 (unsigned long)__this_cpu_read(mipsr2bdemustats.divs)); in mipsr2_emul_show() 2255 (unsigned long)__this_cpu_read(mipsr2emustats.dsps), in mipsr2_emul_show() 2256 (unsigned long)__this_cpu_read(mipsr2bdemustats.dsps)); in mipsr2_emul_show() [all …]
|
| /linux/arch/sparc/kernel/ |
| H A D | nmi.c | 107 if (__this_cpu_read(nmi_touch)) { in perfctr_irq() 111 if (!touched && __this_cpu_read(last_irq_sum) == sum) { in perfctr_irq() 113 if (__this_cpu_read(alert_counter) == 30 * nmi_hz) in perfctr_irq() 120 if (__this_cpu_read(wd_enabled)) { in perfctr_irq() 160 if (!__this_cpu_read(wd_enabled)) in stop_nmi_watchdog() 217 if (__this_cpu_read(wd_enabled)) in start_nmi_watchdog() 231 if (!__this_cpu_read(wd_enabled)) in nmi_adjust_hz_one()
|
| /linux/kernel/time/ |
| H A D | tick-oneshot.c | 29 struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); in tick_program_event() 56 struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); in tick_resume_oneshot() 127 ret = __this_cpu_read(tick_cpu_device.mode) == TICKDEV_MODE_ONESHOT; in tick_oneshot_mode_active()
|
| /linux/include/asm-generic/ |
| H A D | irq_regs.h | 21 return __this_cpu_read(__irq_regs); in get_irq_regs() 28 old_regs = __this_cpu_read(__irq_regs); in set_irq_regs()
|
| /linux/include/linux/ |
| H A D | context_tracking_state.h | 110 return __this_cpu_read(context_tracking.nesting); in ct_nesting() 122 return __this_cpu_read(context_tracking.nmi_nesting); in ct_nmi_nesting() 148 return context_tracking_enabled() && __this_cpu_read(context_tracking.active); in context_tracking_enabled_this_cpu()
|
| /linux/arch/powerpc/lib/ |
| H A D | code-patching.c | 292 patching_mm = __this_cpu_read(cpu_patching_context.mm); in __do_patch_mem_mm() 293 text_poke_addr = __this_cpu_read(cpu_patching_context.addr); in __do_patch_mem_mm() 335 text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK; in __do_patch_mem() 338 pte = __this_cpu_read(cpu_patching_context.pte); in __do_patch_mem() 478 patching_mm = __this_cpu_read(cpu_patching_context.mm); in __do_patch_instructions_mm() 479 text_poke_addr = __this_cpu_read(cpu_patching_context.addr); in __do_patch_instructions_mm() 527 text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK; in __do_patch_instructions() 530 pte = __this_cpu_read(cpu_patching_context.pte); in __do_patch_instructions()
|
| /linux/arch/x86/kernel/ |
| H A D | irq.c | 269 struct irq_desc *desc = __this_cpu_read(vector_irq[vector]); in reevaluate_vector() 283 struct irq_desc *desc = __this_cpu_read(vector_irq[vector]); in call_irq_handler() 450 if (unlikely(!__this_cpu_read(posted_msi_handler_active))) in intel_ack_posted_msi_irq() 546 if (IS_ERR_OR_NULL(__this_cpu_read(vector_irq[vector]))) in fixup_irqs() 550 desc = __this_cpu_read(vector_irq[vector]); in fixup_irqs() 561 if (__this_cpu_read(vector_irq[vector]) != VECTOR_RETRIGGERED) in fixup_irqs()
|
| /linux/lib/ |
| H A D | percpu_counter.c | 106 count = __this_cpu_read(*fbc->counters); in percpu_counter_add_batch() 126 count = __this_cpu_read(*fbc->counters) + amount; in percpu_counter_add_batch() 152 count = __this_cpu_read(*fbc->counters); in percpu_counter_sync() 340 count = __this_cpu_read(*fbc->counters); in __percpu_counter_limited_add() 385 count = __this_cpu_read(*fbc->counters); in __percpu_counter_limited_add()
|
| /linux/drivers/xen/events/ |
| H A D | events_2l.c | 123 struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu); in evtchn_2l_unmask() 173 struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu); in evtchn_2l_handle_events() 192 start_word_idx = __this_cpu_read(current_word_idx); in evtchn_2l_handle_events() 193 start_bit_idx = __this_cpu_read(current_bit_idx); in evtchn_2l_handle_events()
|
| /linux/arch/x86/include/asm/ |
| H A D | cpu_entry_area.h | 148 CEA_ESTACK_TOP(__this_cpu_read(cea_exception_stacks), name) 151 CEA_ESTACK_BOT(__this_cpu_read(cea_exception_stacks), name)
|
| H A D | irq_stack.h | 119 call_on_stack(__this_cpu_read(hardirq_stack_ptr), \ 138 if (user_mode(regs) || __this_cpu_read(hardirq_stack_inuse)) { \
|
| H A D | debugreg.h | 123 return __this_cpu_read(cpu_dr7) & DR_GLOBAL_ENABLE_MASK; in hw_breakpoint_active()
|
| /linux/arch/x86/coco/sev/ |
| H A D | noinstr.c | 58 new_ist = old_ist = __this_cpu_read(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC]); in __sev_es_ist_enter() 84 ist = __this_cpu_read(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC]); in __sev_es_ist_exit()
|
| /linux/arch/powerpc/kernel/ |
| H A D | hw_breakpoint.c | 236 struct perf_event *bp = __this_cpu_read(bp_per_reg[i]); in thread_change_pc() 246 info = counter_arch_bp(__this_cpu_read(bp_per_reg[i])); in thread_change_pc() 408 bp[i] = __this_cpu_read(bp_per_reg[i]); in hw_breakpoint_handler() 519 bp = __this_cpu_read(bp_per_reg[i]); in single_step_dabr_instruction()
|
| /linux/arch/arm64/include/asm/ |
| H A D | arch_timer.h | 27 __wa = __this_cpu_read(timer_unstable_counter_workaround); \ 34 __wa = __this_cpu_read(timer_unstable_counter_workaround); \
|
| H A D | percpu.h | 272 #undef __this_cpu_read 273 #define __this_cpu_read raw_cpu_read macro
|
| /linux/drivers/irqchip/ |
| H A D | irq-xtensa-mx.c | 83 mask = __this_cpu_read(cached_irq_mask) & ~mask; in xtensa_mx_irq_mask() 101 mask |= __this_cpu_read(cached_irq_mask); in xtensa_mx_irq_unmask()
|
| /linux/drivers/cpuidle/ |
| H A D | cpuidle-psci.c | 124 struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev); in psci_idle_cpuhp_up() 138 struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev); in psci_idle_cpuhp_down() 221 u32 *state = __this_cpu_read(psci_cpuidle_data.psci_states); in psci_enter_idle_state()
|
| H A D | cpuidle-riscv-sbi.c | 81 u32 *states = __this_cpu_read(sbi_cpuidle_data.states); in sbi_cpuidle_enter_state() 149 struct device *pd_dev = __this_cpu_read(sbi_cpuidle_data.dev); in sbi_cpuidle_cpuhp_up() 159 struct device *pd_dev = __this_cpu_read(sbi_cpuidle_data.dev); in sbi_cpuidle_cpuhp_down()
|
| /linux/arch/s390/kernel/ |
| H A D | vtime.c | 98 u64 mult = __this_cpu_read(mt_scaling_mult); in scale_vtime() 99 u64 div = __this_cpu_read(mt_scaling_div); in scale_vtime() 139 if (smp_cpu_mtid && time_after64(jiffies_64, __this_cpu_read(mt_scaling_jiffies))) in do_account_vtime()
|
| /linux/drivers/clocksource/ |
| H A D | numachip.c | 59 unsigned local_apicid = __this_cpu_read(x86_cpu_to_apicid) & 0xff; in numachip_timer_each()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | time.h | 103 return __this_cpu_read(decrementers_next_tb); in timer_get_next_tb()
|
| /linux/arch/powerpc/perf/ |
| H A D | hv-24x7.c | 1465 txn_flags = __this_cpu_read(hv_24x7_txn_flags); in h_24x7_event_read() 1478 if (__this_cpu_read(hv_24x7_txn_err)) in h_24x7_event_read() 1536 WARN_ON_ONCE(__this_cpu_read(hv_24x7_txn_flags)); in h_24x7_event_start_txn() 1581 txn_flags = __this_cpu_read(hv_24x7_txn_flags); in h_24x7_event_commit_txn() 1588 ret = __this_cpu_read(hv_24x7_txn_err); in h_24x7_event_commit_txn() 1632 WARN_ON_ONCE(!__this_cpu_read(hv_24x7_txn_flags)); in h_24x7_event_cancel_txn()
|
| /linux/arch/arm64/kvm/hyp/vhe/ |
| H A D | switch.c | 141 write_sysreg(__this_cpu_read(kvm_hyp_vector), vbar_el1); in __activate_traps() 187 host_vectors = __this_cpu_read(this_cpu_vector); in __deactivate_traps()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | this_cpu_ops.rst | 204 __this_cpu_read(pcp)
|