Home
last modified time | relevance | path

Searched refs:__this_cpu_read (Results 1 – 25 of 92) sorted by relevance

1234

/linux/arch/sparc/kernel/
H A Dnmi.c107 if (__this_cpu_read(nmi_touch)) { in perfctr_irq()
111 if (!touched && __this_cpu_read(last_irq_sum) == sum) { in perfctr_irq()
113 if (__this_cpu_read(alert_counter) == 30 * nmi_hz) in perfctr_irq()
120 if (__this_cpu_read(wd_enabled)) { in perfctr_irq()
160 if (!__this_cpu_read(wd_enabled)) in stop_nmi_watchdog()
217 if (__this_cpu_read(wd_enabled)) in start_nmi_watchdog()
231 if (!__this_cpu_read(wd_enabled)) in nmi_adjust_hz_one()
/linux/arch/mips/kernel/
H A Dmips-r2-to-r6-emul.c2243 (unsigned long)__this_cpu_read(mipsr2emustats.movs), in mipsr2_emul_show()
2244 (unsigned long)__this_cpu_read(mipsr2bdemustats.movs)); in mipsr2_emul_show()
2246 (unsigned long)__this_cpu_read(mipsr2emustats.hilo), in mipsr2_emul_show()
2247 (unsigned long)__this_cpu_read(mipsr2bdemustats.hilo)); in mipsr2_emul_show()
2249 (unsigned long)__this_cpu_read(mipsr2emustats.muls), in mipsr2_emul_show()
2250 (unsigned long)__this_cpu_read(mipsr2bdemustats.muls)); in mipsr2_emul_show()
2252 (unsigned long)__this_cpu_read(mipsr2emustats.divs), in mipsr2_emul_show()
2253 (unsigned long)__this_cpu_read(mipsr2bdemustats.divs)); in mipsr2_emul_show()
2255 (unsigned long)__this_cpu_read(mipsr2emustats.dsps), in mipsr2_emul_show()
2256 (unsigned long)__this_cpu_read(mipsr2bdemustats.dsps)); in mipsr2_emul_show()
[all …]
/linux/kernel/
H A Dwatchdog.c471 u8 tail = __this_cpu_read(cpustat_tail); in update_cpustat()
477 old_stat = __this_cpu_read(cpustat_old[i]); in update_cpustat()
498 u8 tail = __this_cpu_read(cpustat_tail); in print_cpustat()
515 __this_cpu_read(cpustat_util[group][STATS_SYSTEM]), in print_cpustat()
516 __this_cpu_read(cpustat_util[group][STATS_SOFTIRQ]), in print_cpustat()
517 __this_cpu_read(cpustat_util[group][STATS_HARDIRQ]), in print_cpustat()
518 __this_cpu_read(cpustat_util[group][STATS_IDLE])); in print_cpustat()
551 int tail = __this_cpu_read(cpustat_tail); in need_counting_irqs()
554 util = __this_cpu_read(cpustat_util[tail][STATS_HARDIRQ]); in need_counting_irqs()
560 if (!__this_cpu_read(snapshot_taken)) { in start_counting_irqs()
[all …]
H A Dsoftirq.c78 struct task_struct *tsk = __this_cpu_read(ksoftirqd); in wakeup_softirqd()
153 return __this_cpu_read(softirq_ctrl.cnt) != 0; in local_bh_blocked()
489 if (!force_irqthreads() || !__this_cpu_read(ksoftirqd)) { in invoke_softirq()
701 struct task_struct *tsk = __this_cpu_read(ktimerd); in wake_timersd()
1081 *__this_cpu_read(tasklet_vec.tail) = per_cpu(tasklet_vec, cpu).head; in takeover_tasklets()
1089 *__this_cpu_read(tasklet_hi_vec.tail) = per_cpu(tasklet_hi_vec, cpu).head; in takeover_tasklets()
/linux/kernel/time/
H A Dtick-oneshot.c29 struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); in tick_program_event()
56 struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); in tick_resume_oneshot()
127 ret = __this_cpu_read(tick_cpu_device.mode) == TICKDEV_MODE_ONESHOT; in tick_oneshot_mode_active()
/linux/include/asm-generic/
H A Dirq_regs.h21 return __this_cpu_read(__irq_regs); in get_irq_regs()
28 old_regs = __this_cpu_read(__irq_regs); in set_irq_regs()
/linux/include/linux/
H A Dcontext_tracking_state.h110 return __this_cpu_read(context_tracking.nesting); in ct_nesting()
122 return __this_cpu_read(context_tracking.nmi_nesting); in ct_nmi_nesting()
148 return context_tracking_enabled() && __this_cpu_read(context_tracking.active); in context_tracking_enabled_this_cpu()
/linux/arch/powerpc/lib/
H A Dcode-patching.c292 patching_mm = __this_cpu_read(cpu_patching_context.mm); in __do_patch_mem_mm()
293 text_poke_addr = __this_cpu_read(cpu_patching_context.addr); in __do_patch_mem_mm()
335 text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK; in __do_patch_mem()
338 pte = __this_cpu_read(cpu_patching_context.pte); in __do_patch_mem()
478 patching_mm = __this_cpu_read(cpu_patching_context.mm); in __do_patch_instructions_mm()
479 text_poke_addr = __this_cpu_read(cpu_patching_context.addr); in __do_patch_instructions_mm()
527 text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK; in __do_patch_instructions()
530 pte = __this_cpu_read(cpu_patching_context.pte); in __do_patch_instructions()
/linux/arch/x86/kernel/
H A Dirq.c262 struct irq_desc *desc = __this_cpu_read(vector_irq[vector]); in reevaluate_vector()
276 struct irq_desc *desc = __this_cpu_read(vector_irq[vector]); in call_irq_handler()
433 if (unlikely(!__this_cpu_read(posted_msi_handler_active))) in intel_ack_posted_msi_irq()
532 if (IS_ERR_OR_NULL(__this_cpu_read(vector_irq[vector]))) in fixup_irqs()
536 desc = __this_cpu_read(vector_irq[vector]); in fixup_irqs()
547 if (__this_cpu_read(vector_irq[vector]) != VECTOR_RETRIGGERED) in fixup_irqs()
H A Dkvm.c263 if (__this_cpu_read(async_pf_enabled)) { in kvm_read_and_reset_apf_flags()
264 flags = __this_cpu_read(apf_reason.flags); in kvm_read_and_reset_apf_flags()
314 if (__this_cpu_read(async_pf_enabled)) { in DEFINE_IDTENTRY_SYSVEC()
315 token = __this_cpu_read(apf_reason.token); in DEFINE_IDTENTRY_SYSVEC()
402 if (!__this_cpu_read(async_pf_enabled)) in kvm_pv_disable_apf()
/linux/lib/
H A Dpercpu_counter.c106 count = __this_cpu_read(*fbc->counters); in percpu_counter_add_batch()
126 count = __this_cpu_read(*fbc->counters) + amount; in percpu_counter_add_batch()
152 count = __this_cpu_read(*fbc->counters); in percpu_counter_sync()
340 count = __this_cpu_read(*fbc->counters); in __percpu_counter_limited_add()
385 count = __this_cpu_read(*fbc->counters); in __percpu_counter_limited_add()
/linux/drivers/xen/events/
H A Devents_2l.c123 struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu); in evtchn_2l_unmask()
173 struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu); in evtchn_2l_handle_events()
192 start_word_idx = __this_cpu_read(current_word_idx); in evtchn_2l_handle_events()
193 start_bit_idx = __this_cpu_read(current_bit_idx); in evtchn_2l_handle_events()
/linux/arch/x86/include/asm/
H A Dcpu_entry_area.h148 CEA_ESTACK_TOP(__this_cpu_read(cea_exception_stacks), name)
151 CEA_ESTACK_BOT(__this_cpu_read(cea_exception_stacks), name)
H A Dirq_stack.h119 call_on_stack(__this_cpu_read(hardirq_stack_ptr), \
138 if (user_mode(regs) || __this_cpu_read(hardirq_stack_inuse)) { \
H A Dhardirq.h88 return __this_cpu_read(irq_stat.kvm_cpu_l1tf_flush_l1d); in kvm_get_cpu_l1tf_flush_l1d()
/linux/arch/x86/coco/sev/
H A Dnoinstr.c57 new_ist = old_ist = __this_cpu_read(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC]); in __sev_es_ist_enter()
83 ist = __this_cpu_read(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC]); in __sev_es_ist_exit()
/linux/arch/powerpc/kernel/
H A Dhw_breakpoint.c236 struct perf_event *bp = __this_cpu_read(bp_per_reg[i]); in thread_change_pc()
246 info = counter_arch_bp(__this_cpu_read(bp_per_reg[i])); in thread_change_pc()
408 bp[i] = __this_cpu_read(bp_per_reg[i]); in hw_breakpoint_handler()
519 bp = __this_cpu_read(bp_per_reg[i]); in single_step_dabr_instruction()
/linux/arch/arm64/include/asm/
H A Darch_timer.h27 __wa = __this_cpu_read(timer_unstable_counter_workaround); \
34 __wa = __this_cpu_read(timer_unstable_counter_workaround); \
/linux/arch/x86/xen/
H A Dtime.c53 src = &__this_cpu_read(xen_vcpu)->time; in xen_clocksource_read()
69 src = &__this_cpu_read(xen_vcpu)->time; in xen_sched_clock()
547 pvti = &__this_cpu_read(xen_vcpu)->time; in xen_time_init()
628 if (!__this_cpu_read(xen_vcpu)) { in xen_hvm_init_time_ops()
/linux/drivers/irqchip/
H A Dirq-xtensa-mx.c83 mask = __this_cpu_read(cached_irq_mask) & ~mask; in xtensa_mx_irq_mask()
101 mask |= __this_cpu_read(cached_irq_mask); in xtensa_mx_irq_unmask()
/linux/drivers/cpuidle/
H A Dcpuidle-psci.c124 struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev); in psci_idle_cpuhp_up()
138 struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev); in psci_idle_cpuhp_down()
221 u32 *state = __this_cpu_read(psci_cpuidle_data.psci_states); in psci_enter_idle_state()
H A Dcpuidle-riscv-sbi.c81 u32 *states = __this_cpu_read(sbi_cpuidle_data.states); in sbi_cpuidle_enter_state()
149 struct device *pd_dev = __this_cpu_read(sbi_cpuidle_data.dev); in sbi_cpuidle_cpuhp_up()
159 struct device *pd_dev = __this_cpu_read(sbi_cpuidle_data.dev); in sbi_cpuidle_cpuhp_down()
/linux/mm/
H A Dvmstat.c362 x = delta + __this_cpu_read(*p); in __mod_zone_page_state()
364 t = __this_cpu_read(pcp->stat_threshold); in __mod_zone_page_state()
398 x = delta + __this_cpu_read(*p); in __mod_node_page_state()
400 t = __this_cpu_read(pcp->stat_threshold); in __mod_node_page_state()
445 t = __this_cpu_read(pcp->stat_threshold); in __inc_zone_state()
468 t = __this_cpu_read(pcp->stat_threshold); in __inc_node_state()
501 t = __this_cpu_read(pcp->stat_threshold); in __dec_zone_state()
524 t = __this_cpu_read(pcp->stat_threshold); in __dec_node_state()
855 if (!__this_cpu_read(pcp->expire) || in refresh_cpu_vm_stats()
856 !__this_cpu_read(pcp->count)) in refresh_cpu_vm_stats()
[all …]
/linux/drivers/clocksource/
H A Dnumachip.c59 unsigned local_apicid = __this_cpu_read(x86_cpu_to_apicid) & 0xff; in numachip_timer_each()
/linux/arch/powerpc/include/asm/
H A Dtime.h103 return __this_cpu_read(decrementers_next_tb); in timer_get_next_tb()

1234