| /linux/arch/x86/mm/ |
| H A D | tlb.c | 207 if (asid == this_cpu_read(cpu_tlbstate.loaded_mm_asid)) in clear_asid_other() 250 if (this_cpu_read(cpu_tlbstate.invalidate_other)) in choose_new_asid() 254 if (this_cpu_read(cpu_tlbstate.ctxs[asid].ctx_id) != in choose_new_asid() 259 ns.need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < next_tlb_gen); in choose_new_asid() 587 struct mm_struct *loaded_mm = this_cpu_read(cpu_tlbstate.loaded_mm); in leave_mm() 601 WARN_ON(!this_cpu_read(cpu_tlbstate_shared.is_lazy)); in leave_mm() 645 if (this_cpu_read(cpu_info.smt_active)) { in l1d_flush_evaluate() 674 prev_mm = this_cpu_read(cpu_tlbstate.last_user_mm_spec); in cond_mitigation() 770 cr4_update_pce_mm(this_cpu_read(cpu_tlbstate.loaded_mm)); in cr4_update_pce() 786 struct mm_struct *prev = this_cpu_read(cpu_tlbstate.loaded_mm); in switch_mm_irqs_off() [all …]
|
| /linux/kernel/trace/ |
| H A D | trace_preemptirq.c | 61 if (this_cpu_read(tracing_irq_cpu)) { in trace_hardirqs_on_prepare() 72 if (this_cpu_read(tracing_irq_cpu)) { in trace_hardirqs_on() 92 if (!this_cpu_read(tracing_irq_cpu)) { in trace_hardirqs_off_finish() 106 if (!this_cpu_read(tracing_irq_cpu)) { in trace_hardirqs_off()
|
| /linux/drivers/irqchip/ |
| H A D | irq-csky-mpintc.c | 56 (this_cpu_read(intcl_reg) + INTCL_CFGR) : (INTCG_base + INTCG_CICFGR))) 75 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_handler() 83 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_unmask() 92 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_mask() 99 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_eoi() 215 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_send_ipi()
|
| /linux/kernel/ |
| H A D | watchdog_perf.c | 151 WARN_ONCE(this_cpu_read(watchdog_ev), "unexpected watchdog_ev leak"); in hardlockup_detector_event_create() 172 perf_event_enable(this_cpu_read(watchdog_ev)); in watchdog_hardlockup_enable() 181 struct perf_event *event = this_cpu_read(watchdog_ev); in watchdog_hardlockup_disable() 200 struct perf_event *event = this_cpu_read(watchdog_ev); in hardlockup_detector_perf_adjust_period() 276 perf_event_release_kernel(this_cpu_read(watchdog_ev)); in watchdog_hardlockup_probe()
|
| H A D | softirq.c | 176 DEBUG_LOCKS_WARN_ON(this_cpu_read(softirq_ctrl.cnt)); in __local_bh_disable_ip() 224 this_cpu_read(softirq_ctrl.cnt)); in __local_bh_enable() 272 curcnt = this_cpu_read(softirq_ctrl.cnt); in __local_bh_enable_ip() 335 return !this_cpu_read(softirq_ctrl.cnt); in should_wake_ksoftirqd()
|
| /linux/arch/x86/include/asm/ |
| H A D | kvmclock.h | 11 return &this_cpu_read(hv_clock_per_cpu)->pvti; in this_cpu_pvti() 16 return this_cpu_read(hv_clock_per_cpu); in this_cpu_hvclock()
|
| H A D | entry-common.h | 99 this_cpu_read(x86_ibpb_exit_to_user)) { in arch_exit_to_user_mode_prepare()
|
| /linux/arch/loongarch/kernel/ |
| H A D | kfpu.c | 45 WARN_ON(this_cpu_read(in_kernel_fpu)); in kernel_fpu_begin() 73 WARN_ON(!this_cpu_read(in_kernel_fpu)); in kernel_fpu_end()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | this_cpu_ops.rst | 46 this_cpu_read(pcp) 75 z = this_cpu_read(x); 232 且更容易与 ``()`` 结合。第二种形式也与 ``this_cpu_read()`` 和大家的使用方式一致。
|
| /linux/arch/x86/coco/sev/ |
| H A D | noinstr.c | 123 data = this_cpu_read(runtime_data); in __sev_get_ghcb() 166 data = this_cpu_read(runtime_data); in __sev_put_ghcb()
|
| /linux/kernel/trace/rv/monitors/opid/ |
| H A D | opid.c | 90 if (this_cpu_read(hardirq_context)) in handle_sched_need_resched() 99 if (this_cpu_read(hardirq_context)) in handle_sched_waking()
|
| /linux/tools/testing/shared/linux/ |
| H A D | percpu.h | 7 #define this_cpu_read(var) var macro
|
| /linux/include/crypto/internal/ |
| H A D | simd.h | 54 (may_use_simd() && !this_cpu_read(crypto_simd_disabled_for_test))
|
| /linux/drivers/accessibility/speakup/ |
| H A D | fakekey.c | 86 return this_cpu_read(reporting_keystroke); in speakup_fake_key_pressed()
|
| /linux/arch/x86/kernel/cpu/resctrl/ |
| H A D | pseudo_lock.c | 167 closid_p = this_cpu_read(pqr_state.cur_closid); in resctrl_arch_pseudo_lock_fn() 168 rmid_p = this_cpu_read(pqr_state.cur_rmid); in resctrl_arch_pseudo_lock_fn()
|
| /linux/arch/x86/kernel/cpu/mce/ |
| H A D | core.c | 831 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in machine_check_poll() 997 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in mce_no_way_out() 1315 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in mce_clear_state() 1366 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in __mc_scan_banks() 1849 u8 n_banks = this_cpu_read(mce_num_banks); in __mcheck_cpu_mce_banks_init() 1916 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in __mcheck_cpu_init_prepare_banks() 2045 if (this_cpu_read(mce_num_banks) > 8) in mce_zhaoxin_feature_init() 2330 if (bank >= this_cpu_read(mce_num_banks)) { in mce_disable_bank() 2419 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in mce_disable_error_reporting() 2775 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in mce_reenable_cpu()
|
| /linux/arch/x86/kernel/ |
| H A D | doublefault_32.c | 16 #define TSS(x) this_cpu_read(cpu_tss_rw.x86_tss.x)
|
| /linux/arch/powerpc/lib/ |
| H A D | code-patching.c | 135 free_vm_area(this_cpu_read(cpu_patching_context.area)); in text_area_cpu_down() 195 put_patching_mm(this_cpu_read(cpu_patching_context.mm), in text_area_cpu_down_mm() 196 this_cpu_read(cpu_patching_context.addr)); in text_area_cpu_down_mm()
|
| /linux/arch/x86/kernel/fpu/ |
| H A D | core.c | 90 if (!this_cpu_read(kernel_fpu_allowed)) in irq_fpu_usable() 480 WARN_ON_FPU(!this_cpu_read(kernel_fpu_allowed)); in kernel_fpu_begin_mask() 502 WARN_ON_FPU(this_cpu_read(kernel_fpu_allowed)); in kernel_fpu_end()
|
| /linux/mm/kasan/ |
| H A D | sw_tags.c | 67 u32 state = this_cpu_read(prng_state); in kasan_random_tag()
|
| /linux/arch/powerpc/mm/nohash/ |
| H A D | e500_hugetlbpage.c | 84 index = this_cpu_read(next_tlbcam_idx); in tlb1_next()
|
| /linux/arch/x86/platform/uv/ |
| H A D | uv_nmi.c | 825 while (this_cpu_read(uv_cpu_nmi.state) != UV_NMI_STATE_DUMP) in uv_nmi_dump_state() 947 if (!this_cpu_read(uv_cpu_nmi.pinging) && !uv_check_nmi(hub_nmi)) { in uv_handle_nmi() 1018 if (!this_cpu_read(uv_cpu_nmi.pinging)) { in uv_handle_nmi_ping()
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | intel_epb.c | 94 u64 val = this_cpu_read(saved_epb); in intel_epb_restore()
|
| /linux/kernel/irq/ |
| H A D | irqdesc.c | 953 this_cpu_write(desc->kstat_irqs->ref, this_cpu_read(desc->kstat_irqs->cnt)); in kstat_snapshot_irqs() 963 return this_cpu_read(desc->kstat_irqs->cnt) - this_cpu_read(desc->kstat_irqs->ref); in kstat_get_irq_since_snapshot()
|
| /linux/arch/x86/kernel/apic/ |
| H A D | x2apic_cluster.c | 93 struct cpumask *cmsk = this_cpu_read(cluster_masks); in init_x2apic_ldr()
|