| /linux/lib/tests/ |
| H A D | test_fprobe.c | 39 KUNIT_EXPECT_FALSE(current_test, preemptible()); in fp_entry_handler() 60 KUNIT_EXPECT_FALSE(current_test, preemptible()); in fp_exit_handler() 199 KUNIT_EXPECT_FALSE(current_test, preemptible()); in entry_only_handler() 213 KUNIT_EXPECT_FALSE(current_test, preemptible()); in fprobe_entry_multi_handler() 229 KUNIT_EXPECT_FALSE(current_test, preemptible()); in fprobe_exit_multi_handler()
|
| /linux/kernel/ |
| H A D | Kconfig.preempt | 66 preemptible. This allows reaction to interactive events by 99 preemptible priority-inheritance aware variants, enforcing 101 non-preemptible sections. This makes the kernel, except for very 103 level interrupt handling) fully preemptible and brings most 114 the softirq is preemptible. This enforces the same per-CPU BLK 119 this if you suspect an error with preemptible softirq and want test
|
| H A D | softirq.c | 167 if (preemptible()) { in __local_bh_disable_ip() 260 bool preempt_on = preemptible(); in __local_bh_enable_ip()
|
| /linux/drivers/irqchip/ |
| H A D | irq-gic-v4.c | 240 WARN_ON(preemptible()); in its_make_vpe_non_resident() 266 WARN_ON(preemptible()); in its_make_vpe_resident() 291 WARN_ON(preemptible()); in its_commit_vpe()
|
| /linux/arch/x86/xen/ |
| H A D | multicalls.c | 151 BUG_ON(preemptible()); in xen_mc_flush() 227 BUG_ON(preemptible()); in __xen_mc_entry() 259 BUG_ON(preemptible()); in xen_mc_extend_args()
|
| H A D | enlighten_pv.c | 425 BUG_ON(preemptible()); in xen_start_context_switch() 435 BUG_ON(preemptible()); in xen_end_context_switch()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | local_ops.rst | 150 /* Increment the counter from a non preemptible context */ 155 * preemptible context (it disables preemption) :
|
| /linux/include/linux/ |
| H A D | seqlock.h | 142 * @preemptible: preemptibility of above locktype 145 #define SEQCOUNT_LOCKNAME(lockname, locktype, preemptible, lockbase) \ 166 if (preemptible && unlikely(seq & 1)) { \ 184 return preemptible; \ 263 #define seqprop_preemptible(s) __seqprop(s, preemptible)(s) 487 * non-preemptible. Preemption will be automatically disabled if and 489 * preemptible. If readers can be invoked from hardirq or softirq 512 * the seqcount write serialization lock is associated, and preemptible. 607 * As the write sections are fully preemptible, no special handling for 871 * automatically serialized and non-preemptible 144 SEQCOUNT_LOCKNAME(lockname,locktype,preemptible,lockbase) global() argument [all...] |
| H A D | seqlock_types.h | 14 * Write side critical sections must be serialized and non-preemptible. 46 * To remain preemptible while avoiding a possible livelock caused by the 62 #define SEQCOUNT_LOCKNAME(lockname, locktype, preemptible, lockbase) \ argument
|
| H A D | kprobes.h | 587 if (preemptible()) in kprobe_page_fault()
|
| /linux/arch/arm64/kernel/ |
| H A D | debug-monitors.c | 88 WARN_ON(preemptible()); in enable_debug_monitors() 109 WARN_ON(preemptible()); in disable_debug_monitors()
|
| H A D | fpsimd.c | 365 WARN_ON(preemptible()); in task_fpsimd_load() 451 WARN_ON(preemptible()); in fpsimd_save_user_state() 1788 WARN_ON(preemptible()); in fpsimd_save_and_flush_cpu_state() 1824 WARN_ON((preemptible() || in_serving_softirq()) && !state); in kernel_neon_begin()
|
| H A D | stacktrace.c | 309 ((task == current) && !preemptible()) \
|
| /linux/arch/arm64/include/asm/ |
| H A D | archrandom.h | 64 if (unlikely(!system_capabilities_finalized() && !preemptible())) in __cpu_has_rng()
|
| /linux/drivers/firmware/ |
| H A D | arm_sdei.c | 318 WARN_ON_ONCE(preemptible()); in _ipi_mask_cpu() 338 WARN_ON_ONCE(preemptible()); in _ipi_unmask_cpu() 346 WARN_ON_ONCE(preemptible()); in _ipi_private_reset() 716 WARN_ON_ONCE(preemptible()); in sdei_pm_notifier()
|
| /linux/Documentation/core-api/ |
| H A D | local_ops.rst | 103 In preemptible context, use ``get_cpu_var()`` and ``put_cpu_var()`` around 156 /* Increment the counter from a non preemptible context */ 161 * preemptible context (it disables preemption) :
|
| /linux/kernel/rcu/ |
| H A D | update.c | 126 return lock_is_held(&rcu_sched_lock_map) || !preemptible(); in rcu_read_lock_sched_held() 390 return !preemptible(); in rcu_read_lock_any_held()
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_tsn.c | 168 tx_ring->preemptible = false; in igc_fpe_clear_preempt_queue() 201 tx_ring->preemptible = !!(preemptible_queue & BIT(i)); in igc_fpe_save_preempt_queue() 508 } else if (ring->preemptible) { in igc_tsn_enable_offload()
|
| /linux/drivers/xen/ |
| H A D | time.c | 61 BUG_ON(preemptible()); in xen_get_runstate_snapshot_cpu_delta()
|
| /linux/Documentation/locking/ |
| H A D | preempt-locking.rst | 12 A preemptible kernel creates new locking issues. The issues are the same as 13 those under SMP: concurrency and reentrancy. Thankfully, the Linux preemptible
|
| /linux/arch/arm64/kvm/ |
| H A D | debug.c | 256 return (WARN_ON_ONCE(preemptible()) || skip_condition || in skip_trbe_access()
|
| /linux/Documentation/RCU/ |
| H A D | UP.rst | 110 preemptible RCU? 146 running preemptible RCU?
|
| /linux/Documentation/mm/ |
| H A D | highmem.rst | 125 Each call of kmap_atomic() in the kernel creates a non-preemptible section 149 available. Therefore, kmap() is only callable from preemptible context.
|
| /linux/Documentation/admin-guide/thermal/ |
| H A D | intel_powerclamp.rst | 90 scheme to work for both preemptible and non-preemptible kernels.
|
| /linux/arch/x86/kernel/ |
| H A D | ldt.c | 113 DEBUG_LOCKS_WARN_ON(preemptible()); in switch_ldt()
|