| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | active_mm.rst | 64 计数器,即有多少 “真正的地址空间用户”,另一个是 “mm_count”计数器,即 “lazy” 68 一个lazy的用户仍在活动,所以你实际上得到的情况是,你有一个地址空间 **只** 69 被lazy的用户使用。这通常是一个短暂的生命周期状态,因为一旦这个线程被安排给一 73 “init_mm”应该被认为只是一个 “没有其他上下文时的lazy上下文”,事实上,它主
|
| /linux/kernel/rcu/ |
| H A D | tree_nocb.h | 246 * can elapse before lazy callbacks are flushed. Lazy callbacks 248 * however, LAZY_FLUSH_JIFFIES will ensure no lazy callbacks are 315 unsigned long j, bool lazy) in rcu_nocb_do_flush_bypass() 332 * If the new CB requested was a lazy one, queue it onto the main in rcu_nocb_do_flush_bypass() 335 * the lazy CB is ordered with the existing CBs in the bypass list. in rcu_nocb_do_flush_bypass() 337 if (lazy && rhp) { in rcu_nocb_do_flush_bypass() 359 unsigned long j, bool lazy) in rcu_nocb_flush_bypass() 365 return rcu_nocb_do_flush_bypass(rdp, rhp, j, lazy); 401 bool lazy) in rcu_nocb_try_bypass() 445 if (rdp->nocb_nobypass_count < nocb_nobypass_lim_per_jiffy && !lazy) { in rcu_nocb_try_bypass() 309 rcu_nocb_do_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp_in,unsigned long j,bool lazy) rcu_nocb_do_flush_bypass() argument 353 rcu_nocb_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp,unsigned long j,bool lazy) rcu_nocb_flush_bypass() argument 395 rcu_nocb_try_bypass(struct rcu_data * rdp,struct rcu_head * rhp,bool * was_alldone,unsigned long flags,bool lazy) rcu_nocb_try_bypass() argument 596 call_rcu_nocb(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func,unsigned long flags,bool lazy) call_rcu_nocb() argument 656 bool lazy = false; nocb_gp_wait() local 1666 rcu_nocb_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp,unsigned long j,bool lazy) rcu_nocb_flush_bypass() argument 1672 call_rcu_nocb(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func,unsigned long flags,bool lazy) call_rcu_nocb() argument [all...] |
| H A D | tree.h | 295 long lazy_len; /* Length of buffered lazy callbacks. */ 504 unsigned long j, bool lazy); 506 rcu_callback_t func, unsigned long flags, bool lazy);
|
| H A D | tree.c | 3106 bool lazy; in __call_rcu_common() 3137 lazy = lazy_in && !rcu_async_should_hurry(); in __call_rcu_common() 3153 call_rcu_nocb(rdp, head, func, flags, lazy); 3165 * flush all lazy callbacks (including the new one) to the main ->cblist while 3194 * By default the callbacks are 'lazy' and are kept hidden from the main 3771 * if it's fully lazy. in rcu_barrier_entrain() 3815 * pending lazy RCU callbacks. in rcu_barrier() 3094 bool lazy; __call_rcu_common() local
|
| /linux/Documentation/mm/ |
| H A D | active_mm.rst | 5 Note, the mm_count refcount may no longer include the "lazy" users 7 with CONFIG_MMU_LAZY_TLB_REFCOUNT=n. Taking and releasing these lazy 63 and a "mm_count" counter that is the number of "lazy" users (ie anonymous 67 user exited on another CPU while a lazy user was still active, so you do 69 lazy users. That is often a short-lived state, because once that thread 74 more. "init_mm" should be considered just a "lazy context when no other
|
| /linux/drivers/opp/ |
| H A D | of.c | 139 list_del(&opp_table->lazy); in _opp_table_free_required_tables() 151 bool lazy = false; in _opp_table_alloc_required_tables() local 188 lazy = true; in _opp_table_alloc_required_tables() 192 if (lazy) { in _opp_table_alloc_required_tables() 198 list_add(&opp_table->lazy, &lazy_opp_tables); in _opp_table_alloc_required_tables() 355 list_for_each_entry_safe(opp_table, temp, &lazy_opp_tables, lazy) { in lazy_link_required_opp_table() 356 bool lazy = false; in lazy_link_required_opp_table() local 380 lazy = true; in lazy_link_required_opp_table() 390 lazy = false; in lazy_link_required_opp_table() 396 if (!lazy) { in lazy_link_required_opp_table() [all …]
|
| H A D | opp.h | 207 struct list_head node, lazy; member 268 return unlikely(!list_empty(&opp_table->lazy)); in lazy_linking_pending()
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | icp_qat_hw_20_comp.h | 69 __u16 lazy; member 102 QAT_FIELD_SET(val32, csr.lazy, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_UPPER()
|
| /linux/Documentation/arch/arm/ |
| H A D | kernel_mode_neon.rst | 30 The NEON/VFP register file is managed using lazy preserve (on UP systems) and 31 lazy restore (on both SMP and UP systems). This means that the register file is 45 mode will hit the lazy restore trap upon next use. This is handled by the
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_fence.h | 32 int nouveau_fence_wait(struct nouveau_fence *, bool lazy, bool intr);
|
| H A D | nouveau_fence.c | 323 nouveau_fence_wait(struct nouveau_fence *fence, bool lazy, bool intr) in nouveau_fence_wait() argument 327 if (!lazy) in nouveau_fence_wait()
|
| /linux/mm/ |
| H A D | vmalloc.c | 944 struct rb_list lazy; member 2363 if (RB_EMPTY_ROOT(&vn->lazy.root)) in __purge_vmap_area_lazy() 2366 spin_lock(&vn->lazy.lock); in __purge_vmap_area_lazy() 2367 WRITE_ONCE(vn->lazy.root.rb_node, NULL); in __purge_vmap_area_lazy() 2368 list_replace_init(&vn->lazy.head, &vn->purge_list); in __purge_vmap_area_lazy() 2369 spin_unlock(&vn->lazy.lock); in __purge_vmap_area_lazy() 2466 spin_lock(&vn->lazy.lock); in free_vmap_area_noflush() 2467 insert_vmap_area(va, &vn->lazy.root, &vn->lazy.head); in free_vmap_area_noflush() 2468 spin_unlock(&vn->lazy.lock); in free_vmap_area_noflush() 5206 spin_lock(&vn->lazy.lock); in show_purge_info() [all …]
|
| H A D | Kconfig | 1458 The architecture uses the lazy MMU mode. This allows changes to 1463 tristate "KUnit tests for the lazy MMU mode" if !KUNIT_ALL_TESTS 1468 Enable this option to check that the lazy MMU mode interface behaves
|
| /linux/tools/perf/Documentation/ |
| H A D | perf-probe.txt | 165 3) Define event based on source file with lazy pattern 176 …ine, and '%return' means that it probes function return. And ';PTN' means lazy matching pattern (s… 177 …ber or lazy matching by using 'SRC:ALN' or 'SRC;PTN' syntax, where 'SRC' is the source file path, … 229 The lazy line matching is similar to glob matching but ignoring spaces in both of pattern and targe…
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_fence.c | 229 int vmw_fence_obj_wait(struct vmw_fence_obj *fence, bool lazy, in vmw_fence_obj_wait() argument 464 ret = vmw_fence_obj_wait(fence, arg->lazy, true, timeout); in vmw_fence_obj_wait_ioctl()
|
| /linux/include/uapi/drm/ |
| H A D | vmwgfx_drm.h | 647 __s32 lazy; member
|
| /linux/Documentation/arch/parisc/ |
| H A D | registers.rst | 18 CR10 (CCR) lazy FPU saving*
|
| /linux/Documentation/filesystems/fuse/ |
| H A D | fuse.rst | 27 umounted. Note that detaching (or lazy umounting) the filesystem 213 filesystem is still attached (it hasn't been lazy unmounted)
|
| /linux/Documentation/filesystems/ |
| H A D | autofs-mount-control.rst | 23 Currently autofs uses "umount -l" (lazy umount) to clear active mounts 24 at restart. While using lazy umount works for most cases, anything that
|
| /linux/Documentation/trace/rv/ |
| H A D | monitor_sched.rst | 187 This is not valid for the *lazy* variant of the flag, which causes only
|
| /linux/Documentation/arch/powerpc/ |
| H A D | transactional_memory.rst | 94 Examples are glibc's getpid() and lazy symbol resolution.
|
| /linux/include/hyperv/ |
| H A D | hvgdk_mini.h | 684 u64 lazy : 1; member
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | numa_memory_policy.rst | 138 support allocation at fault time--a.k.a lazy allocation--so hugetlbfs 140 Although hugetlbfs segments now support lazy allocation, their support
|
| /linux/arch/sparc/lib/ |
| H A D | checksum_32.S | 411 addx %g5, %g0, %g5 ! I am now to lazy to optimize this (question it
|
| /linux/Documentation/locking/ |
| H A D | ww-mutex-design.rst | 350 The Wound-Wait preemption is implemented with a lazy-preemption scheme:
|