| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_bo_evict.c | 175 &xe->pinned.late.evicted, xe_bo_evict_pinned); in xe_bo_evict_all() 188 &xe->pinned.early.evicted, in xe_bo_evict_all() 228 return xe_bo_apply_to_pinned(xe, &xe->pinned.early.evicted, in xe_bo_restore_early() 248 ret = xe_bo_apply_to_pinned(xe, &xe->pinned.late.evicted, in xe_bo_restore_late() 346 INIT_LIST_HEAD(&xe->pinned.early.evicted); in xe_bo_pinned_init() 348 INIT_LIST_HEAD(&xe->pinned.late.evicted); in xe_bo_pinned_init()
|
| H A D | xe_device_types.h | 328 /** @pinned.early.evicted: pinned BO that have been evicted */ 329 struct list_head evicted; 335 /** @pinned.late.evicted: pinned BO that have been evicted */ 336 struct list_head evicted; 494 struct list_head evicted; global() member 501 struct list_head evicted; global() member
|
| H A D | xe_vm.c | 342 vm_bo->evicted = false; in xe_gpuvm_validate() 353 vm_bo->evicted = false; 364 * Validates all evicted gem objects and rebinds their vmas. Note that in xe_vm_validate_rebind() 1914 * but even so, the unbind when evicted would still in find_ufence_get()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_vm.h | 333 /* buffers that requested this placement but are currently evicted */ 334 uint64_t evicted; member 358 * evicted -> relocated (PDs, PTs) or moved (per VM BOs) -> idle 362 struct list_head evicted; member 544 void amdgpu_vm_bo_invalidate(struct amdgpu_bo *bo, bool evicted); 549 bool evicted);
|
| H A D | amdgpu_vm.c | 156 * amdgpu_vm_bo_evicted - vm_bo is evicted in amdgpu_vm_bo_evicted() 158 * @vm_bo: vm_bo which is evicted in amdgpu_vm_bo_evicted() 172 list_move(&vm_bo->vm_status, &vm->evicted); in amdgpu_vm_bo_moved() 174 list_move_tail(&vm_bo->vm_status, &vm->evicted); in amdgpu_vm_bo_moved() 226 * amdgpu_vm_bo_evicted_user - vm_bo is evicted 228 * @vm_bo: vm_bo which is evicted 390 vm->stats[bo_memtype].evicted += size; in amdgpu_vm_update_stats() 458 * is currently evicted. add the bo to the evicted list to make sure it 591 * amdgpu_vm_validate - validate evicted BO 2281 amdgpu_vm_bo_invalidate(struct amdgpu_bo * bo,bool evicted) amdgpu_vm_bo_invalidate() argument 2316 amdgpu_vm_bo_move(struct amdgpu_bo * bo,struct ttm_resource * new_mem,bool evicted) amdgpu_vm_bo_move() argument [all...] |
| /linux/Documentation/admin-guide/mm/ |
| H A D | multigen_lru.rst | 72 ``N`` milliseconds from getting evicted. The OOM killer is triggered 92 evicted generations in this file. 153 the active list) and therefore cannot be evicted. ``swappiness``
|
| /linux/drivers/gpu/drm/ttm/ |
| H A D | ttm_bo.c | 503 /** @evicted: Number of successful evictions. */ 504 unsigned long evicted; 538 evict_walk->evicted++; in ttm_bo_evict_cb() 600 evict_walk.evicted = 0; in ttm_bo_evict_alloc() 602 } while (!lret && evict_walk.evicted); in ttm_bo_evict_alloc() 621 * Make sure the buffer is not evicted any more during memory pressure. 641 * Allows the buffer object to be evicted again during memory pressure. 505 unsigned long evicted; global() member
|
| /linux/Documentation/gpu/ |
| H A D | drm-vm-bind-locking.rst | 99 Furthermore, it typically protects the gpu_vm's list of evicted and 109 to be able to update the gpu_vm evicted- and external object 259 option is to just mark the gpu_vm_bos of the evicted gem object with 260 an ``evicted`` bool that is inspected before the next time the 261 corresponding gpu_vm evicted list needs to be traversed. For example, when 264 gpu_vm_bo marked evicted, can then be added to the gpu_vm's list of 265 evicted gpu_vm_bos. The ``evicted`` bool is formally protected by the
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_device_queue_manager.c | 648 * Eviction state logic: mark all queues as evicted, even ones in create_queue_nocpsch() 652 q->properties.is_evicted = !!qpd->evicted; in create_queue_nocpsch() 1149 if (qpd->evicted++ > 0) /* already evicted, do nothing */ in evict_process_queues_nocpsch() 1157 /* Mark all queues as evicted. Deactivate all active queues on in evict_process_queues_nocpsch() 1199 if (qpd->evicted++ > 0) /* already evicted, do nothing */ in evict_process_queues_cpsch() 1217 /* Mark all queues as evicted. Deactivate all active queues on in evict_process_queues_cpsch() 1268 if (WARN_ON_ONCE(!qpd->evicted)) /* already restored, do nothing */ in restore_process_queues_nocpsch() 1270 if (qpd->evicted > in restore_process_queues_nocpsch() [all...] |
| H A D | kfd_priv.h | 486 * @is_evicted: Defines if the queue is evicted. Only active queues 487 * are evicted, rendering them inactive. 674 unsigned int evicted; /* eviction counter, 0=active */ member 724 /* Approx. wait time before attempting to restore evicted BOs */
|
| H A D | kfd_process.c | 60 /* Ordered, single-threaded workqueue for restoring evicted 64 * remain evicted indefinitely. 1721 pdd->qpd.evicted = 0; in kfd_create_process_device_data() 2006 * we would like to set all the queues to be in evicted state to prevent in kfd_process_evict_queues() 2178 * before KFD BOs are unreserved. If not, the process can be evicted in restore_process_worker()
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | cache-policies.rst | 15 e.g. to start writing back dirty blocks that are going to be evicted
|
| /linux/mm/ |
| H A D | workingset.c | 45 * are evicted from the inactive list every time before another access 93 * 1. When a page is finally evicted from memory, the number of 133 * had (R - E) more page slots, the page would not have been evicted 157 * used once will be evicted from memory. 178 * slot of the evicted page. This is called a shadow entry. 259 atomic_long_add(delta, &lrugen->evicted[hist][type][tier]); 265 * Tests if the shadow entry is for a folio that was recently evicted. in lru_gen_test_recent() 380 * @folio: the folio being evicted 383 * of the evicted @folio so that a later refault can be detected. in workingset_eviction() 413 * recently evicted [all...] |
| H A D | vmscan.c | 2146 * are not likely to be evicted by use-once streaming in shrink_active_list() 3127 * The P term is refaulted/(evicted+protected) from a tier in the generation in iterate_mm_list_nowalk() 3128 * currently being evicted; the I term is the exponential moving average of the in iterate_mm_list_nowalk() 3129 * P term over the generations previously evicted, using the smoothing factor 3164 atomic_long_read(&lrugen->evicted[hist][type][i]); in read_ctrl_pos() 3192 atomic_long_read(&lrugen->evicted[hist][type][tier]); in reset_ctrl_pos() 3198 atomic_long_set(&lrugen->evicted[hist][type][tier], 0); in reset_ctrl_pos() 4900 * If too many file cache in the coldest generation can't be evicted in try_to_shrink_lruvec() 5386 n[1] = atomic_long_read(&lrugen->evicted[hist][type][tier]); in lru_gen_seq_show_full()
|
| H A D | Kconfig | 1389 from evicted generations for debugging purpose.
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_gpuvm.c | 1257 if (vm_bo->evicted) in drm_gpuvm_prepare_objects_locked() 1504 if (!vm_bo->evicted) in drm_gpuvm_validate_locked() 1965 vm_bo->evicted = evict; in drm_gpuvm_bo_evict()
|
| /linux/Documentation/mm/ |
| H A D | multigen_lru.rst | 87 pages can be evicted regardless of swap constraints. These three 151 from getting evicted. The OOM killer is triggered if this working set
|
| H A D | unevictable-lru.rst | 127 may be evicted. To prevent eviction of any such pages, the AS_UNEVICTABLE
|
| /linux/drivers/firmware/efi/ |
| H A D | Kconfig | 180 have been evicted, since otherwise it will trigger even on clean
|
| /linux/include/linux/ |
| H A D | mmzone.h | 482 * comparisons of refaulted/(evicted+protected) from the first tier and the rest 541 /* whether to keep historical stats from evicted generations */ 573 /* the exponential moving average of evicted+protected */ 578 atomic_long_t evicted[NR_HIST_GENS][ANON_AND_FILE][MAX_NR_TIERS]; 508 atomic_long_t evicted[NR_HIST_GENS][ANON_AND_FILE][MAX_NR_TIERS]; global() member
|
| /linux/Documentation/arch/x86/ |
| H A D | sgx.rst | 64 number for a page evicted from the EPC.
|
| /linux/Documentation/filesystems/ |
| H A D | vfat.rst | 207 evicted from the inode cache. However, it means that operations
|
| /linux/init/ |
| H A D | Kconfig | 2030 pages marked for writeback, (recently) evicted pages).
|