| /linux/drivers/gpu/drm/ |
| H A D | drm_gpuvm.c | 1107 INIT_LIST_HEAD(&gpuvm->evict.list); in drm_gpuvm_init() 1108 spin_lock_init(&gpuvm->evict.lock); in drm_gpuvm_init() 1151 drm_WARN(gpuvm->drm, !list_empty(&gpuvm->evict.list), in drm_gpuvm_fini() 1258 drm_gpuvm_bo_list_add(vm_bo, evict, false); in drm_gpuvm_prepare_objects_locked() 1470 LIST_HEAD(evict); in __drm_gpuvm_validate() 1473 for_each_vm_bo_in_list(gpuvm, evict, &evict, vm_bo) { in __drm_gpuvm_validate() 1480 restore_vm_bo_list(gpuvm, evict); in __drm_gpuvm_validate() 1494 list_for_each_entry_safe(vm_bo, next, &gpuvm->evict.list, in drm_gpuvm_validate_locked() 1495 list.entry.evict) { in drm_gpuvm_validate_locked() 1505 drm_gpuvm_bo_list_del_init(vm_bo, evict, false); in drm_gpuvm_validate_locked() [all …]
|
| H A D | drm_gem_vram_helper.c | 504 bool evict, in drm_gem_vram_bo_driver_move() argument 767 bool evict, in bo_driver_move() argument 787 return drm_gem_vram_bo_driver_move(gbo, evict, ctx, new_mem); in bo_driver_move()
|
| H A D | drm_gem.c | 1723 if (obj->funcs->evict) in drm_gem_evict_locked() 1724 return obj->funcs->evict(obj); in drm_gem_evict_locked()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_bo.c | 548 bool force = false, evict = false; in nouveau_bo_pin_locked() local 558 evict = true; in nouveau_bo_pin_locked() 563 bool error = evict; in nouveau_bo_pin_locked() 586 if (evict) { in nouveau_bo_pin_locked() 936 nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, in nouveau_bo_move_m2mf() argument 982 ret = ttm_bo_move_accel_cleanup(bo, &fence->base, evict, false, in nouveau_bo_move_m2mf() 1144 nouveau_bo_move(struct ttm_buffer_object *bo, bool evict, in nouveau_bo_move() argument 1162 drm_gpuvm_bo_gem_evict(obj, evict); in nouveau_bo_move() 1207 ret = nouveau_bo_move_m2mf(bo, evict, ctx, in nouveau_bo_move() 1227 drm_gpuvm_bo_gem_evict(obj, !evict); in nouveau_bo_move()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_ttm_move.h | 33 int i915_ttm_move(struct ttm_buffer_object *bo, bool evict,
|
| H A D | i915_gem_ttm_move.c | 570 int i915_ttm_move(struct ttm_buffer_object *bo, bool evict, in i915_ttm_move() argument 660 if (I915_SELFTEST_ONLY(evict && fail_gpu_migration)) in i915_ttm_move() 663 ret = ttm_bo_move_accel_cleanup(bo, migration_fence, evict, in i915_ttm_move()
|
| /linux/drivers/gpu/drm/i915/selftests/ |
| H A D | i915_mock_selftests.h | 32 selftest(evict, i915_gem_evict_mock_selftests)
|
| H A D | i915_live_selftests.h | 39 selftest(evict, i915_gem_evict_live_selftests)
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | mmu_rb.h | 27 int (*evict)(void *ops_arg, struct mmu_rb_node *mnode, member
|
| H A D | mmu_rb.c | 244 if (handler->ops->evict(handler->ops_arg, rbnode, evict_arg, in hfi1_mmu_rb_evict()
|
| H A D | pin_system.c | 31 .evict = sdma_rb_evict,
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_ttm.c | 134 bool evict, in radeon_move_blit() argument 183 r = ttm_bo_move_accel_cleanup(bo, &fence->base, evict, false, new_mem); in radeon_move_blit() 188 static int radeon_bo_move(struct ttm_buffer_object *bo, bool evict, in radeon_bo_move() argument 238 r = radeon_move_blit(bo, evict, new_mem, old_mem); in radeon_bo_move()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_ttm.c | 383 bool evict, in amdgpu_move_blit() argument 430 r = ttm_bo_move_accel_cleanup(bo, fence, evict, true, new_mem); in amdgpu_move_blit() 498 static int amdgpu_bo_move(struct ttm_buffer_object *bo, bool evict, in amdgpu_bo_move() argument 520 amdgpu_bo_move_notify(bo, evict, new_mem); in amdgpu_bo_move() 527 amdgpu_bo_move_notify(bo, evict, new_mem); in amdgpu_bo_move() 539 amdgpu_bo_move_notify(bo, evict, new_mem); in amdgpu_bo_move() 556 amdgpu_bo_move_notify(bo, evict, new_mem); in amdgpu_bo_move() 582 amdgpu_bo_move_notify(bo, evict, new_mem); in amdgpu_bo_move() 584 r = amdgpu_move_blit(bo, evict, new_mem, old_mem); in amdgpu_bo_move() 602 if (evict) in amdgpu_bo_move()
|
| /linux/Documentation/mm/ |
| H A D | multigen_lru.rst | 106 evict or protect. The desired effect is to balance refault percentages 133 evict from, it first compares ``min_seq[]`` to select the older type. 215 decides which type to evict when both types are available from the 269 to select types to evict and tiers to protect.
|
| /linux/drivers/gpu/drm/qxl/ |
| H A D | qxl_ttm.c | 137 static int qxl_bo_move(struct ttm_buffer_object *bo, bool evict, in qxl_bo_move() argument
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | multigen_lru.rst | 146 Users can write the following command to ``lru_gen`` to evict 156 of anonymous memory. ``nr_to_reclaim`` limits the number of pages to evict.
|
| H A D | zswap.rst | 53 evict pages from its own compressed pool on an LRU basis and write them back to
|
| /linux/Documentation/gpu/ |
| H A D | drm-compute.rst | 32 not to evict when creating a new job (any kind). If all of userspace opts in
|
| H A D | drm-vm-bind-locking.rst | 219 The gpu_vm_bos marked for eviction are put on the gpu_vm's evict list, 222 the gpu_vm's dma_resv protecting the gpu_vm's evict list is locked. 253 object is bound to need to be put on their gpu_vm's evict list. 327 gpu_vm's evict list and external objects lists. However, there are
|
| /linux/drivers/gpu/drm/ttm/tests/ |
| H A D | ttm_kunit_helpers.c | 63 static int mock_move(struct ttm_buffer_object *bo, bool evict, in mock_move() argument
|
| /linux/drivers/gpu/drm/ttm/ |
| H A D | ttm_bo_util.c | 702 bool evict, in ttm_bo_move_accel_cleanup() argument 712 if (!evict) in ttm_bo_move_accel_cleanup()
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_ttm_buffer.c | 497 bool evict, in vmw_move() argument
|
| /linux/drivers/gpu/drm/loongson/ |
| H A D | lsdc_ttm.c | 160 bool evict, in lsdc_bo_move() argument
|
| /linux/fs/ |
| H A D | inode.c | 823 static void evict(struct inode *inode) in evict() function 888 evict(inode); in dispose_list() 1966 evict(inode); in iput_final()
|
| /linux/Documentation/bpf/ |
| H A D | map_hash.rst | 29 will automatically evict the least recently used entries when the hash
|