/linux/drivers/gpu/drm/ttm/ |
H A D | ttm_device.c | 242 spin_lock_init(&bdev->lru_lock); in ttm_device_init() 269 spin_lock(&bdev->lru_lock); in ttm_device_fini() 273 spin_unlock(&bdev->lru_lock); in ttm_device_fini() 285 spin_lock(&bdev->lru_lock); in ttm_device_clear_lru_dma_mappings() 294 spin_unlock(&bdev->lru_lock); in ttm_device_clear_lru_dma_mappings() 300 spin_lock(&bdev->lru_lock); in ttm_device_clear_lru_dma_mappings() 302 spin_unlock(&bdev->lru_lock); in ttm_device_clear_lru_dma_mappings()
|
H A D | ttm_bo_util.c | 925 spinlock_t *lru_lock = &curs->res_curs.man->bdev->lru_lock; in ttm_bo_lru_cursor_fini() local 928 spin_lock(lru_lock); in ttm_bo_lru_cursor_fini() 930 spin_unlock(lru_lock); in ttm_bo_lru_cursor_fini() 960 spinlock_t *lru_lock = &curs->res_curs.man->bdev->lru_lock; in __ttm_bo_lru_cursor_next() local 968 spin_lock(lru_lock); in __ttm_bo_lru_cursor_next() 995 spin_unlock(lru_lock); in __ttm_bo_lru_cursor_next() 1014 spin_lock(lru_lock); in __ttm_bo_lru_cursor_next() 1017 spin_unlock(lru_lock); in __ttm_bo_lru_cursor_next() 1099 spin_lock(&bo->bdev->lru_lock); in ttm_bo_shrink() 1101 spin_unlock(&bo->bdev->lru_lock); in ttm_bo_shrink() [all …]
|
/linux/Documentation/bpf/ |
H A D | map_lru_hash_update.dot | 13 lru_lock [shape=rectangle,fillcolor=2,label="LRU lock"] 151 local_lock -> lru_lock [style=invis] 152 lru_lock -> hash_lock [style=invis]
|
/linux/fs/gfs2/ |
H A D | glock.c | 70 static DEFINE_SPINLOCK(lru_lock); 186 spin_lock(&lru_lock); in gfs2_glock_free_later() 188 spin_unlock(&lru_lock); in gfs2_glock_free_later() 221 spin_lock(&lru_lock); in gfs2_glock_add_to_lru() 229 spin_unlock(&lru_lock); in gfs2_glock_add_to_lru() 234 spin_lock(&lru_lock); in gfs2_glock_remove_from_lru() 240 spin_unlock(&lru_lock); in gfs2_glock_remove_from_lru() 1983 __releases(&lru_lock) in gfs2_dispose_glock_lru() 1984 __acquires(&lru_lock) in gfs2_dispose_glock_lru() 2011 cond_resched_lock(&lru_lock); in gfs2_dispose_glock_lru() [all …]
|
/linux/mm/ |
H A D | vmscan.c | 1935 spin_unlock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1937 spin_lock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1959 spin_unlock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1962 spin_lock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1981 spin_unlock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1984 spin_lock_irq(&lruvec->lru_lock); in move_folios_to_lru() 2033 spin_lock_irq(&lruvec->lru_lock); in shrink_inactive_list() 2045 spin_unlock_irq(&lruvec->lru_lock); in shrink_inactive_list() 2053 spin_lock_irq(&lruvec->lru_lock); in shrink_inactive_list() 2143 spin_lock_irq(&lruvec->lru_lock); in shrink_active_list() [all …]
|
H A D | mmzone.c | 80 spin_lock_init(&lruvec->lru_lock); in lruvec_init()
|
H A D | swap.c | 242 __releases(lruvec->lru_lock) in lru_note_cost() 255 spin_unlock_irq(&lruvec->lru_lock); in lru_note_cost() 286 spin_unlock_irq(&lruvec->lru_lock); in lru_note_cost() 290 spin_lock_irq(&lruvec->lru_lock); in lru_note_cost()
|
H A D | memcontrol.c | 1208 spin_lock(&lruvec->lru_lock); in folio_lruvec_lock() 1230 spin_lock_irq(&lruvec->lru_lock); in folio_lruvec_lock_irq() 1254 spin_lock_irqsave(&lruvec->lru_lock, *flags); in folio_lruvec_lock_irqsave() 1267 * This function must be called under lru_lock, just before a page is added
|
H A D | compaction.c | 1080 * so avoid taking lru_lock and isolating it unnecessarily in an in isolate_migratepages_block() 1163 compact_lock_irqsave(&lruvec->lru_lock, &flags, cc); in isolate_migratepages_block() 1227 /* Avoid potential deadlock in freeing page under lru_lock */ in isolate_migratepages_block()
|
H A D | huge_memory.c | 3239 lockdep_assert_held(&lruvec->lru_lock); in unmap_huge_pmd_locked() 3879 * requires taking the lru_lock so we do the put_page in __folio_split()
|
/linux/include/drm/ttm/ |
H A D | ttm_device.h | 252 spinlock_t lru_lock; member
|
/linux/drivers/gpu/drm/xe/tests/ |
H A D | xe_bo.c | 521 spin_lock(&bo->ttm.bdev->lru_lock); in shrink_test_run_device() 523 spin_unlock(&bo->ttm.bdev->lru_lock); in shrink_test_run_device()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vram_mgr.c | 319 spin_lock(&man->bdev->lru_lock); in amdgpu_vram_mgr_do_reserve() 321 spin_unlock(&man->bdev->lru_lock); in amdgpu_vram_mgr_do_reserve()
|
/linux/Documentation/locking/ |
H A D | lockstat.rst | 195 …&(&zone->lru_lock)->rlock: 93000 94934 0.16 59.18 188253.…
|
/linux/include/linux/ |
H A D | mm_inline.h | 44 lockdep_assert_held(&lruvec->lru_lock); in __update_lru_size()
|
/linux/Documentation/translations/it_IT/locking/ |
H A D | lockstat.rst | 221 …&(&zone->lru_lock)->rlock: 93000 94934 0.16 59.18 188253.…
|
/linux/Documentation/trace/ |
H A D | events-kmem.rst | 72 contention on the lruvec->lru_lock.
|
/linux/Documentation/admin-guide/cgroup-v1/ |
H A D | memcg_test.rst | 138 each LRU handled under a single lru_lock for that memcg and node.
|
/linux/Documentation/mm/ |
H A D | unevictable-lru.rst | 321 mlock folio batch, to batch up the rest of the work to be done under lru_lock by 382 under lru_lock by __munlock_folio(). __munlock_folio() decrements the 494 under lru_lock by __munlock_folio(). __munlock_folio() decrements the
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_bo.c | 1448 spin_lock(&ttm_bo->bdev->lru_lock); in xe_ttm_bo_lock_in_destructor() 1450 spin_unlock(&ttm_bo->bdev->lru_lock); in xe_ttm_bo_lock_in_destructor()
|
H A D | xe_vm.c | 558 spin_lock(&vm->xe->ttm.lru_lock); in preempt_rebind_work_func() 560 spin_unlock(&vm->xe->ttm.lru_lock); in preempt_rebind_work_func()
|