/linux/drivers/gpu/drm/ttm/ |
H A D | ttm_bo.c | 74 * object. This function must be called with struct ttm_global::lru_lock 108 spin_lock(&bo->bdev->lru_lock); in ttm_bo_set_bulk_move() 114 spin_unlock(&bo->bdev->lru_lock); in ttm_bo_set_bulk_move() 206 * the resv object while holding the lru_lock. in ttm_bo_individualize_resv() 208 spin_lock(&bo->bdev->lru_lock); in ttm_bo_individualize_resv() 210 spin_unlock(&bo->bdev->lru_lock); in ttm_bo_individualize_resv() 284 spin_lock(&bo->bdev->lru_lock); in ttm_bo_release() 300 spin_unlock(&bo->bdev->lru_lock); in ttm_bo_release() 455 spin_lock(&bdev->lru_lock); in ttm_bo_evict_first() 467 spin_unlock(&bdev->lru_lock); in ttm_bo_evict_first() [all...] |
H A D | ttm_bo_util.c | 925 spinlock_t *lru_lock = &curs->res_curs.man->bdev->lru_lock; in ttm_bo_lru_cursor_fini() local 928 spin_lock(lru_lock); in ttm_bo_lru_cursor_fini() 930 spin_unlock(lru_lock); in ttm_bo_lru_cursor_fini() 960 spinlock_t *lru_lock = &curs->res_curs.man->bdev->lru_lock; in __ttm_bo_lru_cursor_next() local 968 spin_lock(lru_lock); in __ttm_bo_lru_cursor_next() 995 spin_unlock(lru_lock); in __ttm_bo_lru_cursor_next() 1014 spin_lock(lru_lock); in __ttm_bo_lru_cursor_next() 1017 spin_unlock(lru_lock); in __ttm_bo_lru_cursor_next() [all...] |
/linux/Documentation/bpf/ |
H A D | map_lru_hash_update.dot | 13 lru_lock [shape=rectangle,fillcolor=2,label="LRU lock"] 151 local_lock -> lru_lock [style=invis] 152 lru_lock -> hash_lock [style=invis]
|
/linux/mm/ |
H A D | swap.c | 242 __releases(lruvec->lru_lock) in lru_note_cost_unlock_irq() 255 spin_unlock_irq(&lruvec->lru_lock); in lru_note_cost_unlock_irq() 286 spin_unlock_irq(&lruvec->lru_lock); in lru_note_cost_unlock_irq() 290 spin_lock_irq(&lruvec->lru_lock); in lru_note_cost_unlock_irq()
|
H A D | mmzone.c | 80 spin_lock_init(&lruvec->lru_lock); in lruvec_init()
|
H A D | vmscan.c | 865 * Context: lru_lock must not be held, interrupts must be enabled. 1719 * lruvec->lru_lock is heavily contended. Some of the functions that 1848 * (2) The lru_lock must not be held. 1935 spin_unlock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1937 spin_lock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1947 * !lru //skip lru_lock in move_folios_to_lru() 1959 spin_unlock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1962 spin_lock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1981 spin_unlock_irq(&lruvec->lru_lock); in move_folios_to_lru() 1984 spin_lock_irq(&lruvec->lru_lock); in move_folios_to_lru() [all...] |
H A D | memcontrol.c | 1208 spin_lock(&lruvec->lru_lock); in folio_lruvec_lock() 1230 spin_lock_irq(&lruvec->lru_lock); in folio_lruvec_lock_irq() 1254 spin_lock_irqsave(&lruvec->lru_lock, *flags); in folio_lruvec_lock_irqsave() 1267 * This function must be called under lru_lock, just before a page is added
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_exec.c | 329 spin_lock(&xe->ttm.lru_lock); in xe_exec_ioctl() 331 spin_unlock(&xe->ttm.lru_lock); in xe_exec_ioctl()
|
H A D | xe_bo.c | 1443 * lru_lock, which will immediately be unlocked again since in xe_ttm_bo_lock_in_destructor() 1447 spin_lock(&ttm_bo->bdev->lru_lock); in xe_ttm_bo_lock_in_destructor() 1449 spin_unlock(&ttm_bo->bdev->lru_lock); in xe_ttm_bo_lock_in_destructor()
|
/linux/fs/gfs2/ |
H A D | glock.c | 70 static DEFINE_SPINLOCK(lru_lock); 186 spin_lock(&lru_lock); in gfs2_glock_free_later() 188 spin_unlock(&lru_lock); in gfs2_glock_free_later() 221 spin_lock(&lru_lock); in gfs2_glock_add_to_lru() 229 spin_unlock(&lru_lock); in gfs2_glock_add_to_lru() 234 spin_lock(&lru_lock); in gfs2_glock_remove_from_lru() 240 spin_unlock(&lru_lock); in gfs2_glock_remove_from_lru() 1976 * Must be called under the lru_lock, but may drop and retake this 1977 * lock. While the lru_lock is dropped, entries may vanish from the 1983 __releases(&lru_lock) in gfs2_dispose_glock_lru() [all...] |
/linux/include/drm/ttm/ |
H A D | ttm_device.h | 250 * @lru_lock: Protection for the per manager LRU and ddestroy lists. 252 spinlock_t lru_lock; member
|
/linux/drivers/gpu/drm/xe/tests/ |
H A D | xe_bo.c | 521 spin_lock(&bo->ttm.bdev->lru_lock); in shrink_test_run_device() 523 spin_unlock(&bo->ttm.bdev->lru_lock); in shrink_test_run_device()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vram_mgr.c | 319 spin_lock(&man->bdev->lru_lock); in amdgpu_vram_mgr_do_reserve() 321 spin_unlock(&man->bdev->lru_lock); in amdgpu_vram_mgr_do_reserve()
|
H A D | amdgpu_vm.c | 499 spin_lock(&adev->mman.bdev.lru_lock); in amdgpu_vm_move_to_lru_tail() 501 spin_unlock(&adev->mman.bdev.lru_lock); in amdgpu_vm_move_to_lru_tail()
|
/linux/Documentation/locking/ |
H A D | lockstat.rst | 195 &(&zone->lru_lock)->rlock: 93000 94934 0.16 59.18 188253.78 1.98 1199912 3809894 0.15 391.40 3559518.81 0.93
|
/linux/include/linux/ |
H A D | swap.h | 381 __releases(lruvec->lru_lock);
|
H A D | mm_inline.h | 44 lockdep_assert_held(&lruvec->lru_lock); in __update_lru_size()
|
H A D | mmzone.h | 656 /* per lruvec lru_lock for memcg */ 657 spinlock_t lru_lock; member
|
/linux/Documentation/translations/it_IT/locking/ |
H A D | lockstat.rst | 221 &(&zone->lru_lock)->rlock: 93000 94934 0.16 59.18 188253.78 1.98 1199912 3809894 0.15 391.40 3559518.81 0.93
|
/linux/Documentation/trace/ |
H A D | events-kmem.rst | 72 contention on the lruvec->lru_lock.
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_ttm.c | 981 spin_lock(&bo->bdev->lru_lock); in i915_ttm_adjust_lru() 1010 spin_unlock(&bo->bdev->lru_lock); in i915_ttm_adjust_lru()
|
/linux/Documentation/admin-guide/cgroup-v1/ |
H A D | memcg_test.rst | 138 each LRU handled under a single lru_lock for that memcg and node.
|
H A D | memory.rst | 316 lruvec->lru_lock. 319 lruvec->lru_lock; the folio LRU flag is cleared before 320 isolating a page from its LRU under lruvec->lru_lock.
|
/linux/Documentation/mm/ |
H A D | unevictable-lru.rst | 321 mlock folio batch, to batch up the rest of the work to be done under lru_lock by 382 under lru_lock by __munlock_folio(). __munlock_folio() decrements the 494 under lru_lock by __munlock_folio(). __munlock_folio() decrements the
|
H A D | process_addrs.rst | 431 lruvec->lru_lock (in folio_lruvec_lock_irq) 479 ->lruvec->lru_lock (follow_page_mask->mark_page_accessed) 480 ->lruvec->lru_lock (check_pte_range->folio_isolate_lru)
|