Symbol: lru
arch/mips/mm/cerr-sb1.c
326
uint8_t lru;
arch/mips/mm/cerr-sb1.c
348
lru = (taghi >> 14) & 0xff;
arch/mips/mm/cerr-sb1.c
352
(lru & 0x3),
arch/mips/mm/cerr-sb1.c
353
((lru >> 2) & 0x3),
arch/mips/mm/cerr-sb1.c
354
((lru >> 4) & 0x3),
arch/mips/mm/cerr-sb1.c
355
((lru >> 6) & 0x3));
arch/mips/mm/cerr-sb1.c
481
uint8_t ecc, lru;
arch/mips/mm/cerr-sb1.c
503
lru = (taghi >> 14) & 0xff;
arch/mips/mm/cerr-sb1.c
507
(lru & 0x3),
arch/mips/mm/cerr-sb1.c
508
((lru >> 2) & 0x3),
arch/mips/mm/cerr-sb1.c
509
((lru >> 4) & 0x3),
arch/mips/mm/cerr-sb1.c
510
((lru >> 6) & 0x3));
arch/x86/mm/fault.c
275
list_for_each_entry(page, &pgd_list, lru) {
arch/x86/mm/init_64.c
149
list_for_each_entry(page, &pgd_list, lru) {
arch/x86/mm/init_64.c
190
list_for_each_entry(page, &pgd_list, lru) {
arch/x86/mm/pat/set_memory.c
1297
list_for_each_entry(page, &pgd_list, lru) {
arch/x86/mm/pat/set_memory.c
889
list_for_each_entry(page, &pgd_list, lru) {
arch/x86/xen/mmu_pv.c
836
list_for_each_entry(page, &pgd_list, lru) {
arch/x86/xen/mmu_pv.c
947
list_for_each_entry(page, &pgd_list, lru) {
block/blk-mq-tag.c
592
page = list_first_entry(&tags->page_list, struct page, lru);
block/blk-mq-tag.c
593
list_del_init(&page->lru);
block/blk-mq.c
3458
list_for_each_entry(page, &tags->page_list, lru) {
block/blk-mq.c
3640
list_add_tail(&page->lru, &tags->page_list);
drivers/android/binder_alloc.c
1069
int lru = 0;
drivers/android/binder_alloc.c
1085
lru++;
drivers/android/binder_alloc.c
1089
seq_printf(m, " pages: %d:%d:%d\n", active, lru, free);
drivers/android/binder_alloc.c
1135
struct list_lru_one *lru,
drivers/android/binder_alloc.c
1137
__must_hold(&lru->lock)
drivers/android/binder_alloc.c
1139
struct binder_shrinker_mdata *mdata = container_of(item, typeof(*mdata), lru);
drivers/android/binder_alloc.c
1182
list_lru_isolate(lru, item);
drivers/android/binder_alloc.c
1183
spin_unlock(&lru->lock);
drivers/android/binder_alloc.c
300
INIT_LIST_HEAD(&mdata->lru);
drivers/android/binder_alloc.h
125
struct list_lru_one *lru,
drivers/android/binder_alloc.h
67
struct list_head lru;
drivers/android/binder_alloc.h
78
return &mdata->lru;
drivers/block/xen-blkfront.c
1221
list_for_each_entry_safe(indirect_page, n, &rinfo->indirect_pages, lru) {
drivers/block/xen-blkfront.c
1222
list_del(&indirect_page->lru);
drivers/block/xen-blkfront.c
1502
list_add(&indirect_page->lru, &rinfo->indirect_pages);
drivers/block/xen-blkfront.c
2204
list_add(&indirect_page->lru, &rinfo->indirect_pages);
drivers/block/xen-blkfront.c
2240
list_for_each_entry_safe(indirect_page, n, &rinfo->indirect_pages, lru) {
drivers/block/xen-blkfront.c
2241
list_del(&indirect_page->lru);
drivers/block/xen-blkfront.c
417
struct page, lru);
drivers/block/xen-blkfront.c
418
list_del(&indirect_page->lru);
drivers/dma-buf/heaps/system_heap.c
382
list_add_tail(&page->lru, &pages);
drivers/dma-buf/heaps/system_heap.c
393
list_for_each_entry_safe(page, tmp_page, &pages, lru) {
drivers/dma-buf/heaps/system_heap.c
396
list_del(&page->lru);
drivers/dma-buf/heaps/system_heap.c
420
list_for_each_entry_safe(page, tmp_page, &pages, lru)
drivers/gpu/drm/drm_gem.c
1529
drm_gem_lru_init(struct drm_gem_lru *lru, struct mutex *lock)
drivers/gpu/drm/drm_gem.c
1531
lru->lock = lock;
drivers/gpu/drm/drm_gem.c
1532
lru->count = 0;
drivers/gpu/drm/drm_gem.c
1533
INIT_LIST_HEAD(&lru->list);
drivers/gpu/drm/drm_gem.c
1540
obj->lru->count -= obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem.c
1541
WARN_ON(obj->lru->count < 0);
drivers/gpu/drm/drm_gem.c
1543
obj->lru = NULL;
drivers/gpu/drm/drm_gem.c
1556
struct drm_gem_lru *lru = obj->lru;
drivers/gpu/drm/drm_gem.c
1558
if (!lru)
drivers/gpu/drm/drm_gem.c
1561
mutex_lock(lru->lock);
drivers/gpu/drm/drm_gem.c
1563
mutex_unlock(lru->lock);
drivers/gpu/drm/drm_gem.c
1576
drm_gem_lru_move_tail_locked(struct drm_gem_lru *lru, struct drm_gem_object *obj)
drivers/gpu/drm/drm_gem.c
1578
lockdep_assert_held_once(lru->lock);
drivers/gpu/drm/drm_gem.c
1580
if (obj->lru)
drivers/gpu/drm/drm_gem.c
1583
lru->count += obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem.c
1584
list_add_tail(&obj->lru_node, &lru->list);
drivers/gpu/drm/drm_gem.c
1585
obj->lru = lru;
drivers/gpu/drm/drm_gem.c
1600
drm_gem_lru_move_tail(struct drm_gem_lru *lru, struct drm_gem_object *obj)
drivers/gpu/drm/drm_gem.c
1602
mutex_lock(lru->lock);
drivers/gpu/drm/drm_gem.c
1603
drm_gem_lru_move_tail_locked(lru, obj);
drivers/gpu/drm/drm_gem.c
1604
mutex_unlock(lru->lock);
drivers/gpu/drm/drm_gem.c
1625
drm_gem_lru_scan(struct drm_gem_lru *lru,
drivers/gpu/drm/drm_gem.c
1635
drm_gem_lru_init(&still_in_lru, lru->lock);
drivers/gpu/drm/drm_gem.c
1637
mutex_lock(lru->lock);
drivers/gpu/drm/drm_gem.c
1640
obj = list_first_entry_or_null(&lru->list, typeof(*obj), lru_node);
drivers/gpu/drm/drm_gem.c
1660
mutex_unlock(lru->lock);
drivers/gpu/drm/drm_gem.c
1683
WARN_ON(obj->lru == &still_in_lru);
drivers/gpu/drm/drm_gem.c
1684
WARN_ON(obj->lru == lru);
drivers/gpu/drm/drm_gem.c
1694
mutex_lock(lru->lock);
drivers/gpu/drm/drm_gem.c
1702
obj->lru = lru;
drivers/gpu/drm/drm_gem.c
1703
list_splice_tail(&still_in_lru.list, &lru->list);
drivers/gpu/drm/drm_gem.c
1704
lru->count += still_in_lru.count;
drivers/gpu/drm/drm_gem.c
1706
mutex_unlock(lru->lock);
drivers/gpu/drm/i915/i915_gpu_error.c
1074
list_for_each_entry_safe(page, n, &vma->page_list, lru) {
drivers/gpu/drm/i915/i915_gpu_error.c
1075
list_del_init(&page->lru);
drivers/gpu/drm/i915/i915_gpu_error.c
1250
list_for_each_entry_safe_reverse(page, n, &dst->page_list, lru) {
drivers/gpu/drm/i915/i915_gpu_error.c
1251
list_del_init(&page->lru);
drivers/gpu/drm/i915/i915_gpu_error.c
297
list_add_tail(&page->lru, &dst->page_list);
drivers/gpu/drm/i915/i915_gpu_error.c
411
list_add_tail(&virt_to_page(ptr)->lru, &dst->page_list);
drivers/gpu/drm/i915/i915_gpu_error.c
632
list_for_each_entry(page, &vma->page_list, lru) {
drivers/gpu/drm/i915/i915_gpu_error.c
637
if (page == list_last_entry(&vma->page_list, typeof(*page), lru))
drivers/gpu/drm/msm/msm_drv.c
131
mutex_init(&priv->lru.lock);
drivers/gpu/drm/msm/msm_drv.c
132
drm_gem_lru_init(&priv->lru.unbacked, &priv->lru.lock);
drivers/gpu/drm/msm/msm_drv.c
133
drm_gem_lru_init(&priv->lru.pinned, &priv->lru.lock);
drivers/gpu/drm/msm/msm_drv.c
134
drm_gem_lru_init(&priv->lru.willneed, &priv->lru.lock);
drivers/gpu/drm/msm/msm_drv.c
135
drm_gem_lru_init(&priv->lru.dontneed, &priv->lru.lock);
drivers/gpu/drm/msm/msm_drv.c
143
might_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_drv.h
160
} lru;
drivers/gpu/drm/msm/msm_gem.c
1265
drm_gem_lru_move_tail(&priv->lru.unbacked, obj);
drivers/gpu/drm/msm/msm_gem.c
1319
drm_gem_lru_move_tail(&priv->lru.pinned, obj);
drivers/gpu/drm/msm/msm_gem.c
152
drm_gem_lru_move_tail_locked(&priv->lru.pinned, obj);
drivers/gpu/drm/msm/msm_gem.c
154
drm_gem_lru_move_tail_locked(&priv->lru.willneed, obj);
drivers/gpu/drm/msm/msm_gem.c
158
drm_gem_lru_move_tail_locked(&priv->lru.dontneed, obj);
drivers/gpu/drm/msm/msm_gem.c
172
drm_gem_lru_move_tail_locked(&priv->lru.unbacked, obj);
drivers/gpu/drm/msm/msm_gem.c
182
mutex_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
184
mutex_unlock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
290
drm_gem_lru_move_tail_locked(&priv->lru.pinned, obj);
drivers/gpu/drm/msm/msm_gem.c
297
mutex_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
299
mutex_unlock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
495
mutex_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
499
mutex_unlock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
802
mutex_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
814
mutex_unlock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
839
mutex_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem.c
842
mutex_unlock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem_shrinker.c
170
struct drm_gem_lru *lru;
drivers/gpu/drm/msm/msm_gem_shrinker.c
177
{ &priv->lru.dontneed, purge, true },
drivers/gpu/drm/msm/msm_gem_shrinker.c
178
{ &priv->lru.willneed, evict, can_swap() },
drivers/gpu/drm/msm/msm_gem_shrinker.c
179
{ &priv->lru.dontneed, active_purge, can_block(sc) },
drivers/gpu/drm/msm/msm_gem_shrinker.c
180
{ &priv->lru.willneed, active_evict, can_swap() && can_block(sc) },
drivers/gpu/drm/msm/msm_gem_shrinker.c
190
drm_gem_lru_scan(stages[i].lru, nr,
drivers/gpu/drm/msm/msm_gem_shrinker.c
250
&priv->lru.dontneed,
drivers/gpu/drm/msm/msm_gem_shrinker.c
251
&priv->lru.willneed,
drivers/gpu/drm/msm/msm_gem_shrinker.c
252
&priv->lru.pinned,
drivers/gpu/drm/msm/msm_gem_shrinker.c
38
unsigned count = priv->lru.dontneed.count;
drivers/gpu/drm/msm/msm_gem_shrinker.c
41
count += priv->lru.willneed.count;
drivers/gpu/drm/msm/msm_gem_submit.c
382
mutex_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem_submit.c
386
mutex_unlock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem_vma.c
1308
mutex_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_gem_vma.c
1311
mutex_unlock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_ringbuffer.c
25
mutex_lock(&priv->lru.lock);
drivers/gpu/drm/msm/msm_ringbuffer.c
35
mutex_unlock(&priv->lru.lock);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
143
eobj = list_first_entry_or_null(&imem->lru, typeof(*eobj), lru);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
149
list_del_init(&eobj->lru);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
205
if (likely(iobj->lru.next) && iobj->map) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
206
BUG_ON(!list_empty(&iobj->lru));
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
207
list_add_tail(&iobj->lru, &imem->lru);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
249
if (likely(iobj->lru.next))
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
250
list_del_init(&iobj->lru);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
275
if (likely(iobj->lru.next)) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
276
list_del_init(&iobj->lru);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
277
iobj->lru.next = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
303
iobj->lru.next = NULL; /* Exclude from eviction. */
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
325
if (likely(iobj->lru.next))
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
326
list_del(&iobj->lru);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
370
INIT_LIST_HEAD(&iobj->lru);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
38
struct list_head lru;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
437
INIT_LIST_HEAD(&imem->lru);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
53
struct list_head lru;
drivers/gpu/drm/ttm/tests/ttm_bo_test.c
274
list_is_last(&res1->lru.link, &man->lru[bo->priority]), 1);
drivers/gpu/drm/ttm/tests/ttm_bo_test.c
311
list_is_last(&res2->lru.link, &priv->ttm_dev->unevictable), 1);
drivers/gpu/drm/ttm/tests/ttm_bo_test.c
315
list_is_last(&res1->lru.link, &priv->ttm_dev->unevictable), 1);
drivers/gpu/drm/ttm/tests/ttm_bo_validate_test.c
453
KUNIT_ASSERT_TRUE(test, list_is_singular(&man->lru[bo->priority]));
drivers/gpu/drm/ttm/tests/ttm_device_test.c
93
KUNIT_ASSERT_TRUE(test, list_empty(&man->lru[0]));
drivers/gpu/drm/ttm/tests/ttm_resource_test.c
130
KUNIT_ASSERT_TRUE(test, list_empty(&man->lru[bo->priority]));
drivers/gpu/drm/ttm/tests/ttm_resource_test.c
146
KUNIT_ASSERT_TRUE(test, list_is_singular(&man->lru[bo->priority]));
drivers/gpu/drm/ttm/tests/ttm_resource_test.c
201
KUNIT_ASSERT_TRUE(test, list_empty(&res->lru.link));
drivers/gpu/drm/ttm/tests/ttm_resource_test.c
224
KUNIT_ASSERT_TRUE(test, list_empty(&man->lru[i]));
drivers/gpu/drm/ttm/tests/ttm_resource_test.c
311
KUNIT_ASSERT_TRUE(test, list_empty(&man->lru[bo->priority]));
drivers/gpu/drm/ttm/ttm_device.c
269
if (list_empty(&man->lru[0]))
drivers/gpu/drm/ttm/ttm_device.c
291
list_del_init(&bo->resource->lru.link);
drivers/gpu/drm/ttm/ttm_device.c
316
ttm_device_clear_lru_dma_mappings(bdev, &man->lru[j]);
drivers/gpu/drm/ttm/ttm_pool.c
1181
list_for_each_entry(p, &pt->pages, lru)
drivers/gpu/drm/ttm/ttm_pool.c
297
list_add(&p->lru, &pt->pages);
drivers/gpu/drm/ttm/ttm_pool.c
308
p = list_first_entry_or_null(&pt->pages, typeof(*p), lru);
drivers/gpu/drm/ttm/ttm_pool.c
311
list_del(&p->lru);
drivers/gpu/drm/ttm/ttm_resource.c
175
list_bulk_move_tail(&man->lru[j], &pos->first->lru.link,
drivers/gpu/drm/ttm/ttm_resource.c
176
&pos->last->lru.link);
drivers/gpu/drm/ttm/ttm_resource.c
192
struct ttm_lru_item *lru = &cur->lru;
drivers/gpu/drm/ttm/ttm_resource.c
195
lru = list_prev_entry(lru, link);
drivers/gpu/drm/ttm/ttm_resource.c
196
} while (!ttm_lru_item_is_res(lru));
drivers/gpu/drm/ttm/ttm_resource.c
198
return ttm_lru_item_to_res(lru);
drivers/gpu/drm/ttm/ttm_resource.c
204
struct ttm_lru_item *lru = &cur->lru;
drivers/gpu/drm/ttm/ttm_resource.c
207
lru = list_next_entry(lru, link);
drivers/gpu/drm/ttm/ttm_resource.c
208
} while (!ttm_lru_item_is_res(lru));
drivers/gpu/drm/ttm/ttm_resource.c
210
return ttm_lru_item_to_res(lru);
drivers/gpu/drm/ttm/ttm_resource.c
220
list_move(&res->lru.link, &pos->last->lru.link);
drivers/gpu/drm/ttm/ttm_resource.c
255
list_move(&res->lru.link, &pos->last->lru.link);
drivers/gpu/drm/ttm/ttm_resource.c
304
list_move_tail(&res->lru.link, &bdev->unevictable);
drivers/gpu/drm/ttm/ttm_resource.c
315
list_move_tail(&res->lru.link, &man->lru[bo->priority]);
drivers/gpu/drm/ttm/ttm_resource.c
346
list_add_tail(&res->lru.link, &bo->bdev->unevictable);
drivers/gpu/drm/ttm/ttm_resource.c
348
list_add_tail(&res->lru.link, &man->lru[bo->priority]);
drivers/gpu/drm/ttm/ttm_resource.c
370
list_del_init(&res->lru.link);
drivers/gpu/drm/ttm/ttm_resource.c
529
INIT_LIST_HEAD(&man->lru[i]);
drivers/gpu/drm/ttm/ttm_resource.c
65
list_move(&cursor->hitch.link, &pos->last->lru.link);
drivers/gpu/drm/ttm/ttm_resource.c
662
list_move(&cursor->hitch.link, &man->lru[cursor->priority]);
drivers/gpu/drm/ttm/ttm_resource.c
677
struct ttm_lru_item *lru;
drivers/gpu/drm/ttm/ttm_resource.c
682
lru = &cursor->hitch;
drivers/gpu/drm/ttm/ttm_resource.c
683
list_for_each_entry_continue(lru, &man->lru[cursor->priority], link) {
drivers/gpu/drm/ttm/ttm_resource.c
684
if (ttm_lru_item_is_res(lru)) {
drivers/gpu/drm/ttm/ttm_resource.c
685
ttm_resource_cursor_check_bulk(cursor, lru);
drivers/gpu/drm/ttm/ttm_resource.c
686
list_move(&cursor->hitch.link, &lru->link);
drivers/gpu/drm/ttm/ttm_resource.c
687
return ttm_lru_item_to_res(lru);
drivers/gpu/drm/ttm/ttm_resource.c
694
list_move(&cursor->hitch.link, &man->lru[cursor->priority]);
drivers/gpu/drm/ttm/ttm_resource.c
710
struct ttm_lru_item *lru;
drivers/gpu/drm/ttm/ttm_resource.c
712
list_for_each_entry(lru, head, link) {
drivers/gpu/drm/ttm/ttm_resource.c
713
if (ttm_lru_item_is_res(lru))
drivers/gpu/drm/ttm/ttm_resource.c
714
return ttm_lru_item_to_res(lru);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
116
list_add_tail(&page->lru, &ctx->page_list);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
139
list_for_each_entry_safe(entry, next, &ctx->page_list, lru) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
140
list_del_init(&entry->lru);
drivers/iommu/iommu-pages.c
15
IOPTDESC_MATCH(lru, iopt_freelist_elm); /* Ensure bit 0 is clear */
drivers/md/bcache/bcache.h
289
struct list_head lru;
drivers/md/bcache/request.c
435
i = list_first_entry(&dc->io_lru, struct io, lru);
drivers/md/bcache/request.c
449
list_move_tail(&i->lru, &dc->io_lru);
drivers/md/bcache/super.c
1422
list_add(&io->lru, &dc->io_lru);
drivers/md/dm-bufio.c
100
static void lru_destroy(struct lru *lru)
drivers/md/dm-bufio.c
102
WARN_ON_ONCE(lru->cursor);
drivers/md/dm-bufio.c
103
WARN_ON_ONCE(!list_empty(&lru->iterators));
drivers/md/dm-bufio.c
109
static void lru_insert(struct lru *lru, struct lru_entry *le)
drivers/md/dm-bufio.c
117
if (lru->cursor) {
drivers/md/dm-bufio.c
118
list_add_tail(&le->list, lru->cursor);
drivers/md/dm-bufio.c
121
lru->cursor = &le->list;
drivers/md/dm-bufio.c
123
lru->count++;
drivers/md/dm-bufio.c
139
static void lru_iter_begin(struct lru *lru, struct lru_iter *it)
drivers/md/dm-bufio.c
141
it->lru = lru;
drivers/md/dm-bufio.c
142
it->stop = lru->cursor ? to_le(lru->cursor->prev) : NULL;
drivers/md/dm-bufio.c
143
it->e = lru->cursor ? to_le(lru->cursor) : NULL;
drivers/md/dm-bufio.c
144
list_add(&it->list, &lru->iterators);
drivers/md/dm-bufio.c
1641
list_del(&b->lru.list);
drivers/md/dm-bufio.c
1679
list_add(&b->lru.list, &c->reserved_buffers);
drivers/md/dm-bufio.c
188
static void lru_iter_invalidate(struct lru *lru, struct lru_entry *e)
drivers/md/dm-bufio.c
192
list_for_each_entry(it, &lru->iterators, list) {
drivers/md/dm-bufio.c
214
static void lru_remove(struct lru *lru, struct lru_entry *le)
drivers/md/dm-bufio.c
216
lru_iter_invalidate(lru, le);
drivers/md/dm-bufio.c
2163
lru_iter_begin(&c->cache.lru[LIST_DIRTY], &it);
drivers/md/dm-bufio.c
217
if (lru->count == 1) {
drivers/md/dm-bufio.c
218
lru->cursor = NULL;
drivers/md/dm-bufio.c
220
if (lru->cursor == &le->list)
drivers/md/dm-bufio.c
221
lru->cursor = lru->cursor->next;
drivers/md/dm-bufio.c
224
lru->count--;
drivers/md/dm-bufio.c
249
static struct lru_entry *lru_evict(struct lru *lru, le_predicate pred, void *context, bool no_sleep)
drivers/md/dm-bufio.c
252
struct list_head *h = lru->cursor;
drivers/md/dm-bufio.c
2617
list_del(&b->lru.list);
drivers/md/dm-bufio.c
262
while (tested < lru->count) {
drivers/md/dm-bufio.c
2659
list_del(&b->lru.list);
drivers/md/dm-bufio.c
275
lru->cursor = le->list.next;
drivers/md/dm-bufio.c
283
lru->cursor = le->list.next;
drivers/md/dm-bufio.c
387
struct lru lru[LIST_SIZE];
drivers/md/dm-bufio.c
558
lru_init(&bc->lru[LIST_CLEAN]);
drivers/md/dm-bufio.c
559
lru_init(&bc->lru[LIST_DIRTY]);
drivers/md/dm-bufio.c
569
lru_destroy(&bc->lru[LIST_CLEAN]);
drivers/md/dm-bufio.c
570
lru_destroy(&bc->lru[LIST_DIRTY]);
drivers/md/dm-bufio.c
580
return bc->lru[list_mode].count;
drivers/md/dm-bufio.c
629
lru_reference(&b->lru);
drivers/md/dm-bufio.c
677
le = lru_evict(&bc->lru[list_mode], __evict_pred, &w, bc->no_sleep);
drivers/md/dm-bufio.c
712
lru_remove(&bc->lru[b->list_mode], &b->lru);
drivers/md/dm-bufio.c
714
lru_insert(&bc->lru[b->list_mode], &b->lru);
drivers/md/dm-bufio.c
733
le = lru_evict(&bc->lru[old_mode], __evict_pred, &w, bc->no_sleep);
drivers/md/dm-bufio.c
739
lru_insert(&bc->lru[b->list_mode], &b->lru);
drivers/md/dm-bufio.c
774
struct lru *lru = &bc->lru[list_mode];
drivers/md/dm-bufio.c
777
if (!lru->cursor)
drivers/md/dm-bufio.c
78
struct lru *lru;
drivers/md/dm-bufio.c
780
first = le = to_le(lru->cursor);
drivers/md/dm-bufio.c
855
lru_insert(&bc->lru[b->list_mode], &b->lru);
drivers/md/dm-bufio.c
882
lru_remove(&bc->lru[b->list_mode], &b->lru);
drivers/md/dm-bufio.c
93
static void lru_init(struct lru *lru)
drivers/md/dm-bufio.c
938
lru_remove(&bc->lru[b->list_mode], &b->lru);
drivers/md/dm-bufio.c
95
lru->cursor = NULL;
drivers/md/dm-bufio.c
96
lru->count = 0;
drivers/md/dm-bufio.c
97
INIT_LIST_HEAD(&lru->iterators);
drivers/md/dm-writecache.c
1022
INIT_LIST_HEAD(&wc->lru);
drivers/md/dm-writecache.c
110
struct list_head lru;
drivers/md/dm-writecache.c
1700
INIT_LIST_HEAD(&e->lru);
drivers/md/dm-writecache.c
1736
INIT_LIST_HEAD(&e->lru);
drivers/md/dm-writecache.c
1838
e = container_of(wbl->list.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
1839
list_del(&e->lru);
drivers/md/dm-writecache.c
1865
f = container_of(wbl->list.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
1872
list_del(&f->lru);
drivers/md/dm-writecache.c
1902
e = container_of(wbl->list.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
1903
list_del(&e->lru);
drivers/md/dm-writecache.c
1921
f = container_of(wbl->list.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
1923
list_del(&f->lru);
drivers/md/dm-writecache.c
1992
while (!list_empty(&wc->lru) &&
drivers/md/dm-writecache.c
1995
(jiffies - container_of(wc->lru.prev, struct wc_entry, lru)->age >=
drivers/md/dm-writecache.c
2013
e = container_of(wc->lru.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
2024
list_move(&e->lru, &skipped);
drivers/md/dm-writecache.c
2030
list_move(&e->lru, &wbl.list);
drivers/md/dm-writecache.c
2065
list_move(&g->lru, &wbl.list);
drivers/md/dm-writecache.c
2084
list_splice_tail(&skipped, &wc->lru);
drivers/md/dm-writecache.c
668
list_add(&ins->lru, &wc->lru);
drivers/md/dm-writecache.c
674
list_del(&e->lru);
drivers/md/dm-writecache.c
695
list_add_tail(&e->lru, &wc->freelist);
drivers/md/dm-writecache.c
736
e = container_of(wc->freelist.next, struct wc_entry, lru);
drivers/md/dm-writecache.c
739
list_del(&e->lru);
drivers/md/dm-writecache.c
775
wc->lru.next = LIST_POISON1;
drivers/md/dm-writecache.c
776
wc->lru.prev = LIST_POISON2;
drivers/md/dm-writecache.c
801
if (list_empty(&wc->lru))
drivers/md/dm-writecache.c
804
e = container_of(wc->lru.next, struct wc_entry, lru);
drivers/md/dm-writecache.c
815
if (unlikely(e->lru.next == &wc->lru))
drivers/md/dm-writecache.c
817
e2 = container_of(e->lru.next, struct wc_entry, lru);
drivers/md/dm-writecache.c
83
struct list_head lru;
drivers/md/dm-writecache.c
847
if (unlikely(e->lru.prev == &wc->lru))
drivers/md/dm-writecache.c
849
e = container_of(e->lru.prev, struct wc_entry, lru);
drivers/md/raid5-cache.c
1363
BUG_ON(list_empty(&sh->lru));
drivers/md/raid5-cache.c
1374
list_del_init(&sh->lru);
drivers/md/raid5-cache.c
1404
list_for_each_entry_safe(sh, next, &conf->r5c_full_stripe_list, lru) {
drivers/md/raid5-cache.c
1412
&conf->r5c_partial_stripe_list, lru) {
drivers/md/raid5-cache.c
1476
if (!list_empty(&sh->lru) &&
drivers/md/raid5-cache.c
1933
list_for_each_entry(sh, list, lru)
drivers/md/raid5-cache.c
1945
list_for_each_entry_safe(sh, next, cached_stripe_list, lru) {
drivers/md/raid5-cache.c
1947
list_del_init(&sh->lru);
drivers/md/raid5-cache.c
1958
list_for_each_entry_safe(sh, next, cached_stripe_list, lru)
drivers/md/raid5-cache.c
1961
list_del_init(&sh->lru);
drivers/md/raid5-cache.c
2108
list_del_init(&sh->lru);
drivers/md/raid5-cache.c
2163
list_add_tail(&sh->lru, cached_stripe_list);
drivers/md/raid5-cache.c
2170
list_move_tail(&sh->lru, cached_stripe_list);
drivers/md/raid5-cache.c
2258
list_for_each_entry(sh, &ctx->cached_list, lru) {
drivers/md/raid5-cache.c
2354
list_for_each_entry(sh, &ctx->cached_list, lru) {
drivers/md/raid5-cache.c
2427
list_for_each_entry_safe(sh, next, &ctx->cached_list, lru) {
drivers/md/raid5-cache.c
2430
list_del_init(&sh->lru);
drivers/md/raid5.c
191
if (list_empty(&sh->lru)) {
drivers/md/raid5.c
195
list_add_tail(&sh->lru, &group->loprio_list);
drivers/md/raid5.c
197
list_add_tail(&sh->lru, &group->handle_list);
drivers/md/raid5.c
232
BUG_ON(!list_empty(&sh->lru));
drivers/md/raid5.c
2359
INIT_LIST_HEAD(&sh->lru);
drivers/md/raid5.c
2558
list_add(&nsh->lru, &newstripes);
drivers/md/raid5.c
2563
nsh = list_entry(newstripes.next, struct stripe_head, lru);
drivers/md/raid5.c
2564
list_del(&nsh->lru);
drivers/md/raid5.c
257
list_add_tail(&sh->lru, &conf->delayed_list);
drivers/md/raid5.c
2577
list_for_each_entry(nsh, &newstripes, lru) {
drivers/md/raid5.c
260
list_add_tail(&sh->lru, &conf->bitmap_list);
drivers/md/raid5.c
2641
nsh = list_entry(newstripes.next, struct stripe_head, lru);
drivers/md/raid5.c
2642
list_del_init(&nsh->lru);
drivers/md/raid5.c
266
list_add_tail(&sh->lru,
drivers/md/raid5.c
269
list_add_tail(&sh->lru,
drivers/md/raid5.c
286
list_add_tail(&sh->lru, temp_inactive_list);
drivers/md/raid5.c
290
list_add_tail(&sh->lru, temp_inactive_list);
drivers/md/raid5.c
297
list_add_tail(&sh->lru, &conf->r5c_full_stripe_list);
drivers/md/raid5.c
305
list_add_tail(&sh->lru, &conf->r5c_partial_stripe_list);
drivers/md/raid5.c
455
sh = list_entry(first, struct stripe_head, lru);
drivers/md/raid5.c
5308
sh = list_entry(l, struct stripe_head, lru);
drivers/md/raid5.c
5313
list_add_tail(&sh->lru, &conf->hold_list);
drivers/md/raid5.c
5327
struct stripe_head *sh = list_entry(head.next, struct stripe_head, lru);
drivers/md/raid5.c
5329
list_del_init(&sh->lru);
drivers/md/raid5.c
5562
sh = list_entry(handle_list->next, typeof(*sh), lru);
drivers/md/raid5.c
5581
list_for_each_entry(tmp, &conf->hold_list, lru) {
drivers/md/raid5.c
5611
list_del_init(&sh->lru);
drivers/md/raid5.c
5635
sh = list_first_entry(&cb->list, struct stripe_head, lru);
drivers/md/raid5.c
5636
list_del_init(&sh->lru);
drivers/md/raid5.c
5684
list_add_tail(&sh->lru, &cb->list);
drivers/md/raid5.c
6426
list_add(&sh->lru, &stripes);
drivers/md/raid5.c
6460
sh = list_entry(stripes.next, struct stripe_head, lru);
drivers/md/raid5.c
6461
list_del_init(&sh->lru);
drivers/md/raid5.c
658
BUG_ON(list_empty(&sh->lru) &&
drivers/md/raid5.c
663
list_del_init(&sh->lru);
drivers/md/raid5.h
202
struct list_head lru; /* inactive_list or handle_list */
drivers/misc/vmw_balloon.c
1006
list_for_each_entry(page, pages, lru) {
drivers/misc/vmw_balloon.c
1050
list_for_each_entry_safe(page, tmp, &b->huge_pages, lru) {
drivers/misc/vmw_balloon.c
1053
list_move(&page->lru, pages);
drivers/misc/vmw_balloon.c
1081
list_for_each_entry_safe(page, tmp, &ctl->refused_pages, lru) {
drivers/misc/vmw_balloon.c
1082
list_del(&page->lru);
drivers/misc/vmw_balloon.c
1085
list_add(&page[i].lru, &ctl->prealloc_pages);
drivers/misc/vmw_balloon.c
670
struct page, lru);
drivers/misc/vmw_balloon.c
671
list_del(&page->lru);
drivers/misc/vmw_balloon.c
685
list_add(&page->lru, &ctl->pages);
drivers/misc/vmw_balloon.c
863
list_for_each_entry(page, &ctl->pages, lru)
drivers/misc/vmw_balloon.c
895
list_move(&page->lru, &ctl->refused_pages);
drivers/misc/vmw_balloon.c
920
list_for_each_entry_safe(page, tmp, page_list, lru) {
drivers/misc/vmw_balloon.c
921
list_del(&page->lru);
drivers/virtio/virtio_balloon.c
265
list_add(&page->lru, &pages);
drivers/virtio/virtio_balloon.c
272
list_for_each_entry_safe(page, next, &pages, lru) {
drivers/virtio/virtio_balloon.c
273
list_del(&page->lru);
drivers/virtio/virtio_balloon.c
295
list_for_each_entry_safe(page, next, pages, lru) {
drivers/virtio/virtio_balloon.c
296
list_del(&page->lru);
drivers/virtio/virtio_balloon.c
320
list_add(&page->lru, &pages);
drivers/virtio/virtio_balloon.c
486
list_for_each_entry_safe(page, next, &vb->free_page_list, lru) {
drivers/virtio/virtio_balloon.c
487
list_del(&page->lru);
drivers/virtio/virtio_balloon.c
725
list_add(&page->lru, &vb->free_page_list);
drivers/xen/balloon.c
154
list_add_tail(&page->lru, &ballooned_pages);
drivers/xen/balloon.c
157
list_add(&page->lru, &ballooned_pages);
drivers/xen/balloon.c
173
page = list_entry(ballooned_pages.next, struct page, lru);
drivers/xen/balloon.c
176
list_del(&page->lru);
drivers/xen/balloon.c
191
struct list_head *next = page->lru.next;
drivers/xen/balloon.c
194
return list_entry(next, struct page, lru);
drivers/xen/balloon.c
399
page = list_first_entry_or_null(&ballooned_pages, struct page, lru);
drivers/xen/balloon.c
453
list_add(&page->lru, &pages);
drivers/xen/balloon.c
470
list_for_each_entry_safe(page, tmp, &pages, lru) {
drivers/xen/balloon.c
475
list_del(&page->lru);
drivers/xen/grant-table.c
946
page = list_first_entry(&cache->pages, struct page, lru);
drivers/xen/grant-table.c
947
list_del(&page->lru);
drivers/xen/grant-table.c
954
list_add(&page->lru, &cache->pages);
drivers/xen/privcmd.c
122
list_for_each_entry_safe(p, n, pages, lru)
drivers/xen/privcmd.c
157
list_add_tail(&page->lru, pagelist);
drivers/xen/privcmd.c
197
page = list_entry(pos, struct page, lru);
drivers/xen/privcmd.c
231
page = list_entry(pos, struct page, lru);
drivers/xen/privcmd.c
310
struct page, lru);
fs/btrfs/compression.c
171
struct page *page = list_entry(tmp, struct page, lru);
fs/btrfs/compression.c
193
folio = list_first_entry(&compr_pool.list, struct folio, lru);
fs/btrfs/compression.c
194
list_del_init(&folio->lru);
fs/btrfs/compression.c
218
list_add(&folio->lru, &compr_pool.list);
fs/btrfs/send.c
2766
struct btrfs_lru_cache_entry *lru;
fs/btrfs/send.c
2769
lru = btrfs_lru_cache_lru_entry(&sctx->dir_utimes_cache);
fs/btrfs/send.c
2770
ASSERT(lru != NULL);
fs/btrfs/send.c
2772
ret = send_utimes(sctx, lru->key, lru->gen);
fs/btrfs/send.c
2776
btrfs_lru_cache_remove(&sctx->dir_utimes_cache, lru);
fs/ceph/mds_client.c
4567
page = list_first_entry(&recon_state->pagelist->head, struct page, lru);
fs/ceph/mds_client.c
5019
struct page, lru);
fs/ceph/mds_client.h
402
struct list_head lru;
fs/ceph/snap.c
1207
list_del_init(&exist->lru);
fs/ceph/snap.c
1229
INIT_LIST_HEAD(&sm->lru);
fs/ceph/snap.c
1250
list_del_init(&exist->lru);
fs/ceph/snap.c
1276
list_add_tail(&sm->lru, &mdsc->snapid_map_lru);
fs/ceph/snap.c
1299
struct ceph_snapid_map, lru);
fs/ceph/snap.c
1304
list_move(&sm->lru, &to_free);
fs/ceph/snap.c
1309
sm = list_first_entry(&to_free, struct ceph_snapid_map, lru);
fs/ceph/snap.c
1310
list_del(&sm->lru);
fs/ceph/snap.c
1329
list_move(&sm->lru, &to_free);
fs/ceph/snap.c
1334
sm = list_first_entry(&to_free, struct ceph_snapid_map, lru);
fs/ceph/snap.c
1335
list_del(&sm->lru);
fs/ceph/xattr.c
1420
struct page, lru);
fs/dcache.c
1180
struct list_lru_one *lru, void *arg)
fs/dcache.c
1200
d_lru_isolate(lru, dentry);
fs/dcache.c
1231
d_lru_shrink_move(lru, dentry, freeable);
fs/dcache.c
1261
struct list_lru_one *lru, void *arg)
fs/dcache.c
1274
d_lru_shrink_move(lru, dentry, freeable);
fs/dcache.c
534
static void d_lru_isolate(struct list_lru_one *lru, struct dentry *dentry)
fs/dcache.c
541
list_lru_isolate(lru, &dentry->d_lru);
fs/dcache.c
544
static void d_lru_shrink_move(struct list_lru_one *lru, struct dentry *dentry,
fs/dcache.c
551
list_lru_isolate_move(lru, &dentry->d_lru, list);
fs/gfs2/quota.c
152
struct list_lru_one *lru, void *arg)
fs/gfs2/quota.c
165
list_lru_isolate_move(lru, &qd->qd_lru, dispose);
fs/inode.c
1000
spin_unlock(&lru->lock);
fs/inode.c
1016
list_lru_isolate_move(lru, &inode->i_lru, freeable);
fs/inode.c
958
struct list_lru_one *lru, void *arg)
fs/inode.c
979
list_lru_isolate(lru, &inode->i_lru);
fs/nfs/dir.c
2897
cache = list_entry(head->next, struct nfs_access_entry, lru);
fs/nfs/dir.c
2898
list_del(&cache->lru);
fs/nfs/dir.c
2922
struct nfs_access_entry, lru);
fs/nfs/dir.c
2923
list_move(&cache->lru, &head);
fs/nfs/dir.c
2987
list_move(&entry->lru, head);
fs/nfs/dir.c
3122
list_move_tail(&cache->lru, &nfsi->access_cache_entry_lru);
fs/nfs/dir.c
3148
cache = list_entry(lh, struct nfs_access_entry, lru);
fs/nfs/dir.c
3205
list_add_tail(&set->lru, &nfsi->access_cache_entry_lru);
fs/nfs/dir.c
3210
list_add_tail(&set->lru, &nfsi->access_cache_entry_lru);
fs/nfs/dir.c
3211
list_del(&entry->lru);
fs/nfs/nfs42xattr.c
1008
struct list_lru *lru)
fs/nfs/nfs42xattr.c
1011
list_lru_destroy(lru);
fs/nfs/nfs42xattr.c
130
struct list_lru *lru;
fs/nfs/nfs42xattr.c
132
lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ?
fs/nfs/nfs42xattr.c
135
return list_lru_add_obj(lru, &entry->lru);
fs/nfs/nfs42xattr.c
141
struct list_lru *lru;
fs/nfs/nfs42xattr.c
143
lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ?
fs/nfs/nfs42xattr.c
146
return list_lru_del_obj(lru, &entry->lru);
fs/nfs/nfs42xattr.c
239
INIT_LIST_HEAD(&entry->lru);
fs/nfs/nfs42xattr.c
261
if (WARN_ON(!list_empty(&entry->lru)))
fs/nfs/nfs42xattr.c
352
list_lru_del_obj(&nfs4_xattr_cache_lru, &oldcache->lru);
fs/nfs/nfs42xattr.c
477
list_lru_add_obj(&nfs4_xattr_cache_lru, &cache->lru);
fs/nfs/nfs42xattr.c
71
struct list_head lru;
fs/nfs/nfs42xattr.c
805
struct list_lru_one *lru, void *arg)
fs/nfs/nfs42xattr.c
810
struct nfs4_xattr_cache, lru);
fs/nfs/nfs42xattr.c
82
struct list_head lru;
fs/nfs/nfs42xattr.c
830
list_lru_isolate(lru, &cache->lru);
fs/nfs/nfs42xattr.c
870
struct list_lru_one *lru, void *arg)
fs/nfs/nfs42xattr.c
876
struct nfs4_xattr_entry, lru);
fs/nfs/nfs42xattr.c
899
list_lru_isolate(lru, &entry->lru);
fs/nfs/nfs42xattr.c
910
list_lru_isolate(lru, &entry->lru);
fs/nfs/nfs42xattr.c
925
struct list_lru *lru;
fs/nfs/nfs42xattr.c
927
lru = (shrink == nfs4_xattr_large_entry_shrinker) ?
fs/nfs/nfs42xattr.c
930
freed = list_lru_shrink_walk(lru, sc, entry_lru_isolate, &dispose);
fs/nfs/nfs42xattr.c
953
struct list_lru *lru;
fs/nfs/nfs42xattr.c
955
lru = (shrink == nfs4_xattr_large_entry_shrinker) ?
fs/nfs/nfs42xattr.c
958
count = list_lru_shrink_count(lru, sc);
fs/nfs/nfs42xattr.c
971
INIT_LIST_HEAD(&cache->lru);
fs/nfs/nfs42xattr.c
981
struct list_lru *lru, const char *name,
fs/nfs/nfs42xattr.c
991
ret = list_lru_init_memcg(lru, *shrinker);
fs/nfsd/filecache.c
1389
unsigned long lru = 0, total_age = 0;
fs/nfsd/filecache.c
1397
lru = list_lru_count(&nfsd_file_lru);
fs/nfsd/filecache.c
1419
seq_printf(m, "lru entries: %lu\n", lru);
fs/nfsd/filecache.c
495
nfsd_file_lru_cb(struct list_head *item, struct list_lru_one *lru,
fs/nfsd/filecache.c
530
list_lru_isolate(lru, &nf->nf_lru);
fs/nfsd/filecache.c
538
nfsd_file_gc_cb(struct list_head *item, struct list_lru_one *lru,
fs/nfsd/filecache.c
552
return nfsd_file_lru_cb(item, lru, arg);
fs/proc/meminfo.c
42
int lru;
fs/proc/meminfo.c
53
for (lru = LRU_BASE; lru < NR_LRU_LISTS; lru++)
fs/proc/meminfo.c
54
pages[lru] = global_node_page_state(NR_LRU_BASE + lru);
fs/xfs/xfs_buf.c
1508
struct list_lru_one *lru,
fs/xfs/xfs_buf.c
1530
list_lru_isolate_move(lru, item, dispose);
fs/xfs/xfs_buf.c
1606
struct list_lru_one *lru,
fs/xfs/xfs_buf.c
1629
list_lru_isolate_move(lru, item, dispose);
fs/xfs/xfs_qm.c
416
struct list_lru_one *lru,
fs/xfs/xfs_qm.c
418
__releases(&lru->lock) __acquires(&lru->lock)
fs/xfs/xfs_qm.c
454
list_lru_isolate(lru, &dqp->q_lru);
fs/xfs/xfs_qm.c
477
list_lru_isolate_move(lru, &dqp->q_lru, &isol->dispose);
include/drm/drm_gem.h
457
struct drm_gem_lru *lru;
include/drm/drm_gem.h
613
void drm_gem_lru_init(struct drm_gem_lru *lru, struct mutex *lock);
include/drm/drm_gem.h
615
void drm_gem_lru_move_tail_locked(struct drm_gem_lru *lru, struct drm_gem_object *obj);
include/drm/drm_gem.h
616
void drm_gem_lru_move_tail(struct drm_gem_lru *lru, struct drm_gem_object *obj);
include/drm/drm_gem.h
618
drm_gem_lru_scan(struct drm_gem_lru *lru,
include/drm/ttm/ttm_resource.h
218
struct list_head lru[TTM_MAX_BO_PRIORITY];
include/drm/ttm/ttm_resource.h
276
struct ttm_lru_item lru;
include/drm/ttm/ttm_resource.h
288
return container_of(item, struct ttm_resource, lru);
include/drm/ttm/ttm_resource.h
406
WARN_ON(!list_empty(&man->lru[i]));
include/linux/list_lru.h
115
bool list_lru_add(struct list_lru *lru, struct list_head *item, int nid,
include/linux/list_lru.h
129
bool list_lru_add_obj(struct list_lru *lru, struct list_head *item);
include/linux/list_lru.h
155
bool list_lru_del(struct list_lru *lru, struct list_head *item, int nid,
include/linux/list_lru.h
169
bool list_lru_del_obj(struct list_lru *lru, struct list_head *item);
include/linux/list_lru.h
183
unsigned long list_lru_count_one(struct list_lru *lru,
include/linux/list_lru.h
185
unsigned long list_lru_count_node(struct list_lru *lru, int nid);
include/linux/list_lru.h
187
static inline unsigned long list_lru_shrink_count(struct list_lru *lru,
include/linux/list_lru.h
190
return list_lru_count_one(lru, sc->nid, sc->memcg);
include/linux/list_lru.h
193
static inline unsigned long list_lru_count(struct list_lru *lru)
include/linux/list_lru.h
199
count += list_lru_count_node(lru, nid);
include/linux/list_lru.h
47
struct list_lru_one lru;
include/linux/list_lru.h
64
void list_lru_destroy(struct list_lru *lru);
include/linux/list_lru.h
65
int __list_lru_init(struct list_lru *lru, bool memcg_aware,
include/linux/list_lru.h
68
#define list_lru_init(lru) \
include/linux/list_lru.h
69
__list_lru_init((lru), false, NULL)
include/linux/list_lru.h
70
#define list_lru_init_memcg(lru, shrinker) \
include/linux/list_lru.h
71
__list_lru_init((lru), true, shrinker)
include/linux/list_lru.h
73
static inline int list_lru_init_memcg_key(struct list_lru *lru, struct shrinker *shrinker,
include/linux/list_lru.h
77
lru->key = key;
include/linux/list_lru.h
79
return list_lru_init_memcg(lru, shrinker);
include/linux/list_lru.h
82
int memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru,
include/linux/lru_cache.h
166
struct list_head lru;
include/linux/memcontrol.h
1323
enum lru_list lru, int zone_idx)
include/linux/memcontrol.h
898
void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru,
include/linux/memcontrol.h
903
enum lru_list lru, int zone_idx)
include/linux/memcontrol.h
908
return READ_ONCE(mz->lru_zone_size[zone_idx][lru]);
include/linux/mm.h
251
return list_entry((head)->prev, struct folio, lru);
include/linux/mm.h
2756
return (uintptr_t)page->lru.next & BIT(1);
include/linux/mm.h
2771
return (uintptr_t)folio->lru.next & BIT(1);
include/linux/mm.h
2780
page->lru.next = (void *)BIT(1);
include/linux/mm.h
2785
page->lru.next = NULL;
include/linux/mm_inline.h
100
return lru;
include/linux/mm_inline.h
180
enum lru_list lru = type * LRU_INACTIVE_FILE;
include/linux/mm_inline.h
197
lru += LRU_ACTIVE;
include/linux/mm_inline.h
198
__update_lru_size(lruvec, lru, zone, delta);
include/linux/mm_inline.h
205
lru += LRU_ACTIVE;
include/linux/mm_inline.h
206
__update_lru_size(lruvec, lru, zone, -delta);
include/linux/mm_inline.h
212
__update_lru_size(lruvec, lru, zone, -delta);
include/linux/mm_inline.h
213
__update_lru_size(lruvec, lru + LRU_ACTIVE, zone, delta);
include/linux/mm_inline.h
277
list_add_tail(&folio->lru, &lrugen->folios[gen][type][zone]);
include/linux/mm_inline.h
279
list_add(&folio->lru, &lrugen->folios[gen][type][zone]);
include/linux/mm_inline.h
301
list_del(&folio->lru);
include/linux/mm_inline.h
343
enum lru_list lru = folio_lru_list(folio);
include/linux/mm_inline.h
348
update_lru_size(lruvec, lru, folio_zonenum(folio),
include/linux/mm_inline.h
350
if (lru != LRU_UNEVICTABLE)
include/linux/mm_inline.h
351
list_add(&folio->lru, &lruvec->lists[lru]);
include/linux/mm_inline.h
357
enum lru_list lru = folio_lru_list(folio);
include/linux/mm_inline.h
362
update_lru_size(lruvec, lru, folio_zonenum(folio),
include/linux/mm_inline.h
365
list_add_tail(&folio->lru, &lruvec->lists[lru]);
include/linux/mm_inline.h
371
enum lru_list lru = folio_lru_list(folio);
include/linux/mm_inline.h
376
if (lru != LRU_UNEVICTABLE)
include/linux/mm_inline.h
377
list_del(&folio->lru);
include/linux/mm_inline.h
378
update_lru_size(lruvec, lru, folio_zonenum(folio),
include/linux/mm_inline.h
39
enum lru_list lru, enum zone_type zid,
include/linux/mm_inline.h
47
mod_lruvec_state(lruvec, NR_LRU_BASE + lru, nr_pages);
include/linux/mm_inline.h
49
NR_ZONE_LRU_BASE + lru, nr_pages);
include/linux/mm_inline.h
53
enum lru_list lru, enum zone_type zid,
include/linux/mm_inline.h
56
__update_lru_size(lruvec, lru, zid, nr_pages);
include/linux/mm_inline.h
58
mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages);
include/linux/mm_inline.h
89
enum lru_list lru;
include/linux/mm_inline.h
96
lru = folio_is_file_lru(folio) ? LRU_INACTIVE_FILE : LRU_INACTIVE_ANON;
include/linux/mm_inline.h
98
lru += LRU_ACTIVE;
include/linux/mm_types.h
408
struct list_head lru;
include/linux/mm_types.h
511
FOLIO_MATCH(lru, lru);
include/linux/mm_types.h
513
FOLIO_MATCH(compound_head, lru);
include/linux/mm_types.h
96
struct list_head lru;
include/linux/mmzone.h
333
#define for_each_lru(lru) for (lru = 0; lru < NR_LRU_LISTS; lru++)
include/linux/mmzone.h
335
#define for_each_evictable_lru(lru) for (lru = 0; lru <= LRU_ACTIVE_FILE; lru++)
include/linux/mmzone.h
337
static inline bool is_file_lru(enum lru_list lru)
include/linux/mmzone.h
339
return (lru == LRU_INACTIVE_FILE || lru == LRU_ACTIVE_FILE);
include/linux/mmzone.h
342
static inline bool is_active_lru(enum lru_list lru)
include/linux/mmzone.h
344
return (lru == LRU_ACTIVE_ANON || lru == LRU_ACTIVE_FILE);
include/linux/nfs_fs.h
62
struct list_head lru;
include/linux/page-flags.h
558
PAGEFLAG(LRU, lru, PF_HEAD) __CLEARPAGEFLAG(LRU, lru, PF_HEAD)
include/linux/page-flags.h
559
TESTCLEARFLAG(LRU, lru, PF_HEAD)
include/linux/slab.h
770
void *kmem_cache_alloc_lru_noprof(struct kmem_cache *s, struct list_lru *lru,
include/linux/vmstat.h
501
static inline const char *lru_list_name(enum lru_list lru)
include/linux/vmstat.h
503
return node_stat_name(NR_LRU_BASE + lru) + 3; // skip "nr_"
include/linux/xarray.h
1701
static inline void xas_set_lru(struct xa_state *xas, struct list_lru *lru)
include/linux/xarray.h
1703
xas->xa_lru = lru;
include/trace/events/mmflags.h
148
DEF_PAGEFLAG_NAME(lru), \
include/trace/events/pagemap.h
37
__field(enum lru_list, lru )
include/trace/events/pagemap.h
44
__entry->lru = folio_lru_list(folio);
include/trace/events/pagemap.h
52
__entry->lru,
include/trace/events/vmscan.h
298
int lru),
include/trace/events/vmscan.h
300
TP_ARGS(highest_zoneidx, order, nr_requested, nr_scanned, nr_skipped, nr_taken, lru),
include/trace/events/vmscan.h
309
__field(int, lru)
include/trace/events/vmscan.h
319
__entry->lru = lru;
include/trace/events/vmscan.h
333
__print_symbolic(__entry->lru, LRU_NAMES))
kernel/bpf/bpf_lru_list.c
136
static void __bpf_lru_list_rotate_active(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
150
if (++i == lru->nr_scans || node == first_node)
kernel/bpf/bpf_lru_list.c
163
static void __bpf_lru_list_rotate_inactive(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
179
while (i < lru->nr_scans) {
kernel/bpf/bpf_lru_list.c
203
__bpf_lru_list_shrink_inactive(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
217
} else if (lru->del_from_htab(lru->del_arg, node)) {
kernel/bpf/bpf_lru_list.c
224
if (++i == lru->nr_scans)
kernel/bpf/bpf_lru_list.c
234
static void __bpf_lru_list_rotate(struct bpf_lru *lru, struct bpf_lru_list *l)
kernel/bpf/bpf_lru_list.c
237
__bpf_lru_list_rotate_active(lru, l);
kernel/bpf/bpf_lru_list.c
239
__bpf_lru_list_rotate_inactive(lru, l);
kernel/bpf/bpf_lru_list.c
252
static unsigned int __bpf_lru_list_shrink(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
263
nshrinked = __bpf_lru_list_shrink_inactive(lru, l, tgt_nshrink,
kernel/bpf/bpf_lru_list.c
276
if (lru->del_from_htab(lru->del_arg, node)) {
kernel/bpf/bpf_lru_list.c
315
static void bpf_lru_list_pop_free_to_local(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
318
struct bpf_lru_list *l = &lru->common_lru.lru_list;
kernel/bpf/bpf_lru_list.c
326
__bpf_lru_list_rotate(lru, l);
kernel/bpf/bpf_lru_list.c
332
if (++nfree == lru->target_free)
kernel/bpf/bpf_lru_list.c
336
if (nfree < lru->target_free)
kernel/bpf/bpf_lru_list.c
337
__bpf_lru_list_shrink(lru, l, lru->target_free - nfree,
kernel/bpf/bpf_lru_list.c
344
static void __local_list_add_pending(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
350
*(u32 *)((void *)node + lru->hash_offset) = hash;
kernel/bpf/bpf_lru_list.c
372
__local_list_pop_pending(struct bpf_lru *lru, struct bpf_lru_locallist *loc_l)
kernel/bpf/bpf_lru_list.c
382
lru->del_from_htab(lru->del_arg, node)) {
kernel/bpf/bpf_lru_list.c
396
static struct bpf_lru_node *bpf_percpu_lru_pop_free(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
405
l = per_cpu_ptr(lru->percpu_lru, cpu);
kernel/bpf/bpf_lru_list.c
409
__bpf_lru_list_rotate(lru, l);
kernel/bpf/bpf_lru_list.c
413
__bpf_lru_list_shrink(lru, l, PERCPU_FREE_TARGET, free_list,
kernel/bpf/bpf_lru_list.c
418
*(u32 *)((void *)node + lru->hash_offset) = hash;
kernel/bpf/bpf_lru_list.c
428
static struct bpf_lru_node *bpf_common_lru_pop_free(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
432
struct bpf_common_lru *clru = &lru->common_lru;
kernel/bpf/bpf_lru_list.c
444
bpf_lru_list_pop_free_to_local(lru, loc_l);
kernel/bpf/bpf_lru_list.c
449
__local_list_add_pending(lru, loc_l, cpu, node, hash);
kernel/bpf/bpf_lru_list.c
473
node = __local_list_pop_pending(lru, steal_loc_l);
kernel/bpf/bpf_lru_list.c
484
__local_list_add_pending(lru, loc_l, cpu, node, hash);
kernel/bpf/bpf_lru_list.c
491
struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash)
kernel/bpf/bpf_lru_list.c
493
if (lru->percpu)
kernel/bpf/bpf_lru_list.c
494
return bpf_percpu_lru_pop_free(lru, hash);
kernel/bpf/bpf_lru_list.c
496
return bpf_common_lru_pop_free(lru, hash);
kernel/bpf/bpf_lru_list.c
499
static void bpf_common_lru_push_free(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
512
loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu);
kernel/bpf/bpf_lru_list.c
530
bpf_lru_list_push_free(&lru->common_lru.lru_list, node);
kernel/bpf/bpf_lru_list.c
533
static void bpf_percpu_lru_push_free(struct bpf_lru *lru,
kernel/bpf/bpf_lru_list.c
539
l = per_cpu_ptr(lru->percpu_lru, node->cpu);
kernel/bpf/bpf_lru_list.c
548
void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node)
kernel/bpf/bpf_lru_list.c
550
if (lru->percpu)
kernel/bpf/bpf_lru_list.c
551
bpf_percpu_lru_push_free(lru, node);
kernel/bpf/bpf_lru_list.c
553
bpf_common_lru_push_free(lru, node);
kernel/bpf/bpf_lru_list.c
556
static void bpf_common_lru_populate(struct bpf_lru *lru, void *buf,
kernel/bpf/bpf_lru_list.c
560
struct bpf_lru_list *l = &lru->common_lru.lru_list;
kernel/bpf/bpf_lru_list.c
573
lru->target_free = clamp((nr_elems / num_possible_cpus()) / 2,
kernel/bpf/bpf_lru_list.c
577
static void bpf_percpu_lru_populate(struct bpf_lru *lru, void *buf,
kernel/bpf/bpf_lru_list.c
592
l = per_cpu_ptr(lru->percpu_lru, cpu);
kernel/bpf/bpf_lru_list.c
608
void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset,
kernel/bpf/bpf_lru_list.c
611
if (lru->percpu)
kernel/bpf/bpf_lru_list.c
612
bpf_percpu_lru_populate(lru, buf, node_offset, elem_size,
kernel/bpf/bpf_lru_list.c
615
bpf_common_lru_populate(lru, buf, node_offset, elem_size,
kernel/bpf/bpf_lru_list.c
646
int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
kernel/bpf/bpf_lru_list.c
652
lru->percpu_lru = alloc_percpu(struct bpf_lru_list);
kernel/bpf/bpf_lru_list.c
653
if (!lru->percpu_lru)
kernel/bpf/bpf_lru_list.c
659
l = per_cpu_ptr(lru->percpu_lru, cpu);
kernel/bpf/bpf_lru_list.c
662
lru->nr_scans = PERCPU_NR_SCANS;
kernel/bpf/bpf_lru_list.c
664
struct bpf_common_lru *clru = &lru->common_lru;
kernel/bpf/bpf_lru_list.c
678
lru->nr_scans = LOCAL_NR_SCANS;
kernel/bpf/bpf_lru_list.c
681
lru->percpu = percpu;
kernel/bpf/bpf_lru_list.c
682
lru->del_from_htab = del_from_htab;
kernel/bpf/bpf_lru_list.c
683
lru->del_arg = del_arg;
kernel/bpf/bpf_lru_list.c
684
lru->hash_offset = hash_offset;
kernel/bpf/bpf_lru_list.c
689
void bpf_lru_destroy(struct bpf_lru *lru)
kernel/bpf/bpf_lru_list.c
691
if (lru->percpu)
kernel/bpf/bpf_lru_list.c
692
free_percpu(lru->percpu_lru);
kernel/bpf/bpf_lru_list.c
694
free_percpu(lru->common_lru.local_list);
kernel/bpf/bpf_lru_list.h
72
int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
kernel/bpf/bpf_lru_list.h
74
void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset,
kernel/bpf/bpf_lru_list.h
76
void bpf_lru_destroy(struct bpf_lru *lru);
kernel/bpf/bpf_lru_list.h
77
struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash);
kernel/bpf/bpf_lru_list.h
78
void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node);
kernel/bpf/hashtab.c
1271
bpf_lru_push_free(&htab->lru, &elem->lru_node);
kernel/bpf/hashtab.c
1476
bpf_lru_push_free(&htab->lru, &l_new->lru_node);
kernel/bpf/hashtab.c
2318
bool lru = htab_is_lru(htab);
kernel/bpf/hashtab.c
2333
else if (!lru)
kernel/bpf/hashtab.c
301
struct bpf_lru_node *node = bpf_lru_pop_free(&htab->lru, hash);
kernel/bpf/hashtab.c
345
err = bpf_lru_init(&htab->lru,
kernel/bpf/hashtab.c
358
bpf_lru_populate(&htab->lru, htab->elems,
kernel/bpf/hashtab.c
378
bpf_lru_destroy(&htab->lru);
kernel/bpf/hashtab.c
411
bool lru = (attr->map_type == BPF_MAP_TYPE_LRU_HASH ||
kernel/bpf/hashtab.c
434
if (!lru && percpu_lru)
kernel/bpf/hashtab.c
437
if (lru && !prealloc)
kernel/bpf/hashtab.c
93
struct bpf_lru lru;
kernel/kexec_core.c
324
list_for_each_entry_safe(page, next, list, lru) {
kernel/kexec_core.c
325
list_del(&page->lru);
kernel/kexec_core.c
368
list_add(&pages->lru, &extra_pages);
kernel/kexec_core.c
375
list_add(&pages->lru, &image->control_pages);
kernel/kexec_core.c
675
list_for_each_entry(page, &image->dest_pages, lru) {
kernel/kexec_core.c
678
list_del(&page->lru);
kernel/kexec_core.c
693
list_add(&page->lru, &image->unusable_pages);
kernel/kexec_core.c
736
list_add(&page->lru, &image->dest_pages);
kernel/power/wakelock.c
100
list_move(&wl->lru, &wakelocks_lru_list);
kernel/power/wakelock.c
111
list_for_each_entry_safe_reverse(wl, aux, &wakelocks_lru_list, lru) {
kernel/power/wakelock.c
126
list_del(&wl->lru);
kernel/power/wakelock.c
32
struct list_head lru;
kernel/power/wakelock.c
95
list_add(&wl->lru, &wakelocks_lru_list);
kernel/vmcore_info.c
198
VMCOREINFO_OFFSET(page, lru);
lib/lru_cache.c
109
INIT_LIST_HEAD(&lc->lru);
lib/lru_cache.c
188
INIT_LIST_HEAD(&lc->lru);
lib/lru_cache.c
318
else if (!list_empty(&lc->lru))
lib/lru_cache.c
319
n = lc->lru.prev;
lib/lru_cache.c
339
if (!list_empty(&lc->lru))
lib/lru_cache.c
557
list_move(&e->list, &lc->lru);
lib/test_lockup.c
315
list_add(&page->lru, pages);
lib/test_lockup.c
323
list_for_each_entry_safe(page, next, pages, lru)
mm/balloon.c
119
list_for_each_entry_safe(page, tmp, &b_dev_info->pages, lru) {
mm/balloon.c
122
list_del(&page->lru);
mm/balloon.c
127
list_add(&page->lru, pages);
mm/balloon.c
226
return list_first_entry(&pages, struct page, lru);
mm/balloon.c
252
list_del(&page->lru);
mm/balloon.c
273
list_add(&page->lru, &b_dev_info->pages);
mm/balloon.c
36
list_add(&page->lru, &balloon->pages);
mm/balloon.c
83
list_for_each_entry_safe(page, tmp, pages, lru) {
mm/balloon.c
84
list_del(&page->lru);
mm/bootmem_info.c
37
INIT_LIST_HEAD(&page->lru);
mm/compaction.c
102
list_del(&page->lru);
mm/compaction.c
1206
list_add(&folio->lru, &cc->migratepages);
mm/compaction.c
1608
list_add_tail(&page->lru, &cc->freepages[order]);
mm/compaction.c
1822
lru);
mm/compaction.c
1825
list_del(&freepage->lru);
mm/compaction.c
1831
list_add(&freepage[size].lru, &cc->freepages[start_order]);
mm/compaction.c
1863
list_add(&dst->lru, &cc->freepages[order]);
mm/compaction.c
634
list_add_tail(&page->lru, &freelist[order]);
mm/compaction.c
99
list_for_each_entry_safe(page, next, &freepages[order], lru) {
mm/damon/ops-common.c
345
list_del(&folio->lru);
mm/damon/ops-common.c
351
list_add(&folio->lru, &migrate_folios);
mm/damon/ops-common.c
355
list_add(&folio->lru, &ret_folios);
mm/damon/ops-common.c
375
list_del(&folio->lru);
mm/damon/ops-common.c
403
list_move(&folio->lru, &node_folio_list);
mm/damon/paddr.c
196
list_add(&folio->lru, &folio_list);
mm/damon/paddr.c
280
list_add(&folio->lru, &folio_list);
mm/damon/vaddr.c
697
list_add(&folio->lru, &migration_lists[i]);
mm/gup.c
2306
list_add_tail(&folio->lru, movable_folio_list);
mm/huge_memory.c
3459
list_add_tail(&new_folio->lru, list);
mm/huge_memory.c
3466
list_add_tail(&new_folio->lru, &folio->lru);
mm/hugetlb.c
1294
list_move(&folio->lru, &h->hugepage_freelists[nid]);
mm/hugetlb.c
1307
list_for_each_entry(folio, &h->hugepage_freelists[nid], lru) {
mm/hugetlb.c
1317
list_move(&folio->lru, &h->hugepage_activelist);
mm/hugetlb.c
1457
list_del(&folio->lru);
mm/hugetlb.c
1490
INIT_LIST_HEAD(&folio->lru);
mm/hugetlb.c
1648
list_for_each_entry_safe(folio, t_folio, non_hvo_folios, lru) {
mm/hugetlb.c
1649
list_del(&folio->lru);
mm/hugetlb.c
1667
list_for_each_entry_safe(folio, t_folio, folio_list, lru)
mm/hugetlb.c
1669
list_del(&folio->lru);
mm/hugetlb.c
1674
list_del(&folio->lru);
mm/hugetlb.c
1715
list_for_each_entry(folio, &non_hvo_folios, lru)
mm/hugetlb.c
1720
list_for_each_entry_safe(folio, t_folio, &non_hvo_folios, lru) {
mm/hugetlb.c
1818
INIT_LIST_HEAD(&folio->lru);
mm/hugetlb.c
1938
list_for_each_entry_safe(folio, tmp_f, folio_list, lru) {
mm/hugetlb.c
1991
struct folio, lru);
mm/hugetlb.c
2340
list_add(&folio->lru, &surplus_list);
mm/hugetlb.c
2375
list_for_each_entry_safe(folio, tmp, &surplus_list, lru) {
mm/hugetlb.c
2388
list_for_each_entry_safe(folio, tmp, &surplus_list, lru)
mm/hugetlb.c
2437
list_add(&folio->lru, &page_list);
mm/hugetlb.c
2990
list_add(&folio->lru, &h->hugepage_activelist);
mm/hugetlb.c
3255
list_for_each_entry_safe(folio, tmp_f, folio_list, lru) {
mm/hugetlb.c
3379
list_add(&folio->lru, &folio_list);
mm/hugetlb.c
3447
list_add(&folio->lru, &folio_list);
mm/hugetlb.c
3516
list_move(&folio->lru, &folio_list);
mm/hugetlb.c
3724
list_for_each_entry_safe(folio, next, freel, lru) {
mm/hugetlb.c
3730
list_add(&folio->lru, &page_list);
mm/hugetlb.c
3884
list_add(&folio->lru, &page_list);
mm/hugetlb.c
3947
list_add(&folio->lru, &page_list);
mm/hugetlb.c
3990
list_for_each_entry_safe(folio, next, src_list, lru) {
mm/hugetlb.c
3999
list_del(&folio->lru);
mm/hugetlb.c
4017
list_add(&new_folio->lru, &dst_list);
mm/hugetlb.c
4050
list_for_each_entry_safe(folio, next, &src->hugepage_freelists[node], lru) {
mm/hugetlb.c
4055
list_add(&folio->lru, &list);
mm/hugetlb.c
4067
list_for_each_entry_safe(folio, next, &list, lru) {
mm/hugetlb.c
4068
list_del(&folio->lru);
mm/hugetlb.c
7148
list_move_tail(&folio->lru, list);
mm/hugetlb.c
7199
list_move_tail(&folio->lru, &(folio_hstate(folio))->hugepage_activelist);
mm/hugetlb.c
7250
list_move_tail(&new_folio->lru, &(folio_hstate(new_folio))->hugepage_activelist);
mm/hugetlb_cgroup.c
231
list_for_each_entry(folio, &h->hugepage_activelist, lru)
mm/hugetlb_cgroup.c
906
list_move(&new_folio->lru, &h->hugepage_activelist);
mm/hugetlb_vmemmap.c
204
list_for_each_entry_safe(page, next, list, lru)
mm/hugetlb_vmemmap.c
222
list_del(&walk->reuse_page->lru);
mm/hugetlb_vmemmap.c
233
list_add(&page->lru, walk->vmemmap_pages);
mm/hugetlb_vmemmap.c
265
page = list_first_entry(walk->vmemmap_pages, struct page, lru);
mm/hugetlb_vmemmap.c
266
list_del(&page->lru);
mm/hugetlb_vmemmap.c
346
list_add(&walk.reuse_page->lru, vmemmap_pages);
mm/hugetlb_vmemmap.c
400
list_add(&page->lru, list);
mm/hugetlb_vmemmap.c
406
list_for_each_entry_safe(page, next, list, lru)
mm/hugetlb_vmemmap.c
527
list_for_each_entry_safe(folio, t_folio, folio_list, lru) {
mm/hugetlb_vmemmap.c
539
list_move(&folio->lru, non_hvo_folios);
mm/hugetlb_vmemmap.c
663
list_for_each_entry(folio, folio_list, lru) {
mm/hugetlb_vmemmap.c
709
list_for_each_entry(folio, folio_list, lru) {
mm/khugepaged.c
2057
list_add_tail(&folio->lru, &pagelist);
mm/khugepaged.c
2105
list_for_each_entry(folio, &pagelist, lru) {
mm/khugepaged.c
2227
list_for_each_entry_safe(folio, tmp, &pagelist, lru) {
mm/khugepaged.c
2228
list_del(&folio->lru);
mm/khugepaged.c
2252
list_for_each_entry_safe(folio, tmp, &pagelist, lru) {
mm/khugepaged.c
2253
list_del(&folio->lru);
mm/khugepaged.c
525
list_for_each_entry_safe(folio, tmp, compound_pagelist, lru) {
mm/khugepaged.c
526
list_del(&folio->lru);
mm/khugepaged.c
592
list_for_each_entry(f, compound_pagelist, lru) {
mm/khugepaged.c
642
list_add_tail(&folio->lru, compound_pagelist);
mm/khugepaged.c
724
list_for_each_entry_safe(src, tmp, compound_pagelist, lru) {
mm/khugepaged.c
725
list_del(&src->lru);
mm/list_lru.c
113
static void list_lru_register(struct list_lru *lru)
mm/list_lru.c
117
static void list_lru_unregister(struct list_lru *lru)
mm/list_lru.c
121
static int lru_shrinker_id(struct list_lru *lru)
mm/list_lru.c
126
static inline bool list_lru_memcg_aware(struct list_lru *lru)
mm/list_lru.c
132
list_lru_from_memcg_idx(struct list_lru *lru, int nid, int idx)
mm/list_lru.c
134
return &lru->node[nid].lru;
mm/list_lru.c
138
lock_list_lru_of_memcg(struct list_lru *lru, int nid, struct mem_cgroup *memcg,
mm/list_lru.c
141
struct list_lru_one *l = &lru->node[nid].lru;
mm/list_lru.c
161
bool list_lru_add(struct list_lru *lru, struct list_head *item, int nid,
mm/list_lru.c
164
struct list_lru_node *nlru = &lru->node[nid];
mm/list_lru.c
167
l = lock_list_lru_of_memcg(lru, nid, memcg, false, false);
mm/list_lru.c
174
set_shrinker_bit(memcg, nid, lru_shrinker_id(lru));
mm/list_lru.c
183
bool list_lru_add_obj(struct list_lru *lru, struct list_head *item)
mm/list_lru.c
188
if (list_lru_memcg_aware(lru)) {
mm/list_lru.c
190
ret = list_lru_add(lru, item, nid, mem_cgroup_from_virt(item));
mm/list_lru.c
193
ret = list_lru_add(lru, item, nid, NULL);
mm/list_lru.c
201
bool list_lru_del(struct list_lru *lru, struct list_head *item, int nid,
mm/list_lru.c
204
struct list_lru_node *nlru = &lru->node[nid];
mm/list_lru.c
206
l = lock_list_lru_of_memcg(lru, nid, memcg, false, false);
mm/list_lru.c
22
static inline bool list_lru_memcg_aware(struct list_lru *lru)
mm/list_lru.c
220
bool list_lru_del_obj(struct list_lru *lru, struct list_head *item)
mm/list_lru.c
225
if (list_lru_memcg_aware(lru)) {
mm/list_lru.c
227
ret = list_lru_del(lru, item, nid, mem_cgroup_from_virt(item));
mm/list_lru.c
230
ret = list_lru_del(lru, item, nid, NULL);
mm/list_lru.c
24
return lru->memcg_aware;
mm/list_lru.c
252
unsigned long list_lru_count_one(struct list_lru *lru,
mm/list_lru.c
259
l = list_lru_from_memcg_idx(lru, nid, memcg_kmem_id(memcg));
mm/list_lru.c
27
static void list_lru_register(struct list_lru *lru)
mm/list_lru.c
270
unsigned long list_lru_count_node(struct list_lru *lru, int nid)
mm/list_lru.c
274
nlru = &lru->node[nid];
mm/list_lru.c
280
__list_lru_walk_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg,
mm/list_lru.c
284
struct list_lru_node *nlru = &lru->node[nid];
mm/list_lru.c
29
if (!list_lru_memcg_aware(lru))
mm/list_lru.c
290
l = lock_list_lru_of_memcg(lru, nid, memcg, irq_off, true);
mm/list_lru.c
33
list_add(&lru->list, &memcg_list_lrus);
mm/list_lru.c
337
list_lru_walk_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg,
mm/list_lru.c
341
return __list_lru_walk_one(lru, nid, memcg, isolate,
mm/list_lru.c
347
list_lru_walk_one_irq(struct list_lru *lru, int nid, struct mem_cgroup *memcg,
mm/list_lru.c
351
return __list_lru_walk_one(lru, nid, memcg, isolate,
mm/list_lru.c
355
unsigned long list_lru_walk_node(struct list_lru *lru, int nid,
mm/list_lru.c
361
isolated += list_lru_walk_one(lru, nid, NULL, isolate, cb_arg,
mm/list_lru.c
365
if (*nr_to_walk > 0 && list_lru_memcg_aware(lru)) {
mm/list_lru.c
37
static void list_lru_unregister(struct list_lru *lru)
mm/list_lru.c
370
xa_for_each(&lru->xa, index, mlru) {
mm/list_lru.c
378
isolated += __list_lru_walk_one(lru, nid, memcg,
mm/list_lru.c
39
if (!list_lru_memcg_aware(lru))
mm/list_lru.c
393
static void init_one_lru(struct list_lru *lru, struct list_lru_one *l)
mm/list_lru.c
399
if (lru->key)
mm/list_lru.c
400
lockdep_set_class(&l->lock, lru->key);
mm/list_lru.c
405
static struct list_lru_memcg *memcg_init_list_lru_one(struct list_lru *lru, gfp_t gfp)
mm/list_lru.c
415
init_one_lru(lru, &mlru->node[nid]);
mm/list_lru.c
420
static inline void memcg_init_list_lru(struct list_lru *lru, bool memcg_aware)
mm/list_lru.c
423
xa_init_flags(&lru->xa, XA_FLAGS_LOCK_IRQ);
mm/list_lru.c
424
lru->memcg_aware = memcg_aware;
mm/list_lru.c
427
static void memcg_destroy_list_lru(struct list_lru *lru)
mm/list_lru.c
429
XA_STATE(xas, &lru->xa, 0);
mm/list_lru.c
43
list_del(&lru->list);
mm/list_lru.c
432
if (!list_lru_memcg_aware(lru))
mm/list_lru.c
443
static void memcg_reparent_list_lru_one(struct list_lru *lru, int nid,
mm/list_lru.c
451
dst = list_lru_from_memcg_idx(lru, nid, dst_idx);
mm/list_lru.c
458
set_shrinker_bit(dst_memcg, nid, lru_shrinker_id(lru));
mm/list_lru.c
469
struct list_lru *lru;
mm/list_lru.c
47
static int lru_shrinker_id(struct list_lru *lru)
mm/list_lru.c
473
list_for_each_entry(lru, &memcg_list_lrus, list) {
mm/list_lru.c
475
XA_STATE(xas, &lru->xa, memcg->kmemcg_id);
mm/list_lru.c
49
return lru->shrinker_id;
mm/list_lru.c
493
memcg_reparent_list_lru_one(lru, i, &mlru->node[i], parent);
mm/list_lru.c
506
struct list_lru *lru)
mm/list_lru.c
510
return idx < 0 || xa_load(&lru->xa, idx);
mm/list_lru.c
513
int memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru,
mm/list_lru.c
519
XA_STATE(xas, &lru->xa, 0);
mm/list_lru.c
521
if (!list_lru_memcg_aware(lru) || memcg_list_lru_allocated(memcg, lru))
mm/list_lru.c
53
list_lru_from_memcg_idx(struct list_lru *lru, int nid, int idx)
mm/list_lru.c
537
while (!memcg_list_lru_allocated(parent, lru)) {
mm/list_lru.c
543
mlru = memcg_init_list_lru_one(lru, gfp);
mm/list_lru.c
55
if (list_lru_memcg_aware(lru) && idx >= 0) {
mm/list_lru.c
56
struct list_lru_memcg *mlru = xa_load(&lru->xa, idx);
mm/list_lru.c
565
static inline void memcg_init_list_lru(struct list_lru *lru, bool memcg_aware)
mm/list_lru.c
569
static void memcg_destroy_list_lru(struct list_lru *lru)
mm/list_lru.c
574
int __list_lru_init(struct list_lru *lru, bool memcg_aware, struct shrinker *shrinker)
mm/list_lru.c
580
lru->shrinker_id = shrinker->id;
mm/list_lru.c
582
lru->shrinker_id = -1;
mm/list_lru.c
588
lru->node = kzalloc_objs(*lru->node, nr_node_ids);
mm/list_lru.c
589
if (!lru->node)
mm/list_lru.c
593
init_one_lru(lru, &lru->node[i].lru);
mm/list_lru.c
595
memcg_init_list_lru(lru, memcg_aware);
mm/list_lru.c
596
list_lru_register(lru);
mm/list_lru.c
60
return &lru->node[nid].lru;
mm/list_lru.c
602
void list_lru_destroy(struct list_lru *lru)
mm/list_lru.c
605
if (!lru->node)
mm/list_lru.c
608
list_lru_unregister(lru);
mm/list_lru.c
610
memcg_destroy_list_lru(lru);
mm/list_lru.c
611
kfree(lru->node);
mm/list_lru.c
612
lru->node = NULL;
mm/list_lru.c
615
lru->shrinker_id = -1;
mm/list_lru.c
80
lock_list_lru_of_memcg(struct list_lru *lru, int nid, struct mem_cgroup *memcg,
mm/list_lru.c
87
l = list_lru_from_memcg_idx(lru, nid, memcg_kmem_id(memcg));
mm/madvise.c
435
list_add(&folio->lru, &folio_list);
mm/madvise.c
550
list_add(&folio->lru, &folio_list);
mm/memcontrol-v1.c
1767
enum lru_list lru;
mm/memcontrol-v1.c
1771
for_each_lru(lru) {
mm/memcontrol-v1.c
1772
if (!(BIT(lru) & lru_mask))
mm/memcontrol-v1.c
1775
nr += lruvec_page_state(lruvec, NR_LRU_BASE + lru);
mm/memcontrol-v1.c
1777
nr += lruvec_page_state_local(lruvec, NR_LRU_BASE + lru);
mm/memcontrol-v1.c
1787
enum lru_list lru;
mm/memcontrol-v1.c
1789
for_each_lru(lru) {
mm/memcontrol-v1.c
1790
if (!(BIT(lru) & lru_mask))
mm/memcontrol-v1.c
1793
nr += memcg_page_state(memcg, NR_LRU_BASE + lru);
mm/memcontrol-v1.c
1795
nr += memcg_page_state_local(memcg, NR_LRU_BASE + lru);
mm/memcontrol.c
1290
void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru,
mm/memcontrol.c
1301
lru_size = &mz->lru_zone_size[zid][lru];
mm/memcontrol.c
1309
__func__, lruvec, lru, nr_pages, size)) {
mm/memcontrol.c
3187
bool __memcg_slab_post_alloc_hook(struct kmem_cache *s, struct list_lru *lru,
mm/memcontrol.c
3215
if (lru) {
mm/memcontrol.c
3220
ret = memcg_list_lru_alloc(memcg, lru, flags);
mm/memory-failure.c
1234
{ lru|dirty, lru|dirty, MF_MSG_DIRTY_LRU, me_pagecache_dirty },
mm/memory-failure.c
1235
{ lru|dirty, lru, MF_MSG_CLEAN_LRU, me_pagecache_clean },
mm/memory_hotplug.c
1856
mtc.nid = folio_nid(list_first_entry(&source, struct folio, lru));
mm/memory_hotplug.c
1869
list_for_each_entry(folio, &source, lru) {
mm/mempolicy.c
1261
list_add_tail(&folio->lru, foliolist);
mm/mempolicy.c
1587
list_for_each_entry(folio, &pagelist, lru) {
mm/mempolicy.c
1591
if (!list_entry_is_head(folio, &pagelist, lru)) {
mm/migrate.c
1174
list_move_tail(&src->lru, ret);
mm/migrate.c
1364
prev = dst->lru.prev;
mm/migrate.c
1365
list_del(&dst->lru);
mm/migrate.c
1408
list_del(&src->lru);
mm/migrate.c
1422
list_add(&dst->lru, prev);
mm/migrate.c
1548
list_move_tail(&src->lru, ret);
mm/migrate.c
1577
list_move_tail(&folio->lru, split_folios);
mm/migrate.c
1627
list_for_each_entry_safe(folio, folio2, from, lru) {
mm/migrate.c
1645
list_move_tail(&folio->lru, ret_folios);
mm/migrate.c
1713
dst = list_first_entry(dst_folios, struct folio, lru);
mm/migrate.c
1714
dst2 = list_next_entry(dst, lru);
mm/migrate.c
1715
list_for_each_entry_safe(folio, folio2, src_folios, lru) {
mm/migrate.c
1747
dst2 = list_next_entry(dst, lru);
mm/migrate.c
1758
dst = list_first_entry(dst_folios, struct folio, lru);
mm/migrate.c
1759
dst2 = list_next_entry(dst, lru);
mm/migrate.c
1760
list_for_each_entry_safe(folio, folio2, src_folios, lru) {
mm/migrate.c
1767
list_del(&dst->lru);
mm/migrate.c
1770
dst2 = list_next_entry(dst, lru);
mm/migrate.c
1810
list_for_each_entry_safe(folio, folio2, from, lru) {
mm/migrate.c
1871
list_move_tail(&folio->lru, ret_folios);
mm/migrate.c
1883
list_del(&folio->lru);
mm/migrate.c
1946
list_move_tail(&folio->lru, &unmap_folios);
mm/migrate.c
1947
list_add_tail(&dst->lru, &dst_folios);
mm/migrate.c
2095
list_for_each_entry_safe(folio, folio2, from, lru) {
mm/migrate.c
2098
list_move_tail(&folio->lru, &ret_folios);
mm/migrate.c
2107
list_cut_before(&folios, from, &folio2->lru);
mm/migrate.c
2256
list_add_tail(&folio->lru, pagelist);
mm/migrate.c
262
list_for_each_entry_safe(folio, folio2, l, lru) {
mm/migrate.c
267
list_del(&folio->lru);
mm/migrate.c
2731
list_add(&folio->lru, &migratepages);
mm/migrate.c
294
list_add(&folio->lru, list);
mm/mm_init.c
591
INIT_LIST_HEAD(&page->lru);
mm/mmzone.c
77
enum lru_list lru;
mm/mmzone.c
83
for_each_lru(lru)
mm/mmzone.c
84
INIT_LIST_HEAD(&lruvec->lists[lru]);
mm/page_alloc.c
6837
list_for_each_entry(page, page_list, lru)
mm/page_alloc.c
6910
list_for_each_entry_safe(page, next, &list[order], lru) {
mm/page_alloc.c
6920
list_del(&page->lru);
mm/page_alloc.c
6922
list_add_tail(&page[i].lru, &list[0]);
mm/page_alloc.c
7636
list_del(&page->lru);
mm/page_alloc.c
7669
struct page, lru);
mm/page_alloc.c
7729
list_add_tail(&page->lru, &zone->unaccepted_pages);
mm/page_reporting.c
183
list_for_each_entry_safe(page, next, list, lru) {
mm/page_reporting.c
218
if (!list_is_first(&page->lru, list))
mm/page_reporting.c
219
list_rotate_to_front(&page->lru, list);
mm/page_reporting.c
243
next = list_first_entry(list, struct page, lru);
mm/page_reporting.c
251
if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list))
mm/page_reporting.c
252
list_rotate_to_front(&next->lru, list);
mm/pgtable-generic.c
173
INIT_LIST_HEAD(&pgtable->lru);
mm/pgtable-generic.c
175
list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru);
mm/pgtable-generic.c
190
pmd_huge_pte(mm, pmdp) = list_first_entry_or_null(&pgtable->lru,
mm/pgtable-generic.c
191
struct page, lru);
mm/pgtable-generic.c
193
list_del(&pgtable->lru);
mm/shuffle.c
135
list_swap(&page_i->lru, &page_j->lru);
mm/slab.h
634
bool __memcg_slab_post_alloc_hook(struct kmem_cache *s, struct list_lru *lru,
mm/slub.c
2448
bool memcg_slab_post_alloc_hook(struct kmem_cache *s, struct list_lru *lru,
mm/slub.c
2457
if (likely(__memcg_slab_post_alloc_hook(s, lru, flags, size, p)))
mm/slub.c
2550
struct list_lru *lru,
mm/slub.c
4498
bool slab_post_alloc_hook(struct kmem_cache *s, struct list_lru *lru,
mm/slub.c
4549
return memcg_slab_post_alloc_hook(s, lru, flags, size, p);
mm/slub.c
4837
static __fastpath_inline void *slab_alloc_node(struct kmem_cache *s, struct list_lru *lru,
mm/slub.c
4866
slab_post_alloc_hook(s, lru, gfpflags, 1, &object, init, orig_size);
mm/slub.c
4882
void *kmem_cache_alloc_lru_noprof(struct kmem_cache *s, struct list_lru *lru,
mm/slub.c
4885
void *ret = slab_alloc_node(s, lru, gfpflags, NUMA_NO_NODE, _RET_IP_,
mm/swapfile.c
1780
page = list_next_entry(page, lru);
mm/swapfile.c
3757
INIT_LIST_HEAD(&head->lru);
mm/swapfile.c
3762
list_for_each_entry(list_page, &head->lru, lru) {
mm/swapfile.c
3784
list_add_tail(&page->lru, &head->lru);
mm/swapfile.c
3822
page = list_next_entry(head, lru);
mm/swapfile.c
3834
page = list_next_entry(page, lru);
mm/swapfile.c
3840
page = list_next_entry(page, lru);
mm/swapfile.c
3850
while ((page = list_prev_entry(page, lru)) != head) {
mm/swapfile.c
3864
page = list_next_entry(page, lru);
mm/swapfile.c
3873
while ((page = list_prev_entry(page, lru)) != head) {
mm/swapfile.c
3900
list_for_each_entry_safe(page, next, &head->lru, lru) {
mm/swapfile.c
3901
list_del(&page->lru);
mm/vmscan.c
1112
list_del(&folio->lru);
mm/vmscan.c
1251
list_add_tail(&folio->lru, folio_list);
mm/vmscan.c
1276
list_add(&folio->lru, &demote_folios);
mm/vmscan.c
1569
list_add(&folio->lru, &ret_folios);
mm/vmscan.c
1633
list_for_each_entry_safe(folio, next, folio_list, lru) {
mm/vmscan.c
1640
list_move(&folio->lru, &clean_folios);
mm/vmscan.c
1676
enum lru_list lru, unsigned long *nr_zone_taken)
mm/vmscan.c
1684
update_lru_size(lruvec, lru, zid, -nr_zone_taken[zid]);
mm/vmscan.c
1713
enum lru_list lru)
mm/vmscan.c
1715
struct list_head *src = &lruvec->lists[lru];
mm/vmscan.c
1775
list_move(&folio->lru, move_to);
mm/vmscan.c
1799
total_scan, skipped, nr_taken, lru);
mm/vmscan.c
1800
update_lru_sizes(lruvec, lru, nr_zone_taken);
mm/vmscan.c
187
if ((_folio)->lru.prev != _base) { \
mm/vmscan.c
190
prev = lru_to_folio(&(_folio->lru)); \
mm/vmscan.c
1906
list_del(&folio->lru);
mm/vmscan.c
1979
enum lru_list lru)
mm/vmscan.c
1986
bool file = is_file_lru(lru);
mm/vmscan.c
2009
&nr_scanned, sc, lru);
mm/vmscan.c
2101
enum lru_list lru)
mm/vmscan.c
2111
bool file = is_file_lru(lru);
mm/vmscan.c
2119
&nr_scanned, sc, lru);
mm/vmscan.c
2134
list_del(&folio->lru);
mm/vmscan.c
2163
list_add(&folio->lru, &l_active);
mm/vmscan.c
2170
list_add(&folio->lru, &l_inactive);
mm/vmscan.c
2208
list_del(&folio->lru);
mm/vmscan.c
2234
list_move(&folio->lru, &node_folio_list);
mm/vmscan.c
2249
static unsigned long shrink_list(enum lru_list lru, unsigned long nr_to_scan,
mm/vmscan.c
2252
if (is_active_lru(lru)) {
mm/vmscan.c
2253
if (sc->may_deactivate & (1 << is_file_lru(lru)))
mm/vmscan.c
2254
shrink_active_list(nr_to_scan, lruvec, sc, lru);
mm/vmscan.c
2260
return shrink_inactive_list(nr_to_scan, lruvec, sc, lru);
mm/vmscan.c
2536
enum lru_list lru;
mm/vmscan.c
2595
for_each_evictable_lru(lru) {
mm/vmscan.c
2596
bool file = is_file_lru(lru);
mm/vmscan.c
2600
lruvec_size = lruvec_lru_size(lruvec, lru, sc->reclaim_idx);
mm/vmscan.c
2639
nr[lru] = scan;
mm/vmscan.c
3309
enum lru_list lru = type * LRU_INACTIVE_FILE;
mm/vmscan.c
3320
lru += LRU_ACTIVE;
mm/vmscan.c
3321
__update_lru_size(lruvec, lru, zone, delta);
mm/vmscan.c
3877
list_move_tail(&folio->lru, &lrugen->folios[new_gen][type][zone]);
mm/vmscan.c
3997
enum lru_list lru = type * LRU_INACTIVE_FILE;
mm/vmscan.c
4004
__update_lru_size(lruvec, lru, zone, delta);
mm/vmscan.c
4005
__update_lru_size(lruvec, lru + LRU_ACTIVE, zone, -delta);
mm/vmscan.c
412
static unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru,
mm/vmscan.c
421
size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid);
mm/vmscan.c
423
size += zone_page_state(zone, NR_ZONE_LRU_BASE + lru);
mm/vmscan.c
4463
list_move(&folio->lru, &lrugen->folios[gen][type][zone]);
mm/vmscan.c
4470
list_move(&folio->lru, &lrugen->folios[gen][type][zone]);
mm/vmscan.c
4485
list_move_tail(&folio->lru, &lrugen->folios[gen][type][zone]);
mm/vmscan.c
4500
list_move(&folio->lru, &lrugen->folios[gen][type][zone]);
mm/vmscan.c
4583
list_add(&folio->lru, list);
mm/vmscan.c
4586
list_move(&folio->lru, &moved);
mm/vmscan.c
4725
list_for_each_entry_safe_reverse(folio, next, &list, lru) {
mm/vmscan.c
4729
list_del(&folio->lru);
mm/vmscan.c
4737
list_move(&folio->lru, &clean);
mm/vmscan.c
5089
enum lru_list lru;
mm/vmscan.c
5091
for_each_evictable_lru(lru) {
mm/vmscan.c
5092
if (!list_empty(&lruvec->lists[lru]))
mm/vmscan.c
5109
enum lru_list lru;
mm/vmscan.c
5112
for_each_evictable_lru(lru) {
mm/vmscan.c
5113
int type = is_file_lru(lru);
mm/vmscan.c
5114
bool active = is_active_lru(lru);
mm/vmscan.c
5115
struct list_head *head = &lruvec->lists[lru];
mm/vmscan.c
5777
enum lru_list lru;
mm/vmscan.c
5813
for_each_evictable_lru(lru) {
mm/vmscan.c
5814
if (nr[lru]) {
mm/vmscan.c
5815
nr_to_scan = min(nr[lru], SWAP_CLUSTER_MAX);
mm/vmscan.c
5816
nr[lru] -= nr_to_scan;
mm/vmscan.c
5818
nr_reclaimed += shrink_list(lru, nr_to_scan,
mm/vmscan.c
5850
lru = LRU_BASE;
mm/vmscan.c
5855
lru = LRU_FILE;
mm/vmscan.c
5860
nr[lru] = 0;
mm/vmscan.c
5861
nr[lru + LRU_ACTIVE] = 0;
mm/vmscan.c
5867
lru = (lru == LRU_FILE) ? LRU_BASE : LRU_FILE;
mm/vmscan.c
5868
nr_scanned = targets[lru] - nr[lru];
mm/vmscan.c
5869
nr[lru] = targets[lru] * (100 - percentage) / 100;
mm/vmscan.c
5870
nr[lru] -= min(nr[lru], nr_scanned);
mm/vmscan.c
5872
lru += LRU_ACTIVE;
mm/vmscan.c
5873
nr_scanned = targets[lru] - nr[lru];
mm/vmscan.c
5874
nr[lru] = targets[lru] * (100 - percentage) / 100;
mm/vmscan.c
5875
nr[lru] -= min(nr[lru], nr_scanned);
mm/workingset.c
699
struct list_lru_one *lru,
mm/workingset.c
700
void *arg) __must_hold(lru->lock)
mm/workingset.c
722
spin_unlock_irq(&lru->lock);
mm/workingset.c
731
spin_unlock_irq(&lru->lock);
mm/workingset.c
737
list_lru_isolate(lru, item);
mm/workingset.c
740
spin_unlock(&lru->lock);
mm/zpdesc.h
34
struct list_head lru;
mm/zpdesc.h
55
ZPDESC_MATCH(lru, lru);
mm/zswap.c
1103
struct zswap_entry *entry = container_of(item, struct zswap_entry, lru);
mm/zswap.c
1475
INIT_LIST_HEAD(&entry->lru);
mm/zswap.c
197
struct list_head lru;
mm/zswap.c
641
list_lru_add(list_lru, &entry->lru, nid, memcg);
mm/zswap.c
653
list_lru_del(list_lru, &entry->lru, nid, memcg);
net/ceph/messenger.c
913
page = list_first_entry(&pagelist->head, struct page, lru);
net/ceph/messenger.c
968
BUG_ON(list_is_last(&cursor->page->lru, &pagelist->head));
net/ceph/messenger.c
969
cursor->page = list_next_entry(cursor->page, lru);
net/ceph/pagelist.c
113
list_add_tail(&page->lru, &pl->free_list);
net/ceph/pagelist.c
125
struct page, lru);
net/ceph/pagelist.c
126
list_del(&page->lru);
net/ceph/pagelist.c
32
struct page *page = list_entry(pl->head.prev, struct page, lru);
net/ceph/pagelist.c
45
lru);
net/ceph/pagelist.c
46
list_del(&page->lru);
net/ceph/pagelist.c
61
page = list_first_entry(&pl->free_list, struct page, lru);
net/ceph/pagelist.c
62
list_del(&page->lru);
net/ceph/pagelist.c
69
list_add_tail(&page->lru, &pl->head);
rust/helpers/binder.c
10
__rust_helper unsigned long rust_helper_list_lru_count(struct list_lru *lru)
rust/helpers/binder.c
12
return list_lru_count(lru);
rust/helpers/binder.c
15
__rust_helper unsigned long rust_helper_list_lru_walk(struct list_lru *lru,
rust/helpers/binder.c
20
return list_lru_walk(lru, isolate, cb_arg, nr_to_walk);
samples/bpf/test_lru_dist.c
102
lru->map_fd = bpf_map_create(BPF_MAP_TYPE_HASH, NULL,
samples/bpf/test_lru_dist.c
106
assert(lru->map_fd != -1);
samples/bpf/test_lru_dist.c
108
lru->free_nodes = malloc(lru_size * sizeof(struct pfect_lru_node));
samples/bpf/test_lru_dist.c
109
assert(lru->free_nodes);
samples/bpf/test_lru_dist.c
111
INIT_LIST_HEAD(&lru->list);
samples/bpf/test_lru_dist.c
112
lru->cur_size = 0;
samples/bpf/test_lru_dist.c
113
lru->lru_size = lru_size;
samples/bpf/test_lru_dist.c
114
lru->nr_unique = lru->nr_misses = lru->total = 0;
samples/bpf/test_lru_dist.c
117
static void pfect_lru_destroy(struct pfect_lru *lru)
samples/bpf/test_lru_dist.c
119
close(lru->map_fd);
samples/bpf/test_lru_dist.c
120
free(lru->free_nodes);
samples/bpf/test_lru_dist.c
123
static int pfect_lru_lookup_or_insert(struct pfect_lru *lru,
samples/bpf/test_lru_dist.c
129
lru->total++;
samples/bpf/test_lru_dist.c
130
if (!bpf_map_lookup_elem(lru->map_fd, &key, &node)) {
samples/bpf/test_lru_dist.c
132
list_move(&node->list, &lru->list);
samples/bpf/test_lru_dist.c
138
if (lru->cur_size < lru->lru_size) {
samples/bpf/test_lru_dist.c
139
node = &lru->free_nodes[lru->cur_size++];
samples/bpf/test_lru_dist.c
144
node = list_last_entry(&lru->list,
samples/bpf/test_lru_dist.c
147
bpf_map_update_elem(lru->map_fd, &node->key, &null_node, BPF_EXIST);
samples/bpf/test_lru_dist.c
151
list_move(&node->list, &lru->list);
samples/bpf/test_lru_dist.c
153
lru->nr_misses++;
samples/bpf/test_lru_dist.c
155
assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_EXIST));
samples/bpf/test_lru_dist.c
157
lru->nr_unique++;
samples/bpf/test_lru_dist.c
158
assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_NOEXIST));
samples/bpf/test_lru_dist.c
99
static void pfect_lru_init(struct pfect_lru *lru, unsigned int lru_size,
tools/testing/selftests/bpf/progs/timer.c
155
bpf_map_update_elem(&lru, &lru_key, &init, 0);
tools/testing/selftests/bpf/progs/timer.c
156
lru_timer = bpf_map_lookup_elem(&lru, &lru_key);
tools/testing/selftests/bpf/progs/timer.c
159
bpf_timer_init(lru_timer, &lru, CLOCK_MONOTONIC);
tools/testing/selftests/bpf/progs/timer.c
52
} lru SEC(".maps");
tools/testing/selftests/bpf/progs/timer.c
99
lru_timer = bpf_map_lookup_elem(&lru, &lru_key);
tools/testing/selftests/bpf/progs/wq.c
188
return test_elem_callback(&lru, &key, wq_callback);
tools/testing/selftests/bpf/progs/wq.c
51
} lru SEC(".maps");
tools/testing/selftests/bpf/progs/wq.c
66
if (map == &lru &&
tools/testing/selftests/bpf/progs/wq_failures.c
28
} lru SEC(".maps");
tools/testing/selftests/bpf/progs/wq_failures.c
88
if (bpf_wq_init(wq, &lru, 0) != 0)
tools/testing/shared/linux.c
54
void *kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *lru,