Symbol: nr_to_scan
drivers/android/binder_alloc.c
1227
NULL, sc->nr_to_scan);
drivers/gpu/drm/drm_gem.c
1626
unsigned int nr_to_scan,
drivers/gpu/drm/drm_gem.c
1639
while (freed < nr_to_scan) {
drivers/gpu/drm/drm_pagemap_util.c
390
} while (sc->nr_scanned < sc->nr_to_scan);
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
326
sc->nr_to_scan,
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
330
if (sc->nr_scanned < sc->nr_to_scan && current_is_kswapd()) {
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
335
sc->nr_to_scan - sc->nr_scanned,
drivers/gpu/drm/msm/msm_drv.h
241
unsigned long msm_gem_shrinker_shrink(struct drm_device *dev, unsigned long nr_to_scan);
drivers/gpu/drm/msm/msm_gem_shrinker.c
182
long nr = sc->nr_to_scan;
drivers/gpu/drm/msm/msm_gem_shrinker.c
200
trace_msm_gem_shrink(sc->nr_to_scan, stages[0].freed,
drivers/gpu/drm/msm/msm_gem_shrinker.c
210
msm_gem_shrinker_shrink(struct drm_device *dev, unsigned long nr_to_scan)
drivers/gpu/drm/msm/msm_gem_shrinker.c
214
.nr_to_scan = nr_to_scan,
drivers/gpu/drm/msm/msm_gpu_trace.h
119
TP_PROTO(u32 nr_to_scan, u32 purged, u32 evicted,
drivers/gpu/drm/msm/msm_gpu_trace.h
121
TP_ARGS(nr_to_scan, purged, evicted, active_purged, active_evicted),
drivers/gpu/drm/msm/msm_gpu_trace.h
123
__field(u32, nr_to_scan)
drivers/gpu/drm/msm/msm_gpu_trace.h
130
__entry->nr_to_scan = nr_to_scan;
drivers/gpu/drm/msm/msm_gpu_trace.h
137
__entry->nr_to_scan, __entry->purged, __entry->evicted,
drivers/gpu/drm/panfrost/panfrost_gem_shrinker.c
75
if (freed >= sc->nr_to_scan)
drivers/gpu/drm/ttm/ttm_pool.c
1155
while (num_freed < sc->nr_to_scan &&
drivers/gpu/drm/ttm/ttm_pool.c
1291
.nr_to_scan = TTM_SHRINKER_BATCH,
drivers/gpu/drm/xe/xe_shrinker.c
167
unsigned long nr_to_scan, bool can_backup)
drivers/gpu/drm/xe/xe_shrinker.c
177
force = (nr_to_scan > shrinker->purgeable_pages && can_backup);
drivers/gpu/drm/xe/xe_shrinker.c
208
unsigned long nr_to_scan, nr_scanned = 0, freed = 0;
drivers/gpu/drm/xe/xe_shrinker.c
219
nr_to_scan = sc->nr_to_scan;
drivers/gpu/drm/xe/xe_shrinker.c
226
runtime_pm = xe_shrinker_runtime_pm_get(shrinker, false, nr_to_scan, can_backup);
drivers/gpu/drm/xe/xe_shrinker.c
228
if (purgeable && nr_scanned < nr_to_scan) {
drivers/gpu/drm/xe/xe_shrinker.c
230
nr_to_scan, &nr_scanned);
drivers/gpu/drm/xe/xe_shrinker.c
236
if (nr_scanned >= nr_to_scan || !can_backup)
drivers/gpu/drm/xe/xe_shrinker.c
246
nr_to_scan, &nr_scanned);
drivers/md/bcache/btree.c
670
unsigned long i, nr = sc->nr_to_scan;
drivers/md/bcache/sysfs.c
861
sc.nr_to_scan = strtoul_or_return(buf);
drivers/md/dm-bufio.c
2466
atomic_long_add(sc->nr_to_scan, &c->need_shrink);
drivers/md/dm-bufio.c
2469
return sc->nr_to_scan;
drivers/md/dm-zoned-metadata.c
628
count = dmz_shrink_mblock_cache(zmd, sc->nr_to_scan);
drivers/md/raid5.c
7433
while (ret < sc->nr_to_scan &&
drivers/misc/vmw_balloon.c
1530
deflated_frames = vmballoon_deflate(b, sc->nr_to_scan, true);
drivers/virtio/virtio_balloon.c
874
return shrink_free_pages(vb, sc->nr_to_scan);
fs/btrfs/extent_map.c
1130
long nr_to_scan;
fs/btrfs/extent_map.c
1188
if (ctx->scanned >= ctx->nr_to_scan)
fs/btrfs/extent_map.c
1272
if (ctx->scanned >= ctx->nr_to_scan || btrfs_fs_closing(fs_info))
fs/btrfs/extent_map.c
1313
ctx.nr_to_scan = atomic64_read(&fs_info->em_shrinker_nr_to_scan);
fs/btrfs/extent_map.c
1324
while (ctx.scanned < ctx.nr_to_scan && !btrfs_fs_closing(fs_info)) {
fs/btrfs/extent_map.c
1367
void btrfs_free_extent_maps(struct btrfs_fs_info *fs_info, long nr_to_scan)
fs/btrfs/extent_map.c
1384
if (atomic64_cmpxchg(&fs_info->em_shrinker_nr_to_scan, 0, nr_to_scan) != 0)
fs/btrfs/extent_map.h
192
void btrfs_free_extent_maps(struct btrfs_fs_info *fs_info, long nr_to_scan);
fs/btrfs/super.c
2417
const long nr_to_scan = min_t(unsigned long, LONG_MAX, sc->nr_to_scan);
fs/btrfs/super.c
2420
btrfs_free_extent_maps(fs_info, nr_to_scan);
fs/ceph/dir.c
1637
unsigned long nr_to_scan;
fs/ceph/dir.c
1658
if (!lwc->nr_to_scan)
fs/ceph/dir.c
1660
--lwc->nr_to_scan;
fs/ceph/dir.c
1789
lwc.nr_to_scan = CEPH_CAPS_PER_RELEASE * 2;
fs/ceph/dir.c
1791
if (!lwc.nr_to_scan) /* more invalid leases */
fs/ceph/dir.c
1794
if (lwc.nr_to_scan < CEPH_CAPS_PER_RELEASE)
fs/ceph/dir.c
1795
lwc.nr_to_scan = CEPH_CAPS_PER_RELEASE;
fs/ceph/dir.c
1801
if (!lwc.nr_to_scan) /* more to check */
fs/erofs/zutil.c
254
unsigned long nr = sc->nr_to_scan;
fs/ext4/extents_status.c
1672
static int __es_shrink(struct ext4_sb_info *sbi, int nr_to_scan,
fs/ext4/extents_status.c
1719
nr_shrunk += es_reclaim_extents(ei, &nr_to_scan);
fs/ext4/extents_status.c
1722
if (nr_to_scan <= 0)
fs/ext4/extents_status.c
1738
nr_shrunk = es_reclaim_extents(locked_ei, &nr_to_scan);
fs/ext4/extents_status.c
1768
trace_ext4_es_shrink_count(sbi->s_sb, sc->nr_to_scan, nr);
fs/ext4/extents_status.c
1776
int nr_to_scan = sc->nr_to_scan;
fs/ext4/extents_status.c
1780
trace_ext4_es_shrink_scan_enter(sbi->s_sb, nr_to_scan, ret);
fs/ext4/extents_status.c
1782
nr_shrunk = __es_shrink(sbi, nr_to_scan, NULL);
fs/ext4/extents_status.c
185
static int es_reclaim_extents(struct ext4_inode_info *ei, int *nr_to_scan);
fs/ext4/extents_status.c
186
static int __es_shrink(struct ext4_sb_info *sbi, int nr_to_scan,
fs/ext4/extents_status.c
1901
int *nr_to_scan, int *nr_shrunk)
fs/ext4/extents_status.c
1912
while (*nr_to_scan > 0) {
fs/ext4/extents_status.c
1918
(*nr_to_scan)--;
fs/ext4/extents_status.c
1943
static int es_reclaim_extents(struct ext4_inode_info *ei, int *nr_to_scan)
fs/ext4/extents_status.c
1958
if (!es_do_reclaim_extents(ei, EXT_MAX_BLOCKS, nr_to_scan, &nr_shrunk) &&
fs/ext4/extents_status.c
1960
es_do_reclaim_extents(ei, start - 1, nr_to_scan, &nr_shrunk);
fs/f2fs/shrinker.c
82
unsigned long nr = sc->nr_to_scan;
fs/gfs2/glock.c
1985
return gfs2_scan_glock_lru(sc->nr_to_scan);
fs/jbd2/checkpoint.c
420
unsigned long *nr_to_scan)
fs/jbd2/checkpoint.c
463
(*nr_to_scan) -= min(*nr_to_scan, freed);
fs/jbd2/checkpoint.c
464
if (*nr_to_scan == 0)
fs/jbd2/checkpoint.c
481
if (*nr_to_scan && journal->j_shrink_transaction)
fs/jbd2/journal.c
1259
unsigned long nr_to_scan = sc->nr_to_scan;
fs/jbd2/journal.c
1264
trace_jbd2_shrink_scan_enter(journal, sc->nr_to_scan, count);
fs/jbd2/journal.c
1266
nr_shrunk = jbd2_journal_shrink_checkpoint_list(journal, &nr_to_scan);
fs/jbd2/journal.c
1269
trace_jbd2_shrink_scan_exit(journal, nr_to_scan, nr_shrunk, count);
fs/jbd2/journal.c
1288
trace_jbd2_shrink_count(journal, sc->nr_to_scan, count);
fs/mbcache.c
303
unsigned long nr_to_scan)
fs/mbcache.c
309
while (nr_to_scan-- && !list_empty(&cache->c_list)) {
fs/mbcache.c
336
return mb_cache_shrink(cache, sc->nr_to_scan);
fs/mbcache.c
48
unsigned long nr_to_scan);
fs/nfs/dir.c
2904
nfs_do_access_cache_scan(unsigned int nr_to_scan)
fs/nfs/dir.c
2915
if (nr_to_scan-- == 0)
fs/nfs/dir.c
2946
int nr_to_scan = sc->nr_to_scan;
fs/nfs/dir.c
2951
return nfs_do_access_cache_scan(nr_to_scan);
fs/nfs/dir.c
2966
unsigned int nr_to_scan;
fs/nfs/dir.c
2970
nr_to_scan = 100;
fs/nfs/dir.c
2972
if (diff < nr_to_scan)
fs/nfs/dir.c
2973
nr_to_scan = diff;
fs/nfs/dir.c
2974
nfs_do_access_cache_scan(nr_to_scan);
fs/nfsd/nfs4state.c
2216
if (scanned >= sc->nr_to_scan) {
fs/nfsd/nfscache.c
331
if (freed > sc->nr_to_scan)
fs/quota/dquot.c
791
while (!list_empty(&free_dquots) && sc->nr_to_scan) {
fs/quota/dquot.c
797
sc->nr_to_scan--;
fs/super.c
211
dentries = mult_frac(sc->nr_to_scan, dentries, total_objects);
fs/super.c
212
inodes = mult_frac(sc->nr_to_scan, inodes, total_objects);
fs/super.c
213
fs_objects = mult_frac(sc->nr_to_scan, fs_objects, total_objects);
fs/super.c
222
sc->nr_to_scan = dentries + 1;
fs/super.c
224
sc->nr_to_scan = inodes + 1;
fs/super.c
228
sc->nr_to_scan = fs_objects + 1;
fs/ubifs/shrinker.c
282
unsigned long nr = sc->nr_to_scan;
fs/xfs/xfs_icache.c
1096
unsigned long nr_to_scan)
fs/xfs/xfs_icache.c
1100
.icw_scan_limit = min_t(unsigned long, LONG_MAX, nr_to_scan),
fs/xfs/xfs_icache.h
56
long xfs_reclaim_inodes_nr(struct xfs_mount *mp, unsigned long nr_to_scan);
fs/xfs/xfs_super.c
1253
return xfs_reclaim_inodes_nr(XFS_M(sb), sc->nr_to_scan);
fs/xfs/xfs_trace.h
575
__field(unsigned long, nr_to_scan)
fs/xfs/xfs_trace.h
580
__entry->nr_to_scan = sc->nr_to_scan;
fs/xfs/xfs_trace.h
585
__entry->nr_to_scan,
include/drm/drm_gem.h
619
unsigned int nr_to_scan,
include/linux/jbd2.h
1434
unsigned long jbd2_journal_shrink_checkpoint_list(journal_t *journal, unsigned long *nr_to_scan);
include/linux/shrinker.h
45
unsigned long nr_to_scan;
include/trace/events/btrfs.h
2557
__field( long, nr_to_scan )
include/trace/events/btrfs.h
2564
__entry->nr_to_scan = \
include/trace/events/btrfs.h
2572
__entry->nr_to_scan, __entry->nr,
include/trace/events/ext4.h
2389
TP_PROTO(struct super_block *sb, int nr_to_scan, int cache_cnt),
include/trace/events/ext4.h
2391
TP_ARGS(sb, nr_to_scan, cache_cnt),
include/trace/events/ext4.h
2395
__field( int, nr_to_scan )
include/trace/events/ext4.h
2401
__entry->nr_to_scan = nr_to_scan;
include/trace/events/ext4.h
2407
__entry->nr_to_scan, __entry->cache_cnt)
include/trace/events/ext4.h
2411
TP_PROTO(struct super_block *sb, int nr_to_scan, int cache_cnt),
include/trace/events/ext4.h
2413
TP_ARGS(sb, nr_to_scan, cache_cnt)
include/trace/events/ext4.h
2417
TP_PROTO(struct super_block *sb, int nr_to_scan, int cache_cnt),
include/trace/events/ext4.h
2419
TP_ARGS(sb, nr_to_scan, cache_cnt)
include/trace/events/jbd2.h
399
TP_PROTO(journal_t *journal, unsigned long nr_to_scan,
include/trace/events/jbd2.h
402
TP_ARGS(journal, nr_to_scan, count),
include/trace/events/jbd2.h
406
__field(unsigned long, nr_to_scan)
include/trace/events/jbd2.h
412
__entry->nr_to_scan = nr_to_scan;
include/trace/events/jbd2.h
418
__entry->nr_to_scan, __entry->count)
include/trace/events/jbd2.h
423
TP_PROTO(journal_t *journal, unsigned long nr_to_scan, unsigned long count),
include/trace/events/jbd2.h
425
TP_ARGS(journal, nr_to_scan, count)
include/trace/events/jbd2.h
430
TP_PROTO(journal_t *journal, unsigned long nr_to_scan, unsigned long count),
include/trace/events/jbd2.h
432
TP_ARGS(journal, nr_to_scan, count)
include/trace/events/jbd2.h
437
TP_PROTO(journal_t *journal, unsigned long nr_to_scan,
include/trace/events/jbd2.h
440
TP_ARGS(journal, nr_to_scan, nr_shrunk, count),
include/trace/events/jbd2.h
444
__field(unsigned long, nr_to_scan)
include/trace/events/jbd2.h
451
__entry->nr_to_scan = nr_to_scan;
include/trace/events/jbd2.h
458
__entry->nr_to_scan, __entry->nr_shrunk,
kernel/rcu/tree_nocb.h
1250
sc->nr_to_scan -= _count;
kernel/rcu/tree_nocb.h
1252
if (sc->nr_to_scan <= 0)
mm/huge_memory.c
4433
if (!--sc->nr_to_scan)
mm/huge_memory.c
4485
if (sc->nr_to_scan && !list_empty(&ds_queue->split_queue)) {
mm/shmem.c
734
unsigned long batch = sc ? sc->nr_to_scan : 128;
mm/shrinker.c
433
unsigned long nr_to_scan = min(batch_size, total_scan);
mm/shrinker.c
435
shrinkctl->nr_to_scan = nr_to_scan;
mm/shrinker.c
436
shrinkctl->nr_scanned = nr_to_scan;
mm/shrinker_debug.c
109
unsigned long nr_to_scan = 0, read_len;
mm/shrinker_debug.c
123
if (sscanf(kbuf, "%llu %d %lu", &id, &nid, &nr_to_scan) != 3)
mm/shrinker_debug.c
129
if (nr_to_scan == 0)
mm/shrinker_debug.c
147
sc.nr_to_scan = nr_to_scan;
mm/shrinker_debug.c
148
sc.nr_scanned = nr_to_scan;
mm/slab_common.c
2156
sc->nr_to_scan -= count;
mm/slab_common.c
2159
if (sc->nr_to_scan <= 0)
mm/vmscan.c
1710
static unsigned long isolate_lru_folios(unsigned long nr_to_scan,
mm/vmscan.c
1724
while (scan < nr_to_scan && !list_empty(src)) {
mm/vmscan.c
1798
trace_mm_vmscan_lru_isolate(sc->reclaim_idx, sc->order, nr_to_scan,
mm/vmscan.c
1977
static unsigned long shrink_inactive_list(unsigned long nr_to_scan,
mm/vmscan.c
2008
nr_taken = isolate_lru_folios(nr_to_scan, lruvec, &folio_list,
mm/vmscan.c
2098
static void shrink_active_list(unsigned long nr_to_scan,
mm/vmscan.c
2118
nr_taken = isolate_lru_folios(nr_to_scan, lruvec, &l_hold,
mm/vmscan.c
2249
static unsigned long shrink_list(enum lru_list lru, unsigned long nr_to_scan,
mm/vmscan.c
2254
shrink_active_list(nr_to_scan, lruvec, sc, lru);
mm/vmscan.c
2260
return shrink_inactive_list(nr_to_scan, lruvec, sc, lru);
mm/vmscan.c
4540
static int scan_folios(unsigned long nr_to_scan, struct lruvec *lruvec,
mm/vmscan.c
4551
int scan_batch = min(nr_to_scan, MAX_LRU_BATCH);
mm/vmscan.c
4663
static int isolate_folios(unsigned long nr_to_scan, struct lruvec *lruvec,
mm/vmscan.c
4676
scanned = scan_folios(nr_to_scan, lruvec, sc, type, tier, list);
mm/vmscan.c
4686
static int evict_folios(unsigned long nr_to_scan, struct lruvec *lruvec,
mm/vmscan.c
4706
scanned = isolate_folios(nr_to_scan, lruvec, sc, swappiness, &type, &list);
mm/vmscan.c
4778
int swappiness, unsigned long *nr_to_scan)
mm/vmscan.c
4785
*nr_to_scan = 0;
mm/vmscan.c
4801
*nr_to_scan = size;
mm/vmscan.c
4814
unsigned long nr_to_scan;
mm/vmscan.c
4821
success = should_run_aging(lruvec, max_seq, swappiness, &nr_to_scan);
mm/vmscan.c
4824
if (nr_to_scan && !mem_cgroup_online(memcg))
mm/vmscan.c
4825
return nr_to_scan;
mm/vmscan.c
4827
nr_to_scan = apply_proportional_protection(memcg, sc, nr_to_scan);
mm/vmscan.c
4831
return nr_to_scan >> sc->priority;
mm/vmscan.c
4870
long nr_to_scan;
mm/vmscan.c
4877
nr_to_scan = get_nr_to_scan(lruvec, sc, swappiness);
mm/vmscan.c
4878
if (nr_to_scan <= 0)
mm/vmscan.c
4881
delta = evict_folios(nr_to_scan, lruvec, sc, swappiness);
mm/vmscan.c
4886
if (scanned >= nr_to_scan)
mm/vmscan.c
4903
return nr_to_scan < 0;
mm/vmscan.c
5776
unsigned long nr_to_scan;
mm/vmscan.c
5815
nr_to_scan = min(nr[lru], SWAP_CLUSTER_MAX);
mm/vmscan.c
5816
nr[lru] -= nr_to_scan;
mm/vmscan.c
5818
nr_reclaimed += shrink_list(lru, nr_to_scan,
net/sunrpc/auth.c
427
rpcauth_prune_expired(struct list_head *free, int nr_to_scan)
net/sunrpc/auth.c
435
if (nr_to_scan-- == 0)
net/sunrpc/auth.c
458
rpcauth_cache_do_shrink(int nr_to_scan)
net/sunrpc/auth.c
464
freed = rpcauth_prune_expired(&free, nr_to_scan);
net/sunrpc/auth.c
485
return rpcauth_cache_do_shrink(sc->nr_to_scan);
net/sunrpc/auth.c
499
unsigned int nr_to_scan;
net/sunrpc/auth.c
504
nr_to_scan = 100;
net/sunrpc/auth.c
505
if (diff < nr_to_scan)
net/sunrpc/auth.c
506
nr_to_scan = diff;
net/sunrpc/auth.c
507
rpcauth_cache_do_shrink(nr_to_scan);