Symbol: shrink
arch/powerpc/kernel/tau_6xx.c
117
int shrink;
arch/powerpc/kernel/tau_6xx.c
130
shrink = (2 + size - min_window) / 4;
arch/powerpc/kernel/tau_6xx.c
131
if (shrink) {
arch/powerpc/kernel/tau_6xx.c
132
tau[cpu].low += shrink;
arch/powerpc/kernel/tau_6xx.c
133
tau[cpu].high -= shrink;
drivers/android/binder_alloc.c
1218
binder_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
drivers/android/binder_alloc.c
1224
binder_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
drivers/cpuidle/governors/haltpoll.c
102
val /= shrink;
drivers/cpuidle/governors/haltpoll.c
96
unsigned int shrink = guest_halt_poll_shrink;
drivers/cpuidle/governors/haltpoll.c
99
if (shrink == 0) {
drivers/dax/bus.c
966
resource_size_t shrink;
drivers/dax/bus.c
968
shrink = min_t(u64, to_shrink, range_len(range));
drivers/dax/bus.c
969
if (shrink >= range_len(range)) {
drivers/dax/bus.c
973
to_shrink -= shrink;
drivers/dax/bus.c
990
- shrink);
drivers/gpu/drm/drm_gem.c
1628
bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket),
drivers/gpu/drm/drm_gem.c
1675
if (shrink(obj, ticket)) {
drivers/gpu/drm/drm_pagemap_util.c
351
drm_pagemap_shrinker_count(struct shrinker *shrink, struct shrink_control *sc)
drivers/gpu/drm/drm_pagemap_util.c
353
struct drm_pagemap_shrinker *shrinker = shrink->private_data;
drivers/gpu/drm/drm_pagemap_util.c
360
drm_pagemap_shrinker_scan(struct shrinker *shrink, struct shrink_control *sc)
drivers/gpu/drm/drm_pagemap_util.c
362
struct drm_pagemap_shrinker *shrinker = shrink->private_data;
drivers/gpu/drm/drm_pagemap_util.c
402
shrinker_free(shrinker->shrink);
drivers/gpu/drm/drm_pagemap_util.c
420
struct shrinker *shrink;
drivers/gpu/drm/drm_pagemap_util.c
427
shrink = shrinker_alloc(0, "drm-drm_pagemap:%s", drm->unique);
drivers/gpu/drm/drm_pagemap_util.c
428
if (!shrink) {
drivers/gpu/drm/drm_pagemap_util.c
436
shrinker->shrink = shrink;
drivers/gpu/drm/drm_pagemap_util.c
437
shrink->count_objects = drm_pagemap_shrinker_count;
drivers/gpu/drm/drm_pagemap_util.c
438
shrink->scan_objects = drm_pagemap_shrinker_scan;
drivers/gpu/drm/drm_pagemap_util.c
439
shrink->private_data = shrinker;
drivers/gpu/drm/drm_pagemap_util.c
440
shrinker_register(shrink);
drivers/gpu/drm/drm_pagemap_util.c
55
struct shrinker *shrink;
drivers/gpu/drm/i915/gem/i915_gem_object_types.h
81
int (*shrink)(struct drm_i915_gem_object *obj, unsigned int flags);
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
106
const unsigned int shrink[] = {
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
109
}, *s = shrink;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
495
.shrink = shmem_shrink,
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
108
unsigned int shrink)
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
130
trace_i915_gem_shrink(i915, target, shrink);
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
137
if (shrink & I915_SHRINK_BOUND) {
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
140
shrink &= ~I915_SHRINK_BOUND;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
153
if (shrink & I915_SHRINK_ACTIVE) {
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
183
if ((shrink & phase->bit) == 0)
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
202
if (shrink & I915_SHRINK_VMAPS &&
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
206
if (!(shrink & I915_SHRINK_ACTIVE) &&
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
228
if (drop_pages(obj, shrink, trylock_vm) &&
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
230
!try_to_writeback(obj, shrink))
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
250
if (shrink & I915_SHRINK_BOUND)
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
42
unsigned long shrink, bool trylock_vm)
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
47
if (shrink & I915_SHRINK_ACTIVE)
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
49
if (!(shrink & I915_SHRINK_BOUND))
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
62
if (obj->ops->shrink) {
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
71
return obj->ops->shrink(obj, shrink_flags);
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1233
.shrink = i915_ttm_shrink,
drivers/gpu/drm/i915/selftests/scatterlist.c
272
shrink,
drivers/gpu/drm/msm/msm_gem_shrinker.c
171
bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket);
drivers/gpu/drm/msm/msm_gem_shrinker.c
192
stages[i].shrink,
drivers/gpu/drm/ttm/ttm_pool.c
1148
static unsigned long ttm_pool_shrinker_scan(struct shrinker *shrink,
drivers/gpu/drm/ttm/ttm_pool.c
1164
static unsigned long ttm_pool_shrinker_count(struct shrinker *shrink,
drivers/gpu/drm/xe/xe_shrinker.c
139
xe_shrinker_count(struct shrinker *shrink, struct shrink_control *sc)
drivers/gpu/drm/xe/xe_shrinker.c
141
struct xe_shrinker *shrinker = to_xe_shrinker(shrink);
drivers/gpu/drm/xe/xe_shrinker.c
201
static unsigned long xe_shrinker_scan(struct shrinker *shrink, struct shrink_control *sc)
drivers/gpu/drm/xe/xe_shrinker.c
203
struct xe_shrinker *shrinker = to_xe_shrinker(shrink);
drivers/gpu/drm/xe/xe_shrinker.c
272
shrinker_free(shrinker->shrink);
drivers/gpu/drm/xe/xe_shrinker.c
290
shrinker->shrink = shrinker_alloc(0, "drm-xe_gem:%s", xe->drm.unique);
drivers/gpu/drm/xe/xe_shrinker.c
291
if (!shrinker->shrink) {
drivers/gpu/drm/xe/xe_shrinker.c
299
shrinker->shrink->count_objects = xe_shrinker_count;
drivers/gpu/drm/xe/xe_shrinker.c
300
shrinker->shrink->scan_objects = xe_shrinker_scan;
drivers/gpu/drm/xe/xe_shrinker.c
301
shrinker->shrink->private_data = shrinker;
drivers/gpu/drm/xe/xe_shrinker.c
302
shrinker_register(shrinker->shrink);
drivers/gpu/drm/xe/xe_shrinker.c
31
struct shrinker *shrink;
drivers/gpu/drm/xe/xe_shrinker.c
35
static struct xe_shrinker *to_xe_shrinker(struct shrinker *shrink)
drivers/gpu/drm/xe/xe_shrinker.c
37
return shrink->private_data;
drivers/md/bcache/bcache.h
547
struct shrinker *shrink;
drivers/md/bcache/btree.c
665
static unsigned long bch_mca_scan(struct shrinker *shrink,
drivers/md/bcache/btree.c
668
struct cache_set *c = shrink->private_data;
drivers/md/bcache/btree.c
732
static unsigned long bch_mca_count(struct shrinker *shrink,
drivers/md/bcache/btree.c
735
struct cache_set *c = shrink->private_data;
drivers/md/bcache/btree.c
753
if (c->shrink)
drivers/md/bcache/btree.c
754
shrinker_free(c->shrink);
drivers/md/bcache/btree.c
829
c->shrink = shrinker_alloc(0, "md-bcache:%pU", c->set_uuid);
drivers/md/bcache/btree.c
830
if (!c->shrink) {
drivers/md/bcache/btree.c
835
c->shrink->count_objects = bch_mca_count;
drivers/md/bcache/btree.c
836
c->shrink->scan_objects = bch_mca_scan;
drivers/md/bcache/btree.c
837
c->shrink->seeks = 4;
drivers/md/bcache/btree.c
838
c->shrink->batch = c->btree_pages * 2;
drivers/md/bcache/btree.c
839
c->shrink->private_data = c;
drivers/md/bcache/btree.c
841
shrinker_register(c->shrink);
drivers/md/bcache/sysfs.c
862
if (c->shrink)
drivers/md/bcache/sysfs.c
863
c->shrink->scan_objects(c->shrink, &sc);
drivers/md/dm-bufio.c
2461
static unsigned long dm_bufio_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
drivers/md/dm-bufio.c
2465
c = shrink->private_data;
drivers/md/dm-bufio.c
2472
static unsigned long dm_bufio_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
drivers/md/dm-bufio.c
2474
struct dm_bufio_client *c = shrink->private_data;
drivers/md/dm-zoned-metadata.c
610
static unsigned long dmz_mblock_shrinker_count(struct shrinker *shrink,
drivers/md/dm-zoned-metadata.c
613
struct dmz_metadata *zmd = shrink->private_data;
drivers/md/dm-zoned-metadata.c
621
static unsigned long dmz_mblock_shrinker_scan(struct shrinker *shrink,
drivers/md/dm-zoned-metadata.c
624
struct dmz_metadata *zmd = shrink->private_data;
drivers/md/persistent-data/dm-array.c
687
r = ((new_size > old_size) ? grow : shrink)(&resize);
drivers/md/raid5.c
7425
static unsigned long raid5_cache_scan(struct shrinker *shrink,
drivers/md/raid5.c
7428
struct r5conf *conf = shrink->private_data;
drivers/md/raid5.c
7446
static unsigned long raid5_cache_count(struct shrinker *shrink,
drivers/md/raid5.c
7449
struct r5conf *conf = shrink->private_data;
fs/erofs/zutil.c
242
static unsigned long erofs_shrink_count(struct shrinker *shrink,
fs/erofs/zutil.c
248
static unsigned long erofs_shrink_scan(struct shrinker *shrink,
fs/ext4/extents_status.c
1760
static unsigned long ext4_es_count(struct shrinker *shrink,
fs/ext4/extents_status.c
1766
sbi = shrink->private_data;
fs/ext4/extents_status.c
1772
static unsigned long ext4_es_scan(struct shrinker *shrink,
fs/ext4/extents_status.c
1775
struct ext4_sb_info *sbi = shrink->private_data;
fs/ext4/inode.c
5869
int shrink = (attr->ia_size < inode->i_size);
fs/ext4/inode.c
5897
if (shrink) {
fs/ext4/inode.c
5933
if (ext4_handle_valid(handle) && shrink) {
fs/ext4/inode.c
5942
if (!shrink) {
fs/ext4/inode.c
5950
if (shrink)
fs/ext4/inode.c
5983
if (!shrink) {
fs/f2fs/f2fs.h
4520
unsigned long f2fs_shrink_count(struct shrinker *shrink,
fs/f2fs/f2fs.h
4522
unsigned long f2fs_shrink_scan(struct shrinker *shrink,
fs/f2fs/shrinker.c
40
unsigned long f2fs_shrink_count(struct shrinker *shrink,
fs/f2fs/shrinker.c
79
unsigned long f2fs_shrink_scan(struct shrinker *shrink,
fs/gfs2/glock.c
1980
static unsigned long gfs2_glock_shrink_scan(struct shrinker *shrink,
fs/gfs2/glock.c
1988
static unsigned long gfs2_glock_shrink_count(struct shrinker *shrink,
fs/gfs2/quota.c
173
static unsigned long gfs2_qd_shrink_scan(struct shrinker *shrink,
fs/gfs2/quota.c
190
static unsigned long gfs2_qd_shrink_count(struct shrinker *shrink,
fs/jbd2/journal.c
1255
static unsigned long jbd2_journal_shrink_scan(struct shrinker *shrink,
fs/jbd2/journal.c
1258
journal_t *journal = shrink->private_data;
fs/jbd2/journal.c
1281
static unsigned long jbd2_journal_shrink_count(struct shrinker *shrink,
fs/jbd2/journal.c
1284
journal_t *journal = shrink->private_data;
fs/mbcache.c
293
static unsigned long mb_cache_count(struct shrinker *shrink,
fs/mbcache.c
296
struct mb_cache *cache = shrink->private_data;
fs/mbcache.c
332
static unsigned long mb_cache_scan(struct shrinker *shrink,
fs/mbcache.c
335
struct mb_cache *cache = shrink->private_data;
fs/nfs/dir.c
2944
nfs_access_cache_scan(struct shrinker *shrink, struct shrink_control *sc)
fs/nfs/dir.c
2956
nfs_access_cache_count(struct shrinker *shrink, struct shrink_control *sc)
fs/nfs/internal.h
389
extern unsigned long nfs_access_cache_count(struct shrinker *shrink,
fs/nfs/internal.h
391
extern unsigned long nfs_access_cache_scan(struct shrinker *shrink,
fs/nfs/nfs42xattr.c
790
static unsigned long nfs4_xattr_cache_count(struct shrinker *shrink,
fs/nfs/nfs42xattr.c
792
static unsigned long nfs4_xattr_entry_count(struct shrinker *shrink,
fs/nfs/nfs42xattr.c
794
static unsigned long nfs4_xattr_cache_scan(struct shrinker *shrink,
fs/nfs/nfs42xattr.c
796
static unsigned long nfs4_xattr_entry_scan(struct shrinker *shrink,
fs/nfs/nfs42xattr.c
839
nfs4_xattr_cache_scan(struct shrinker *shrink, struct shrink_control *sc)
fs/nfs/nfs42xattr.c
860
nfs4_xattr_cache_count(struct shrinker *shrink, struct shrink_control *sc)
fs/nfs/nfs42xattr.c
920
nfs4_xattr_entry_scan(struct shrinker *shrink, struct shrink_control *sc)
fs/nfs/nfs42xattr.c
927
lru = (shrink == nfs4_xattr_large_entry_shrinker) ?
fs/nfs/nfs42xattr.c
950
nfs4_xattr_entry_count(struct shrinker *shrink, struct shrink_control *sc)
fs/nfs/nfs42xattr.c
955
lru = (shrink == nfs4_xattr_large_entry_shrinker) ?
fs/nfsd/nfs4state.c
4877
nfsd4_state_shrinker_count(struct shrinker *shrink, struct shrink_control *sc)
fs/nfsd/nfs4state.c
4879
struct nfsd_net *nn = shrink->private_data;
fs/nfsd/nfs4state.c
4891
nfsd4_state_shrinker_scan(struct shrinker *shrink, struct shrink_control *sc)
fs/nfsd/nfscache.c
294
nfsd_reply_cache_count(struct shrinker *shrink, struct shrink_control *sc)
fs/nfsd/nfscache.c
296
struct nfsd_net *nn = shrink->private_data;
fs/nfsd/nfscache.c
313
nfsd_reply_cache_scan(struct shrinker *shrink, struct shrink_control *sc)
fs/nfsd/nfscache.c
315
struct nfsd_net *nn = shrink->private_data;
fs/nfsd/nfscache.c
41
static unsigned long nfsd_reply_cache_count(struct shrinker *shrink,
fs/nfsd/nfscache.c
43
static unsigned long nfsd_reply_cache_scan(struct shrinker *shrink,
fs/ocfs2/reservations.c
657
unsigned int shrink;
fs/ocfs2/reservations.c
659
shrink = min_bits;
fs/ocfs2/reservations.c
661
shrink = lru_resv->r_len / 2;
fs/ocfs2/reservations.c
663
lru_resv->r_len -= shrink;
fs/ocfs2/reservations.c
666
resv->r_len = shrink;
fs/quota/dquot.c
785
dqcache_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
fs/quota/dquot.c
805
dqcache_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
fs/super.c
179
static unsigned long super_cache_scan(struct shrinker *shrink,
fs/super.c
189
sb = shrink->private_data;
fs/super.c
236
static unsigned long super_cache_count(struct shrinker *shrink,
fs/super.c
242
sb = shrink->private_data;
fs/ubifs/shrinker.c
267
unsigned long ubifs_shrink_count(struct shrinker *shrink,
fs/ubifs/shrinker.c
279
unsigned long ubifs_shrink_scan(struct shrinker *shrink,
fs/ubifs/ubifs.h
1908
unsigned long ubifs_shrink_scan(struct shrinker *shrink,
fs/ubifs/ubifs.h
1910
unsigned long ubifs_shrink_count(struct shrinker *shrink,
fs/xfs/xfs_buf.c
1636
struct shrinker *shrink,
fs/xfs/xfs_buf.c
1639
struct xfs_buftarg *btp = shrink->private_data;
fs/xfs/xfs_buf.c
1658
struct shrinker *shrink,
fs/xfs/xfs_buf.c
1661
struct xfs_buftarg *btp = shrink->private_data;
fs/xfs/xfs_icache.c
2279
struct shrinker *shrink,
fs/xfs/xfs_icache.c
2282
struct xfs_mount *mp = shrink->private_data;
fs/xfs/xfs_icache.c
2300
struct shrinker *shrink,
fs/xfs/xfs_icache.c
2303
struct xfs_mount *mp = shrink->private_data;
fs/xfs/xfs_qm.c
493
struct shrinker *shrink,
fs/xfs/xfs_qm.c
496
struct xfs_quotainfo *qi = shrink->private_data;
fs/xfs/xfs_qm.c
527
struct shrinker *shrink,
fs/xfs/xfs_qm.c
530
struct xfs_quotainfo *qi = shrink->private_data;
include/drm/drm_gem.h
621
bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket),
include/trace/events/vmscan.h
220
__field(void *, shrink)
include/trace/events/vmscan.h
232
__entry->shrink = shr->scan_objects;
include/trace/events/vmscan.h
243
__entry->shrink,
include/trace/events/vmscan.h
264
__field(void *, shrink)
include/trace/events/vmscan.h
274
__entry->shrink = shr->scan_objects;
include/trace/events/vmscan.h
282
__entry->shrink,
kernel/rcu/tree_nocb.h
1178
lazy_rcu_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
kernel/rcu/tree_nocb.h
1203
lazy_rcu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
mm/huge_memory.c
279
static unsigned long shrink_huge_zero_folio_count(struct shrinker *shrink,
mm/huge_memory.c
286
static unsigned long shrink_huge_zero_folio_scan(struct shrinker *shrink,
mm/huge_memory.c
4367
static unsigned long deferred_split_count(struct shrinker *shrink,
mm/huge_memory.c
4407
static unsigned long deferred_split_scan(struct shrinker *shrink,
mm/huge_memory.c
71
static unsigned long deferred_split_count(struct shrinker *shrink,
mm/huge_memory.c
73
static unsigned long deferred_split_scan(struct shrinker *shrink,
mm/slab_common.c
2126
kfree_rcu_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
mm/slab_common.c
2144
kfree_rcu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
mm/slub.c
9150
SLAB_ATTR(shrink);
mm/vmalloc.c
5400
vmap_node_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
mm/vmalloc.c
5415
vmap_node_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
net/core/filter.c
12407
u32 shrink = min_t(u32, delta, skb_frag_size(frag));
net/core/filter.c
12409
memcpy(xdp->data_end, skb_frag_address(frag), shrink);
net/core/filter.c
12411
xdp->data_end += shrink;
net/core/filter.c
12412
sinfo->xdp_frags_size -= shrink;
net/core/filter.c
12413
delta -= shrink;
net/core/filter.c
12414
if (bpf_xdp_shrink_data(xdp, frag, shrink, false))
net/core/filter.c
3670
bool shrink = len_diff < 0;
net/core/filter.c
3678
if (!shrink) {
net/core/filter.c
3714
bool shrink = len_diff < 0;
net/core/filter.c
3739
if (!shrink)
net/core/filter.c
3755
if ((shrink && (len_diff_abs >= len_cur ||
net/core/filter.c
3757
(!shrink && (skb->len + len_diff_abs > len_max &&
net/core/filter.c
3761
ret = shrink ? bpf_skb_net_shrink(skb, off, len_diff_abs, flags) :
net/core/filter.c
4180
static struct xdp_buff *bpf_xdp_shrink_data_zc(struct xdp_buff *xdp, int shrink,
net/core/filter.c
4190
zc_frag->data_end -= shrink;
net/core/filter.c
4192
zc_frag->data += shrink;
net/core/filter.c
4199
int shrink, bool tail)
net/core/filter.c
4202
bool release = skb_frag_size(frag) == shrink;
net/core/filter.c
4208
zc_frag = bpf_xdp_shrink_data_zc(xdp, shrink, tail, release);
net/core/filter.c
4215
skb_frag_off_add(frag, shrink);
net/core/filter.c
4216
skb_frag_size_sub(frag, shrink);
net/core/filter.c
4232
int shrink = min_t(int, offset, skb_frag_size(frag));
net/core/filter.c
4234
len_free += shrink;
net/core/filter.c
4235
offset -= shrink;
net/core/filter.c
4236
if (bpf_xdp_shrink_data(xdp, frag, shrink, true))
net/sunrpc/auth.c
475
rpcauth_cache_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
net/sunrpc/auth.c
489
rpcauth_cache_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
tools/mm/slabinfo.c
1344
else if (shrink)
tools/mm/slabinfo.c
1488
shrink = 1;
tools/mm/slabinfo.c
1523
&& !validate && !shrink && !set_debug && !show_ops)
tools/mm/slabinfo.c
74
int shrink;
virt/kvm/kvm_main.c
3607
unsigned int old, val, shrink, grow_start;
virt/kvm/kvm_main.c
3610
shrink = READ_ONCE(halt_poll_ns_shrink);
virt/kvm/kvm_main.c
3612
if (shrink == 0)
virt/kvm/kvm_main.c
3615
val /= shrink;