Symbol: prealloc
arch/powerpc/platforms/ps3/setup.c
132
#define prealloc_ps3fb_videomemory() prealloc(&ps3fb_videomemory)
arch/powerpc/platforms/ps3/setup.c
155
#define prealloc_ps3flash_bounce_buffer() prealloc(&ps3flash_bounce_buffer)
drivers/gpu/drm/i915/display/intel_fbdev.c
274
bool prealloc = false;
drivers/gpu/drm/i915/display/intel_fbdev.c
305
prealloc = true;
drivers/gpu/drm/i915/display/intel_fbdev.c
341
if (!intel_bo_is_shmem(obj) && !prealloc)
drivers/gpu/drm/msm/msm_gem_vma.c
1150
mmu->funcs->prealloc_count(mmu, &job->prealloc, start_iova, end_iova - start_iova);
drivers/gpu/drm/msm/msm_gem_vma.c
118
struct msm_mmu_prealloc prealloc;
drivers/gpu/drm/msm/msm_gem_vma.c
1218
atomic_add(job->prealloc.count, &vm->prealloc_throttle.in_flight);
drivers/gpu/drm/msm/msm_gem_vma.c
1346
ret = mmu->funcs->prealloc_allocate(mmu, &job->prealloc);
drivers/gpu/drm/msm/msm_gem_vma.c
706
vm->mmu->prealloc = &job->prealloc;
drivers/gpu/drm/msm/msm_gem_vma.c
731
vm->mmu->prealloc = NULL;
drivers/gpu/drm/msm/msm_gem_vma.c
758
vm->mmu->funcs->prealloc_cleanup(vm->mmu, &job->prealloc);
drivers/gpu/drm/msm/msm_gem_vma.c
760
atomic_sub(job->prealloc.count, &vm->prealloc_throttle.in_flight);
drivers/gpu/drm/msm/msm_iommu.c
384
struct msm_mmu_prealloc *p = pagetable->base.prealloc;
drivers/gpu/drm/msm/msm_mmu.h
66
struct msm_mmu_prealloc *prealloc;
drivers/md/dm-cache-target.c
1403
struct dm_bio_prison_cell_v2 *prealloc;
drivers/md/dm-cache-target.c
1405
prealloc = alloc_prison_cell(cache);
drivers/md/dm-cache-target.c
1415
prealloc, &mg->cell);
drivers/md/dm-cache-target.c
1417
free_prison_cell(cache, prealloc);
drivers/md/dm-cache-target.c
1422
if (mg->cell != prealloc)
drivers/md/dm-cache-target.c
1423
free_prison_cell(cache, prealloc);
drivers/md/dm-cache-target.c
1533
struct dm_bio_prison_cell_v2 *prealloc;
drivers/md/dm-cache-target.c
1535
prealloc = alloc_prison_cell(cache);
drivers/md/dm-cache-target.c
1539
READ_WRITE_LOCK_LEVEL, prealloc, &mg->cell);
drivers/md/dm-cache-target.c
1541
free_prison_cell(cache, prealloc);
drivers/md/dm-cache-target.c
1546
if (mg->cell != prealloc)
drivers/md/dm-cache-target.c
1547
free_prison_cell(cache, prealloc);
drivers/media/platform/renesas/vsp1/vsp1_dl.c
1114
unsigned int prealloc)
drivers/media/platform/renesas/vsp1/vsp1_dl.c
1149
dlm->pool = vsp1_dl_body_pool_create(vsp1, prealloc + 1,
drivers/media/platform/renesas/vsp1/vsp1_dl.c
1154
for (i = 0; i < prealloc; ++i) {
drivers/media/platform/renesas/vsp1/vsp1_dl.c
1171
dlm->list_count = prealloc;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
1175
VSP1_EXTCMD_AUTOFLD, prealloc);
drivers/media/platform/renesas/vsp1/vsp1_dl.h
56
unsigned int prealloc);
drivers/usb/gadget/function/u_ether.c
360
status = prealloc(&dev->tx_reqs, link->in_ep, n);
drivers/usb/gadget/function/u_ether.c
363
status = prealloc(&dev->rx_reqs, link->out_ep, n);
fs/btrfs/extent-io-tree.c
1050
struct extent_state *prealloc = NULL;
fs/btrfs/extent-io-tree.c
1068
if (!prealloc) {
fs/btrfs/extent-io-tree.c
1076
prealloc = alloc_extent_state(mask);
fs/btrfs/extent-io-tree.c
1095
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
1096
if (!prealloc)
fs/btrfs/extent-io-tree.c
1098
prealloc->start = start;
fs/btrfs/extent-io-tree.c
1099
prealloc->end = end;
fs/btrfs/extent-io-tree.c
1100
insert_state_fast(tree, prealloc, p, parent, bits, changeset);
fs/btrfs/extent-io-tree.c
1101
cache_state(prealloc, cached_state);
fs/btrfs/extent-io-tree.c
1102
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1168
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
1169
if (!prealloc)
fs/btrfs/extent-io-tree.c
1171
ret = split_state(tree, state, prealloc, start);
fs/btrfs/extent-io-tree.c
1175
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1201
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
1202
if (!prealloc)
fs/btrfs/extent-io-tree.c
1209
prealloc->start = start;
fs/btrfs/extent-io-tree.c
1211
prealloc->end = end;
fs/btrfs/extent-io-tree.c
1213
prealloc->end = last_start - 1;
fs/btrfs/extent-io-tree.c
1215
inserted_state = insert_state(tree, prealloc, bits, changeset);
fs/btrfs/extent-io-tree.c
1218
extent_io_tree_panic(tree, prealloc, "insert", ret);
fs/btrfs/extent-io-tree.c
1223
if (inserted_state == prealloc)
fs/btrfs/extent-io-tree.c
1224
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1258
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
1259
if (!prealloc)
fs/btrfs/extent-io-tree.c
1261
ret = split_state(tree, state, prealloc, end + 1);
fs/btrfs/extent-io-tree.c
1264
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1268
set_state_bits(tree, prealloc, bits, changeset);
fs/btrfs/extent-io-tree.c
1269
cache_state(prealloc, cached_state);
fs/btrfs/extent-io-tree.c
1270
merge_state(tree, prealloc);
fs/btrfs/extent-io-tree.c
1271
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1285
btrfs_free_extent_state(prealloc);
fs/btrfs/extent-io-tree.c
1320
struct extent_state *prealloc = NULL;
fs/btrfs/extent-io-tree.c
1333
if (!prealloc) {
fs/btrfs/extent-io-tree.c
1341
prealloc = alloc_extent_state(GFP_NOFS);
fs/btrfs/extent-io-tree.c
1342
if (!prealloc && !first_iteration)
fs/btrfs/extent-io-tree.c
1360
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
1361
if (!prealloc) {
fs/btrfs/extent-io-tree.c
1365
prealloc->start = start;
fs/btrfs/extent-io-tree.c
1366
prealloc->end = end;
fs/btrfs/extent-io-tree.c
1367
insert_state_fast(tree, prealloc, p, parent, bits, NULL);
fs/btrfs/extent-io-tree.c
1368
cache_state(prealloc, cached_state);
fs/btrfs/extent-io-tree.c
1369
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1410
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
1411
if (!prealloc) {
fs/btrfs/extent-io-tree.c
1415
ret = split_state(tree, state, prealloc, start);
fs/btrfs/extent-io-tree.c
1416
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1443
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
1444
if (!prealloc) {
fs/btrfs/extent-io-tree.c
1453
prealloc->start = start;
fs/btrfs/extent-io-tree.c
1455
prealloc->end = end;
fs/btrfs/extent-io-tree.c
1457
prealloc->end = last_start - 1;
fs/btrfs/extent-io-tree.c
1459
inserted_state = insert_state(tree, prealloc, bits, NULL);
fs/btrfs/extent-io-tree.c
1462
extent_io_tree_panic(tree, prealloc, "insert", ret);
fs/btrfs/extent-io-tree.c
1466
if (inserted_state == prealloc)
fs/btrfs/extent-io-tree.c
1467
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1494
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
1495
if (!prealloc) {
fs/btrfs/extent-io-tree.c
1500
ret = split_state(tree, state, prealloc, end + 1);
fs/btrfs/extent-io-tree.c
1503
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1507
set_state_bits(tree, prealloc, bits, NULL);
fs/btrfs/extent-io-tree.c
1508
cache_state(prealloc, cached_state);
fs/btrfs/extent-io-tree.c
1509
clear_state_bit(tree, prealloc, clear_bits, 0, end, NULL);
fs/btrfs/extent-io-tree.c
1510
prealloc = NULL;
fs/btrfs/extent-io-tree.c
1524
btrfs_free_extent_state(prealloc);
fs/btrfs/extent-io-tree.c
166
static struct extent_state *alloc_extent_state_atomic(struct extent_state *prealloc)
fs/btrfs/extent-io-tree.c
168
if (!prealloc)
fs/btrfs/extent-io-tree.c
169
prealloc = alloc_extent_state(GFP_ATOMIC);
fs/btrfs/extent-io-tree.c
171
return prealloc;
fs/btrfs/extent-io-tree.c
501
struct extent_state *prealloc, u64 split)
fs/btrfs/extent-io-tree.c
509
prealloc->start = orig->start;
fs/btrfs/extent-io-tree.c
510
prealloc->end = split - 1;
fs/btrfs/extent-io-tree.c
511
prealloc->state = orig->state;
fs/btrfs/extent-io-tree.c
522
if (prealloc->end < entry->start) {
fs/btrfs/extent-io-tree.c
524
} else if (prealloc->end > entry->end) {
fs/btrfs/extent-io-tree.c
527
btrfs_free_extent_state(prealloc);
fs/btrfs/extent-io-tree.c
532
rb_link_node(&prealloc->rb_node, parent, node);
fs/btrfs/extent-io-tree.c
533
rb_insert_color(&prealloc->rb_node, &tree->state);
fs/btrfs/extent-io-tree.c
615
struct extent_state *prealloc = NULL;
fs/btrfs/extent-io-tree.c
636
if (!prealloc) {
fs/btrfs/extent-io-tree.c
644
prealloc = alloc_extent_state(mask);
fs/btrfs/extent-io-tree.c
699
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
700
if (!prealloc)
fs/btrfs/extent-io-tree.c
702
ret = split_state(tree, state, prealloc, start);
fs/btrfs/extent-io-tree.c
703
prealloc = NULL;
fs/btrfs/extent-io-tree.c
727
prealloc = alloc_extent_state_atomic(prealloc);
fs/btrfs/extent-io-tree.c
728
if (!prealloc)
fs/btrfs/extent-io-tree.c
730
ret = split_state(tree, state, prealloc, end + 1);
fs/btrfs/extent-io-tree.c
733
prealloc = NULL;
fs/btrfs/extent-io-tree.c
740
clear_state_bit(tree, prealloc, bits, wake, end, changeset);
fs/btrfs/extent-io-tree.c
742
prealloc = NULL;
fs/btrfs/extent-io-tree.c
762
btrfs_free_extent_state(prealloc);
fs/btrfs/extent_io.c
3328
struct btrfs_folio_state *prealloc,
fs/btrfs/extent_io.c
3389
ret = attach_extent_buffer_folio(eb, eb->folios[i], prealloc);
fs/btrfs/extent_io.c
3411
struct btrfs_folio_state *prealloc = NULL;
fs/btrfs/extent_io.c
3456
prealloc = btrfs_alloc_folio_state(fs_info, PAGE_SIZE, BTRFS_SUBPAGE_METADATA);
fs/btrfs/extent_io.c
3457
if (IS_ERR(prealloc)) {
fs/btrfs/extent_io.c
3458
ret = PTR_ERR(prealloc);
fs/btrfs/extent_io.c
3467
btrfs_free_folio_state(prealloc);
fs/btrfs/extent_io.c
3475
ret = attach_eb_folio_to_filemap(eb, i, prealloc, &existing_eb);
fs/btrfs/extent_io.c
871
struct btrfs_folio_state *prealloc)
fs/btrfs/extent_io.c
895
btrfs_free_folio_state(prealloc);
fs/btrfs/extent_io.c
899
if (prealloc)
fs/btrfs/extent_io.c
901
folio_attach_private(folio, prealloc);
fs/btrfs/inode.c
4011
static int btrfs_add_inode_to_root(struct btrfs_inode *inode, bool prealloc)
fs/btrfs/inode.c
4021
if (prealloc) {
fs/btrfs/ioctl.c
3591
struct btrfs_qgroup_list *prealloc = NULL;
fs/btrfs/ioctl.c
3613
prealloc = kzalloc_obj(*prealloc);
fs/btrfs/ioctl.c
3614
if (!prealloc) {
fs/btrfs/ioctl.c
3632
ret = btrfs_add_qgroup_relation(trans, sa->src, sa->dst, prealloc);
fs/btrfs/ioctl.c
3633
prealloc = NULL;
fs/btrfs/ioctl.c
3651
kfree(prealloc);
fs/btrfs/qgroup.c
1010
struct btrfs_qgroup *prealloc = NULL;
fs/btrfs/qgroup.c
1142
ASSERT(prealloc == NULL);
fs/btrfs/qgroup.c
1143
prealloc = kzalloc_obj(*prealloc, GFP_NOFS);
fs/btrfs/qgroup.c
1144
if (unlikely(!prealloc)) {
fs/btrfs/qgroup.c
1157
qgroup = add_qgroup_rb(fs_info, prealloc, found_key.offset);
fs/btrfs/qgroup.c
1158
prealloc = NULL;
fs/btrfs/qgroup.c
1199
ASSERT(prealloc == NULL);
fs/btrfs/qgroup.c
1200
prealloc = kzalloc_obj(*prealloc, GFP_NOFS);
fs/btrfs/qgroup.c
1201
if (!prealloc) {
fs/btrfs/qgroup.c
1205
qgroup = add_qgroup_rb(fs_info, prealloc, BTRFS_FS_TREE_OBJECTID);
fs/btrfs/qgroup.c
1206
prealloc = NULL;
fs/btrfs/qgroup.c
1281
kfree(prealloc);
fs/btrfs/qgroup.c
1546
struct btrfs_qgroup_list *prealloc)
fs/btrfs/qgroup.c
1554
ASSERT(prealloc);
fs/btrfs/qgroup.c
1558
kfree(prealloc);
fs/btrfs/qgroup.c
1593
ret = __add_relation_rb(prealloc, member, parent);
fs/btrfs/qgroup.c
1594
prealloc = NULL;
fs/btrfs/qgroup.c
1603
kfree(prealloc);
fs/btrfs/qgroup.c
1681
struct btrfs_qgroup *prealloc = NULL;
fs/btrfs/qgroup.c
1696
prealloc = kzalloc_obj(*prealloc, GFP_NOFS);
fs/btrfs/qgroup.c
1697
if (!prealloc) {
fs/btrfs/qgroup.c
1707
qgroup = add_qgroup_rb(fs_info, prealloc, qgroupid);
fs/btrfs/qgroup.c
1709
prealloc = NULL;
fs/btrfs/qgroup.c
1714
kfree(prealloc);
fs/btrfs/qgroup.c
202
struct btrfs_qgroup *prealloc,
fs/btrfs/qgroup.c
208
ASSERT(prealloc);
fs/btrfs/qgroup.c
210
prealloc->qgroupid = qgroupid;
fs/btrfs/qgroup.c
211
node = rb_find_add(&prealloc->node, &fs_info->qgroup_tree, btrfs_qgroup_qgroupid_cmp);
fs/btrfs/qgroup.c
213
kfree(prealloc);
fs/btrfs/qgroup.c
217
INIT_LIST_HEAD(&prealloc->groups);
fs/btrfs/qgroup.c
218
INIT_LIST_HEAD(&prealloc->members);
fs/btrfs/qgroup.c
219
INIT_LIST_HEAD(&prealloc->dirty);
fs/btrfs/qgroup.c
220
INIT_LIST_HEAD(&prealloc->iterator);
fs/btrfs/qgroup.c
221
INIT_LIST_HEAD(&prealloc->nested_iterator);
fs/btrfs/qgroup.c
223
return prealloc;
fs/btrfs/qgroup.c
271
static int __add_relation_rb(struct btrfs_qgroup_list *prealloc,
fs/btrfs/qgroup.c
276
kfree(prealloc);
fs/btrfs/qgroup.c
280
prealloc->group = parent;
fs/btrfs/qgroup.c
281
prealloc->member = member;
fs/btrfs/qgroup.c
282
list_add_tail(&prealloc->next_group, &member->groups);
fs/btrfs/qgroup.c
283
list_add_tail(&prealloc->next_member, &parent->members);
fs/btrfs/qgroup.c
298
struct btrfs_qgroup_list *prealloc,
fs/btrfs/qgroup.c
307
return __add_relation_rb(prealloc, member, parent);
fs/btrfs/qgroup.c
3339
struct btrfs_qgroup *prealloc;
fs/btrfs/qgroup.c
3349
prealloc = kzalloc_obj(*prealloc, GFP_NOFS);
fs/btrfs/qgroup.c
3350
if (!prealloc)
fs/btrfs/qgroup.c
3449
dstgroup = add_qgroup_rb(fs_info, prealloc, objectid);
fs/btrfs/qgroup.c
3450
prealloc = NULL;
fs/btrfs/qgroup.c
3579
kfree(prealloc);
fs/btrfs/qgroup.c
495
struct btrfs_qgroup *prealloc;
fs/btrfs/qgroup.c
498
prealloc = kzalloc_obj(*prealloc);
fs/btrfs/qgroup.c
499
if (!prealloc) {
fs/btrfs/qgroup.c
503
qgroup = add_qgroup_rb(fs_info, prealloc, found_key.offset);
fs/btrfs/qgroup.h
342
struct btrfs_qgroup_list *prealloc);
fs/btrfs/ulist.c
113
if (!ulist->prealloc)
fs/btrfs/ulist.c
114
ulist->prealloc = kzalloc_obj(*ulist->prealloc, gfp_mask);
fs/btrfs/ulist.c
218
if (ulist->prealloc) {
fs/btrfs/ulist.c
219
node = ulist->prealloc;
fs/btrfs/ulist.c
220
ulist->prealloc = NULL;
fs/btrfs/ulist.c
53
ulist->prealloc = NULL;
fs/btrfs/ulist.c
72
kfree(ulist->prealloc);
fs/btrfs/ulist.c
73
ulist->prealloc = NULL;
fs/btrfs/ulist.h
44
struct ulist_node *prealloc;
fs/ext4/extents_status.c
1484
struct extent_status *prealloc)
fs/ext4/extents_status.c
1537
err = __es_insert_extent(inode, &newes, prealloc);
fs/ext4/extents_status.c
180
struct extent_status *prealloc);
fs/ext4/extents_status.c
184
struct extent_status *prealloc);
fs/ext4/extents_status.c
190
struct pending_reservation **prealloc);
fs/ext4/extents_status.c
2076
struct pending_reservation **prealloc)
fs/ext4/extents_status.c
2102
if (likely(*prealloc == NULL)) {
fs/ext4/extents_status.c
2109
pr = *prealloc;
fs/ext4/extents_status.c
2110
*prealloc = NULL;
fs/ext4/extents_status.c
2315
struct pending_reservation **prealloc)
fs/ext4/extents_status.c
2346
ret = __insert_pending(inode, first, prealloc);
fs/ext4/extents_status.c
2358
ret = __insert_pending(inode, last, prealloc);
fs/ext4/extents_status.c
2371
ret = __insert_pending(inode, first, prealloc);
fs/ext4/extents_status.c
2383
ret = __insert_pending(inode, last, prealloc);
fs/ext4/extents_status.c
843
struct extent_status *prealloc)
fs/ext4/extents_status.c
883
if (prealloc)
fs/ext4/extents_status.c
884
es = prealloc;
fs/f2fs/data.c
1625
blkcnt_t prealloc;
fs/f2fs/data.c
1674
prealloc = 0;
fs/f2fs/data.c
1699
prealloc++;
fs/f2fs/data.c
1795
err = f2fs_reserve_new_blocks(&dn, prealloc);
fs/f2fs/data.c
1800
if (prealloc && dn.ofs_in_node != last_ofs_in_node + 1) {
fs/kernfs/file.c
687
if (ops->prealloc && ops->seq_show)
fs/kernfs/file.c
689
if (ops->prealloc) {
fs/namespace.c
2249
struct path *prealloc, unsigned count)
fs/namespace.c
2253
struct path *res = prealloc, *to_free = NULL;
fs/namespace.c
2282
void drop_collected_paths(const struct path *paths, const struct path *prealloc)
fs/namespace.c
2286
if (paths != prealloc)
fs/ntfs3/attrib.c
601
sbi->options->prealloc) {
fs/ntfs3/file.c
1412
if (ni->mi.sbi->options->prealloc) {
fs/ntfs3/file.c
416
&new_valid, ni->mi.sbi->options->prealloc, NULL,
fs/ntfs3/ntfs_fs.h
110
unsigned prealloc : 1; /* Preallocate space when file is growing. */
fs/ntfs3/super.c
1840
opts->prealloc = 1;
fs/ntfs3/super.c
409
opts->prealloc = 1;
fs/ntfs3/super.c
412
opts->prealloc = result.boolean;
fs/ntfs3/super.c
795
if (opts->prealloc)
fs/sysfs/file.c
238
.prealloc = true,
fs/sysfs/file.c
243
.prealloc = true,
fs/sysfs/file.c
249
.prealloc = true,
fs/xfs/xfs_dquot.c
125
int prealloc = 0;
fs/xfs/xfs_dquot.c
132
prealloc = 1;
fs/xfs/xfs_dquot.c
136
prealloc = 1;
fs/xfs/xfs_dquot.c
147
if (prealloc)
fs/xfs/xfs_iomap.c
1462
xfs_filblks_t prealloc,
fs/xfs/xfs_iomap.c
1475
whichfork == XFS_COW_FORK && !prealloc;
fs/xfs/xfs_iomap.c
1483
alen = XFS_FILBLKS_MIN(len + prealloc, XFS_MAX_BMBT_EXTLEN);
fs/xfs/xfs_iomap.c
1486
if (prealloc && alen >= len)
fs/xfs/xfs_iomap.c
1487
prealloc = alen - len;
fs/xfs/xfs_iomap.c
1555
if (whichfork == XFS_DATA_FORK && prealloc)
fs/xfs/xfs_iomap.c
1557
if (whichfork == XFS_COW_FORK && (prealloc || aoff < off || alen > len))
fs/xfs/xfs_iomap.c
1572
if (prealloc || use_cowextszhint) {
fs/xfs/xfs_iomap.c
1575
prealloc = 0;
include/linux/kernfs.h
318
bool prealloc;
kernel/bpf/hashtab.c
1079
bool prealloc = htab_is_prealloc(htab);
kernel/bpf/hashtab.c
1083
if (prealloc) {
kernel/bpf/hashtab.c
1119
if (prealloc) {
kernel/bpf/hashtab.c
1136
if (!prealloc)
kernel/bpf/hashtab.c
2316
bool prealloc = htab_is_prealloc(htab);
kernel/bpf/hashtab.c
2324
if (prealloc) {
kernel/bpf/hashtab.c
419
bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC);
kernel/bpf/hashtab.c
437
if (lru && !prealloc)
kernel/bpf/hashtab.c
552
bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC);
kernel/bpf/hashtab.c
635
if (prealloc) {
lib/stackdepot.c
293
static bool depot_init_pool(void **prealloc)
lib/stackdepot.c
306
if (!new_pool && *prealloc) {
lib/stackdepot.c
308
WRITE_ONCE(new_pool, *prealloc);
lib/stackdepot.c
309
*prealloc = NULL;
lib/stackdepot.c
341
static void depot_keep_new_pool(void **prealloc)
lib/stackdepot.c
353
WRITE_ONCE(new_pool, *prealloc);
lib/stackdepot.c
354
*prealloc = NULL;
lib/stackdepot.c
361
static struct stack_record *depot_pop_free_pool(void **prealloc, size_t size)
lib/stackdepot.c
371
if (!depot_init_pool(prealloc))
lib/stackdepot.c
434
depot_alloc_stack(unsigned long *entries, unsigned int nr_entries, u32 hash, depot_flags_t flags, void **prealloc)
lib/stackdepot.c
462
stack = depot_pop_free_pool(prealloc, record_size);
lib/stackdepot.c
650
void *prealloc = NULL;
lib/stackdepot.c
688
prealloc = page_address(page);
lib/stackdepot.c
706
depot_alloc_stack(entries, nr_entries, hash, depot_flags, &prealloc);
lib/stackdepot.c
718
if (prealloc) {
lib/stackdepot.c
724
depot_keep_new_pool(&prealloc);
lib/stackdepot.c
730
if (prealloc) {
lib/stackdepot.c
733
free_pages_nolock(virt_to_page(prealloc), DEPOT_POOL_ORDER);
lib/stackdepot.c
735
free_pages((unsigned long)prealloc, DEPOT_POOL_ORDER);
mm/memory.c
1062
struct folio **prealloc, struct page *page)
mm/memory.c
1067
new_folio = *prealloc;
mm/memory.c
1079
*prealloc = NULL;
mm/memory.c
1128
int max_nr, int *rss, struct folio **prealloc)
mm/memory.c
1146
if (unlikely(!*prealloc && folio_test_large(folio) && max_nr != 1)) {
mm/memory.c
1183
addr, rss, prealloc, page);
mm/memory.c
1235
struct folio *prealloc = NULL;
mm/memory.c
1318
ptent, addr, max_nr, rss, &prealloc);
mm/memory.c
1326
if (unlikely(prealloc)) {
mm/memory.c
1333
folio_put(prealloc);
mm/memory.c
1334
prealloc = NULL;
mm/memory.c
1357
prealloc = folio_prealloc(src_mm, src_vma, addr, false);
mm/memory.c
1358
if (!prealloc)
mm/memory.c
1370
if (unlikely(prealloc))
mm/memory.c
1371
folio_put(prealloc);
net/sched/sch_gred.c
484
struct gred_sched_data **prealloc,
net/sched/sch_gred.c
496
table->tab[dp] = q = *prealloc;
net/sched/sch_gred.c
497
*prealloc = NULL;
net/sched/sch_gred.c
651
struct gred_sched_data *prealloc;
net/sched/sch_gred.c
703
prealloc = kzalloc_obj(*prealloc);
net/sched/sch_gred.c
706
err = gred_change_vq(sch, ctl->DP, ctl, prio, stab, max_P, &prealloc,
net/sched/sch_gred.c
721
kfree(prealloc);
net/sched/sch_gred.c
728
kfree(prealloc);