kmem_cache_zalloc
uf->flist = kmem_cache_zalloc(flash_block_cache, GFP_KERNEL);
fl->next = kmem_cache_zalloc(flash_block_cache, GFP_KERNEL);
p = kmem_cache_zalloc(flash_block_cache, GFP_KERNEL);
pte = kmem_cache_zalloc(hpte_cache, GFP_KERNEL);
datap = kmem_cache_zalloc(cache, GFP_KERNEL);
zdev->fmb = kmem_cache_zalloc(zdev_fmb_cache, GFP_KERNEL);
sp = kmem_cache_zalloc(mmu_page_header_cache, GFP_KERNEL_ACCOUNT);
ctxt = kmem_cache_zalloc(x86_emulator_cache, GFP_KERNEL_ACCOUNT);
obj_request = kmem_cache_zalloc(rbd_obj_request_cache, GFP_NOIO);
blkif = kmem_cache_zalloc(xen_blkif_cachep, GFP_KERNEL);
event = kmem_cache_zalloc(mhi_cntrl->ev_ring_el_cache, GFP_KERNEL);
event = kmem_cache_zalloc(mhi_cntrl->ev_ring_el_cache, GFP_KERNEL);
event = kmem_cache_zalloc(mhi_cntrl->ev_ring_el_cache, GFP_KERNEL);
buf_addr = kmem_cache_zalloc(mhi_cntrl->tre_buf_cache, GFP_KERNEL);
item = kmem_cache_zalloc(mhi_cntrl->ring_item_cache, GFP_ATOMIC);
event = kmem_cache_zalloc(mhi_cntrl->ev_ring_el_cache, GFP_KERNEL);
return kmem_cache_zalloc(qi_cache, flags);
desc = kmem_cache_zalloc(chan->ccp->dma_desc_cache, GFP_NOWAIT);
desc = kmem_cache_zalloc(chan->pt->dma_desc_cache, GFP_NOWAIT);
desc = kmem_cache_zalloc(ioat_cache, flags);
desc = kmem_cache_zalloc(d40c->base->desc_slab, GFP_NOWAIT);
block = kmem_cache_zalloc(slab_blocks, GFP_KERNEL);
obj = kmem_cache_zalloc(slab_objects, GFP_KERNEL);
return kmem_cache_zalloc(slab_ce, GFP_KERNEL);
workload = kmem_cache_zalloc(s->workloads, GFP_KERNEL);
return kmem_cache_zalloc(slab_vmas, GFP_KERNEL);
kmem_cache_zalloc(slab_vma_resources, GFP_KERNEL);
task = kmem_cache_zalloc(pipe->task_slab, GFP_KERNEL);
fence = kmem_cache_zalloc(lima_fence_slab, GFP_KERNEL);
fence = kmem_cache_zalloc(sched_fence_slab, GFP_KERNEL);
vbuf = kmem_cache_zalloc(vgdev->vbufs, GFP_KERNEL | __GFP_NOFAIL);
vbuf = kmem_cache_zalloc(vgdev->vbufs, GFP_ATOMIC);
return kmem_cache_zalloc(xe_hw_fence_slab, GFP_KERNEL);
return kmem_cache_zalloc(parallel ? xe_sched_job_parallel_slab :
message = kmem_cache_zalloc(gb_message_cache, gfp_flags);
operation = kmem_cache_zalloc(gb_operation_cache, gfp_flags);
tx_desc = kmem_cache_zalloc(ig.desc_cache, GFP_ATOMIC);
pent = kmem_cache_zalloc(lv2table_kmem_cache, GFP_ATOMIC);
zero_lv2_table = kmem_cache_zalloc(lv2table_kmem_cache, GFP_KERNEL);
info = kmem_cache_zalloc(iommu_devinfo_cache, GFP_ATOMIC);
dma_domain = kmem_cache_zalloc(fsl_pamu_domain_cache, GFP_KERNEL);
table = kmem_cache_zalloc(data->l2_tables, gfp);
return kmem_cache_zalloc(iova_cache, GFP_ATOMIC | __GFP_NOWARN);
iopte = kmem_cache_zalloc(iopte_cachep, GFP_KERNEL);
page_table = kmem_cache_zalloc(iommu->pt_pool, gfp);
event = kmem_cache_zalloc(_dm_event_cache, GFP_ATOMIC);
sh = kmem_cache_zalloc(sc, gfp);
aeb = kmem_cache_zalloc(ai->aeb_slab_cache, GFP_KERNEL);
fte = kmem_cache_zalloc(steering->ftes_cache, GFP_KERNEL);
fg = kmem_cache_zalloc(steering->fgs_cache, GFP_KERNEL);
node = kmem_cache_zalloc(node_cache, GFP_KERNEL);
newnode = kmem_cache_zalloc(node_cache, GFP_KERNEL);
node = kmem_cache_zalloc(node_cache, GFP_KERNEL);
peer = kmem_cache_zalloc(peer_cache, GFP_KERNEL);
return kmem_cache_zalloc(trans_pcie->dev_cmd_pool, GFP_ATOMIC);
cmbe = kmem_cache_zalloc(cmbe_cache, GFP_KERNEL);
q = kmem_cache_zalloc(qdio_q_cache, GFP_KERNEL);
private->io_region = kmem_cache_zalloc(vfio_ccw_io_region,
private->cmd_region = kmem_cache_zalloc(vfio_ccw_cmd_region,
private->schib_region = kmem_cache_zalloc(vfio_ccw_schib_region,
private->crw_region = kmem_cache_zalloc(vfio_ccw_crw_region,
newbuf = kmem_cache_zalloc(qeth_qdio_outbuf_cache, gfp);
buf->aob = kmem_cache_zalloc(qeth_qaob_cache,
fc_req = kmem_cache_zalloc(zfcp_fc_req_cache, GFP_ATOMIC);
fc_req = kmem_cache_zalloc(zfcp_fc_req_cache, GFP_KERNEL);
fc_req = kmem_cache_zalloc(zfcp_fc_req_cache, GFP_KERNEL);
ascb = kmem_cache_zalloc(asd_ascb_cache, gfp_flags);
struct sas_task *task = kmem_cache_zalloc(sas_task_cache, flags);
event = kmem_cache_zalloc(sas_event_cache, gfp_flags);
pla = kmem_cache_zalloc(qla_tgt_plogi_cachep, GFP_ATOMIC);
mq->recvbuf = kmem_cache_zalloc(maple_queue_cache, GFP_KERNEL);
dr = kmem_cache_zalloc(lio_dr_cache, GFP_ATOMIC);
ooo_cmdsn = kmem_cache_zalloc(lio_ooo_cache, GFP_ATOMIC);
qr = kmem_cache_zalloc(lio_qr_cache, GFP_ATOMIC);
r2t = kmem_cache_zalloc(lio_r2t_cache, GFP_ATOMIC);
qr = kmem_cache_zalloc(lio_qr_cache, GFP_ATOMIC);
tl_cmd = kmem_cache_zalloc(tcm_loop_cmd_cache, GFP_KERNEL);
lba_map = kmem_cache_zalloc(t10_alua_lba_map_cache, GFP_KERNEL);
lba_map_mem = kmem_cache_zalloc(t10_alua_lba_map_mem_cache, GFP_KERNEL);
lu_gp = kmem_cache_zalloc(t10_alua_lu_gp_cache, GFP_KERNEL);
lu_gp_mem = kmem_cache_zalloc(t10_alua_lu_gp_mem_cache, GFP_KERNEL);
tg_pt_gp = kmem_cache_zalloc(t10_alua_tg_pt_gp_cache, GFP_KERNEL);
pr_reg = kmem_cache_zalloc(t10_pr_reg_cache, GFP_ATOMIC);
pr_reg = kmem_cache_zalloc(t10_pr_reg_cache, GFP_KERNEL);
se_sess = kmem_cache_zalloc(se_sess_cache, GFP_KERNEL);
ua = kmem_cache_zalloc(se_ua_cache, GFP_ATOMIC);
tcmu_cmd = kmem_cache_zalloc(tcmu_cmd_cache, GFP_NOIO);
qh->desc_list = kmem_cache_zalloc(desc_cache, flags | GFP_DMA);
urbp = kmem_cache_zalloc(uhci_up_cachep, GFP_ATOMIC);
urbp = kmem_cache_zalloc(xenhcd_urbp_cachep, mem_flags);
urb_listitem = kmem_cache_zalloc(urb_listitem_cachep,
qh = kmem_cache_zalloc(qh_cachep, flags);
qtd = kmem_cache_zalloc(qtd_cachep, flags);
priv = kmem_cache_zalloc(stub_priv_cache, GFP_ATOMIC);
ctx = kmem_cache_zalloc(kioctx_cachep, GFP_KERNEL);
aio = kmem_cache_zalloc(backing_aio_cachep, GFP_KERNEL);
aio = kmem_cache_zalloc(backing_aio_cachep, GFP_KERNEL);
return kmem_cache_zalloc(btrfs_path_cachep, GFP_NOFS);
defrag = kmem_cache_zalloc(btrfs_inode_defrag_cachep, GFP_NOFS);
node = kmem_cache_zalloc(delayed_node_cache, GFP_NOFS);
eb = kmem_cache_zalloc(extent_buffer_cache, GFP_NOFS|__GFP_NOFAIL);
em = kmem_cache_zalloc(extent_map_cache, GFP_NOFS);
info = kmem_cache_zalloc(btrfs_free_space_cachep,
info->bitmap = kmem_cache_zalloc(btrfs_free_space_bitmap_cachep,
info = kmem_cache_zalloc(btrfs_free_space_cachep, GFP_NOFS);
info = kmem_cache_zalloc(btrfs_free_space_cachep, GFP_NOFS);
map = kmem_cache_zalloc(btrfs_free_space_bitmap_cachep, GFP_NOFS);
e = kmem_cache_zalloc(btrfs_free_space_cachep,
e->bitmap = kmem_cache_zalloc(
entry = kmem_cache_zalloc(btrfs_ordered_extent_cache, GFP_NOFS);
h = kmem_cache_zalloc(btrfs_trans_handle_cachep, GFP_NOFS);
struct buffer_head *ret = kmem_cache_zalloc(bh_cachep, gfp_flags);
object = kmem_cache_zalloc(cachefiles_object_jar, GFP_KERNEL);
di = kmem_cache_zalloc(ceph_dentry_cachep, GFP_KERNEL);
kmem_cache_zalloc(ceph_dir_file_cachep, GFP_KERNEL);
fi = kmem_cache_zalloc(ceph_file_cachep, GFP_KERNEL);
req = kmem_cache_zalloc(ceph_mds_request_cachep, GFP_NOFS);
capsnap = kmem_cache_zalloc(ceph_cap_snap_cachep, GFP_NOFS);
sd = kmem_cache_zalloc(configfs_dir_cachep, GFP_KERNEL);
crypt_info = kmem_cache_zalloc(fscrypt_inode_info_cachep, GFP_KERNEL);
return kmem_cache_zalloc(lkb_cache, GFP_ATOMIC);
return kmem_cache_zalloc(rsb_cache, GFP_ATOMIC);
file_info = kmem_cache_zalloc(ecryptfs_file_info_cache, GFP_KERNEL);
file_info = kmem_cache_zalloc(ecryptfs_file_info_cache, GFP_KERNEL);
kmem_cache_zalloc(ecryptfs_auth_tok_list_item_cache,
kmem_cache_zalloc(ecryptfs_auth_tok_list_item_cache, GFP_KERNEL);
new_auth_tok = kmem_cache_zalloc(ecryptfs_global_auth_tok_cache,
sbi = kmem_cache_zalloc(ecryptfs_sb_info_cache, GFP_KERNEL);
pcl = kmem_cache_zalloc(pcs->slab, GFP_KERNEL);
to_free = kmem_cache_zalloc(ephead_cache, GFP_KERNEL);
if (!(epi = kmem_cache_zalloc(epi_cache, GFP_KERNEL))) {
return kmem_cache_zalloc(ext4_pending_cachep, GFP_KERNEL | __GFP_NOFAIL);
return kmem_cache_zalloc(ext4_es_cachep, GFP_KERNEL | __GFP_NOFAIL);
meta_group_info[i] = kmem_cache_zalloc(cachep, GFP_NOFS);
pa = kmem_cache_zalloc(ext4_pspace_cachep, GFP_NOFS);
ac = kmem_cache_zalloc(ext4_ac_cachep, GFP_NOFS);
ext4_io_end_t *io_end = kmem_cache_zalloc(io_end_cachep, flags);
io_end_vec = kmem_cache_zalloc(io_end_vec_cachep, GFP_NOFS);
struct fuse_req *req = kmem_cache_zalloc(fuse_req_cachep, flags);
qd = kmem_cache_zalloc(gfs2_quotad_cachep, GFP_NOFS);
tmp = kmem_cache_zalloc(gfs2_qadata_cachep, GFP_NOFS);
rgd = kmem_cache_zalloc(gfs2_rgrpd_cachep, GFP_NOFS);
tr = kmem_cache_zalloc(gfs2_trans_cachep, GFP_NOFS);
bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL);
ret = kmem_cache_zalloc(jbd2_journal_head_cache, GFP_NOFS);
ret = kmem_cache_zalloc(jbd2_journal_head_cache,
new_transaction = kmem_cache_zalloc(transaction_cache,
xd = kmem_cache_zalloc(xattr_datum_cache, GFP_KERNEL);
ref = kmem_cache_zalloc(xattr_ref_cache, GFP_KERNEL);
kn = kmem_cache_zalloc(kernfs_node_cache, GFP_KERNEL);
ret = kmem_cache_zalloc(kernfs_iattrs_cache, GFP_KERNEL);
struct file_lock *fl = kmem_cache_zalloc(filelock_cache, GFP_KERNEL);
struct file_lease *fl = kmem_cache_zalloc(filelease_cache, GFP_KERNEL);
struct mount *mnt = kmem_cache_zalloc(mnt_cache, GFP_KERNEL);
cookie = kmem_cache_zalloc(fscache_cookie_jar, GFP_KERNEL);
dreq = kmem_cache_zalloc(nfs_direct_cachep, GFP_KERNEL);
kmem_cache_zalloc(nfs_page_cachep, nfs_io_gfp_mask());
struct nfs_pgio_header *p = kmem_cache_zalloc(nfs_rdata_cachep, GFP_KERNEL);
p = kmem_cache_zalloc(nfs_cdata_cachep, nfs_io_gfp_mask());
p = kmem_cache_zalloc(nfs_wdata_cachep, nfs_io_gfp_mask());
clp = kmem_cache_zalloc(client_slab, GFP_KERNEL);
co = kmem_cache_zalloc(odstate_slab, GFP_KERNEL);
stid = kmem_cache_zalloc(slab, GFP_KERNEL);
lock = kmem_cache_zalloc(dlm_lock_cache, GFP_NOFS);
res = kmem_cache_zalloc(dlm_lockres_cache, GFP_NOFS);
res->lockname.name = kmem_cache_zalloc(dlm_lockname_cache, GFP_NOFS);
kmem_cache_zalloc(ocfs2_dquot_cachep, GFP_NOFS);
new_op = kmem_cache_zalloc(op_cache, GFP_KERNEL);
xattrs = kmem_cache_zalloc(pidfs_xattr_cachep, GFP_KERNEL);
new_attr = kmem_cache_zalloc(pidfs_attr_cachep, GFP_KERNEL);
ent = kmem_cache_zalloc(proc_dir_entry_cache, GFP_KERNEL);
netd = kmem_cache_zalloc(proc_dir_entry_cache, GFP_KERNEL);
return kmem_cache_zalloc(dquot_cachep, GFP_NOFS);
p = kmem_cache_zalloc(seq_file_cache, GFP_KERNEL);
ce = kmem_cache_zalloc(cache_slab, GFP_KERNEL);
struct ksmbd_work *work = kmem_cache_zalloc(work_cache, KSMBD_DEFAULT_GFP);
fp = kmem_cache_zalloc(filp_cache, KSMBD_DEFAULT_GFP);
vi = kmem_cache_zalloc(fsverity_info_cachep, GFP_KERNEL);
xefi = kmem_cache_zalloc(xfs_extfree_item_cache,
cur = kmem_cache_zalloc(cache,
state = kmem_cache_zalloc(xfs_da_state_cache,
dfp = kmem_cache_zalloc(xfs_defer_pending_cache,
xmi = kmem_cache_zalloc(xfs_exchmaps_intent_cache,
ip->i_cowfp = kmem_cache_zalloc(xfs_ifork_cache,
*ppargsp = kmem_cache_zalloc(xfs_parent_args_cache, GFP_KERNEL);
ifp = kmem_cache_zalloc(xfs_ifork_cache, XCHK_GFP_FLAGS);
ifp = kmem_cache_zalloc(xfs_ifork_cache, XCHK_GFP_FLAGS);
attrip = kmem_cache_zalloc(xfs_attri_cache, GFP_KERNEL | __GFP_NOFAIL);
attrdp = kmem_cache_zalloc(xfs_attrd_cache, GFP_KERNEL | __GFP_NOFAIL);
new = kmem_cache_zalloc(xfs_attr_intent_cache,
buip = kmem_cache_zalloc(xfs_bui_cache, GFP_KERNEL | __GFP_NOFAIL);
budp = kmem_cache_zalloc(xfs_bud_cache, GFP_KERNEL | __GFP_NOFAIL);
bi = kmem_cache_zalloc(xfs_bmap_intent_cache,
bp = kmem_cache_zalloc(xfs_buf_cache,
bip = kmem_cache_zalloc(xfs_buf_item_cache, GFP_KERNEL | __GFP_NOFAIL);
dqp = kmem_cache_zalloc(xfs_dquot_cache, GFP_KERNEL | __GFP_NOFAIL);
xmi_lip = kmem_cache_zalloc(xfs_xmi_cache, GFP_KERNEL | __GFP_NOFAIL);
xmd_lip = kmem_cache_zalloc(xfs_xmd_cache, GFP_KERNEL | __GFP_NOFAIL);
efip = kmem_cache_zalloc(xfs_efi_cache,
efdp = kmem_cache_zalloc(xfs_efd_cache,
xefi = kmem_cache_zalloc(xfs_extfree_item_cache,
icp = kmem_cache_zalloc(xfs_icreate_cache, GFP_KERNEL | __GFP_NOFAIL);
iip = ip->i_itemp = kmem_cache_zalloc(xfs_ili_cache,
iup = kmem_cache_zalloc(xfs_iunlink_cache, GFP_KERNEL | __GFP_NOFAIL);
tic = kmem_cache_zalloc(xfs_log_ticket_cache,
cuip = kmem_cache_zalloc(xfs_cui_cache,
cudp = kmem_cache_zalloc(xfs_cud_cache, GFP_KERNEL | __GFP_NOFAIL);
ruip = kmem_cache_zalloc(xfs_rui_cache,
rudp = kmem_cache_zalloc(xfs_rud_cache, GFP_KERNEL | __GFP_NOFAIL);
tp = kmem_cache_zalloc(xfs_trans_cache, GFP_KERNEL | __GFP_NOFAIL);
ntp = kmem_cache_zalloc(xfs_trans_cache, GFP_KERNEL | __GFP_NOFAIL);
tp->t_dqinfo = kmem_cache_zalloc(xfs_dqtrx_cache,
kmem_cache_zalloc(_cache, irqs_disabled() ? GFP_ATOMIC : GFP_KERNEL)
((handle_t *)kmem_cache_zalloc(jbd2_handle_cache, _gfp_flags))
amark = kmem_cache_zalloc(audit_tree_mark_cachep, GFP_KERNEL);
new = kmem_cache_zalloc(cred_jar, GFP_KERNEL);
tsk->delays = kmem_cache_zalloc(delayacct_cache, GFP_KERNEL);
cd->data = kmem_cache_zalloc(ctx_cache, GFP_KERNEL);
sig = kmem_cache_zalloc(signal_cachep, GFP_KERNEL);
ns = kmem_cache_zalloc(pid_ns_cachep, GFP_KERNEL);
stats_new = kmem_cache_zalloc(taskstats_cache, GFP_KERNEL);
tmr = kmem_cache_zalloc(posix_timers_cache, GFP_KERNEL);
fault = kmem_cache_zalloc(fault_cache, GFP_NOWAIT);
new = kmem_cache_zalloc(uid_cachep, GFP_KERNEL);
ns = kmem_cache_zalloc(user_ns_cachep, GFP_KERNEL);
ns = kmem_cache_zalloc(uts_ns_cache, GFP_KERNEL);
obj = kmem_cache_zalloc(cache, gfp);
rmap_item = kmem_cache_zalloc(rmap_item_cache, GFP_KERNEL |
memcg = kmem_cache_zalloc(memcg_cachep, GFP_KERNEL);
return kmem_cache_zalloc(cache, GFP_KERNEL);
region = kmem_cache_zalloc(vm_region_jar, GFP_KERNEL);
s = kmem_cache_zalloc(kmem_cache, GFP_KERNEL);
struct kmem_cache *s = kmem_cache_zalloc(kmem_cache, GFP_NOWAIT);
struct kmem_cache *s = kmem_cache_zalloc(kmem_cache, GFP_NOWAIT);
return kmem_cache_zalloc(swap_table_cachep, gfp);
vas[area] = kmem_cache_zalloc(vmap_area_cachep, GFP_KERNEL);
vas[area] = kmem_cache_zalloc(
free = kmem_cache_zalloc(vmap_area_cachep, GFP_NOWAIT);
free = kmem_cache_zalloc(vmap_area_cachep, GFP_NOWAIT);
va = kmem_cache_zalloc(vmap_area_cachep, GFP_NOWAIT);
return kmem_cache_zalloc(zspage_cachep, gfp);
orig_entry = kmem_cache_zalloc(batadv_tt_orig_cache, GFP_ATOMIC);
tt_global_entry = kmem_cache_zalloc(batadv_tg_cache,
m = kmem_cache_zalloc(ceph_msg_cache, flags);
net = kmem_cache_zalloc(net_cachep, GFP_KERNEL);
q = kmem_cache_zalloc(f->frags_cachep, GFP_ATOMIC);
struct mfc_cache *c = kmem_cache_zalloc(mrt_cachep, GFP_KERNEL);
struct mfc_cache *c = kmem_cache_zalloc(mrt_cachep, GFP_ATOMIC);
fn = kmem_cache_zalloc(fib6_node_kmem, GFP_ATOMIC);
struct mfc6_cache *c = kmem_cache_zalloc(mrt_cachep, GFP_KERNEL);
struct mfc6_cache *c = kmem_cache_zalloc(mrt_cachep, GFP_ATOMIC);
psock = kmem_cache_zalloc(kcm_psockp, GFP_KERNEL);
mux = kmem_cache_zalloc(kcm_muxp, GFP_KERNEL);
flow = kmem_cache_zalloc(flow_offload_cachep, GFP_ATOMIC);
flow = kmem_cache_zalloc(flow_cache, GFP_KERNEL);
conn = kmem_cache_zalloc(rds_conn_slab, gfp);
tc = kmem_cache_zalloc(rds_tcp_conn_slab, gfp);
call = kmem_cache_zalloc(rxrpc_call_jar, gfp);
f = kmem_cache_zalloc(fq_flow_cachep, GFP_ATOMIC | __GFP_NOWARN);
retval = kmem_cache_zalloc(sctp_chunk_cachep, gfp);
xsk_addr = kmem_cache_zalloc(xsk_tx_generic_cache,
xsk_addr = kmem_cache_zalloc(xsk_tx_generic_cache,
x = kmem_cache_zalloc(xfrm_state_cache, GFP_ATOMIC);
key = kmem_cache_zalloc(key_jar, GFP_KERNEL);
file->f_security = kmem_cache_zalloc(lsm_file_cache, GFP_KERNEL);
inode->i_security = kmem_cache_zalloc(lsm_inode_cache, gfp);
xpd_node = kmem_cache_zalloc(avc_xperms_decision_cachep, GFP_NOWAIT);
xpd->allowed = kmem_cache_zalloc(avc_xperms_data_cachep,
xpd->auditallow = kmem_cache_zalloc(avc_xperms_data_cachep,
xpd->dontaudit = kmem_cache_zalloc(avc_xperms_data_cachep,
xp_node = kmem_cache_zalloc(avc_xperms_cachep, GFP_NOWAIT);
node = kmem_cache_zalloc(avc_node_cachep, GFP_NOWAIT);
newnode = kmem_cache_zalloc(avtab_node_cachep, GFP_KERNEL);
xperms = kmem_cache_zalloc(avtab_xperms_cachep, GFP_KERNEL);
e_iter = kmem_cache_zalloc(ebitmap_node_cachep,
new = kmem_cache_zalloc(ebitmap_node_cachep, GFP_ATOMIC);
tmp = kmem_cache_zalloc(ebitmap_node_cachep,
new = kmem_cache_zalloc(ebitmap_node_cachep, GFP_ATOMIC);
tmp = kmem_cache_zalloc(hashtab_node_cachep,
newnode = kmem_cache_zalloc(hashtab_node_cachep, GFP_KERNEL);
nrp = kmem_cache_zalloc(smack_rule_cache, gfp);
sp = kmem_cache_zalloc(smack_rule_cache, GFP_KERNEL);
work = kmem_cache_zalloc(async_pf_cache, GFP_NOWAIT);
work = kmem_cache_zalloc(async_pf_cache, GFP_ATOMIC);
vcpu = kmem_cache_zalloc(kvm_vcpu_cache, GFP_KERNEL_ACCOUNT);