KMEM_CACHE
pte_list_desc_cache = KMEM_CACHE(pte_list_desc, SLAB_ACCOUNT);
bfq_pool = KMEM_CACHE(bfq_queue, 0);
blk_requestq_cachep = KMEM_CACHE(request_queue, SLAB_PANIC);
bio_fallback_crypt_ctx_cache = KMEM_CACHE(bio_fallback_crypt_ctx, 0);
bio_crypt_ctx_cache = KMEM_CACHE(bio_crypt_ctx, 0);
rbd_img_request_cache = KMEM_CACHE(rbd_img_request, 0);
rbd_obj_request_cache = KMEM_CACHE(rbd_obj_request, 0);
slab_fences = KMEM_CACHE(mock_fence,
slab_fences = KMEM_CACHE(mock_fence,
ioat_sed_cache = KMEM_CACHE(ioat_sed_ent, 0);
amdgpu_mux_chunk_slab = KMEM_CACHE(amdgpu_mux_chunk, SLAB_HWCACHE_ALIGN);
amdgpu_sync_slab = KMEM_CACHE(amdgpu_sync_entry, SLAB_HWCACHE_ALIGN);
slab_blocks = KMEM_CACHE(drm_buddy_block, 0);
slab_luts = KMEM_CACHE(i915_lut_handle, 0);
slab_objects = KMEM_CACHE(drm_i915_gem_object, SLAB_HWCACHE_ALIGN);
slab_ce = KMEM_CACHE(intel_context, SLAB_HWCACHE_ALIGN);
slab_cache = KMEM_CACHE(active_node, SLAB_HWCACHE_ALIGN);
slab_execute_cbs = KMEM_CACHE(execute_cb,
slab_dependencies = KMEM_CACHE(i915_dependency,
slab_priorities = KMEM_CACHE(i915_priolist, 0);
slab_vmas = KMEM_CACHE(i915_vma, SLAB_HWCACHE_ALIGN);
slab_vma_resources = KMEM_CACHE(i915_vma_resource, SLAB_HWCACHE_ALIGN);
sched_fence_slab = KMEM_CACHE(drm_sched_fence, SLAB_HWCACHE_ALIGN);
bch_search_cache = KMEM_CACHE(search, 0);
_cell_cache = KMEM_CACHE(dm_bio_prison_cell, 0);
_cell_cache = KMEM_CACHE(dm_bio_prison_cell_v2, 0);
migration_cache = KMEM_CACHE(dm_cache_migration, 0);
_hydration_cache = KMEM_CACHE(dm_clone_region_hydration, 0);
_dm_io_cache = KMEM_CACHE(io, 0);
_flush_entry_cache = KMEM_CACHE(dm_dirty_log_flush_entry, 0);
backing_req_cache = KMEM_CACHE(pcache_backing_dev_req, 0);
key_cache = KMEM_CACHE(pcache_cache_key, 0);
exception_cache = KMEM_CACHE(dm_exception, 0);
pending_cache = KMEM_CACHE(dm_snap_pending_exception, 0);
_new_mapping_cache = KMEM_CACHE(dm_thin_new_mapping, 0);
_dm_event_cache = KMEM_CACHE(dm_uevent, 0);
log->io_kc = KMEM_CACHE(r5l_io_unit, 0);
ppl_conf->io_kc = KMEM_CACHE(ppl_io_unit, 0);
node_cache = KMEM_CACHE(allowedips_node, 0);
peer_cache = KMEM_CACHE(wg_peer, 0);
entry_cache = KMEM_CACHE(ratelimiter_entry, 0);
sas_task_cache = KMEM_CACHE(sas_task, SLAB_HWCACHE_ALIGN);
sas_event_cache = KMEM_CACHE(asd_sas_event, SLAB_HWCACHE_ALIGN);
virtscsi_cmd_cache = KMEM_CACHE(virtio_scsi_cmd, 0);
maple_queue_cache = KMEM_CACHE(maple_buffer, SLAB_HWCACHE_ALIGN);
stub_priv_cache = KMEM_CACHE(stub_priv, SLAB_HWCACHE_ALIGN);
kiocb_cachep = KMEM_CACHE(aio_kiocb, SLAB_HWCACHE_ALIGN|SLAB_PANIC);
kioctx_cachep = KMEM_CACHE(kioctx,SLAB_HWCACHE_ALIGN|SLAB_PANIC);
backing_aio_cachep = KMEM_CACHE(backing_aio, SLAB_HWCACHE_ALIGN);
btrfs_path_cachep = KMEM_CACHE(btrfs_path, 0);
delayed_node_cache = KMEM_CACHE(btrfs_delayed_node, 0);
btrfs_delayed_ref_head_cachep = KMEM_CACHE(btrfs_delayed_ref_head, 0);
btrfs_delayed_ref_node_cachep = KMEM_CACHE(btrfs_delayed_ref_node, 0);
btrfs_delayed_extent_op_cachep = KMEM_CACHE(btrfs_delayed_extent_op, 0);
btrfs_free_space_cachep = KMEM_CACHE(btrfs_free_space, 0);
btrfs_ordered_extent_cache = KMEM_CACHE(btrfs_ordered_extent, 0);
btrfs_trans_handle_cachep = KMEM_CACHE(btrfs_trans_handle, SLAB_TEMPORARY);
bh_cachep = KMEM_CACHE(buffer_head,
ceph_cap_cachep = KMEM_CACHE(ceph_cap, 0);
ceph_cap_snap_cachep = KMEM_CACHE(ceph_cap_snap, 0);
ceph_cap_flush_cachep = KMEM_CACHE(ceph_cap_flush,
ceph_dentry_cachep = KMEM_CACHE(ceph_dentry_info,
ceph_file_cachep = KMEM_CACHE(ceph_file_info, 0);
ceph_dir_file_cachep = KMEM_CACHE(ceph_dir_file_info, 0);
ceph_mds_request_cachep = KMEM_CACHE(ceph_mds_request, 0);
fscrypt_inode_info_cachep = KMEM_CACHE(fscrypt_inode_info,
dio_cache = KMEM_CACHE(dio, SLAB_PANIC);
return KMEM_CACHE(dlm_msg, 0);
return KMEM_CACHE(dlm_mhandle, 0);
ext4_system_zone_cachep = KMEM_CACHE(ext4_system_zone, 0);
ext4_es_cachep = KMEM_CACHE(extent_status, SLAB_RECLAIM_ACCOUNT);
ext4_pending_cachep = KMEM_CACHE(pending_reservation, SLAB_RECLAIM_ACCOUNT);
ext4_fc_dentry_cachep = KMEM_CACHE(ext4_fc_dentry_update,
ext4_pspace_cachep = KMEM_CACHE(ext4_prealloc_space,
ext4_ac_cachep = KMEM_CACHE(ext4_allocation_context,
ext4_free_data_cachep = KMEM_CACHE(ext4_free_data,
io_end_cachep = KMEM_CACHE(ext4_io_end, SLAB_RECLAIM_ACCOUNT);
io_end_vec_cachep = KMEM_CACHE(ext4_io_end_vec, 0);
bio_post_read_ctx_cache = KMEM_CACHE(bio_post_read_ctx, SLAB_RECLAIM_ACCOUNT);
hostfs_inode_cache = KMEM_CACHE(hostfs_inode_info, 0);
jbd2_inode_cache = KMEM_CACHE(jbd2_inode, 0);
jbd2_handle_cache = KMEM_CACHE(jbd2_journal_handle, SLAB_TEMPORARY);
jbd2_revoke_record_cache = KMEM_CACHE(jbd2_revoke_record_s,
jbd2_revoke_table_cache = KMEM_CACHE(jbd2_revoke_table_s,
full_dnode_slab = KMEM_CACHE(jffs2_full_dnode, 0);
raw_dirent_slab = KMEM_CACHE(jffs2_raw_dirent, SLAB_HWCACHE_ALIGN);
raw_inode_slab = KMEM_CACHE(jffs2_raw_inode, SLAB_HWCACHE_ALIGN);
tmp_dnode_info_slab = KMEM_CACHE(jffs2_tmp_dnode_info, 0);
node_frag_slab = KMEM_CACHE(jffs2_node_frag, 0);
inode_cache_slab = KMEM_CACHE(jffs2_inode_cache, 0);
xattr_datum_cache = KMEM_CACHE(jffs2_xattr_datum, 0);
xattr_ref_cache = KMEM_CACHE(jffs2_xattr_ref, 0);
mb_entry_cache = KMEM_CACHE(mb_cache_entry, SLAB_RECLAIM_ACCOUNT);
nfsd_file_slab = KMEM_CACHE(nfsd_file, 0);
nfsd_file_mark_slab = KMEM_CACHE(nfsd_file_mark, 0);
nfs4_layout_cache = KMEM_CACHE(nfs4_layout, 0);
nfs4_layout_stateid_cache = KMEM_CACHE(nfs4_layout_stateid, 0);
client_slab = KMEM_CACHE(nfs4_client, 0);
openowner_slab = KMEM_CACHE(nfs4_openowner, 0);
lockowner_slab = KMEM_CACHE(nfs4_lockowner, 0);
file_slab = KMEM_CACHE(nfs4_file, 0);
stateid_slab = KMEM_CACHE(nfs4_ol_stateid, 0);
deleg_slab = KMEM_CACHE(nfs4_delegation, 0);
odstate_slab = KMEM_CACHE(nfs4_clnt_odstate, 0);
drc_slab = KMEM_CACHE(nfsd_cacherep, 0);
dnotify_struct_cache = KMEM_CACHE(dnotify_struct,
dnotify_mark_cache = KMEM_CACHE(dnotify_mark, SLAB_PANIC|SLAB_ACCOUNT);
fanotify_mark_cache = KMEM_CACHE(fanotify_mark,
fanotify_fid_event_cachep = KMEM_CACHE(fanotify_fid_event,
fanotify_path_event_cachep = KMEM_CACHE(fanotify_path_event,
KMEM_CACHE(fanotify_perm_event, SLAB_PANIC);
fanotify_mnt_event_cachep = KMEM_CACHE(fanotify_mnt_event, SLAB_PANIC);
inotify_inode_mark_cachep = KMEM_CACHE(inotify_inode_mark,
fsnotify_mark_connector_cachep = KMEM_CACHE(fsnotify_mark_connector,
fsnotify_inode_mark_connector_cachep = KMEM_CACHE(
seq_file_cache = KMEM_CACHE(seq_file, SLAB_ACCOUNT|SLAB_PANIC);
audit_buffer_cache = KMEM_CACHE(audit_buffer, SLAB_PANIC);
audit_tree_mark_cachep = KMEM_CACHE(audit_tree_mark, SLAB_PANIC);
cred_jar = KMEM_CACHE(cred,
delayacct_cache = KMEM_CACHE(task_delay_info, SLAB_PANIC|SLAB_ACCOUNT);
perf_event_cache = KMEM_CACHE(perf_event, SLAB_PANIC);
nsproxy_cachep = KMEM_CACHE(nsproxy, SLAB_PANIC|SLAB_ACCOUNT);
pid_ns_cachep = KMEM_CACHE(pid_namespace, SLAB_PANIC | SLAB_ACCOUNT);
task_group_cache = KMEM_CACHE(task_group, 0);
sigqueue_cachep = KMEM_CACHE(sigqueue, SLAB_PANIC | SLAB_ACCOUNT);
taskstats_cache = KMEM_CACHE(taskstats, SLAB_PANIC);
field_cachep = KMEM_CACHE(ftrace_event_field, SLAB_PANIC);
file_cachep = KMEM_CACHE(trace_event_file, SLAB_PANIC);
fault_cache = KMEM_CACHE(user_event_enabler_fault, 0);
user_ns_cachep = KMEM_CACHE(user_namespace, SLAB_PANIC | SLAB_ACCOUNT);
pwq_cache = KMEM_CACHE(pool_workqueue, SLAB_PANIC);
damon_region_cache = KMEM_CACHE(damon_region, 0);
mm_slot_cache = KMEM_CACHE(mm_slot, 0);
object_cache = KMEM_CACHE(kmemleak_object, SLAB_NOLEAKTRACE);
scan_area_cache = KMEM_CACHE(kmemleak_scan_area, SLAB_NOLEAKTRACE);
rmap_item_cache = KMEM_CACHE(ksm_rmap_item, 0);
stable_node_cache = KMEM_CACHE(ksm_stable_node, 0);
mm_slot_cache = KMEM_CACHE(ksm_mm_slot, 0);
memcg_pn_cachep = KMEM_CACHE(mem_cgroup_per_node,
vm_region_jar = KMEM_CACHE(vm_region, SLAB_PANIC|SLAB_ACCOUNT);
anon_vma_chain_cachep = KMEM_CACHE(anon_vma_chain,
vmap_area_cachep = KMEM_CACHE(vmap_area, SLAB_PANIC);
zswap_entry_cache = KMEM_CACHE(zswap_entry, 0);
p9_req_cache = KMEM_CACHE(p9_req_t, SLAB_TYPESAFE_BY_RCU);
br_fdb_cache = KMEM_CACHE(net_bridge_fdb_entry, SLAB_HWCACHE_ALIGN);
ceph_msg_cache = KMEM_CACHE(ceph_msg, 0);
peer_cachep = KMEM_CACHE(inet_peer, SLAB_HWCACHE_ALIGN | SLAB_PANIC);
mrt_cachep = KMEM_CACHE(mfc_cache, SLAB_HWCACHE_ALIGN | SLAB_PANIC);
ipv4_dst_ops.kmem_cachep = KMEM_CACHE(rtable,
fib6_node_kmem = KMEM_CACHE(fib6_node,
mrt_cachep = KMEM_CACHE(mfc6_cache, SLAB_HWCACHE_ALIGN);
xfrm6_tunnel_spi_kmem = KMEM_CACHE(xfrm6_tunnel_spi, SLAB_HWCACHE_ALIGN);
kcm_muxp = KMEM_CACHE(kcm_mux, SLAB_HWCACHE_ALIGN);
kcm_psockp = KMEM_CACHE(kcm_psock, SLAB_HWCACHE_ALIGN);
ip_vs_conn_cachep = KMEM_CACHE(ip_vs_conn, SLAB_HWCACHE_ALIGN);
conncount_conn_cachep = KMEM_CACHE(nf_conncount_tuple, 0);
conncount_rb_cachep = KMEM_CACHE(nf_conncount_rb, 0);
nf_ct_expect_cachep = KMEM_CACHE(nf_conntrack_expect, 0);
flow_offload_cachep = KMEM_CACHE(flow_offload, SLAB_HWCACHE_ALIGN);
rds_conn_slab = KMEM_CACHE(rds_connection, 0);
rds_tcp_conn_slab = KMEM_CACHE(rds_tcp_connection, 0);
rds_tcp_incoming_slab = KMEM_CACHE(rds_tcp_incoming, 0);
sctp_bucket_cachep = KMEM_CACHE(sctp_bind_bucket, SLAB_HWCACHE_ALIGN);
sctp_chunk_cachep = KMEM_CACHE(sctp_chunk, SLAB_HWCACHE_ALIGN);
xfrm_dst_cache = KMEM_CACHE(xfrm_dst, SLAB_HWCACHE_ALIGN | SLAB_PANIC);
xfrm_state_cache = KMEM_CACHE(xfrm_state,
avc_node_cachep = KMEM_CACHE(avc_node, SLAB_PANIC);
avc_xperms_cachep = KMEM_CACHE(avc_xperms_node, SLAB_PANIC);
avc_xperms_decision_cachep = KMEM_CACHE(avc_xperms_decision_node, SLAB_PANIC);
avc_xperms_data_cachep = KMEM_CACHE(extended_perms_data, SLAB_PANIC);
avtab_node_cachep = KMEM_CACHE(avtab_node, SLAB_PANIC);
avtab_xperms_cachep = KMEM_CACHE(avtab_extended_perms, SLAB_PANIC);
ebitmap_node_cachep = KMEM_CACHE(ebitmap_node, SLAB_PANIC);
hashtab_node_cachep = KMEM_CACHE(hashtab_node, SLAB_PANIC);
smack_rule_cache = KMEM_CACHE(smack_rule, 0);
async_pf_cache = KMEM_CACHE(kvm_async_pf, 0);