SLAB_HWCACHE_ALIGN
SLAB_HWCACHE_ALIGN|SLAB_ACCOUNT, spufs_init_once);
SLAB_PANIC | SLAB_HWCACHE_ALIGN, NULL);
SLAB_PANIC | SLAB_HWCACHE_ALIGN, NULL);
0, (SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT|
SLAB_HWCACHE_ALIGN | SLAB_PANIC, NULL);
SLAB_HWCACHE_ALIGN | SLAB_PANIC, NULL);
SLAB_HWCACHE_ALIGN | SLAB_TYPESAFE_BY_RCU, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN | SLAB_RECLAIM_ACCOUNT | SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN,
amdgpu_mux_chunk_slab = KMEM_CACHE(amdgpu_mux_chunk, SLAB_HWCACHE_ALIGN);
amdgpu_sync_slab = KMEM_CACHE(amdgpu_sync_entry, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN,
slab_objects = KMEM_CACHE(drm_i915_gem_object, SLAB_HWCACHE_ALIGN);
slab_ce = KMEM_CACHE(intel_context, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN,
slab_cache = KMEM_CACHE(active_node, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN |
SLAB_HWCACHE_ALIGN |
SLAB_HWCACHE_ALIGN |
slab_vmas = KMEM_CACHE(i915_vma, SLAB_HWCACHE_ALIGN);
slab_vma_resources = KMEM_CACHE(i915_vma_resource, SLAB_HWCACHE_ALIGN);
0, SLAB_HWCACHE_ALIGN, sizeof(struct lima_sched_task),
0, SLAB_HWCACHE_ALIGN, sizeof(struct lima_sched_task),
SLAB_HWCACHE_ALIGN, NULL);
sched_fence_slab = KMEM_CACHE(drm_sched_fence, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN,
SLAB_HWCACHE_ALIGN,
SLAB_HWCACHE_ALIGN,
SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
const slab_flags_t flags = SLAB_HWCACHE_ALIGN;
SLAB_HWCACHE_ALIGN | SLAB_CACHE_DMA32,
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
CHAP_BUF_SIZE, 0, SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN,
SLAB_HWCACHE_ALIGN,
SLAB_HWCACHE_ALIGN,
SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
sas_task_cache = KMEM_CACHE(sas_task, SLAB_HWCACHE_ALIGN);
sas_event_cache = KMEM_CACHE(asd_sas_event, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN, NULL);
sizeof(struct qedf_io_work), 0, SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SCSI_SENSE_BUFFERSIZE, 0, SLAB_HWCACHE_ALIGN,
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
maple_queue_cache = KMEM_CACHE(maple_buffer, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN, NULL);
stub_priv_cache = KMEM_CACHE(stub_priv, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN|SLAB_ACCOUNT,
kiocb_cachep = KMEM_CACHE(aio_kiocb, SLAB_HWCACHE_ALIGN|SLAB_PANIC);
kioctx_cachep = KMEM_CACHE(kioctx,SLAB_HWCACHE_ALIGN|SLAB_PANIC);
backing_aio_cachep = KMEM_CACHE(backing_aio, SLAB_HWCACHE_ALIGN);
0, SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN | info->flags, info->ctor);
0, SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_ACCOUNT, NULL);
SLAB_HWCACHE_ALIGN | SLAB_PANIC |
&args, SLAB_HWCACHE_ALIGN | SLAB_PANIC |
SLAB_HWCACHE_ALIGN|SLAB_ACCOUNT|SLAB_RECLAIM_ACCOUNT,
SLAB_HWCACHE_ALIGN|SLAB_ACCOUNT, hfs_init_once);
SLAB_HWCACHE_ALIGN, NULL);
HFSPLUS_INODE_SIZE, 0, SLAB_HWCACHE_ALIGN|SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN|SLAB_TEMPORARY);
SLAB_HWCACHE_ALIGN|SLAB_TEMPORARY,
raw_dirent_slab = KMEM_CACHE(jffs2_raw_dirent, SLAB_HWCACHE_ALIGN);
raw_inode_slab = KMEM_CACHE(jffs2_raw_inode, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN|SLAB_PANIC, offsetof(struct filename, iname),
0, SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_ACCOUNT, NULL);
SLAB_HWCACHE_ALIGN | SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN | SLAB_ACCOUNT,
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
0, (SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT|
(SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT|
SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT,
0, SLAB_HWCACHE_ALIGN, NULL);
(SLAB_HWCACHE_ALIGN | SLAB_RECLAIM_ACCOUNT |
(SLAB_HWCACHE_ALIGN | SLAB_RECLAIM_ACCOUNT |
(SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT|
SLAB_HWCACHE_ALIGN, 0,
MAX_CIFS_SMALL_BUFFER_SIZE, 0, SLAB_HWCACHE_ALIGN,
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
&response_args, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT|SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN|SLAB_PANIC,
SLAB_HWCACHE_ALIGN |
(SLAB_HWCACHE_ALIGN |
SLAB_HWCACHE_ALIGN | SLAB_PANIC | SLAB_ACCOUNT |
SLAB_HWCACHE_ALIGN|SLAB_ACCOUNT, init_once);
SLAB_HWCACHE_ALIGN | SLAB_PANIC | SLAB_ACCOUNT);
SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_TYPESAFE_BY_RCU|
SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN | SLAB_PANIC | SLAB_ACCOUNT,
SLAB_HWCACHE_ALIGN | SLAB_ACCOUNT, NULL);
0, SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_PANIC | SLAB_HWCACHE_ALIGN, NULL);
SLAB_PANIC | SLAB_HWCACHE_ALIGN);
#define SLAB_CORE_FLAGS (SLAB_HWCACHE_ALIGN | SLAB_CACHE_DMA | \
if (flags & SLAB_HWCACHE_ALIGN) {
SLAB_HWCACHE_ALIGN | SLAB_NO_OBJ_EXT, 0, 0);
SLAB_HWCACHE_ALIGN | SLAB_NO_OBJ_EXT, 0, 0);
return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_HWCACHE_ALIGN));
SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_TYPESAFE_BY_RCU|
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
br_fdb_cache = KMEM_CACHE(net_bridge_fdb_entry, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN|SLAB_PANIC,
SLAB_HWCACHE_ALIGN|SLAB_PANIC|
SLAB_HWCACHE_ALIGN|SLAB_PANIC,
SLAB_HWCACHE_ALIGN | SLAB_PANIC,
SLAB_HWCACHE_ALIGN | SLAB_ACCOUNT |
peer_cachep = KMEM_CACHE(inet_peer, SLAB_HWCACHE_ALIGN | SLAB_PANIC);
mrt_cachep = KMEM_CACHE(mfc_cache, SLAB_HWCACHE_ALIGN | SLAB_PANIC);
SLAB_HWCACHE_ALIGN | SLAB_PANIC);
SLAB_HWCACHE_ALIGN | SLAB_PANIC |
SLAB_HWCACHE_ALIGN | SLAB_PANIC |
SLAB_HWCACHE_ALIGN | SLAB_ACCOUNT);
mrt_cachep = KMEM_CACHE(mfc6_cache, SLAB_HWCACHE_ALIGN);
SLAB_HWCACHE_ALIGN | SLAB_ACCOUNT, NULL);
xfrm6_tunnel_spi_kmem = KMEM_CACHE(xfrm6_tunnel_spi, SLAB_HWCACHE_ALIGN);
kcm_muxp = KMEM_CACHE(kcm_mux, SLAB_HWCACHE_ALIGN);
kcm_psockp = KMEM_CACHE(kcm_psock, SLAB_HWCACHE_ALIGN);
ip_vs_conn_cachep = KMEM_CACHE(ip_vs_conn, SLAB_HWCACHE_ALIGN);
SLAB_TYPESAFE_BY_RCU | SLAB_HWCACHE_ALIGN, NULL);
flow_offload_cachep = KMEM_CACHE(flow_offload, SLAB_HWCACHE_ALIGN);
0, SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN, NULL);
SLAB_HWCACHE_ALIGN, NULL);
0, SLAB_HWCACHE_ALIGN, NULL);
sctp_bucket_cachep = KMEM_CACHE(sctp_bind_bucket, SLAB_HWCACHE_ALIGN);
sctp_chunk_cachep = KMEM_CACHE(sctp_chunk, SLAB_HWCACHE_ALIGN);
(SLAB_HWCACHE_ALIGN |
0, (SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT|
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN,
0, SLAB_HWCACHE_ALIGN, NULL);
xfrm_dst_cache = KMEM_CACHE(xfrm_dst, SLAB_HWCACHE_ALIGN | SLAB_PANIC);
SLAB_HWCACHE_ALIGN | SLAB_PANIC);
0, SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);