slab_size
bslab->slab_size = size;
unsigned int slab_size = bs_bio_slab_size(bs);
bslab = xa_load(&bio_slabs, slab_size);
xa_erase(&bio_slabs, slab_size);
unsigned int slab_size;
.slab_size = __cpu_to_le64(config.slab_size),
.slab_size = __le64_to_cpu(config.slab_size),
result = VDO_ASSERT(config->slab_size > 0, "slab size unspecified");
result = VDO_ASSERT(is_power_of_2(config->slab_size),
result = VDO_ASSERT(config->slab_size <= (1 << MAX_VDO_SLAB_BITS),
result = VDO_ASSERT(config->slab_journal_blocks <= config->slab_size,
result = vdo_configure_slab(config->slab_size, config->slab_journal_blocks,
block_count_t slab_size = slab_config.slab_blocks;
(unsigned long long) slab_size, zone_count);
slab_count = (partition->count / slab_size);
int vdo_configure_slab(block_count_t slab_size, block_count_t slab_journal_blocks,
if (slab_journal_blocks >= slab_size)
ref_blocks = vdo_get_saved_reference_count_size(slab_size - slab_journal_blocks);
if (meta_blocks >= slab_size)
data_blocks = slab_size - meta_blocks;
.slab_blocks = slab_size,
int __must_check vdo_configure_slab(block_count_t slab_size,
block_count_t slab_size; /* number of blocks in a slab */
__le64 slab_size;
write_block_count_t("slabSize : ", config->slab_size, ", ", buf, maxlen);
block_count_t slab_size;
slab_size = depot->slab_config.slab_blocks;
slab_origin = depot->first_block + (depot->slab_count * slab_size);
depot->new_slab_count++, slab_origin += slab_size) {
block_count_t slab_size = state.slab_config.slab_blocks;
if (!is_power_of_2(slab_size)) {
slab_size_shift = ilog2(slab_size);
int slab_size;
slab_size = offsetof(struct ext4_group_info,
slab_size, 0, SLAB_RECLAIM_ACCOUNT,
size_t slab_size;
slab_size = 1 << (i+10);
jbd2_slab[i] = kmem_cache_create(jbd2_slab_names[i], slab_size,
slab_size, 0, NULL);
static int jbd2_journal_create_slab(size_t slab_size);
length = slab_size(slab);
memset(kasan_reset_tag(addr), POISON_INUSE, slab_size(slab));
unsigned int slab_size = (unsigned int)PAGE_SIZE << order;
rem = slab_size % size;
if (rem <= slab_size / fract_leftover)
return objext_offset + objext_size <= slab_size(slab);
end = start + slab_size(slab);
SLAB_ATTR_RO(slab_size);
if (slab_size(s1) == slab_size(s2))
result = slab_size(s1) < slab_size(s2);
slab->slab_size = get_obj("slab_size");
unsigned int sanity_checks, slab_size, store_user, trace;
return slab_size(s) - s->objects * s->object_size;
s->slab_size, s->slabs - s->partial - s->cpu_slabs,
s->slab_size - s->object_size, s->cpu_slabs, onoff(s->store_user),
(s->slab_size - s->object_size) * s->objects);
((page_size << s->order) - s->objs_per_slab * s->slab_size) *
store_size(size_str, slab_size(s));
size = slab_size(s);
objwaste = s->slab_size - s->object_size;
if (s->slab_size < min_memobj)
min_memobj = s->slab_size;
if (s->slab_size > max_memobj)
max_memobj = s->slab_size;
total_objsize += s->objects * s->slab_size;