ksize
size_t ksize, size_t kzsize);
size_t ksize,
free_mem_ptr = (ulg)output_start + ksize;
free_mem_end_ptr = (ulg)output_start + ksize + 0x200000;
unsigned int kstart, ksize;
ksize = PAGE_ALIGN(((u32)_end - (u32)CONFIG_KERNEL_START));
memblock_reserve(kstart, ksize);
static int opal_get_variable(const char *key, u64 ksize, u8 *data, u64 *dsize)
rc = opal_secvar_get(key, ksize, data, dsize);
static int opal_set_variable(const char *key, u64 ksize, u8 *data, u64 dsize)
rc = opal_secvar_enqueue_update(key, ksize, data, dsize);
static int fpstate_realloc(u64 xfeatures, unsigned int ksize,
fpsize = ksize + ALIGN(offsetof(struct fpstate, regs), 64);
newfps->size = ksize;
unsigned int ksize, usize;
ksize = xstate_calculate_size(mask, compacted);
perm->__state_size = ksize;
unsigned int ksize, usize;
ksize = perm->__state_size;
if (fpstate_realloc(xfd_event, ksize, usize, guest_fpu))
if (vec->ksize) {
err = do_setkey(crypto_shash_setkey, tfm, vec->key, vec->ksize,
if (vec->ksize) {
err = do_setkey(crypto_ahash_setkey, tfm, vec->key, vec->ksize,
vec->ksize = 0;
vec->ksize = maxkeysize;
vec->ksize = prandom_u32_inclusive(rng, 1, maxkeysize);
generate_random_bytes(rng, (u8 *)vec->key, vec->ksize);
crypto_ahash_reqtfm(req), vec->key, vec->ksize);
vec->psize, vec->ksize);
if (template[nr_unkeyed].ksize)
if (!template[nr_unkeyed + nr_keyed].ksize) {
template[nr_unkeyed + nr_keyed].ksize);
static int prepare_keybuf(const u8 *key, unsigned int ksize,
keybuf = kmalloc(key_offset + ksize, GFP_KERNEL);
memcpy(keyptr, key, ksize);
#define do_setkey(setkey_f, tfm, key, ksize, cfg, alignmask) \
err = prepare_keybuf((key), (ksize), (cfg), (alignmask), \
err = setkey_f((tfm), keyptr, (ksize)); \
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 8,
.ksize = 8,
.ksize = 8,
.ksize = 8,
.ksize = 8,
.ksize = 8,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 4,
.ksize = 8,
.ksize = 8,
.ksize = 8,
.ksize = 8,
.ksize = 32,
.ksize = 64,
.ksize = 1,
.ksize = 32,
.ksize = 64,
.ksize = 1,
.ksize = 64,
.ksize = 32,
.ksize = 1,
.ksize = 64,
.ksize = 32,
.ksize = 1,
.ksize = 64,
.ksize = 32,
.ksize = 1,
.ksize = 64,
.ksize = 1,
.ksize = 32,
.ksize = 64,
.ksize = 1,
.ksize = 32,
unsigned short ksize;
.ksize = 32,
.ksize = 32,
.ksize = 32,
.ksize = 37,
.ksize = 32,
.ksize = 4,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 4,
.ksize = 16,
.ksize = 25,
.ksize = 16,
.ksize = 80,
.ksize = 80,
.ksize = 20,
.ksize = 4,
.ksize = 20,
.ksize = 25,
.ksize = 20,
.ksize = 80,
.ksize = 80,
.ksize = 20,
.ksize = 4,
.ksize = 20,
.ksize = 25,
.ksize = 20,
.ksize = 80,
.ksize = 80,
.ksize = 20,
.ksize = 4,
.ksize = 131,
.ksize = 131,
.ksize = 32,
.ksize = 32,
.ksize = 32,
.ksize = 32,
.ksize = 4,
.ksize = 32,
.ksize = 37,
.ksize = 32,
.ksize = 80,
.ksize = 80,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 32,
.ksize = 32,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 32,
.ksize = 24,
.ksize = 24,
.ksize = 24,
.ksize = 24,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 16,
.ksize = 20,
.ksize = 4,
.ksize = 131,
.ksize = 131,
.ksize = 20,
.ksize = 4,
.ksize = 131,
.ksize = 131,
.ksize = 20,
.ksize = 4,
.ksize = 131,
.ksize = 131,
.ksize = 20,
.ksize = 4,
.ksize = 131,
.ksize = 131,
.ksize = 20,
.ksize = 4,
.ksize = 131,
.ksize = 131,
.ksize = 20,
.ksize = 4,
.ksize = 131,
.ksize = 131,
total_old_size = ksize(container_of(ptr, struct devres, data));
ctx->cword.encrypt.ksize = (key_len - 16) / 8;
ctx->cword.decrypt.ksize = ctx->cword.encrypt.ksize;
ksize:2;
unsigned int in_size, out_size, drv_size, ksize;
ksize = max(max(in_size, out_size), drv_size);
if (ksize > sizeof(stack_kdata)) {
kdata = kmalloc(ksize, GFP_KERNEL);
if (ksize > in_size)
memset(kdata + in_size, 0, ksize - in_size);
KUNIT_EXPECT_GE(test, ksize(ctl->cache), ctl->len);
unsigned int in_size, out_size, drv_size, ksize;
ksize = max(max(in_size, out_size), drv_size);
if (ksize <= sizeof(stack_kdata)) {
kdata = kmalloc(ksize, GFP_KERNEL);
if (ksize > in_size)
memset(kdata + in_size, 0, ksize - in_size);
WARN_ON(dev + 1 > (struct drm_device *) (container + ksize(container)));
add_kmalloc_block(ksize(p));
add_kmalloc_block(ksize(p));
remove_kmalloc_block(ksize(ptr));
int ksize;
ksize = sizeof(*cdev->kmap.addr) * bsize;
cdev->kmap.addr = kvzalloc(ksize, GFP_KERNEL);
} else if (size > ksize(q_vector)) {
ksize = kend - offset;
rc = __nvmem_reg_write(nvmem, offset, val, ksize);
rc = __nvmem_reg_read(nvmem, offset, val, ksize);
offset += ksize;
val += ksize;
ksize = kend - offset;
memset(val, keepout->value, ksize);
val += ksize;
offset += ksize;
ksize = end - offset;
return __nvmem_reg_write(nvmem, offset, val, ksize);
return __nvmem_reg_read(nvmem, offset, val, ksize);
unsigned int kend, ksize;
size_t ksize(const void *objp);
static inline int copy_safe_from_sockptr(void *dst, size_t ksize,
if (optlen < ksize)
if (copy_from_sockptr(dst, optval, ksize))
static inline int copy_struct_from_sockptr(void *dst, size_t ksize,
size_t size = min(ksize, usize);
size_t rest = max(ksize, usize) - size;
return copy_struct_from_user(dst, ksize, src.user, size);
if (usize < ksize) {
} else if (usize > ksize) {
copy_struct_from_user(void *dst, size_t ksize, const void __user *src,
size_t size = min(ksize, usize);
size_t rest = max(ksize, usize) - size;
if (WARN_ON_ONCE(ksize > __builtin_object_size(dst, 1)))
if (usize < ksize) {
} else if (usize > ksize) {
size_t ksize, bool *ignored_trailing)
size_t size = min(ksize, usize);
size_t rest = max(ksize, usize) - size;
if (WARN_ON_ONCE(ksize > __builtin_object_size(src, 1)))
if (usize > ksize) {
*ignored_trailing = ksize < usize &&
u32 (*ksize)(u32 size);
alloc_bytes = max(ksize(orig), kmalloc_size_roundup(bytes));
actual_bytes_used = ksize(stream);
actual_bytes_used += ksize(frag_container);
actual_bytes_used += ksize(frag_container->fragment);
size_t ksize, usize;
ksize = size;
memcpy(expected, umem_src, ksize);
KUNIT_EXPECT_EQ_MSG(test, copy_struct_from_user(kmem, ksize, umem, usize), 0,
KUNIT_EXPECT_MEMEQ_MSG(test, kmem, expected, ksize,
ksize = size;
memset(expected + usize, 0x0, ksize - usize);
KUNIT_EXPECT_EQ_MSG(test, copy_struct_from_user(kmem, ksize, umem, usize), 0,
KUNIT_EXPECT_MEMEQ_MSG(test, kmem, expected, ksize,
ksize = size / 2;
KUNIT_EXPECT_EQ_MSG(test, copy_struct_from_user(kmem, ksize, umem, usize), -E2BIG,
ksize = size / 2;
memcpy(expected, umem_src, ksize);
KUNIT_EXPECT_EQ_MSG(test, clear_user(umem + ksize, usize - ksize), 0,
KUNIT_EXPECT_EQ_MSG(test, copy_struct_from_user(kmem, ksize, umem, usize), 0,
KUNIT_EXPECT_MEMEQ_MSG(test, kmem, expected, ksize,
real_size = ksize(ptr);
KUNIT_EXPECT_KASAN_FAIL(test, ksize(ptr));
KUNIT_EXPECT_EQ(test, ksize(buf), size); /* Precise size match after KFENCE alloc. */
KUNIT_EXPECT_GE(test, ksize(buf), size * 3);
KUNIT_EXPECT_GE(test, ksize(buf), size * 2);
return ksize(objp);
ks = ksize(mem);
EXPORT_SYMBOL(ksize);
memcpy(n, kasan_reset_tag(p), ksize(p));
*size = ksize(data);
if (ops->ksize)
return ops->ksize(desc->size);
.ksize = nft_rbtree_ksize,
if (req_size <= (ksize(*sfa) - next_offset))
new_acts_size = max(next_offset + req_size, ksize(*sfa) * 2);
int i, ksize;
ksize = parm->nkeys * sizeof(struct tc_pedit_key);
if (nla_len(pattr) < sizeof(*parm) + ksize) {
nparms->tcfp_keys = kmemdup(parm->keys, ksize, GFP_KERNEL);
copy_min_struct_from_user(void *const dst, const size_t ksize,
BUILD_BUG_ON(ksize <= 0);
BUILD_BUG_ON(ksize < ksize_min);
return copy_struct_from_user(dst, ksize, src, usize);
tomoyo_memory_used[TOMOYO_MEMORY_POLICY] -= ksize(ptr);
const size_t s = ksize(ptr);
KUNIT_ASSERT_GE_MSG(test, ksize(buf), priv->cal_blob->size, "Buffer to small");
KUNIT_ASSERT_GE_MSG(test, ksize(buf), priv->cal_blob->size, "Buffer to small");
long usize, ksize;
ksize = bpf_get_stack(ctx, raw_data + usize, max_len - usize, 0);
if (ksize < 0)
total_size = usize + ksize;