ksize
memcpy(&ksize, p, sizeof(ksize));
p += ksize;
u_int32_t ksize;
ksize = GETBLEAF(h, i)->ksize;
p += ksize;
u_int32_t ksize;
"size %03d pgno %03d", bi->ksize, bi->pgno);
else if (bi->ksize)
" {%.*s}", (int)bi->ksize, bi->bytes);
} else if (bl->ksize)
(int)bl->ksize, bl->bytes);
memcpy(&pgno, bl->bytes + bl->ksize,
memcpy(&sz, bl->bytes + bl->ksize +
(int)bl->dsize, bl->bytes + bl->ksize);
nksize = NBINTERNAL(bi->ksize);
__ovfl_delete(t, bl->bytes + bl->ksize) == RET_ERROR)
nbytes = NBINTERNAL(bi->ksize);
nbytes = NBINTERNAL(bl->ksize);
a.size = tbl->ksize;
b.size = bl->ksize;
WR_BINTERNAL(dest, nksize ? nksize : bl->ksize,
memmove(dest, bl->bytes, nksize ? nksize : bl->ksize);
nbytes = NBINTERNAL(bl->ksize);
WR_BINTERNAL(dest, bl->ksize, r->pgno, 0);
memmove(dest, bl->bytes, bl->ksize);
nbytes = NBINTERNAL(bi->ksize);
nbytes = NBINTERNAL(bi->ksize);
nbytes = NBINTERNAL(bi->ksize);
key->size = bl->ksize;
if (__ovfl_get(t, bl->bytes + bl->ksize,
memmove(rdata->data, bl->bytes + bl->ksize, bl->dsize);
data->data = bl->bytes + bl->ksize;
k2.size = bl->ksize;
k2.size = bi->ksize;
if (bl->ksize > rkey->size) {
malloc(bl->ksize) : realloc(rkey->data, bl->ksize));
rkey->size = bl->ksize;
memmove(rkey->data, bl->bytes, bl->ksize);
key->size = bl->ksize;
u_int32_t ksize; /* key size */
u_int32_t ksize; /* size of key */
#define NBLEAF(p) NBLEAFDBT((p)->ksize, (p)->dsize)
#define NBLEAFDBT(ksize, dsize) \
(ksize) + (dsize))
int32_t ksize;
ksize = size;
while ((ksize > 0) && (BIGKEYLEN(pagep))) {
if (ksize < KEY_OFF(pagep, 0) ||
ksize -= BIGKEYLEN(pagep);
assert(ksize >= 0);
if (ksize != 0) {
#define IMPLEMENT_ARIA_CFBR(ksize, cbits) \
IMPLEMENT_CFBR(aria, aria, EVP_ARIA_KEY, ks, ksize, cbits, 16, 0)
int ksize : 2;
#define DECLARE_AES_EVP(ksize, lmode, umode) \
static EVP_CIPHER *_hidden_aes_##ksize##_##lmode = NULL; \
static const EVP_CIPHER *padlock_aes_##ksize##_##lmode(void) \
if (_hidden_aes_##ksize##_##lmode == NULL \
&& ((_hidden_aes_##ksize##_##lmode = EVP_CIPHER_meth_new(NID_aes_##ksize##_##lmode, \
AES_KEY_SIZE_##ksize)) \
|| !EVP_CIPHER_meth_set_iv_length(_hidden_aes_##ksize##_##lmode, \
|| !EVP_CIPHER_meth_set_flags(_hidden_aes_##ksize##_##lmode, \
|| !EVP_CIPHER_meth_set_init(_hidden_aes_##ksize##_##lmode, \
|| !EVP_CIPHER_meth_set_do_cipher(_hidden_aes_##ksize##_##lmode, \
|| !EVP_CIPHER_meth_set_impl_ctx_size(_hidden_aes_##ksize##_##lmode, \
|| !EVP_CIPHER_meth_set_set_asn1_params(_hidden_aes_##ksize##_##lmode, \
|| !EVP_CIPHER_meth_set_get_asn1_params(_hidden_aes_##ksize##_##lmode, \
EVP_CIPHER_meth_free(_hidden_aes_##ksize##_##lmode); \
_hidden_aes_##ksize##_##lmode = NULL; \
return _hidden_aes_##ksize##_##lmode; \
cdata->cword.b.ksize = (key_len - 128) / 64;
"size %03d pgno %03d", bi->ksize, bi->pgno);
else if (bi->ksize)
" {%.*s}", (int)bi->ksize, bi->bytes);
else if (bl->ksize)
bl->ksize, bl->bytes);
*(pgno_t *)(bl->bytes + bl->ksize),
*(u_int32_t *)(bl->bytes + bl->ksize +
(int)bl->dsize, bl->bytes + bl->ksize);
nksize = NBINTERNAL(bi->ksize);
__ovfl_delete(t, bl->bytes + bl->ksize) == RET_ERROR)
nbytes = NBINTERNAL(bi->ksize);
nbytes = NBINTERNAL(bl->ksize);
a.size = tbl->ksize;
b.size = bl->ksize;
WR_BINTERNAL(dest, nksize ? nksize : bl->ksize,
memmove(dest, bl->bytes, nksize ? nksize : bl->ksize);
nbytes = NBINTERNAL(bl->ksize);
WR_BINTERNAL(dest, bl->ksize, r->pgno, 0);
memmove(dest, bl->bytes, bl->ksize);
nbytes = NBINTERNAL(bi->ksize);
nbytes = NBINTERNAL(bi->ksize);
nbytes = NBINTERNAL(bi->ksize);
if (__ovfl_get(t, bl->bytes + bl->ksize,
memmove(rdata->data, bl->bytes + bl->ksize, bl->dsize);
data->data = bl->bytes + bl->ksize;
k2.size = bl->ksize;
k2.size = bi->ksize;
if (bl->ksize > rkey->size) {
p = realloc(rkey->data, bl->ksize);
rkey->size = bl->ksize;
memmove(rkey->data, bl->bytes, bl->ksize);
key->size = bl->ksize;
key->size = bl->ksize;
u_int32_t ksize; /* key size */
u_int32_t ksize; /* size of key */
#define NBLEAF(p) NBLEAFDBT((p)->ksize, (p)->dsize)
#define NBLEAFDBT(ksize, dsize) \
(ksize) + (dsize))
int ksize;
ksize = size;
ksize -= bytes;
if (bytes != ksize || memcmp(p + bp[ndx], kkey, bytes)) {
unsigned long ksize = kva_layout.km_high - kva_layout.km_low;
return sysctl_handle_long(oidp, &ksize, 0, req);
unsigned long ksize = vm_max_kernel_address - KERNBASE;
return (sysctl_handle_long(oidp, &ksize, 0, req));
unsigned long ksize = VM_MAX_KERNEL_ADDRESS - VM_MIN_KERNEL_ADDRESS;
return sysctl_handle_long(oidp, &ksize, 0, req);
osize = ksize(ptr);
unsigned long ksize;
ksize = VM_MAX_KERNEL_ADDRESS - KERNBASE;
return (sysctl_handle_long(oidp, &ksize, 0, req));
size_t ksize; /* given buffer size */
memset(sd->kbuf, 0, sd->ksize);
if (sd->ktotal + sd->ksize < sd->valsize)
sd->kavail = sd->ksize;
sdata.ksize = sizeof(xbuf);
sdata.ksize = size;
if ((error = sooptcopyin(sopt, sdata.kbuf, sdata.ksize,
unsigned long ksize = VM_MAX_KERNEL_ADDRESS - VM_MIN_KERNEL_ADDRESS;
return sysctl_handle_long(oidp, &ksize, 0, req);
size_t ksize, zsize, size;
ksize = sizeof(struct uma_keg) +
ksize = roundup(ksize, UMA_SUPER_ALIGN);
size = (zsize * 2) + ksize;
args.size = ksize;