overhead
len -= (*mech->overhead)(app_data, data_prot, len);
int (*overhead)(void *, int, int);
int (*overhead)(void *, int, int);
len -= (*mech->overhead)(app_data, data_prot, len);
overhead = 7 + ctx->mech_used->length;
if (ohlen+overhead < req_output_size)
*max_input_size = (req_output_size - ohlen - overhead) & (~(OM_uint32)7);
int overhead;
size_t overhead, int send)
max_early_data += overhead;
size_t overhead = ((block_size == 1) ? 0 : 1) /* padding length byte */
if (overhead > *reclen)
good = constant_time_ge_s(*reclen, overhead + padding_length);
const size_t overhead = 1 /* padding length byte */ + mac_size;
if (overhead > *reclen)
good = constant_time_ge_s(*reclen, padding_length + overhead);
size_t overhead, int send)
max_early_data += overhead;
size_t len, frag_off, overhead, used_len;
overhead = s->rlayer.wrlmethod->get_max_record_overhead(s->rlayer.wrl);
used_len = BIO_wpending(s->wbio) + overhead;
if (s->d1->mtu > overhead + DTLS1_HM_HEADER_LENGTH) {
curr_mtu = s->d1->mtu - overhead;
return (((caddr_t)cp - sizeof(union overhead)));
union overhead *op;
union overhead ov1;
mem = __crt_malloc(size + align + offset + sizeof(union overhead));
x = roundup2((uintptr_t)mem + sizeof(union overhead), align);
ov1.ov_index = x - (uintptr_t)mem + sizeof(union overhead);
union overhead *op;
op = (union overhead *)pagepool_start;
op->ov_next = (union overhead *)((caddr_t)op + sz);
op = (union overhead *)((caddr_t)op + sz);
union overhead *op, op1;
union overhead *op;
union overhead *ov_next; /* when free */
static union overhead *nextf[NBUCKETS];
uint32_t overhead, overhead_per_group, ngdb;
overhead = le32toh(fs->e2fs->e2fs_first_dblock) +
overhead += ngroups * (1 /* superblock */ + ngdb);
sbp->f_blocks = fs->e2fs_bcount - overhead;
OM_uint32 overhead;
overhead = 5 + 2 + krb5_mech_oid.length;
overhead += 8 + 8 + ec->ec_msgblocklen;
overhead += ec->ec_checksumlen;
overhead = 16 + ec->ec_blocklen;
overhead += ec->ec_msgblocklen - 1;
overhead += 16;
overhead += ec->ec_checksumlen;
overhead = 16 + ec->ec_checksumlen;
*max_input_size = req_output_size - overhead;
uint32_t overhead, rate, errrate;
overhead = ieee80211_compute_duration(ic->ic_rt,
res = (overhead + (nbits / rate)) *
+ priv->overhead ) * hinfo->run.fifo_queues * \
cfg->overhead = priv->overhead;
priv->overhead = 0;
priv->overhead = 8+4+12; /* Ethernet */
priv->overhead = 10; /* HDLC */
if (cfg->overhead == -1)
priv->overhead = 0;
else if (cfg->overhead > 0 &&
cfg->overhead < MAX_OHSIZE)
priv->overhead = cfg->overhead;
((uint64_t) m->m_pkthdr.len + priv->overhead ) *
hinfo->ber_p[priv->overhead + m->m_pkthdr.len]) {
u_int32_t overhead;
u_int32_t overhead;
uint32_t p_length, overhead;
overhead = SCTP_MIN_OVERHEAD;
overhead = SCTP_MIN_V4_OVERHEAD;
overhead += sizeof(struct sctp_asconf_chunk);
overhead += sctp_get_auth_chunk_len(stcb->asoc.peer_hmac_id);
if (stcb->asoc.smallest_mtu <= overhead) {
if ((SCTP_BUF_LEN(m_asconf) + p_length > stcb->asoc.smallest_mtu - overhead) ||
uint32_t frag_point, overhead;
overhead = SCTP_MIN_OVERHEAD;
overhead = SCTP_MIN_V4_OVERHEAD;
overhead += sizeof(struct sctp_idata_chunk);
overhead += sctp_get_auth_chunk_len(asoc->peer_hmac_id);
overhead += sizeof(struct sctp_data_chunk);
overhead += sctp_get_auth_chunk_len(asoc->peer_hmac_id);
KASSERT(overhead % 4 == 0,
("overhead (%u) not a multiple of 4", overhead));
overhead += (asoc->smallest_mtu % 4);
KASSERT(asoc->smallest_mtu > overhead,
asoc->smallest_mtu, overhead));
frag_point = asoc->smallest_mtu - overhead;
uint32_t overhead;
uint32_t overhead;
overhead = SCTP_MIN_OVERHEAD;
overhead = SCTP_MIN_V4_OVERHEAD;
overhead += sctp_get_auth_chunk_len(asoc->peer_hmac_id);
overhead += sctp_get_auth_chunk_len(asoc->peer_hmac_id);
KASSERT(overhead % 4 == 0,
("overhead (%u) not a multiple of 4", overhead));
if (((uint32_t)chk->send_size + overhead) > mtu) {
if (((uint32_t)chk->send_size + overhead) > mtu) {
sctp_log_rwnd(uint8_t from, uint32_t peers_rwnd, uint32_t snd_size, uint32_t overhead)
sctp_clog.x.rwnd.overhead = overhead;
sctp_log_rwnd_set(uint8_t from, uint32_t peers_rwnd, uint32_t flight_size, uint32_t overhead, uint32_t a_rwndval)
sctp_clog.x.rwnd.overhead = overhead;
int64_t overhead;
overhead = 0;
overhead += tsc2 - tsc1;
overhead /= count;
tsc_freq = (tsc2 - tsc1 - overhead) * 10;
le64enc(&hdr.overhead, sec);
uint64_t overhead;
size_t blocks, fs_nindir, overhead;
overhead = howmany(file_len -
switch (overhead) {
indir_level = overhead = 0;
overhead += blocks + 1;
indir_level, (uintmax_t)overhead);
int sp, overhead, maxoverhead;
overhead = ccp_MTUOverhead(&dl->physical->link.ccp);
if (maxoverhead < overhead)
maxoverhead = overhead;
overhead = ccp_MTUOverhead(&bundle->ncp.mp.link.ccp);
if (maxoverhead < overhead)
maxoverhead = overhead;