c_len
ec->loader = kmalloc(c_len(&excd),
c_len(&excd), useld);
(int)c_start(&excd), c_len(&excd),
u32 c_len;
copy_size = c_req->c_len + aead_req->assoclen;
copy_size = c_req->c_len;
sec_sqe->type2.clen_ivhlen |= cpu_to_le32(c_req->c_len);
sec_sqe3->c_len_ivin |= cpu_to_le32(c_req->c_len);
sec_sqe->type2.alen_ivllen = cpu_to_le32(c_req->c_len + aq->assoclen);
sqe3->a_len_key = cpu_to_le32(c_req->c_len + aq->assoclen);
sreq->c_req.c_len = sk_req->cryptlen;
if (unlikely(ctx->sec->qm.ver == QM_HW_V2 && !sreq->c_req.c_len))
sreq->c_req.c_len & (AES_BLOCK_SIZE - 1))) {
req->c_req.c_len = a_req->cryptlen - (req->c_req.encrypt ? 0 : sz);
copy_size = c_req->c_len;
u32 len, c_len;
c_len = offsetof(struct vsc_rom_cmd, data.dl_start.crc);
cmd->data.dl_start.crc = cpu_to_le32(vsc_sum_crc(cmd, c_len));
u32 c_len;
c_len = offsetof(struct vsc_fw_cmd, data.dl_set.payload[index]);
cmd->data.dl_set.payload[index] = cpu_to_le32(vsc_sum_crc(cmd, c_len));
c_len = offsetof(struct vsc_fw_cmd, data.dl_start.crc);
cmd->data.dl_start.crc = cpu_to_le32(vsc_sum_crc(cmd, c_len));
c_len = offsetof(struct vsc_fw_cmd, data.dl_start.crc);
cmd->data.boot.crc = cpu_to_le32(vsc_sum_crc(cmd, c_len));
unsigned int *c_len)
*c_len += sprintf(c_str + *c_len, "CHAP_C=0x%s", challenge_asciihex);
*c_len += 1;
if (!from_dnode && i >= ei.c_len)
if (back->c_len && back->len != back->c_len)
if (front->c_len && front->len != front->c_len)
unsigned int c_len)
blkaddr, c_len);
ei.c_len = c_len;
ei->c_len = 0;
unsigned int llen, unsigned int c_len);
unsigned int llen, unsigned int c_len) { }
unsigned int c_len;
unsigned int c_len;
c_len = f2fs_cluster_blocks_are_contiguous(dn, ofs_in_node);
if (!c_len)
fofs, blkaddr, cluster_size, c_len);
u32 c_len)
unsigned int i, b_len = ocfs2_clusters_to_blocks(sb, 1) * c_len;
u32 c_len);
else if (cmp < 0 || len < tmp->c_len)
else if (cmp < 0 || len < p->c_len)
const char *c_name, int c_len,
p->c_len = c_len;
p->c_len = len;
const char *c_name, int c_len,
if (ovl_cache_entry_find_link(c_name, c_len, &newp, &parent))
p = ovl_cache_entry_new(rdd, name, len, c_name, c_len, ino, d_type);
const char *c_name, int c_len,
p = ovl_cache_entry_find(rdd->root, c_name, c_len);
p = ovl_cache_entry_new(rdd, name, namelen, c_name, c_len,
int c_len = 0, ret;
int c_len;
c_len = ovl_casefold(rdd, name, namelen, &cf_name);
if (c_len <= 0) {
c_len = namelen;
ret = ovl_cache_entry_add_rb(rdd, name, namelen, c_name, c_len, ino, d_type);
ret = ovl_fill_lowest(rdd, name, namelen, c_name, c_len, offset, ino, d_type);
unsigned int c_len),
TP_ARGS(inode, pgofs, len, blkaddr, c_len),
__field(unsigned int, c_len)
__entry->c_len = c_len;
__entry->c_len)
int buflen, c_len;
c_len = *datalen - matchoff + strlen("v=");
buflen = sprintf(buffer, "%u", c_len);
int c_len;
c_len = qdisc_qlen(sch) - qdisc_qlen(q->l_queue);
if (qdisc_qlen(q->l_queue) && (!c_len || q->c_protection_credit <= 0)) {
if (c_len)
} else if (c_len) {