GCM_AES_IV_SIZE
u8 iv[GCM_AES_IV_SIZE];
u8 iv[GCM_AES_IV_SIZE];
.ivsize = GCM_AES_IV_SIZE,
put_unaligned_be32(2, iv + GCM_AES_IV_SIZE);
put_unaligned_be32(2, iv + GCM_AES_IV_SIZE);
memcpy(iv, req->iv, GCM_AES_IV_SIZE);
memcpy(iv, req->iv, GCM_AES_IV_SIZE);
.ivsize = GCM_AES_IV_SIZE,
u8 iv[GCM_AES_IV_SIZE] = {};
u8 iv[GCM_AES_IV_SIZE] = {};
.ivsize = GCM_AES_IV_SIZE, \
if (crypto_aead_alg_ivsize(alg) != GCM_AES_IV_SIZE)
memcpy(pctx->iv, req->iv, GCM_AES_IV_SIZE);
memcpy(pctx->iv + GCM_AES_IV_SIZE, &counter, 4);
inst->alg.ivsize = GCM_AES_IV_SIZE;
scatterwalk_map_and_copy(iv + GCM_AES_IV_SIZE, req->src, 0, req->assoclen - 8, 0);
sg_set_buf(rctx->src, iv + GCM_AES_IV_SIZE, req->assoclen - 8);
sg_set_buf(rctx->dst, iv + GCM_AES_IV_SIZE, req->assoclen - 8);
if (crypto_aead_alg_ivsize(alg) != GCM_AES_IV_SIZE)
align + GCM_AES_IV_SIZE);
crypto4xx_memcpy_to_le32(iv, req->iv, GCM_AES_IV_SIZE);
.ivsize = GCM_AES_IV_SIZE,
if (likely(ivsize == GCM_AES_IV_SIZE)) {
.ivsize = GCM_AES_IV_SIZE,
memcpy(req_ctx->hw_ctx.J0 + GCM_AES_IV_SIZE, "\x00\x00\x00\x01", 4);
.ivsize = GCM_AES_IV_SIZE,
.ivsize = GCM_AES_IV_SIZE,
bool generic_gcm = (ivsize == GCM_AES_IV_SIZE);
FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1 | GCM_AES_IV_SIZE | last);
.ivsize = GCM_AES_IV_SIZE,
rctx->ivsize = GCM_AES_IV_SIZE - GCM_AES_SALT_SIZE;
rctx->ivsize = GCM_AES_IV_SIZE - GCM_AES_SALT_SIZE;
.ivsize = GCM_AES_IV_SIZE,
memcpy(rctx->iv, req->iv, GCM_AES_IV_SIZE);
rctx->iv[i + GCM_AES_IV_SIZE] = 0;
.ivsize = GCM_AES_IV_SIZE,
memcpy(ivptr, req->iv, GCM_AES_IV_SIZE);
.ivsize = GCM_AES_IV_SIZE,
.ivsize = GCM_AES_IV_SIZE,
.base.ivsize = GCM_AES_IV_SIZE,
.base.ivsize = GCM_AES_IV_SIZE,
if (iv_size != GCM_AES_IV_SIZE)
if (!iv || iv_size != GCM_AES_IV_SIZE)
GCM_AES_IV_SIZE, aad_dma_list,
memcpy(iv, req->iv, GCM_AES_IV_SIZE);
memcpy(iv, req->iv, GCM_AES_IV_SIZE);
.ivsize = GCM_AES_IV_SIZE,
memcpy(rctx->iv + GCM_AES_IV_SIZE, &counter, 4);
memcpy(rctx->iv, req->iv, GCM_AES_IV_SIZE);
memcpy(rctx->iv, req->iv, GCM_AES_IV_SIZE);
.ivsize = GCM_AES_IV_SIZE,
.base.ivsize = GCM_AES_IV_SIZE,
writel(GCM_AES_IV_SIZE, cryp->base + STARFIVE_AES_IVLEN);
memcpy(rctx->iv, req->iv, GCM_AES_IV_SIZE);
.ivsize = GCM_AES_IV_SIZE,
dmabuf = kmalloc(sizeof(*hwreq) + GCM_AES_IV_SIZE, GFP_KERNEL);
memcpy(dmabuf + sizeof(struct zynqmp_aead_hw_req), req->iv, GCM_AES_IV_SIZE);
GCM_AES_IV_SIZE,
GCM_AES_IV_SIZE, DMA_TO_DEVICE);
dma_unmap_single(dev, dma_addr_hw_req, sizeof(struct zynqmp_aead_hw_req) + GCM_AES_IV_SIZE,
memzero_explicit(dmabuf, sizeof(struct zynqmp_aead_hw_req) + GCM_AES_IV_SIZE);
GCM_AES_IV_SIZE;
sizeof(struct versal_in_params), req->iv, GCM_AES_IV_SIZE);
.ivsize = GCM_AES_IV_SIZE,
.ivsize = GCM_AES_IV_SIZE,
.ivsize = GCM_AES_IV_SIZE,
.ivsize = GCM_AES_IV_SIZE,
const u8 iv[GCM_AES_IV_SIZE], u8 *authtag);
int assoc_len, const u8 iv[GCM_AES_IV_SIZE],
const u8 iv[GCM_AES_IV_SIZE], u8 *authtag)
memcpy(ctr, iv, GCM_AES_IV_SIZE);
int assoc_len, const u8 iv[GCM_AES_IV_SIZE],
memcpy(ctr, iv, GCM_AES_IV_SIZE);
u8 iv[GCM_AES_IV_SIZE] __nonstring;