a_ctx
struct chcr_dev *dev = a_ctx(tfm)->dev;
struct uld_ctx *u_ctx = ULD_CTX(a_ctx(tfm));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
error = chcr_aead_dma_map(&ULD_CTX(a_ctx(tfm))->lldi.pdev->dev, req,
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct chcr_context *ctx = a_ctx(tfm);
create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, size,
struct chcr_context *ctx = a_ctx(tfm);
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct chcr_context *ctx = a_ctx(tfm);
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct adapter *adap = padap(a_ctx(tfm)->dev);
create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, 0,
struct chcr_context *ctx = a_ctx(tfm);
create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, size,
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
return chcr_device_init(a_ctx(tfm));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(authenc));
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(authenc));
struct chcr_context *ctx = a_ctx(tfm);
cdev = a_ctx(tfm)->dev;
struct chcr_context *ctx = a_ctx(tfm);
struct chcr_context *ctx = a_ctx(tfm);
struct sec_auth_ctx a_ctx;
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
return crypto_aead_setauthsize(a_ctx->fallback_aead_tfm, authsize);
static int sec_aead_fallback_setkey(struct sec_auth_ctx *a_ctx,
crypto_aead_clear_flags(a_ctx->fallback_aead_tfm, CRYPTO_TFM_REQ_MASK);
crypto_aead_set_flags(a_ctx->fallback_aead_tfm,
return crypto_aead_setkey(a_ctx->fallback_aead_tfm, key, keylen);
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
return sec_aead_fallback_setkey(a_ctx, tfm, key, keylen);
ctx->a_ctx.a_alg = a_alg;
return sec_aead_fallback_setkey(a_ctx, tfm, key, keylen);
ret = sec_aead_auth_set_key(&ctx->a_ctx, &keys);
ret = sec_aead_fallback_setkey(a_ctx, tfm, key, keylen);
struct sec_auth_ctx *auth_ctx = &ctx->a_ctx;
struct sec_auth_ctx *auth_ctx = &ctx->a_ctx;
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
a_ctx->hash_tfm = crypto_alloc_shash(hash_name, 0, 0);
if (IS_ERR(a_ctx->hash_tfm)) {
return PTR_ERR(a_ctx->hash_tfm);
a_ctx->fallback_aead_tfm = crypto_alloc_aead(aead_name, 0,
if (IS_ERR(a_ctx->fallback_aead_tfm)) {
crypto_free_shash(ctx->a_ctx.hash_tfm);
return PTR_ERR(a_ctx->fallback_aead_tfm);
crypto_free_aead(ctx->a_ctx.fallback_aead_tfm);
crypto_free_shash(ctx->a_ctx.hash_tfm);
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
a_ctx->fallback_aead_tfm = crypto_alloc_aead(aead_name, 0,
if (IS_ERR(a_ctx->fallback_aead_tfm)) {
return PTR_ERR(a_ctx->fallback_aead_tfm);
crypto_free_aead(ctx->a_ctx.fallback_aead_tfm);
if (unlikely(ctx->a_ctx.a_key_len & WORD_MASK))
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
subreq = aead_request_alloc(a_ctx->fallback_aead_tfm, GFP_KERNEL);
aead_request_set_tfm(subreq, a_ctx->fallback_aead_tfm);
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
a_ctx->a_key = dma_alloc_coherent(ctx->dev, SEC_MAX_AKEY_SIZE,
&a_ctx->a_key_dma, GFP_KERNEL);
if (!a_ctx->a_key)
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
memzero_explicit(a_ctx->a_key, SEC_MAX_AKEY_SIZE);
a_ctx->a_key, a_ctx->a_key_dma);
struct aa_sk_ctx *a_ctx = aa_sock(socka->sk);
if (rcu_access_pointer(a_ctx->label) != label) {
aa_put_label(rcu_dereference_protected(a_ctx->label, true));
rcu_assign_pointer(a_ctx->label, aa_get_label(label));
unix_connect_peers(a_ctx, b_ctx);