Symbol: a_ctx
drivers/crypto/chelsio/chcr_algo.c
214
struct chcr_dev *dev = a_ctx(tfm)->dev;
drivers/crypto/chelsio/chcr_algo.c
2282
struct uld_ctx *u_ctx = ULD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
2290
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
2305
error = chcr_aead_dma_map(&ULD_CTX(a_ctx(tfm))->lldi.pdev->dev, req,
drivers/crypto/chelsio/chcr_algo.c
2334
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
2352
struct chcr_context *ctx = a_ctx(tfm);
drivers/crypto/chelsio/chcr_algo.c
2478
create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, size,
drivers/crypto/chelsio/chcr_algo.c
2626
struct chcr_context *ctx = a_ctx(tfm);
drivers/crypto/chelsio/chcr_algo.c
2852
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
2878
struct chcr_context *ctx = a_ctx(tfm);
drivers/crypto/chelsio/chcr_algo.c
2953
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
2967
struct adapter *adap = padap(a_ctx(tfm)->dev);
drivers/crypto/chelsio/chcr_algo.c
3033
create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, 0,
drivers/crypto/chelsio/chcr_algo.c
3050
struct chcr_context *ctx = a_ctx(tfm);
drivers/crypto/chelsio/chcr_algo.c
3151
create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, size,
drivers/crypto/chelsio/chcr_algo.c
3165
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
3177
return chcr_device_init(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
3182
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
3190
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
3199
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
3237
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
3274
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
3298
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
drivers/crypto/chelsio/chcr_algo.c
3339
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
drivers/crypto/chelsio/chcr_algo.c
3369
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
drivers/crypto/chelsio/chcr_algo.c
3384
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
drivers/crypto/chelsio/chcr_algo.c
3405
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
drivers/crypto/chelsio/chcr_algo.c
3463
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(authenc));
drivers/crypto/chelsio/chcr_algo.c
3546
struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(authenc));
drivers/crypto/chelsio/chcr_algo.c
3610
struct chcr_context *ctx = a_ctx(tfm);
drivers/crypto/chelsio/chcr_algo.c
3615
cdev = a_ctx(tfm)->dev;
drivers/crypto/chelsio/chcr_algo.c
3660
struct chcr_context *ctx = a_ctx(tfm);
drivers/crypto/chelsio/chcr_algo.c
3688
struct chcr_context *ctx = a_ctx(tfm);
drivers/crypto/hisilicon/sec2/sec.h
196
struct sec_auth_ctx a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1381
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1383
return crypto_aead_setauthsize(a_ctx->fallback_aead_tfm, authsize);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1386
static int sec_aead_fallback_setkey(struct sec_auth_ctx *a_ctx,
drivers/crypto/hisilicon/sec2/sec_crypto.c
1390
crypto_aead_clear_flags(a_ctx->fallback_aead_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1391
crypto_aead_set_flags(a_ctx->fallback_aead_tfm,
drivers/crypto/hisilicon/sec2/sec_crypto.c
1393
return crypto_aead_setkey(a_ctx->fallback_aead_tfm, key, keylen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1403
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1409
return sec_aead_fallback_setkey(a_ctx, tfm, key, keylen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1411
ctx->a_ctx.a_alg = a_alg;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1423
return sec_aead_fallback_setkey(a_ctx, tfm, key, keylen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1438
ret = sec_aead_auth_set_key(&ctx->a_ctx, &keys);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1444
ret = sec_aead_fallback_setkey(a_ctx, tfm, key, keylen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1845
struct sec_auth_ctx *auth_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1900
struct sec_auth_ctx *auth_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
2141
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
2151
a_ctx->hash_tfm = crypto_alloc_shash(hash_name, 0, 0);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2152
if (IS_ERR(a_ctx->hash_tfm)) {
drivers/crypto/hisilicon/sec2/sec_crypto.c
2155
return PTR_ERR(a_ctx->hash_tfm);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2158
a_ctx->fallback_aead_tfm = crypto_alloc_aead(aead_name, 0,
drivers/crypto/hisilicon/sec2/sec_crypto.c
2160
if (IS_ERR(a_ctx->fallback_aead_tfm)) {
drivers/crypto/hisilicon/sec2/sec_crypto.c
2162
crypto_free_shash(ctx->a_ctx.hash_tfm);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2164
return PTR_ERR(a_ctx->fallback_aead_tfm);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2174
crypto_free_aead(ctx->a_ctx.fallback_aead_tfm);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2175
crypto_free_shash(ctx->a_ctx.hash_tfm);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2183
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
2193
a_ctx->fallback_aead_tfm = crypto_alloc_aead(aead_name, 0,
drivers/crypto/hisilicon/sec2/sec_crypto.c
2196
if (IS_ERR(a_ctx->fallback_aead_tfm)) {
drivers/crypto/hisilicon/sec2/sec_crypto.c
2199
return PTR_ERR(a_ctx->fallback_aead_tfm);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2209
crypto_free_aead(ctx->a_ctx.fallback_aead_tfm);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2482
if (unlikely(ctx->a_ctx.a_key_len & WORD_MASK))
drivers/crypto/hisilicon/sec2/sec_crypto.c
2533
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
2537
subreq = aead_request_alloc(a_ctx->fallback_aead_tfm, GFP_KERNEL);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2541
aead_request_set_tfm(subreq, a_ctx->fallback_aead_tfm);
drivers/crypto/hisilicon/sec2/sec_crypto.c
741
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
743
a_ctx->a_key = dma_alloc_coherent(ctx->dev, SEC_MAX_AKEY_SIZE,
drivers/crypto/hisilicon/sec2/sec_crypto.c
744
&a_ctx->a_key_dma, GFP_KERNEL);
drivers/crypto/hisilicon/sec2/sec_crypto.c
745
if (!a_ctx->a_key)
drivers/crypto/hisilicon/sec2/sec_crypto.c
753
struct sec_auth_ctx *a_ctx = &ctx->a_ctx;
drivers/crypto/hisilicon/sec2/sec_crypto.c
758
memzero_explicit(a_ctx->a_key, SEC_MAX_AKEY_SIZE);
drivers/crypto/hisilicon/sec2/sec_crypto.c
760
a_ctx->a_key, a_ctx->a_key_dma);
security/apparmor/lsm.c
1318
struct aa_sk_ctx *a_ctx = aa_sock(socka->sk);
security/apparmor/lsm.c
1324
if (rcu_access_pointer(a_ctx->label) != label) {
security/apparmor/lsm.c
1326
aa_put_label(rcu_dereference_protected(a_ctx->label, true));
security/apparmor/lsm.c
1327
rcu_assign_pointer(a_ctx->label, aa_get_label(label));
security/apparmor/lsm.c
1337
unix_connect_peers(a_ctx, b_ctx);