Symbol: GHASH_BLOCK_SIZE
arch/arm/crypto/ghash-ce-glue.c
119
u8 buf[GHASH_BLOCK_SIZE] = {};
arch/arm/crypto/ghash-ce-glue.c
144
if (keylen != GHASH_BLOCK_SIZE)
arch/arm/crypto/ghash-ce-glue.c
148
memcpy(&key->k, inkey, GHASH_BLOCK_SIZE);
arch/arm/crypto/ghash-ce-glue.c
181
.base.cra_blocksize = GHASH_BLOCK_SIZE,
arch/arm/crypto/ghash-ce-glue.c
250
int buf_added = min(count, GHASH_BLOCK_SIZE - *buf_count);
arch/arm/crypto/ghash-ce-glue.c
259
if (count >= GHASH_BLOCK_SIZE || *buf_count == GHASH_BLOCK_SIZE) {
arch/arm/crypto/ghash-ce-glue.c
260
int blocks = count / GHASH_BLOCK_SIZE;
arch/arm/crypto/ghash-ce-glue.c
265
src += blocks * GHASH_BLOCK_SIZE;
arch/arm/crypto/ghash-ce-glue.c
266
count %= GHASH_BLOCK_SIZE;
arch/arm/crypto/ghash-ce-glue.c
280
u8 buf[GHASH_BLOCK_SIZE];
arch/arm/crypto/ghash-ce-glue.c
302
memset(&buf[buf_count], 0, GHASH_BLOCK_SIZE - buf_count);
arch/arm/crypto/ghash-ce-glue.c
87
blocks = len / GHASH_BLOCK_SIZE;
arch/arm/crypto/ghash-ce-glue.c
89
return len - blocks * GHASH_BLOCK_SIZE;
arch/arm64/crypto/ghash-ce-glue.c
101
src += chunk * GHASH_BLOCK_SIZE;
arch/arm64/crypto/ghash-ce-glue.c
133
u8 buf[GHASH_BLOCK_SIZE] = {};
arch/arm64/crypto/ghash-ce-glue.c
159
if (keylen != GHASH_BLOCK_SIZE)
arch/arm64/crypto/ghash-ce-glue.c
163
memcpy(&key->k, inkey, GHASH_BLOCK_SIZE);
arch/arm64/crypto/ghash-ce-glue.c
174
.base.cra_blocksize = GHASH_BLOCK_SIZE,
arch/arm64/crypto/ghash-ce-glue.c
193
u8 key[GHASH_BLOCK_SIZE];
arch/arm64/crypto/ghash-ce-glue.c
204
memcpy(&ctx->ghash_key.k, key, GHASH_BLOCK_SIZE);
arch/arm64/crypto/ghash-ce-glue.c
230
int buf_added = min(count, GHASH_BLOCK_SIZE - *buf_count);
arch/arm64/crypto/ghash-ce-glue.c
239
if (count >= GHASH_BLOCK_SIZE || *buf_count == GHASH_BLOCK_SIZE) {
arch/arm64/crypto/ghash-ce-glue.c
240
int blocks = count / GHASH_BLOCK_SIZE;
arch/arm64/crypto/ghash-ce-glue.c
246
src += blocks * GHASH_BLOCK_SIZE;
arch/arm64/crypto/ghash-ce-glue.c
247
count %= GHASH_BLOCK_SIZE;
arch/arm64/crypto/ghash-ce-glue.c
261
u8 buf[GHASH_BLOCK_SIZE];
arch/arm64/crypto/ghash-ce-glue.c
277
memset(&buf[buf_count], 0, GHASH_BLOCK_SIZE - buf_count);
arch/arm64/crypto/ghash-ce-glue.c
83
#define MAX_BLOCKS (SZ_64K / GHASH_BLOCK_SIZE)
arch/arm64/crypto/ghash-ce-glue.c
92
blocks = len / GHASH_BLOCK_SIZE;
arch/arm64/crypto/ghash-ce-glue.c
93
len -= blocks * GHASH_BLOCK_SIZE;
arch/arm64/crypto/sm4-ce-gcm-glue.c
107
nblocks = n / GHASH_BLOCK_SIZE;
arch/arm64/crypto/sm4-ce-gcm-glue.c
111
p += nblocks * GHASH_BLOCK_SIZE;
arch/arm64/crypto/sm4-ce-gcm-glue.c
114
buflen = n % GHASH_BLOCK_SIZE;
arch/arm64/crypto/sm4-ce-gcm-glue.c
125
memset(&buffer[buflen], 0, GHASH_BLOCK_SIZE - buflen);
arch/arm64/crypto/sm4-ce-gcm-glue.c
75
u8 __aligned(8) buffer[GHASH_BLOCK_SIZE];
arch/arm64/crypto/sm4-ce-gcm-glue.c
90
if (n + buflen < GHASH_BLOCK_SIZE) {
arch/arm64/crypto/sm4-ce-gcm-glue.c
97
unsigned int l = GHASH_BLOCK_SIZE - buflen;
arch/powerpc/crypto/ghash.c
108
crypto_xor((u8 *)dctx->shash, src, GHASH_BLOCK_SIZE);
arch/powerpc/crypto/ghash.c
110
srclen -= GHASH_BLOCK_SIZE;
arch/powerpc/crypto/ghash.c
111
src += GHASH_BLOCK_SIZE;
arch/powerpc/crypto/ghash.c
134
u8 buf[GHASH_BLOCK_SIZE] = {};
arch/powerpc/crypto/ghash.c
156
.cra_blocksize = GHASH_BLOCK_SIZE,
arch/powerpc/crypto/ghash.c
56
if (keylen != GHASH_BLOCK_SIZE)
arch/powerpc/crypto/ghash.c
67
memcpy(&ctx->key, key, GHASH_BLOCK_SIZE);
arch/powerpc/crypto/ghash.c
80
gcm_ghash_p8(dctx->shash, ctx->htable, src, GHASH_BLOCK_SIZE);
arch/powerpc/crypto/ghash.c
85
crypto_xor((u8 *)dctx->shash, src, GHASH_BLOCK_SIZE);
arch/powerpc/crypto/ghash.c
94
int remain = srclen - round_down(srclen, GHASH_BLOCK_SIZE);
arch/riscv/crypto/ghash-riscv64-glue.c
100
riscv64_ghash_blocks(tctx, dctx, buf, GHASH_BLOCK_SIZE);
arch/riscv/crypto/ghash-riscv64-glue.c
116
.cra_blocksize = GHASH_BLOCK_SIZE,
arch/riscv/crypto/ghash-riscv64-glue.c
41
if (keylen != GHASH_BLOCK_SIZE)
arch/riscv/crypto/ghash-riscv64-glue.c
44
memcpy(&tctx->key, key, GHASH_BLOCK_SIZE);
arch/riscv/crypto/ghash-riscv64-glue.c
71
GHASH_BLOCK_SIZE);
arch/riscv/crypto/ghash-riscv64-glue.c
73
src += GHASH_BLOCK_SIZE;
arch/riscv/crypto/ghash-riscv64-glue.c
74
srclen -= GHASH_BLOCK_SIZE;
arch/riscv/crypto/ghash-riscv64-glue.c
86
round_down(srclen, GHASH_BLOCK_SIZE));
arch/riscv/crypto/ghash-riscv64-glue.c
87
return srclen - round_down(srclen, GHASH_BLOCK_SIZE);
arch/riscv/crypto/ghash-riscv64-glue.c
97
u8 buf[GHASH_BLOCK_SIZE] = {};
arch/s390/crypto/aes_s390.c
838
u8 j0[GHASH_BLOCK_SIZE];/* initial counter value */
arch/s390/crypto/aes_s390.c
933
.ivsize = GHASH_BLOCK_SIZE - sizeof(u32),
arch/s390/crypto/ghash_s390.c
100
memcpy(dctx->key, ctx->key, GHASH_BLOCK_SIZE);
arch/s390/crypto/ghash_s390.c
119
.cra_blocksize = GHASH_BLOCK_SIZE,
arch/s390/crypto/ghash_s390.c
21
u8 key[GHASH_BLOCK_SIZE];
arch/s390/crypto/ghash_s390.c
25
u8 icv[GHASH_BLOCK_SIZE];
arch/s390/crypto/ghash_s390.c
26
u8 key[GHASH_BLOCK_SIZE];
arch/s390/crypto/ghash_s390.c
35
memcpy(dctx->key, ctx->key, GHASH_BLOCK_SIZE);
arch/s390/crypto/ghash_s390.c
45
if (keylen != GHASH_BLOCK_SIZE)
arch/s390/crypto/ghash_s390.c
48
memcpy(ctx->key, key, GHASH_BLOCK_SIZE);
arch/s390/crypto/ghash_s390.c
59
n = srclen & ~(GHASH_BLOCK_SIZE - 1);
arch/s390/crypto/ghash_s390.c
68
u8 buf[GHASH_BLOCK_SIZE] = {};
arch/s390/crypto/ghash_s390.c
71
cpacf_kimd(CPACF_KIMD_GHASH, dctx, buf, GHASH_BLOCK_SIZE);
arch/s390/crypto/ghash_s390.c
82
memcpy(dst, dctx->icv, GHASH_BLOCK_SIZE);
arch/x86/crypto/ghash-clmulni-intel_glue.c
116
memcpy(dst, buf, GHASH_BLOCK_SIZE);
arch/x86/crypto/ghash-clmulni-intel_glue.c
133
.cra_blocksize = GHASH_BLOCK_SIZE,
arch/x86/crypto/ghash-clmulni-intel_glue.c
46
if (keylen != GHASH_BLOCK_SIZE)
crypto/ghash-generic.c
115
memcpy(dst, buf, GHASH_BLOCK_SIZE);
crypto/ghash-generic.c
139
.cra_blocksize = GHASH_BLOCK_SIZE,
crypto/ghash-generic.c
61
if (keylen != GHASH_BLOCK_SIZE)
crypto/ghash-generic.c
67
BUILD_BUG_ON(sizeof(k) != GHASH_BLOCK_SIZE);
crypto/ghash-generic.c
68
memcpy(&k, key, GHASH_BLOCK_SIZE); /* avoid violating alignment rules */
crypto/ghash-generic.c
70
memzero_explicit(&k, GHASH_BLOCK_SIZE);
crypto/ghash-generic.c
86
crypto_xor(dst, src, GHASH_BLOCK_SIZE);
crypto/ghash-generic.c
88
src += GHASH_BLOCK_SIZE;
crypto/ghash-generic.c
89
srclen -= GHASH_BLOCK_SIZE;
crypto/ghash-generic.c
90
} while (srclen >= GHASH_BLOCK_SIZE);
drivers/crypto/inside-secure/safexcel_cipher.c
2559
ctx->state_sz = GHASH_BLOCK_SIZE;
include/crypto/ghash.h
21
u8 buffer[GHASH_BLOCK_SIZE];
lib/crypto/aesgcm.c
67
crypto_xor((u8 *)ghash, src, min(len, GHASH_BLOCK_SIZE));
lib/crypto/aesgcm.c
70
src += GHASH_BLOCK_SIZE;
lib/crypto/aesgcm.c
71
len -= GHASH_BLOCK_SIZE;