SHA256_BLOCK_SIZE
case SHA256_BLOCK_SIZE:
#define PHMAC_SHA256_PK_SIZE (SHA256_BLOCK_SIZE + 32)
case SHA256_BLOCK_SIZE:
.base.cra_blocksize = SHA256_BLOCK_SIZE,
partial = ctx.bytecount % SHA256_BLOCK_SIZE;
.base.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
u8 key[SHA256_BLOCK_SIZE];
rctx->block_size = SHA256_BLOCK_SIZE;
.cra_blocksize = SHA256_BLOCK_SIZE,
.halg.base.cra_blocksize = SHA256_BLOCK_SIZE,
ctx->block_size = SHA256_BLOCK_SIZE;
.halg.base.cra_blocksize = SHA256_BLOCK_SIZE,
ctx->block_size = SHA256_BLOCK_SIZE;
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
char partial_buffer[SHA256_BLOCK_SIZE];
char partial_buffer_out[SHA256_BLOCK_SIZE];
char key_buffer[SHA256_BLOCK_SIZE];
char pad_buffer[SHA256_BLOCK_SIZE + 32];
char partial_buffer[SHA256_BLOCK_SIZE];
char hmac_key[SHA256_BLOCK_SIZE];
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.blocksize = SHA256_BLOCK_SIZE,
.blocksize = SHA256_BLOCK_SIZE,
.block_size = SHA256_BLOCK_SIZE,
block_size = SHA256_BLOCK_SIZE;
#define MAX_HMAC_BLOCK_SIZE (SHA256_BLOCK_SIZE)
SHA256_BLOCK_SIZE, NS_BIT);
blocksize = SHA256_BLOCK_SIZE;
.blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
u8 ipad[SHA256_BLOCK_SIZE], opad[SHA256_BLOCK_SIZE];
if (keylen > SHA256_BLOCK_SIZE) {
memset(ipad + keylen, 0, SHA256_BLOCK_SIZE - keylen);
memcpy(opad, ipad, SHA256_BLOCK_SIZE);
for (i = 0; i < SHA256_BLOCK_SIZE; i++) {
memcpy(dest_ipad, ipad, SHA256_BLOCK_SIZE);
sg_init_one(&sg[0], ipad, SHA256_BLOCK_SIZE);
ahash_request_set_crypt(req, sg, dest_ipad, SHA256_BLOCK_SIZE);
sg_init_one(&sg[0], opad, SHA256_BLOCK_SIZE);
ahash_request_set_crypt(req, sg, dest_opad, SHA256_BLOCK_SIZE);
memcpy(rctx->data, ctx->ipad, SHA256_BLOCK_SIZE);
rctx->data_used = SHA256_BLOCK_SIZE;
rctx->len += SHA256_BLOCK_SIZE;
SHA256_BLOCK_SIZE, DMA_TO_DEVICE);
max_read = SHA256_BLOCK_SIZE - offset;
max_read = SHA256_BLOCK_SIZE;
SHA256_BLOCK_SIZE, wait_req);
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
u8 ipad[SHA256_BLOCK_SIZE] __aligned(sizeof(u32));
u8 data[SHA256_BLOCK_SIZE] __aligned(sizeof(u32));
u8 data[SHA256_BLOCK_SIZE] __aligned(sizeof(u32));
u8 data[SHA256_BLOCK_SIZE] __aligned(sizeof(u32));
req->block_sz = SHA256_BLOCK_SIZE;
.cra_blocksize = SHA256_BLOCK_SIZE,
req->block_sz = SHA256_BLOCK_SIZE;
req->len = SHA256_BLOCK_SIZE;
req->processed = SHA256_BLOCK_SIZE;
req->block_sz = SHA256_BLOCK_SIZE;
req->len = SHA256_BLOCK_SIZE;
req->processed = SHA256_BLOCK_SIZE;
req->block_sz = SHA256_BLOCK_SIZE;
.cra_blocksize = SHA256_BLOCK_SIZE,
rctx->blk_sz = SHA256_BLOCK_SIZE;
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
SHA256_DIGEST_SIZE, SHA256_BLOCK_SIZE);
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
} while (leftover >= SHA256_BLOCK_SIZE);
.cra_blocksize = SHA256_BLOCK_SIZE,
to_process = total & ~(SHA256_BLOCK_SIZE - 1);
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
d = SHA256_BLOCK_SIZE;
bs = SHA256_BLOCK_SIZE;
.cra_blocksize = SHA256_BLOCK_SIZE,
int blocks = len / SHA256_BLOCK_SIZE;
len -= blocks * SHA256_BLOCK_SIZE;
state->count += blocks * SHA256_BLOCK_SIZE;
.cra_blocksize = SHA256_BLOCK_SIZE,
.blocksize = SHA256_BLOCK_SIZE,
.blocksize = SHA256_BLOCK_SIZE,
#define QCE_SHA_MAX_BLOCKSIZE SHA256_BLOCK_SIZE
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
#define SAHARA_MAX_SHA_BLOCK_SIZE SHA256_BLOCK_SIZE
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
#define TALITOS_MAX_KEY_SIZE (AES_MAX_KEY_SIZE + SHA256_BLOCK_SIZE)
.cra_blocksize = SHA256_BLOCK_SIZE,
.cra_blocksize = SHA256_BLOCK_SIZE,
.block_size = SHA256_BLOCK_SIZE,
u8 buf[SHA256_BLOCK_SIZE];
u8 buf[SHA256_BLOCK_SIZE] __aligned(__alignof__(__be64));
data += (nblocks - rem) * SHA256_BLOCK_SIZE;
static_branch_likely(&have_ce) && len >= SHA256_BLOCK_SIZE &&
data += SHA256_BLOCK_SIZE;
MAX_BYTES / SHA256_BLOCK_SIZE);
data += unit * SHA256_BLOCK_SIZE;
nblocks * SHA256_BLOCK_SIZE);
data += SHA256_BLOCK_SIZE;
size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE;
if (partial + len >= SHA256_BLOCK_SIZE) {
size_t l = SHA256_BLOCK_SIZE - partial;
nblocks = len / SHA256_BLOCK_SIZE;
len %= SHA256_BLOCK_SIZE;
data += nblocks * SHA256_BLOCK_SIZE;
size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE;
if (partial > SHA256_BLOCK_SIZE - 8) {
memset(&ctx->buf[partial], 0, SHA256_BLOCK_SIZE - partial);
memset(&ctx->buf[partial], 0, SHA256_BLOCK_SIZE - 8 - partial);
*(__be64 *)&ctx->buf[SHA256_BLOCK_SIZE - 8] = cpu_to_be64(bitcount);
u8 b[SHA256_BLOCK_SIZE];
unsigned long w[SHA256_BLOCK_SIZE / sizeof(unsigned long)];
if (unlikely(raw_key_len > SHA256_BLOCK_SIZE)) {
__sha256_init(&ctx->sha_ctx, &key->istate, SHA256_BLOCK_SIZE);
ctx->ctx.sha_ctx.bytecount = SHA256_BLOCK_SIZE;
ctx->ctx.sha_ctx.bytecount = SHA256_BLOCK_SIZE;
SHA256_BLOCK_SIZE - digest_size);
*(__be32 *)&ctx->sha_ctx.buf[SHA256_BLOCK_SIZE - 4] =
cpu_to_be32(8 * (SHA256_BLOCK_SIZE + digest_size));
const size_t data_len = 4 * SHA256_BLOCK_SIZE;
for (size_t align = 0; align < SHA256_BLOCK_SIZE; align++) {
u8 salt[SHA256_BLOCK_SIZE];
if (static_branch_likely(&have_sha_ni) && len >= SHA256_BLOCK_SIZE &&