CAMELLIA_BLOCK_SIZE
round_down(nbytes, CAMELLIA_BLOCK_SIZE), key);
err = skcipher_walk_done(&walk, nbytes % CAMELLIA_BLOCK_SIZE);
round_down(nbytes, CAMELLIA_BLOCK_SIZE), key, walk.iv);
err = skcipher_walk_done(&walk, nbytes % CAMELLIA_BLOCK_SIZE);
round_down(nbytes, CAMELLIA_BLOCK_SIZE), key, walk.iv);
err = skcipher_walk_done(&walk, nbytes % CAMELLIA_BLOCK_SIZE);
.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.ivsize = CAMELLIA_BLOCK_SIZE,
ECB_WALK_START(req, CAMELLIA_BLOCK_SIZE, CAMELLIA_AESNI_PARALLEL_BLOCKS);
ECB_WALK_START(req, CAMELLIA_BLOCK_SIZE, CAMELLIA_AESNI_PARALLEL_BLOCKS);
CBC_WALK_START(req, CAMELLIA_BLOCK_SIZE, -1);
CBC_WALK_START(req, CAMELLIA_BLOCK_SIZE, CAMELLIA_AESNI_PARALLEL_BLOCKS);
.base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.ivsize = CAMELLIA_BLOCK_SIZE,
ECB_WALK_START(req, CAMELLIA_BLOCK_SIZE, CAMELLIA_AESNI_PARALLEL_BLOCKS);
ECB_WALK_START(req, CAMELLIA_BLOCK_SIZE, CAMELLIA_AESNI_PARALLEL_BLOCKS);
CBC_WALK_START(req, CAMELLIA_BLOCK_SIZE, -1);
CBC_WALK_START(req, CAMELLIA_BLOCK_SIZE, CAMELLIA_AESNI_PARALLEL_BLOCKS);
.base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.ivsize = CAMELLIA_BLOCK_SIZE,
u8 buf[CAMELLIA_BLOCK_SIZE];
crypto_xor(dst + CAMELLIA_BLOCK_SIZE, iv, CAMELLIA_BLOCK_SIZE);
ECB_WALK_START(req, CAMELLIA_BLOCK_SIZE, -1);
ECB_WALK_START(req, CAMELLIA_BLOCK_SIZE, -1);
CBC_WALK_START(req, CAMELLIA_BLOCK_SIZE, -1);
CBC_WALK_START(req, CAMELLIA_BLOCK_SIZE, -1);
.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.base.cra_blocksize = CAMELLIA_BLOCK_SIZE,
.ivsize = CAMELLIA_BLOCK_SIZE,
.cra_blocksize = CAMELLIA_BLOCK_SIZE,