AES_XTS_BLOCKSIZE
u_int8_t block[AES_XTS_BLOCKSIZE];
for (i = 0; i < (len / AES_XTS_BLOCKSIZE); i++) {
bcopy(in, block, AES_XTS_BLOCKSIZE);
in += AES_XTS_BLOCKSIZE;
bcopy(block, out, AES_XTS_BLOCKSIZE);
out += AES_XTS_BLOCKSIZE;
u_int8_t tweak[AES_XTS_BLOCKSIZE];
for (j = 0; j < DEV_BSIZE; j += AES_XTS_BLOCKSIZE)
for (j = 0; j < DEV_BSIZE; j += AES_XTS_BLOCKSIZE)
for (j = 0; j < DEV_BSIZE; j += AES_XTS_BLOCKSIZE)
for (j = 0; j < DEV_BSIZE; j += AES_XTS_BLOCKSIZE)
for (j = 0; j < DEV_BSIZE; j += AES_XTS_BLOCKSIZE)
u_int8_t tweak[AES_XTS_BLOCKSIZE];
u_int8_t block[AES_XTS_BLOCKSIZE];
for (i = 0; i < AES_XTS_BLOCKSIZE; i++)
for (i = 0; i < AES_XTS_BLOCKSIZE; i++)
for (i = 0; i < AES_XTS_BLOCKSIZE; i++) {
if ((ulud->paddr & (AES_XTS_BLOCKSIZE - 1)) != 0 ||
(ulud->length % AES_XTS_BLOCKSIZE) != 0)
for (j = 0; j < DEV_BSIZE; j += AES_XTS_BLOCKSIZE)
u_int8_t block[AES_XTS_BLOCKSIZE];
for (i = 0; i < AES_XTS_BLOCKSIZE; i++)
for (i = 0; i < AES_XTS_BLOCKSIZE; i++)
for (i = 0; i < AES_XTS_BLOCKSIZE; i++) {
u_int8_t tweak[AES_XTS_BLOCKSIZE];
addr &= ~(AES_XTS_BLOCKSIZE - 1);
roundup(vmr->vmr_size, AES_XTS_BLOCKSIZE))) {
if (addr & (AES_XTS_BLOCKSIZE - 1)) {
size += (addr & (AES_XTS_BLOCKSIZE - 1));