Lines Matching refs:AES_BLOCK_SIZE
57 u8 rk[13 * (8 * AES_BLOCK_SIZE) + 32];
59 } __aligned(AES_BLOCK_SIZE);
108 while (walk.nbytes >= AES_BLOCK_SIZE) { in __ecb_crypt()
109 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in __ecb_crypt()
113 walk.stride / AES_BLOCK_SIZE); in __ecb_crypt()
120 walk.nbytes - blocks * AES_BLOCK_SIZE); in __ecb_crypt()
168 while (walk.nbytes >= AES_BLOCK_SIZE) { in cbc_encrypt()
169 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in cbc_encrypt()
177 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in cbc_encrypt()
191 while (walk.nbytes >= AES_BLOCK_SIZE) { in cbc_decrypt()
192 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in cbc_decrypt()
196 walk.stride / AES_BLOCK_SIZE); in cbc_decrypt()
204 walk.nbytes - blocks * AES_BLOCK_SIZE); in cbc_decrypt()
234 u8 buf[AES_BLOCK_SIZE]; in ctr_encrypt()
240 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in ctr_encrypt()
241 u8 *final = (walk.total % AES_BLOCK_SIZE) ? buf : NULL; in ctr_encrypt()
245 walk.stride / AES_BLOCK_SIZE); in ctr_encrypt()
255 u8 *dst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; in ctr_encrypt()
256 u8 *src = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; in ctr_encrypt()
259 walk.total % AES_BLOCK_SIZE); in ctr_encrypt()
265 walk.nbytes - blocks * AES_BLOCK_SIZE); in ctr_encrypt()
324 int tail = req->cryptlen % (8 * AES_BLOCK_SIZE); in __xts_crypt()
333 if (req->cryptlen < AES_BLOCK_SIZE) in __xts_crypt()
337 if (unlikely(tail > 0 && tail < AES_BLOCK_SIZE)) { in __xts_crypt()
339 AES_BLOCK_SIZE) - 2; in __xts_crypt()
346 xts_blocks * AES_BLOCK_SIZE, in __xts_crypt()
357 while (walk.nbytes >= AES_BLOCK_SIZE) { in __xts_crypt()
358 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in __xts_crypt()
360 if (walk.nbytes < walk.total || walk.nbytes % AES_BLOCK_SIZE) in __xts_crypt()
362 walk.stride / AES_BLOCK_SIZE); in __xts_crypt()
379 out += blocks * AES_BLOCK_SIZE; in __xts_crypt()
380 in += blocks * AES_BLOCK_SIZE; in __xts_crypt()
381 nbytes -= blocks * AES_BLOCK_SIZE; in __xts_crypt()
399 skcipher_request_set_crypt(req, src, dst, AES_BLOCK_SIZE + tail, in __xts_crypt()
437 .base.cra_blocksize = AES_BLOCK_SIZE,
444 .walksize = 8 * AES_BLOCK_SIZE,
452 .base.cra_blocksize = AES_BLOCK_SIZE,
459 .walksize = 8 * AES_BLOCK_SIZE,
460 .ivsize = AES_BLOCK_SIZE,
475 .chunksize = AES_BLOCK_SIZE,
476 .walksize = 8 * AES_BLOCK_SIZE,
477 .ivsize = AES_BLOCK_SIZE,
491 .chunksize = AES_BLOCK_SIZE,
492 .walksize = 8 * AES_BLOCK_SIZE,
493 .ivsize = AES_BLOCK_SIZE,
501 .base.cra_blocksize = AES_BLOCK_SIZE,
508 .walksize = 8 * AES_BLOCK_SIZE,
509 .ivsize = AES_BLOCK_SIZE,