Lines Matching refs:AES_BLOCK_SIZE

54 	u8 b[AES_BLOCK_SIZE];
187 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_encrypt()
192 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ecb_encrypt()
207 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_decrypt()
212 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ecb_decrypt()
225 while ((blocks = (walk->nbytes / AES_BLOCK_SIZE))) { in cbc_encrypt_walk()
231 err = skcipher_walk_done(walk, walk->nbytes % AES_BLOCK_SIZE); in cbc_encrypt_walk()
255 while ((blocks = (walk->nbytes / AES_BLOCK_SIZE))) { in cbc_decrypt_walk()
261 err = skcipher_walk_done(walk, walk->nbytes % AES_BLOCK_SIZE); in cbc_decrypt_walk()
281 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_encrypt()
292 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_encrypt()
293 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_encrypt()
300 cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_encrypt()
308 if (req->cryptlen == AES_BLOCK_SIZE) in cts_cbc_encrypt()
319 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_encrypt()
339 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_decrypt()
350 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_decrypt()
351 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_decrypt()
358 cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_decrypt()
366 if (req->cryptlen == AES_BLOCK_SIZE) in cts_cbc_decrypt()
377 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_decrypt()
402 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ctr_encrypt()
408 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ctr_encrypt()
411 u8 __aligned(8) tail[AES_BLOCK_SIZE]; in ctr_encrypt()
459 int tail = req->cryptlen % AES_BLOCK_SIZE; in xts_encrypt()
465 if (req->cryptlen < AES_BLOCK_SIZE) in xts_encrypt()
472 AES_BLOCK_SIZE) - 2; in xts_encrypt()
481 xts_blocks * AES_BLOCK_SIZE, in xts_encrypt()
489 for (first = 1; walk.nbytes >= AES_BLOCK_SIZE; first = 0) { in xts_encrypt()
493 nbytes &= ~(AES_BLOCK_SIZE - 1); in xts_encrypt()
510 skcipher_request_set_crypt(req, src, dst, AES_BLOCK_SIZE + tail, in xts_encrypt()
531 int tail = req->cryptlen % AES_BLOCK_SIZE; in xts_decrypt()
537 if (req->cryptlen < AES_BLOCK_SIZE) in xts_decrypt()
544 AES_BLOCK_SIZE) - 2; in xts_decrypt()
553 xts_blocks * AES_BLOCK_SIZE, in xts_decrypt()
561 for (first = 1; walk.nbytes >= AES_BLOCK_SIZE; first = 0) { in xts_decrypt()
565 nbytes &= ~(AES_BLOCK_SIZE - 1); in xts_decrypt()
582 skcipher_request_set_crypt(req, src, dst, AES_BLOCK_SIZE + tail, in xts_decrypt()
603 .base.cra_blocksize = AES_BLOCK_SIZE,
617 .base.cra_blocksize = AES_BLOCK_SIZE,
623 .ivsize = AES_BLOCK_SIZE,
632 .base.cra_blocksize = AES_BLOCK_SIZE,
638 .ivsize = AES_BLOCK_SIZE,
639 .walksize = 2 * AES_BLOCK_SIZE,
654 .ivsize = AES_BLOCK_SIZE,
655 .chunksize = AES_BLOCK_SIZE,
669 .ivsize = AES_BLOCK_SIZE,
670 .chunksize = AES_BLOCK_SIZE,
679 .base.cra_blocksize = AES_BLOCK_SIZE,
685 .ivsize = AES_BLOCK_SIZE,
686 .walksize = 2 * AES_BLOCK_SIZE,