Lines Matching refs:AES_BLOCK_SIZE

102 	u8 dg[AES_BLOCK_SIZE];
142 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_encrypt()
147 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ecb_encrypt()
162 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_decrypt()
167 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ecb_decrypt()
182 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in cbc_encrypt()
187 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in cbc_encrypt()
202 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in cbc_decrypt()
207 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in cbc_decrypt()
222 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ctr_encrypt()
227 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ctr_encrypt()
230 u8 __aligned(8) tail[AES_BLOCK_SIZE]; in ctr_encrypt()
272 for (first = 1; (blocks = (walk.nbytes / AES_BLOCK_SIZE)); first = 0) { in xts_encrypt()
278 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in xts_encrypt()
294 for (first = 1; (blocks = (walk.nbytes / AES_BLOCK_SIZE)); first = 0) { in xts_decrypt()
300 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in xts_decrypt()
312 .cra_blocksize = AES_BLOCK_SIZE,
327 .cra_blocksize = AES_BLOCK_SIZE,
333 .ivsize = AES_BLOCK_SIZE,
349 .ivsize = AES_BLOCK_SIZE,
350 .chunksize = AES_BLOCK_SIZE,
365 .ivsize = AES_BLOCK_SIZE,
366 .chunksize = AES_BLOCK_SIZE,
376 .cra_blocksize = AES_BLOCK_SIZE,
382 .ivsize = AES_BLOCK_SIZE,
425 aes_ecb_encrypt(ctx->consts, (u8[AES_BLOCK_SIZE]){}, rk, rounds, 1); in cmac_setkey()
437 static u8 const ks[3][AES_BLOCK_SIZE] = { in xcbc_setkey()
438 { [0 ... AES_BLOCK_SIZE - 1] = 0x1 }, in xcbc_setkey()
439 { [0 ... AES_BLOCK_SIZE - 1] = 0x2 }, in xcbc_setkey()
440 { [0 ... AES_BLOCK_SIZE - 1] = 0x3 }, in xcbc_setkey()
446 u8 key[AES_BLOCK_SIZE]; in xcbc_setkey()
465 memset(ctx->dg, 0, AES_BLOCK_SIZE); in mac_init()
486 crypto_xor(dg, in, AES_BLOCK_SIZE); in mac_do_update()
487 in += AES_BLOCK_SIZE; in mac_do_update()
504 if ((ctx->len % AES_BLOCK_SIZE) == 0 && in mac_update()
505 (ctx->len + len) > AES_BLOCK_SIZE) { in mac_update()
507 int blocks = len / AES_BLOCK_SIZE; in mac_update()
509 len %= AES_BLOCK_SIZE; in mac_update()
514 p += blocks * AES_BLOCK_SIZE; in mac_update()
517 ctx->len = AES_BLOCK_SIZE; in mac_update()
523 l = min(len, AES_BLOCK_SIZE - ctx->len); in mac_update()
525 if (l <= AES_BLOCK_SIZE) { in mac_update()
543 memcpy(out, ctx->dg, AES_BLOCK_SIZE); in cbcmac_final()
554 if (ctx->len != AES_BLOCK_SIZE) { in cmac_final()
556 consts += AES_BLOCK_SIZE; in cmac_final()
561 memcpy(out, ctx->dg, AES_BLOCK_SIZE); in cmac_final()
570 .base.cra_blocksize = AES_BLOCK_SIZE,
572 2 * AES_BLOCK_SIZE,
575 .digestsize = AES_BLOCK_SIZE,
585 .base.cra_blocksize = AES_BLOCK_SIZE,
587 2 * AES_BLOCK_SIZE,
590 .digestsize = AES_BLOCK_SIZE,
604 .digestsize = AES_BLOCK_SIZE,