Lines Matching refs:AES_BLOCK_SIZE

64 	__be32 *n = (__be32 *)&maciv[AES_BLOCK_SIZE - 8];  in ccm_init_mac()
82 memcpy(maciv, req->iv, AES_BLOCK_SIZE - l); in ccm_init_mac()
95 memset(&req->iv[AES_BLOCK_SIZE - l], 0, l); in ccm_init_mac()
108 if (*macp > 0 && *macp < AES_BLOCK_SIZE) { in ccm_update_mac()
109 int added = min(abytes, AES_BLOCK_SIZE - *macp); in ccm_update_mac()
118 while (abytes >= AES_BLOCK_SIZE) { in ccm_update_mac()
120 crypto_xor(mac, in, AES_BLOCK_SIZE); in ccm_update_mac()
122 in += AES_BLOCK_SIZE; in ccm_update_mac()
123 abytes -= AES_BLOCK_SIZE; in ccm_update_mac()
177 u8 buf[AES_BLOCK_SIZE]; in ccm_crypt_fallback()
181 int blocks = walk->nbytes / AES_BLOCK_SIZE; in ccm_crypt_fallback()
182 u32 tail = walk->nbytes % AES_BLOCK_SIZE; in ccm_crypt_fallback()
193 u32 bsize = AES_BLOCK_SIZE; in ccm_crypt_fallback()
195 if (nbytes < AES_BLOCK_SIZE) in ccm_crypt_fallback()
198 crypto_inc(walk->iv, AES_BLOCK_SIZE); in ccm_crypt_fallback()
217 crypto_xor(mac, buf, AES_BLOCK_SIZE); in ccm_crypt_fallback()
227 u8 __aligned(8) mac[AES_BLOCK_SIZE]; in ccm_encrypt()
228 u8 buf[AES_BLOCK_SIZE]; in ccm_encrypt()
240 memcpy(buf, req->iv, AES_BLOCK_SIZE); in ccm_encrypt()
246 u32 tail = walk.nbytes % AES_BLOCK_SIZE; in ccm_encrypt()
285 u8 __aligned(8) mac[AES_BLOCK_SIZE]; in ccm_decrypt()
286 u8 buf[AES_BLOCK_SIZE]; in ccm_decrypt()
298 memcpy(buf, req->iv, AES_BLOCK_SIZE); in ccm_decrypt()
304 u32 tail = walk.nbytes % AES_BLOCK_SIZE; in ccm_decrypt()
350 .ivsize = AES_BLOCK_SIZE,
351 .chunksize = AES_BLOCK_SIZE,
352 .maxauthsize = AES_BLOCK_SIZE,