Lines Matching refs:AES_BLOCK_SIZE

70 	__be32 *n = (__be32 *)&maciv[AES_BLOCK_SIZE - 8];  in ccm_init_mac()
88 memcpy(maciv, req->iv, AES_BLOCK_SIZE - l); in ccm_init_mac()
101 memset(&req->iv[AES_BLOCK_SIZE - l], 0, l); in ccm_init_mac()
114 if (*macp > 0 && *macp < AES_BLOCK_SIZE) { in ccm_update_mac()
115 int added = min(abytes, AES_BLOCK_SIZE - *macp); in ccm_update_mac()
124 while (abytes >= AES_BLOCK_SIZE) { in ccm_update_mac()
126 crypto_xor(mac, in, AES_BLOCK_SIZE); in ccm_update_mac()
128 in += AES_BLOCK_SIZE; in ccm_update_mac()
129 abytes -= AES_BLOCK_SIZE; in ccm_update_mac()
183 u8 buf[AES_BLOCK_SIZE]; in ccm_crypt_fallback()
187 int blocks = walk->nbytes / AES_BLOCK_SIZE; in ccm_crypt_fallback()
188 u32 tail = walk->nbytes % AES_BLOCK_SIZE; in ccm_crypt_fallback()
199 u32 bsize = AES_BLOCK_SIZE; in ccm_crypt_fallback()
201 if (nbytes < AES_BLOCK_SIZE) in ccm_crypt_fallback()
204 crypto_inc(walk->iv, AES_BLOCK_SIZE); in ccm_crypt_fallback()
223 crypto_xor(mac, buf, AES_BLOCK_SIZE); in ccm_crypt_fallback()
233 u8 __aligned(8) mac[AES_BLOCK_SIZE]; in ccm_encrypt()
234 u8 buf[AES_BLOCK_SIZE]; in ccm_encrypt()
246 memcpy(buf, req->iv, AES_BLOCK_SIZE); in ccm_encrypt()
252 u32 tail = walk.nbytes % AES_BLOCK_SIZE; in ccm_encrypt()
291 u8 __aligned(8) mac[AES_BLOCK_SIZE]; in ccm_decrypt()
292 u8 buf[AES_BLOCK_SIZE]; in ccm_decrypt()
304 memcpy(buf, req->iv, AES_BLOCK_SIZE); in ccm_decrypt()
310 u32 tail = walk.nbytes % AES_BLOCK_SIZE; in ccm_decrypt()
356 .ivsize = AES_BLOCK_SIZE,
357 .chunksize = AES_BLOCK_SIZE,
358 .maxauthsize = AES_BLOCK_SIZE,