Lines Matching refs:iv_size

157 	unsigned int iv_size;  member
293 memset(iv, 0, cc->iv_size); in crypt_iv_plain_gen()
302 memset(iv, 0, cc->iv_size); in crypt_iv_plain64_gen()
311 memset(iv, 0, cc->iv_size); in crypt_iv_plain64be_gen()
313 *(__be64 *)&iv[cc->iv_size - sizeof(u64)] = cpu_to_be64(dmreq->iv_sector); in crypt_iv_plain64be_gen()
325 memset(iv, 0, cc->iv_size); in crypt_iv_essiv_gen()
364 memset(iv, 0, cc->iv_size - sizeof(u64)); /* rest is cleared below */ in crypt_iv_benbi_gen()
367 put_unaligned(val, (__be64 *)(iv + cc->iv_size - sizeof(u64))); in crypt_iv_benbi_gen()
375 memset(iv, 0, cc->iv_size); in crypt_iv_null_gen()
490 memcpy(iv, &md5state.hash, cc->iv_size); in crypt_iv_lmk_one()
508 memset(iv, 0, cc->iv_size); in crypt_iv_lmk_gen()
529 crypto_xor(dst + sg->offset, iv, cc->iv_size); in crypt_iv_lmk_post()
559 if (cc->key_size <= (cc->iv_size + TCW_WHITENING_SIZE)) { in crypt_iv_tcw_ctr()
570 tcw->iv_seed = kzalloc(cc->iv_size, GFP_KERNEL); in crypt_iv_tcw_ctr()
584 int key_offset = cc->key_size - cc->iv_size - TCW_WHITENING_SIZE; in crypt_iv_tcw_init()
586 memcpy(tcw->iv_seed, &cc->key[key_offset], cc->iv_size); in crypt_iv_tcw_init()
587 memcpy(tcw->whitening, &cc->key[key_offset + cc->iv_size], in crypt_iv_tcw_init()
597 memset(tcw->iv_seed, 0, cc->iv_size); in crypt_iv_tcw_wipe()
660 if (cc->iv_size > 8) in crypt_iv_tcw_gen()
662 cc->iv_size - 8); in crypt_iv_tcw_gen()
690 get_random_bytes(iv, cc->iv_size); in crypt_iv_random_gen()
702 if (crypto_skcipher_blocksize(any_tfm(cc)) != cc->iv_size) { in crypt_iv_eboiv_ctr()
724 memset(buf, 0, cc->iv_size); in crypt_iv_eboiv_gen()
727 sg_init_one(&src, page_address(ZERO_PAGE(0)), cc->iv_size); in crypt_iv_eboiv_gen()
728 sg_init_one(&dst, iv, cc->iv_size); in crypt_iv_eboiv_gen()
729 skcipher_request_set_crypt(req, &src, &dst, cc->iv_size, buf); in crypt_iv_eboiv_gen()
926 return iv_of_dmreq(cc, dmreq) + cc->iv_size; in org_iv_of_dmreq()
932 u8 *ptr = iv_of_dmreq(cc, dmreq) + cc->iv_size + cc->iv_size; in org_sector_of_dmreq()
939 u8 *ptr = iv_of_dmreq(cc, dmreq) + cc->iv_size + in org_tag_of_dmreq()
940 cc->iv_size + sizeof(uint64_t); in org_tag_of_dmreq()
972 BUG_ON(cc->integrity_iv_size && cc->integrity_iv_size != cc->iv_size); in crypt_convert_block_aead()
1001 sg_set_buf(&dmreq->sg_in[1], org_iv, cc->iv_size); in crypt_convert_block_aead()
1007 sg_set_buf(&dmreq->sg_out[1], org_iv, cc->iv_size); in crypt_convert_block_aead()
1014 memcpy(org_iv, tag_iv, cc->iv_size); in crypt_convert_block_aead()
1021 memcpy(tag_iv, org_iv, cc->iv_size); in crypt_convert_block_aead()
1024 memcpy(iv, org_iv, cc->iv_size); in crypt_convert_block_aead()
1027 aead_request_set_ad(req, sizeof(uint64_t) + cc->iv_size); in crypt_convert_block_aead()
1111 memcpy(iv, org_iv, cc->iv_size); in crypt_convert_block_skcipher()
2164 cc->iv_size = crypto_aead_ivsize(any_tfm_aead(cc)); in crypt_ctr_ivmode()
2166 cc->iv_size = crypto_skcipher_ivsize(any_tfm(cc)); in crypt_ctr_ivmode()
2168 if (cc->iv_size) in crypt_ctr_ivmode()
2170 cc->iv_size = max(cc->iv_size, in crypt_ctr_ivmode()
2209 cc->key_extra_size = cc->iv_size + TCW_WHITENING_SIZE; in crypt_ctr_ivmode()
2213 cc->integrity_iv_size = cc->iv_size; in crypt_ctr_ivmode()
2327 cc->iv_size = crypto_aead_ivsize(any_tfm_aead(cc)); in crypt_ctr_cipher_new()
2329 cc->iv_size = crypto_skcipher_ivsize(any_tfm(cc)); in crypt_ctr_cipher_new()
2633 iv_size_padding + cc->iv_size + in crypt_ctr()
2634 cc->iv_size + in crypt_ctr()