Lines Matching refs:sector_size

177 	unsigned short sector_size;  member
430 if (cc->sector_size != (1 << SECTOR_SHIFT)) { in crypt_iv_lmk_ctr()
588 if (cc->sector_size != (1 << SECTOR_SHIFT)) { in crypt_iv_tcw_ctr()
767 *(__le64 *)buf = cpu_to_le64(dmreq->iv_sector * cc->sector_size); in crypt_iv_eboiv_gen()
969 *(__le64 *)es = cpu_to_le64(dmreq->iv_sector * cc->sector_size); in crypt_iv_elephant()
995 memcpy(data_offset, data2 + sg2->offset, cc->sector_size); in crypt_iv_elephant()
1000 diffuser_disk_to_cpu((u32 *)data_offset, cc->sector_size / sizeof(u32)); in crypt_iv_elephant()
1001 diffuser_b_decrypt((u32 *)data_offset, cc->sector_size / sizeof(u32)); in crypt_iv_elephant()
1002 diffuser_a_decrypt((u32 *)data_offset, cc->sector_size / sizeof(u32)); in crypt_iv_elephant()
1003 diffuser_cpu_to_disk((__le32 *)data_offset, cc->sector_size / sizeof(u32)); in crypt_iv_elephant()
1006 for (i = 0; i < (cc->sector_size / 32); i++) in crypt_iv_elephant()
1010 diffuser_disk_to_cpu((u32 *)data_offset, cc->sector_size / sizeof(u32)); in crypt_iv_elephant()
1011 diffuser_a_encrypt((u32 *)data_offset, cc->sector_size / sizeof(u32)); in crypt_iv_elephant()
1012 diffuser_b_encrypt((u32 *)data_offset, cc->sector_size / sizeof(u32)); in crypt_iv_elephant()
1013 diffuser_cpu_to_disk((__le32 *)data_offset, cc->sector_size / sizeof(u32)); in crypt_iv_elephant()
1191 if (1 << bi->interval_exp != cc->sector_size) { in crypt_integrity_ctr()
1312 if (unlikely(bv_in.bv_len & (cc->sector_size - 1))) in crypt_convert_block_aead()
1339 sg_set_page(&dmreq->sg_in[2], bv_in.bv_page, cc->sector_size, bv_in.bv_offset); in crypt_convert_block_aead()
1345 sg_set_page(&dmreq->sg_out[2], bv_out.bv_page, cc->sector_size, bv_out.bv_offset); in crypt_convert_block_aead()
1367 cc->sector_size, iv); in crypt_convert_block_aead()
1374 cc->sector_size + cc->integrity_tag_size, iv); in crypt_convert_block_aead()
1390 bio_advance_iter(ctx->bio_in, &ctx->iter_in, cc->sector_size); in crypt_convert_block_aead()
1391 bio_advance_iter(ctx->bio_out, &ctx->iter_out, cc->sector_size); in crypt_convert_block_aead()
1410 if (unlikely(bv_in.bv_len & (cc->sector_size - 1))) in crypt_convert_block_skcipher()
1433 sg_set_page(sg_in, bv_in.bv_page, cc->sector_size, bv_in.bv_offset); in crypt_convert_block_skcipher()
1436 sg_set_page(sg_out, bv_out.bv_page, cc->sector_size, bv_out.bv_offset); in crypt_convert_block_skcipher()
1457 skcipher_request_set_crypt(req, sg_in, sg_out, cc->sector_size, iv); in crypt_convert_block_skcipher()
1467 bio_advance_iter(ctx->bio_in, &ctx->iter_in, cc->sector_size); in crypt_convert_block_skcipher()
1468 bio_advance_iter(ctx->bio_out, &ctx->iter_out, cc->sector_size); in crypt_convert_block_skcipher()
1563 unsigned int sector_step = cc->sector_size >> SECTOR_SHIFT; in crypt_convert()
3153 } else if (sscanf(opt_string, "sector_size:%hu%c", &cc->sector_size, &dummy) == 1) { in crypt_ctr_optional()
3154 if (cc->sector_size < (1 << SECTOR_SHIFT) || in crypt_ctr_optional()
3155 cc->sector_size > 4096 || in crypt_ctr_optional()
3156 (cc->sector_size & (cc->sector_size - 1))) { in crypt_ctr_optional()
3160 if (ti->len & ((cc->sector_size >> SECTOR_SHIFT) - 1)) { in crypt_ctr_optional()
3164 cc->sector_shift = __ffs(cc->sector_size) - SECTOR_SHIFT; in crypt_ctr_optional()
3222 cc->sector_size = (1 << SECTOR_SHIFT); in crypt_ctr()
3304 (tmpll & ((cc->sector_size >> SECTOR_SHIFT) - 1))) { in crypt_ctr()
3439 if (unlikely((bio->bi_iter.bi_sector & ((cc->sector_size >> SECTOR_SHIFT) - 1)) != 0)) in crypt_map()
3442 if (unlikely(bio->bi_iter.bi_size & (cc->sector_size - 1))) in crypt_map()
3518 num_feature_args += cc->sector_size != (1 << SECTOR_SHIFT); in crypt_status()
3536 if (cc->sector_size != (1 << SECTOR_SHIFT)) in crypt_status()
3537 DMEMIT(" sector_size:%d", cc->sector_size); in crypt_status()
3559 if (cc->sector_size != (1 << SECTOR_SHIFT)) in crypt_status()
3560 DMEMIT(",sector_size=%d", cc->sector_size); in crypt_status()
3665 max_t(unsigned int, limits->logical_block_size, cc->sector_size); in crypt_io_hints()
3667 max_t(unsigned int, limits->physical_block_size, cc->sector_size); in crypt_io_hints()
3668 limits->io_min = max_t(unsigned int, limits->io_min, cc->sector_size); in crypt_io_hints()