/Linux-v5.10/block/ |
D | blk-crypto-internal.h | 34 return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req), in bio_crypt_ctx_back_mergeable() 42 bio->bi_iter.bi_size, req->crypt_ctx); in bio_crypt_ctx_front_mergeable() 48 return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req), in bio_crypt_ctx_merge_rq() 49 next->crypt_ctx); in bio_crypt_ctx_merge_rq() 54 rq->crypt_ctx = NULL; in blk_crypto_rq_set_defaults() 60 return rq->crypt_ctx; in blk_crypto_rq_is_encrypted() 117 memcpy(rq->crypt_ctx->bc_dun, bio->bi_crypt_context->bc_dun, in bio_crypt_do_front_merge() 118 sizeof(rq->crypt_ctx->bc_dun)); in bio_crypt_do_front_merge()
|
D | blk-crypto.c | 186 return bio_crypt_ctx_compatible(rq->crypt_ctx, bio->bi_crypt_context); in bio_crypt_rq_ctx_compatible() 221 return blk_ksm_get_slot_for_key(rq->q->ksm, rq->crypt_ctx->bc_key, in __blk_crypto_init_request() 237 mempool_free(rq->crypt_ctx, bio_crypt_ctx_pool); in __blk_crypto_free_request() 297 if (!rq->crypt_ctx) { in __blk_crypto_rq_bio_prep() 298 rq->crypt_ctx = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in __blk_crypto_rq_bio_prep() 299 if (!rq->crypt_ctx) in __blk_crypto_rq_bio_prep() 302 *rq->crypt_ctx = *bio->bi_crypt_context; in __blk_crypto_rq_bio_prep()
|
D | blk-crypto-fallback.c | 40 struct bio_crypt_ctx crypt_ctx; member 378 struct bio_crypt_ctx *bc = &f_ctx->crypt_ctx; in blk_crypto_fallback_decrypt_bio() 512 f_ctx->crypt_ctx = *bc; in blk_crypto_fallback_bio_prep()
|
/Linux-v5.10/arch/x86/crypto/ |
D | twofish_avx_glue.c | 60 struct twofish_ctx crypt_ctx; member 74 err = __twofish_setkey(&ctx->crypt_ctx, key, keylen / 2); in xts_twofish_setkey() 203 &ctx->tweak_ctx, &ctx->crypt_ctx, false); in xts_encrypt() 212 &ctx->tweak_ctx, &ctx->crypt_ctx, true); in xts_decrypt()
|
D | cast6_avx_glue.c | 169 struct cast6_ctx crypt_ctx; member 183 err = __cast6_setkey(&ctx->crypt_ctx, key, keylen / 2); in xts_cast6_setkey() 197 &ctx->tweak_ctx, &ctx->crypt_ctx, false); in xts_encrypt() 206 &ctx->tweak_ctx, &ctx->crypt_ctx, true); in xts_decrypt()
|
D | glue_helper.c | 268 void *crypt_ctx, bool decrypt) in glue_xts_req_128bit() argument 309 nbytes = __glue_xts_req_128bit(gctx, crypt_ctx, &walk); in glue_xts_req_128bit() 337 __glue_xts_req_128bit(gctx, crypt_ctx, &walk)); in glue_xts_req_128bit() 352 __glue_xts_req_128bit(gctx, crypt_ctx, &walk)); in glue_xts_req_128bit()
|
D | serpent_avx_glue.c | 90 err = __serpent_setkey(&ctx->crypt_ctx, key, keylen / 2); in xts_serpent_setkey() 209 &ctx->crypt_ctx, false); in xts_encrypt() 219 &ctx->crypt_ctx, true); in xts_decrypt()
|
D | camellia_aesni_avx_glue.c | 186 err = __camellia_setkey(&ctx->crypt_ctx, key, keylen / 2); in xts_camellia_setkey() 201 &ctx->tweak_ctx, &ctx->crypt_ctx, false); in xts_encrypt() 210 &ctx->tweak_ctx, &ctx->crypt_ctx, true); in xts_decrypt()
|
D | serpent_avx2_glue.c | 167 &ctx->crypt_ctx, false); in xts_encrypt() 177 &ctx->crypt_ctx, true); in xts_decrypt()
|
D | camellia_aesni_avx2_glue.c | 179 &ctx->tweak_ctx, &ctx->crypt_ctx, false); in xts_encrypt() 188 &ctx->tweak_ctx, &ctx->crypt_ctx, true); in xts_decrypt()
|
/Linux-v5.10/arch/x86/include/asm/crypto/ |
D | serpent-avx.h | 15 struct serpent_ctx crypt_ctx; member
|
D | glue_helper.h | 112 void *crypt_ctx, bool decrypt);
|
D | camellia.h | 24 struct camellia_ctx crypt_ctx; member
|
/Linux-v5.10/drivers/scsi/ufs/ |
D | ufshcd-crypto.h | 22 lrbp->data_unit_num = rq->crypt_ctx->bc_dun[0]; in ufshcd_prepare_lrbp_crypto()
|
/Linux-v5.10/Documentation/block/ |
D | inline-encryption.rst | 127 which sets the request's ``crypt_ctx`` to a copy of the bio's 129 bio is merged to the front of the request, which updates the ``crypt_ctx`` of
|
/Linux-v5.10/include/linux/ |
D | blkdev.h | 225 struct bio_crypt_ctx *crypt_ctx; member
|