/Linux-v5.4/drivers/crypto/ccp/ |
D | ccp-crypto-aes-xts.c | 171 rctx->cmd.u.xts.type = CCP_AES_TYPE_128; in ccp_aes_xts_crypt() 172 rctx->cmd.u.xts.action = (encrypt) ? CCP_AES_ACTION_ENCRYPT in ccp_aes_xts_crypt() 174 rctx->cmd.u.xts.unit_size = unit_size; in ccp_aes_xts_crypt() 175 rctx->cmd.u.xts.key = &ctx->u.aes.key_sg; in ccp_aes_xts_crypt() 176 rctx->cmd.u.xts.key_len = ctx->u.aes.key_len; in ccp_aes_xts_crypt() 177 rctx->cmd.u.xts.iv = &rctx->iv_sg; in ccp_aes_xts_crypt() 178 rctx->cmd.u.xts.iv_len = AES_BLOCK_SIZE; in ccp_aes_xts_crypt() 179 rctx->cmd.u.xts.src = req->src; in ccp_aes_xts_crypt() 180 rctx->cmd.u.xts.src_len = req->nbytes; in ccp_aes_xts_crypt() 181 rctx->cmd.u.xts.dst = req->dst; in ccp_aes_xts_crypt()
|
D | ccp-ops.c | 1075 struct ccp_xts_aes_engine *xts = &cmd->u.xts; in ccp_run_xts_aes_cmd() local 1085 switch (xts->unit_size) { in ccp_run_xts_aes_cmd() 1106 if (xts->key_len == AES_KEYSIZE_128) in ccp_run_xts_aes_cmd() 1108 else if (xts->key_len == AES_KEYSIZE_256) in ccp_run_xts_aes_cmd() 1113 if (!xts->final && (xts->src_len & (AES_BLOCK_SIZE - 1))) in ccp_run_xts_aes_cmd() 1116 if (xts->iv_len != AES_BLOCK_SIZE) in ccp_run_xts_aes_cmd() 1119 if (!xts->key || !xts->iv || !xts->src || !xts->dst) in ccp_run_xts_aes_cmd() 1132 op.u.xts.type = aestype; in ccp_run_xts_aes_cmd() 1133 op.u.xts.action = xts->action; in ccp_run_xts_aes_cmd() 1134 op.u.xts.unit_size = xts->unit_size; in ccp_run_xts_aes_cmd() [all …]
|
D | Makefile | 17 ccp-crypto-aes-xts.o \
|
D | ccp-dev-v3.c | 185 | (op->u.xts.action << REQ1_AES_ACTION_SHIFT) in ccp_perform_xts_aes() 186 | (op->u.xts.unit_size << REQ1_XTS_AES_SIZE_SHIFT) in ccp_perform_xts_aes()
|
D | ccp-dev-v5.c | 342 CCP_XTS_TYPE(&function) = op->u.xts.type; in ccp5_perform_xts_aes() 343 CCP_XTS_ENCRYPT(&function) = op->u.xts.action; in ccp5_perform_xts_aes() 344 CCP_XTS_SIZE(&function) = op->u.xts.unit_size; in ccp5_perform_xts_aes()
|
D | ccp-dev.h | 547 struct ccp_xts_aes_op xts; member
|
/Linux-v5.4/arch/x86/crypto/ |
D | serpent_avx2_glue.c | 79 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc_16way) } 82 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc_8way_avx) } 85 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc) } 127 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec_16way) } 130 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec_8way_avx) } 133 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec) }
|
D | camellia_aesni_avx2_glue.c | 81 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc_32way) } 84 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc_16way) } 87 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc) } 135 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec_32way) } 138 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec_16way) } 141 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec) }
|
D | cast6_avx_glue.c | 99 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(cast6_xts_enc_8way) } 102 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(cast6_xts_enc) } 138 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(cast6_xts_dec_8way) } 141 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(cast6_xts_dec) }
|
D | serpent_avx_glue.c | 131 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc_8way_avx) } 134 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc) } 170 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec_8way_avx) } 173 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec) }
|
D | twofish_avx_glue.c | 128 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(twofish_xts_enc_8way) } 131 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(twofish_xts_enc) } 173 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(twofish_xts_dec_8way) } 176 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(twofish_xts_dec) }
|
D | camellia_aesni_avx_glue.c | 97 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc_16way) } 100 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_enc) } 142 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec_16way) } 145 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(camellia_xts_dec) }
|
D | aesni-intel_glue.c | 584 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(aesni_xts_enc8) } 587 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(aesni_xts_enc) } 597 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(aesni_xts_dec8) } 600 .fn_u = { .xts = GLUE_XTS_FUNC_CAST(aesni_xts_dec) }
|
D | glue_helper.c | 243 gctx->funcs[i].fn_u.xts(ctx, dst, src, in __glue_xts_req_128bit()
|
/Linux-v5.4/Documentation/admin-guide/device-mapper/ |
D | dm-crypt.rst | 26 aes-xts-plain64 27 serpent-xts-plain64 41 capi:xts(aes)-plain64 46 capi:authenc(hmac(sha256),xts(aes))-random
|
D | dm-init.rst | 95 0 1048576 crypt aes-xts-plain64
|
/Linux-v5.4/Documentation/crypto/ |
D | api-samples.rst | 33 tfm = crypto_alloc_skcipher("xts(aes)", 0, 0); 35 pr_err("Error allocating xts(aes) handle: %ld\n", PTR_ERR(tfm));
|
/Linux-v5.4/arch/x86/include/asm/crypto/ |
D | glue_helper.h | 32 common_glue_xts_func_t xts; member
|
/Linux-v5.4/include/linux/ |
D | ccp.h | 652 struct ccp_xts_aes_engine xts; member
|
/Linux-v5.4/arch/s390/crypto/ |
D | aes_s390.c | 553 u8 xts[16]; in xts_aes_crypt() member 564 memset(pcc_param.xts, 0, sizeof(pcc_param.xts)); in xts_aes_crypt() 570 memcpy(xts_param.init, pcc_param.xts, 16); in xts_aes_crypt()
|
D | paes_s390.c | 444 u8 xts[16]; in xts_paes_crypt() member 461 memcpy(xts_param.init, pcc_param.xts, 16); in xts_paes_crypt()
|
/Linux-v5.4/crypto/ |
D | Makefile | 84 obj-$(CONFIG_CRYPTO_XTS) += xts.o
|
D | Kconfig | 433 XTS: IEEE1619/D16 narrow block cipher use with aes-xts-plain,
|
/Linux-v5.4/drivers/crypto/inside-secure/ |
D | safexcel_cipher.c | 49 unsigned int key_len, xts; member 465 switch (ctx->key_len >> ctx->xts) { in safexcel_context_control() 480 ctx->key_len >> ctx->xts); in safexcel_context_control() 2018 ctx->xts = 1; in safexcel_skcipher_aes_xts_cra_init()
|
/Linux-v5.4/Documentation/dev-tools/ |
D | coccinelle.rst | 381 /home/user/linux/crypto/xts.c:227:9-16: ERR_CAST can be used with alg 505 …* TODO [[view:/home/user/linux/crypto/xts.c::face=ovl-face1::linb=227::colb=9::cole=16][ERR_CAST c…
|