/Linux-v5.4/drivers/crypto/nx/ |
D | nx-aes-xcbc.c | 63 struct nx_sg *in_sg, *out_sg; in nx_xcbc_empty() local 81 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) keys, &len, in nx_xcbc_empty() 93 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_xcbc_empty() 108 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) keys[1], &len, in nx_xcbc_empty() 121 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_xcbc_empty() 172 struct nx_sg *in_sg; in nx_xcbc_update() local 195 in_sg = nx_ctx->in_sg; in nx_xcbc_update() 230 in_sg = nx_build_sg_list(nx_ctx->in_sg, in nx_xcbc_update() 241 in_sg = nx_build_sg_list(in_sg, in nx_xcbc_update() 251 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * in nx_xcbc_update() [all …]
|
D | nx-sha256.c | 104 struct nx_sg *in_sg = nx_ctx->in_sg; in nx_sha256_update() local 108 in_sg = nx_build_sg_list(in_sg, in nx_sha256_update() 117 used_sgs = in_sg - nx_ctx->in_sg; in nx_sha256_update() 132 in_sg = nx_build_sg_list(in_sg, (u8 *) data, in nx_sha256_update() 135 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_sha256_update() 181 struct nx_sg *in_sg, *out_sg; in nx_sha256_final() local 210 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) sctx->buf, in nx_sha256_final() 226 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_sha256_final()
|
D | nx-sha512.c | 104 struct nx_sg *in_sg = nx_ctx->in_sg; in nx_sha512_update() local 108 in_sg = nx_build_sg_list(in_sg, in nx_sha512_update() 116 used_sgs = in_sg - nx_ctx->in_sg; in nx_sha512_update() 131 in_sg = nx_build_sg_list(in_sg, (u8 *) data, in nx_sha512_update() 134 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_sha512_update() 184 struct nx_sg *in_sg, *out_sg; in nx_sha512_final() local 219 in_sg = nx_build_sg_list(nx_ctx->in_sg, sctx->buf, &len, in nx_sha512_final() 231 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_sha512_final()
|
D | nx-aes-gcm.c | 107 struct nx_sg *nx_sg = nx_ctx->in_sg; in nx_gca() 137 nx_sg = nx_walk_and_build(nx_ctx->in_sg, max_sg_len, in nx_gca() 145 nx_ctx->op_aead.inlen = (nx_ctx->in_sg - nx_sg) in nx_gca() 205 nx_sg = nx_walk_and_build(nx_ctx->in_sg, max_sg_len, in gmac() 213 nx_ctx->op.inlen = (nx_ctx->in_sg - nx_sg) in gmac() 251 struct nx_sg *in_sg, *out_sg; in gcm_empty() local 271 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) desc->info, in gcm_empty() 284 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in gcm_empty()
|
D | nx.c | 268 struct nx_sg *nx_insg = nx_ctx->in_sg; in nx_build_sg_lists() 293 nx_ctx->op.inlen = trim_sg_list(nx_ctx->in_sg, nx_insg, delta, nbytes); in nx_build_sg_lists() 313 nx_ctx->op.in = __pa(nx_ctx->in_sg); in nx_ctx_init() 321 nx_ctx->op_aead.in = __pa(nx_ctx->in_sg); in nx_ctx_init() 675 nx_ctx->in_sg = (struct nx_sg *)((u8 *)nx_ctx->csbcpb + NX_PAGE_SIZE); in nx_crypto_ctx_init() 676 nx_ctx->out_sg = (struct nx_sg *)((u8 *)nx_ctx->in_sg + NX_PAGE_SIZE); in nx_crypto_ctx_init() 751 nx_ctx->in_sg = NULL; in nx_crypto_ctx_exit()
|
D | nx-aes-ccm.c | 164 struct nx_sg *nx_insg = nx_ctx->in_sg; in generate_pat() 252 nx_ctx->op.inlen = (nx_ctx->in_sg - nx_insg) * in generate_pat() 285 nx_insg = nx_walk_and_build(nx_ctx->in_sg, in generate_pat() 299 nx_ctx->op_aead.inlen = (nx_ctx->in_sg - nx_insg) * in generate_pat()
|
D | nx.h | 126 struct nx_sg *in_sg; /* aligned pointer into kmem to an sg list */ member
|
/Linux-v5.4/lib/ |
D | sg_split.c | 80 struct scatterlist *in_sg, *out_sg; in sg_split_phys() local 84 in_sg = split->in_sg0; in sg_split_phys() 87 *out_sg = *in_sg; in sg_split_phys() 96 in_sg = sg_next(in_sg); in sg_split_phys() 106 struct scatterlist *in_sg, *out_sg; in sg_split_mapped() local 110 in_sg = split->in_sg0; in sg_split_mapped() 113 sg_dma_address(out_sg) = sg_dma_address(in_sg); in sg_split_mapped() 114 sg_dma_len(out_sg) = sg_dma_len(in_sg); in sg_split_mapped() 119 in_sg = sg_next(in_sg); in sg_split_mapped()
|
/Linux-v5.4/drivers/crypto/ |
D | omap-des.c | 151 struct scatterlist *in_sg; member 377 struct scatterlist *in_sg, struct scatterlist *out_sg, in omap_des_crypt_dma() argument 387 scatterwalk_start(&dd->in_walk, dd->in_sg); in omap_des_crypt_dma() 396 dma_sync_sg_for_device(dd->dev, dd->in_sg, in_sg_len, DMA_TO_DEVICE); in omap_des_crypt_dma() 415 tx_in = dmaengine_prep_slave_sg(dd->dma_lch_in, in_sg, in_sg_len, in omap_des_crypt_dma() 466 err = dma_map_sg(dd->dev, dd->in_sg, dd->in_sg_len, in omap_des_crypt_dma_start() 481 err = omap_des_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, in omap_des_crypt_dma_start() 484 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); in omap_des_crypt_dma_start() 543 dd->in_sg = req->src; in omap_des_prepare_req() 551 ret = omap_crypto_align_sg(&dd->in_sg, dd->total, DES_BLOCK_SIZE, in omap_des_prepare_req() [all …]
|
D | omap-aes.c | 268 struct scatterlist *in_sg, in omap_aes_crypt_dma() argument 277 scatterwalk_start(&dd->in_walk, dd->in_sg); in omap_aes_crypt_dma() 286 dma_sync_sg_for_device(dd->dev, dd->in_sg, in_sg_len, DMA_TO_DEVICE); in omap_aes_crypt_dma() 305 tx_in = dmaengine_prep_slave_sg(dd->dma_lch_in, in_sg, in_sg_len, in omap_aes_crypt_dma() 357 err = dma_map_sg(dd->dev, dd->in_sg, dd->in_sg_len, in omap_aes_crypt_dma_start() 372 err = omap_aes_crypt_dma(dd, dd->in_sg, dd->out_sg, dd->in_sg_len, in omap_aes_crypt_dma_start() 375 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); in omap_aes_crypt_dma_start() 432 dd->in_sg = req->src; in omap_aes_prepare_req() 440 ret = omap_crypto_align_sg(&dd->in_sg, dd->total, AES_BLOCK_SIZE, in omap_aes_prepare_req() 452 dd->in_sg_len = sg_nents_for_len(dd->in_sg, dd->total); in omap_aes_prepare_req() [all …]
|
D | sahara.c | 183 struct scatterlist *in_sg; member 221 struct scatterlist *in_sg; member 470 dev->nb_in_sg = sg_nents_for_len(dev->in_sg, dev->total); in sahara_hw_descriptor_create() 486 ret = dma_map_sg(dev->device, dev->in_sg, dev->nb_in_sg, in sahara_hw_descriptor_create() 501 sg = dev->in_sg; in sahara_hw_descriptor_create() 544 dma_unmap_sg(dev->device, dev->in_sg, dev->nb_in_sg, in sahara_hw_descriptor_create() 565 dev->in_sg = req->src; in sahara_aes_process() 594 dma_unmap_sg(dev->device, dev->in_sg, dev->nb_in_sg, in sahara_aes_process() 802 dev->in_sg = rctx->in_sg; in sahara_sha_hw_links_create() 804 dev->nb_in_sg = sg_nents_for_len(dev->in_sg, rctx->total); in sahara_sha_hw_links_create() [all …]
|
D | atmel-tdes.c | 112 struct scatterlist *in_sg; member 326 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); in atmel_tdes_crypt_pdc_stop() 513 in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) && in atmel_tdes_crypt_start() 514 IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size); in atmel_tdes_crypt_start() 519 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) in atmel_tdes_crypt_start() 525 count = min_t(size_t, dd->total, sg_dma_len(dd->in_sg)); in atmel_tdes_crypt_start() 528 err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); in atmel_tdes_crypt_start() 538 dma_unmap_sg(dd->dev, dd->in_sg, 1, in atmel_tdes_crypt_start() 543 addr_in = sg_dma_address(dd->in_sg); in atmel_tdes_crypt_start() 550 count = atmel_tdes_sg_copy(&dd->in_sg, &dd->in_offset, in atmel_tdes_crypt_start() [all …]
|
D | omap-aes-gcm.c | 33 dd->in_sg = NULL; in omap_aes_gcm_finish_req() 53 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); in omap_aes_gcm_done_task() 137 dd->in_sg = dd->in_sgl; in omap_aes_gcm_copy_buffers() 158 dd->in_sg_len = sg_nents_for_len(dd->in_sg, alen + clen); in omap_aes_gcm_copy_buffers()
|
D | omap-aes.h | 178 struct scatterlist *in_sg; member
|
/Linux-v5.4/drivers/crypto/stm32/ |
D | stm32-cryp.c | 107 #define _walked_in (cryp->in_walk.offset - cryp->in_sg->offset) 151 struct scatterlist *in_sg; member 317 ret = stm32_cryp_check_aligned(cryp->in_sg, cryp->total_in, in stm32_cryp_check_io_aligned() 366 sg_copy_buf(buf_in, cryp->in_sg, 0, cryp->total_in, 0); in stm32_cryp_copy_sgs() 369 cryp->in_sg = &cryp->in_sgl; in stm32_cryp_copy_sgs() 981 cryp->in_sg = req ? req->src : areq->src; in stm32_cryp_prepare_req() 985 cryp->in_sg_len = sg_nents_for_len(cryp->in_sg, cryp->total_in); in stm32_cryp_prepare_req() 1003 scatterwalk_start(&cryp->in_walk, cryp->in_sg); in stm32_cryp_prepare_req() 1090 if (unlikely(cryp->in_sg->length == _walked_in)) { in stm32_cryp_next_in() 1091 cryp->in_sg = sg_next(cryp->in_sg); in stm32_cryp_next_in() [all …]
|
/Linux-v5.4/drivers/crypto/sunxi-ss/ |
D | sun4i-ss-hash.c | 183 struct scatterlist *in_sg = areq->src; in sun4i_hash() local 247 while (in_sg && i == 1) { in sun4i_hash() 248 if (in_sg->length % 4) in sun4i_hash() 250 in_sg = sg_next(in_sg); in sun4i_hash()
|
D | sun4i-ss-cipher.c | 147 struct scatterlist *in_sg = areq->src; in sun4i_ss_cipher_poll() local 188 while (in_sg && no_chunk == 1) { in sun4i_ss_cipher_poll() 189 if (in_sg->length % 4) in sun4i_ss_cipher_poll() 191 in_sg = sg_next(in_sg); in sun4i_ss_cipher_poll()
|
/Linux-v5.4/crypto/asymmetric_keys/ |
D | public_key.c | 170 struct scatterlist in_sg, out_sg; in software_key_eds_op() local 209 sg_init_one(&in_sg, in, params->in_len); in software_key_eds_op() 211 akcipher_request_set_crypt(req, &in_sg, &out_sg, params->in_len, in software_key_eds_op()
|
D | asym_tpm.c | 467 struct scatterlist in_sg, out_sg; in tpm_key_encrypt() local 493 sg_init_one(&in_sg, in, params->in_len); in tpm_key_encrypt() 495 akcipher_request_set_crypt(req, &in_sg, &out_sg, params->in_len, in tpm_key_encrypt()
|
/Linux-v5.4/crypto/ |
D | rsa-pkcs1pad.c | 101 struct scatterlist in_sg[2], out_sg[1]; member 258 pkcs1pad_sg_set_buf(req_ctx->in_sg, req_ctx->in_buf, in pkcs1pad_encrypt() 266 akcipher_request_set_crypt(&req_ctx->child_req, req_ctx->in_sg, in pkcs1pad_encrypt() 417 pkcs1pad_sg_set_buf(req_ctx->in_sg, req_ctx->in_buf, in pkcs1pad_sign() 425 akcipher_request_set_crypt(&req_ctx->child_req, req_ctx->in_sg, in pkcs1pad_sign()
|
/Linux-v5.4/net/vmw_vsock/ |
D | virtio_transport.c | 149 int ret, in_sg = 0, out_sg = 0; in virtio_transport_send_pkt_work() local 174 ret = virtqueue_add_sgs(vq, sgs, out_sg, in_sg, pkt, GFP_KERNEL); in virtio_transport_send_pkt_work()
|
/Linux-v5.4/drivers/staging/wusbcore/ |
D | wa-xfer.c | 1005 static struct scatterlist *wa_xfer_create_subset_sg(struct scatterlist *in_sg, in wa_xfer_create_subset_sg() argument 1012 struct scatterlist *current_xfer_sg = in_sg; in wa_xfer_create_subset_sg()
|