/Linux-v4.19/lib/ |
D | sg_split.c | 19 struct scatterlist *out_sg; member 82 struct scatterlist *in_sg, *out_sg; in sg_split_phys() local 87 out_sg = split->out_sg; in sg_split_phys() 88 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys() 89 *out_sg = *in_sg; in sg_split_phys() 91 out_sg->offset += split->skip_sg0; in sg_split_phys() 92 out_sg->length -= split->skip_sg0; in sg_split_phys() 94 out_sg->offset = 0; in sg_split_phys() 96 sg_dma_address(out_sg) = 0; in sg_split_phys() 97 sg_dma_len(out_sg) = 0; in sg_split_phys() [all …]
|
/Linux-v4.19/drivers/crypto/nx/ |
D | nx-aes-xcbc.c | 75 struct nx_sg *in_sg, *out_sg; in nx_xcbc_empty() local 99 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) keys, &len, in nx_xcbc_empty() 106 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_empty() 128 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_xcbc_empty() 135 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_empty() 187 struct nx_sg *out_sg; in nx_xcbc_update() local 216 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_xcbc_update() 224 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_update() 312 struct nx_sg *in_sg, *out_sg; in nx_xcbc_final() local 348 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_xcbc_final() [all …]
|
D | nx-sha512.c | 74 struct nx_sg *out_sg; in nx_sha512_update() local 105 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_sha512_update() 107 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha512_update() 197 struct nx_sg *in_sg, *out_sg; in nx_sha512_final() local 241 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_sha512_final() 245 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha512_final()
|
D | nx-sha256.c | 74 struct nx_sg *out_sg; in nx_sha256_update() local 105 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_sha256_update() 107 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha256_update() 194 struct nx_sg *in_sg, *out_sg; in nx_sha256_final() local 232 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, max_sg_len); in nx_sha256_final() 240 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha256_final()
|
D | nx.c | 281 struct nx_sg *nx_outsg = nx_ctx->out_sg; in nx_build_sg_lists() 306 nx_ctx->op.outlen = trim_sg_list(nx_ctx->out_sg, nx_outsg, delta, nbytes); in nx_build_sg_lists() 326 nx_ctx->op.out = __pa(nx_ctx->out_sg); in nx_ctx_init() 334 nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); in nx_ctx_init() 690 nx_ctx->out_sg = (struct nx_sg *)((u8 *)nx_ctx->in_sg + NX_PAGE_SIZE); in nx_crypto_ctx_init() 694 (struct nx_csbcpb *)((u8 *)nx_ctx->out_sg + in nx_crypto_ctx_init() 766 nx_ctx->out_sg = NULL; in nx_crypto_ctx_exit()
|
D | nx-aes-gcm.c | 263 struct nx_sg *in_sg, *out_sg; in gcm_empty() local 290 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) out, &len, in gcm_empty() 297 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in gcm_empty()
|
D | nx.h | 138 struct nx_sg *out_sg; /* aligned pointer into kmem to an sg list */ member
|
D | nx-aes-ccm.c | 177 struct nx_sg *nx_outsg = nx_ctx->out_sg; in generate_pat() 266 nx_ctx->op.outlen = (nx_ctx->out_sg - nx_outsg) * in generate_pat()
|
/Linux-v4.19/drivers/crypto/ |
D | omap-des.c | 156 struct scatterlist *out_sg; member 381 struct scatterlist *in_sg, struct scatterlist *out_sg, in omap_des_crypt_dma() argument 392 scatterwalk_start(&dd->out_walk, dd->out_sg); in omap_des_crypt_dma() 438 tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, out_sg_len, in omap_des_crypt_dma() 477 err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_des_crypt_dma_start() 485 err = omap_des_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, in omap_des_crypt_dma_start() 489 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_des_crypt_dma_start() 548 dd->out_sg = req->dst; in omap_des_prepare_req() 561 ret = omap_crypto_align_sg(&dd->out_sg, dd->total, DES_BLOCK_SIZE, in omap_des_prepare_req() 571 dd->out_sg_len = sg_nents_for_len(dd->out_sg, dd->total); in omap_des_prepare_req() [all …]
|
D | omap-aes-gcm.c | 38 dd->out_sg = NULL; in omap_aes_gcm_finish_req() 55 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_gcm_done_task() 58 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, DMA_FROM_DEVICE); in omap_aes_gcm_done_task() 61 omap_crypto_cleanup(dd->out_sg, dd->orig_out, in omap_aes_gcm_done_task() 146 dd->out_sg = req->dst; in omap_aes_gcm_copy_buffers() 149 dd->out_sg = scatterwalk_ffwd(sg_arr, req->dst, assoclen); in omap_aes_gcm_copy_buffers() 152 if (req->src == req->dst || dd->out_sg == sg_arr) in omap_aes_gcm_copy_buffers() 155 ret = omap_crypto_align_sg(&dd->out_sg, cryptlen, in omap_aes_gcm_copy_buffers() 163 dd->out_sg_len = sg_nents_for_len(dd->out_sg, clen); in omap_aes_gcm_copy_buffers()
|
D | omap-aes.c | 273 struct scatterlist *out_sg, in omap_aes_crypt_dma() argument 282 scatterwalk_start(&dd->out_walk, dd->out_sg); in omap_aes_crypt_dma() 328 tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, out_sg_len, in omap_aes_crypt_dma() 368 err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_crypt_dma_start() 376 err = omap_aes_crypt_dma(dd, dd->in_sg, dd->out_sg, dd->in_sg_len, in omap_aes_crypt_dma_start() 380 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_crypt_dma_start() 437 dd->out_sg = req->dst; in omap_aes_prepare_req() 450 ret = omap_crypto_align_sg(&dd->out_sg, dd->total, AES_BLOCK_SIZE, in omap_aes_prepare_req() 460 dd->out_sg_len = sg_nents_for_len(dd->out_sg, dd->total); in omap_aes_prepare_req() 493 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_done_task() [all …]
|
D | atmel-tdes.c | 118 struct scatterlist *out_sg; member 328 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); in atmel_tdes_crypt_pdc_stop() 335 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, in atmel_tdes_crypt_pdc_stop() 518 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) && in atmel_tdes_crypt_start() 519 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size); in atmel_tdes_crypt_start() 522 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) in atmel_tdes_crypt_start() 529 count = min_t(size_t, count, sg_dma_len(dd->out_sg)); in atmel_tdes_crypt_start() 537 err = dma_map_sg(dd->dev, dd->out_sg, 1, in atmel_tdes_crypt_start() 547 addr_out = sg_dma_address(dd->out_sg); in atmel_tdes_crypt_start() 571 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE); in atmel_tdes_crypt_start() [all …]
|
D | sahara.c | 226 struct scatterlist *out_sg; member 478 dev->nb_out_sg = sg_nents_for_len(dev->out_sg, dev->total); in sahara_hw_descriptor_create() 495 ret = dma_map_sg(dev->device, dev->out_sg, dev->nb_out_sg, in sahara_hw_descriptor_create() 518 sg = dev->out_sg; in sahara_hw_descriptor_create() 544 dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, in sahara_hw_descriptor_create() 569 dev->out_sg = req->dst; in sahara_aes_process() 595 dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, in sahara_aes_process()
|
D | omap-aes.h | 183 struct scatterlist *out_sg; member
|
/Linux-v4.19/drivers/crypto/stm32/ |
D | stm32-cryp.c | 108 #define _walked_out (cryp->out_walk.offset - cryp->out_sg->offset) 153 struct scatterlist *out_sg; member 323 ret = stm32_cryp_check_aligned(cryp->out_sg, cryp->total_out, in stm32_cryp_check_io_aligned() 374 cryp->out_sg_save = cryp->out_sg; in stm32_cryp_copy_sgs() 375 cryp->out_sg = &cryp->out_sgl; in stm32_cryp_copy_sgs() 972 cryp->out_sg = req ? req->dst : areq->dst; in stm32_cryp_prepare_req() 973 cryp->out_sg_save = cryp->out_sg; in stm32_cryp_prepare_req() 982 cryp->out_sg_len = sg_nents_for_len(cryp->out_sg, cryp->total_out); in stm32_cryp_prepare_req() 994 scatterwalk_start(&cryp->out_walk, cryp->out_sg); in stm32_cryp_prepare_req() 1068 if (unlikely(cryp->out_sg->length == _walked_out)) { in stm32_cryp_next_out() [all …]
|
/Linux-v4.19/crypto/ |
D | rsa-pkcs1pad.c | 104 struct scatterlist in_sg[2], out_sg[1]; member 270 pkcs1pad_sg_set_buf(req_ctx->out_sg, req_ctx->out_buf, in pkcs1pad_encrypt() 371 pkcs1pad_sg_set_buf(req_ctx->out_sg, req_ctx->out_buf, in pkcs1pad_decrypt() 380 req_ctx->out_sg, req->src_len, in pkcs1pad_decrypt() 545 pkcs1pad_sg_set_buf(req_ctx->out_sg, req_ctx->out_buf, in pkcs1pad_verify() 554 req_ctx->out_sg, req->src_len, in pkcs1pad_verify()
|
/Linux-v4.19/drivers/crypto/sunxi-ss/ |
D | sun4i-ss-cipher.c | 134 struct scatterlist *out_sg = areq->dst; in sun4i_ss_cipher_poll() local 179 while (out_sg && no_chunk == 1) { in sun4i_ss_cipher_poll() 180 if (out_sg->length % 4) in sun4i_ss_cipher_poll() 182 out_sg = sg_next(out_sg); in sun4i_ss_cipher_poll()
|
/Linux-v4.19/net/vmw_vsock/ |
D | virtio_transport.c | 133 int ret, in_sg = 0, out_sg = 0; in virtio_transport_send_pkt_work() local 152 sgs[out_sg++] = &hdr; in virtio_transport_send_pkt_work() 155 sgs[out_sg++] = &buf; in virtio_transport_send_pkt_work() 158 ret = virtqueue_add_sgs(vq, sgs, out_sg, in_sg, pkt, GFP_KERNEL); in virtio_transport_send_pkt_work()
|
/Linux-v4.19/net/tls/ |
D | tls_sw.c | 668 struct scatterlist *out_sg, in decrypt_internal() argument 682 if (*zc && (out_iov || out_sg)) { in decrypt_internal() 686 n_sgout = sg_nents(out_sg); in decrypt_internal() 758 } else if (out_sg) { in decrypt_internal() 759 memcpy(sgout, out_sg, n_sgout * sizeof(*sgout)); in decrypt_internal()
|
/Linux-v4.19/drivers/usb/wusbcore/ |
D | wa-xfer.c | 1009 struct scatterlist *out_sg; in wa_xfer_create_subset_sg() local 1039 out_sg = kmalloc((sizeof(struct scatterlist) * nents), GFP_ATOMIC); in wa_xfer_create_subset_sg() 1040 if (out_sg) { in wa_xfer_create_subset_sg() 1041 sg_init_table(out_sg, nents); in wa_xfer_create_subset_sg() 1045 last_seg_sg = current_seg_sg = out_sg; in wa_xfer_create_subset_sg() 1079 return out_sg; in wa_xfer_create_subset_sg()
|