/Linux-v5.4/drivers/crypto/rockchip/ |
D | rk3288_crypto.c | 67 static int check_alignment(struct scatterlist *sg_src, in check_alignment() argument 73 in = IS_ALIGNED((uint32_t)sg_src->offset, 4) && in check_alignment() 74 IS_ALIGNED((uint32_t)sg_src->length, align_mask); in check_alignment() 81 return (align && (sg_src->length == sg_dst->length)); in check_alignment() 85 struct scatterlist *sg_src, in rk_load_data() argument 91 check_alignment(sg_src, sg_dst, dev->align_size) : in rk_load_data() 94 count = min(dev->left_bytes, sg_src->length); in rk_load_data() 97 if (!dma_map_sg(dev->dev, sg_src, 1, DMA_TO_DEVICE)) { in rk_load_data() 102 dev->addr_in = sg_dma_address(sg_src); in rk_load_data() 109 dma_unmap_sg(dev->dev, sg_src, 1, in rk_load_data() [all …]
|
D | rk3288_crypto_ablkcipher.c | 252 u8 *src_last_blk = page_address(sg_page(dev->sg_src)) + in rk_set_data_start() 253 dev->sg_src->offset + dev->sg_src->length - ivsize; in rk_set_data_start() 264 err = dev->load_data(dev, dev->sg_src, dev->sg_dst); in rk_set_data_start() 279 dev->sg_src = req->src; in rk_ablk_start() 358 if (sg_is_last(dev->sg_src)) { in rk_ablk_rx() 364 dev->sg_src = sg_next(dev->sg_src); in rk_ablk_rx()
|
D | rk3288_crypto_ahash.c | 187 err = dev->load_data(dev, dev->sg_src, NULL); in rk_ahash_set_data_start() 204 dev->sg_src = req->src; in rk_ahash_start() 238 if (sg_is_last(dev->sg_src)) { in rk_ahash_crypto_rx() 244 dev->sg_src = sg_next(dev->sg_src); in rk_ahash_crypto_rx()
|
D | rk3288_crypto.h | 202 struct scatterlist *sg_src; member 223 struct scatterlist *sg_src,
|
/Linux-v5.4/arch/arm/crypto/ |
D | aes-ce-glue.c | 283 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local 311 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt() 341 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local 369 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt() 460 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local 506 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt() 532 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local 578 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
|
/Linux-v5.4/drivers/crypto/nx/ |
D | nx.c | 150 struct scatterlist *sg_src, in nx_walk_and_build() argument 161 scatterwalk_start(&walk, sg_src); in nx_walk_and_build() 163 if (start < offset + sg_src->length) in nx_walk_and_build() 166 offset += sg_src->length; in nx_walk_and_build() 167 sg_src = sg_next(sg_src); in nx_walk_and_build()
|
/Linux-v5.4/arch/arm64/crypto/ |
D | aes-glue.c | 296 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local 323 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt() 353 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local 380 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt() 531 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local 577 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt() 603 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local 649 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
|
D | aes-neonbs-glue.c | 324 struct scatterlist sg_src[2], sg_dst[2]; in __xts_crypt() local 394 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in __xts_crypt()
|
/Linux-v5.4/arch/x86/crypto/ |
D | aegis128-aesni-glue.c | 70 struct aegis_state *state, struct scatterlist *sg_src, in crypto_aegis128_aesni_process_ad() argument 77 scatterwalk_start(&walk, sg_src); in crypto_aegis128_aesni_process_ad()
|
/Linux-v5.4/drivers/crypto/ux500/cryp/ |
D | cryp.h | 215 struct scatterlist *sg_src; member
|
D | cryp_core.c | 540 ctx->device->dma.sg_src = sg; in cryp_set_dma_transfer() 542 ctx->device->dma.sg_src, in cryp_set_dma_transfer() 557 ctx->device->dma.sg_src, in cryp_set_dma_transfer() 617 dma_unmap_sg(chan->device->dev, ctx->device->dma.sg_src, in cryp_dma_done()
|
/Linux-v5.4/drivers/crypto/ |
D | s5p-sss.c | 308 struct scatterlist *sg_src; member 513 dma_unmap_sg(dev->dev, dev->sg_src, 1, DMA_TO_DEVICE); in s5p_unset_indata() 563 dev->sg_src = sg; in s5p_set_indata() 603 if (!sg_is_last(dev->sg_src)) { in s5p_aes_rx() 604 ret = s5p_set_indata(dev, sg_next(dev->sg_src)); in s5p_aes_rx() 748 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_interrupt() 1983 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_crypt_start()
|
/Linux-v5.4/crypto/ |
D | aegis128-core.c | 282 struct scatterlist *sg_src, in crypto_aegis128_process_ad() argument 289 scatterwalk_start(&walk, sg_src); in crypto_aegis128_process_ad()
|
/Linux-v5.4/drivers/dma/ |
D | ste_dma40.c | 2111 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_log() argument 2120 ret = d40_log_sg_to_lli(sg_src, sg_len, in d40_prep_sg_log() 2139 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_phy() argument 2152 ret = d40_phy_sg_to_lli(sg_src, sg_len, src_dev_addr, in d40_prep_sg_phy() 2209 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, in d40_prep_sg() argument 2229 desc = d40_prep_desc(chan, sg_src, sg_len, dma_flags); in d40_prep_sg() 2233 if (sg_next(&sg_src[sg_len - 1]) == sg_src) in d40_prep_sg() 2244 ret = d40_prep_sg_log(chan, desc, sg_src, sg_dst, in d40_prep_sg() 2247 ret = d40_prep_sg_phy(chan, desc, sg_src, sg_dst, in d40_prep_sg()
|