Home
last modified time | relevance | path

Searched refs:sg_src (Results 1 – 14 of 14) sorted by relevance

/Linux-v5.4/drivers/crypto/rockchip/
Drk3288_crypto.c67 static int check_alignment(struct scatterlist *sg_src, in check_alignment() argument
73 in = IS_ALIGNED((uint32_t)sg_src->offset, 4) && in check_alignment()
74 IS_ALIGNED((uint32_t)sg_src->length, align_mask); in check_alignment()
81 return (align && (sg_src->length == sg_dst->length)); in check_alignment()
85 struct scatterlist *sg_src, in rk_load_data() argument
91 check_alignment(sg_src, sg_dst, dev->align_size) : in rk_load_data()
94 count = min(dev->left_bytes, sg_src->length); in rk_load_data()
97 if (!dma_map_sg(dev->dev, sg_src, 1, DMA_TO_DEVICE)) { in rk_load_data()
102 dev->addr_in = sg_dma_address(sg_src); in rk_load_data()
109 dma_unmap_sg(dev->dev, sg_src, 1, in rk_load_data()
[all …]
Drk3288_crypto_ablkcipher.c252 u8 *src_last_blk = page_address(sg_page(dev->sg_src)) + in rk_set_data_start()
253 dev->sg_src->offset + dev->sg_src->length - ivsize; in rk_set_data_start()
264 err = dev->load_data(dev, dev->sg_src, dev->sg_dst); in rk_set_data_start()
279 dev->sg_src = req->src; in rk_ablk_start()
358 if (sg_is_last(dev->sg_src)) { in rk_ablk_rx()
364 dev->sg_src = sg_next(dev->sg_src); in rk_ablk_rx()
Drk3288_crypto_ahash.c187 err = dev->load_data(dev, dev->sg_src, NULL); in rk_ahash_set_data_start()
204 dev->sg_src = req->src; in rk_ahash_start()
238 if (sg_is_last(dev->sg_src)) { in rk_ahash_crypto_rx()
244 dev->sg_src = sg_next(dev->sg_src); in rk_ahash_crypto_rx()
Drk3288_crypto.h202 struct scatterlist *sg_src; member
223 struct scatterlist *sg_src,
/Linux-v5.4/arch/arm/crypto/
Daes-ce-glue.c283 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local
311 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt()
341 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local
369 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt()
460 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local
506 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt()
532 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local
578 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
/Linux-v5.4/drivers/crypto/nx/
Dnx.c150 struct scatterlist *sg_src, in nx_walk_and_build() argument
161 scatterwalk_start(&walk, sg_src); in nx_walk_and_build()
163 if (start < offset + sg_src->length) in nx_walk_and_build()
166 offset += sg_src->length; in nx_walk_and_build()
167 sg_src = sg_next(sg_src); in nx_walk_and_build()
/Linux-v5.4/arch/arm64/crypto/
Daes-glue.c296 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local
323 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt()
353 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local
380 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt()
531 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local
577 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt()
603 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local
649 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
Daes-neonbs-glue.c324 struct scatterlist sg_src[2], sg_dst[2]; in __xts_crypt() local
394 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in __xts_crypt()
/Linux-v5.4/arch/x86/crypto/
Daegis128-aesni-glue.c70 struct aegis_state *state, struct scatterlist *sg_src, in crypto_aegis128_aesni_process_ad() argument
77 scatterwalk_start(&walk, sg_src); in crypto_aegis128_aesni_process_ad()
/Linux-v5.4/drivers/crypto/ux500/cryp/
Dcryp.h215 struct scatterlist *sg_src; member
Dcryp_core.c540 ctx->device->dma.sg_src = sg; in cryp_set_dma_transfer()
542 ctx->device->dma.sg_src, in cryp_set_dma_transfer()
557 ctx->device->dma.sg_src, in cryp_set_dma_transfer()
617 dma_unmap_sg(chan->device->dev, ctx->device->dma.sg_src, in cryp_dma_done()
/Linux-v5.4/drivers/crypto/
Ds5p-sss.c308 struct scatterlist *sg_src; member
513 dma_unmap_sg(dev->dev, dev->sg_src, 1, DMA_TO_DEVICE); in s5p_unset_indata()
563 dev->sg_src = sg; in s5p_set_indata()
603 if (!sg_is_last(dev->sg_src)) { in s5p_aes_rx()
604 ret = s5p_set_indata(dev, sg_next(dev->sg_src)); in s5p_aes_rx()
748 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_interrupt()
1983 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_crypt_start()
/Linux-v5.4/crypto/
Daegis128-core.c282 struct scatterlist *sg_src, in crypto_aegis128_process_ad() argument
289 scatterwalk_start(&walk, sg_src); in crypto_aegis128_process_ad()
/Linux-v5.4/drivers/dma/
Dste_dma40.c2111 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_log() argument
2120 ret = d40_log_sg_to_lli(sg_src, sg_len, in d40_prep_sg_log()
2139 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_phy() argument
2152 ret = d40_phy_sg_to_lli(sg_src, sg_len, src_dev_addr, in d40_prep_sg_phy()
2209 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, in d40_prep_sg() argument
2229 desc = d40_prep_desc(chan, sg_src, sg_len, dma_flags); in d40_prep_sg()
2233 if (sg_next(&sg_src[sg_len - 1]) == sg_src) in d40_prep_sg()
2244 ret = d40_prep_sg_log(chan, desc, sg_src, sg_dst, in d40_prep_sg()
2247 ret = d40_prep_sg_phy(chan, desc, sg_src, sg_dst, in d40_prep_sg()