Lines Matching refs:src_nents
146 struct scatterlist *dst, int src_nents, in caam_unmap() argument
152 if (src_nents) in caam_unmap()
153 dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE); in caam_unmap()
157 dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL); in caam_unmap()
359 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local
380 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc()
381 if (unlikely(src_nents < 0)) { in aead_edesc_alloc()
385 return ERR_PTR(src_nents); in aead_edesc_alloc()
396 if (src_nents) { in aead_edesc_alloc()
397 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc()
413 dma_unmap_sg(dev, req->src, src_nents, in aead_edesc_alloc()
425 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc()
426 if (unlikely(src_nents < 0)) { in aead_edesc_alloc()
430 return ERR_PTR(src_nents); in aead_edesc_alloc()
433 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc()
473 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in aead_edesc_alloc()
488 caam_unmap(dev, req->src, req->dst, src_nents, in aead_edesc_alloc()
495 edesc->src_nents = src_nents; in aead_edesc_alloc()
512 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in aead_edesc_alloc()
534 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in aead_edesc_alloc()
1116 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in skcipher_edesc_alloc() local
1124 src_nents = sg_nents_for_len(req->src, req->cryptlen); in skcipher_edesc_alloc()
1125 if (unlikely(src_nents < 0)) { in skcipher_edesc_alloc()
1128 return ERR_PTR(src_nents); in skcipher_edesc_alloc()
1139 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in skcipher_edesc_alloc()
1150 dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE); in skcipher_edesc_alloc()
1154 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in skcipher_edesc_alloc()
1183 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1192 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1205 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1211 edesc->src_nents = src_nents; in skcipher_edesc_alloc()
1229 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in skcipher_edesc_alloc()
1261 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in aead_unmap()
1273 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in skcipher_unmap()
1404 edesc->src_nents > 1 ? 100 : ivsize, 1); in skcipher_encrypt_done()
1442 edesc->src_nents > 1 ? 100 : ivsize, 1); in skcipher_decrypt_done()
3357 if (edesc->src_nents) in ahash_unmap()
3358 dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE); in ahash_unmap()
3528 int src_nents, mapped_nents, qm_sg_bytes, qm_sg_src_index; in ahash_update_ctx() local
3539 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_ctx()
3540 if (src_nents < 0) { in ahash_update_ctx()
3542 return src_nents; in ahash_update_ctx()
3545 if (src_nents) { in ahash_update_ctx()
3546 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_ctx()
3559 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_ctx()
3564 edesc->src_nents = src_nents; in ahash_update_ctx()
3716 int src_nents, mapped_nents; in ahash_finup_ctx() local
3722 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_finup_ctx()
3723 if (src_nents < 0) { in ahash_finup_ctx()
3725 return src_nents; in ahash_finup_ctx()
3728 if (src_nents) { in ahash_finup_ctx()
3729 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_finup_ctx()
3742 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_finup_ctx()
3746 edesc->src_nents = src_nents; in ahash_finup_ctx()
3809 int src_nents, mapped_nents; in ahash_digest() local
3815 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_digest()
3816 if (src_nents < 0) { in ahash_digest()
3818 return src_nents; in ahash_digest()
3821 if (src_nents) { in ahash_digest()
3822 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_digest()
3835 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_digest()
3839 edesc->src_nents = src_nents; in ahash_digest()
3980 int qm_sg_bytes, src_nents, mapped_nents; in ahash_update_no_ctx() local
3991 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_no_ctx()
3992 if (src_nents < 0) { in ahash_update_no_ctx()
3994 return src_nents; in ahash_update_no_ctx()
3997 if (src_nents) { in ahash_update_no_ctx()
3998 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_no_ctx()
4011 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_no_ctx()
4016 edesc->src_nents = src_nents; in ahash_update_no_ctx()
4098 int qm_sg_bytes, src_nents, mapped_nents; in ahash_finup_no_ctx() local
4104 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_finup_no_ctx()
4105 if (src_nents < 0) { in ahash_finup_no_ctx()
4107 return src_nents; in ahash_finup_no_ctx()
4110 if (src_nents) { in ahash_finup_no_ctx()
4111 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_finup_no_ctx()
4124 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_finup_no_ctx()
4128 edesc->src_nents = src_nents; in ahash_finup_no_ctx()
4197 int src_nents, mapped_nents; in ahash_update_first() local
4209 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_first()
4210 if (src_nents < 0) { in ahash_update_first()
4212 return src_nents; in ahash_update_first()
4215 if (src_nents) { in ahash_update_first()
4216 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_first()
4229 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_first()
4234 edesc->src_nents = src_nents; in ahash_update_first()