Searched refs:drv_ctx (Results 1 – 5 of 5) sorted by relevance
/Linux-v6.6/drivers/crypto/caam/ |
D | qi.c | 125 ret = qman_enqueue(req->drv_ctx->req_fq, &fd); in caam_qi_enqueue() 127 refcount_inc(&req->drv_ctx->refcnt); in caam_qi_enqueue() 159 refcount_dec(&drv_req->drv_ctx->refcnt); in caam_fq_ern_cb() 166 dma_unmap_single(drv_req->drv_ctx->qidev, qm_fd_addr(fd), in caam_fq_ern_cb() 293 static int empty_caam_fq(struct qman_fq *fq, struct caam_drv_ctx *drv_ctx) in empty_caam_fq() argument 313 if (refcount_read(&drv_ctx->refcnt) == 1) in empty_caam_fq() 320 dev_warn_once(drv_ctx->qidev, "%d frames from FQID %u still pending in CAAM\n", in empty_caam_fq() 321 refcount_read(&drv_ctx->refcnt), fq->fqid); in empty_caam_fq() 326 int caam_drv_ctx_update(struct caam_drv_ctx *drv_ctx, u32 *sh_desc) in caam_drv_ctx_update() argument 331 struct device *qidev = drv_ctx->qidev; in caam_drv_ctx_update() [all …]
|
D | qi.h | 92 struct caam_drv_ctx *drv_ctx; member 135 bool caam_drv_ctx_busy(struct caam_drv_ctx *drv_ctx); 147 int caam_drv_ctx_update(struct caam_drv_ctx *drv_ctx, u32 *sh_desc); 153 void caam_drv_ctx_rel(struct caam_drv_ctx *drv_ctx);
|
D | caamalg_qi.c | 76 struct caam_drv_ctx *drv_ctx[NUM_OP]; member 263 if (ctx->drv_ctx[ENCRYPT]) { in aead_setkey() 264 ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT], in aead_setkey() 272 if (ctx->drv_ctx[DECRYPT]) { in aead_setkey() 273 ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT], in aead_setkey() 387 if (ctx->drv_ctx[ENCRYPT]) { in gcm_setkey() 388 ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT], in gcm_setkey() 396 if (ctx->drv_ctx[DECRYPT]) { in gcm_setkey() 397 ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT], in gcm_setkey() 495 if (ctx->drv_ctx[ENCRYPT]) { in rfc4106_setkey() [all …]
|
/Linux-v6.6/drivers/crypto/xilinx/ |
D | zynqmp-aes-gcm.c | 283 struct zynqmp_aead_drv_ctx *drv_ctx; in zynqmp_aes_aead_encrypt() local 289 drv_ctx = container_of(alg, struct zynqmp_aead_drv_ctx, alg.aead.base); in zynqmp_aes_aead_encrypt() 291 return crypto_transfer_aead_request_to_engine(drv_ctx->engine, req); in zynqmp_aes_aead_encrypt() 296 struct zynqmp_aead_drv_ctx *drv_ctx; in zynqmp_aes_aead_decrypt() local 302 drv_ctx = container_of(alg, struct zynqmp_aead_drv_ctx, alg.aead.base); in zynqmp_aes_aead_decrypt() 304 return crypto_transfer_aead_request_to_engine(drv_ctx->engine, req); in zynqmp_aes_aead_decrypt() 312 struct zynqmp_aead_drv_ctx *drv_ctx; in zynqmp_aes_aead_init() local 315 drv_ctx = container_of(alg, struct zynqmp_aead_drv_ctx, alg.aead.base); in zynqmp_aes_aead_init() 316 tfm_ctx->dev = drv_ctx->dev; in zynqmp_aes_aead_init() 318 tfm_ctx->fbk_cipher = crypto_alloc_aead(drv_ctx->alg.aead.base.base.cra_name, in zynqmp_aes_aead_init() [all …]
|
D | zynqmp-sha.c | 52 struct zynqmp_sha_drv_ctx *drv_ctx; in zynqmp_sha_init_tfm() local 54 drv_ctx = container_of(alg, struct zynqmp_sha_drv_ctx, sha3_384); in zynqmp_sha_init_tfm() 55 tfm_ctx->dev = drv_ctx->dev; in zynqmp_sha_init_tfm()
|