Lines Matching refs:pdb

43 	struct rsa_pub_pdb *pdb = &edesc->pdb.pub;  in rsa_pub_unmap()  local
45 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE); in rsa_pub_unmap()
46 dma_unmap_single(dev, pdb->e_dma, key->e_sz, DMA_TO_DEVICE); in rsa_pub_unmap()
55 struct rsa_priv_f1_pdb *pdb = &edesc->pdb.priv_f1; in rsa_priv_f1_unmap() local
57 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE); in rsa_priv_f1_unmap()
58 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE); in rsa_priv_f1_unmap()
67 struct rsa_priv_f2_pdb *pdb = &edesc->pdb.priv_f2; in rsa_priv_f2_unmap() local
71 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE); in rsa_priv_f2_unmap()
72 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE); in rsa_priv_f2_unmap()
73 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE); in rsa_priv_f2_unmap()
74 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL); in rsa_priv_f2_unmap()
75 dma_unmap_single(dev, pdb->tmp2_dma, q_sz, DMA_BIDIRECTIONAL); in rsa_priv_f2_unmap()
84 struct rsa_priv_f3_pdb *pdb = &edesc->pdb.priv_f3; in rsa_priv_f3_unmap() local
88 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE); in rsa_priv_f3_unmap()
89 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE); in rsa_priv_f3_unmap()
90 dma_unmap_single(dev, pdb->dp_dma, p_sz, DMA_TO_DEVICE); in rsa_priv_f3_unmap()
91 dma_unmap_single(dev, pdb->dq_dma, q_sz, DMA_TO_DEVICE); in rsa_priv_f3_unmap()
92 dma_unmap_single(dev, pdb->c_dma, p_sz, DMA_TO_DEVICE); in rsa_priv_f3_unmap()
93 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL); in rsa_priv_f3_unmap()
94 dma_unmap_single(dev, pdb->tmp2_dma, q_sz, DMA_BIDIRECTIONAL); in rsa_priv_f3_unmap()
308 struct rsa_pub_pdb *pdb = &edesc->pdb.pub; in set_rsa_pub_pdb() local
311 pdb->n_dma = dma_map_single(dev, key->n, key->n_sz, DMA_TO_DEVICE); in set_rsa_pub_pdb()
312 if (dma_mapping_error(dev, pdb->n_dma)) { in set_rsa_pub_pdb()
317 pdb->e_dma = dma_map_single(dev, key->e, key->e_sz, DMA_TO_DEVICE); in set_rsa_pub_pdb()
318 if (dma_mapping_error(dev, pdb->e_dma)) { in set_rsa_pub_pdb()
320 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE); in set_rsa_pub_pdb()
325 pdb->sgf |= RSA_PDB_SGF_F; in set_rsa_pub_pdb()
326 pdb->f_dma = edesc->sec4_sg_dma; in set_rsa_pub_pdb()
329 pdb->f_dma = sg_dma_address(req->src); in set_rsa_pub_pdb()
333 pdb->sgf |= RSA_PDB_SGF_G; in set_rsa_pub_pdb()
334 pdb->g_dma = edesc->sec4_sg_dma + in set_rsa_pub_pdb()
337 pdb->g_dma = sg_dma_address(req->dst); in set_rsa_pub_pdb()
340 pdb->sgf |= (key->e_sz << RSA_PDB_E_SHIFT) | key->n_sz; in set_rsa_pub_pdb()
341 pdb->f_len = req->src_len; in set_rsa_pub_pdb()
353 struct rsa_priv_f1_pdb *pdb = &edesc->pdb.priv_f1; in set_rsa_priv_f1_pdb() local
356 pdb->n_dma = dma_map_single(dev, key->n, key->n_sz, DMA_TO_DEVICE); in set_rsa_priv_f1_pdb()
357 if (dma_mapping_error(dev, pdb->n_dma)) { in set_rsa_priv_f1_pdb()
362 pdb->d_dma = dma_map_single(dev, key->d, key->d_sz, DMA_TO_DEVICE); in set_rsa_priv_f1_pdb()
363 if (dma_mapping_error(dev, pdb->d_dma)) { in set_rsa_priv_f1_pdb()
365 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE); in set_rsa_priv_f1_pdb()
370 pdb->sgf |= RSA_PRIV_PDB_SGF_G; in set_rsa_priv_f1_pdb()
371 pdb->g_dma = edesc->sec4_sg_dma; in set_rsa_priv_f1_pdb()
374 pdb->g_dma = sg_dma_address(req->src); in set_rsa_priv_f1_pdb()
378 pdb->sgf |= RSA_PRIV_PDB_SGF_F; in set_rsa_priv_f1_pdb()
379 pdb->f_dma = edesc->sec4_sg_dma + in set_rsa_priv_f1_pdb()
382 pdb->f_dma = sg_dma_address(req->dst); in set_rsa_priv_f1_pdb()
385 pdb->sgf |= (key->d_sz << RSA_PDB_D_SHIFT) | key->n_sz; in set_rsa_priv_f1_pdb()
397 struct rsa_priv_f2_pdb *pdb = &edesc->pdb.priv_f2; in set_rsa_priv_f2_pdb() local
402 pdb->d_dma = dma_map_single(dev, key->d, key->d_sz, DMA_TO_DEVICE); in set_rsa_priv_f2_pdb()
403 if (dma_mapping_error(dev, pdb->d_dma)) { in set_rsa_priv_f2_pdb()
408 pdb->p_dma = dma_map_single(dev, key->p, p_sz, DMA_TO_DEVICE); in set_rsa_priv_f2_pdb()
409 if (dma_mapping_error(dev, pdb->p_dma)) { in set_rsa_priv_f2_pdb()
414 pdb->q_dma = dma_map_single(dev, key->q, q_sz, DMA_TO_DEVICE); in set_rsa_priv_f2_pdb()
415 if (dma_mapping_error(dev, pdb->q_dma)) { in set_rsa_priv_f2_pdb()
420 pdb->tmp1_dma = dma_map_single(dev, key->tmp1, p_sz, DMA_BIDIRECTIONAL); in set_rsa_priv_f2_pdb()
421 if (dma_mapping_error(dev, pdb->tmp1_dma)) { in set_rsa_priv_f2_pdb()
426 pdb->tmp2_dma = dma_map_single(dev, key->tmp2, q_sz, DMA_BIDIRECTIONAL); in set_rsa_priv_f2_pdb()
427 if (dma_mapping_error(dev, pdb->tmp2_dma)) { in set_rsa_priv_f2_pdb()
433 pdb->sgf |= RSA_PRIV_PDB_SGF_G; in set_rsa_priv_f2_pdb()
434 pdb->g_dma = edesc->sec4_sg_dma; in set_rsa_priv_f2_pdb()
437 pdb->g_dma = sg_dma_address(req->src); in set_rsa_priv_f2_pdb()
441 pdb->sgf |= RSA_PRIV_PDB_SGF_F; in set_rsa_priv_f2_pdb()
442 pdb->f_dma = edesc->sec4_sg_dma + in set_rsa_priv_f2_pdb()
445 pdb->f_dma = sg_dma_address(req->dst); in set_rsa_priv_f2_pdb()
448 pdb->sgf |= (key->d_sz << RSA_PDB_D_SHIFT) | key->n_sz; in set_rsa_priv_f2_pdb()
449 pdb->p_q_len = (q_sz << RSA_PDB_Q_SHIFT) | p_sz; in set_rsa_priv_f2_pdb()
454 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL); in set_rsa_priv_f2_pdb()
456 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE); in set_rsa_priv_f2_pdb()
458 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE); in set_rsa_priv_f2_pdb()
460 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE); in set_rsa_priv_f2_pdb()
472 struct rsa_priv_f3_pdb *pdb = &edesc->pdb.priv_f3; in set_rsa_priv_f3_pdb() local
477 pdb->p_dma = dma_map_single(dev, key->p, p_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
478 if (dma_mapping_error(dev, pdb->p_dma)) { in set_rsa_priv_f3_pdb()
483 pdb->q_dma = dma_map_single(dev, key->q, q_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
484 if (dma_mapping_error(dev, pdb->q_dma)) { in set_rsa_priv_f3_pdb()
489 pdb->dp_dma = dma_map_single(dev, key->dp, p_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
490 if (dma_mapping_error(dev, pdb->dp_dma)) { in set_rsa_priv_f3_pdb()
495 pdb->dq_dma = dma_map_single(dev, key->dq, q_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
496 if (dma_mapping_error(dev, pdb->dq_dma)) { in set_rsa_priv_f3_pdb()
501 pdb->c_dma = dma_map_single(dev, key->qinv, p_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
502 if (dma_mapping_error(dev, pdb->c_dma)) { in set_rsa_priv_f3_pdb()
507 pdb->tmp1_dma = dma_map_single(dev, key->tmp1, p_sz, DMA_BIDIRECTIONAL); in set_rsa_priv_f3_pdb()
508 if (dma_mapping_error(dev, pdb->tmp1_dma)) { in set_rsa_priv_f3_pdb()
513 pdb->tmp2_dma = dma_map_single(dev, key->tmp2, q_sz, DMA_BIDIRECTIONAL); in set_rsa_priv_f3_pdb()
514 if (dma_mapping_error(dev, pdb->tmp2_dma)) { in set_rsa_priv_f3_pdb()
520 pdb->sgf |= RSA_PRIV_PDB_SGF_G; in set_rsa_priv_f3_pdb()
521 pdb->g_dma = edesc->sec4_sg_dma; in set_rsa_priv_f3_pdb()
524 pdb->g_dma = sg_dma_address(req->src); in set_rsa_priv_f3_pdb()
528 pdb->sgf |= RSA_PRIV_PDB_SGF_F; in set_rsa_priv_f3_pdb()
529 pdb->f_dma = edesc->sec4_sg_dma + in set_rsa_priv_f3_pdb()
532 pdb->f_dma = sg_dma_address(req->dst); in set_rsa_priv_f3_pdb()
535 pdb->sgf |= key->n_sz; in set_rsa_priv_f3_pdb()
536 pdb->p_q_len = (q_sz << RSA_PDB_Q_SHIFT) | p_sz; in set_rsa_priv_f3_pdb()
541 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL); in set_rsa_priv_f3_pdb()
543 dma_unmap_single(dev, pdb->c_dma, p_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
545 dma_unmap_single(dev, pdb->dq_dma, q_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
547 dma_unmap_single(dev, pdb->dp_dma, p_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
549 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
551 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE); in set_rsa_priv_f3_pdb()
585 init_rsa_pub_desc(edesc->hw_desc, &edesc->pdb.pub); in caam_rsa_enc()
618 init_rsa_priv_f1_desc(edesc->hw_desc, &edesc->pdb.priv_f1); in caam_rsa_dec_priv_f1()
651 init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2); in caam_rsa_dec_priv_f2()
684 init_rsa_priv_f3_desc(edesc->hw_desc, &edesc->pdb.priv_f3); in caam_rsa_dec_priv_f3()