Lines Matching refs:crypt

350 static void finish_scattered_hmac(struct crypt_ctl *crypt)  in finish_scattered_hmac()  argument
352 struct aead_request *req = crypt->data.aead_req; in finish_scattered_hmac()
362 dma_pool_free(buffer_pool, req_ctx->hmac_virt, crypt->icv_rev_aes); in finish_scattered_hmac()
368 struct crypt_ctl *crypt; in one_packet() local
374 crypt = crypt_phys2virt(phys); in one_packet()
376 switch (crypt->ctl_flags & CTL_FLAG_MASK) { in one_packet()
378 struct aead_request *req = crypt->data.aead_req; in one_packet()
381 free_buf_chain(dev, req_ctx->src, crypt->src_buf); in one_packet()
382 free_buf_chain(dev, req_ctx->dst, crypt->dst_buf); in one_packet()
384 finish_scattered_hmac(crypt); in one_packet()
390 struct skcipher_request *req = crypt->data.ablk_req; in one_packet()
408 free_buf_chain(dev, req_ctx->dst, crypt->dst_buf); in one_packet()
410 free_buf_chain(dev, req_ctx->src, crypt->src_buf); in one_packet()
415 ctx = crypto_tfm_ctx(crypt->data.tfm); in one_packet()
416 dma_pool_free(ctx_pool, crypt->regist_ptr, in one_packet()
417 crypt->regist_buf->phys_addr); in one_packet()
418 dma_pool_free(buffer_pool, crypt->regist_buf, crypt->src_buf); in one_packet()
423 ctx = crypto_tfm_ctx(crypt->data.tfm); in one_packet()
431 crypt->ctl_flags = CTL_FLAG_UNUSED; in one_packet()
684 struct crypt_ctl *crypt; in register_chain_var() local
699 crypt = get_crypt_desc_emerg(); in register_chain_var()
700 if (!crypt) { in register_chain_var()
711 crypt->data.tfm = tfm; in register_chain_var()
712 crypt->regist_ptr = pad; in register_chain_var()
713 crypt->regist_buf = buf; in register_chain_var()
715 crypt->auth_offs = 0; in register_chain_var()
716 crypt->auth_len = HMAC_PAD_BLOCKLEN; in register_chain_var()
717 crypt->crypto_ctx = ctx_addr; in register_chain_var()
718 crypt->src_buf = buf_phys; in register_chain_var()
719 crypt->icv_rev_aes = target; in register_chain_var()
720 crypt->mode = NPE_OP_HASH_GEN_ICV; in register_chain_var()
721 crypt->init_len = init_len; in register_chain_var()
722 crypt->ctl_flags |= CTL_FLAG_GEN_ICV; in register_chain_var()
730 qmgr_put_entry(send_qid, crypt_virt2phys(crypt)); in register_chain_var()
784 struct crypt_ctl *crypt; in gen_rev_aes_key() local
788 crypt = get_crypt_desc_emerg(); in gen_rev_aes_key()
789 if (!crypt) in gen_rev_aes_key()
794 crypt->data.tfm = tfm; in gen_rev_aes_key()
795 crypt->crypt_offs = 0; in gen_rev_aes_key()
796 crypt->crypt_len = AES_BLOCK128; in gen_rev_aes_key()
797 crypt->src_buf = 0; in gen_rev_aes_key()
798 crypt->crypto_ctx = dir->npe_ctx_phys; in gen_rev_aes_key()
799 crypt->icv_rev_aes = dir->npe_ctx_phys + sizeof(u32); in gen_rev_aes_key()
800 crypt->mode = NPE_OP_ENC_GEN_KEY; in gen_rev_aes_key()
801 crypt->init_len = dir->npe_ctx_idx; in gen_rev_aes_key()
802 crypt->ctl_flags |= CTL_FLAG_GEN_REVAES; in gen_rev_aes_key()
805 qmgr_put_entry(send_qid, crypt_virt2phys(crypt)); in gen_rev_aes_key()
978 struct crypt_ctl *crypt; in ablk_perform() local
999 crypt = get_crypt_desc(); in ablk_perform()
1000 if (!crypt) in ablk_perform()
1003 crypt->data.ablk_req = req; in ablk_perform()
1004 crypt->crypto_ctx = dir->npe_ctx_phys; in ablk_perform()
1005 crypt->mode = dir->npe_mode; in ablk_perform()
1006 crypt->init_len = dir->npe_ctx_idx; in ablk_perform()
1008 crypt->crypt_offs = 0; in ablk_perform()
1009 crypt->crypt_len = nbytes; in ablk_perform()
1012 memcpy(crypt->iv, req->iv, ivsize); in ablk_perform()
1020 crypt->mode |= NPE_OP_NOT_IN_PLACE; in ablk_perform()
1029 crypt->dst_buf = dst_hook.phys_next; in ablk_perform()
1039 crypt->src_buf = src_hook.phys_next; in ablk_perform()
1040 crypt->ctl_flags |= CTL_FLAG_PERFORM_ABLK; in ablk_perform()
1041 qmgr_put_entry(send_qid, crypt_virt2phys(crypt)); in ablk_perform()
1046 free_buf_chain(dev, req_ctx->src, crypt->src_buf); in ablk_perform()
1049 free_buf_chain(dev, req_ctx->dst, crypt->dst_buf); in ablk_perform()
1051 crypt->ctl_flags = CTL_FLAG_UNUSED; in ablk_perform()
1095 struct crypt_ctl *crypt; in aead_perform() local
1119 crypt = get_crypt_desc(); in aead_perform()
1120 if (!crypt) in aead_perform()
1123 crypt->data.aead_req = req; in aead_perform()
1124 crypt->crypto_ctx = dir->npe_ctx_phys; in aead_perform()
1125 crypt->mode = dir->npe_mode; in aead_perform()
1126 crypt->init_len = dir->npe_ctx_idx; in aead_perform()
1128 crypt->crypt_offs = cryptoffset; in aead_perform()
1129 crypt->crypt_len = eff_cryptlen; in aead_perform()
1131 crypt->auth_offs = 0; in aead_perform()
1132 crypt->auth_len = req->assoclen + cryptlen; in aead_perform()
1134 memcpy(crypt->iv, req->iv, ivsize); in aead_perform()
1136 buf = chainup_buffers(dev, req->src, crypt->auth_len, in aead_perform()
1139 crypt->src_buf = src_hook.phys_next; in aead_perform()
1145 crypt->icv_rev_aes = buf->phys_addr + in aead_perform()
1153 crypt->mode |= NPE_OP_NOT_IN_PLACE; in aead_perform()
1156 buf = chainup_buffers(dev, req->dst, crypt->auth_len, in aead_perform()
1159 crypt->dst_buf = dst_hook.phys_next; in aead_perform()
1167 crypt->icv_rev_aes = buf->phys_addr + in aead_perform()
1176 &crypt->icv_rev_aes); in aead_perform()
1188 crypt->ctl_flags |= CTL_FLAG_PERFORM_AEAD; in aead_perform()
1189 qmgr_put_entry(send_qid, crypt_virt2phys(crypt)); in aead_perform()
1194 free_buf_chain(dev, req_ctx->dst, crypt->dst_buf); in aead_perform()
1196 free_buf_chain(dev, req_ctx->src, crypt->src_buf); in aead_perform()
1197 crypt->ctl_flags = CTL_FLAG_UNUSED; in aead_perform()