Lines Matching refs:IV
73 #define IV AES_BLOCK_SIZE macro
844 chcr_req->sec_cpl.pldlen = htonl(IV + wrparam->bytes); in create_cipher_wr()
846 FILL_SEC_CPL_CIPHERSTOP_HI(0, 0, IV + 1, 0); in create_cipher_wr()
852 0, 0, IV >> 1); in create_cipher_wr()
884 temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + kctx_len + IV in create_cipher_wr()
1203 memcpy(req->iv, reqctx->init_iv, IV); in chcr_handle_cipher_resp()
1294 reqctx->imm = (transhdr_len + IV + req->cryptlen) <= in process_cipher()
1296 bytes = IV + req->cryptlen; in process_cipher()
1324 memcpy(reqctx->init_iv, reqctx->iv, IV); in process_cipher()
1328 memcpy(reqctx->iv, req->iv, IV); in process_cipher()
1329 memcpy(reqctx->init_iv, req->iv, IV); in process_cipher()
2388 reqctx->scratch_pad = reqctx->iv + IV; in chcr_aead_common_init()
2510 chcr_req->sec_cpl.pldlen = htonl(req->assoclen + IV + req->cryptlen); in create_authenc_wr()
2512 null ? 0 : 1 + IV, in create_authenc_wr()
2513 null ? 0 : IV + req->assoclen, in create_authenc_wr()
2514 req->assoclen + IV + 1, in create_authenc_wr()
2518 null ? 0 : req->assoclen + IV + 1, in create_authenc_wr()
2529 IV >> 1); in create_authenc_wr()
2547 ulptx = (struct ulptx_sgl *)(ivptr + IV); in create_authenc_wr()
2556 memcpy(ivptr, req->iv, IV); in create_authenc_wr()
2561 temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV + in create_authenc_wr()
2599 reqctx->iv_dma = dma_map_single(dev, reqctx->iv, (IV + reqctx->b0_len), in chcr_aead_dma_map()
2604 reqctx->b0_dma = reqctx->iv_dma + IV; in chcr_aead_dma_map()
2632 dma_unmap_single(dev, reqctx->iv_dma, IV, DMA_BIDIRECTIONAL); in chcr_aead_dma_map()
2661 dma_unmap_single(dev, reqctx->iv_dma, (IV + reqctx->b0_len), in chcr_aead_dma_unmap()
2716 dsgl_walk_add_page(&dsgl_walk, IV + reqctx->b0_len, reqctx->iv_dma); in chcr_add_aead_dst_ent()
2731 memcpy(buf, reqctx->iv, IV); in chcr_add_cipher_src_ent()
2732 buf += IV; in chcr_add_cipher_src_ent()
2977 (req->assoclen + IV + 1 + ccm_xtra) : 0; in fill_sec_cpl_for_aead()
2987 htonl(req->assoclen + IV + req->cryptlen + ccm_xtra); in fill_sec_cpl_for_aead()
2990 1 + IV, IV + assoclen + ccm_xtra, in fill_sec_cpl_for_aead()
2991 req->assoclen + IV + 1 + ccm_xtra, 0); in fill_sec_cpl_for_aead()
3000 aeadctx->hmac_ctrl, IV >> 1); in fill_sec_cpl_for_aead()
3100 ulptx = (struct ulptx_sgl *)(ivptr + IV); in create_aead_ccm_wr()
3108 temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV + in create_aead_ccm_wr()
3187 htonl(req->assoclen + IV + req->cryptlen); in create_gcm_wr()
3189 assoclen ? 1 + IV : 0, in create_gcm_wr()
3190 assoclen ? IV + assoclen : 0, in create_gcm_wr()
3191 req->assoclen + IV + 1, 0); in create_gcm_wr()
3193 FILL_SEC_CPL_AUTHINSERT(0, req->assoclen + IV + 1, in create_gcm_wr()
3200 aeadctx->hmac_ctrl, IV >> 1); in create_gcm_wr()
3225 temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV + in create_gcm_wr()