Lines Matching refs:chunksize
147 unsigned int chunksize, u32 stat_pad_len) in spu_ablkcipher_rx_sg_create() argument
172 rctx->dst_nents, chunksize); in spu_ablkcipher_rx_sg_create()
173 if (datalen < chunksize) { in spu_ablkcipher_rx_sg_create()
175 __func__, chunksize, datalen); in spu_ablkcipher_rx_sg_create()
214 u8 tx_frag_num, unsigned int chunksize, u32 pad_len) in spu_ablkcipher_tx_sg_create() argument
240 rctx->src_nents, chunksize); in spu_ablkcipher_tx_sg_create()
241 if (unlikely(datalen < chunksize)) { in spu_ablkcipher_tx_sg_create()
323 unsigned int chunksize = 0; /* Num bytes of request to submit */ in handle_ablkcipher_req() local
355 chunksize = ctx->max_payload; in handle_ablkcipher_req()
357 chunksize = remaining; in handle_ablkcipher_req()
359 rctx->src_sent += chunksize; in handle_ablkcipher_req()
363 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, chunksize); in handle_ablkcipher_req()
364 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_ablkcipher_req()
401 add_to_ctr(rctx->msg_buf.iv_ctr, chunksize >> 4); in handle_ablkcipher_req()
433 rctx->src_sent, chunk_start, remaining, chunksize); in handle_ablkcipher_req()
446 &cipher_parms, update_key, chunksize); in handle_ablkcipher_req()
448 atomic64_add(chunksize, &iproc_priv.bytes_out); in handle_ablkcipher_req()
450 stat_pad_len = spu->spu_wordalign_padlen(chunksize); in handle_ablkcipher_req()
464 dump_sg(rctx->src_sg, rctx->src_skip, chunksize); in handle_ablkcipher_req()
482 err = spu_ablkcipher_rx_sg_create(mssg, rctx, rx_frag_num, chunksize, in handle_ablkcipher_req()
496 err = spu_ablkcipher_tx_sg_create(mssg, rctx, tx_frag_num, chunksize, in handle_ablkcipher_req()
713 unsigned int chunksize = 0; /* length of hash carry + new data */ in handle_ahash_req() local
774 chunksize = nbytes_to_hash; in handle_ahash_req()
776 (chunksize > ctx->max_payload)) in handle_ahash_req()
777 chunksize = ctx->max_payload; in handle_ahash_req()
788 rem = chunksize % blocksize; /* remainder */ in handle_ahash_req()
791 chunksize -= rem; in handle_ahash_req()
792 if (chunksize == 0) { in handle_ahash_req()
813 new_data_len = chunksize - local_nbuf; in handle_ahash_req()
830 rctx->total_sent += chunksize; in handle_ahash_req()
837 chunksize, in handle_ahash_req()
850 atomic64_add(chunksize, &iproc_priv.bytes_out); in handle_ahash_req()
860 flow_log("chunk_start: %u chunk_size: %u\n", chunk_start, chunksize); in handle_ahash_req()
881 data_pad_len = spu->spu_gcm_ccm_pad_len(ctx->cipher.mode, chunksize); in handle_ahash_req()
1218 unsigned int chunksize, in spu_aead_tx_sg_create() argument
1260 datalen = chunksize; in spu_aead_tx_sg_create()
1261 if ((chunksize > ctx->digestsize) && incl_icv) in spu_aead_tx_sg_create()
1315 unsigned int chunksize; in handle_aead_req() local
1336 chunksize = rctx->total_todo; in handle_aead_req()
1338 flow_log("%s: chunksize %u\n", __func__, chunksize); in handle_aead_req()
1392 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, chunksize); in handle_aead_req()
1393 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_aead_req()
1400 rctx->total_sent = chunksize; in handle_aead_req()
1401 rctx->src_sent = chunksize; in handle_aead_req()
1420 chunksize); in handle_aead_req()
1438 chunksize - digestsize); in handle_aead_req()
1442 chunksize, rctx->is_encrypt, in handle_aead_req()
1455 aead_parms.assoc_size + chunksize - in handle_aead_req()
1460 aead_parms.assoc_size + chunksize); in handle_aead_req()
1474 atomic64_add(chunksize, &iproc_priv.bytes_out); in handle_aead_req()
1476 flow_log("%s()-sent chunksize:%u\n", __func__, chunksize); in handle_aead_req()
1484 &aead_parms, chunksize); in handle_aead_req()
1488 chunksize, aead_parms.aad_pad_len, in handle_aead_req()
1509 dump_sg(rctx->src_sg, rctx->src_skip, chunksize); in handle_aead_req()
1522 resp_len = chunksize; in handle_aead_req()
1561 assoc_nents, aead_parms.iv_len, chunksize, in handle_aead_req()