/Linux-v5.4/drivers/dma/sh/ |
D | rcar-dmac.c | 81 unsigned int nchunks; member 356 chan->index, desc, desc->nchunks, &desc->hwdescs.dma); in rcar_dmac_chan_start_xfer() 370 RCAR_DMACHCRB_DCNT(desc->nchunks - 1) | in rcar_dmac_chan_start_xfer() 727 rcar_dmac_realloc_hwdesc(chan, desc, desc->nchunks * sizeof(*hwdesc)); in rcar_dmac_fill_hwdesc() 902 unsigned int nchunks = 0; in rcar_dmac_chan_prep_sg() local 992 nchunks++; in rcar_dmac_chan_prep_sg() 996 desc->nchunks = nchunks; in rcar_dmac_chan_prep_sg() 1007 desc->hwdescs.use = !cross_boundary && nchunks > 1; in rcar_dmac_chan_prep_sg() 1367 dptr = desc->nchunks; in rcar_dmac_chan_get_residue() 1369 WARN_ON(dptr >= desc->nchunks); in rcar_dmac_chan_get_residue() [all …]
|
/Linux-v5.4/net/sunrpc/xprtrdma/ |
D | rpc_rdma.c | 448 int nsegs, nchunks; in rpcrdma_encode_write_list() local 468 nchunks = 0; in rpcrdma_encode_write_list() 480 nchunks++; in rpcrdma_encode_write_list() 485 *segcount = cpu_to_be32(nchunks); in rpcrdma_encode_write_list() 511 int nsegs, nchunks; in rpcrdma_encode_reply_chunk() local 529 nchunks = 0; in rpcrdma_encode_reply_chunk() 541 nchunks++; in rpcrdma_encode_reply_chunk() 546 *segcount = cpu_to_be32(nchunks); in rpcrdma_encode_reply_chunk()
|
/Linux-v5.4/lib/ |
D | bitmap.c | 380 int c, old_c, totaldigits, ndigits, nchunks, nbits; in __bitmap_parse() local 386 nchunks = nbits = totaldigits = c = 0; in __bitmap_parse() 432 if (nchunks == 0 && chunk == 0) in __bitmap_parse() 437 nchunks++; in __bitmap_parse() 438 nbits += (nchunks == 1) ? nbits_to_hold_value(chunk) : CHUNKSZ; in __bitmap_parse()
|
/Linux-v5.4/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_dev_api.h | 461 u32 nchunks; member 483 u32 nchunks; member 515 u32 nchunks; member
|
D | pvrdma_mr.c | 171 cmd->nchunks = npages; in pvrdma_reg_user_mr() 243 cmd->nchunks = max_num_sg; in pvrdma_alloc_mr()
|
D | pvrdma_srq.c | 176 cmd->nchunks = srq->npages; in pvrdma_create_srq()
|
D | pvrdma_cq.c | 181 cmd->nchunks = npages; in pvrdma_create_cq()
|
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_cs.c | 145 p->nchunks = cs->in.num_chunks; in amdgpu_cs_parser_init() 146 p->chunks = kmalloc_array(p->nchunks, sizeof(struct amdgpu_cs_chunk), in amdgpu_cs_parser_init() 153 for (i = 0; i < p->nchunks; i++) { in amdgpu_cs_parser_init() 248 i = p->nchunks - 1; in amdgpu_cs_parser_init() 254 p->nchunks = 0; in amdgpu_cs_parser_init() 772 for (i = 0; i < parser->nchunks; i++) in amdgpu_cs_parser_fini() 799 for (i = 0, j = 0; i < p->nchunks && j < p->job->num_ibs; i++) { in amdgpu_cs_vm_handling() 949 for (i = 0, j = 0; i < parser->nchunks && j < parser->job->num_ibs; i++) { in amdgpu_cs_ib_fill() 1216 for (i = 0; i < p->nchunks; ++i) { in amdgpu_cs_dependencies()
|
D | amdgpu.h | 462 unsigned nchunks; member
|
/Linux-v5.4/net/sctp/ |
D | auth.c | 764 __u16 nchunks; in sctp_auth_ep_add_chunkid() local 773 nchunks = param_len - sizeof(struct sctp_paramhdr); in sctp_auth_ep_add_chunkid() 774 if (nchunks == SCTP_NUM_CHUNK_TYPES) in sctp_auth_ep_add_chunkid() 777 p->chunks[nchunks] = chunk_id; in sctp_auth_ep_add_chunkid()
|
/Linux-v5.4/drivers/mtd/nand/raw/ |
D | marvell_nand.c | 257 int nchunks; member 272 .nchunks = nc, \ 1212 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page_raw() 1271 else if (chunk < lt->nchunks - 1) in marvell_nfc_hw_ecc_bch_read_chunk() 1329 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page() 1372 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page() 1465 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_write_page_raw() 1525 if (lt->nchunks == 1) in marvell_nfc_hw_ecc_bch_write_chunk() 1535 } else if (chunk < lt->nchunks - 1) { in marvell_nfc_hw_ecc_bch_write_chunk() 1542 if (chunk == lt->nchunks - 1) in marvell_nfc_hw_ecc_bch_write_chunk() [all …]
|
D | sunxi_nand.c | 347 int chunksize, int nchunks, in sunxi_nfc_dma_op_prepare() argument 361 sg_init_one(sg, buf, nchunks * chunksize); in sunxi_nfc_dma_op_prepare() 374 writel(nchunks, nfc->regs + NFC_REG_SECTOR_NUM); in sunxi_nfc_dma_op_prepare() 377 writel(chunksize * nchunks, nfc->regs + NFC_REG_MDMA_CNT); in sunxi_nfc_dma_op_prepare() 905 int nchunks) in sunxi_nfc_hw_ecc_read_chunks_dma() argument 920 ret = sunxi_nfc_dma_op_prepare(nfc, buf, ecc->size, nchunks, in sunxi_nfc_hw_ecc_read_chunks_dma() 951 for (i = 0; i < nchunks; i++) { in sunxi_nfc_hw_ecc_read_chunks_dma() 983 for (i = 0; i < nchunks; i++) { in sunxi_nfc_hw_ecc_read_chunks_dma() 1185 int nchunks = DIV_ROUND_UP(data_offs + readlen, nand->ecc.size); in sunxi_nfc_hw_ecc_read_subpage_dma() local 1192 ret = sunxi_nfc_hw_ecc_read_chunks_dma(nand, buf, false, page, nchunks); in sunxi_nfc_hw_ecc_read_subpage_dma()
|
/Linux-v5.4/drivers/crypto/cavium/cpt/ |
D | cptvf.h | 54 u32 nchunks; /* Number of command chunks */ member
|
D | cptvf_main.c | 195 queue->nchunks = 0; in free_command_queues() 242 i, queue->nchunks); in alloc_command_queues() 248 if (queue->nchunks == 0) { in alloc_command_queues() 257 queue->nchunks++; in alloc_command_queues()
|
/Linux-v5.4/drivers/gpu/drm/radeon/ |
D | radeon_cs.c | 302 p->nchunks = cs->num_chunks; in radeon_cs_parser_init() 303 p->chunks = kcalloc(p->nchunks, sizeof(struct radeon_cs_chunk), GFP_KERNEL); in radeon_cs_parser_init() 307 for (i = 0; i < p->nchunks; i++) { in radeon_cs_parser_init() 452 for (i = 0; i < parser->nchunks; i++) in radeon_cs_parser_fini()
|
D | radeon.h | 1062 unsigned nchunks; member
|
/Linux-v5.4/sound/drivers/vx/ |
D | vx_pcm.c | 677 struct vx_pipe *pipe, int nchunks) in vx_pcm_playback_transfer() argument 684 for (i = 0; i < nchunks; i++) { in vx_pcm_playback_transfer()
|