/Linux-v5.15/drivers/dma/sh/ |
D | rcar-dmac.c | 81 unsigned int nchunks; member 394 chan->index, desc, desc->nchunks, &desc->hwdescs.dma); in rcar_dmac_chan_start_xfer() 408 RCAR_DMACHCRB_DCNT(desc->nchunks - 1) | in rcar_dmac_chan_start_xfer() 765 rcar_dmac_realloc_hwdesc(chan, desc, desc->nchunks * sizeof(*hwdesc)); in rcar_dmac_fill_hwdesc() 936 unsigned int nchunks = 0; in rcar_dmac_chan_prep_sg() local 1026 nchunks++; in rcar_dmac_chan_prep_sg() 1030 desc->nchunks = nchunks; in rcar_dmac_chan_prep_sg() 1041 desc->hwdescs.use = !cross_boundary && nchunks > 1; in rcar_dmac_chan_prep_sg() 1401 dptr = desc->nchunks; in rcar_dmac_chan_get_residue() 1403 WARN_ON(dptr >= desc->nchunks); in rcar_dmac_chan_get_residue() [all …]
|
/Linux-v5.15/net/sunrpc/xprtrdma/ |
D | rpc_rdma.c | 410 int nsegs, nchunks; in rpcrdma_encode_write_list() local 430 nchunks = 0; in rpcrdma_encode_write_list() 442 nchunks++; in rpcrdma_encode_write_list() 447 *segcount = cpu_to_be32(nchunks); in rpcrdma_encode_write_list() 475 int nsegs, nchunks; in rpcrdma_encode_reply_chunk() local 496 nchunks = 0; in rpcrdma_encode_reply_chunk() 508 nchunks++; in rpcrdma_encode_reply_chunk() 513 *segcount = cpu_to_be32(nchunks); in rpcrdma_encode_reply_chunk()
|
/Linux-v5.15/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_dev_api.h | 462 u32 nchunks; member 484 u32 nchunks; member 516 u32 nchunks; member
|
D | pvrdma_mr.c | 171 cmd->nchunks = npages; in pvrdma_reg_user_mr() 243 cmd->nchunks = max_num_sg; in pvrdma_alloc_mr()
|
D | pvrdma_srq.c | 176 cmd->nchunks = srq->npages; in pvrdma_create_srq()
|
D | pvrdma_cq.c | 184 cmd->nchunks = npages; in pvrdma_create_cq()
|
/Linux-v5.15/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_cs.c | 146 p->nchunks = cs->in.num_chunks; in amdgpu_cs_parser_init() 147 p->chunks = kvmalloc_array(p->nchunks, sizeof(struct amdgpu_cs_chunk), in amdgpu_cs_parser_init() 154 for (i = 0; i < p->nchunks; i++) { in amdgpu_cs_parser_init() 249 i = p->nchunks - 1; in amdgpu_cs_parser_init() 255 p->nchunks = 0; in amdgpu_cs_parser_init() 710 for (i = 0; i < parser->nchunks; i++) in amdgpu_cs_parser_fini() 737 for (i = 0, j = 0; i < p->nchunks && j < p->job->num_ibs; i++) { in amdgpu_cs_vm_handling() 880 for (i = 0, j = 0; i < parser->nchunks && j < parser->job->num_ibs; i++) { in amdgpu_cs_ib_fill() 1153 for (i = 0; i < p->nchunks; ++i) { in amdgpu_cs_dependencies()
|
D | amdgpu.h | 531 unsigned nchunks; member
|
/Linux-v5.15/net/sctp/ |
D | auth.c | 759 __u16 nchunks; in sctp_auth_ep_add_chunkid() local 768 nchunks = param_len - sizeof(struct sctp_paramhdr); in sctp_auth_ep_add_chunkid() 769 if (nchunks == SCTP_NUM_CHUNK_TYPES) in sctp_auth_ep_add_chunkid() 772 p->chunks[nchunks] = chunk_id; in sctp_auth_ep_add_chunkid()
|
/Linux-v5.15/drivers/mtd/nand/raw/ |
D | marvell_nand.c | 259 int nchunks; member 274 .nchunks = nc, \ 1251 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page_raw() 1310 else if (chunk < lt->nchunks - 1) in marvell_nfc_hw_ecc_bch_read_chunk() 1368 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page() 1410 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page() 1503 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_write_page_raw() 1563 if (lt->nchunks == 1) in marvell_nfc_hw_ecc_bch_write_chunk() 1573 } else if (chunk < lt->nchunks - 1) { in marvell_nfc_hw_ecc_bch_write_chunk() 1580 if (chunk == lt->nchunks - 1) in marvell_nfc_hw_ecc_bch_write_chunk() [all …]
|
D | sunxi_nand.c | 351 int chunksize, int nchunks, in sunxi_nfc_dma_op_prepare() argument 365 sg_init_one(sg, buf, nchunks * chunksize); in sunxi_nfc_dma_op_prepare() 380 writel(nchunks, nfc->regs + NFC_REG_SECTOR_NUM); in sunxi_nfc_dma_op_prepare() 386 writel(chunksize * nchunks, nfc->regs + NFC_REG_MDMA_CNT); in sunxi_nfc_dma_op_prepare() 916 int nchunks) in sunxi_nfc_hw_ecc_read_chunks_dma() argument 931 ret = sunxi_nfc_dma_op_prepare(nfc, buf, ecc->size, nchunks, in sunxi_nfc_hw_ecc_read_chunks_dma() 967 for (i = 0; i < nchunks; i++) { in sunxi_nfc_hw_ecc_read_chunks_dma() 999 for (i = 0; i < nchunks; i++) { in sunxi_nfc_hw_ecc_read_chunks_dma() 1201 int nchunks = DIV_ROUND_UP(data_offs + readlen, nand->ecc.size); in sunxi_nfc_hw_ecc_read_subpage_dma() local 1208 ret = sunxi_nfc_hw_ecc_read_chunks_dma(nand, buf, false, page, nchunks); in sunxi_nfc_hw_ecc_read_subpage_dma()
|
/Linux-v5.15/drivers/crypto/cavium/cpt/ |
D | cptvf.h | 54 u32 nchunks; /* Number of command chunks */ member
|
D | cptvf_main.c | 195 queue->nchunks = 0; in free_command_queues() 242 i, queue->nchunks); in alloc_command_queues() 248 if (queue->nchunks == 0) { in alloc_command_queues() 257 queue->nchunks++; in alloc_command_queues()
|
/Linux-v5.15/drivers/gpu/drm/radeon/ |
D | radeon_cs.c | 301 p->nchunks = cs->num_chunks; in radeon_cs_parser_init() 302 p->chunks = kvcalloc(p->nchunks, sizeof(struct radeon_cs_chunk), GFP_KERNEL); in radeon_cs_parser_init() 306 for (i = 0; i < p->nchunks; i++) { in radeon_cs_parser_init() 453 for (i = 0; i < parser->nchunks; i++) in radeon_cs_parser_fini()
|
D | radeon.h | 1061 unsigned nchunks; member
|
/Linux-v5.15/drivers/vhost/ |
D | vdpa.c | 694 unsigned long lock_limit, sz2pin, nchunks, i; in vhost_vdpa_pa_map() local 723 nchunks = 0; in vhost_vdpa_pa_map() 738 nchunks++; in vhost_vdpa_pa_map() 769 nchunks = 0; in vhost_vdpa_pa_map() 784 if (nchunks) { in vhost_vdpa_pa_map()
|
/Linux-v5.15/sound/drivers/vx/ |
D | vx_pcm.c | 677 struct vx_pipe *pipe, int nchunks) in vx_pcm_playback_transfer() argument 684 for (i = 0; i < nchunks; i++) { in vx_pcm_playback_transfer()
|