/Linux-v4.19/net/9p/ |
D | trans_rdma.c | 167 struct p9_trans_rdma *rdma = clnt->trans; in p9_rdma_show_options() local 169 if (rdma->port != P9_PORT) in p9_rdma_show_options() 170 seq_printf(m, ",port=%u", rdma->port); in p9_rdma_show_options() 171 if (rdma->sq_depth != P9_RDMA_SQ_DEPTH) in p9_rdma_show_options() 172 seq_printf(m, ",sq=%u", rdma->sq_depth); in p9_rdma_show_options() 173 if (rdma->rq_depth != P9_RDMA_RQ_DEPTH) in p9_rdma_show_options() 174 seq_printf(m, ",rq=%u", rdma->rq_depth); in p9_rdma_show_options() 175 if (rdma->timeout != P9_RDMA_TIMEOUT) in p9_rdma_show_options() 176 seq_printf(m, ",timeout=%lu", rdma->timeout); in p9_rdma_show_options() 177 if (rdma->privport) in p9_rdma_show_options() [all …]
|
/Linux-v4.19/net/sunrpc/xprtrdma/ |
D | svc_rdma_sendto.c | 127 svc_rdma_send_ctxt_alloc(struct svcxprt_rdma *rdma) in svc_rdma_send_ctxt_alloc() argument 136 size += rdma->sc_max_send_sges * sizeof(struct ib_sge); in svc_rdma_send_ctxt_alloc() 140 buffer = kmalloc(rdma->sc_max_req_size, GFP_KERNEL); in svc_rdma_send_ctxt_alloc() 143 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_send_ctxt_alloc() 144 rdma->sc_max_req_size, DMA_TO_DEVICE); in svc_rdma_send_ctxt_alloc() 145 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_send_ctxt_alloc() 156 for (i = 0; i < rdma->sc_max_send_sges; i++) in svc_rdma_send_ctxt_alloc() 157 ctxt->sc_sges[i].lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_send_ctxt_alloc() 173 void svc_rdma_send_ctxts_destroy(struct svcxprt_rdma *rdma) in svc_rdma_send_ctxts_destroy() argument 177 while ((ctxt = svc_rdma_next_send_ctxt(&rdma->sc_send_ctxts))) { in svc_rdma_send_ctxts_destroy() [all …]
|
D | svc_rdma_recvfrom.c | 121 svc_rdma_recv_ctxt_alloc(struct svcxprt_rdma *rdma) in svc_rdma_recv_ctxt_alloc() argument 130 buffer = kmalloc(rdma->sc_max_req_size, GFP_KERNEL); in svc_rdma_recv_ctxt_alloc() 133 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_recv_ctxt_alloc() 134 rdma->sc_max_req_size, DMA_FROM_DEVICE); in svc_rdma_recv_ctxt_alloc() 135 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_recv_ctxt_alloc() 144 ctxt->rc_recv_sge.length = rdma->sc_max_req_size; in svc_rdma_recv_ctxt_alloc() 145 ctxt->rc_recv_sge.lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_recv_ctxt_alloc() 158 static void svc_rdma_recv_ctxt_destroy(struct svcxprt_rdma *rdma, in svc_rdma_recv_ctxt_destroy() argument 161 ib_dma_unmap_single(rdma->sc_pd->device, ctxt->rc_recv_sge.addr, in svc_rdma_recv_ctxt_destroy() 172 void svc_rdma_recv_ctxts_destroy(struct svcxprt_rdma *rdma) in svc_rdma_recv_ctxts_destroy() argument [all …]
|
D | svc_rdma_rw.c | 55 svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) in svc_rdma_get_rw_ctxt() argument 59 spin_lock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 61 ctxt = svc_rdma_next_ctxt(&rdma->sc_rw_ctxts); in svc_rdma_get_rw_ctxt() 64 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 66 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 85 static void svc_rdma_put_rw_ctxt(struct svcxprt_rdma *rdma, in svc_rdma_put_rw_ctxt() argument 90 spin_lock(&rdma->sc_rw_ctxt_lock); in svc_rdma_put_rw_ctxt() 91 list_add(&ctxt->rw_list, &rdma->sc_rw_ctxts); in svc_rdma_put_rw_ctxt() 92 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_put_rw_ctxt() 100 void svc_rdma_destroy_rw_ctxts(struct svcxprt_rdma *rdma) in svc_rdma_destroy_rw_ctxts() argument [all …]
|
D | svc_rdma_transport.c | 152 struct svcxprt_rdma *rdma = in svc_rdma_bc_free() local 157 kfree(rdma); in svc_rdma_bc_free() 324 struct svcxprt_rdma *rdma = cma_id->context; in rdma_cma_handler() local 325 struct svc_xprt *xprt = &rdma->sc_xprt; in rdma_cma_handler() 335 clear_bit(RDMAXPRT_CONN_PENDING, &rdma->sc_flags); in rdma_cma_handler() 633 struct svcxprt_rdma *rdma = in svc_rdma_detach() local 637 rdma_disconnect(rdma->sc_cm_id); in svc_rdma_detach() 642 struct svcxprt_rdma *rdma = in __svc_rdma_free() local 644 struct svc_xprt *xprt = &rdma->sc_xprt; in __svc_rdma_free() 648 if (rdma->sc_qp && !IS_ERR(rdma->sc_qp)) in __svc_rdma_free() [all …]
|
D | svc_rdma_backchannel.c | 117 static int svc_rdma_bc_sendto(struct svcxprt_rdma *rdma, in svc_rdma_bc_sendto() argument 123 ret = svc_rdma_map_reply_msg(rdma, ctxt, &rqst->rq_snd_buf, NULL); in svc_rdma_bc_sendto() 132 return svc_rdma_send(rdma, &ctxt->sc_send_wr); in svc_rdma_bc_sendto() 175 rpcrdma_bc_send_request(struct svcxprt_rdma *rdma, struct rpc_rqst *rqst) in rpcrdma_bc_send_request() argument 183 ctxt = svc_rdma_send_ctxt_get(rdma); in rpcrdma_bc_send_request() 195 svc_rdma_sync_reply_hdr(rdma, ctxt, RPCRDMA_HDRLEN_MIN); in rpcrdma_bc_send_request() 201 rc = svc_rdma_bc_sendto(rdma, rqst, ctxt); in rpcrdma_bc_send_request() 203 svc_rdma_send_ctxt_put(rdma, ctxt); in rpcrdma_bc_send_request() 222 struct svcxprt_rdma *rdma; in xprt_rdma_bc_send_request() local 236 rdma = container_of(sxprt, struct svcxprt_rdma, sc_xprt); in xprt_rdma_bc_send_request() [all …]
|
/Linux-v4.19/include/linux/sunrpc/ |
D | svc_rdma.h | 157 extern void svc_rdma_recv_ctxts_destroy(struct svcxprt_rdma *rdma); 158 extern bool svc_rdma_post_recvs(struct svcxprt_rdma *rdma); 159 extern void svc_rdma_recv_ctxt_put(struct svcxprt_rdma *rdma, 161 extern void svc_rdma_flush_recv_queues(struct svcxprt_rdma *rdma); 165 extern void svc_rdma_destroy_rw_ctxts(struct svcxprt_rdma *rdma); 166 extern int svc_rdma_recv_read_chunk(struct svcxprt_rdma *rdma, 169 extern int svc_rdma_send_write_chunk(struct svcxprt_rdma *rdma, 171 extern int svc_rdma_send_reply_chunk(struct svcxprt_rdma *rdma, 176 extern void svc_rdma_send_ctxts_destroy(struct svcxprt_rdma *rdma); 178 svc_rdma_send_ctxt_get(struct svcxprt_rdma *rdma); [all …]
|
/Linux-v4.19/Documentation/cgroup-v1/ |
D | rdma.txt | 27 Currently user space applications can easily take away all the rdma verb 30 rdma resources. This can leads to service unavailability. 33 of processes can be limited. Through this controller different rdma 42 by rdma cgroup, which can be extended later if required. 51 of the child processes which shares the address space, rdma resources are 55 rdma resources. Linking resources around css also ensures that cgroups can be 59 Whenever RDMA resource charging occurs, owner rdma cgroup is returned to 60 the caller. Same rdma cgroup should be passed while uncharging the resource. 80 IB stack honors limits enforced by the rdma controller. When application 85 Following resources can be accounted by rdma controller. [all …]
|
/Linux-v4.19/drivers/gpu/drm/mediatek/ |
D | mtk_disp_rdma.c | 55 #define RDMA_FIFO_SIZE(rdma) ((rdma)->data->fifo_size) argument 83 struct mtk_ddp_comp *rdma = &priv->ddp_comp; in mtk_disp_rdma_irq_handler() local 86 writel(0x0, rdma->regs + DISP_REG_RDMA_INT_STATUS); in mtk_disp_rdma_irq_handler() 91 mtk_crtc_ddp_irq(priv->crtc, rdma); in mtk_disp_rdma_irq_handler() 108 struct mtk_disp_rdma *rdma = comp_to_rdma(comp); in mtk_rdma_enable_vblank() local 110 rdma->crtc = crtc; in mtk_rdma_enable_vblank() 117 struct mtk_disp_rdma *rdma = comp_to_rdma(comp); in mtk_rdma_disable_vblank() local 119 rdma->crtc = NULL; in mtk_rdma_disable_vblank() 140 struct mtk_disp_rdma *rdma = comp_to_rdma(comp); in mtk_rdma_config() local 153 RDMA_FIFO_PSEUDO_SIZE(RDMA_FIFO_SIZE(rdma)) | in mtk_rdma_config() [all …]
|
/Linux-v4.19/Documentation/ABI/stable/ |
D | sysfs-driver-ib_srp | 4 Contact: linux-rdma@vger.kernel.org 74 Contact: linux-rdma@vger.kernel.org 80 Contact: linux-rdma@vger.kernel.org 86 Contact: linux-rdma@vger.kernel.org 94 Contact: linux-rdma@vger.kernel.org 101 Contact: linux-rdma@vger.kernel.org 108 Contact: linux-rdma@vger.kernel.org 114 Contact: linux-rdma@vger.kernel.org 121 Contact: linux-rdma@vger.kernel.org 128 Contact: linux-rdma@vger.kernel.org [all …]
|
D | sysfs-class-infiniband | 9 Contact: linux-rdma@vger.kernel.org 22 Contact: linux-rdma@vger.kernel.org 32 Contact: linux-rdma@vger.kernel.org 47 Contact: linux-rdma@vger.kernel.org 76 Contact: linux-rdma@vger.kernel.org 104 Contact: linux-rdma@vger.kernel.org 202 Contact: linux-rdma@vger.kernel.org 221 Contact: linux-rdma@vger.kernel.org 228 Contact: linux-rdma@vger.kernel.org 240 Contact: linux-rdma@vger.kernel.org [all …]
|
D | sysfs-transport-srp | 4 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 11 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 20 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 36 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 50 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org
|
/Linux-v4.19/drivers/net/ethernet/seeq/ |
D | sgiseeq.c | 68 volatile struct hpc_dma_desc rdma; member 208 sp->rx_desc[i].rdma.pbuf = dma_addr; in seeq_init_ring() 210 sp->rx_desc[i].rdma.cntinfo = RCNTINFO_INIT; in seeq_init_ring() 213 sp->rx_desc[i - 1].rdma.cntinfo |= HPCDMA_EOR; in seeq_init_ring() 258 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 259 r[i].rdma.pnext); in sgiseeq_dump_rings() 262 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 263 r[i].rdma.pnext); in sgiseeq_dump_rings() 350 while (!(rd->rdma.cntinfo & HPCDMA_OWN)) { in sgiseeq_rx() 351 len = PKT_BUF_SZ - (rd->rdma.cntinfo & HPCDMA_BCNT) - 3; in sgiseeq_rx() [all …]
|
/Linux-v4.19/drivers/nvme/target/ |
D | Makefile | 5 obj-$(CONFIG_NVME_TARGET_RDMA) += nvmet-rdma.o 12 nvmet-rdma-y += rdma.o
|
/Linux-v4.19/drivers/nvme/host/ |
D | Makefile | 8 obj-$(CONFIG_NVME_RDMA) += nvme-rdma.o 21 nvme-rdma-y += rdma.o
|
/Linux-v4.19/Documentation/devicetree/bindings/media/ |
D | mediatek-mdp.txt | 12 "mediatek,mt8173-mdp-rdma" - read DMA 24 "mediatek,mt8173-mdp-rdma" 35 mdp_rdma0: rdma@14001000 { 36 compatible = "mediatek,mt8173-mdp-rdma"; 47 mdp_rdma1: rdma@14002000 { 48 compatible = "mediatek,mt8173-mdp-rdma";
|
/Linux-v4.19/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_offload.c | 281 struct rdma_info *rdma = data; in cxgb_rdma_ctl() local 284 rdma->udbell_physbase = pci_resource_start(pdev, 2); in cxgb_rdma_ctl() 285 rdma->udbell_len = pci_resource_len(pdev, 2); in cxgb_rdma_ctl() 286 rdma->tpt_base = in cxgb_rdma_ctl() 288 rdma->tpt_top = t3_read_reg(adapter, A_ULPTX_TPT_ULIMIT); in cxgb_rdma_ctl() 289 rdma->pbl_base = in cxgb_rdma_ctl() 291 rdma->pbl_top = t3_read_reg(adapter, A_ULPTX_PBL_ULIMIT); in cxgb_rdma_ctl() 292 rdma->rqt_base = t3_read_reg(adapter, A_ULPRX_RQ_LLIMIT); in cxgb_rdma_ctl() 293 rdma->rqt_top = t3_read_reg(adapter, A_ULPRX_RQ_ULIMIT); in cxgb_rdma_ctl() 294 rdma->kdb_addr = adapter->regs + A_SG_KDOORBELL; in cxgb_rdma_ctl() [all …]
|
/Linux-v4.19/drivers/macintosh/ |
D | rack-meter.c | 148 struct rackmeter_dma *rdma = rm->dma_buf_v; in rackmeter_do_pause() local 157 memset(rdma->buf1, 0, sizeof(rdma->buf1)); in rackmeter_do_pause() 158 memset(rdma->buf2, 0, sizeof(rdma->buf2)); in rackmeter_do_pause() 373 struct resource ri2s, rdma; in rackmeter_probe() local 432 of_address_to_resource(i2s, 1, &rdma)) { in rackmeter_probe() 442 pr_debug(" dma @0x%08x\n", (unsigned int)rdma.start); in rackmeter_probe() 476 rm->dma_regs = ioremap(rdma.start, 0x100); in rackmeter_probe()
|
/Linux-v4.19/drivers/infiniband/ |
D | Kconfig | 23 need libibumad from rdma-core 24 <https://github.com/linux-rdma/rdma-core>. 36 rdma-core <https://github.com/linux-rdma/rdma-core>.
|
/Linux-v4.19/Documentation/devicetree/bindings/display/mediatek/ |
D | mediatek,disp.txt | 31 "mediatek,<chip>-disp-rdma" - read DMA / line buffer 56 "mediatek,<chip>-disp-rdma" 94 rdma0: rdma@1400e000 { 95 compatible = "mediatek,mt8173-disp-rdma"; 104 rdma1: rdma@1400f000 { 105 compatible = "mediatek,mt8173-disp-rdma"; 114 rdma2: rdma@14010000 { 115 compatible = "mediatek,mt8173-disp-rdma";
|
/Linux-v4.19/include/trace/events/ |
D | rpcrdma.h | 1184 const struct svcxprt_rdma *rdma, 1188 TP_ARGS(rdma, page), 1192 __string(device, rdma->sc_cm_id->device->name) 1193 __string(addr, rdma->sc_xprt.xpt_remotebuf) 1198 __assign_str(device, rdma->sc_cm_id->device->name); 1199 __assign_str(addr, rdma->sc_xprt.xpt_remotebuf); 1209 const struct svcxprt_rdma *rdma, 1213 TP_ARGS(rdma, status), 1217 __string(device, rdma->sc_cm_id->device->name) 1218 __string(addr, rdma->sc_xprt.xpt_remotebuf) [all …]
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/ |
D | eq.c | 251 be32_to_cpu(pf_eqe->rdma.pftype_token) >> 24; in eq_pf_process() 253 be32_to_cpu(pf_eqe->rdma.pftype_token) & in eq_pf_process() 255 pfault->rdma.r_key = in eq_pf_process() 256 be32_to_cpu(pf_eqe->rdma.r_key); in eq_pf_process() 257 pfault->rdma.packet_size = in eq_pf_process() 258 be16_to_cpu(pf_eqe->rdma.packet_length); in eq_pf_process() 259 pfault->rdma.rdma_op_len = in eq_pf_process() 260 be32_to_cpu(pf_eqe->rdma.rdma_op_len); in eq_pf_process() 261 pfault->rdma.rdma_va = in eq_pf_process() 262 be64_to_cpu(pf_eqe->rdma.rdma_va); in eq_pf_process() [all …]
|
/Linux-v4.19/net/rds/ |
D | send.c | 269 (rm->rdma.op_active && in rds_send_xmit() 298 if (rm->rdma.op_active && !cp->cp_xmit_rdma_sent) { in rds_send_xmit() 299 rm->m_final_op = &rm->rdma; in rds_send_xmit() 304 ret = conn->c_trans->xmit_rdma(conn, &rm->rdma); in rds_send_xmit() 341 ops_present = (rm->atomic.op_active || rm->rdma.op_active); in rds_send_xmit() 344 if (rm->rdma.op_active && !rm->rdma.op_silent) in rds_send_xmit() 498 notify = rm->rdma.op_notify | rm->data.op_notify; in rds_rdma_send_complete() 499 ro = &rm->rdma; in rds_rdma_send_complete() 570 ro = &rm->rdma; in __rds_send_complete() 636 struct rm_rdma_op *ro = &rm->rdma; in rds_send_remove_from_sock() [all …]
|
D | ib_send.c | 103 rds_ib_send_complete(container_of(op, struct rds_message, rdma), in rds_ib_send_unmap_rdma() 143 if (rm->rdma.op_active && rm->data.op_notify) in rds_ib_send_unmap_data() 144 rds_ib_send_unmap_rdma(ic, &rm->rdma, wc_status); in rds_ib_send_unmap_data() 171 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op() 582 if (rm->rdma.op_active) { in rds_ib_xmit() 585 ext_hdr.h_rdma_rkey = cpu_to_be32(rm->rdma.op_rkey); in rds_ib_xmit() 618 if (rm->rdma.op_active && rm->rdma.op_fence) in rds_ib_xmit() 950 rds_message_addref(container_of(op, struct rds_message, rdma)); in rds_ib_xmit_rdma()
|
/Linux-v4.19/drivers/infiniband/sw/rxe/ |
D | Kconfig | 27 https://github.com/linux-rdma/rdma-core/blob/master/Documentation/rxe.md
|