/Linux-v5.4/net/9p/ |
D | trans_rdma.c | 152 struct p9_trans_rdma *rdma = clnt->trans; in p9_rdma_show_options() local 154 if (rdma->port != P9_PORT) in p9_rdma_show_options() 155 seq_printf(m, ",port=%u", rdma->port); in p9_rdma_show_options() 156 if (rdma->sq_depth != P9_RDMA_SQ_DEPTH) in p9_rdma_show_options() 157 seq_printf(m, ",sq=%u", rdma->sq_depth); in p9_rdma_show_options() 158 if (rdma->rq_depth != P9_RDMA_RQ_DEPTH) in p9_rdma_show_options() 159 seq_printf(m, ",rq=%u", rdma->rq_depth); in p9_rdma_show_options() 160 if (rdma->timeout != P9_RDMA_TIMEOUT) in p9_rdma_show_options() 161 seq_printf(m, ",timeout=%lu", rdma->timeout); in p9_rdma_show_options() 162 if (rdma->privport) in p9_rdma_show_options() [all …]
|
/Linux-v5.4/net/sunrpc/xprtrdma/ |
D | svc_rdma_sendto.c | 127 svc_rdma_send_ctxt_alloc(struct svcxprt_rdma *rdma) in svc_rdma_send_ctxt_alloc() argument 136 size += rdma->sc_max_send_sges * sizeof(struct ib_sge); in svc_rdma_send_ctxt_alloc() 140 buffer = kmalloc(rdma->sc_max_req_size, GFP_KERNEL); in svc_rdma_send_ctxt_alloc() 143 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_send_ctxt_alloc() 144 rdma->sc_max_req_size, DMA_TO_DEVICE); in svc_rdma_send_ctxt_alloc() 145 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_send_ctxt_alloc() 156 for (i = 0; i < rdma->sc_max_send_sges; i++) in svc_rdma_send_ctxt_alloc() 157 ctxt->sc_sges[i].lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_send_ctxt_alloc() 173 void svc_rdma_send_ctxts_destroy(struct svcxprt_rdma *rdma) in svc_rdma_send_ctxts_destroy() argument 177 while ((ctxt = svc_rdma_next_send_ctxt(&rdma->sc_send_ctxts))) { in svc_rdma_send_ctxts_destroy() [all …]
|
D | svc_rdma_rw.c | 55 svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) in svc_rdma_get_rw_ctxt() argument 59 spin_lock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 61 ctxt = svc_rdma_next_ctxt(&rdma->sc_rw_ctxts); in svc_rdma_get_rw_ctxt() 64 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 66 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 85 static void svc_rdma_put_rw_ctxt(struct svcxprt_rdma *rdma, in svc_rdma_put_rw_ctxt() argument 90 spin_lock(&rdma->sc_rw_ctxt_lock); in svc_rdma_put_rw_ctxt() 91 list_add(&ctxt->rw_list, &rdma->sc_rw_ctxts); in svc_rdma_put_rw_ctxt() 92 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_put_rw_ctxt() 100 void svc_rdma_destroy_rw_ctxts(struct svcxprt_rdma *rdma) in svc_rdma_destroy_rw_ctxts() argument [all …]
|
D | svc_rdma_recvfrom.c | 121 svc_rdma_recv_ctxt_alloc(struct svcxprt_rdma *rdma) in svc_rdma_recv_ctxt_alloc() argument 130 buffer = kmalloc(rdma->sc_max_req_size, GFP_KERNEL); in svc_rdma_recv_ctxt_alloc() 133 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_recv_ctxt_alloc() 134 rdma->sc_max_req_size, DMA_FROM_DEVICE); in svc_rdma_recv_ctxt_alloc() 135 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_recv_ctxt_alloc() 144 ctxt->rc_recv_sge.length = rdma->sc_max_req_size; in svc_rdma_recv_ctxt_alloc() 145 ctxt->rc_recv_sge.lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_recv_ctxt_alloc() 158 static void svc_rdma_recv_ctxt_destroy(struct svcxprt_rdma *rdma, in svc_rdma_recv_ctxt_destroy() argument 161 ib_dma_unmap_single(rdma->sc_pd->device, ctxt->rc_recv_sge.addr, in svc_rdma_recv_ctxt_destroy() 172 void svc_rdma_recv_ctxts_destroy(struct svcxprt_rdma *rdma) in svc_rdma_recv_ctxts_destroy() argument [all …]
|
D | svc_rdma_transport.c | 269 struct svcxprt_rdma *rdma = cma_id->context; in rdma_cma_handler() local 270 struct svc_xprt *xprt = &rdma->sc_xprt; in rdma_cma_handler() 280 clear_bit(RDMAXPRT_CONN_PENDING, &rdma->sc_flags); in rdma_cma_handler() 579 struct svcxprt_rdma *rdma = in svc_rdma_detach() local 583 rdma_disconnect(rdma->sc_cm_id); in svc_rdma_detach() 588 struct svcxprt_rdma *rdma = in __svc_rdma_free() local 590 struct svc_xprt *xprt = &rdma->sc_xprt; in __svc_rdma_free() 594 if (rdma->sc_qp && !IS_ERR(rdma->sc_qp)) in __svc_rdma_free() 595 ib_drain_qp(rdma->sc_qp); in __svc_rdma_free() 597 svc_rdma_flush_recv_queues(rdma); in __svc_rdma_free() [all …]
|
D | svc_rdma_backchannel.c | 114 static int svc_rdma_bc_sendto(struct svcxprt_rdma *rdma, in svc_rdma_bc_sendto() argument 120 ret = svc_rdma_map_reply_msg(rdma, ctxt, &rqst->rq_snd_buf, NULL); in svc_rdma_bc_sendto() 129 return svc_rdma_send(rdma, &ctxt->sc_send_wr); in svc_rdma_bc_sendto() 172 rpcrdma_bc_send_request(struct svcxprt_rdma *rdma, struct rpc_rqst *rqst) in rpcrdma_bc_send_request() argument 180 ctxt = svc_rdma_send_ctxt_get(rdma); in rpcrdma_bc_send_request() 192 svc_rdma_sync_reply_hdr(rdma, ctxt, RPCRDMA_HDRLEN_MIN); in rpcrdma_bc_send_request() 198 rc = svc_rdma_bc_sendto(rdma, rqst, ctxt); in rpcrdma_bc_send_request() 200 svc_rdma_send_ctxt_put(rdma, ctxt); in rpcrdma_bc_send_request() 217 struct svcxprt_rdma *rdma; in xprt_rdma_bc_send_request() local 226 rdma = container_of(sxprt, struct svcxprt_rdma, sc_xprt); in xprt_rdma_bc_send_request() [all …]
|
/Linux-v5.4/include/linux/sunrpc/ |
D | svc_rdma.h | 160 extern void svc_rdma_recv_ctxts_destroy(struct svcxprt_rdma *rdma); 161 extern bool svc_rdma_post_recvs(struct svcxprt_rdma *rdma); 162 extern void svc_rdma_recv_ctxt_put(struct svcxprt_rdma *rdma, 164 extern void svc_rdma_flush_recv_queues(struct svcxprt_rdma *rdma); 168 extern void svc_rdma_destroy_rw_ctxts(struct svcxprt_rdma *rdma); 169 extern int svc_rdma_recv_read_chunk(struct svcxprt_rdma *rdma, 172 extern int svc_rdma_send_write_chunk(struct svcxprt_rdma *rdma, 174 extern int svc_rdma_send_reply_chunk(struct svcxprt_rdma *rdma, 179 extern void svc_rdma_send_ctxts_destroy(struct svcxprt_rdma *rdma); 181 svc_rdma_send_ctxt_get(struct svcxprt_rdma *rdma); [all …]
|
/Linux-v5.4/Documentation/admin-guide/cgroup-v1/ |
D | rdma.rst | 28 Currently user space applications can easily take away all the rdma verb 31 rdma resources. This can lead to service unavailability. 34 of processes can be limited. Through this controller different rdma 43 by rdma cgroup, which can be extended later if required. 52 of the child processes which shares the address space, rdma resources are 56 rdma resources. Linking resources around css also ensures that cgroups can be 60 Whenever RDMA resource charging occurs, owner rdma cgroup is returned to 61 the caller. Same rdma cgroup should be passed while uncharging the resource. 81 IB stack honors limits enforced by the rdma controller. When application 86 Following resources can be accounted by rdma controller. [all …]
|
D | index.rst | 21 rdma
|
/Linux-v5.4/drivers/gpu/drm/mediatek/ |
D | mtk_disp_rdma.c | 47 #define RDMA_FIFO_SIZE(rdma) ((rdma)->data->fifo_size) argument 75 struct mtk_ddp_comp *rdma = &priv->ddp_comp; in mtk_disp_rdma_irq_handler() local 78 writel(0x0, rdma->regs + DISP_REG_RDMA_INT_STATUS); in mtk_disp_rdma_irq_handler() 83 mtk_crtc_ddp_irq(priv->crtc, rdma); in mtk_disp_rdma_irq_handler() 100 struct mtk_disp_rdma *rdma = comp_to_rdma(comp); in mtk_rdma_enable_vblank() local 102 rdma->crtc = crtc; in mtk_rdma_enable_vblank() 109 struct mtk_disp_rdma *rdma = comp_to_rdma(comp); in mtk_rdma_disable_vblank() local 111 rdma->crtc = NULL; in mtk_rdma_disable_vblank() 132 struct mtk_disp_rdma *rdma = comp_to_rdma(comp); in mtk_rdma_config() local 145 RDMA_FIFO_PSEUDO_SIZE(RDMA_FIFO_SIZE(rdma)) | in mtk_rdma_config() [all …]
|
/Linux-v5.4/Documentation/ABI/stable/ |
D | sysfs-driver-ib_srp | 4 Contact: linux-rdma@vger.kernel.org 74 Contact: linux-rdma@vger.kernel.org 80 Contact: linux-rdma@vger.kernel.org 86 Contact: linux-rdma@vger.kernel.org 94 Contact: linux-rdma@vger.kernel.org 101 Contact: linux-rdma@vger.kernel.org 108 Contact: linux-rdma@vger.kernel.org 114 Contact: linux-rdma@vger.kernel.org 121 Contact: linux-rdma@vger.kernel.org 128 Contact: linux-rdma@vger.kernel.org [all …]
|
D | sysfs-class-infiniband | 9 Contact: linux-rdma@vger.kernel.org 22 Contact: linux-rdma@vger.kernel.org 32 Contact: linux-rdma@vger.kernel.org 47 Contact: linux-rdma@vger.kernel.org 76 Contact: linux-rdma@vger.kernel.org 104 Contact: linux-rdma@vger.kernel.org 202 Contact: linux-rdma@vger.kernel.org 221 Contact: linux-rdma@vger.kernel.org 228 Contact: linux-rdma@vger.kernel.org 240 Contact: linux-rdma@vger.kernel.org [all …]
|
D | sysfs-transport-srp | 4 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 11 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 20 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 36 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 50 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org
|
/Linux-v5.4/drivers/net/ethernet/seeq/ |
D | sgiseeq.c | 69 volatile struct hpc_dma_desc rdma; member 209 sp->rx_desc[i].rdma.pbuf = dma_addr; in seeq_init_ring() 211 sp->rx_desc[i].rdma.cntinfo = RCNTINFO_INIT; in seeq_init_ring() 214 sp->rx_desc[i - 1].rdma.cntinfo |= HPCDMA_EOR; in seeq_init_ring() 259 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 260 r[i].rdma.pnext); in sgiseeq_dump_rings() 263 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 264 r[i].rdma.pnext); in sgiseeq_dump_rings() 351 while (!(rd->rdma.cntinfo & HPCDMA_OWN)) { in sgiseeq_rx() 352 len = PKT_BUF_SZ - (rd->rdma.cntinfo & HPCDMA_BCNT) - 3; in sgiseeq_rx() [all …]
|
/Linux-v5.4/Documentation/devicetree/bindings/media/ |
D | mediatek-mdp.txt | 12 "mediatek,mt8173-mdp-rdma" - read DMA 24 "mediatek,mt8173-mdp-rdma" 35 mdp_rdma0: rdma@14001000 { 36 compatible = "mediatek,mt8173-mdp-rdma"; 47 mdp_rdma1: rdma@14002000 { 48 compatible = "mediatek,mt8173-mdp-rdma";
|
/Linux-v5.4/drivers/nvme/host/ |
D | Makefile | 8 obj-$(CONFIG_NVME_RDMA) += nvme-rdma.o 22 nvme-rdma-y += rdma.o
|
/Linux-v5.4/drivers/nvme/target/ |
D | Makefile | 7 obj-$(CONFIG_NVME_TARGET_RDMA) += nvmet-rdma.o 15 nvmet-rdma-y += rdma.o
|
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_offload.c | 281 struct rdma_info *rdma = data; in cxgb_rdma_ctl() local 284 rdma->udbell_physbase = pci_resource_start(pdev, 2); in cxgb_rdma_ctl() 285 rdma->udbell_len = pci_resource_len(pdev, 2); in cxgb_rdma_ctl() 286 rdma->tpt_base = in cxgb_rdma_ctl() 288 rdma->tpt_top = t3_read_reg(adapter, A_ULPTX_TPT_ULIMIT); in cxgb_rdma_ctl() 289 rdma->pbl_base = in cxgb_rdma_ctl() 291 rdma->pbl_top = t3_read_reg(adapter, A_ULPTX_PBL_ULIMIT); in cxgb_rdma_ctl() 292 rdma->rqt_base = t3_read_reg(adapter, A_ULPRX_RQ_LLIMIT); in cxgb_rdma_ctl() 293 rdma->rqt_top = t3_read_reg(adapter, A_ULPRX_RQ_ULIMIT); in cxgb_rdma_ctl() 294 rdma->kdb_addr = adapter->regs + A_SG_KDOORBELL; in cxgb_rdma_ctl() [all …]
|
/Linux-v5.4/drivers/infiniband/ |
D | Kconfig | 25 need libibumad from rdma-core 26 <https://github.com/linux-rdma/rdma-core>. 38 rdma-core <https://github.com/linux-rdma/rdma-core>.
|
/Linux-v5.4/drivers/macintosh/ |
D | rack-meter.c | 146 struct rackmeter_dma *rdma = rm->dma_buf_v; in rackmeter_do_pause() local 155 memset(rdma->buf1, 0, sizeof(rdma->buf1)); in rackmeter_do_pause() 156 memset(rdma->buf2, 0, sizeof(rdma->buf2)); in rackmeter_do_pause() 371 struct resource ri2s, rdma; in rackmeter_probe() local 431 of_address_to_resource(i2s, 1, &rdma)) { in rackmeter_probe() 441 pr_debug(" dma @0x%08x\n", (unsigned int)rdma.start); in rackmeter_probe() 475 rm->dma_regs = ioremap(rdma.start, 0x100); in rackmeter_probe()
|
/Linux-v5.4/Documentation/devicetree/bindings/display/mediatek/ |
D | mediatek,disp.txt | 31 "mediatek,<chip>-disp-rdma" - read DMA / line buffer 56 "mediatek,<chip>-disp-rdma" 94 rdma0: rdma@1400e000 { 95 compatible = "mediatek,mt8173-disp-rdma"; 104 rdma1: rdma@1400f000 { 105 compatible = "mediatek,mt8173-disp-rdma"; 114 rdma2: rdma@14010000 { 115 compatible = "mediatek,mt8173-disp-rdma";
|
/Linux-v5.4/drivers/infiniband/hw/mlx5/ |
D | odp.c | 73 } rdma; member 1268 u32 rkey = pfault->rdma.r_key; in mlx5_ib_mr_rdma_pfault_handler() 1277 pfault->rdma.rdma_va += pfault->bytes_committed; in mlx5_ib_mr_rdma_pfault_handler() 1278 pfault->rdma.rdma_op_len -= min(pfault->bytes_committed, in mlx5_ib_mr_rdma_pfault_handler() 1279 pfault->rdma.rdma_op_len); in mlx5_ib_mr_rdma_pfault_handler() 1282 address = pfault->rdma.rdma_va; in mlx5_ib_mr_rdma_pfault_handler() 1283 length = pfault->rdma.rdma_op_len; in mlx5_ib_mr_rdma_pfault_handler() 1290 length = pfault->rdma.packet_size; in mlx5_ib_mr_rdma_pfault_handler() 1387 be32_to_cpu(pf_eqe->rdma.pftype_token) >> 24; in mlx5_ib_eq_pf_process() 1389 be32_to_cpu(pf_eqe->rdma.pftype_token) & in mlx5_ib_eq_pf_process() [all …]
|
/Linux-v5.4/include/trace/events/ |
D | rpcrdma.h | 1503 const struct svcxprt_rdma *rdma, 1507 TP_ARGS(rdma, page), 1511 __string(device, rdma->sc_cm_id->device->name) 1512 __string(addr, rdma->sc_xprt.xpt_remotebuf) 1517 __assign_str(device, rdma->sc_cm_id->device->name); 1518 __assign_str(addr, rdma->sc_xprt.xpt_remotebuf); 1528 const struct svcxprt_rdma *rdma, 1532 TP_ARGS(rdma, status), 1536 __string(device, rdma->sc_cm_id->device->name) 1537 __string(addr, rdma->sc_xprt.xpt_remotebuf) [all …]
|
/Linux-v5.4/net/rds/ |
D | send.c | 281 (rm->rdma.op_active && in rds_send_xmit() 310 if (rm->rdma.op_active && !cp->cp_xmit_rdma_sent) { in rds_send_xmit() 311 rm->m_final_op = &rm->rdma; in rds_send_xmit() 316 ret = conn->c_trans->xmit_rdma(conn, &rm->rdma); in rds_send_xmit() 353 ops_present = (rm->atomic.op_active || rm->rdma.op_active); in rds_send_xmit() 356 if (rm->rdma.op_active && !rm->rdma.op_silent) in rds_send_xmit() 509 ro = &rm->rdma; in rds_rdma_send_complete() 580 ro = &rm->rdma; in __rds_send_complete() 646 struct rm_rdma_op *ro = &rm->rdma; in rds_send_remove_from_sock() 659 rm->rdma.op_notifier = NULL; in rds_send_remove_from_sock() [all …]
|
/Linux-v5.4/drivers/infiniband/sw/rxe/ |
D | Kconfig | 28 https://github.com/linux-rdma/rdma-core/blob/master/Documentation/rxe.md
|