/Linux-v4.19/drivers/infiniband/hw/mthca/ |
D | mthca_provider.c | 66 struct mthca_dev *mdev = to_mdev(ibdev); in mthca_query_device() 156 err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, in mthca_query_port() 168 props->gid_tbl_len = to_mdev(ibdev)->limits.gid_table_len; in mthca_query_port() 170 props->pkey_tbl_len = to_mdev(ibdev)->limits.pkey_table_len; in mthca_query_port() 195 if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) in mthca_modify_device() 199 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mthca_modify_device() 213 if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) in mthca_modify_port() 226 err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port); in mthca_modify_port() 230 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mthca_modify_port() 250 err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, in mthca_query_pkey() [all …]
|
D | mthca_mad.c | 121 mthca_update_rate(to_mdev(ibdev), port_num); in smp_snoop() 122 update_sm_ah(to_mdev(ibdev), port_num, in smp_snoop() 156 mutex_lock(&to_mdev(dev)->cap_mask_mutex); in node_desc_override() 159 mutex_unlock(&to_mdev(dev)->cap_mask_mutex); in node_desc_override() 221 forward_trap(to_mdev(ibdev), port_num, in_mad); in mthca_process_mad() 261 err = mthca_MAD_IFC(to_mdev(ibdev), in mthca_process_mad() 268 mthca_err(to_mdev(ibdev), "MAD_IFC returned %d\n", err); in mthca_process_mad()
|
D | mthca_cq.c | 340 if (!mthca_is_memfree(to_mdev(cq->ibcq.device)) && in mthca_cq_resize_copy_cqes() 662 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_poll_cq() 735 mthca_write64(dbhi, 0xffffffff, to_mdev(cq->device)->kar + MTHCA_CQ_DOORBELL, in mthca_tavor_arm_cq() 736 MTHCA_GET_DOORBELL_LOCK(&to_mdev(cq->device)->doorbell_lock)); in mthca_tavor_arm_cq() 767 to_mdev(ibcq->device)->kar + MTHCA_CQ_DOORBELL, in mthca_arbel_arm_cq() 768 MTHCA_GET_DOORBELL_LOCK(&to_mdev(ibcq->device)->doorbell_lock)); in mthca_arbel_arm_cq()
|
D | mthca_srq.c | 367 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_modify_srq() 390 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_query_srq() 478 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_tavor_post_srq_recv() 578 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_arbel_post_srq_recv()
|
D | mthca_mcg.c | 122 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_attach() 216 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_detach()
|
/Linux-v4.19/drivers/infiniband/hw/mlx4/ |
D | mr.c | 67 err = mlx4_mr_alloc(to_mdev(pd->device)->dev, to_mpd(pd)->pdn, 0, in mlx4_ib_get_dma_mr() 72 err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_get_dma_mr() 82 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_get_dma_mr() 408 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_reg_user_mr() 449 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_reg_user_mr() 465 struct mlx4_ib_dev *dev = to_mdev(mr->device); in mlx4_ib_rereg_user_mr() 606 ret = mlx4_mr_free(to_mdev(ibmr->device)->dev, &mr->mmr); in mlx4_ib_dereg_mr() 619 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_alloc_mw() 653 mlx4_mw_free(to_mdev(ibmw->device)->dev, &mw->mmw); in mlx4_ib_dealloc_mw() 663 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_alloc_mr() [all …]
|
D | main.c | 134 struct mlx4_ib_dev *ibdev = to_mdev(device); in mlx4_ib_get_netdev() 251 struct mlx4_ib_dev *ibdev = to_mdev(attr->device); in mlx4_ib_add_gid() 329 struct mlx4_ib_dev *ibdev = to_mdev(attr->device); in mlx4_ib_del_gid() 425 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_query_device() 460 err = mlx4_MAD_IFC(to_mdev(ibdev), MLX4_MAD_IFC_IGNORE_KEYS, in mlx4_ib_query_device() 639 struct mlx4_dev *dev = to_mdev(device)->dev; in mlx4_ib_port_link_layer() 663 if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) in ib_link_query_port() 666 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, in ib_link_query_port() 682 props->gid_tbl_len = to_mdev(ibdev)->dev->caps.gid_table_len[port]; in ib_link_query_port() 683 props->max_msg_sz = to_mdev(ibdev)->dev->caps.max_msg_sz; in ib_link_query_port() [all …]
|
D | cm.c | 146 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in id_map_find_by_sl_id() 194 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_find_del() 212 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in sl_id_map_add() 247 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_alloc() 256 ent->dev = to_mdev(ibdev); in id_map_alloc() 260 spin_lock(&to_mdev(ibdev)->sriov.id_map_lock); in id_map_alloc() 285 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_get() 301 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in schedule_delayed()
|
D | srq.c | 75 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_create_srq() 232 struct mlx4_ib_dev *dev = to_mdev(ibsrq->device); in mlx4_ib_modify_srq() 257 struct mlx4_ib_dev *dev = to_mdev(ibsrq->device); in mlx4_ib_query_srq() 275 struct mlx4_ib_dev *dev = to_mdev(srq->device); in mlx4_ib_destroy_srq() 320 struct mlx4_ib_dev *mdev = to_mdev(ibsrq->device); in mlx4_ib_post_srq_recv()
|
D | cq.c | 95 struct mlx4_ib_dev *dev = to_mdev(cq->device); in mlx4_ib_modify_cq() 181 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_create_cq() 383 struct mlx4_ib_dev *dev = to_mdev(ibcq->device); in mlx4_ib_resize_cq() 484 struct mlx4_ib_dev *dev = to_mdev(cq->device); in mlx4_ib_destroy_cq() 703 struct mlx4_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx4_ib_poll_one() 723 mqp = __mlx4_qp_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one() 735 msrq = mlx4_srq_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one() 845 if (mlx4_is_mfunc(to_mdev(cq->ibcq.device)->dev)) { in mlx4_ib_poll_one() 890 struct mlx4_ib_dev *mdev = to_mdev(cq->ibcq.device); in mlx4_ib_poll_cq() 916 to_mdev(ibcq->device)->uar_map, in mlx4_ib_arm_cq() [all …]
|
D | qp.c | 762 err = create_qp_rss(to_mdev(pd->device), init_attr, &ucmd, qp); in _mlx4_ib_create_qp_rss() 781 struct mlx4_ib_dev *dev = to_mdev(context->ibucontext.device); in mlx4_ib_alloc_wqn() 831 struct mlx4_ib_dev *dev = to_mdev(context->ibucontext.device); in mlx4_ib_release_wqn() 1493 if (!(to_mdev(pd->device)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC)) in _mlx4_ib_create_qp() 1508 err = create_qp_common(to_mdev(pd->device), pd, MLX4_IB_QP_SRC, in _mlx4_ib_create_qp() 1529 int res = mlx4_qp_reserve_range(to_mdev(pd->device)->dev, in _mlx4_ib_create_qp() 1536 sqpn = get_sqp_num(to_mdev(pd->device), init_attr); in _mlx4_ib_create_qp() 1539 err = create_qp_common(to_mdev(pd->device), pd, MLX4_IB_QP_SRC, in _mlx4_ib_create_qp() 1562 struct mlx4_ib_dev *dev = to_mdev(device); in mlx4_ib_create_qp() 1593 struct mlx4_ib_dev *dev = to_mdev(qp->device); in _mlx4_ib_destroy_qp() [all …]
|
D | mad.c | 230 struct mlx4_ib_dev *dev = to_mdev(ibdev); in smp_snoop() 376 spin_lock_irqsave(&to_mdev(dev)->sm_lock, flags); in node_desc_override() 379 spin_unlock_irqrestore(&to_mdev(dev)->sm_lock, flags); in node_desc_override() 434 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_find_real_gid() 668 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_demux_mad() 832 forward_trap(to_mdev(ibdev), port_num, in_mad); in ib_process_mad() 865 err = mlx4_MAD_IFC(to_mdev(ibdev), in ib_process_mad() 876 if (!mlx4_is_slave(to_mdev(ibdev)->dev)) in ib_process_mad() 940 struct mlx4_ib_dev *dev = to_mdev(ibdev); in iboe_process_mad() 989 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_process_mad() [all …]
|
D | ah.c | 47 struct mlx4_dev *dev = to_mdev(pd->device)->dev; in create_ib_ah() 84 struct mlx4_ib_dev *ibdev = to_mdev(pd->device); in create_iboe_ah()
|
/Linux-v4.19/drivers/infiniband/hw/mlx5/ |
D | ib_virt.c | 54 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_get_vf_config() 96 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_set_vf_link_state() 130 dev = to_mdev(device); in mlx5_ib_get_vf_stats() 154 struct mlx5_ib_dev *dev = to_mdev(device); in set_vf_node_guid() 175 struct mlx5_ib_dev *dev = to_mdev(device); in set_vf_port_guid()
|
D | gsi.c | 118 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_gsi_create_qp() 215 struct mlx5_ib_dev *dev = to_mdev(qp->device); in mlx5_ib_gsi_destroy_qp() 275 struct mlx5_ib_dev *dev = to_mdev(qp->device); in modify_to_rts() 315 struct mlx5_ib_dev *dev = to_mdev(device); in setup_qp() 376 struct mlx5_ib_dev *dev = to_mdev(qp->device); in mlx5_ib_gsi_modify_qp() 417 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); in mlx5_ib_add_outstanding_wr() 468 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); in get_tx_qp()
|
D | mad.c | 108 err = mlx5_MAD_IFC(to_mdev(ibdev), in process_mad() 256 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_process_mad() 327 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, 1, NULL, NULL, in_mad, in mlx5_query_mad_ifc_smp_attr_node_info() 466 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_pkey() 495 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_gids() 506 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_gids() 522 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_mad_ifc_port()
|
D | srq.c | 79 struct mlx5_ib_dev *dev = to_mdev(pd->device); in create_srq_user() 242 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_create_srq() 370 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_modify_srq() 395 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_query_srq() 419 struct mlx5_ib_dev *dev = to_mdev(srq->device); in mlx5_ib_destroy_srq() 455 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_post_srq_recv()
|
D | cq.c | 49 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event() 168 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.device); in handle_responder() 512 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_poll_one() 642 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in poll_soft_wc() 669 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_poll_cq() 704 struct mlx5_core_dev *mdev = to_mdev(ibcq->device)->mdev; in mlx5_ib_arm_cq() 985 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_create_cq() 1096 struct mlx5_ib_dev *dev = to_mdev(cq->device); in mlx5_ib_destroy_cq() 1182 struct mlx5_ib_dev *dev = to_mdev(cq->device); in mlx5_ib_modify_cq() 1273 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in copy_resize_cqes() [all …]
|
D | flow.c | 54 struct mlx5_ib_dev *dev = to_mdev(uobj->context->device); in UVERBS_HANDLER() 135 struct mlx5_ib_dev *dev = to_mdev(uobj->context->device); in UVERBS_HANDLER()
|
D | main.c | 135 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_port_link_layer() 238 struct mlx5_ib_dev *ibdev = to_mdev(device); in mlx5_ib_get_netdev() 387 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_query_port_roce() 516 return set_roce_addr(to_mdev(attr->device), attr->port_num, in mlx5_ib_add_gid() 523 return set_roce_addr(to_mdev(attr->device), attr->port_num, in mlx5_ib_del_gid() 551 if (mlx5_use_mad_ifc(to_mdev(ibdev))) in mlx5_get_vport_access_method() 609 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_system_image_guid() 641 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_max_pkeys() 662 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_vendor_id() 727 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_query_device() [all …]
|
D | mr.c | 779 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_get_dma_mr() 855 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mr_umem_get() 932 struct mlx5_ib_dev *dev = to_mdev(pd->device); in alloc_mr_from_cache() 1134 struct mlx5_ib_dev *dev = to_mdev(pd->device); in reg_create() 1228 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_get_memic_mr() 1303 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_reg_user_mr() 1421 struct mlx5_ib_dev *dev = to_mdev(pd->device); in rereg_umr() 1445 struct mlx5_ib_dev *dev = to_mdev(ib_mr->device); in mlx5_ib_rereg_user_mr() 1670 dereg_mr(to_mdev(ibmr->device), to_mmr(ibmr)); in mlx5_ib_dereg_mr() 1678 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_alloc_mr() [all …]
|
D | devx.c | 429 dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 473 dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 505 dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 725 struct mlx5_ib_dev *dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 927 struct mlx5_ib_dev *dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER()
|
D | ah.c | 79 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_create_ah()
|
/Linux-v4.19/drivers/dma/ |
D | altera-msgdma.c | 203 #define to_mdev(chan) container_of(chan, struct msgdma_device, dmachan) macro 307 struct msgdma_device *mdev = to_mdev(tx->chan); in msgdma_tx_submit() 336 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_memcpy() 394 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_slave_sg() 465 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_dma_config() 572 struct msgdma_device *mdev = to_mdev(chan); in msgdma_issue_pending() 641 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_free_chan_resources() 658 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_alloc_chan_resources()
|
/Linux-v4.19/drivers/staging/most/usb/ |
D | usb.c | 124 #define to_mdev(d) container_of(d, struct most_dev, iface) macro 230 struct most_dev *mdev = to_mdev(iface); in hdm_poison_channel() 339 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_write_completion() 488 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_read_completion() 565 mdev = to_mdev(iface); in hdm_enqueue() 633 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_dma_alloc() 641 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_dma_free() 667 struct most_dev *mdev = to_mdev(iface); in hdm_configure_channel() 749 mdev = to_mdev(iface); in hdm_request_netinfo()
|