| /Linux-v5.4/drivers/infiniband/hw/cxgb3/ |
| D | iwch_ev.c | 51 xa_lock(&rnicp->qps); in post_qp_event() 52 qhp = xa_load(&rnicp->qps, CQE_QPID(rsp_msg->cqe)); in post_qp_event() 58 xa_unlock(&rnicp->qps); in post_qp_event() 68 xa_unlock(&rnicp->qps); in post_qp_event() 79 xa_unlock(&rnicp->qps); in post_qp_event() 117 xa_lock(&rnicp->qps); in iwch_ev_dispatch() 119 qhp = xa_load(&rnicp->qps, CQE_QPID(rsp_msg->cqe)); in iwch_ev_dispatch() 126 xa_unlock(&rnicp->qps); in iwch_ev_dispatch() 131 xa_unlock(&rnicp->qps); in iwch_ev_dispatch()
|
| D | iwch.c | 70 xa_lock_irq(&rnicp->qps); in disable_dbs() 71 xa_for_each(&rnicp->qps, index, qhp) in disable_dbs() 73 xa_unlock_irq(&rnicp->qps); in disable_dbs() 81 xa_lock_irq(&rnicp->qps); in enable_dbs() 82 xa_for_each(&rnicp->qps, index, qhp) { in enable_dbs() 88 xa_unlock_irq(&rnicp->qps); in enable_dbs() 102 xa_init_flags(&rnicp->qps, XA_FLAGS_LOCK_IRQ); in rnic_init() 186 WARN_ON(!xa_empty(&dev->qps)); in close_rnic_dev()
|
| D | iwch.h | 110 struct xarray qps; member 143 return xa_load(&rhp->qps, qpid); in get_qhp()
|
| D | iwch_provider.c | 679 xa_erase_irq(&rhp->qps, qhp->wq.qpid); in iwch_destroy_qp() 795 if (xa_store_irq(&rhp->qps, qhp->wq.qpid, qhp, GFP_KERNEL)) { in iwch_create_qp()
|
| /Linux-v5.4/drivers/net/ethernet/huawei/hinic/ |
| D | hinic_hw_io.c | 126 qp = &func_to_io->qps[i]; in write_sq_ctxts() 170 qp = &func_to_io->qps[i]; in write_rq_ctxts() 377 qps_size = num_qps * sizeof(*func_to_io->qps); in hinic_io_create_qps() 378 func_to_io->qps = devm_kzalloc(&pdev->dev, qps_size, GFP_KERNEL); in hinic_io_create_qps() 379 if (!func_to_io->qps) in hinic_io_create_qps() 415 err = init_qp(func_to_io, &func_to_io->qps[i], i, in hinic_io_create_qps() 440 destroy_qp(func_to_io, &func_to_io->qps[j]); in hinic_io_create_qps() 455 devm_kfree(&pdev->dev, func_to_io->qps); in hinic_io_create_qps() 474 destroy_qp(func_to_io, &func_to_io->qps[i]); in hinic_io_destroy_qps() 484 devm_kfree(&pdev->dev, func_to_io->qps); in hinic_io_destroy_qps() [all …]
|
| D | hinic_hw_io.h | 58 struct hinic_qp *qps; member
|
| D | hinic_hw_dev.c | 904 struct hinic_qp *qp = &func_to_io->qps[i]; in hinic_hwdev_get_sq() 922 struct hinic_qp *qp = &func_to_io->qps[i]; in hinic_hwdev_get_rq()
|
| /Linux-v5.4/drivers/infiniband/hw/cxgb4/ |
| D | device.c | 344 xa_for_each(&qpd->devp->qps, index, qp) in qp_open() 354 xa_lock_irq(&qpd->devp->qps); in qp_open() 355 xa_for_each(&qpd->devp->qps, index, qp) in qp_open() 357 xa_unlock_irq(&qpd->devp->qps); in qp_open() 940 WARN_ON(!xa_empty(&ctx->dev->qps)); in c4iw_dealloc() 1049 xa_init_flags(&devp->qps, XA_FLAGS_LOCK_IRQ); in c4iw_alloc() 1272 xa_lock_irqsave(&ctx->dev->qps, flags); in stop_queues() 1276 xa_for_each(&ctx->dev->qps, index, qp) in stop_queues() 1281 xa_unlock_irqrestore(&ctx->dev->qps, flags); in stop_queues() 1311 xa_lock_irq(&ctx->dev->qps); in resume_queues() [all …]
|
| D | ev.c | 126 xa_lock_irq(&dev->qps); in c4iw_ev_dispatch() 127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch() 134 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch() 149 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch() 155 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch()
|
| D | qp.c | 66 xa_lock_irq(&dev->qps); in alloc_ird() 71 xa_unlock_irq(&dev->qps); in alloc_ird() 82 xa_lock_irq(&dev->qps); in free_ird() 84 xa_unlock_irq(&dev->qps); in free_ird() 915 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_sq_db() 924 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_sq_db() 932 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_rq_db() 941 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_rq_db() 2088 xa_lock_irq(&rhp->qps); in c4iw_destroy_qp() 2089 __xa_erase(&rhp->qps, qhp->wq.sq.qid); in c4iw_destroy_qp() [all …]
|
| D | iw_cxgb4.h | 319 struct xarray qps; member 356 return xa_load(&rhp->qps, qpid); in get_qhp()
|
| /Linux-v5.4/drivers/gpu/drm/amd/display/dc/dsc/ |
| D | rc_calc.h | 76 const qp_set qps; member
|
| D | rc_calc.c | 44 void get_qp_set(qp_set qps, enum colour_mode cm, enum bits_per_comp bpc, enum max_min max_min, floa… in get_qp_set() argument 86 memcpy(qps, table[index].qps, sizeof(qp_set)); in get_qp_set()
|
| /Linux-v5.4/drivers/net/ethernet/mellanox/mlx4/ |
| D | en_rx.c | 1166 &rss_map->qps[i]); in mlx4_en_config_rss_steer() 1174 rss_map->indir_qp = &rss_map->qps[0]; in mlx4_en_config_rss_steer() 1250 MLX4_QP_STATE_RST, NULL, 0, 0, &rss_map->qps[i]); in mlx4_en_config_rss_steer() 1251 mlx4_qp_remove(mdev->dev, &rss_map->qps[i]); in mlx4_en_config_rss_steer() 1252 mlx4_qp_free(mdev->dev, &rss_map->qps[i]); in mlx4_en_config_rss_steer() 1276 MLX4_QP_STATE_RST, NULL, 0, 0, &rss_map->qps[i]); in mlx4_en_release_rss_steer() 1277 mlx4_qp_remove(mdev->dev, &rss_map->qps[i]); in mlx4_en_release_rss_steer() 1278 mlx4_qp_free(mdev->dev, &rss_map->qps[i]); in mlx4_en_release_rss_steer()
|
| D | en_main.c | 139 &priv->rss_map.qps[i], in mlx4_en_update_loopback_state()
|
| D | mlx4_en.h | 444 struct mlx4_qp qps[MAX_RX_RINGS]; member
|
| /Linux-v5.4/drivers/infiniband/hw/qedr/ |
| D | qedr_iw_cm.c | 494 qp = xa_load(&dev->qps, conn_param->qpn); in qedr_iw_connect() 684 qp = xa_load(&dev->qps, conn_param->qpn); in qedr_iw_accept() 742 xa_erase_irq(&qp->dev->qps, qp->qp_id); in qedr_iw_qp_rem_ref() 751 return xa_load(&dev->qps, qpn); in qedr_iw_get_qp()
|
| D | qedr.h | 171 struct xarray qps; member
|
| D | main.c | 362 xa_init_flags(&dev->qps, XA_FLAGS_LOCK_IRQ); in qedr_alloc_resources()
|
| /Linux-v5.4/drivers/infiniband/hw/qib/ |
| D | qib_verbs.c | 670 struct rvt_qp *qps[20]; in qib_verbs_sdma_desc_avail() local 683 if (n == ARRAY_SIZE(qps)) in qib_verbs_sdma_desc_avail() 690 qps[n++] = qp; in qib_verbs_sdma_desc_avail() 696 qp = qps[i]; in qib_verbs_sdma_desc_avail() 1146 struct rvt_qp *qps[5]; in qib_ib_piobufavail() local 1163 if (n == ARRAY_SIZE(qps)) in qib_ib_piobufavail() 1169 qps[n++] = qp; in qib_ib_piobufavail() 1176 qp = qps[i]; in qib_ib_piobufavail()
|
| /Linux-v5.4/drivers/infiniband/hw/ocrdma/ |
| D | ocrdma_hw.h | 104 enum ib_qp_state get_ibqp_state(enum ocrdma_qp_state qps);
|
| D | ocrdma_hw.c | 151 enum ib_qp_state get_ibqp_state(enum ocrdma_qp_state qps) in get_ibqp_state() argument 153 switch (qps) { in get_ibqp_state() 173 static enum ocrdma_qp_state get_ocrdma_qp_state(enum ib_qp_state qps) in get_ocrdma_qp_state() argument 175 switch (qps) { in get_ocrdma_qp_state()
|
| /Linux-v5.4/drivers/infiniband/hw/hfi1/ |
| D | pio.c | 1613 struct rvt_qp *qps[PIO_WAIT_BATCH_SIZE]; in sc_piobufavail() local 1633 if (n == ARRAY_SIZE(qps)) in sc_piobufavail() 1642 priv = qps[top_idx]->priv; in sc_piobufavail() 1649 qps[n++] = qp; in sc_piobufavail() 1664 hfi1_qp_wakeup(qps[top_idx], in sc_piobufavail() 1668 hfi1_qp_wakeup(qps[i], in sc_piobufavail()
|
| /Linux-v5.4/drivers/net/ethernet/intel/i40e/ |
| D | i40e_virtchnl_pf.c | 804 u16 vsi_id, qps; in i40e_map_pf_queues_to_vsi() local 812 qps = vf->ch[i].num_qps; in i40e_map_pf_queues_to_vsi() 815 qps = pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs; in i40e_map_pf_queues_to_vsi() 820 if (j * 2 >= qps) { in i40e_map_pf_queues_to_vsi() 851 u32 qps, num_tc = 1; /* VF has at least one traffic class */ in i40e_map_pf_to_vf_queues() local 860 qps = vf->ch[i].num_qps; in i40e_map_pf_to_vf_queues() 863 qps = pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs; in i40e_map_pf_to_vf_queues() 867 for (j = 0; j < qps; j++) { in i40e_map_pf_to_vf_queues()
|
| /Linux-v5.4/drivers/infiniband/hw/mlx5/ |
| D | mlx5_ib.h | 931 int qps; member
|