| /Linux-v5.4/drivers/gpu/drm/amd/display/dc/dsc/ |
| D | qp_tables.h | 29 const qp_table qp_table_422_10bpc_min = { 62 const qp_table qp_table_444_8bpc_max = { 103 const qp_table qp_table_420_12bpc_max = { 136 const qp_table qp_table_444_10bpc_min = { 189 const qp_table qp_table_420_8bpc_max = { 210 const qp_table qp_table_444_8bpc_min = { 251 const qp_table qp_table_444_12bpc_min = { 316 const qp_table qp_table_420_12bpc_min = { 349 const qp_table qp_table_422_12bpc_min = { 390 const qp_table qp_table_422_12bpc_max = { [all …]
|
| D | rc_calc.h | 79 typedef struct qp_entry qp_table[]; typedef
|
| /Linux-v5.4/drivers/net/ethernet/mellanox/mlx4/ |
| D | qp.c | 51 struct mlx4_qp_table *qp_table = &mlx4_priv(dev)->qp_table; in mlx4_qp_event() local 54 spin_lock(&qp_table->lock); in mlx4_qp_event() 60 spin_unlock(&qp_table->lock); in mlx4_qp_event() 226 struct mlx4_qp_table *qp_table = &priv->qp_table; in __mlx4_qp_reserve_range() local 239 *base = mlx4_zone_alloc_entries(qp_table->zones, uid, cnt, align, in __mlx4_qp_reserve_range() 278 struct mlx4_qp_table *qp_table = &priv->qp_table; in __mlx4_qp_release_range() local 282 mlx4_zone_free_entries_unique(qp_table->zones, base_qpn, cnt); in __mlx4_qp_release_range() 311 struct mlx4_qp_table *qp_table = &priv->qp_table; in __mlx4_qp_alloc_icm() local 314 err = mlx4_table_get(dev, &qp_table->qp_table, qpn); in __mlx4_qp_alloc_icm() 318 err = mlx4_table_get(dev, &qp_table->auxc_table, qpn); in __mlx4_qp_alloc_icm() [all …]
|
| D | profile.c | 189 for (priv->qp_table.rdmarc_shift = 0; in mlx4_make_profile() 190 request->num_qp << priv->qp_table.rdmarc_shift < profile[i].num; in mlx4_make_profile() 191 ++priv->qp_table.rdmarc_shift) in mlx4_make_profile() 193 dev->caps.max_qp_dest_rdma = 1 << priv->qp_table.rdmarc_shift; in mlx4_make_profile() 194 priv->qp_table.rdmarc_base = (u32) profile[i].start; in mlx4_make_profile() 196 init_hca->log_rd_per_qp = priv->qp_table.rdmarc_shift; in mlx4_make_profile()
|
| D | main.c | 1589 err = mlx4_init_icm_table(dev, &priv->qp_table.cmpt_table, in mlx4_init_cmpt_table() 1635 mlx4_cleanup_icm_table(dev, &priv->qp_table.cmpt_table); in mlx4_init_cmpt_table() 1719 err = mlx4_init_icm_table(dev, &priv->qp_table.qp_table, in mlx4_init_icm() 1730 err = mlx4_init_icm_table(dev, &priv->qp_table.auxc_table, in mlx4_init_icm() 1741 err = mlx4_init_icm_table(dev, &priv->qp_table.altc_table, in mlx4_init_icm() 1752 err = mlx4_init_icm_table(dev, &priv->qp_table.rdmarc_table, in mlx4_init_icm() 1754 dev_cap->rdmarc_entry_sz << priv->qp_table.rdmarc_shift, in mlx4_init_icm() 1810 mlx4_cleanup_icm_table(dev, &priv->qp_table.rdmarc_table); in mlx4_init_icm() 1813 mlx4_cleanup_icm_table(dev, &priv->qp_table.altc_table); in mlx4_init_icm() 1816 mlx4_cleanup_icm_table(dev, &priv->qp_table.auxc_table); in mlx4_init_icm() [all …]
|
| D | mlx4.h | 722 struct mlx4_icm_table qp_table; member 898 struct mlx4_qp_table qp_table; member
|
| /Linux-v5.4/drivers/infiniband/hw/mthca/ |
| D | mthca_qp.c | 198 return qp->qpn >= dev->qp_table.sqp_start && in is_sqp() 199 qp->qpn <= dev->qp_table.sqp_start + 3; in is_sqp() 204 return qp->qpn >= dev->qp_table.sqp_start && in is_qp0() 205 qp->qpn <= dev->qp_table.sqp_start + 1; in is_qp0() 244 spin_lock(&dev->qp_table.lock); in mthca_qp_event() 245 qp = mthca_array_get(&dev->qp_table.qp, qpn & (dev->limits.num_qps - 1)); in mthca_qp_event() 248 spin_unlock(&dev->qp_table.lock); in mthca_qp_event() 265 spin_lock(&dev->qp_table.lock); in mthca_qp_event() 268 spin_unlock(&dev->qp_table.lock); in mthca_qp_event() 770 cpu_to_be32(dev->qp_table.rdb_base + in __mthca_modify_qp() [all …]
|
| D | mthca_main.c | 443 mdev->qp_table.qp_table = mthca_alloc_icm_table(mdev, init_hca->qpc_base, in mthca_init_icm() 448 if (!mdev->qp_table.qp_table) { in mthca_init_icm() 454 mdev->qp_table.eqp_table = mthca_alloc_icm_table(mdev, init_hca->eqpc_base, in mthca_init_icm() 459 if (!mdev->qp_table.eqp_table) { in mthca_init_icm() 465 mdev->qp_table.rdb_table = mthca_alloc_icm_table(mdev, init_hca->rdb_base, in mthca_init_icm() 468 mdev->qp_table.rdb_shift, 0, in mthca_init_icm() 470 if (!mdev->qp_table.rdb_table) { in mthca_init_icm() 530 mthca_free_icm_table(mdev, mdev->qp_table.rdb_table); in mthca_init_icm() 533 mthca_free_icm_table(mdev, mdev->qp_table.eqp_table); in mthca_init_icm() 536 mthca_free_icm_table(mdev, mdev->qp_table.qp_table); in mthca_init_icm() [all …]
|
| D | mthca_profile.c | 206 for (dev->qp_table.rdb_shift = 0; in mthca_make_profile() 207 request->num_qp << dev->qp_table.rdb_shift < profile[i].num; in mthca_make_profile() 208 ++dev->qp_table.rdb_shift) in mthca_make_profile() 210 dev->qp_table.rdb_base = (u32) profile[i].start; in mthca_make_profile()
|
| D | mthca_dev.h | 260 struct mthca_icm_table *qp_table; member 345 struct mthca_qp_table qp_table; member
|
| D | mthca_cq.c | 520 *cur_qp = mthca_array_get(&dev->qp_table.qp, in mthca_poll_one()
|
| D | mthca_provider.c | 107 props->max_qp_rd_atom = 1 << mdev->qp_table.rdb_shift; in mthca_query_device()
|
| /Linux-v5.4/drivers/infiniband/hw/hns/ |
| D | hns_roce_qp.c | 114 struct hns_roce_qp_table *qp_table = &hr_dev->qp_table; in hns_roce_reserve_range_qp() local 116 return hns_roce_bitmap_alloc_range(&qp_table->bitmap, cnt, align, in hns_roce_reserve_range_qp() 166 struct hns_roce_qp_table *qp_table = &hr_dev->qp_table; in hns_roce_qp_alloc() local 176 ret = hns_roce_table_get(hr_dev, &qp_table->qp_table, hr_qp->qpn); in hns_roce_qp_alloc() 183 ret = hns_roce_table_get(hr_dev, &qp_table->irrl_table, hr_qp->qpn); in hns_roce_qp_alloc() 191 ret = hns_roce_table_get(hr_dev, &qp_table->trrl_table, in hns_roce_qp_alloc() 201 ret = hns_roce_table_get(hr_dev, &qp_table->sccc_table, in hns_roce_qp_alloc() 217 hns_roce_table_put(hr_dev, &qp_table->sccc_table, in hns_roce_qp_alloc() 222 hns_roce_table_put(hr_dev, &qp_table->trrl_table, hr_qp->qpn); in hns_roce_qp_alloc() 225 hns_roce_table_put(hr_dev, &qp_table->irrl_table, hr_qp->qpn); in hns_roce_qp_alloc() [all …]
|
| D | hns_roce_main.c | 614 ret = hns_roce_init_hem_table(hr_dev, &hr_dev->qp_table.qp_table, in hns_roce_init_hem() 622 ret = hns_roce_init_hem_table(hr_dev, &hr_dev->qp_table.irrl_table, in hns_roce_init_hem() 634 &hr_dev->qp_table.trrl_table, in hns_roce_init_hem() 694 &hr_dev->qp_table.sccc_table, in hns_roce_init_hem() 741 &hr_dev->qp_table.sccc_table); in hns_roce_init_hem() 763 &hr_dev->qp_table.trrl_table); in hns_roce_init_hem() 766 hns_roce_cleanup_hem_table(hr_dev, &hr_dev->qp_table.irrl_table); in hns_roce_init_hem() 769 hns_roce_cleanup_hem_table(hr_dev, &hr_dev->qp_table.qp_table); in hns_roce_init_hem()
|
| D | hns_roce_hem.c | 1105 &hr_dev->qp_table.sccc_table); in hns_roce_cleanup_hem() 1108 &hr_dev->qp_table.trrl_table); in hns_roce_cleanup_hem() 1109 hns_roce_cleanup_hem_table(hr_dev, &hr_dev->qp_table.irrl_table); in hns_roce_cleanup_hem() 1110 hns_roce_cleanup_hem_table(hr_dev, &hr_dev->qp_table.qp_table); in hns_roce_cleanup_hem()
|
| D | hns_roce_device.h | 552 struct hns_roce_hem_table qp_table; member 1032 struct hns_roce_qp_table qp_table; member
|
| D | hns_roce_hw_v2.c | 3695 mtts_2 = hns_roce_table_find(hr_dev, &hr_dev->qp_table.irrl_table, in modify_qp_init_to_rtr() 3703 mtts_3 = hns_roce_table_find(hr_dev, &hr_dev->qp_table.trrl_table, in modify_qp_init_to_rtr() 4750 mutex_lock(&hr_dev->qp_table.scc_mutex); in hns_roce_v2_qp_flow_control_init() 4791 mutex_unlock(&hr_dev->qp_table.scc_mutex); in hns_roce_v2_qp_flow_control_init()
|
| /Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/ |
| D | qp.c | 158 priv = container_of(table, struct mlx5_priv, qp_table); in rsc_event_notifier() 200 struct mlx5_qp_table *table = &dev->priv.qp_table; in create_resource_common() 222 struct mlx5_qp_table *table = &dev->priv.qp_table; in destroy_resource_common() 526 struct mlx5_qp_table *table = &dev->priv.qp_table; in mlx5_init_qp_table() 539 struct mlx5_qp_table *table = &dev->priv.qp_table; in mlx5_cleanup_qp_table() 727 struct mlx5_qp_table *table = &dev->priv.qp_table; in mlx5_core_res_hold()
|
| /Linux-v5.4/drivers/infiniband/sw/rdmavt/ |
| D | qp.c | 424 rdi->qp_dev->qp_table = in rvt_driver_qp_init() 426 sizeof(*rdi->qp_dev->qp_table), in rvt_driver_qp_init() 428 if (!rdi->qp_dev->qp_table) in rvt_driver_qp_init() 432 RCU_INIT_POINTER(rdi->qp_dev->qp_table[i], NULL); in rvt_driver_qp_init() 445 kfree(rdi->qp_dev->qp_table); in rvt_driver_qp_init() 479 qp = rcu_dereference_protected(rdi->qp_dev->qp_table[n], in rvt_free_all_qps() 481 RCU_INIT_POINTER(rdi->qp_dev->qp_table[n], NULL); in rvt_free_all_qps() 508 kfree(rdi->qp_dev->qp_table); in rvt_qp_exit() 785 qpp = &rdi->qp_dev->qp_table[n]; in rvt_remove_qp() 1405 qp->next = rdi->qp_dev->qp_table[n]; in rvt_insert_qp() [all …]
|
| /Linux-v5.4/include/rdma/ |
| D | rdmavt_qp.h | 494 struct rvt_qp __rcu **qp_table; member 725 for (qp = rcu_dereference(rdi->qp_dev->qp_table[n]); qp; in rvt_lookup_qpn()
|
| /Linux-v5.4/drivers/infiniband/hw/i40iw/ |
| D | i40iw_hw.c | 94 iwdev->qp_table = (struct i40iw_qp **)(&iwdev->allocated_arps[BITS_TO_LONGS(arp_table_size)]); in i40iw_initialize_hw_resources() 308 iwqp = iwdev->qp_table[info->qp_cq_id]; in i40iw_process_aeq()
|
| D | i40iw.h | 248 struct i40iw_qp **qp_table; member
|
| D | i40iw_utils.c | 654 iwdev->qp_table[qp_num] = NULL; in i40iw_rem_ref() 689 return &iwdev->qp_table[qpn]->ibqp; in i40iw_get_qp()
|
| /Linux-v5.4/include/linux/mlx5/ |
| D | qp.h | 558 return radix_tree_lookup(&dev->priv.qp_table.tree, qpn); in __mlx5_qp_lookup()
|
| D | driver.h | 552 struct mlx5_qp_table qp_table; member
|