Home
last modified time | relevance | path

Searched refs:cons_index (Results 1 – 21 of 21) sorted by relevance

/Linux-v5.4/drivers/infiniband/hw/mthca/
Dmthca_cq.c181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw()
208 *cq->set_ci_db = cpu_to_be32(cq->cons_index); in update_cons_index()
289 for (prod_index = cq->cons_index; in mthca_cq_clean()
292 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean()
297 qpn, cq->cqn, cq->cons_index, prod_index); in mthca_cq_clean()
303 while ((int) --prod_index - (int) cq->cons_index >= 0) { in mthca_cq_clean()
316 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean()
318 cq->cons_index += nfreed; in mthca_cq_clean()
337 cq->cons_index &= cq->ibcq.cqe; in mthca_cq_resize_copy_cqes()
339 cq->cons_index -= cq->ibcq.cqe + 1; in mthca_cq_resize_copy_cqes()
[all …]
Dmthca_eq.c237 eqe = get_eqe(eq, eq->cons_index); in next_eqe_sw()
363 ++eq->cons_index; in mthca_eq_int()
379 set_eq_ci(dev, eq, eq->cons_index); in mthca_eq_int()
411 dev->eq_table.eq[i].cons_index); in mthca_tavor_interrupt()
424 tavor_set_eq_ci(dev, eq, eq->cons_index); in mthca_tavor_msi_x_interrupt()
444 dev->eq_table.eq[i].cons_index); in mthca_arbel_interrupt()
458 arbel_set_eq_ci(dev, eq, eq->cons_index); in mthca_arbel_msi_x_interrupt()
555 eq->cons_index = 0; in mthca_create_eq()
Dmthca_provider.h109 u32 cons_index; member
206 u32 cons_index; member
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/lib/
Deq.h28 u32 cons_index; member
57 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); in next_eqe_sw()
59 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; in next_eqe_sw()
65 u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in eq_update_ci()
/Linux-v5.4/include/linux/mlx5/
Dcq.h52 u32 cons_index; member
144 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx5_cq_set_ci()
154 u32 cons_index) in mlx5_cq_arm() argument
161 ci = cons_index & 0xffffff; in mlx5_cq_arm()
/Linux-v5.4/include/linux/mlx4/
Dcq.h151 ci = cq->cons_index & 0xffffff; in mlx4_cq_arm()
169 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx4_cq_set_ci()
Ddevice.h743 u32 cons_index; member
/Linux-v5.4/drivers/infiniband/hw/mlx4/
Dcq.c90 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw()
348 i = cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes()
352 return i - cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes()
362 i = cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes()
377 ++cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes()
685 ++cq->mcq.cons_index; in mlx4_ib_poll_one()
934 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); ++prod_index) in __mlx4_ib_cq_clean()
935 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe) in __mlx4_ib_cq_clean()
942 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) { in __mlx4_ib_cq_clean()
962 cq->mcq.cons_index += nfreed; in __mlx4_ib_cq_clean()
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx4/
Deq.c99 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | in eq_set_ci()
123 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); in next_eqe_sw()
124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw()
543 eq->eqn, eq->cons_index, ret); in mlx4_eq_int()
573 eq->eqn, eq->cons_index, ret); in mlx4_eq_int()
697 eq->eqn, eq->cons_index, ret); in mlx4_eq_int()
782 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int()
785 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int()
807 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int()
809 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int()
[all …]
Den_tx.c395 u32 cons_index = mcq->cons_index; in mlx4_en_process_tx_cq() local
412 index = cons_index & size_mask; in mlx4_en_process_tx_cq()
421 cons_index & size) && (done < budget)) { in mlx4_en_process_tx_cq()
465 ++cons_index; in mlx4_en_process_tx_cq()
466 index = cons_index & size_mask; in mlx4_en_process_tx_cq()
474 mcq->cons_index = cons_index; in mlx4_en_process_tx_cq()
Den_rx.c691 index = cq->mcq.cons_index & ring->size_mask; in mlx4_en_process_rx_cq()
696 cq->mcq.cons_index & cq->size)) { in mlx4_en_process_rx_cq()
897 ++cq->mcq.cons_index; in mlx4_en_process_rx_cq()
898 index = (cq->mcq.cons_index) & ring->size_mask; in mlx4_en_process_rx_cq()
914 ring->cons = cq->mcq.cons_index; in mlx4_en_process_rx_cq()
Dcq.c409 cq->cons_index = 0; in mlx4_cq_alloc()
Dmlx4.h395 u32 cons_index; member
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/
Deq.c162 ++eq->cons_index; in mlx5_eq_comp_int()
185 count_eqe = eq->core.cons_index; in mlx5_eq_poll_irq_disabled()
187 count_eqe = eq->core.cons_index - count_eqe; in mlx5_eq_poll_irq_disabled()
221 ++eq->cons_index; in mlx5_eq_async_int()
263 eq->cons_index = 0; in create_map_eq()
735 u32 ci = eq->cons_index + cc; in mlx5_eq_get_eqe()
755 eq->cons_index += cc; in mlx5_eq_update_ci()
756 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in mlx5_eq_update_ci()
Dcq.c110 cq->cons_index = 0; in mlx5_core_create_cq()
/Linux-v5.4/drivers/infiniband/hw/mlx5/
Dcq.c95 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw()
442 ++cq->mcq.cons_index; in mlx5_poll_one()
629 uar_page, to_mcq(ibcq)->mcq.cons_index); in mlx5_ib_arm_cq()
1028 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); prod_index++) in __mlx5_ib_cq_clean()
1029 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe) in __mlx5_ib_cq_clean()
1035 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) { in __mlx5_ib_cq_clean()
1053 cq->mcq.cons_index += nfreed; in __mlx5_ib_cq_clean()
1171 i = cq->mcq.cons_index; in copy_resize_cqes()
1202 ++cq->mcq.cons_index; in copy_resize_cqes()
/Linux-v5.4/drivers/infiniband/hw/hns/
Dhns_roce_hw_v1.c1997 return get_sw_cqe(hr_cq, hr_cq->cons_index); in next_cqe_sw()
2000 static void hns_roce_v1_cq_set_ci(struct hns_roce_cq *hr_cq, u32 cons_index) in hns_roce_v1_cq_set_ci() argument
2004 doorbell[0] = cpu_to_le32(cons_index & ((hr_cq->cq_depth << 1) - 1)); in hns_roce_v1_cq_set_ci()
2025 for (prod_index = hr_cq->cons_index; get_sw_cqe(hr_cq, prod_index); in __hns_roce_v1_cq_clean()
2027 if (prod_index == hr_cq->cons_index + hr_cq->ib_cq.cqe) in __hns_roce_v1_cq_clean()
2035 while ((int) --prod_index - (int) hr_cq->cons_index >= 0) { in __hns_roce_v1_cq_clean()
2054 hr_cq->cons_index += nfreed; in __hns_roce_v1_cq_clean()
2061 hns_roce_v1_cq_set_ci(hr_cq, hr_cq->cons_index); in __hns_roce_v1_cq_clean()
2173 cpu_to_le32(hr_cq->cons_index & ((hr_cq->cq_depth << 1) - 1)); in hns_roce_v1_req_notify_cq()
2208 ++hr_cq->cons_index; in hns_roce_v1_poll_one()
[all …]
Dhns_roce_hw_v2.c2465 return get_sw_cqe_v2(hr_cq, hr_cq->cons_index); in next_cqe_sw_v2()
2484 static void hns_roce_v2_cq_set_ci(struct hns_roce_cq *hr_cq, u32 cons_index) in hns_roce_v2_cq_set_ci() argument
2486 *hr_cq->set_ci_db = cons_index & 0xffffff; in hns_roce_v2_cq_set_ci()
2498 for (prod_index = hr_cq->cons_index; get_sw_cqe_v2(hr_cq, prod_index); in __hns_roce_v2_cq_clean()
2500 if (prod_index > hr_cq->cons_index + hr_cq->ib_cq.cqe) in __hns_roce_v2_cq_clean()
2508 while ((int) --prod_index - (int) hr_cq->cons_index >= 0) { in __hns_roce_v2_cq_clean()
2533 hr_cq->cons_index += nfreed; in __hns_roce_v2_cq_clean()
2539 hns_roce_v2_cq_set_ci(hr_cq, hr_cq->cons_index); in __hns_roce_v2_cq_clean()
2646 hr_cq->cons_index & ((hr_cq->cq_depth << 1) - 1)); in hns_roce_v2_req_notify_cq()
2713 ++hr_cq->cons_index; in hns_roce_v2_poll_one()
[all …]
Dhns_roce_device.h502 u32 cons_index; member
765 int cons_index; member
Dhns_roce_cq.c154 hr_cq->cons_index = 0; in hns_roce_cq_alloc()
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/
Dhealth.c183 eq->core.eqn, eq->core.cons_index, eq->core.irqn); in mlx5e_health_channel_eq_recover()