Home
last modified time | relevance | path

Searched refs:cons_idx (Results 1 – 10 of 10) sorted by relevance

/Linux-v5.4/include/linux/qed/
Dqed_chain.h91 u16 cons_idx; member
97 u32 cons_idx; member
194 return p_chain->u.chain16.cons_idx; in qed_chain_get_cons_idx()
199 return p_chain->u.chain32.cons_idx; in qed_chain_get_cons_idx_u32()
208 (u32)p_chain->u.chain16.cons_idx); in qed_chain_get_elem_left()
211 p_chain->u.chain16.cons_idx / p_chain->elem_per_page; in qed_chain_get_elem_left()
222 (u64)p_chain->u.chain32.cons_idx); in qed_chain_get_elem_left_u32()
225 p_chain->u.chain32.cons_idx / p_chain->elem_per_page; in qed_chain_get_elem_left_u32()
329 p_chain->u.chain16.cons_idx++; in qed_chain_return_produced()
331 p_chain->u.chain32.cons_idx++; in qed_chain_return_produced()
[all …]
/Linux-v5.4/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_qp.c101 ci_start = atomic_read(&wq->cons_idx); in hinic_sq_prepare_ctxt()
158 ci_start = atomic_read(&wq->cons_idx); in hinic_rq_prepare_ctxt()
711 unsigned int *wqe_size, u16 *cons_idx) in hinic_sq_read_wqebb() argument
720 hw_wqe = hinic_read_wqe(sq->wq, sizeof(*ctrl), cons_idx); in hinic_sq_read_wqebb()
724 *skb = sq->saved_skb[*cons_idx]; in hinic_sq_read_wqebb()
749 unsigned int wqe_size, u16 *cons_idx) in hinic_sq_read_wqe() argument
753 hw_wqe = hinic_read_wqe(sq->wq, wqe_size, cons_idx); in hinic_sq_read_wqe()
754 *skb = sq->saved_skb[*cons_idx]; in hinic_sq_read_wqe()
837 struct sk_buff **skb, u16 *cons_idx) in hinic_rq_read_wqe() argument
844 hw_wqe = hinic_read_wqe(rq->wq, wqe_size, cons_idx); in hinic_rq_read_wqe()
[all …]
Dhinic_hw_qp.h186 unsigned int wqe_size, u16 *cons_idx);
190 unsigned int *wqe_size, u16 *cons_idx);
205 struct sk_buff **skb, u16 *cons_idx);
210 u16 *cons_idx);
212 void hinic_rq_put_wqe(struct hinic_rq *rq, u16 cons_idx,
216 u16 cons_idx, struct hinic_sge *sge);
Dhinic_hw_wq.h44 atomic_t cons_idx; member
104 u16 *cons_idx);
106 struct hinic_hw_wqe *hinic_read_wqe_direct(struct hinic_wq *wq, u16 cons_idx);
Dhinic_hw_wq.c564 atomic_set(&wq->cons_idx, 0); in hinic_wq_allocate()
668 atomic_set(&wq[i].cons_idx, 0); in hinic_wqs_cmdq_alloc()
807 atomic_add(num_wqebbs, &wq->cons_idx); in hinic_put_wqe()
821 u16 *cons_idx) in hinic_read_wqe() argument
831 curr_cons_idx = atomic_read(&wq->cons_idx); in hinic_read_wqe()
839 *cons_idx = curr_cons_idx; in hinic_read_wqe()
844 copy_wqe_to_shadow(wq, shadow_addr, num_wqebbs, *cons_idx); in hinic_read_wqe()
848 return WQ_PAGE_ADDR(wq, *cons_idx) + WQE_PAGE_OFF(wq, *cons_idx); in hinic_read_wqe()
858 struct hinic_hw_wqe *hinic_read_wqe_direct(struct hinic_wq *wq, u16 cons_idx) in hinic_read_wqe_direct() argument
860 return WQ_PAGE_ADDR(wq, cons_idx) + WQE_PAGE_OFF(wq, cons_idx); in hinic_read_wqe_direct()
Dhinic_hw_eqs.c57 #define GET_CURR_AEQ_ELEM(eq) GET_AEQ_ELEM(eq, (eq)->cons_idx)
59 #define GET_CURR_CEQ_ELEM(eq) GET_CEQ_ELEM(eq, (eq)->cons_idx)
203 val |= HINIC_EQ_CI_SET(eq->cons_idx, IDX) | in eq_update_ci()
264 eq->cons_idx++; in aeq_irq_handler()
266 if (eq->cons_idx == eq->q_len) { in aeq_irq_handler()
267 eq->cons_idx = 0; in aeq_irq_handler()
328 eq->cons_idx++; in ceq_irq_handler()
330 if (eq->cons_idx == eq->q_len) { in ceq_irq_handler()
331 eq->cons_idx = 0; in ceq_irq_handler()
677 eq->cons_idx = 0; in init_eq()
Dhinic_hw_api_cmd.c129 chain->cons_idx = get_hw_cons_idx(chain); in chain_busy()
133 if (chain->cons_idx == MASKED_IDX(chain, prod_idx + 1)) { in chain_busy()
288 chain->cons_idx = HINIC_API_CMD_STATUS_GET(status, CONS_IDX); in api_cmd_status_update()
307 if (chain->cons_idx == chain->prod_idx) { in wait_for_status_poll()
806 chain->cons_idx = 0; in api_chain_init()
Dhinic_hw_api_cmd.h177 u32 cons_idx; member
Dhinic_hw_eqs.h180 u32 cons_idx; member
Dhinic_hw_cmdq.c600 static void cmdq_sync_cmd_handler(struct hinic_cmdq *cmdq, u16 cons_idx, in cmdq_sync_cmd_handler() argument
603 u16 prod_idx = cons_idx; in cmdq_sync_cmd_handler()
721 HINIC_CMDQ_CTXT_BLOCK_INFO_SET(atomic_read(&wq->cons_idx), CI); in cmdq_init_queue_ctxt()