Searched refs:cons_idx (Results 1 – 10 of 10) sorted by relevance
/Linux-v5.4/include/linux/qed/ |
D | qed_chain.h | 91 u16 cons_idx; member 97 u32 cons_idx; member 194 return p_chain->u.chain16.cons_idx; in qed_chain_get_cons_idx() 199 return p_chain->u.chain32.cons_idx; in qed_chain_get_cons_idx_u32() 208 (u32)p_chain->u.chain16.cons_idx); in qed_chain_get_elem_left() 211 p_chain->u.chain16.cons_idx / p_chain->elem_per_page; in qed_chain_get_elem_left() 222 (u64)p_chain->u.chain32.cons_idx); in qed_chain_get_elem_left_u32() 225 p_chain->u.chain32.cons_idx / p_chain->elem_per_page; in qed_chain_get_elem_left_u32() 329 p_chain->u.chain16.cons_idx++; in qed_chain_return_produced() 331 p_chain->u.chain32.cons_idx++; in qed_chain_return_produced() [all …]
|
/Linux-v5.4/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_qp.c | 101 ci_start = atomic_read(&wq->cons_idx); in hinic_sq_prepare_ctxt() 158 ci_start = atomic_read(&wq->cons_idx); in hinic_rq_prepare_ctxt() 711 unsigned int *wqe_size, u16 *cons_idx) in hinic_sq_read_wqebb() argument 720 hw_wqe = hinic_read_wqe(sq->wq, sizeof(*ctrl), cons_idx); in hinic_sq_read_wqebb() 724 *skb = sq->saved_skb[*cons_idx]; in hinic_sq_read_wqebb() 749 unsigned int wqe_size, u16 *cons_idx) in hinic_sq_read_wqe() argument 753 hw_wqe = hinic_read_wqe(sq->wq, wqe_size, cons_idx); in hinic_sq_read_wqe() 754 *skb = sq->saved_skb[*cons_idx]; in hinic_sq_read_wqe() 837 struct sk_buff **skb, u16 *cons_idx) in hinic_rq_read_wqe() argument 844 hw_wqe = hinic_read_wqe(rq->wq, wqe_size, cons_idx); in hinic_rq_read_wqe() [all …]
|
D | hinic_hw_qp.h | 186 unsigned int wqe_size, u16 *cons_idx); 190 unsigned int *wqe_size, u16 *cons_idx); 205 struct sk_buff **skb, u16 *cons_idx); 210 u16 *cons_idx); 212 void hinic_rq_put_wqe(struct hinic_rq *rq, u16 cons_idx, 216 u16 cons_idx, struct hinic_sge *sge);
|
D | hinic_hw_wq.h | 44 atomic_t cons_idx; member 104 u16 *cons_idx); 106 struct hinic_hw_wqe *hinic_read_wqe_direct(struct hinic_wq *wq, u16 cons_idx);
|
D | hinic_hw_wq.c | 564 atomic_set(&wq->cons_idx, 0); in hinic_wq_allocate() 668 atomic_set(&wq[i].cons_idx, 0); in hinic_wqs_cmdq_alloc() 807 atomic_add(num_wqebbs, &wq->cons_idx); in hinic_put_wqe() 821 u16 *cons_idx) in hinic_read_wqe() argument 831 curr_cons_idx = atomic_read(&wq->cons_idx); in hinic_read_wqe() 839 *cons_idx = curr_cons_idx; in hinic_read_wqe() 844 copy_wqe_to_shadow(wq, shadow_addr, num_wqebbs, *cons_idx); in hinic_read_wqe() 848 return WQ_PAGE_ADDR(wq, *cons_idx) + WQE_PAGE_OFF(wq, *cons_idx); in hinic_read_wqe() 858 struct hinic_hw_wqe *hinic_read_wqe_direct(struct hinic_wq *wq, u16 cons_idx) in hinic_read_wqe_direct() argument 860 return WQ_PAGE_ADDR(wq, cons_idx) + WQE_PAGE_OFF(wq, cons_idx); in hinic_read_wqe_direct()
|
D | hinic_hw_eqs.c | 57 #define GET_CURR_AEQ_ELEM(eq) GET_AEQ_ELEM(eq, (eq)->cons_idx) 59 #define GET_CURR_CEQ_ELEM(eq) GET_CEQ_ELEM(eq, (eq)->cons_idx) 203 val |= HINIC_EQ_CI_SET(eq->cons_idx, IDX) | in eq_update_ci() 264 eq->cons_idx++; in aeq_irq_handler() 266 if (eq->cons_idx == eq->q_len) { in aeq_irq_handler() 267 eq->cons_idx = 0; in aeq_irq_handler() 328 eq->cons_idx++; in ceq_irq_handler() 330 if (eq->cons_idx == eq->q_len) { in ceq_irq_handler() 331 eq->cons_idx = 0; in ceq_irq_handler() 677 eq->cons_idx = 0; in init_eq()
|
D | hinic_hw_api_cmd.c | 129 chain->cons_idx = get_hw_cons_idx(chain); in chain_busy() 133 if (chain->cons_idx == MASKED_IDX(chain, prod_idx + 1)) { in chain_busy() 288 chain->cons_idx = HINIC_API_CMD_STATUS_GET(status, CONS_IDX); in api_cmd_status_update() 307 if (chain->cons_idx == chain->prod_idx) { in wait_for_status_poll() 806 chain->cons_idx = 0; in api_chain_init()
|
D | hinic_hw_api_cmd.h | 177 u32 cons_idx; member
|
D | hinic_hw_eqs.h | 180 u32 cons_idx; member
|
D | hinic_hw_cmdq.c | 600 static void cmdq_sync_cmd_handler(struct hinic_cmdq *cmdq, u16 cons_idx, in cmdq_sync_cmd_handler() argument 603 u16 prod_idx = cons_idx; in cmdq_sync_cmd_handler() 721 HINIC_CMDQ_CTXT_BLOCK_INFO_SET(atomic_read(&wq->cons_idx), CI); in cmdq_init_queue_ctxt()
|