Searched refs:num_cqe (Results 1 – 12 of 12) sorted by relevance
55 cqe = &cq->queue[cq->cq_get % cq->num_cqe]; in siw_reap_cqe()76 cq->cq_get % cq->num_cqe, cqe->opcode, in siw_reap_cqe()
1153 cq->num_cqe = size; in siw_create_cq()1190 uresp.num_cqe = size; in siw_create_cq()1228 int siw_poll_cq(struct ib_cq *base_cq, int num_cqe, struct ib_wc *wc) in siw_poll_cq() argument1233 for (i = 0; i < num_cqe; i++) { in siw_poll_cq()
1077 idx = cq->cq_put % cq->num_cqe; in siw_sqe_complete()1134 idx = cq->cq_put % cq->num_cqe; in siw_rqe_complete()
210 u32 num_cqe; member
22 __u32 num_cqe; member
19 __u32 num_cqe; member
333 u32 num_cqe = cq->queue_size / GDMA_CQE_SIZE; in mana_gd_ring_cq() local335 u32 head = cq->head % (num_cqe << GDMA_CQE_OWNER_BITS); in mana_gd_ring_cq()1165 unsigned int num_cqe = cq->queue_size / sizeof(struct gdma_cqe); in mana_gd_read_cqe() local1170 cqe = &cq_cqe[cq->head % num_cqe]; in mana_gd_read_cqe()1173 old_bits = (cq->head / num_cqe - 1) & GDMA_CQE_OWNER_MASK; in mana_gd_read_cqe()1178 new_bits = (cq->head / num_cqe) & GDMA_CQE_OWNER_MASK; in mana_gd_read_cqe()1195 int mana_gd_poll_cq(struct gdma_queue *cq, struct gdma_comp *comp, int num_cqe) in mana_gd_poll_cq() argument1200 for (cqe_idx = 0; cqe_idx < num_cqe; cqe_idx++) { in mana_gd_poll_cq()
423 int mana_gd_poll_cq(struct gdma_queue *cq, struct gdma_comp *comp, int num_cqe);
519 u32 num_cqe);
1343 u32 num_cqe) in bnx2fc_indicate_kcqe() argument1349 while (i < num_cqe) { in bnx2fc_indicate_kcqe()
2476 u32 num_cqe) in bnx2i_indicate_kcqe() argument2482 while (i < num_cqe) { in bnx2i_indicate_kcqe()
1656 uresp.num_cqe = depth; in erdma_create_cq()