/Linux-v5.4/drivers/net/ethernet/amazon/ena/ |
D | ena_eth_com.h | 82 int ena_com_rx_pkt(struct ena_com_io_cq *io_cq, 90 bool ena_com_cq_empty(struct ena_com_io_cq *io_cq); 92 static inline void ena_com_unmask_intr(struct ena_com_io_cq *io_cq, in ena_com_unmask_intr() argument 95 writel(intr_reg->intr_control, io_cq->unmask_reg); in ena_com_unmask_intr() 193 static inline int ena_com_update_dev_comp_head(struct ena_com_io_cq *io_cq) in ena_com_update_dev_comp_head() argument 198 if (unlikely(io_cq->cq_head_db_reg)) { in ena_com_update_dev_comp_head() 199 head = io_cq->head; in ena_com_update_dev_comp_head() 200 unreported_comp = head - io_cq->last_head_update; in ena_com_update_dev_comp_head() 201 need_update = unreported_comp > (io_cq->q_depth / ENA_COMP_HEAD_THRESH); in ena_com_update_dev_comp_head() 205 io_cq->qid, head); in ena_com_update_dev_comp_head() [all …]
|
D | ena_eth_com.c | 36 struct ena_com_io_cq *io_cq) in ena_com_get_next_rx_cdesc() argument 42 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_get_next_rx_cdesc() 43 expected_phase = io_cq->phase; in ena_com_get_next_rx_cdesc() 45 cdesc = (struct ena_eth_io_rx_cdesc_base *)(io_cq->cdesc_addr.virt_addr in ena_com_get_next_rx_cdesc() 46 + (head_masked * io_cq->cdesc_entry_size_in_bytes)); in ena_com_get_next_rx_cdesc() 243 ena_com_rx_cdesc_idx_to_ptr(struct ena_com_io_cq *io_cq, u16 idx) in ena_com_rx_cdesc_idx_to_ptr() argument 245 idx &= (io_cq->q_depth - 1); in ena_com_rx_cdesc_idx_to_ptr() 247 ((uintptr_t)io_cq->cdesc_addr.virt_addr + in ena_com_rx_cdesc_idx_to_ptr() 248 idx * io_cq->cdesc_entry_size_in_bytes); in ena_com_rx_cdesc_idx_to_ptr() 251 static u16 ena_com_cdesc_rx_pkt_get(struct ena_com_io_cq *io_cq, in ena_com_cdesc_rx_pkt_get() argument [all …]
|
D | ena_com.c | 414 struct ena_com_io_cq *io_cq) in ena_com_init_io_cq() argument 419 memset(&io_cq->cdesc_addr, 0x0, sizeof(io_cq->cdesc_addr)); in ena_com_init_io_cq() 422 io_cq->cdesc_entry_size_in_bytes = in ena_com_init_io_cq() 423 (io_cq->direction == ENA_COM_IO_QUEUE_DIRECTION_TX) ? in ena_com_init_io_cq() 427 size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth; in ena_com_init_io_cq() 431 io_cq->cdesc_addr.virt_addr = in ena_com_init_io_cq() 433 &io_cq->cdesc_addr.phys_addr, GFP_KERNEL); in ena_com_init_io_cq() 435 if (!io_cq->cdesc_addr.virt_addr) { in ena_com_init_io_cq() 436 io_cq->cdesc_addr.virt_addr = in ena_com_init_io_cq() 438 &io_cq->cdesc_addr.phys_addr, in ena_com_init_io_cq() [all …]
|
D | ena_com.h | 470 struct ena_com_io_cq **io_cq); 846 struct ena_com_io_cq *io_cq); 857 struct ena_com_io_cq *io_cq);
|
/Linux-v5.4/block/ |
D | blk-ioc.c | 34 struct io_cq *icq = container_of(head, struct io_cq, __rcu_head); in icq_free_icq_rcu() 43 static void ioc_exit_icq(struct io_cq *icq) in ioc_exit_icq() 60 static void ioc_destroy_icq(struct io_cq *icq) in ioc_destroy_icq() 109 struct io_cq *icq = hlist_entry(ioc->icq_list.first, in ioc_release_fn() 110 struct io_cq, ioc_node); in ioc_release_fn() 173 struct io_cq *icq; in put_io_context_active() 216 struct io_cq *icq = list_entry(icq_list->next, in __ioc_clear_queue() 217 struct io_cq, q_node); in __ioc_clear_queue() 325 struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q) in ioc_lookup_icq() 327 struct io_cq *icq; in ioc_lookup_icq() [all …]
|
D | blk.h | 276 struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q); 277 struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
|
D | elevator.c | 534 if (WARN_ON(e->icq_size < sizeof(struct io_cq)) || in elv_register() 535 WARN_ON(e->icq_align < __alignof__(struct io_cq))) in elv_register()
|
D | bfq-iosched.h | 388 struct io_cq icq; /* must be the first member */
|
D | blk-mq-sched.c | 39 struct io_cq *icq; in blk_mq_sched_assign_ioc()
|
D | bfq-iosched.c | 387 static struct bfq_io_cq *icq_to_bic(struct io_cq *icq) in icq_to_bic() 4945 static void bfq_exit_icq(struct io_cq *icq) in bfq_exit_icq()
|
/Linux-v5.4/include/linux/ |
D | elevator.h | 10 struct io_cq; 51 void (*init_icq)(struct io_cq *); 52 void (*exit_icq)(struct io_cq *);
|
D | iocontext.h | 72 struct io_cq { struct 115 struct io_cq __rcu *icq_hint;
|
D | blkdev.h | 184 struct io_cq *icq;
|
/Linux-v5.4/drivers/scsi/lpfc/ |
D | lpfc_debugfs.h | 463 cq = phba->sli4_hba.hdwq[wqidx].io_cq; in lpfc_debug_dump_cq() 640 if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid) in lpfc_debug_dump_cq_by_id() 645 lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq); in lpfc_debug_dump_cq_by_id()
|
D | lpfc_sli4.h | 645 struct lpfc_queue *io_cq; /* Fast-path FCP & NVME compl queue */ member
|
D | lpfc_init.c | 8670 phba->sli4_hba.hdwq[idx].io_cq = qdesc; in lpfc_alloc_io_wq_cq() 9117 lpfc_sli4_queue_free(hdwq[idx].io_cq); in lpfc_sli4_release_hdwq() 9119 hdwq[idx].io_cq = NULL; in lpfc_sli4_release_hdwq() 9456 qp[qidx].io_cq, in lpfc_sli4_queue_setup() 9765 lpfc_cq_destroy(phba, qp->io_cq); in lpfc_sli4_queue_unset()
|
D | lpfc_debugfs.c | 3763 qp = phba->sli4_hba.hdwq[eqidx].io_cq; in lpfc_idiag_cqs_for_eq() 4226 qp = phba->sli4_hba.hdwq[qidx].io_cq; in lpfc_idiag_queacc_write()
|
D | lpfc_sli.c | 5557 sli4_hba->sli4_write_cq_db(phba, qp->io_cq, 0, in lpfc_sli4_arm_cqeq_intr()
|