/Linux-v5.4/drivers/net/ethernet/brocade/bna/ |
D | bfa_msgq.c | 311 static void bfa_msgq_rspq_dbell(struct bfa_msgq_rspq *rspq); 322 bfa_fsm_state_decl(rspq, stopped, struct bfa_msgq_rspq, enum rspq_event); 323 bfa_fsm_state_decl(rspq, init_wait, struct bfa_msgq_rspq, 325 bfa_fsm_state_decl(rspq, ready, struct bfa_msgq_rspq, enum rspq_event); 326 bfa_fsm_state_decl(rspq, dbell_wait, struct bfa_msgq_rspq, 330 rspq_sm_stopped_entry(struct bfa_msgq_rspq *rspq) in rspq_sm_stopped_entry() argument 332 rspq->producer_index = 0; in rspq_sm_stopped_entry() 333 rspq->consumer_index = 0; in rspq_sm_stopped_entry() 334 rspq->flags = 0; in rspq_sm_stopped_entry() 338 rspq_sm_stopped(struct bfa_msgq_rspq *rspq, enum rspq_event event) in rspq_sm_stopped() argument [all …]
|
D | bfa_msgq.h | 105 struct bfa_msgq_rspq rspq; member
|
D | bfi.h | 421 struct bfi_msgq rspq; member
|
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 1570 struct adapter *adapter = rxq->rspq.adapter; in do_gro() 1576 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 1588 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 1596 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 1614 int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp, in t4vf_ethrx_handler() argument 1620 (rspq->netdev->features & NETIF_F_RXCSUM); in t4vf_ethrx_handler() 1621 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); in t4vf_ethrx_handler() 1622 struct adapter *adapter = rspq->adapter; in t4vf_ethrx_handler() 1631 (rspq->netdev->features & NETIF_F_GRO) && csum_ok && in t4vf_ethrx_handler() 1647 skb->protocol = eth_type_trans(skb, rspq->netdev); in t4vf_ethrx_handler() [all …]
|
D | cxgb4vf_main.c | 403 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 412 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 429 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs() 435 static void qenable(struct sge_rspq *rspq) in qenable() argument 437 napi_enable(&rspq->napi); in qenable() 443 t4_write_reg(rspq->adapter, T4VF_SGE_BASE_ADDR + SGE_VF_GTS, in qenable() 445 SEINTARM_V(rspq->intr_params) | in qenable() 446 INGRESSQID_V(rspq->cntxt_id)); in qenable() 458 qenable(&s->ethrxq[rxq].rspq); in enable_rx() 482 napi_disable(&s->ethrxq[rxq].rspq.napi); in quiesce_rx() [all …]
|
D | adapter.h | 212 struct sge_rspq rspq; /* Response Queue */ member
|
D | t4vf_hw.c | 1249 int start, int n, const u16 *rspq, int nrspq) in t4vf_config_rss_range() argument 1251 const u16 *rsp = rspq; in t4vf_config_rss_range() 1252 const u16 *rsp_end = rspq+nrspq; in t4vf_config_rss_range() 1311 rsp = rspq; in t4vf_config_rss_range()
|
/Linux-v5.4/drivers/scsi/bfa/ |
D | bfa_hw_cb.c | 46 bfa_hwcb_rspq_ack_msix(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack_msix() argument 48 writel(__HFN_INT_RME_Q0 << RME_Q_NUM(bfa_ioc_pcifn(&bfa->ioc), rspq), in bfa_hwcb_rspq_ack_msix() 51 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack_msix() 54 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack_msix() 55 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack_msix() 59 bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack() argument 61 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack() 64 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack() 65 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack()
|
D | bfa_hw_ct.c | 67 bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct_rspq_ack() argument 71 r32 = readl(bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack() 72 writel(r32, bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack() 74 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct_rspq_ack() 75 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct_rspq_ack() 84 bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct2_rspq_ack() argument 86 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct2_rspq_ack() 87 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct2_rspq_ack()
|
D | bfa.h | 179 void (*hw_rspq_ack)(struct bfa_s *bfa, int rspq, u32 ci); 313 void bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci); 325 void bfa_hwct_reqq_ack(struct bfa_s *bfa, int rspq); 326 void bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci); 327 void bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
|
D | bfi.h | 593 struct bfi_msgq_s rspq; member
|
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_uld.c | 107 struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); in uldrx_handler() 169 err = t4_sge_alloc_rxq(adap, &q->rspq, false, in alloc_uld_rxqs() 182 ids[i] = q->rspq.abs_id; in alloc_uld_rxqs() 188 if (q->rspq.desc) in alloc_uld_rxqs() 189 free_rspq_fl(adap, &q->rspq, in alloc_uld_rxqs() 220 cmplqid = rxq_info->uldrxq[i].rspq.cntxt_id; in setup_sge_queues_uld() 235 if (q->rspq.desc) in t4_free_uld_rxqs() 236 free_rspq_fl(adap, &q->rspq, in t4_free_uld_rxqs() 320 init_rspq(adap, &r->rspq, 5, 1, uld_info->rxq_size, 64); in cfg_queues_uld() 321 r->rspq.uld = uld_type; in cfg_queues_uld() [all …]
|
D | cxgb4_debugfs.c | 2713 rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A"); in sge_qinfo_show() 2728 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show() 2729 R("RspQ size:", rspq.size); in sge_qinfo_show() 2730 R("RspQE size:", rspq.iqe_len); in sge_qinfo_show() 2731 R("RspQ CIDX:", rspq.cidx); in sge_qinfo_show() 2732 R("RspQ Gen:", rspq.gen); in sge_qinfo_show() 2733 S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq)); in sge_qinfo_show() 2734 S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]); in sge_qinfo_show() 2791 rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A"); in sge_qinfo_show() 2792 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show() [all …]
|
D | sge.c | 2754 struct adapter *adapter = rxq->rspq.adap; in do_gro() 2760 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 2774 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 2779 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) in do_gro() 2787 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 2897 static void t4_tx_completion_handler(struct sge_rspq *rspq, in t4_tx_completion_handler() argument 2902 struct port_info *pi = netdev_priv(rspq->netdev); in t4_tx_completion_handler() 2903 struct adapter *adapter = rspq->adap; in t4_tx_completion_handler() 2926 txq = &s->ethtxq[pi->first_qset + rspq->idx]; in t4_tx_completion_handler() 2964 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); in t4_ethrx_handler() [all …]
|
D | cxgb4_ethtool.c | 834 e->rx_mini_pending = s->ethrxq[pi->first_qset].rspq.size; in get_sge_param() 859 s->ethrxq[pi->first_qset + i].rspq.size = e->rx_mini_pending; in set_sge_param() 881 err = cxgb4_set_rspq_intr_params(&q->rspq, us, cnt); in set_rx_intr_params() 896 q->rspq.adaptive_rx = adaptive_rx; in set_adaptive_rx_setting() 907 return q->rspq.adaptive_rx; in get_adaptive_rx_setting() 1100 const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; in get_coalesce()
|
D | cxgb4_ptp.c | 187 c.u.init.absid = cpu_to_be16(receive_q->rspq.abs_id); in cxgb4_ptp_redirect_rx_packet()
|
D | cxgb4_main.c | 756 &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs() 771 free_irq(minfo->vec, &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs() 787 free_irq(minfo->vec, &s->ethrxq[i].rspq); in free_msix_queue_irqs() 848 rss[i] = rxq[*queues].rspq.abs_id; in cxgb4_write_rss() 1005 err = t4_sge_alloc_rxq(adap, &q->rspq, false, dev, in setup_sge_queues() 1013 q->rspq.idx = j; in setup_sge_queues() 1021 q->rspq.cntxt_id, in setup_sge_queues() 1033 cmplqid = rxq_info->uldrxq[i].rspq.cntxt_id; in setup_sge_queues() 1053 QUEUENUMBER_V(s->ethrxq[0].rspq.abs_id)); in setup_sge_queues() 3050 t4_sge_intr_msix(0, &rx->rspq); in cxgb_netpoll() [all …]
|
D | cxgb4.h | 711 struct sge_rspq rspq; member 724 struct sge_rspq rspq; member 1687 int start, int n, const u16 *rspq, unsigned int nrspq);
|
D | t4_hw.c | 5157 int start, int n, const u16 *rspq, unsigned int nrspq) in t4_config_rss_range() argument 5160 const u16 *rsp = rspq; in t4_config_rss_range() 5161 const u16 *rsp_end = rspq + nrspq; in t4_config_rss_range() 5186 rsp = rspq; in t4_config_rss_range() 5189 rsp = rspq; in t4_config_rss_range() 5192 rsp = rspq; in t4_config_rss_range()
|
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 174 return container_of(q, struct sge_qset, rspq); in rspq_to_qset() 656 memset(&q->rspq, 0, sizeof(q->rspq)); in t3_reset_qset() 711 if (q->rspq.desc) { in t3_free_qset() 713 t3_sge_disable_rspcntxt(adapter, q->rspq.cntxt_id); in t3_free_qset() 716 q->rspq.size * sizeof(struct rsp_desc), in t3_free_qset() 717 q->rspq.desc, q->rspq.phys_addr); in t3_free_qset() 732 qs->rspq.cntxt_id = id; in init_qset_cntxt() 1898 struct sge_rspq *q = &qs->rspq; in ofld_poll() 2323 struct sge_rspq *q = &qs->rspq; in process_responses() 2498 t3_write_reg(adap, A_SG_GTS, V_RSPQ(qs->rspq.cntxt_id) | in napi_rx_handler() [all …]
|
D | adapter.h | 202 struct sge_rspq rspq; member
|
D | common.h | 699 const u8 * cpus, const u16 *rspq); 748 unsigned int size, int rspq, int ovfl_mode,
|
D | cxgb3_main.c | 412 rspq.polling), 0, in request_msix_data_irqs() 448 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies() 461 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity() 1173 struct sge_rspq *q = &adap->sge.qs[i].rspq; in t3_synchronize_rx() 1279 adap->sge.qs[0].rspq. in cxgb_up() 2629 t3_intr_handler(adapter, qs->rspq.polling) (0, source); in cxgb_netpoll()
|
D | t3_hw.c | 2364 unsigned int size, int rspq, int ovfl_mode, in t3_sge_init_cqcntxt() argument 2377 V_CQ_BASE_HI((u32) base_addr) | V_CQ_RSPQ(rspq) | in t3_sge_init_cqcntxt() 2535 const u8 * cpus, const u16 *rspq) in t3_config_rss() argument 2551 if (rspq) in t3_config_rss() 2554 (i << 16) | rspq[q_idx++]); in t3_config_rss() 2555 if (rspq[q_idx] == 0xffff) in t3_config_rss()
|
D | t3_cpl.h | 1441 __u8 rspq:3; member 1445 __u8 rspq:3; member
|