Home
last modified time | relevance | path

Searched refs:rspq (Results 1 – 25 of 26) sorted by relevance

12

/Linux-v4.19/drivers/net/ethernet/brocade/bna/
Dbfa_msgq.c319 static void bfa_msgq_rspq_dbell(struct bfa_msgq_rspq *rspq);
330 bfa_fsm_state_decl(rspq, stopped, struct bfa_msgq_rspq, enum rspq_event);
331 bfa_fsm_state_decl(rspq, init_wait, struct bfa_msgq_rspq,
333 bfa_fsm_state_decl(rspq, ready, struct bfa_msgq_rspq, enum rspq_event);
334 bfa_fsm_state_decl(rspq, dbell_wait, struct bfa_msgq_rspq,
338 rspq_sm_stopped_entry(struct bfa_msgq_rspq *rspq) in rspq_sm_stopped_entry() argument
340 rspq->producer_index = 0; in rspq_sm_stopped_entry()
341 rspq->consumer_index = 0; in rspq_sm_stopped_entry()
342 rspq->flags = 0; in rspq_sm_stopped_entry()
346 rspq_sm_stopped(struct bfa_msgq_rspq *rspq, enum rspq_event event) in rspq_sm_stopped() argument
[all …]
Dbfa_msgq.h113 struct bfa_msgq_rspq rspq; member
Dbfi.h429 struct bfi_msgq rspq; member
/Linux-v4.19/drivers/net/ethernet/chelsio/cxgb4vf/
Dsge.c1570 struct adapter *adapter = rxq->rspq.adapter; in do_gro()
1576 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1588 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1596 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1614 int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp, in t4vf_ethrx_handler() argument
1620 (rspq->netdev->features & NETIF_F_RXCSUM); in t4vf_ethrx_handler()
1621 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); in t4vf_ethrx_handler()
1622 struct adapter *adapter = rspq->adapter; in t4vf_ethrx_handler()
1631 (rspq->netdev->features & NETIF_F_GRO) && csum_ok && in t4vf_ethrx_handler()
1647 skb->protocol = eth_type_trans(skb, rspq->netdev); in t4vf_ethrx_handler()
[all …]
Dcxgb4vf_main.c347 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
356 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
373 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs()
379 static void qenable(struct sge_rspq *rspq) in qenable() argument
381 napi_enable(&rspq->napi); in qenable()
387 t4_write_reg(rspq->adapter, T4VF_SGE_BASE_ADDR + SGE_VF_GTS, in qenable()
389 SEINTARM_V(rspq->intr_params) | in qenable()
390 INGRESSQID_V(rspq->cntxt_id)); in qenable()
402 qenable(&s->ethrxq[rxq].rspq); in enable_rx()
426 napi_disable(&s->ethrxq[rxq].rspq.napi); in quiesce_rx()
[all …]
Dadapter.h212 struct sge_rspq rspq; /* Response Queue */ member
Dt4vf_hw.c1239 int start, int n, const u16 *rspq, int nrspq) in t4vf_config_rss_range() argument
1241 const u16 *rsp = rspq; in t4vf_config_rss_range()
1242 const u16 *rsp_end = rspq+nrspq; in t4vf_config_rss_range()
1301 rsp = rspq; in t4vf_config_rss_range()
/Linux-v4.19/drivers/scsi/bfa/
Dbfa_hw_cb.c54 bfa_hwcb_rspq_ack_msix(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack_msix() argument
56 writel(__HFN_INT_RME_Q0 << RME_Q_NUM(bfa_ioc_pcifn(&bfa->ioc), rspq), in bfa_hwcb_rspq_ack_msix()
59 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack_msix()
62 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack_msix()
63 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack_msix()
68 bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack() argument
70 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack()
73 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack()
74 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack()
Dbfa_hw_ct.c75 bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct_rspq_ack() argument
79 r32 = readl(bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack()
80 writel(r32, bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack()
82 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct_rspq_ack()
83 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct_rspq_ack()
93 bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct2_rspq_ack() argument
95 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct2_rspq_ack()
96 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct2_rspq_ack()
Dbfa.h188 void (*hw_rspq_ack)(struct bfa_s *bfa, int rspq, u32 ci);
322 void bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
334 void bfa_hwct_reqq_ack(struct bfa_s *bfa, int rspq);
335 void bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
336 void bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
Dbfi.h601 struct bfi_msgq_s rspq; member
/Linux-v4.19/drivers/net/ethernet/chelsio/cxgb4/
Dcxgb4_uld.c107 struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); in uldrx_handler()
166 err = t4_sge_alloc_rxq(adap, &q->rspq, false, in alloc_uld_rxqs()
179 ids[i] = q->rspq.abs_id; in alloc_uld_rxqs()
185 if (q->rspq.desc) in alloc_uld_rxqs()
186 free_rspq_fl(adap, &q->rspq, in alloc_uld_rxqs()
217 cmplqid = rxq_info->uldrxq[i].rspq.cntxt_id; in setup_sge_queues_uld()
232 if (q->rspq.desc) in t4_free_uld_rxqs()
233 free_rspq_fl(adap, &q->rspq, in t4_free_uld_rxqs()
317 init_rspq(adap, &r->rspq, 5, 1, uld_info->rxq_size, 64); in cfg_queues_uld()
318 r->rspq.uld = uld_type; in cfg_queues_uld()
[all …]
Dcxgb4_debugfs.c2753 rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A"); in sge_qinfo_show()
2768 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show()
2769 R("RspQ size:", rspq.size); in sge_qinfo_show()
2770 R("RspQE size:", rspq.iqe_len); in sge_qinfo_show()
2771 R("RspQ CIDX:", rspq.cidx); in sge_qinfo_show()
2772 R("RspQ Gen:", rspq.gen); in sge_qinfo_show()
2773 S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq)); in sge_qinfo_show()
2774 S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]); in sge_qinfo_show()
2830 rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A"); in sge_qinfo_show()
2831 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show()
[all …]
Dcxgb4_ethtool.c853 e->rx_mini_pending = s->ethrxq[pi->first_qset].rspq.size; in get_sge_param()
878 s->ethrxq[pi->first_qset + i].rspq.size = e->rx_mini_pending; in set_sge_param()
900 err = cxgb4_set_rspq_intr_params(&q->rspq, us, cnt); in set_rx_intr_params()
915 q->rspq.adaptive_rx = adaptive_rx; in set_adaptive_rx_setting()
926 return q->rspq.adaptive_rx; in get_adaptive_rx_setting()
940 const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; in get_coalesce()
Dsge.c2668 struct adapter *adapter = rxq->rspq.adap; in do_gro()
2674 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
2688 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
2693 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) in do_gro()
2701 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
2810 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); in t4_ethrx_handler()
3004 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); in process_responses()
3267 if (napi_reschedule(&rxq->rspq.napi)) in sge_rx_timer_cb()
3819 if (q->rspq.desc) in t4_free_ofld_rxqs()
3820 free_rspq_fl(adap, &q->rspq, in t4_free_ofld_rxqs()
[all …]
Dcxgb4_ptp.c187 c.u.init.absid = cpu_to_be16(receive_q->rspq.abs_id); in cxgb4_ptp_redirect_rx_packet()
Dcxgb4_main.c673 &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs()
683 &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs()
695 free_irq(adap->msix_info[msi_index++].vec, &s->ethrxq[i].rspq); in free_msix_queue_irqs()
721 rss[i] = rxq[*queues].rspq.abs_id; in cxgb4_write_rss()
878 err = t4_sge_alloc_rxq(adap, &q->rspq, false, dev, in setup_sge_queues()
886 q->rspq.idx = j; in setup_sge_queues()
903 cmplqid = rxq_info->uldrxq[i].rspq.cntxt_id; in setup_sge_queues()
923 QUEUENUMBER_V(s->ethrxq[0].rspq.abs_id)); in setup_sge_queues()
2866 t4_sge_intr_msix(0, &rx->rspq); in cxgb_netpoll()
4934 init_rspq(adap, &r->rspq, 5, 10, 1024, 64); in cfg_queues()
Dcxgb4.h691 struct sge_rspq rspq; member
704 struct sge_rspq rspq; member
1639 int start, int n, const u16 *rspq, unsigned int nrspq);
Dt4_hw.c5078 int start, int n, const u16 *rspq, unsigned int nrspq) in t4_config_rss_range() argument
5081 const u16 *rsp = rspq; in t4_config_rss_range()
5082 const u16 *rsp_end = rspq + nrspq; in t4_config_rss_range()
5107 rsp = rspq; in t4_config_rss_range()
5110 rsp = rspq; in t4_config_rss_range()
5113 rsp = rspq; in t4_config_rss_range()
/Linux-v4.19/drivers/net/ethernet/chelsio/cxgb3/
Dsge.c174 return container_of(q, struct sge_qset, rspq); in rspq_to_qset()
656 memset(&q->rspq, 0, sizeof(q->rspq)); in t3_reset_qset()
711 if (q->rspq.desc) { in t3_free_qset()
713 t3_sge_disable_rspcntxt(adapter, q->rspq.cntxt_id); in t3_free_qset()
716 q->rspq.size * sizeof(struct rsp_desc), in t3_free_qset()
717 q->rspq.desc, q->rspq.phys_addr); in t3_free_qset()
732 qs->rspq.cntxt_id = id; in init_qset_cntxt()
1898 struct sge_rspq *q = &qs->rspq; in ofld_poll()
2323 struct sge_rspq *q = &qs->rspq; in process_responses()
2498 t3_write_reg(adap, A_SG_GTS, V_RSPQ(qs->rspq.cntxt_id) | in napi_rx_handler()
[all …]
Dadapter.h202 struct sge_rspq rspq; member
Dcommon.h699 const u8 * cpus, const u16 *rspq);
748 unsigned int size, int rspq, int ovfl_mode,
Dcxgb3_main.c413 rspq.polling), 0, in request_msix_data_irqs()
449 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies()
462 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity()
1174 struct sge_rspq *q = &adap->sge.qs[i].rspq; in t3_synchronize_rx()
1280 adap->sge.qs[0].rspq. in cxgb_up()
2630 t3_intr_handler(adapter, qs->rspq.polling) (0, source); in cxgb_netpoll()
Dt3_hw.c2364 unsigned int size, int rspq, int ovfl_mode, in t3_sge_init_cqcntxt() argument
2377 V_CQ_BASE_HI((u32) base_addr) | V_CQ_RSPQ(rspq) | in t3_sge_init_cqcntxt()
2535 const u8 * cpus, const u16 *rspq) in t3_config_rss() argument
2551 if (rspq) in t3_config_rss()
2554 (i << 16) | rspq[q_idx++]); in t3_config_rss()
2555 if (rspq[q_idx] == 0xffff) in t3_config_rss()
Dt3_cpl.h1441 __u8 rspq:3; member
1445 __u8 rspq:3; member

12