Home
last modified time | relevance | path

Searched refs:rxq (Results 1 – 25 of 131) sorted by relevance

123456

/Linux-v4.19/drivers/net/ethernet/huawei/hinic/
Dhinic_rx.c51 void hinic_rxq_clean_stats(struct hinic_rxq *rxq) in hinic_rxq_clean_stats() argument
53 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_clean_stats()
66 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) in hinic_rxq_get_stats() argument
68 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_get_stats()
84 static void rxq_stats_init(struct hinic_rxq *rxq) in rxq_stats_init() argument
86 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in rxq_stats_init()
89 hinic_rxq_clean_stats(rxq); in rxq_stats_init()
99 static struct sk_buff *rx_alloc_skb(struct hinic_rxq *rxq, in rx_alloc_skb() argument
102 struct hinic_dev *nic_dev = netdev_priv(rxq->netdev); in rx_alloc_skb()
110 skb = netdev_alloc_skb_ip_align(rxq->netdev, rxq->rq->buf_sz); in rx_alloc_skb()
[all …]
Dhinic_rx.h46 void hinic_rxq_clean_stats(struct hinic_rxq *rxq);
48 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats);
50 int hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq,
53 void hinic_clean_rxq(struct hinic_rxq *rxq);
/Linux-v4.19/drivers/net/wireless/intel/iwlwifi/pcie/
Drx.c143 static int iwl_rxq_space(const struct iwl_rxq *rxq) in iwl_rxq_space() argument
146 WARN_ON(rxq->queue_size & (rxq->queue_size - 1)); in iwl_rxq_space()
154 return (rxq->read - rxq->write - 1) & (rxq->queue_size - 1); in iwl_rxq_space()
191 struct iwl_rxq *rxq) in iwl_pcie_rxq_inc_wr_ptr() argument
195 lockdep_assert_held(&rxq->lock); in iwl_pcie_rxq_inc_wr_ptr()
211 rxq->need_update = true; in iwl_pcie_rxq_inc_wr_ptr()
216 rxq->write_actual = round_down(rxq->write, 8); in iwl_pcie_rxq_inc_wr_ptr()
219 (rxq->write_actual | in iwl_pcie_rxq_inc_wr_ptr()
220 ((FIRST_RX_QUEUE + rxq->id) << 16))); in iwl_pcie_rxq_inc_wr_ptr()
222 iwl_write32(trans, RFH_Q_FRBDCB_WIDX_TRG(rxq->id), in iwl_pcie_rxq_inc_wr_ptr()
[all …]
Dctxt-info-gen3.c94 cpu_to_le64(trans_pcie->rxq->bd_dma); in iwl_pcie_ctxt_info_gen3_init()
137 cpu_to_le64(trans_pcie->rxq->rb_stts_dma); in iwl_pcie_ctxt_info_gen3_init()
139 cpu_to_le64(trans_pcie->rxq->tr_tail_dma); in iwl_pcie_ctxt_info_gen3_init()
141 cpu_to_le64(trans_pcie->rxq->cr_tail_dma); in iwl_pcie_ctxt_info_gen3_init()
149 cpu_to_le64(trans_pcie->rxq->used_bd_dma); in iwl_pcie_ctxt_info_gen3_init()
/Linux-v4.19/drivers/net/ethernet/qlogic/qede/
Dqede_fp.c51 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) in qede_alloc_rx_buffer() argument
62 if (allow_lazy && likely(rxq->filled_buffers > 12)) { in qede_alloc_rx_buffer()
63 rxq->filled_buffers--; in qede_alloc_rx_buffer()
74 mapping = dma_map_page(rxq->dev, data, 0, in qede_alloc_rx_buffer()
75 PAGE_SIZE, rxq->data_direction); in qede_alloc_rx_buffer()
76 if (unlikely(dma_mapping_error(rxq->dev, mapping))) { in qede_alloc_rx_buffer()
81 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer()
87 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring); in qede_alloc_rx_buffer()
91 rxq->rx_headroom); in qede_alloc_rx_buffer()
93 rxq->sw_rx_prod++; in qede_alloc_rx_buffer()
[all …]
Dqede_main.c872 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq)) in qede_free_fp_array()
873 xdp_rxq_info_unreg(&fp->rxq->xdp_rxq); in qede_free_fp_array()
874 kfree(fp->rxq); in qede_free_fp_array()
933 fp->rxq = kzalloc(sizeof(*fp->rxq), GFP_KERNEL); in qede_alloc_fp_array()
934 if (!fp->rxq) in qede_alloc_fp_array()
1286 struct qede_rx_queue *rxq) in qede_free_rx_buffers() argument
1290 for (i = rxq->sw_rx_cons; i != rxq->sw_rx_prod; i++) { in qede_free_rx_buffers()
1294 rx_buf = &rxq->sw_rx_ring[i & NUM_RX_BDS_MAX]; in qede_free_rx_buffers()
1298 rx_buf->mapping, PAGE_SIZE, rxq->data_direction); in qede_free_rx_buffers()
1305 static void qede_free_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) in qede_free_mem_rxq() argument
[all …]
/Linux-v4.19/drivers/net/ethernet/marvell/
Dmvneta.c124 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq) argument
744 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add() argument
751 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
757 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
763 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get() argument
767 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get()
775 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update() argument
783 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
803 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
809 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) in mvneta_rxq_next_desc_get() argument
[all …]
Dmv643xx_eth.c410 struct rx_queue rxq[8]; member
452 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument
454 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp()
462 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument
464 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable()
465 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable()
468 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument
470 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_disable()
471 u8 mask = 1 << rxq->index; in rxq_disable()
517 static int rxq_process(struct rx_queue *rxq, int budget) in rxq_process() argument
[all …]
/Linux-v4.19/drivers/infiniband/hw/hfi1/
Dvnic_main.c437 static inline int hfi1_vnic_decap_skb(struct hfi1_vnic_rx_queue *rxq, in hfi1_vnic_decap_skb() argument
440 struct hfi1_vnic_vport_info *vinfo = rxq->vinfo; in hfi1_vnic_decap_skb()
448 vinfo->stats[rxq->idx].rx_oversize++; in hfi1_vnic_decap_skb()
450 vinfo->stats[rxq->idx].rx_runt++; in hfi1_vnic_decap_skb()
456 static inline struct sk_buff *hfi1_vnic_get_skb(struct hfi1_vnic_rx_queue *rxq) in hfi1_vnic_get_skb() argument
461 skb = skb_dequeue(&rxq->skbq); in hfi1_vnic_get_skb()
474 static void hfi1_vnic_handle_rx(struct hfi1_vnic_rx_queue *rxq, in hfi1_vnic_handle_rx() argument
477 struct hfi1_vnic_vport_info *vinfo = rxq->vinfo; in hfi1_vnic_handle_rx()
485 skb = hfi1_vnic_get_skb(rxq); in hfi1_vnic_handle_rx()
489 rc = hfi1_vnic_decap_skb(rxq, skb); in hfi1_vnic_handle_rx()
[all …]
/Linux-v4.19/drivers/atm/
Dambassador.c688 amb_rxq * rxq = &dev->rxq[pool]; in rx_give() local
693 spin_lock_irqsave (&rxq->lock, flags); in rx_give()
695 if (rxq->pending < rxq->maximum) { in rx_give()
696 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); in rx_give()
698 *rxq->in.ptr = *rx; in rx_give()
699 rxq->pending++; in rx_give()
700 rxq->in.ptr = NEXTQ (rxq->in.ptr, rxq->in.start, rxq->in.limit); in rx_give()
702 wr_mem (dev, offsetof(amb_mem, mb.adapter.rx_address[pool]), virt_to_bus (rxq->in.ptr)); in rx_give()
704 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give()
707 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give()
[all …]
/Linux-v4.19/drivers/net/ethernet/atheros/alx/
Dmain.c74 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local
80 next = cur = rxq->write_idx; in alx_refill_rx_ring()
83 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
85 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring()
86 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring()
128 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
135 rxq->write_idx = cur; in alx_refill_rx_ring()
217 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) in alx_clean_rx_irq() argument
226 alx = netdev_priv(rxq->netdev); in alx_clean_rx_irq()
229 rrd = &rxq->rrd[rxq->rrd_read_idx]; in alx_clean_rx_irq()
[all …]
/Linux-v4.19/drivers/net/ethernet/hisilicon/
Dhisi_femac.c133 struct hisi_femac_queue rxq; member
224 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx_refill() local
230 pos = rxq->head; in hisi_femac_rx_refill()
232 if (!CIRC_SPACE(pos, rxq->tail, rxq->num)) in hisi_femac_rx_refill()
234 if (unlikely(rxq->skb[pos])) { in hisi_femac_rx_refill()
236 pos, rxq->skb[pos]); in hisi_femac_rx_refill()
249 rxq->dma_phys[pos] = addr; in hisi_femac_rx_refill()
250 rxq->skb[pos] = skb; in hisi_femac_rx_refill()
252 pos = (pos + 1) % rxq->num; in hisi_femac_rx_refill()
254 rxq->head = pos; in hisi_femac_rx_refill()
[all …]
/Linux-v4.19/drivers/vhost/
Dnet.c125 struct vhost_net_buf rxq; member
144 static void *vhost_net_buf_get_ptr(struct vhost_net_buf *rxq) in vhost_net_buf_get_ptr() argument
146 if (rxq->tail != rxq->head) in vhost_net_buf_get_ptr()
147 return rxq->queue[rxq->head]; in vhost_net_buf_get_ptr()
152 static int vhost_net_buf_get_size(struct vhost_net_buf *rxq) in vhost_net_buf_get_size() argument
154 return rxq->tail - rxq->head; in vhost_net_buf_get_size()
157 static int vhost_net_buf_is_empty(struct vhost_net_buf *rxq) in vhost_net_buf_is_empty() argument
159 return rxq->tail == rxq->head; in vhost_net_buf_is_empty()
162 static void *vhost_net_buf_consume(struct vhost_net_buf *rxq) in vhost_net_buf_consume() argument
164 void *ret = vhost_net_buf_get_ptr(rxq); in vhost_net_buf_consume()
[all …]
/Linux-v4.19/drivers/net/wireless/intel/iwlegacy/
D3945-mac.c951 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_queue_restock() local
956 spin_lock_irqsave(&rxq->lock, flags); in il3945_rx_queue_restock()
957 while (il_rx_queue_space(rxq) > 0 && rxq->free_count) { in il3945_rx_queue_restock()
959 element = rxq->rx_free.next; in il3945_rx_queue_restock()
964 rxq->bd[rxq->write] = in il3945_rx_queue_restock()
966 rxq->queue[rxq->write] = rxb; in il3945_rx_queue_restock()
967 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in il3945_rx_queue_restock()
968 rxq->free_count--; in il3945_rx_queue_restock()
970 spin_unlock_irqrestore(&rxq->lock, flags); in il3945_rx_queue_restock()
973 if (rxq->free_count <= RX_LOW_WATERMARK) in il3945_rx_queue_restock()
[all …]
D4965-mac.c100 il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq) in il4965_rx_queue_reset() argument
104 spin_lock_irqsave(&rxq->lock, flags); in il4965_rx_queue_reset()
105 INIT_LIST_HEAD(&rxq->rx_free); in il4965_rx_queue_reset()
106 INIT_LIST_HEAD(&rxq->rx_used); in il4965_rx_queue_reset()
111 if (rxq->pool[i].page != NULL) { in il4965_rx_queue_reset()
112 pci_unmap_page(il->pci_dev, rxq->pool[i].page_dma, in il4965_rx_queue_reset()
115 __il_free_pages(il, rxq->pool[i].page); in il4965_rx_queue_reset()
116 rxq->pool[i].page = NULL; in il4965_rx_queue_reset()
118 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il4965_rx_queue_reset()
122 rxq->queue[i] = NULL; in il4965_rx_queue_reset()
[all …]
/Linux-v4.19/samples/bpf/
Dxdp_rxq_info_user.c179 struct record *rxq; member
226 rec->rxq = alloc_record_per_rxq(); in alloc_stats_record()
228 rec->rxq[i].cpu = alloc_record_per_cpu(); in alloc_stats_record()
240 free(r->rxq[i].cpu); in free_stats_record()
242 free(r->rxq); in free_stats_record()
286 map_collect_percpu(fd, i, &rec->rxq[i]); in stats_collect()
335 int rxq; in stats_print() local
375 for (rxq = 0; rxq < nr_rxqs; rxq++) { in stats_print()
379 int rxq_ = rxq; in stats_print()
385 rec = &stats_rec->rxq[rxq]; in stats_print()
[all …]
/Linux-v4.19/drivers/net/ethernet/chelsio/cxgb4vf/
Dcxgb4vf_main.c329 int rxq, msi, err; in request_msix_queue_irqs() local
343 for_each_ethrxq(s, rxq) { in request_msix_queue_irqs()
347 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
355 while (--rxq >= 0) in request_msix_queue_irqs()
356 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
367 int rxq, msi; in free_msix_queue_irqs() local
371 for_each_ethrxq(s, rxq) in free_msix_queue_irqs()
373 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs()
398 int rxq; in enable_rx() local
401 for_each_ethrxq(s, rxq) in enable_rx()
[all …]
Dsge.c1567 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument
1570 struct adapter *adapter = rxq->rspq.adapter; in do_gro()
1576 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1579 rxq->stats.rx_drops++; in do_gro()
1588 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1594 rxq->stats.vlan_ex++; in do_gro()
1596 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1599 rxq->stats.lro_pkts++; in do_gro()
1601 rxq->stats.lro_merged++; in do_gro()
1602 rxq->stats.pkts++; in do_gro()
[all …]
/Linux-v4.19/drivers/net/ethernet/marvell/mvpp2/
Dmvpp2_main.c756 int rxq; in mvpp2_swf_bm_pool_init() local
780 for (rxq = 0; rxq < port->nrxqs; rxq++) in mvpp2_swf_bm_pool_init()
781 mvpp2_rxq_long_pool_set(port, rxq, port->pool_long->id); in mvpp2_swf_bm_pool_init()
793 for (rxq = 0; rxq < port->nrxqs; rxq++) in mvpp2_swf_bm_pool_init()
794 mvpp2_rxq_short_pool_set(port, rxq, in mvpp2_swf_bm_pool_init()
1579 mvpp2_rxq_next_desc_get(struct mvpp2_rx_queue *rxq) in mvpp2_rxq_next_desc_get() argument
1581 int rx_desc = rxq->next_desc_to_proc; in mvpp2_rxq_next_desc_get()
1583 rxq->next_desc_to_proc = MVPP2_QUEUE_NEXT_DESC(rxq, rx_desc); in mvpp2_rxq_next_desc_get()
1584 prefetch(rxq->descs + rxq->next_desc_to_proc); in mvpp2_rxq_next_desc_get()
1585 return rxq->descs + rx_desc; in mvpp2_rxq_next_desc_get()
[all …]
/Linux-v4.19/drivers/net/ethernet/alacritech/
Dslicoss.c130 static unsigned int slic_get_free_rx_descs(struct slic_rx_queue *rxq) in slic_get_free_rx_descs() argument
132 return slic_get_free_queue_descs(rxq->put_idx, rxq->done_idx, rxq->len); in slic_get_free_rx_descs()
408 struct slic_rx_queue *rxq = &sdev->rxq; in slic_refill_rx_queue() local
417 while (slic_get_free_rx_descs(rxq) > SLIC_MAX_REQ_RX_DESCS) { in slic_refill_rx_queue()
449 buff = &rxq->rxbuffs[rxq->put_idx]; in slic_refill_rx_queue()
458 rxq->put_idx = slic_next_queue_idx(rxq->put_idx, rxq->len); in slic_refill_rx_queue()
561 struct slic_rx_queue *rxq = &sdev->rxq; in slic_handle_receive() local
571 while (todo && (rxq->done_idx != rxq->put_idx)) { in slic_handle_receive()
572 buff = &rxq->rxbuffs[rxq->done_idx]; in slic_handle_receive()
625 rxq->done_idx = slic_next_queue_idx(rxq->done_idx, rxq->len); in slic_handle_receive()
[all …]
/Linux-v4.19/drivers/target/iscsi/cxgbit/
Dcxgbit_main.c103 spin_lock_bh(&csk->rxq.lock); in cxgbit_close_conn()
104 __skb_queue_tail(&csk->rxq, skb); in cxgbit_close_conn()
105 if (skb_queue_len(&csk->rxq) == 1) in cxgbit_close_conn()
107 spin_unlock_bh(&csk->rxq.lock); in cxgbit_close_conn()
339 spin_lock(&csk->rxq.lock); in cxgbit_queue_lro_skb()
340 __skb_queue_tail(&csk->rxq, skb); in cxgbit_queue_lro_skb()
341 if (skb_queue_len(&csk->rxq) == 1) in cxgbit_queue_lro_skb()
343 spin_unlock(&csk->rxq.lock); in cxgbit_queue_lro_skb()
568 spin_lock(&csk->rxq.lock); in cxgbit_update_dcb_priority()
569 __skb_queue_tail(&csk->rxq, skb); in cxgbit_update_dcb_priority()
[all …]
/Linux-v4.19/drivers/net/ethernet/freescale/
Dgianfar.h1311 static inline int gfar_rxbd_unused(struct gfar_priv_rx_q *rxq) in gfar_rxbd_unused() argument
1313 if (rxq->next_to_clean > rxq->next_to_use) in gfar_rxbd_unused()
1314 return rxq->next_to_clean - rxq->next_to_use - 1; in gfar_rxbd_unused()
1316 return rxq->rx_ring_size + rxq->next_to_clean - rxq->next_to_use - 1; in gfar_rxbd_unused()
1319 static inline u32 gfar_rxbd_dma_lastfree(struct gfar_priv_rx_q *rxq) in gfar_rxbd_dma_lastfree() argument
1325 i = rxq->next_to_use ? rxq->next_to_use - 1 : rxq->rx_ring_size - 1; in gfar_rxbd_dma_lastfree()
1326 bdp = &rxq->rx_bd_base[i]; in gfar_rxbd_dma_lastfree()
1327 bdp_dma = lower_32_bits(rxq->rx_bd_dma_base); in gfar_rxbd_dma_lastfree()
1328 bdp_dma += (uintptr_t)bdp - (uintptr_t)rxq->rx_bd_base; in gfar_rxbd_dma_lastfree()
/Linux-v4.19/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c329 priv->hw->desc->init_rx_desc(&priv->rxq[j]->dma_rx[i], in sxgbe_clear_descriptors()
584 priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings()
593 priv->rxq[queue_num]->priv_ptr = priv; in init_dma_desc_rings()
607 free_rx_ring(priv->device, priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings()
659 free_rx_ring(priv->device, priv->rxq[queue_num], rx_rsize); in free_dma_desc_resources()
682 priv->rxq[queue_num] = devm_kmalloc(priv->device, in rxring_mem_alloc()
684 if (!priv->rxq[queue_num]) in rxring_mem_alloc()
964 (priv->rxq[queue_num])->dma_rx_phy, in sxgbe_init_dma_engine()
1138 (priv->rxq[queue_num])->irq_no, in sxgbe_open()
1140 dev->name, priv->rxq[queue_num]); in sxgbe_open()
[all …]
/Linux-v4.19/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_vfpf.c636 req->rxq.rcq_addr = fp->rx_comp_mapping; in bnx2x_vfpf_setup_q()
637 req->rxq.rcq_np_addr = fp->rx_comp_mapping + BCM_PAGE_SIZE; in bnx2x_vfpf_setup_q()
638 req->rxq.rxq_addr = fp->rx_desc_mapping; in bnx2x_vfpf_setup_q()
639 req->rxq.sge_addr = fp->rx_sge_mapping; in bnx2x_vfpf_setup_q()
640 req->rxq.vf_sb = fp_idx; in bnx2x_vfpf_setup_q()
641 req->rxq.sb_index = HC_INDEX_ETH_RX_CQ_CONS; in bnx2x_vfpf_setup_q()
642 req->rxq.hc_rate = bp->rx_ticks ? 1000000/bp->rx_ticks : 0; in bnx2x_vfpf_setup_q()
643 req->rxq.mtu = bp->dev->mtu; in bnx2x_vfpf_setup_q()
644 req->rxq.buf_sz = fp->rx_buf_size; in bnx2x_vfpf_setup_q()
645 req->rxq.sge_buf_sz = BCM_PAGE_SIZE * PAGES_PER_SGE; in bnx2x_vfpf_setup_q()
[all …]
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/
Den_arfs.c59 int rxq; member
384 arfs_rule->rxq, arfs_rule->flow_id, in arfs_may_expire_flow()
542 dest.tir_num = priv->direct_tir[arfs_rule->rxq].tirn; in arfs_add_rule()
547 __func__, arfs_rule->filter_id, arfs_rule->rxq, err); in arfs_add_rule()
556 struct mlx5_flow_handle *rule, u16 rxq) in arfs_modify_rule_rq() argument
562 dst.tir_num = priv->direct_tir[rxq].tirn; in arfs_modify_rule_rq()
566 "Failed to modify aRFS rule destination to rq=%d\n", rxq); in arfs_modify_rule_rq()
596 arfs_rule->rxq); in arfs_handle_work()
627 u16 rxq, u32 flow_id) in arfs_alloc_rule() argument
637 rule->rxq = rxq; in arfs_alloc_rule()
[all …]

123456