Home
last modified time | relevance | path

Searched refs:rx_desc (Results 1 – 25 of 104) sorted by relevance

12345

/Linux-v5.4/drivers/net/ethernet/intel/fm10k/
Dfm10k_main.c114 union fm10k_rx_desc *rx_desc; in fm10k_alloc_rx_buffers() local
122 rx_desc = FM10K_RX_DESC(rx_ring, i); in fm10k_alloc_rx_buffers()
133 rx_desc->q.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in fm10k_alloc_rx_buffers()
135 rx_desc++; in fm10k_alloc_rx_buffers()
139 rx_desc = FM10K_RX_DESC(rx_ring, 0); in fm10k_alloc_rx_buffers()
145 rx_desc->d.staterr = 0; in fm10k_alloc_rx_buffers()
253 union fm10k_rx_desc *rx_desc, in fm10k_add_rx_frag() argument
300 union fm10k_rx_desc *rx_desc, in fm10k_fetch_rx_buffer() argument
303 unsigned int size = le16_to_cpu(rx_desc->w.length); in fm10k_fetch_rx_buffer()
344 if (fm10k_add_rx_frag(rx_buffer, size, rx_desc, skb)) { in fm10k_fetch_rx_buffer()
[all …]
/Linux-v5.4/drivers/net/ethernet/seeq/
Dsgiseeq.c59 (unsigned long)((sp)->rx_desc)))
95 struct sgiseeq_rx_desc *rx_desc; member
198 if (!sp->rx_desc[i].skb) { in seeq_init_ring()
208 sp->rx_desc[i].skb = skb; in seeq_init_ring()
209 sp->rx_desc[i].rdma.pbuf = dma_addr; in seeq_init_ring()
211 sp->rx_desc[i].rdma.cntinfo = RCNTINFO_INIT; in seeq_init_ring()
212 dma_sync_desc_dev(dev, &sp->rx_desc[i]); in seeq_init_ring()
214 sp->rx_desc[i - 1].rdma.cntinfo |= HPCDMA_EOR; in seeq_init_ring()
215 dma_sync_desc_dev(dev, &sp->rx_desc[i - 1]); in seeq_init_ring()
234 if (sp->rx_desc[i].skb) { in seeq_purge_ring()
[all …]
/Linux-v5.4/drivers/infiniband/ulp/isert/
Dib_isert.c171 struct iser_rx_desc *rx_desc; in isert_alloc_rx_descriptors() local
182 rx_desc = isert_conn->rx_descs; in isert_alloc_rx_descriptors()
184 for (i = 0; i < ISERT_QP_MAX_RECV_DTOS; i++, rx_desc++) { in isert_alloc_rx_descriptors()
185 dma_addr = ib_dma_map_single(ib_dev, (void *)rx_desc, in isert_alloc_rx_descriptors()
190 rx_desc->dma_addr = dma_addr; in isert_alloc_rx_descriptors()
192 rx_sg = &rx_desc->rx_sg; in isert_alloc_rx_descriptors()
193 rx_sg->addr = rx_desc->dma_addr; in isert_alloc_rx_descriptors()
196 rx_desc->rx_cqe.done = isert_recv_done; in isert_alloc_rx_descriptors()
202 rx_desc = isert_conn->rx_descs; in isert_alloc_rx_descriptors()
203 for (j = 0; j < i; j++, rx_desc++) { in isert_alloc_rx_descriptors()
[all …]
/Linux-v5.4/drivers/net/ethernet/hisilicon/hns3/
Dhns3_debugfs.c164 struct hns3_desc *rx_desc, *tx_desc; in hns3_dbg_bd_info() local
220 rx_desc = &ring->desc[rx_index]; in hns3_dbg_bd_info()
222 addr = le64_to_cpu(rx_desc->addr); in hns3_dbg_bd_info()
225 dev_info(dev, "(RX)l234_info: %u\n", rx_desc->rx.l234_info); in hns3_dbg_bd_info()
226 dev_info(dev, "(RX)pkt_len: %u\n", rx_desc->rx.pkt_len); in hns3_dbg_bd_info()
227 dev_info(dev, "(RX)size: %u\n", rx_desc->rx.size); in hns3_dbg_bd_info()
228 dev_info(dev, "(RX)rss_hash: %u\n", rx_desc->rx.rss_hash); in hns3_dbg_bd_info()
229 dev_info(dev, "(RX)fd_id: %u\n", rx_desc->rx.fd_id); in hns3_dbg_bd_info()
230 dev_info(dev, "(RX)vlan_tag: %u\n", rx_desc->rx.vlan_tag); in hns3_dbg_bd_info()
231 dev_info(dev, "(RX)o_dm_vlan_id_fb: %u\n", rx_desc->rx.o_dm_vlan_id_fb); in hns3_dbg_bd_info()
[all …]
/Linux-v5.4/drivers/spi/
Dspi-pxa2xx-dma.c143 struct dma_async_tx_descriptor *tx_desc, *rx_desc; in pxa2xx_spi_dma_prepare() local
154 rx_desc = pxa2xx_spi_dma_prepare_one(drv_data, DMA_DEV_TO_MEM, xfer); in pxa2xx_spi_dma_prepare()
155 if (!rx_desc) { in pxa2xx_spi_dma_prepare()
163 rx_desc->callback = pxa2xx_spi_dma_callback; in pxa2xx_spi_dma_prepare()
164 rx_desc->callback_param = drv_data; in pxa2xx_spi_dma_prepare()
166 dmaengine_submit(rx_desc); in pxa2xx_spi_dma_prepare()
/Linux-v5.4/drivers/net/ethernet/intel/ice/
Dice_txrx.c469 union ice_32b_rx_flex_desc *rx_desc; in ice_alloc_rx_bufs() local
478 rx_desc = ICE_RX_DESC(rx_ring, ntu); in ice_alloc_rx_bufs()
495 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in ice_alloc_rx_bufs()
497 rx_desc++; in ice_alloc_rx_bufs()
501 rx_desc = ICE_RX_DESC(rx_ring, 0); in ice_alloc_rx_bufs()
507 rx_desc->wb.status_error0 = 0; in ice_alloc_rx_bufs()
804 ice_test_staterr(union ice_32b_rx_flex_desc *rx_desc, const u16 stat_err_bits) in ice_test_staterr() argument
806 return !!(rx_desc->wb.status_error0 & in ice_test_staterr()
822 ice_is_non_eop(struct ice_ring *rx_ring, union ice_32b_rx_flex_desc *rx_desc, in ice_is_non_eop() argument
835 if (likely(ice_test_staterr(rx_desc, ICE_RXD_EOF))) in ice_is_non_eop()
[all …]
/Linux-v5.4/drivers/infiniband/ulp/iser/
Diser_initiator.c244 struct iser_rx_desc *rx_desc; in iser_alloc_rx_descriptors() local
267 rx_desc = iser_conn->rx_descs; in iser_alloc_rx_descriptors()
269 for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) { in iser_alloc_rx_descriptors()
270 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, in iser_alloc_rx_descriptors()
275 rx_desc->dma_addr = dma_addr; in iser_alloc_rx_descriptors()
276 rx_desc->cqe.done = iser_task_rsp; in iser_alloc_rx_descriptors()
277 rx_sg = &rx_desc->rx_sg; in iser_alloc_rx_descriptors()
278 rx_sg->addr = rx_desc->dma_addr; in iser_alloc_rx_descriptors()
287 rx_desc = iser_conn->rx_descs; in iser_alloc_rx_descriptors()
288 for (j = 0; j < i; j++, rx_desc++) in iser_alloc_rx_descriptors()
[all …]
/Linux-v5.4/drivers/net/ethernet/intel/iavf/
Diavf_txrx.c882 union iavf_rx_desc *rx_desc; in iavf_alloc_rx_buffers() local
889 rx_desc = IAVF_RX_DESC(rx_ring, ntu); in iavf_alloc_rx_buffers()
905 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in iavf_alloc_rx_buffers()
907 rx_desc++; in iavf_alloc_rx_buffers()
911 rx_desc = IAVF_RX_DESC(rx_ring, 0); in iavf_alloc_rx_buffers()
917 rx_desc->wb.qword1.status_error_len = 0; in iavf_alloc_rx_buffers()
945 union iavf_rx_desc *rx_desc) in iavf_rx_checksum() argument
953 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in iavf_rx_checksum()
1052 union iavf_rx_desc *rx_desc, in iavf_rx_hash() argument
1064 if ((rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) { in iavf_rx_hash()
[all …]
/Linux-v5.4/drivers/net/ethernet/intel/i40e/
Di40e_txrx_common.h8 union i40e_rx_desc *rx_desc, u8 prog_id);
12 union i40e_rx_desc *rx_desc,
15 union i40e_rx_desc *rx_desc, struct sk_buff *skb);
Di40e_xsk.c319 union i40e_rx_desc *rx_desc; in __i40e_alloc_rx_buffers_zc() local
323 rx_desc = I40E_RX_DESC(rx_ring, ntu); in __i40e_alloc_rx_buffers_zc()
335 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma); in __i40e_alloc_rx_buffers_zc()
337 rx_desc++; in __i40e_alloc_rx_buffers_zc()
342 rx_desc = I40E_RX_DESC(rx_ring, 0); in __i40e_alloc_rx_buffers_zc()
347 rx_desc->wb.qword1.status_error_len = 0; in __i40e_alloc_rx_buffers_zc()
543 union i40e_rx_desc *rx_desc; in i40e_clean_rx_irq_zc() local
554 rx_desc = I40E_RX_DESC(rx_ring, rx_ring->next_to_clean); in i40e_clean_rx_irq_zc()
555 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_rx_irq_zc()
563 bi = i40e_clean_programming_status(rx_ring, rx_desc, in i40e_clean_rx_irq_zc()
[all …]
Di40e_txrx.c531 union i40e_rx_desc *rx_desc, u8 prog_id) in i40e_fd_handle_status() argument
539 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_fd_handle_status()
544 pf->fd_inv = le32_to_cpu(rx_desc->wb.qword0.hi_dword.fd_id); in i40e_fd_handle_status()
545 if ((rx_desc->wb.qword0.hi_dword.fd_id != 0) || in i40e_fd_handle_status()
563 if ((rx_desc->wb.qword0.hi_dword.fd_id == 0) && in i40e_fd_handle_status()
592 rx_desc->wb.qword0.hi_dword.fd_id); in i40e_fd_handle_status()
1262 union i40e_rx_desc *rx_desc, in i40e_clean_programming_status() argument
1285 i40e_fd_handle_status(rx_ring, rx_desc, id); in i40e_clean_programming_status()
1571 union i40e_rx_desc *rx_desc; in i40e_alloc_rx_buffers() local
1578 rx_desc = I40E_RX_DESC(rx_ring, ntu); in i40e_alloc_rx_buffers()
[all …]
/Linux-v5.4/drivers/net/ethernet/alteon/
Dacenic.h477 #define RX_STD_RING_SIZE (RX_STD_RING_ENTRIES * sizeof(struct rx_desc))
480 #define RX_JUMBO_RING_SIZE (RX_JUMBO_RING_ENTRIES *sizeof(struct rx_desc))
483 #define RX_MINI_RING_SIZE (RX_MINI_RING_ENTRIES *sizeof(struct rx_desc))
487 sizeof(struct rx_desc))
489 struct rx_desc{ struct
664 struct rx_desc *rx_std_ring;
665 struct rx_desc *rx_jumbo_ring;
666 struct rx_desc *rx_mini_ring;
667 struct rx_desc *rx_return_ring;
/Linux-v5.4/drivers/net/wireless/ti/wl1251/
Drx.c206 struct wl1251_rx_descriptor *rx_desc; in wl1251_rx() local
211 rx_desc = wl->rx_descriptor; in wl1251_rx()
214 wl1251_rx_header(wl, rx_desc); in wl1251_rx()
217 wl1251_rx_body(wl, rx_desc); in wl1251_rx()
/Linux-v5.4/drivers/net/ethernet/intel/ixgbe/
Dixgbe_xsk.c316 union ixgbe_adv_rx_desc *rx_desc; in __ixgbe_alloc_rx_buffers_zc() local
325 rx_desc = IXGBE_RX_DESC(rx_ring, i); in __ixgbe_alloc_rx_buffers_zc()
344 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma); in __ixgbe_alloc_rx_buffers_zc()
346 rx_desc++; in __ixgbe_alloc_rx_buffers_zc()
350 rx_desc = IXGBE_RX_DESC(rx_ring, 0); in __ixgbe_alloc_rx_buffers_zc()
356 rx_desc->wb.upper.length = 0; in __ixgbe_alloc_rx_buffers_zc()
442 union ixgbe_adv_rx_desc *rx_desc; in ixgbe_clean_rx_irq_zc() local
454 rx_desc = IXGBE_RX_DESC(rx_ring, rx_ring->next_to_clean); in ixgbe_clean_rx_irq_zc()
455 size = le16_to_cpu(rx_desc->wb.upper.length); in ixgbe_clean_rx_irq_zc()
467 if (unlikely(!ixgbe_test_staterr(rx_desc, in ixgbe_clean_rx_irq_zc()
[all …]
Dixgbe_txrx_common.h18 union ixgbe_adv_rx_desc *rx_desc,
21 union ixgbe_adv_rx_desc *rx_desc,
/Linux-v5.4/drivers/net/ethernet/
Dec_bhf.c87 struct rx_desc { struct
139 struct rx_desc *rx_descs;
194 static int ec_bhf_pkt_received(struct rx_desc *desc) in ec_bhf_pkt_received()
199 static void ec_bhf_add_rx_desc(struct ec_bhf_priv *priv, struct rx_desc *desc) in ec_bhf_add_rx_desc()
207 struct rx_desc *desc = &priv->rx_descs[priv->rx_dnext]; in ec_bhf_process_rx()
373 priv->rx_dcount = priv->rx_buf.len / sizeof(struct rx_desc); in ec_bhf_setup_rx_descs()
374 priv->rx_descs = (struct rx_desc *)priv->rx_buf.buf; in ec_bhf_setup_rx_descs()
378 struct rx_desc *desc = &priv->rx_descs[i]; in ec_bhf_setup_rx_descs()
401 FIFO_SIZE * sizeof(struct rx_desc)); in ec_bhf_open()
/Linux-v5.4/drivers/net/ethernet/google/gve/
Dgve_rx.c275 static bool gve_rx(struct gve_rx_ring *rx, struct gve_rx_desc *rx_desc, in gve_rx() argument
287 if (unlikely(rx_desc->flags_seq & GVE_RXF_ERR)) in gve_rx()
290 len = be16_to_cpu(rx_desc->len) - GVE_RX_PAD; in gve_rx()
345 if (rx_desc->csum) in gve_rx()
349 skb->csum = csum_unfold(rx_desc->csum); in gve_rx()
354 gve_needs_rss(rx_desc->flags_seq)) in gve_rx()
355 skb_set_hash(skb, be32_to_cpu(rx_desc->rss_hash), in gve_rx()
356 gve_rss_type(rx_desc->flags_seq)); in gve_rx()
/Linux-v5.4/drivers/net/ethernet/marvell/
Dpxa168_eth.c175 struct rx_desc { struct
207 struct rx_desc *p_rx_desc_area;
307 struct rx_desc *p_used_rx_desc; in rxq_refill()
643 (u32) (pep->rx_desc_dma + rx_curr_desc * sizeof(struct rx_desc))); in eth_port_start()
646 (u32) (pep->rx_desc_dma + rx_curr_desc * sizeof(struct rx_desc))); in eth_port_start()
773 struct rx_desc *rx_desc; in rxq_process() local
781 rx_desc = &pep->p_rx_desc_area[rx_curr_desc]; in rxq_process()
782 cmd_sts = rx_desc->cmd_sts; in rxq_process()
797 dma_unmap_single(&pep->pdev->dev, rx_desc->buf_ptr, in rxq_process()
798 rx_desc->buf_size, in rxq_process()
[all …]
Dmvneta.c815 int rx_desc = rxq->next_desc_to_proc; in mvneta_rxq_next_desc_get() local
817 rxq->next_desc_to_proc = MVNETA_QUEUE_NEXT_DESC(rxq, rx_desc); in mvneta_rxq_next_desc_get()
819 return rxq->descs + rx_desc; in mvneta_rxq_next_desc_get()
1612 static void mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, in mvneta_rx_desc_fill() argument
1618 rx_desc->buf_phys_addr = phys_addr; in mvneta_rx_desc_fill()
1619 i = rx_desc - rxq->descs; in mvneta_rx_desc_fill()
1704 struct mvneta_rx_desc *rx_desc) in mvneta_rx_error() argument
1706 u32 status = rx_desc->status; in mvneta_rx_error()
1711 status, rx_desc->data_size); in mvneta_rx_error()
1715 status, rx_desc->data_size); in mvneta_rx_error()
[all …]
/Linux-v5.4/drivers/net/wireless/ti/wlcore/
Dhw_ops.h45 wlcore_hw_get_rx_buf_align(struct wl1271 *wl, u32 rx_desc) in wlcore_hw_get_rx_buf_align() argument
51 return wl->ops->get_rx_buf_align(wl, rx_desc); in wlcore_hw_get_rx_buf_align()
55 wlcore_hw_prepare_read(struct wl1271 *wl, u32 rx_desc, u32 len) in wlcore_hw_prepare_read() argument
58 return wl->ops->prepare_read(wl, rx_desc, len); in wlcore_hw_prepare_read()
/Linux-v5.4/drivers/staging/mt7621-dma/
Dmtk-hsdma.c236 struct hsdma_desc *rx_desc; in hsdma_dump_desc() local
244 rx_desc = &chan->rx_ring[i]; in hsdma_dump_desc()
249 tx_desc->addr1, rx_desc->addr0, rx_desc->flags); in hsdma_dump_desc()
320 struct hsdma_desc *tx_desc, *rx_desc; in mtk_hsdma_start_transfer() local
363 rx_desc = &chan->rx_ring[rx_idx]; in mtk_hsdma_start_transfer()
369 rx_desc->addr0 = dst; in mtk_hsdma_start_transfer()
370 rx_desc->flags = HSDMA_DESC_PLEN0(tlen); in mtk_hsdma_start_transfer()
/Linux-v5.4/drivers/net/ethernet/intel/ixgbevf/
Dixgbevf_main.c444 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_rx_hash() argument
452 rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) & in ixgbevf_rx_hash()
458 skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in ixgbevf_rx_hash()
470 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_rx_checksum() argument
480 if (ixgbevf_test_staterr(rx_desc, IXGBE_RXD_STAT_IPCS) && in ixgbevf_rx_checksum()
481 ixgbevf_test_staterr(rx_desc, IXGBE_RXDADV_ERR_IPE)) { in ixgbevf_rx_checksum()
486 if (!ixgbevf_test_staterr(rx_desc, IXGBE_RXD_STAT_L4CS)) in ixgbevf_rx_checksum()
489 if (ixgbevf_test_staterr(rx_desc, IXGBE_RXDADV_ERR_TCPE)) { in ixgbevf_rx_checksum()
509 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_process_skb_fields() argument
512 ixgbevf_rx_hash(rx_ring, rx_desc, skb); in ixgbevf_process_skb_fields()
[all …]
Dixgbevf.h280 static inline __le32 ixgbevf_test_staterr(union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_test_staterr() argument
283 return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits); in ixgbevf_test_staterr()
467 union ixgbe_adv_rx_desc *rx_desc,
479 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_ipsec_rx() argument
/Linux-v5.4/drivers/atm/
Dhorizon.c713 rx_ch_desc * rx_desc = &memmap->rx_descs[channel]; in hrz_open_rx() local
718 channel_type = rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK; in hrz_open_rx()
746 wr_mem (dev, &rx_desc->wr_buf_type, in hrz_open_rx()
749 wr_mem (dev, &rx_desc->rd_buf_type, buf_ptr); in hrz_open_rx()
787 rx_ch_desc * rx_desc = &memmap->rx_descs[vc]; in hrz_close_rx() local
792 value = rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK; in hrz_close_rx()
806 wr_mem (dev, &rx_desc->wr_buf_type, RX_CHANNEL_DISABLED); in hrz_close_rx()
808 if ((rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK) == RX_CHANNEL_DISABLED) in hrz_close_rx()
852 r1 = rd_mem (dev, &rx_desc->rd_buf_type); in hrz_close_rx()
870 r2 = rd_mem (dev, &rx_desc->rd_buf_type); in hrz_close_rx()
[all …]
/Linux-v5.4/drivers/net/ethernet/intel/ixgb/
Dixgb_main.c1877 struct ixgb_rx_desc *rx_desc, in ixgb_rx_checksum() argument
1883 if ((rx_desc->status & IXGB_RX_DESC_STATUS_IXSM) || in ixgb_rx_checksum()
1884 (!(rx_desc->status & IXGB_RX_DESC_STATUS_TCPCS))) { in ixgb_rx_checksum()
1891 if (rx_desc->errors & IXGB_RX_DESC_ERRORS_TCPE) { in ixgb_rx_checksum()
1938 struct ixgb_rx_desc *rx_desc, *next_rxd; in ixgb_clean_rx_irq() local
1946 rx_desc = IXGB_RX_DESC(*rx_ring, i); in ixgb_clean_rx_irq()
1949 while (rx_desc->status & IXGB_RX_DESC_STATUS_DD) { in ixgb_clean_rx_irq()
1958 status = rx_desc->status; in ixgb_clean_rx_irq()
1986 length = le16_to_cpu(rx_desc->length); in ixgb_clean_rx_irq()
1987 rx_desc->length = 0; in ixgb_clean_rx_irq()
[all …]

12345