Lines Matching +full:0 +full:x0000001
36 #define DMA_ADDR_HIGH(dma_addr) ((u32)(((dma_addr) >> 32) & 0xFFFFFFFF))
38 #define DMA_ADDR_HIGH(dma_addr) ((u32)(0))
40 #define DMA_ADDR_LOW(dma_addr) ((u32)((dma_addr) & 0xFFFFFFFF))
59 if (active != 0) in tsnep_irq()
63 if ((active & ECM_INT_LINK) != 0) in tsnep_irq()
66 /* handle TX/RX queue 0 interrupt */ in tsnep_irq()
67 if ((active & adapter->queue[0].irq_mask) != 0) { in tsnep_irq()
68 tsnep_disable_irq(adapter, adapter->queue[0].irq_mask); in tsnep_irq()
69 napi_schedule(&adapter->queue[0].napi); in tsnep_irq()
103 if (retval != 0) in tsnep_mdiobus_read()
128 if (retval != 0) in tsnep_mdiobus_write()
131 return 0; in tsnep_mdiobus_write()
201 memset(ðtool_eee, 0, sizeof(ethtool_eee)); in tsnep_phy_open()
207 return 0; in tsnep_phy_open()
222 memset(tx->entry, 0, sizeof(tx->entry)); in tsnep_tx_ring_cleanup()
224 for (i = 0; i < TSNEP_RING_PAGE_COUNT; i++) { in tsnep_tx_ring_cleanup()
229 tx->page_dma[i] = 0; in tsnep_tx_ring_cleanup()
242 for (i = 0; i < TSNEP_RING_PAGE_COUNT; i++) { in tsnep_tx_ring_init()
250 for (j = 0; j < TSNEP_RING_ENTRIES_PER_PAGE; j++) { in tsnep_tx_ring_init()
259 for (i = 0; i < TSNEP_RING_SIZE; i++) { in tsnep_tx_ring_init()
265 return 0; in tsnep_tx_ring_init()
277 entry->properties = 0; in tsnep_tx_activate()
311 if (tx->increment_owner_counter < 0) in tsnep_tx_activate()
344 int map_len = 0; in tsnep_tx_map()
347 for (i = 0; i < count; i++) { in tsnep_tx_map()
350 if (i == 0) { in tsnep_tx_map()
358 0, len, DMA_TO_DEVICE); in tsnep_tx_map()
378 int map_len = 0; in tsnep_tx_unmap()
381 for (i = 0; i < count; i++) { in tsnep_tx_unmap()
385 if (i == 0) in tsnep_tx_unmap()
396 entry->len = 0; in tsnep_tx_unmap()
413 if (skb_shinfo(skb)->nr_frags > 0) in tsnep_xmit_frame_ring()
433 if (retval < 0) { in tsnep_xmit_frame_ring()
451 for (i = 0; i < count; i++) in tsnep_xmit_frame_ring()
499 if (skb_shinfo(entry->skb)->nr_frags > 0) in tsnep_tx_poll()
518 memset(&hwtstamps, 0, sizeof(hwtstamps)); in tsnep_tx_poll()
542 return (budget != 0); in tsnep_tx_poll()
572 memset(tx, 0, sizeof(*tx)); in tsnep_tx_open()
581 dma = tx->entry[0].desc_dma | TSNEP_RESET_OWNER_COUNTER; in tsnep_tx_open()
589 return 0; in tsnep_tx_open()
597 ((val & TSNEP_CONTROL_TX_ENABLE) == 0), 10000, in tsnep_tx_close()
609 for (i = 0; i < TSNEP_RING_SIZE; i++) { in tsnep_rx_ring_cleanup()
620 memset(rx->entry, 0, sizeof(rx->entry)); in tsnep_rx_ring_cleanup()
622 for (i = 0; i < TSNEP_RING_PAGE_COUNT; i++) { in tsnep_rx_ring_cleanup()
627 rx->page_dma[i] = 0; in tsnep_rx_ring_cleanup()
646 return 0; in tsnep_rx_alloc_buffer()
653 struct page_pool_params pp_params = { 0 }; in tsnep_rx_ring_init()
658 for (i = 0; i < TSNEP_RING_PAGE_COUNT; i++) { in tsnep_rx_ring_init()
666 for (j = 0; j < TSNEP_RING_ENTRIES_PER_PAGE; j++) { in tsnep_rx_ring_init()
677 pp_params.order = 0; in tsnep_rx_ring_init()
691 for (i = 0; i < TSNEP_RING_SIZE; i++) { in tsnep_rx_ring_init()
701 return 0; in tsnep_rx_ring_init()
720 if (rx->increment_owner_counter < 0) in tsnep_rx_activate()
756 memset(hwtstamps, 0, sizeof(*hwtstamps)); in tsnep_build_skb()
770 int done = 0; in tsnep_rx_poll()
865 memset(rx, 0, sizeof(*rx)); in tsnep_rx_open()
874 dma = rx->entry[0].desc_dma | TSNEP_RESET_OWNER_COUNTER; in tsnep_rx_open()
880 for (i = 0; i < TSNEP_RING_SIZE; i++) in tsnep_rx_open()
888 return 0; in tsnep_rx_open()
897 ((val & TSNEP_CONTROL_RX_ENABLE) == 0), 10000, in tsnep_rx_close()
919 int done = 0; in tsnep_poll()
975 retval = request_irq(queue->irq, handler, 0, queue->name, dev); in tsnep_request_irq()
978 memset(queue->name, 0, sizeof(queue->name)); in tsnep_request_irq()
997 memset(queue->name, 0, sizeof(queue->name)); in tsnep_free_irq()
1005 int tx_queue_index = 0; in tsnep_netdev_open()
1006 int rx_queue_index = 0; in tsnep_netdev_open()
1009 for (i = 0; i < adapter->num_queues; i++) { in tsnep_netdev_open()
1029 retval = tsnep_request_irq(&adapter->queue[i], i == 0); in tsnep_netdev_open()
1052 for (i = 0; i < adapter->num_queues; i++) { in tsnep_netdev_open()
1060 return 0; in tsnep_netdev_open()
1066 for (i = 0; i < adapter->num_queues; i++) { in tsnep_netdev_open()
1067 tsnep_free_irq(&adapter->queue[i], i == 0); in tsnep_netdev_open()
1085 for (i = 0; i < adapter->num_queues; i++) { in tsnep_netdev_close()
1091 tsnep_free_irq(&adapter->queue[i], i == 0); in tsnep_netdev_close()
1099 return 0; in tsnep_netdev_close()
1109 queue_mapping = 0; in tsnep_netdev_xmit_frame()
1128 u16 rx_filter = 0; in tsnep_netdev_set_multicast()
1148 for (i = 0; i < adapter->num_tx_queues; i++) { in tsnep_netdev_get_stats64()
1153 for (i = 0; i < adapter->num_rx_queues; i++) { in tsnep_netdev_get_stats64()
1210 return 0; in tsnep_netdev_set_mac_address()
1219 int retval = 0; in tsnep_netdev_set_features()
1264 iowrite16(0, adapter->addr + TSNEP_RX_FILTER); in tsnep_mac_init()
1288 return 0; in tsnep_mac_init()
1299 return 0; in tsnep_mdio_init()
1321 adapter->mdiobus->phy_mask = 0x0000001; in tsnep_mdio_init()
1342 0); in tsnep_phy_init()
1350 return 0; in tsnep_phy_init()
1362 retval = platform_get_irq(adapter->pdev, 0); in tsnep_queue_init()
1365 if (retval < 0) in tsnep_queue_init()
1370 adapter->queue[0].irq = retval; in tsnep_queue_init()
1371 adapter->queue[0].tx = &adapter->tx[0]; in tsnep_queue_init()
1372 adapter->queue[0].rx = &adapter->rx[0]; in tsnep_queue_init()
1373 adapter->queue[0].irq_mask = irq_mask; in tsnep_queue_init()
1375 adapter->netdev->irq = adapter->queue[0].irq; in tsnep_queue_init()
1383 if (retval < 0) in tsnep_queue_init()
1396 return 0; in tsnep_queue_init()
1432 io = platform_get_resource(pdev, IORESOURCE_MEM, 0); in tsnep_probe()
1500 return 0; in tsnep_probe()
1533 return 0; in tsnep_remove()