Lines Matching refs:vpaths

127 		VXGE_COMPLETE_VPATH_TX(&vdev->vpaths[i].fifo);  in VXGE_COMPLETE_ALL_TX()
137 ring = &vdev->vpaths[i].ring; in VXGE_COMPLETE_ALL_RX()
710 vpath = &vdev->vpaths[mac->vpath_no]; in vxge_add_mac_addr()
738 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
748 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
764 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
860 fifo = &vdev->vpaths[vpath_no].fifo; in vxge_xmit()
1097 vpath = &vdev->vpaths[mac->vpath_no]; in vxge_del_mac_addr()
1144 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1154 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1167 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1186 mcast_cnt = vdev->vpaths[0].mcast_addr_cnt; in vxge_set_multicast()
1187 list_head = &vdev->vpaths[0].mac_addr_list; in vxge_set_multicast()
1189 (vdev->vpaths[0].mac_addr_cnt - mcast_cnt)) > in vxge_set_multicast()
1190 vdev->vpaths[0].max_mac_addr_cnt) in vxge_set_multicast()
1234 mcast_cnt = vdev->vpaths[0].mcast_addr_cnt; in vxge_set_multicast()
1256 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1310 struct vxge_vpath *vpath = &vdev->vpaths[vpath_idx]; in vxge_set_mac_addr()
1356 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_vpath_intr_enable()
1389 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_vpath_intr_disable()
1508 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_reset_vpath()
1585 hw_ring = vdev->vpaths[i].ring.handle; in vxge_config_ci_for_tti_rti()
1592 struct __vxge_hw_fifo *hw_fifo = vdev->vpaths[i].fifo.handle; in vxge_config_ci_for_tti_rti()
1722 if (vdev->vpaths[i].handle) { in do_vxge_reset()
1724 vdev->vpaths[i].handle) in do_vxge_reset()
1745 vxge_restore_vpath_mac_addr(&vdev->vpaths[vp_id]); in do_vxge_reset()
1746 vxge_restore_vpath_vid_table(&vdev->vpaths[vp_id]); in do_vxge_reset()
1762 vxge_hw_vpath_enable(vdev->vpaths[i].handle); in do_vxge_reset()
1764 vxge_hw_vpath_rx_doorbell_init(vdev->vpaths[i].handle); in do_vxge_reset()
1848 ring = &vdev->vpaths[i].ring; in vxge_poll_inta()
1933 "for vpath:%d", vdev->vpaths[0].device_id); in vxge_rth_configure()
1954 vdev->vpaths[index].handle, in vxge_rth_configure()
1961 vdev->vpaths[index].device_id); in vxge_rth_configure()
1977 vpath = &vdev->vpaths[i]; in vxge_reset_all_vpaths()
2008 vpath = &vdev->vpaths[i]; in vxge_close_vpaths()
2029 vpath = &vdev->vpaths[i]; in vxge_open_vpaths()
2270 vxge_hw_vpath_msix_mask(vdev->vpaths[i].handle, msix_id); in vxge_alarm_msix_handle()
2271 vxge_hw_vpath_msix_clear(vdev->vpaths[i].handle, msix_id); in vxge_alarm_msix_handle()
2273 status = vxge_hw_vpath_alarm_process(vdev->vpaths[i].handle, in vxge_alarm_msix_handle()
2276 vxge_hw_vpath_msix_unmask(vdev->vpaths[i].handle, in vxge_alarm_msix_handle()
2391 struct vxge_vpath *vpath = &vdev->vpaths[i]; in vxge_enable_msix()
2480 &vdev->vpaths[vp_idx].fifo); in vxge_add_isr()
2482 &vdev->vpaths[vp_idx].fifo; in vxge_add_isr()
2495 &vdev->vpaths[vp_idx].ring); in vxge_add_isr()
2497 &vdev->vpaths[vp_idx].ring; in vxge_add_isr()
2517 msix_idx += vdev->vpaths[vp_idx].device_id * in vxge_add_isr()
2520 vdev->vpaths[vp_idx].handle, in vxge_add_isr()
2541 &vdev->vpaths[0]); in vxge_add_isr()
2554 msix_idx = (vdev->vpaths[0].handle->vpath->vp_id * in vxge_add_isr()
2556 vxge_hw_vpath_msix_unmask(vdev->vpaths[vp_idx].handle, in vxge_add_isr()
2559 vdev->vxge_entries[intr_cnt].arg = &vdev->vpaths[0]; in vxge_add_isr()
2569 vxge_hw_vpath_tti_ci_set(vdev->vpaths[0].fifo.handle); in vxge_add_isr()
2617 ring = &vdev->vpaths[i].ring; in vxge_poll_vp_lockup()
2632 vpath = &vdev->vpaths[i]; in vxge_poll_vp_lockup()
2740 vpath = &vdev->vpaths[i]; in vxge_open()
2745 vpath = &vdev->vpaths[i]; in vxge_open()
2768 vpath = &vdev->vpaths[i]; in vxge_open()
2790 vpath = &vdev->vpaths[i]; in vxge_open()
2828 vpath = &vdev->vpaths[i]; in vxge_open()
2871 vpath = &vdev->vpaths[i]; in vxge_open()
2893 napi_disable(&vdev->vpaths[i].ring.napi); in vxge_open()
2926 netif_napi_del(&vdev->vpaths[i].ring.napi); in vxge_napi_del_all()
2953 vpath_vector = vxge_mBIT(vdev->vpaths[0].device_id); in do_vxge_close()
3005 napi_disable(&vdev->vpaths[i].ring.napi); in do_vxge_close()
3109 struct vxge_ring_stats *rxstats = &vdev->vpaths[k].ring.stats; in vxge_get_stats64()
3110 struct vxge_fifo_stats *txstats = &vdev->vpaths[k].fifo.stats; in vxge_get_stats64()
3221 vdev->vpaths[i].ring.rx_hwts = vdev->rx_hwts; in vxge_hwtstamp_set()
3308 vpath = &vdev->vpaths[vp_id]; in vxge_vlan_rx_add_vid()
3336 vpath = &vdev->vpaths[vp_id]; in vxge_vlan_rx_kill_vid()
3427 vdev->vpaths = kcalloc(no_of_vpath, sizeof(struct vxge_vpath), in vxge_device_register()
3429 if (!vdev->vpaths) { in vxge_device_register()
3494 kfree(vdev->vpaths); in vxge_device_register()
3525 kfree(vdev->vpaths); in vxge_device_unregister()
3560 vpath = &vdev->vpaths[vpath_idx]; in vxge_callback_crit_err()
4605 vdev->vpaths[j].is_configured = 1; in vxge_probe()
4606 vdev->vpaths[j].device_id = i; in vxge_probe()
4607 vdev->vpaths[j].ring.driver_id = j; in vxge_probe()
4608 vdev->vpaths[j].vdev = vdev; in vxge_probe()
4609 vdev->vpaths[j].max_mac_addr_cnt = max_mac_vpath; in vxge_probe()
4610 memcpy((u8 *)vdev->vpaths[j].macaddr, in vxge_probe()
4615 INIT_LIST_HEAD(&vdev->vpaths[j].mac_addr_list); in vxge_probe()
4617 vdev->vpaths[j].mac_addr_cnt = 0; in vxge_probe()
4618 vdev->vpaths[j].mcast_addr_cnt = 0; in vxge_probe()
4630 macaddr = (u8 *)vdev->vpaths[0].macaddr; in vxge_probe()
4681 memcpy(vdev->ndev->dev_addr, (u8 *)vdev->vpaths[0].macaddr, ETH_ALEN); in vxge_probe()
4695 list_add(&entry->item, &vdev->vpaths[i].mac_addr_list); in vxge_probe()
4696 vdev->vpaths[i].mac_addr_cnt = 1; in vxge_probe()
4733 vxge_free_mac_add_list(&vdev->vpaths[i]); in vxge_probe()
4776 vxge_free_mac_add_list(&vdev->vpaths[i]); in vxge_remove()