Lines Matching refs:ndev_ctx

63 	struct net_device_context *ndev_ctx = netdev_priv(net);  in netvsc_change_rx_flags()  local
64 struct net_device *vf_netdev = rtnl_dereference(ndev_ctx->vf_netdev); in netvsc_change_rx_flags()
83 struct net_device_context *ndev_ctx = netdev_priv(net); in netvsc_set_rx_mode() local
88 vf_netdev = rcu_dereference(ndev_ctx->vf_netdev); in netvsc_set_rx_mode()
94 nvdev = rcu_dereference(ndev_ctx->nvdev); in netvsc_set_rx_mode()
111 struct net_device_context *ndev_ctx = netdev_priv(net); in netvsc_open() local
112 struct net_device *vf_netdev = rtnl_dereference(ndev_ctx->vf_netdev); in netvsc_open()
113 struct netvsc_device *nvdev = rtnl_dereference(ndev_ctx->nvdev); in netvsc_open()
448 struct net_device_context *ndev_ctx = netdev_priv(net); in netvsc_vf_xmit() local
458 = this_cpu_ptr(ndev_ctx->vf_stats); in netvsc_vf_xmit()
465 this_cpu_inc(ndev_ctx->vf_stats->tx_dropped); in netvsc_vf_xmit()
697 struct net_device_context *ndev_ctx = netdev_priv(net); in netvsc_linkstatus_callback() local
731 ndev_ctx->speed = speed; in netvsc_linkstatus_callback()
749 spin_lock_irqsave(&ndev_ctx->lock, flags); in netvsc_linkstatus_callback()
750 list_add_tail(&event->list, &ndev_ctx->reconfig_events); in netvsc_linkstatus_callback()
751 spin_unlock_irqrestore(&ndev_ctx->lock, flags); in netvsc_linkstatus_callback()
753 schedule_delayed_work(&ndev_ctx->dwork, 0); in netvsc_linkstatus_callback()
1010 struct net_device_context *ndev_ctx = netdev_priv(ndev); in netvsc_detach() local
1011 struct hv_device *hdev = ndev_ctx->device_ctx; in netvsc_detach()
1049 struct net_device_context *ndev_ctx = netdev_priv(ndev); in netvsc_attach() local
1050 struct hv_device *hdev = ndev_ctx->device_ctx; in netvsc_attach()
1255 struct net_device_context *ndev_ctx = netdev_priv(net); in netvsc_get_vf_stats() local
1262 = per_cpu_ptr(ndev_ctx->vf_stats, i); in netvsc_get_vf_stats()
1285 struct net_device_context *ndev_ctx = netdev_priv(net); in netvsc_get_pcpu_stats() local
1286 struct netvsc_device *nvdev = rcu_dereference_rtnl(ndev_ctx->nvdev); in netvsc_get_pcpu_stats()
1292 per_cpu_ptr(ndev_ctx->vf_stats, i); in netvsc_get_pcpu_stats()
1344 struct net_device_context *ndev_ctx = netdev_priv(net); in netvsc_get_stats64() local
1351 nvdev = rcu_dereference(ndev_ctx->nvdev); in netvsc_get_stats64()
1982 struct net_device_context *ndev_ctx = netdev_priv(ndev); in netvsc_get_msglevel() local
1984 return ndev_ctx->msg_enable; in netvsc_get_msglevel()
1989 struct net_device_context *ndev_ctx = netdev_priv(ndev); in netvsc_set_msglevel() local
1991 ndev_ctx->msg_enable = val; in netvsc_set_msglevel()
2043 struct net_device_context *ndev_ctx = in netvsc_link_change() local
2045 struct hv_device *device_obj = ndev_ctx->device_ctx; in netvsc_link_change()
2055 schedule_delayed_work(&ndev_ctx->dwork, LINKCHANGE_INT); in netvsc_link_change()
2059 net_device = rtnl_dereference(ndev_ctx->nvdev); in netvsc_link_change()
2065 next_reconfig = ndev_ctx->last_reconfig + LINKCHANGE_INT; in netvsc_link_change()
2073 schedule_delayed_work(&ndev_ctx->dwork, delay); in netvsc_link_change()
2076 ndev_ctx->last_reconfig = jiffies; in netvsc_link_change()
2078 spin_lock_irqsave(&ndev_ctx->lock, flags); in netvsc_link_change()
2079 if (!list_empty(&ndev_ctx->reconfig_events)) { in netvsc_link_change()
2080 event = list_first_entry(&ndev_ctx->reconfig_events, in netvsc_link_change()
2083 reschedule = !list_empty(&ndev_ctx->reconfig_events); in netvsc_link_change()
2085 spin_unlock_irqrestore(&ndev_ctx->lock, flags); in netvsc_link_change()
2119 spin_lock_irqsave(&ndev_ctx->lock, flags); in netvsc_link_change()
2120 list_add(&event->list, &ndev_ctx->reconfig_events); in netvsc_link_change()
2121 spin_unlock_irqrestore(&ndev_ctx->lock, flags); in netvsc_link_change()
2133 schedule_delayed_work(&ndev_ctx->dwork, LINKCHANGE_INT); in netvsc_link_change()
2165 struct net_device_context *ndev_ctx = netdev_priv(ndev); in netvsc_vf_handle_frame() local
2167 = this_cpu_ptr(ndev_ctx->vf_stats); in netvsc_vf_handle_frame()
2188 struct net_device_context *ndev_ctx = netdev_priv(ndev); in netvsc_vf_join() local
2212 schedule_delayed_work(&ndev_ctx->vf_takeover, VF_TAKEOVER_INT); in netvsc_vf_join()
2258 struct net_device_context *ndev_ctx in netvsc_vf_setup() local
2260 struct net_device *ndev = hv_get_drvdata(ndev_ctx->device_ctx); in netvsc_vf_setup()
2264 schedule_delayed_work(&ndev_ctx->vf_takeover, 0); in netvsc_vf_setup()
2268 vf_netdev = rtnl_dereference(ndev_ctx->vf_netdev); in netvsc_vf_setup()
2281 struct net_device_context *ndev_ctx; in get_netvsc_byslot() local
2301 list_for_each_entry(ndev_ctx, &netvsc_dev_list, list) { in get_netvsc_byslot()
2302 if (!ndev_ctx->vf_alloc) in get_netvsc_byslot()
2305 if (ndev_ctx->vf_serial != serial) in get_netvsc_byslot()
2308 ndev = hv_get_drvdata(ndev_ctx->device_ctx); in get_netvsc_byslot()
2321 list_for_each_entry(ndev_ctx, &netvsc_dev_list, list) { in get_netvsc_byslot()
2322 ndev = hv_get_drvdata(ndev_ctx->device_ctx); in get_netvsc_byslot()
2604 struct net_device_context *ndev_ctx; in netvsc_remove() local
2614 ndev_ctx = netdev_priv(net); in netvsc_remove()
2616 cancel_delayed_work_sync(&ndev_ctx->dwork); in netvsc_remove()
2619 nvdev = rtnl_dereference(ndev_ctx->nvdev); in netvsc_remove()
2629 vf_netdev = rtnl_dereference(ndev_ctx->vf_netdev); in netvsc_remove()
2637 list_del(&ndev_ctx->list); in netvsc_remove()
2643 free_percpu(ndev_ctx->vf_stats); in netvsc_remove()
2649 struct net_device_context *ndev_ctx; in netvsc_suspend() local
2656 ndev_ctx = netdev_priv(net); in netvsc_suspend()
2657 cancel_delayed_work_sync(&ndev_ctx->dwork); in netvsc_suspend()
2661 nvdev = rtnl_dereference(ndev_ctx->nvdev); in netvsc_suspend()
2668 ndev_ctx->saved_netvsc_dev_info = netvsc_devinfo_get(nvdev); in netvsc_suspend()
2669 if (!ndev_ctx->saved_netvsc_dev_info) { in netvsc_suspend()