1  // SPDX-License-Identifier: (GPL-2.0 OR MIT)
2  /* Google virtual Ethernet (gve) driver
3   *
4   * Copyright (C) 2015-2021 Google, Inc.
5   */
6  
7  #include <linux/bpf.h>
8  #include <linux/cpumask.h>
9  #include <linux/etherdevice.h>
10  #include <linux/filter.h>
11  #include <linux/interrupt.h>
12  #include <linux/module.h>
13  #include <linux/pci.h>
14  #include <linux/sched.h>
15  #include <linux/timer.h>
16  #include <linux/workqueue.h>
17  #include <linux/utsname.h>
18  #include <linux/version.h>
19  #include <net/sch_generic.h>
20  #include <net/xdp_sock_drv.h>
21  #include "gve.h"
22  #include "gve_dqo.h"
23  #include "gve_adminq.h"
24  #include "gve_register.h"
25  
26  #define GVE_DEFAULT_RX_COPYBREAK	(256)
27  
28  #define DEFAULT_MSG_LEVEL	(NETIF_MSG_DRV | NETIF_MSG_LINK)
29  #define GVE_VERSION		"1.0.0"
30  #define GVE_VERSION_PREFIX	"GVE-"
31  
32  // Minimum amount of time between queue kicks in msec (10 seconds)
33  #define MIN_TX_TIMEOUT_GAP (1000 * 10)
34  
35  char gve_driver_name[] = "gve";
36  const char gve_version_str[] = GVE_VERSION;
37  static const char gve_version_prefix[] = GVE_VERSION_PREFIX;
38  
gve_verify_driver_compatibility(struct gve_priv * priv)39  static int gve_verify_driver_compatibility(struct gve_priv *priv)
40  {
41  	int err;
42  	struct gve_driver_info *driver_info;
43  	dma_addr_t driver_info_bus;
44  
45  	driver_info = dma_alloc_coherent(&priv->pdev->dev,
46  					 sizeof(struct gve_driver_info),
47  					 &driver_info_bus, GFP_KERNEL);
48  	if (!driver_info)
49  		return -ENOMEM;
50  
51  	*driver_info = (struct gve_driver_info) {
52  		.os_type = 1, /* Linux */
53  		.os_version_major = cpu_to_be32(LINUX_VERSION_MAJOR),
54  		.os_version_minor = cpu_to_be32(LINUX_VERSION_SUBLEVEL),
55  		.os_version_sub = cpu_to_be32(LINUX_VERSION_PATCHLEVEL),
56  		.driver_capability_flags = {
57  			cpu_to_be64(GVE_DRIVER_CAPABILITY_FLAGS1),
58  			cpu_to_be64(GVE_DRIVER_CAPABILITY_FLAGS2),
59  			cpu_to_be64(GVE_DRIVER_CAPABILITY_FLAGS3),
60  			cpu_to_be64(GVE_DRIVER_CAPABILITY_FLAGS4),
61  		},
62  	};
63  	strscpy(driver_info->os_version_str1, utsname()->release,
64  		sizeof(driver_info->os_version_str1));
65  	strscpy(driver_info->os_version_str2, utsname()->version,
66  		sizeof(driver_info->os_version_str2));
67  
68  	err = gve_adminq_verify_driver_compatibility(priv,
69  						     sizeof(struct gve_driver_info),
70  						     driver_info_bus);
71  
72  	/* It's ok if the device doesn't support this */
73  	if (err == -EOPNOTSUPP)
74  		err = 0;
75  
76  	dma_free_coherent(&priv->pdev->dev,
77  			  sizeof(struct gve_driver_info),
78  			  driver_info, driver_info_bus);
79  	return err;
80  }
81  
gve_start_xmit(struct sk_buff * skb,struct net_device * dev)82  static netdev_tx_t gve_start_xmit(struct sk_buff *skb, struct net_device *dev)
83  {
84  	struct gve_priv *priv = netdev_priv(dev);
85  
86  	if (gve_is_gqi(priv))
87  		return gve_tx(skb, dev);
88  	else
89  		return gve_tx_dqo(skb, dev);
90  }
91  
gve_get_stats(struct net_device * dev,struct rtnl_link_stats64 * s)92  static void gve_get_stats(struct net_device *dev, struct rtnl_link_stats64 *s)
93  {
94  	struct gve_priv *priv = netdev_priv(dev);
95  	unsigned int start;
96  	u64 packets, bytes;
97  	int num_tx_queues;
98  	int ring;
99  
100  	num_tx_queues = gve_num_tx_queues(priv);
101  	if (priv->rx) {
102  		for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) {
103  			do {
104  				start =
105  				  u64_stats_fetch_begin(&priv->rx[ring].statss);
106  				packets = priv->rx[ring].rpackets;
107  				bytes = priv->rx[ring].rbytes;
108  			} while (u64_stats_fetch_retry(&priv->rx[ring].statss,
109  						       start));
110  			s->rx_packets += packets;
111  			s->rx_bytes += bytes;
112  		}
113  	}
114  	if (priv->tx) {
115  		for (ring = 0; ring < num_tx_queues; ring++) {
116  			do {
117  				start =
118  				  u64_stats_fetch_begin(&priv->tx[ring].statss);
119  				packets = priv->tx[ring].pkt_done;
120  				bytes = priv->tx[ring].bytes_done;
121  			} while (u64_stats_fetch_retry(&priv->tx[ring].statss,
122  						       start));
123  			s->tx_packets += packets;
124  			s->tx_bytes += bytes;
125  		}
126  	}
127  }
128  
gve_alloc_counter_array(struct gve_priv * priv)129  static int gve_alloc_counter_array(struct gve_priv *priv)
130  {
131  	priv->counter_array =
132  		dma_alloc_coherent(&priv->pdev->dev,
133  				   priv->num_event_counters *
134  				   sizeof(*priv->counter_array),
135  				   &priv->counter_array_bus, GFP_KERNEL);
136  	if (!priv->counter_array)
137  		return -ENOMEM;
138  
139  	return 0;
140  }
141  
gve_free_counter_array(struct gve_priv * priv)142  static void gve_free_counter_array(struct gve_priv *priv)
143  {
144  	if (!priv->counter_array)
145  		return;
146  
147  	dma_free_coherent(&priv->pdev->dev,
148  			  priv->num_event_counters *
149  			  sizeof(*priv->counter_array),
150  			  priv->counter_array, priv->counter_array_bus);
151  	priv->counter_array = NULL;
152  }
153  
154  /* NIC requests to report stats */
gve_stats_report_task(struct work_struct * work)155  static void gve_stats_report_task(struct work_struct *work)
156  {
157  	struct gve_priv *priv = container_of(work, struct gve_priv,
158  					     stats_report_task);
159  	if (gve_get_do_report_stats(priv)) {
160  		gve_handle_report_stats(priv);
161  		gve_clear_do_report_stats(priv);
162  	}
163  }
164  
gve_stats_report_schedule(struct gve_priv * priv)165  static void gve_stats_report_schedule(struct gve_priv *priv)
166  {
167  	if (!gve_get_probe_in_progress(priv) &&
168  	    !gve_get_reset_in_progress(priv)) {
169  		gve_set_do_report_stats(priv);
170  		queue_work(priv->gve_wq, &priv->stats_report_task);
171  	}
172  }
173  
gve_stats_report_timer(struct timer_list * t)174  static void gve_stats_report_timer(struct timer_list *t)
175  {
176  	struct gve_priv *priv = from_timer(priv, t, stats_report_timer);
177  
178  	mod_timer(&priv->stats_report_timer,
179  		  round_jiffies(jiffies +
180  		  msecs_to_jiffies(priv->stats_report_timer_period)));
181  	gve_stats_report_schedule(priv);
182  }
183  
gve_alloc_stats_report(struct gve_priv * priv)184  static int gve_alloc_stats_report(struct gve_priv *priv)
185  {
186  	int tx_stats_num, rx_stats_num;
187  
188  	tx_stats_num = (GVE_TX_STATS_REPORT_NUM + NIC_TX_STATS_REPORT_NUM) *
189  		       gve_num_tx_queues(priv);
190  	rx_stats_num = (GVE_RX_STATS_REPORT_NUM + NIC_RX_STATS_REPORT_NUM) *
191  		       priv->rx_cfg.num_queues;
192  	priv->stats_report_len = struct_size(priv->stats_report, stats,
193  					     tx_stats_num + rx_stats_num);
194  	priv->stats_report =
195  		dma_alloc_coherent(&priv->pdev->dev, priv->stats_report_len,
196  				   &priv->stats_report_bus, GFP_KERNEL);
197  	if (!priv->stats_report)
198  		return -ENOMEM;
199  	/* Set up timer for the report-stats task */
200  	timer_setup(&priv->stats_report_timer, gve_stats_report_timer, 0);
201  	priv->stats_report_timer_period = GVE_STATS_REPORT_TIMER_PERIOD;
202  	return 0;
203  }
204  
gve_free_stats_report(struct gve_priv * priv)205  static void gve_free_stats_report(struct gve_priv *priv)
206  {
207  	if (!priv->stats_report)
208  		return;
209  
210  	del_timer_sync(&priv->stats_report_timer);
211  	dma_free_coherent(&priv->pdev->dev, priv->stats_report_len,
212  			  priv->stats_report, priv->stats_report_bus);
213  	priv->stats_report = NULL;
214  }
215  
gve_mgmnt_intr(int irq,void * arg)216  static irqreturn_t gve_mgmnt_intr(int irq, void *arg)
217  {
218  	struct gve_priv *priv = arg;
219  
220  	queue_work(priv->gve_wq, &priv->service_task);
221  	return IRQ_HANDLED;
222  }
223  
gve_intr(int irq,void * arg)224  static irqreturn_t gve_intr(int irq, void *arg)
225  {
226  	struct gve_notify_block *block = arg;
227  	struct gve_priv *priv = block->priv;
228  
229  	iowrite32be(GVE_IRQ_MASK, gve_irq_doorbell(priv, block));
230  	napi_schedule_irqoff(&block->napi);
231  	return IRQ_HANDLED;
232  }
233  
gve_intr_dqo(int irq,void * arg)234  static irqreturn_t gve_intr_dqo(int irq, void *arg)
235  {
236  	struct gve_notify_block *block = arg;
237  
238  	/* Interrupts are automatically masked */
239  	napi_schedule_irqoff(&block->napi);
240  	return IRQ_HANDLED;
241  }
242  
gve_napi_poll(struct napi_struct * napi,int budget)243  static int gve_napi_poll(struct napi_struct *napi, int budget)
244  {
245  	struct gve_notify_block *block;
246  	__be32 __iomem *irq_doorbell;
247  	bool reschedule = false;
248  	struct gve_priv *priv;
249  	int work_done = 0;
250  
251  	block = container_of(napi, struct gve_notify_block, napi);
252  	priv = block->priv;
253  
254  	if (block->tx) {
255  		if (block->tx->q_num < priv->tx_cfg.num_queues)
256  			reschedule |= gve_tx_poll(block, budget);
257  		else
258  			reschedule |= gve_xdp_poll(block, budget);
259  	}
260  
261  	if (block->rx) {
262  		work_done = gve_rx_poll(block, budget);
263  		reschedule |= work_done == budget;
264  	}
265  
266  	if (reschedule)
267  		return budget;
268  
269         /* Complete processing - don't unmask irq if busy polling is enabled */
270  	if (likely(napi_complete_done(napi, work_done))) {
271  		irq_doorbell = gve_irq_doorbell(priv, block);
272  		iowrite32be(GVE_IRQ_ACK | GVE_IRQ_EVENT, irq_doorbell);
273  
274  		/* Ensure IRQ ACK is visible before we check pending work.
275  		 * If queue had issued updates, it would be truly visible.
276  		 */
277  		mb();
278  
279  		if (block->tx)
280  			reschedule |= gve_tx_clean_pending(priv, block->tx);
281  		if (block->rx)
282  			reschedule |= gve_rx_work_pending(block->rx);
283  
284  		if (reschedule && napi_reschedule(napi))
285  			iowrite32be(GVE_IRQ_MASK, irq_doorbell);
286  	}
287  	return work_done;
288  }
289  
gve_napi_poll_dqo(struct napi_struct * napi,int budget)290  static int gve_napi_poll_dqo(struct napi_struct *napi, int budget)
291  {
292  	struct gve_notify_block *block =
293  		container_of(napi, struct gve_notify_block, napi);
294  	struct gve_priv *priv = block->priv;
295  	bool reschedule = false;
296  	int work_done = 0;
297  
298  	if (block->tx)
299  		reschedule |= gve_tx_poll_dqo(block, /*do_clean=*/true);
300  
301  	if (block->rx) {
302  		work_done = gve_rx_poll_dqo(block, budget);
303  		reschedule |= work_done == budget;
304  	}
305  
306  	if (reschedule)
307  		return budget;
308  
309  	if (likely(napi_complete_done(napi, work_done))) {
310  		/* Enable interrupts again.
311  		 *
312  		 * We don't need to repoll afterwards because HW supports the
313  		 * PCI MSI-X PBA feature.
314  		 *
315  		 * Another interrupt would be triggered if a new event came in
316  		 * since the last one.
317  		 */
318  		gve_write_irq_doorbell_dqo(priv, block,
319  					   GVE_ITR_NO_UPDATE_DQO | GVE_ITR_ENABLE_BIT_DQO);
320  	}
321  
322  	return work_done;
323  }
324  
gve_alloc_notify_blocks(struct gve_priv * priv)325  static int gve_alloc_notify_blocks(struct gve_priv *priv)
326  {
327  	int num_vecs_requested = priv->num_ntfy_blks + 1;
328  	unsigned int active_cpus;
329  	int vecs_enabled;
330  	int i, j;
331  	int err;
332  
333  	priv->msix_vectors = kvcalloc(num_vecs_requested,
334  				      sizeof(*priv->msix_vectors), GFP_KERNEL);
335  	if (!priv->msix_vectors)
336  		return -ENOMEM;
337  	for (i = 0; i < num_vecs_requested; i++)
338  		priv->msix_vectors[i].entry = i;
339  	vecs_enabled = pci_enable_msix_range(priv->pdev, priv->msix_vectors,
340  					     GVE_MIN_MSIX, num_vecs_requested);
341  	if (vecs_enabled < 0) {
342  		dev_err(&priv->pdev->dev, "Could not enable min msix %d/%d\n",
343  			GVE_MIN_MSIX, vecs_enabled);
344  		err = vecs_enabled;
345  		goto abort_with_msix_vectors;
346  	}
347  	if (vecs_enabled != num_vecs_requested) {
348  		int new_num_ntfy_blks = (vecs_enabled - 1) & ~0x1;
349  		int vecs_per_type = new_num_ntfy_blks / 2;
350  		int vecs_left = new_num_ntfy_blks % 2;
351  
352  		priv->num_ntfy_blks = new_num_ntfy_blks;
353  		priv->mgmt_msix_idx = priv->num_ntfy_blks;
354  		priv->tx_cfg.max_queues = min_t(int, priv->tx_cfg.max_queues,
355  						vecs_per_type);
356  		priv->rx_cfg.max_queues = min_t(int, priv->rx_cfg.max_queues,
357  						vecs_per_type + vecs_left);
358  		dev_err(&priv->pdev->dev,
359  			"Could not enable desired msix, only enabled %d, adjusting tx max queues to %d, and rx max queues to %d\n",
360  			vecs_enabled, priv->tx_cfg.max_queues,
361  			priv->rx_cfg.max_queues);
362  		if (priv->tx_cfg.num_queues > priv->tx_cfg.max_queues)
363  			priv->tx_cfg.num_queues = priv->tx_cfg.max_queues;
364  		if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues)
365  			priv->rx_cfg.num_queues = priv->rx_cfg.max_queues;
366  	}
367  	/* Half the notification blocks go to TX and half to RX */
368  	active_cpus = min_t(int, priv->num_ntfy_blks / 2, num_online_cpus());
369  
370  	/* Setup Management Vector  - the last vector */
371  	snprintf(priv->mgmt_msix_name, sizeof(priv->mgmt_msix_name), "gve-mgmnt@pci:%s",
372  		 pci_name(priv->pdev));
373  	err = request_irq(priv->msix_vectors[priv->mgmt_msix_idx].vector,
374  			  gve_mgmnt_intr, 0, priv->mgmt_msix_name, priv);
375  	if (err) {
376  		dev_err(&priv->pdev->dev, "Did not receive management vector.\n");
377  		goto abort_with_msix_enabled;
378  	}
379  	priv->irq_db_indices =
380  		dma_alloc_coherent(&priv->pdev->dev,
381  				   priv->num_ntfy_blks *
382  				   sizeof(*priv->irq_db_indices),
383  				   &priv->irq_db_indices_bus, GFP_KERNEL);
384  	if (!priv->irq_db_indices) {
385  		err = -ENOMEM;
386  		goto abort_with_mgmt_vector;
387  	}
388  
389  	priv->ntfy_blocks = kvzalloc(priv->num_ntfy_blks *
390  				     sizeof(*priv->ntfy_blocks), GFP_KERNEL);
391  	if (!priv->ntfy_blocks) {
392  		err = -ENOMEM;
393  		goto abort_with_irq_db_indices;
394  	}
395  
396  	/* Setup the other blocks - the first n-1 vectors */
397  	for (i = 0; i < priv->num_ntfy_blks; i++) {
398  		struct gve_notify_block *block = &priv->ntfy_blocks[i];
399  		int msix_idx = i;
400  
401  		snprintf(block->name, sizeof(block->name), "gve-ntfy-blk%d@pci:%s",
402  			 i, pci_name(priv->pdev));
403  		block->priv = priv;
404  		err = request_irq(priv->msix_vectors[msix_idx].vector,
405  				  gve_is_gqi(priv) ? gve_intr : gve_intr_dqo,
406  				  0, block->name, block);
407  		if (err) {
408  			dev_err(&priv->pdev->dev,
409  				"Failed to receive msix vector %d\n", i);
410  			goto abort_with_some_ntfy_blocks;
411  		}
412  		irq_set_affinity_hint(priv->msix_vectors[msix_idx].vector,
413  				      get_cpu_mask(i % active_cpus));
414  		block->irq_db_index = &priv->irq_db_indices[i].index;
415  	}
416  	return 0;
417  abort_with_some_ntfy_blocks:
418  	for (j = 0; j < i; j++) {
419  		struct gve_notify_block *block = &priv->ntfy_blocks[j];
420  		int msix_idx = j;
421  
422  		irq_set_affinity_hint(priv->msix_vectors[msix_idx].vector,
423  				      NULL);
424  		free_irq(priv->msix_vectors[msix_idx].vector, block);
425  	}
426  	kvfree(priv->ntfy_blocks);
427  	priv->ntfy_blocks = NULL;
428  abort_with_irq_db_indices:
429  	dma_free_coherent(&priv->pdev->dev, priv->num_ntfy_blks *
430  			  sizeof(*priv->irq_db_indices),
431  			  priv->irq_db_indices, priv->irq_db_indices_bus);
432  	priv->irq_db_indices = NULL;
433  abort_with_mgmt_vector:
434  	free_irq(priv->msix_vectors[priv->mgmt_msix_idx].vector, priv);
435  abort_with_msix_enabled:
436  	pci_disable_msix(priv->pdev);
437  abort_with_msix_vectors:
438  	kvfree(priv->msix_vectors);
439  	priv->msix_vectors = NULL;
440  	return err;
441  }
442  
gve_free_notify_blocks(struct gve_priv * priv)443  static void gve_free_notify_blocks(struct gve_priv *priv)
444  {
445  	int i;
446  
447  	if (!priv->msix_vectors)
448  		return;
449  
450  	/* Free the irqs */
451  	for (i = 0; i < priv->num_ntfy_blks; i++) {
452  		struct gve_notify_block *block = &priv->ntfy_blocks[i];
453  		int msix_idx = i;
454  
455  		irq_set_affinity_hint(priv->msix_vectors[msix_idx].vector,
456  				      NULL);
457  		free_irq(priv->msix_vectors[msix_idx].vector, block);
458  	}
459  	free_irq(priv->msix_vectors[priv->mgmt_msix_idx].vector, priv);
460  	kvfree(priv->ntfy_blocks);
461  	priv->ntfy_blocks = NULL;
462  	dma_free_coherent(&priv->pdev->dev, priv->num_ntfy_blks *
463  			  sizeof(*priv->irq_db_indices),
464  			  priv->irq_db_indices, priv->irq_db_indices_bus);
465  	priv->irq_db_indices = NULL;
466  	pci_disable_msix(priv->pdev);
467  	kvfree(priv->msix_vectors);
468  	priv->msix_vectors = NULL;
469  }
470  
gve_setup_device_resources(struct gve_priv * priv)471  static int gve_setup_device_resources(struct gve_priv *priv)
472  {
473  	int err;
474  
475  	err = gve_alloc_counter_array(priv);
476  	if (err)
477  		return err;
478  	err = gve_alloc_notify_blocks(priv);
479  	if (err)
480  		goto abort_with_counter;
481  	err = gve_alloc_stats_report(priv);
482  	if (err)
483  		goto abort_with_ntfy_blocks;
484  	err = gve_adminq_configure_device_resources(priv,
485  						    priv->counter_array_bus,
486  						    priv->num_event_counters,
487  						    priv->irq_db_indices_bus,
488  						    priv->num_ntfy_blks);
489  	if (unlikely(err)) {
490  		dev_err(&priv->pdev->dev,
491  			"could not setup device_resources: err=%d\n", err);
492  		err = -ENXIO;
493  		goto abort_with_stats_report;
494  	}
495  
496  	if (!gve_is_gqi(priv)) {
497  		priv->ptype_lut_dqo = kvzalloc(sizeof(*priv->ptype_lut_dqo),
498  					       GFP_KERNEL);
499  		if (!priv->ptype_lut_dqo) {
500  			err = -ENOMEM;
501  			goto abort_with_stats_report;
502  		}
503  		err = gve_adminq_get_ptype_map_dqo(priv, priv->ptype_lut_dqo);
504  		if (err) {
505  			dev_err(&priv->pdev->dev,
506  				"Failed to get ptype map: err=%d\n", err);
507  			goto abort_with_ptype_lut;
508  		}
509  	}
510  
511  	err = gve_adminq_report_stats(priv, priv->stats_report_len,
512  				      priv->stats_report_bus,
513  				      GVE_STATS_REPORT_TIMER_PERIOD);
514  	if (err)
515  		dev_err(&priv->pdev->dev,
516  			"Failed to report stats: err=%d\n", err);
517  	gve_set_device_resources_ok(priv);
518  	return 0;
519  
520  abort_with_ptype_lut:
521  	kvfree(priv->ptype_lut_dqo);
522  	priv->ptype_lut_dqo = NULL;
523  abort_with_stats_report:
524  	gve_free_stats_report(priv);
525  abort_with_ntfy_blocks:
526  	gve_free_notify_blocks(priv);
527  abort_with_counter:
528  	gve_free_counter_array(priv);
529  
530  	return err;
531  }
532  
533  static void gve_trigger_reset(struct gve_priv *priv);
534  
gve_teardown_device_resources(struct gve_priv * priv)535  static void gve_teardown_device_resources(struct gve_priv *priv)
536  {
537  	int err;
538  
539  	/* Tell device its resources are being freed */
540  	if (gve_get_device_resources_ok(priv)) {
541  		/* detach the stats report */
542  		err = gve_adminq_report_stats(priv, 0, 0x0, GVE_STATS_REPORT_TIMER_PERIOD);
543  		if (err) {
544  			dev_err(&priv->pdev->dev,
545  				"Failed to detach stats report: err=%d\n", err);
546  			gve_trigger_reset(priv);
547  		}
548  		err = gve_adminq_deconfigure_device_resources(priv);
549  		if (err) {
550  			dev_err(&priv->pdev->dev,
551  				"Could not deconfigure device resources: err=%d\n",
552  				err);
553  			gve_trigger_reset(priv);
554  		}
555  	}
556  
557  	kvfree(priv->ptype_lut_dqo);
558  	priv->ptype_lut_dqo = NULL;
559  
560  	gve_free_counter_array(priv);
561  	gve_free_notify_blocks(priv);
562  	gve_free_stats_report(priv);
563  	gve_clear_device_resources_ok(priv);
564  }
565  
gve_add_napi(struct gve_priv * priv,int ntfy_idx,int (* gve_poll)(struct napi_struct *,int))566  static void gve_add_napi(struct gve_priv *priv, int ntfy_idx,
567  			 int (*gve_poll)(struct napi_struct *, int))
568  {
569  	struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
570  
571  	netif_napi_add(priv->dev, &block->napi, gve_poll);
572  }
573  
gve_remove_napi(struct gve_priv * priv,int ntfy_idx)574  static void gve_remove_napi(struct gve_priv *priv, int ntfy_idx)
575  {
576  	struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
577  
578  	netif_napi_del(&block->napi);
579  }
580  
gve_register_xdp_qpls(struct gve_priv * priv)581  static int gve_register_xdp_qpls(struct gve_priv *priv)
582  {
583  	int start_id;
584  	int err;
585  	int i;
586  
587  	start_id = gve_tx_qpl_id(priv, gve_xdp_tx_start_queue_id(priv));
588  	for (i = start_id; i < start_id + gve_num_xdp_qpls(priv); i++) {
589  		err = gve_adminq_register_page_list(priv, &priv->qpls[i]);
590  		if (err) {
591  			netif_err(priv, drv, priv->dev,
592  				  "failed to register queue page list %d\n",
593  				  priv->qpls[i].id);
594  			/* This failure will trigger a reset - no need to clean
595  			 * up
596  			 */
597  			return err;
598  		}
599  	}
600  	return 0;
601  }
602  
gve_register_qpls(struct gve_priv * priv)603  static int gve_register_qpls(struct gve_priv *priv)
604  {
605  	int start_id;
606  	int err;
607  	int i;
608  
609  	start_id = gve_tx_start_qpl_id(priv);
610  	for (i = start_id; i < start_id + gve_num_tx_qpls(priv); i++) {
611  		err = gve_adminq_register_page_list(priv, &priv->qpls[i]);
612  		if (err) {
613  			netif_err(priv, drv, priv->dev,
614  				  "failed to register queue page list %d\n",
615  				  priv->qpls[i].id);
616  			/* This failure will trigger a reset - no need to clean
617  			 * up
618  			 */
619  			return err;
620  		}
621  	}
622  
623  	start_id = gve_rx_start_qpl_id(priv);
624  	for (i = start_id; i < start_id + gve_num_rx_qpls(priv); i++) {
625  		err = gve_adminq_register_page_list(priv, &priv->qpls[i]);
626  		if (err) {
627  			netif_err(priv, drv, priv->dev,
628  				  "failed to register queue page list %d\n",
629  				  priv->qpls[i].id);
630  			/* This failure will trigger a reset - no need to clean
631  			 * up
632  			 */
633  			return err;
634  		}
635  	}
636  	return 0;
637  }
638  
gve_unregister_xdp_qpls(struct gve_priv * priv)639  static int gve_unregister_xdp_qpls(struct gve_priv *priv)
640  {
641  	int start_id;
642  	int err;
643  	int i;
644  
645  	start_id = gve_tx_qpl_id(priv, gve_xdp_tx_start_queue_id(priv));
646  	for (i = start_id; i < start_id + gve_num_xdp_qpls(priv); i++) {
647  		err = gve_adminq_unregister_page_list(priv, priv->qpls[i].id);
648  		/* This failure will trigger a reset - no need to clean up */
649  		if (err) {
650  			netif_err(priv, drv, priv->dev,
651  				  "Failed to unregister queue page list %d\n",
652  				  priv->qpls[i].id);
653  			return err;
654  		}
655  	}
656  	return 0;
657  }
658  
gve_unregister_qpls(struct gve_priv * priv)659  static int gve_unregister_qpls(struct gve_priv *priv)
660  {
661  	int start_id;
662  	int err;
663  	int i;
664  
665  	start_id = gve_tx_start_qpl_id(priv);
666  	for (i = start_id; i < start_id + gve_num_tx_qpls(priv); i++) {
667  		err = gve_adminq_unregister_page_list(priv, priv->qpls[i].id);
668  		/* This failure will trigger a reset - no need to clean up */
669  		if (err) {
670  			netif_err(priv, drv, priv->dev,
671  				  "Failed to unregister queue page list %d\n",
672  				  priv->qpls[i].id);
673  			return err;
674  		}
675  	}
676  
677  	start_id = gve_rx_start_qpl_id(priv);
678  	for (i = start_id; i < start_id + gve_num_rx_qpls(priv); i++) {
679  		err = gve_adminq_unregister_page_list(priv, priv->qpls[i].id);
680  		/* This failure will trigger a reset - no need to clean up */
681  		if (err) {
682  			netif_err(priv, drv, priv->dev,
683  				  "Failed to unregister queue page list %d\n",
684  				  priv->qpls[i].id);
685  			return err;
686  		}
687  	}
688  	return 0;
689  }
690  
gve_create_xdp_rings(struct gve_priv * priv)691  static int gve_create_xdp_rings(struct gve_priv *priv)
692  {
693  	int err;
694  
695  	err = gve_adminq_create_tx_queues(priv,
696  					  gve_xdp_tx_start_queue_id(priv),
697  					  priv->num_xdp_queues);
698  	if (err) {
699  		netif_err(priv, drv, priv->dev, "failed to create %d XDP tx queues\n",
700  			  priv->num_xdp_queues);
701  		/* This failure will trigger a reset - no need to clean
702  		 * up
703  		 */
704  		return err;
705  	}
706  	netif_dbg(priv, drv, priv->dev, "created %d XDP tx queues\n",
707  		  priv->num_xdp_queues);
708  
709  	return 0;
710  }
711  
gve_create_rings(struct gve_priv * priv)712  static int gve_create_rings(struct gve_priv *priv)
713  {
714  	int num_tx_queues = gve_num_tx_queues(priv);
715  	int err;
716  	int i;
717  
718  	err = gve_adminq_create_tx_queues(priv, 0, num_tx_queues);
719  	if (err) {
720  		netif_err(priv, drv, priv->dev, "failed to create %d tx queues\n",
721  			  num_tx_queues);
722  		/* This failure will trigger a reset - no need to clean
723  		 * up
724  		 */
725  		return err;
726  	}
727  	netif_dbg(priv, drv, priv->dev, "created %d tx queues\n",
728  		  num_tx_queues);
729  
730  	err = gve_adminq_create_rx_queues(priv, priv->rx_cfg.num_queues);
731  	if (err) {
732  		netif_err(priv, drv, priv->dev, "failed to create %d rx queues\n",
733  			  priv->rx_cfg.num_queues);
734  		/* This failure will trigger a reset - no need to clean
735  		 * up
736  		 */
737  		return err;
738  	}
739  	netif_dbg(priv, drv, priv->dev, "created %d rx queues\n",
740  		  priv->rx_cfg.num_queues);
741  
742  	if (gve_is_gqi(priv)) {
743  		/* Rx data ring has been prefilled with packet buffers at queue
744  		 * allocation time.
745  		 *
746  		 * Write the doorbell to provide descriptor slots and packet
747  		 * buffers to the NIC.
748  		 */
749  		for (i = 0; i < priv->rx_cfg.num_queues; i++)
750  			gve_rx_write_doorbell(priv, &priv->rx[i]);
751  	} else {
752  		for (i = 0; i < priv->rx_cfg.num_queues; i++) {
753  			/* Post buffers and ring doorbell. */
754  			gve_rx_post_buffers_dqo(&priv->rx[i]);
755  		}
756  	}
757  
758  	return 0;
759  }
760  
add_napi_init_xdp_sync_stats(struct gve_priv * priv,int (* napi_poll)(struct napi_struct * napi,int budget))761  static void add_napi_init_xdp_sync_stats(struct gve_priv *priv,
762  					 int (*napi_poll)(struct napi_struct *napi,
763  							  int budget))
764  {
765  	int start_id = gve_xdp_tx_start_queue_id(priv);
766  	int i;
767  
768  	/* Add xdp tx napi & init sync stats*/
769  	for (i = start_id; i < start_id + priv->num_xdp_queues; i++) {
770  		int ntfy_idx = gve_tx_idx_to_ntfy(priv, i);
771  
772  		u64_stats_init(&priv->tx[i].statss);
773  		priv->tx[i].ntfy_id = ntfy_idx;
774  		gve_add_napi(priv, ntfy_idx, napi_poll);
775  	}
776  }
777  
add_napi_init_sync_stats(struct gve_priv * priv,int (* napi_poll)(struct napi_struct * napi,int budget))778  static void add_napi_init_sync_stats(struct gve_priv *priv,
779  				     int (*napi_poll)(struct napi_struct *napi,
780  						      int budget))
781  {
782  	int i;
783  
784  	/* Add tx napi & init sync stats*/
785  	for (i = 0; i < gve_num_tx_queues(priv); i++) {
786  		int ntfy_idx = gve_tx_idx_to_ntfy(priv, i);
787  
788  		u64_stats_init(&priv->tx[i].statss);
789  		priv->tx[i].ntfy_id = ntfy_idx;
790  		gve_add_napi(priv, ntfy_idx, napi_poll);
791  	}
792  	/* Add rx napi  & init sync stats*/
793  	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
794  		int ntfy_idx = gve_rx_idx_to_ntfy(priv, i);
795  
796  		u64_stats_init(&priv->rx[i].statss);
797  		priv->rx[i].ntfy_id = ntfy_idx;
798  		gve_add_napi(priv, ntfy_idx, napi_poll);
799  	}
800  }
801  
gve_tx_free_rings(struct gve_priv * priv,int start_id,int num_rings)802  static void gve_tx_free_rings(struct gve_priv *priv, int start_id, int num_rings)
803  {
804  	if (gve_is_gqi(priv)) {
805  		gve_tx_free_rings_gqi(priv, start_id, num_rings);
806  	} else {
807  		gve_tx_free_rings_dqo(priv);
808  	}
809  }
810  
gve_alloc_xdp_rings(struct gve_priv * priv)811  static int gve_alloc_xdp_rings(struct gve_priv *priv)
812  {
813  	int start_id;
814  	int err = 0;
815  
816  	if (!priv->num_xdp_queues)
817  		return 0;
818  
819  	start_id = gve_xdp_tx_start_queue_id(priv);
820  	err = gve_tx_alloc_rings(priv, start_id, priv->num_xdp_queues);
821  	if (err)
822  		return err;
823  	add_napi_init_xdp_sync_stats(priv, gve_napi_poll);
824  
825  	return 0;
826  }
827  
gve_alloc_rings(struct gve_priv * priv)828  static int gve_alloc_rings(struct gve_priv *priv)
829  {
830  	int err;
831  
832  	/* Setup tx rings */
833  	priv->tx = kvcalloc(priv->tx_cfg.max_queues, sizeof(*priv->tx),
834  			    GFP_KERNEL);
835  	if (!priv->tx)
836  		return -ENOMEM;
837  
838  	if (gve_is_gqi(priv))
839  		err = gve_tx_alloc_rings(priv, 0, gve_num_tx_queues(priv));
840  	else
841  		err = gve_tx_alloc_rings_dqo(priv);
842  	if (err)
843  		goto free_tx;
844  
845  	/* Setup rx rings */
846  	priv->rx = kvcalloc(priv->rx_cfg.max_queues, sizeof(*priv->rx),
847  			    GFP_KERNEL);
848  	if (!priv->rx) {
849  		err = -ENOMEM;
850  		goto free_tx_queue;
851  	}
852  
853  	if (gve_is_gqi(priv))
854  		err = gve_rx_alloc_rings(priv);
855  	else
856  		err = gve_rx_alloc_rings_dqo(priv);
857  	if (err)
858  		goto free_rx;
859  
860  	if (gve_is_gqi(priv))
861  		add_napi_init_sync_stats(priv, gve_napi_poll);
862  	else
863  		add_napi_init_sync_stats(priv, gve_napi_poll_dqo);
864  
865  	return 0;
866  
867  free_rx:
868  	kvfree(priv->rx);
869  	priv->rx = NULL;
870  free_tx_queue:
871  	gve_tx_free_rings(priv, 0, gve_num_tx_queues(priv));
872  free_tx:
873  	kvfree(priv->tx);
874  	priv->tx = NULL;
875  	return err;
876  }
877  
gve_destroy_xdp_rings(struct gve_priv * priv)878  static int gve_destroy_xdp_rings(struct gve_priv *priv)
879  {
880  	int start_id;
881  	int err;
882  
883  	start_id = gve_xdp_tx_start_queue_id(priv);
884  	err = gve_adminq_destroy_tx_queues(priv,
885  					   start_id,
886  					   priv->num_xdp_queues);
887  	if (err) {
888  		netif_err(priv, drv, priv->dev,
889  			  "failed to destroy XDP queues\n");
890  		/* This failure will trigger a reset - no need to clean up */
891  		return err;
892  	}
893  	netif_dbg(priv, drv, priv->dev, "destroyed XDP queues\n");
894  
895  	return 0;
896  }
897  
gve_destroy_rings(struct gve_priv * priv)898  static int gve_destroy_rings(struct gve_priv *priv)
899  {
900  	int num_tx_queues = gve_num_tx_queues(priv);
901  	int err;
902  
903  	err = gve_adminq_destroy_tx_queues(priv, 0, num_tx_queues);
904  	if (err) {
905  		netif_err(priv, drv, priv->dev,
906  			  "failed to destroy tx queues\n");
907  		/* This failure will trigger a reset - no need to clean up */
908  		return err;
909  	}
910  	netif_dbg(priv, drv, priv->dev, "destroyed tx queues\n");
911  	err = gve_adminq_destroy_rx_queues(priv, priv->rx_cfg.num_queues);
912  	if (err) {
913  		netif_err(priv, drv, priv->dev,
914  			  "failed to destroy rx queues\n");
915  		/* This failure will trigger a reset - no need to clean up */
916  		return err;
917  	}
918  	netif_dbg(priv, drv, priv->dev, "destroyed rx queues\n");
919  	return 0;
920  }
921  
gve_rx_free_rings(struct gve_priv * priv)922  static void gve_rx_free_rings(struct gve_priv *priv)
923  {
924  	if (gve_is_gqi(priv))
925  		gve_rx_free_rings_gqi(priv);
926  	else
927  		gve_rx_free_rings_dqo(priv);
928  }
929  
gve_free_xdp_rings(struct gve_priv * priv)930  static void gve_free_xdp_rings(struct gve_priv *priv)
931  {
932  	int ntfy_idx, start_id;
933  	int i;
934  
935  	start_id = gve_xdp_tx_start_queue_id(priv);
936  	if (priv->tx) {
937  		for (i = start_id; i <  start_id + priv->num_xdp_queues; i++) {
938  			ntfy_idx = gve_tx_idx_to_ntfy(priv, i);
939  			gve_remove_napi(priv, ntfy_idx);
940  		}
941  		gve_tx_free_rings(priv, start_id, priv->num_xdp_queues);
942  	}
943  }
944  
gve_free_rings(struct gve_priv * priv)945  static void gve_free_rings(struct gve_priv *priv)
946  {
947  	int num_tx_queues = gve_num_tx_queues(priv);
948  	int ntfy_idx;
949  	int i;
950  
951  	if (priv->tx) {
952  		for (i = 0; i < num_tx_queues; i++) {
953  			ntfy_idx = gve_tx_idx_to_ntfy(priv, i);
954  			gve_remove_napi(priv, ntfy_idx);
955  		}
956  		gve_tx_free_rings(priv, 0, num_tx_queues);
957  		kvfree(priv->tx);
958  		priv->tx = NULL;
959  	}
960  	if (priv->rx) {
961  		for (i = 0; i < priv->rx_cfg.num_queues; i++) {
962  			ntfy_idx = gve_rx_idx_to_ntfy(priv, i);
963  			gve_remove_napi(priv, ntfy_idx);
964  		}
965  		gve_rx_free_rings(priv);
966  		kvfree(priv->rx);
967  		priv->rx = NULL;
968  	}
969  }
970  
gve_alloc_page(struct gve_priv * priv,struct device * dev,struct page ** page,dma_addr_t * dma,enum dma_data_direction dir,gfp_t gfp_flags)971  int gve_alloc_page(struct gve_priv *priv, struct device *dev,
972  		   struct page **page, dma_addr_t *dma,
973  		   enum dma_data_direction dir, gfp_t gfp_flags)
974  {
975  	*page = alloc_page(gfp_flags);
976  	if (!*page) {
977  		priv->page_alloc_fail++;
978  		return -ENOMEM;
979  	}
980  	*dma = dma_map_page(dev, *page, 0, PAGE_SIZE, dir);
981  	if (dma_mapping_error(dev, *dma)) {
982  		priv->dma_mapping_error++;
983  		put_page(*page);
984  		return -ENOMEM;
985  	}
986  	return 0;
987  }
988  
gve_alloc_queue_page_list(struct gve_priv * priv,u32 id,int pages)989  static int gve_alloc_queue_page_list(struct gve_priv *priv, u32 id,
990  				     int pages)
991  {
992  	struct gve_queue_page_list *qpl = &priv->qpls[id];
993  	int err;
994  	int i;
995  
996  	if (pages + priv->num_registered_pages > priv->max_registered_pages) {
997  		netif_err(priv, drv, priv->dev,
998  			  "Reached max number of registered pages %llu > %llu\n",
999  			  pages + priv->num_registered_pages,
1000  			  priv->max_registered_pages);
1001  		return -EINVAL;
1002  	}
1003  
1004  	qpl->id = id;
1005  	qpl->num_entries = 0;
1006  	qpl->pages = kvcalloc(pages, sizeof(*qpl->pages), GFP_KERNEL);
1007  	/* caller handles clean up */
1008  	if (!qpl->pages)
1009  		return -ENOMEM;
1010  	qpl->page_buses = kvcalloc(pages, sizeof(*qpl->page_buses), GFP_KERNEL);
1011  	/* caller handles clean up */
1012  	if (!qpl->page_buses)
1013  		return -ENOMEM;
1014  
1015  	for (i = 0; i < pages; i++) {
1016  		err = gve_alloc_page(priv, &priv->pdev->dev, &qpl->pages[i],
1017  				     &qpl->page_buses[i],
1018  				     gve_qpl_dma_dir(priv, id), GFP_KERNEL);
1019  		/* caller handles clean up */
1020  		if (err)
1021  			return -ENOMEM;
1022  		qpl->num_entries++;
1023  	}
1024  	priv->num_registered_pages += pages;
1025  
1026  	return 0;
1027  }
1028  
gve_free_page(struct device * dev,struct page * page,dma_addr_t dma,enum dma_data_direction dir)1029  void gve_free_page(struct device *dev, struct page *page, dma_addr_t dma,
1030  		   enum dma_data_direction dir)
1031  {
1032  	if (!dma_mapping_error(dev, dma))
1033  		dma_unmap_page(dev, dma, PAGE_SIZE, dir);
1034  	if (page)
1035  		put_page(page);
1036  }
1037  
gve_free_queue_page_list(struct gve_priv * priv,u32 id)1038  static void gve_free_queue_page_list(struct gve_priv *priv, u32 id)
1039  {
1040  	struct gve_queue_page_list *qpl = &priv->qpls[id];
1041  	int i;
1042  
1043  	if (!qpl->pages)
1044  		return;
1045  	if (!qpl->page_buses)
1046  		goto free_pages;
1047  
1048  	for (i = 0; i < qpl->num_entries; i++)
1049  		gve_free_page(&priv->pdev->dev, qpl->pages[i],
1050  			      qpl->page_buses[i], gve_qpl_dma_dir(priv, id));
1051  
1052  	kvfree(qpl->page_buses);
1053  	qpl->page_buses = NULL;
1054  free_pages:
1055  	kvfree(qpl->pages);
1056  	qpl->pages = NULL;
1057  	priv->num_registered_pages -= qpl->num_entries;
1058  }
1059  
gve_alloc_xdp_qpls(struct gve_priv * priv)1060  static int gve_alloc_xdp_qpls(struct gve_priv *priv)
1061  {
1062  	int start_id;
1063  	int i, j;
1064  	int err;
1065  
1066  	start_id = gve_tx_qpl_id(priv, gve_xdp_tx_start_queue_id(priv));
1067  	for (i = start_id; i < start_id + gve_num_xdp_qpls(priv); i++) {
1068  		err = gve_alloc_queue_page_list(priv, i,
1069  						priv->tx_pages_per_qpl);
1070  		if (err)
1071  			goto free_qpls;
1072  	}
1073  
1074  	return 0;
1075  
1076  free_qpls:
1077  	for (j = start_id; j <= i; j++)
1078  		gve_free_queue_page_list(priv, j);
1079  	return err;
1080  }
1081  
gve_alloc_qpls(struct gve_priv * priv)1082  static int gve_alloc_qpls(struct gve_priv *priv)
1083  {
1084  	int max_queues = priv->tx_cfg.max_queues + priv->rx_cfg.max_queues;
1085  	int page_count;
1086  	int start_id;
1087  	int i, j;
1088  	int err;
1089  
1090  	if (!gve_is_qpl(priv))
1091  		return 0;
1092  
1093  	priv->qpls = kvcalloc(max_queues, sizeof(*priv->qpls), GFP_KERNEL);
1094  	if (!priv->qpls)
1095  		return -ENOMEM;
1096  
1097  	start_id = gve_tx_start_qpl_id(priv);
1098  	page_count = priv->tx_pages_per_qpl;
1099  	for (i = start_id; i < start_id + gve_num_tx_qpls(priv); i++) {
1100  		err = gve_alloc_queue_page_list(priv, i,
1101  						page_count);
1102  		if (err)
1103  			goto free_qpls;
1104  	}
1105  
1106  	start_id = gve_rx_start_qpl_id(priv);
1107  
1108  	/* For GQI_QPL number of pages allocated have 1:1 relationship with
1109  	 * number of descriptors. For DQO, number of pages required are
1110  	 * more than descriptors (because of out of order completions).
1111  	 */
1112  	page_count = priv->queue_format == GVE_GQI_QPL_FORMAT ?
1113  		priv->rx_data_slot_cnt : priv->rx_pages_per_qpl;
1114  	for (i = start_id; i < start_id + gve_num_rx_qpls(priv); i++) {
1115  		err = gve_alloc_queue_page_list(priv, i,
1116  						page_count);
1117  		if (err)
1118  			goto free_qpls;
1119  	}
1120  
1121  	priv->qpl_cfg.qpl_map_size = BITS_TO_LONGS(max_queues) *
1122  				     sizeof(unsigned long) * BITS_PER_BYTE;
1123  	priv->qpl_cfg.qpl_id_map = kvcalloc(BITS_TO_LONGS(max_queues),
1124  					    sizeof(unsigned long), GFP_KERNEL);
1125  	if (!priv->qpl_cfg.qpl_id_map) {
1126  		err = -ENOMEM;
1127  		goto free_qpls;
1128  	}
1129  
1130  	return 0;
1131  
1132  free_qpls:
1133  	for (j = 0; j <= i; j++)
1134  		gve_free_queue_page_list(priv, j);
1135  	kvfree(priv->qpls);
1136  	priv->qpls = NULL;
1137  	return err;
1138  }
1139  
gve_free_xdp_qpls(struct gve_priv * priv)1140  static void gve_free_xdp_qpls(struct gve_priv *priv)
1141  {
1142  	int start_id;
1143  	int i;
1144  
1145  	start_id = gve_tx_qpl_id(priv, gve_xdp_tx_start_queue_id(priv));
1146  	for (i = start_id; i < start_id + gve_num_xdp_qpls(priv); i++)
1147  		gve_free_queue_page_list(priv, i);
1148  }
1149  
gve_free_qpls(struct gve_priv * priv)1150  static void gve_free_qpls(struct gve_priv *priv)
1151  {
1152  	int max_queues = priv->tx_cfg.max_queues + priv->rx_cfg.max_queues;
1153  	int i;
1154  
1155  	if (!priv->qpls)
1156  		return;
1157  
1158  	kvfree(priv->qpl_cfg.qpl_id_map);
1159  	priv->qpl_cfg.qpl_id_map = NULL;
1160  
1161  	for (i = 0; i < max_queues; i++)
1162  		gve_free_queue_page_list(priv, i);
1163  
1164  	kvfree(priv->qpls);
1165  	priv->qpls = NULL;
1166  }
1167  
1168  /* Use this to schedule a reset when the device is capable of continuing
1169   * to handle other requests in its current state. If it is not, do a reset
1170   * in thread instead.
1171   */
gve_schedule_reset(struct gve_priv * priv)1172  void gve_schedule_reset(struct gve_priv *priv)
1173  {
1174  	gve_set_do_reset(priv);
1175  	queue_work(priv->gve_wq, &priv->service_task);
1176  }
1177  
1178  static void gve_reset_and_teardown(struct gve_priv *priv, bool was_up);
1179  static int gve_reset_recovery(struct gve_priv *priv, bool was_up);
1180  static void gve_turndown(struct gve_priv *priv);
1181  static void gve_turnup(struct gve_priv *priv);
1182  
gve_reg_xdp_info(struct gve_priv * priv,struct net_device * dev)1183  static int gve_reg_xdp_info(struct gve_priv *priv, struct net_device *dev)
1184  {
1185  	struct napi_struct *napi;
1186  	struct gve_rx_ring *rx;
1187  	int err = 0;
1188  	int i, j;
1189  	u32 tx_qid;
1190  
1191  	if (!priv->num_xdp_queues)
1192  		return 0;
1193  
1194  	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
1195  		rx = &priv->rx[i];
1196  		napi = &priv->ntfy_blocks[rx->ntfy_id].napi;
1197  
1198  		err = xdp_rxq_info_reg(&rx->xdp_rxq, dev, i,
1199  				       napi->napi_id);
1200  		if (err)
1201  			goto err;
1202  		err = xdp_rxq_info_reg_mem_model(&rx->xdp_rxq,
1203  						 MEM_TYPE_PAGE_SHARED, NULL);
1204  		if (err)
1205  			goto err;
1206  		rx->xsk_pool = xsk_get_pool_from_qid(dev, i);
1207  		if (rx->xsk_pool) {
1208  			err = xdp_rxq_info_reg(&rx->xsk_rxq, dev, i,
1209  					       napi->napi_id);
1210  			if (err)
1211  				goto err;
1212  			err = xdp_rxq_info_reg_mem_model(&rx->xsk_rxq,
1213  							 MEM_TYPE_XSK_BUFF_POOL, NULL);
1214  			if (err)
1215  				goto err;
1216  			xsk_pool_set_rxq_info(rx->xsk_pool,
1217  					      &rx->xsk_rxq);
1218  		}
1219  	}
1220  
1221  	for (i = 0; i < priv->num_xdp_queues; i++) {
1222  		tx_qid = gve_xdp_tx_queue_id(priv, i);
1223  		priv->tx[tx_qid].xsk_pool = xsk_get_pool_from_qid(dev, i);
1224  	}
1225  	return 0;
1226  
1227  err:
1228  	for (j = i; j >= 0; j--) {
1229  		rx = &priv->rx[j];
1230  		if (xdp_rxq_info_is_reg(&rx->xdp_rxq))
1231  			xdp_rxq_info_unreg(&rx->xdp_rxq);
1232  		if (xdp_rxq_info_is_reg(&rx->xsk_rxq))
1233  			xdp_rxq_info_unreg(&rx->xsk_rxq);
1234  	}
1235  	return err;
1236  }
1237  
gve_unreg_xdp_info(struct gve_priv * priv)1238  static void gve_unreg_xdp_info(struct gve_priv *priv)
1239  {
1240  	int i, tx_qid;
1241  
1242  	if (!priv->num_xdp_queues)
1243  		return;
1244  
1245  	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
1246  		struct gve_rx_ring *rx = &priv->rx[i];
1247  
1248  		xdp_rxq_info_unreg(&rx->xdp_rxq);
1249  		if (rx->xsk_pool) {
1250  			xdp_rxq_info_unreg(&rx->xsk_rxq);
1251  			rx->xsk_pool = NULL;
1252  		}
1253  	}
1254  
1255  	for (i = 0; i < priv->num_xdp_queues; i++) {
1256  		tx_qid = gve_xdp_tx_queue_id(priv, i);
1257  		priv->tx[tx_qid].xsk_pool = NULL;
1258  	}
1259  }
1260  
gve_drain_page_cache(struct gve_priv * priv)1261  static void gve_drain_page_cache(struct gve_priv *priv)
1262  {
1263  	struct page_frag_cache *nc;
1264  	int i;
1265  
1266  	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
1267  		nc = &priv->rx[i].page_cache;
1268  		if (nc->va) {
1269  			__page_frag_cache_drain(virt_to_page(nc->va),
1270  						nc->pagecnt_bias);
1271  			nc->va = NULL;
1272  		}
1273  	}
1274  }
1275  
gve_open(struct net_device * dev)1276  static int gve_open(struct net_device *dev)
1277  {
1278  	struct gve_priv *priv = netdev_priv(dev);
1279  	int err;
1280  
1281  	if (priv->xdp_prog)
1282  		priv->num_xdp_queues = priv->rx_cfg.num_queues;
1283  	else
1284  		priv->num_xdp_queues = 0;
1285  
1286  	err = gve_alloc_qpls(priv);
1287  	if (err)
1288  		return err;
1289  
1290  	err = gve_alloc_rings(priv);
1291  	if (err)
1292  		goto free_qpls;
1293  
1294  	err = netif_set_real_num_tx_queues(dev, priv->tx_cfg.num_queues);
1295  	if (err)
1296  		goto free_rings;
1297  	err = netif_set_real_num_rx_queues(dev, priv->rx_cfg.num_queues);
1298  	if (err)
1299  		goto free_rings;
1300  
1301  	err = gve_reg_xdp_info(priv, dev);
1302  	if (err)
1303  		goto free_rings;
1304  
1305  	err = gve_register_qpls(priv);
1306  	if (err)
1307  		goto reset;
1308  
1309  	if (!gve_is_gqi(priv)) {
1310  		/* Hard code this for now. This may be tuned in the future for
1311  		 * performance.
1312  		 */
1313  		priv->data_buffer_size_dqo = GVE_RX_BUFFER_SIZE_DQO;
1314  	}
1315  	err = gve_create_rings(priv);
1316  	if (err)
1317  		goto reset;
1318  
1319  	gve_set_device_rings_ok(priv);
1320  
1321  	if (gve_get_report_stats(priv))
1322  		mod_timer(&priv->stats_report_timer,
1323  			  round_jiffies(jiffies +
1324  				msecs_to_jiffies(priv->stats_report_timer_period)));
1325  
1326  	gve_turnup(priv);
1327  	queue_work(priv->gve_wq, &priv->service_task);
1328  	priv->interface_up_cnt++;
1329  	return 0;
1330  
1331  free_rings:
1332  	gve_free_rings(priv);
1333  free_qpls:
1334  	gve_free_qpls(priv);
1335  	return err;
1336  
1337  reset:
1338  	/* This must have been called from a reset due to the rtnl lock
1339  	 * so just return at this point.
1340  	 */
1341  	if (gve_get_reset_in_progress(priv))
1342  		return err;
1343  	/* Otherwise reset before returning */
1344  	gve_reset_and_teardown(priv, true);
1345  	/* if this fails there is nothing we can do so just ignore the return */
1346  	gve_reset_recovery(priv, false);
1347  	/* return the original error */
1348  	return err;
1349  }
1350  
gve_close(struct net_device * dev)1351  static int gve_close(struct net_device *dev)
1352  {
1353  	struct gve_priv *priv = netdev_priv(dev);
1354  	int err;
1355  
1356  	netif_carrier_off(dev);
1357  	if (gve_get_device_rings_ok(priv)) {
1358  		gve_turndown(priv);
1359  		gve_drain_page_cache(priv);
1360  		err = gve_destroy_rings(priv);
1361  		if (err)
1362  			goto err;
1363  		err = gve_unregister_qpls(priv);
1364  		if (err)
1365  			goto err;
1366  		gve_clear_device_rings_ok(priv);
1367  	}
1368  	del_timer_sync(&priv->stats_report_timer);
1369  
1370  	gve_unreg_xdp_info(priv);
1371  	gve_free_rings(priv);
1372  	gve_free_qpls(priv);
1373  	priv->interface_down_cnt++;
1374  	return 0;
1375  
1376  err:
1377  	/* This must have been called from a reset due to the rtnl lock
1378  	 * so just return at this point.
1379  	 */
1380  	if (gve_get_reset_in_progress(priv))
1381  		return err;
1382  	/* Otherwise reset before returning */
1383  	gve_reset_and_teardown(priv, true);
1384  	return gve_reset_recovery(priv, false);
1385  }
1386  
gve_remove_xdp_queues(struct gve_priv * priv)1387  static int gve_remove_xdp_queues(struct gve_priv *priv)
1388  {
1389  	int err;
1390  
1391  	err = gve_destroy_xdp_rings(priv);
1392  	if (err)
1393  		return err;
1394  
1395  	err = gve_unregister_xdp_qpls(priv);
1396  	if (err)
1397  		return err;
1398  
1399  	gve_unreg_xdp_info(priv);
1400  	gve_free_xdp_rings(priv);
1401  	gve_free_xdp_qpls(priv);
1402  	priv->num_xdp_queues = 0;
1403  	return 0;
1404  }
1405  
gve_add_xdp_queues(struct gve_priv * priv)1406  static int gve_add_xdp_queues(struct gve_priv *priv)
1407  {
1408  	int err;
1409  
1410  	priv->num_xdp_queues = priv->tx_cfg.num_queues;
1411  
1412  	err = gve_alloc_xdp_qpls(priv);
1413  	if (err)
1414  		goto err;
1415  
1416  	err = gve_alloc_xdp_rings(priv);
1417  	if (err)
1418  		goto free_xdp_qpls;
1419  
1420  	err = gve_reg_xdp_info(priv, priv->dev);
1421  	if (err)
1422  		goto free_xdp_rings;
1423  
1424  	err = gve_register_xdp_qpls(priv);
1425  	if (err)
1426  		goto free_xdp_rings;
1427  
1428  	err = gve_create_xdp_rings(priv);
1429  	if (err)
1430  		goto free_xdp_rings;
1431  
1432  	return 0;
1433  
1434  free_xdp_rings:
1435  	gve_free_xdp_rings(priv);
1436  free_xdp_qpls:
1437  	gve_free_xdp_qpls(priv);
1438  err:
1439  	priv->num_xdp_queues = 0;
1440  	return err;
1441  }
1442  
gve_handle_link_status(struct gve_priv * priv,bool link_status)1443  static void gve_handle_link_status(struct gve_priv *priv, bool link_status)
1444  {
1445  	if (!gve_get_napi_enabled(priv))
1446  		return;
1447  
1448  	if (link_status == netif_carrier_ok(priv->dev))
1449  		return;
1450  
1451  	if (link_status) {
1452  		netdev_info(priv->dev, "Device link is up.\n");
1453  		netif_carrier_on(priv->dev);
1454  	} else {
1455  		netdev_info(priv->dev, "Device link is down.\n");
1456  		netif_carrier_off(priv->dev);
1457  	}
1458  }
1459  
gve_set_xdp(struct gve_priv * priv,struct bpf_prog * prog,struct netlink_ext_ack * extack)1460  static int gve_set_xdp(struct gve_priv *priv, struct bpf_prog *prog,
1461  		       struct netlink_ext_ack *extack)
1462  {
1463  	struct bpf_prog *old_prog;
1464  	int err = 0;
1465  	u32 status;
1466  
1467  	old_prog = READ_ONCE(priv->xdp_prog);
1468  	if (!netif_carrier_ok(priv->dev)) {
1469  		WRITE_ONCE(priv->xdp_prog, prog);
1470  		if (old_prog)
1471  			bpf_prog_put(old_prog);
1472  		return 0;
1473  	}
1474  
1475  	gve_turndown(priv);
1476  	if (!old_prog && prog) {
1477  		// Allocate XDP TX queues if an XDP program is
1478  		// being installed
1479  		err = gve_add_xdp_queues(priv);
1480  		if (err)
1481  			goto out;
1482  	} else if (old_prog && !prog) {
1483  		// Remove XDP TX queues if an XDP program is
1484  		// being uninstalled
1485  		err = gve_remove_xdp_queues(priv);
1486  		if (err)
1487  			goto out;
1488  	}
1489  	WRITE_ONCE(priv->xdp_prog, prog);
1490  	if (old_prog)
1491  		bpf_prog_put(old_prog);
1492  
1493  out:
1494  	gve_turnup(priv);
1495  	status = ioread32be(&priv->reg_bar0->device_status);
1496  	gve_handle_link_status(priv, GVE_DEVICE_STATUS_LINK_STATUS_MASK & status);
1497  	return err;
1498  }
1499  
gve_xsk_pool_enable(struct net_device * dev,struct xsk_buff_pool * pool,u16 qid)1500  static int gve_xsk_pool_enable(struct net_device *dev,
1501  			       struct xsk_buff_pool *pool,
1502  			       u16 qid)
1503  {
1504  	struct gve_priv *priv = netdev_priv(dev);
1505  	struct napi_struct *napi;
1506  	struct gve_rx_ring *rx;
1507  	int tx_qid;
1508  	int err;
1509  
1510  	if (qid >= priv->rx_cfg.num_queues) {
1511  		dev_err(&priv->pdev->dev, "xsk pool invalid qid %d", qid);
1512  		return -EINVAL;
1513  	}
1514  	if (xsk_pool_get_rx_frame_size(pool) <
1515  	     priv->dev->max_mtu + sizeof(struct ethhdr)) {
1516  		dev_err(&priv->pdev->dev, "xsk pool frame_len too small");
1517  		return -EINVAL;
1518  	}
1519  
1520  	err = xsk_pool_dma_map(pool, &priv->pdev->dev,
1521  			       DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING);
1522  	if (err)
1523  		return err;
1524  
1525  	/* If XDP prog is not installed, return */
1526  	if (!priv->xdp_prog)
1527  		return 0;
1528  
1529  	rx = &priv->rx[qid];
1530  	napi = &priv->ntfy_blocks[rx->ntfy_id].napi;
1531  	err = xdp_rxq_info_reg(&rx->xsk_rxq, dev, qid, napi->napi_id);
1532  	if (err)
1533  		goto err;
1534  
1535  	err = xdp_rxq_info_reg_mem_model(&rx->xsk_rxq,
1536  					 MEM_TYPE_XSK_BUFF_POOL, NULL);
1537  	if (err)
1538  		goto err;
1539  
1540  	xsk_pool_set_rxq_info(pool, &rx->xsk_rxq);
1541  	rx->xsk_pool = pool;
1542  
1543  	tx_qid = gve_xdp_tx_queue_id(priv, qid);
1544  	priv->tx[tx_qid].xsk_pool = pool;
1545  
1546  	return 0;
1547  err:
1548  	if (xdp_rxq_info_is_reg(&rx->xsk_rxq))
1549  		xdp_rxq_info_unreg(&rx->xsk_rxq);
1550  
1551  	xsk_pool_dma_unmap(pool,
1552  			   DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING);
1553  	return err;
1554  }
1555  
gve_xsk_pool_disable(struct net_device * dev,u16 qid)1556  static int gve_xsk_pool_disable(struct net_device *dev,
1557  				u16 qid)
1558  {
1559  	struct gve_priv *priv = netdev_priv(dev);
1560  	struct napi_struct *napi_rx;
1561  	struct napi_struct *napi_tx;
1562  	struct xsk_buff_pool *pool;
1563  	int tx_qid;
1564  
1565  	pool = xsk_get_pool_from_qid(dev, qid);
1566  	if (!pool)
1567  		return -EINVAL;
1568  	if (qid >= priv->rx_cfg.num_queues)
1569  		return -EINVAL;
1570  
1571  	/* If XDP prog is not installed, unmap DMA and return */
1572  	if (!priv->xdp_prog)
1573  		goto done;
1574  
1575  	tx_qid = gve_xdp_tx_queue_id(priv, qid);
1576  	if (!netif_running(dev)) {
1577  		priv->rx[qid].xsk_pool = NULL;
1578  		xdp_rxq_info_unreg(&priv->rx[qid].xsk_rxq);
1579  		priv->tx[tx_qid].xsk_pool = NULL;
1580  		goto done;
1581  	}
1582  
1583  	napi_rx = &priv->ntfy_blocks[priv->rx[qid].ntfy_id].napi;
1584  	napi_disable(napi_rx); /* make sure current rx poll is done */
1585  
1586  	napi_tx = &priv->ntfy_blocks[priv->tx[tx_qid].ntfy_id].napi;
1587  	napi_disable(napi_tx); /* make sure current tx poll is done */
1588  
1589  	priv->rx[qid].xsk_pool = NULL;
1590  	xdp_rxq_info_unreg(&priv->rx[qid].xsk_rxq);
1591  	priv->tx[tx_qid].xsk_pool = NULL;
1592  	smp_mb(); /* Make sure it is visible to the workers on datapath */
1593  
1594  	napi_enable(napi_rx);
1595  	if (gve_rx_work_pending(&priv->rx[qid]))
1596  		napi_schedule(napi_rx);
1597  
1598  	napi_enable(napi_tx);
1599  	if (gve_tx_clean_pending(priv, &priv->tx[tx_qid]))
1600  		napi_schedule(napi_tx);
1601  
1602  done:
1603  	xsk_pool_dma_unmap(pool,
1604  			   DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING);
1605  	return 0;
1606  }
1607  
gve_xsk_wakeup(struct net_device * dev,u32 queue_id,u32 flags)1608  static int gve_xsk_wakeup(struct net_device *dev, u32 queue_id, u32 flags)
1609  {
1610  	struct gve_priv *priv = netdev_priv(dev);
1611  	int tx_queue_id = gve_xdp_tx_queue_id(priv, queue_id);
1612  
1613  	if (queue_id >= priv->rx_cfg.num_queues || !priv->xdp_prog)
1614  		return -EINVAL;
1615  
1616  	if (flags & XDP_WAKEUP_TX) {
1617  		struct gve_tx_ring *tx = &priv->tx[tx_queue_id];
1618  		struct napi_struct *napi =
1619  			&priv->ntfy_blocks[tx->ntfy_id].napi;
1620  
1621  		if (!napi_if_scheduled_mark_missed(napi)) {
1622  			/* Call local_bh_enable to trigger SoftIRQ processing */
1623  			local_bh_disable();
1624  			napi_schedule(napi);
1625  			local_bh_enable();
1626  		}
1627  
1628  		tx->xdp_xsk_wakeup++;
1629  	}
1630  
1631  	return 0;
1632  }
1633  
verify_xdp_configuration(struct net_device * dev)1634  static int verify_xdp_configuration(struct net_device *dev)
1635  {
1636  	struct gve_priv *priv = netdev_priv(dev);
1637  
1638  	if (dev->features & NETIF_F_LRO) {
1639  		netdev_warn(dev, "XDP is not supported when LRO is on.\n");
1640  		return -EOPNOTSUPP;
1641  	}
1642  
1643  	if (priv->queue_format != GVE_GQI_QPL_FORMAT) {
1644  		netdev_warn(dev, "XDP is not supported in mode %d.\n",
1645  			    priv->queue_format);
1646  		return -EOPNOTSUPP;
1647  	}
1648  
1649  	if (dev->mtu > (PAGE_SIZE / 2) - sizeof(struct ethhdr) - GVE_RX_PAD) {
1650  		netdev_warn(dev, "XDP is not supported for mtu %d.\n",
1651  			    dev->mtu);
1652  		return -EOPNOTSUPP;
1653  	}
1654  
1655  	if (priv->rx_cfg.num_queues != priv->tx_cfg.num_queues ||
1656  	    (2 * priv->tx_cfg.num_queues > priv->tx_cfg.max_queues)) {
1657  		netdev_warn(dev, "XDP load failed: The number of configured RX queues %d should be equal to the number of configured TX queues %d and the number of configured RX/TX queues should be less than or equal to half the maximum number of RX/TX queues %d",
1658  			    priv->rx_cfg.num_queues,
1659  			    priv->tx_cfg.num_queues,
1660  			    priv->tx_cfg.max_queues);
1661  		return -EINVAL;
1662  	}
1663  	return 0;
1664  }
1665  
gve_xdp(struct net_device * dev,struct netdev_bpf * xdp)1666  static int gve_xdp(struct net_device *dev, struct netdev_bpf *xdp)
1667  {
1668  	struct gve_priv *priv = netdev_priv(dev);
1669  	int err;
1670  
1671  	err = verify_xdp_configuration(dev);
1672  	if (err)
1673  		return err;
1674  	switch (xdp->command) {
1675  	case XDP_SETUP_PROG:
1676  		return gve_set_xdp(priv, xdp->prog, xdp->extack);
1677  	case XDP_SETUP_XSK_POOL:
1678  		if (xdp->xsk.pool)
1679  			return gve_xsk_pool_enable(dev, xdp->xsk.pool, xdp->xsk.queue_id);
1680  		else
1681  			return gve_xsk_pool_disable(dev, xdp->xsk.queue_id);
1682  	default:
1683  		return -EINVAL;
1684  	}
1685  }
1686  
gve_adjust_queues(struct gve_priv * priv,struct gve_queue_config new_rx_config,struct gve_queue_config new_tx_config)1687  int gve_adjust_queues(struct gve_priv *priv,
1688  		      struct gve_queue_config new_rx_config,
1689  		      struct gve_queue_config new_tx_config)
1690  {
1691  	int err;
1692  
1693  	if (netif_carrier_ok(priv->dev)) {
1694  		/* To make this process as simple as possible we teardown the
1695  		 * device, set the new configuration, and then bring the device
1696  		 * up again.
1697  		 */
1698  		err = gve_close(priv->dev);
1699  		/* we have already tried to reset in close,
1700  		 * just fail at this point
1701  		 */
1702  		if (err)
1703  			return err;
1704  		priv->tx_cfg = new_tx_config;
1705  		priv->rx_cfg = new_rx_config;
1706  
1707  		err = gve_open(priv->dev);
1708  		if (err)
1709  			goto err;
1710  
1711  		return 0;
1712  	}
1713  	/* Set the config for the next up. */
1714  	priv->tx_cfg = new_tx_config;
1715  	priv->rx_cfg = new_rx_config;
1716  
1717  	return 0;
1718  err:
1719  	netif_err(priv, drv, priv->dev,
1720  		  "Adjust queues failed! !!! DISABLING ALL QUEUES !!!\n");
1721  	gve_turndown(priv);
1722  	return err;
1723  }
1724  
gve_turndown(struct gve_priv * priv)1725  static void gve_turndown(struct gve_priv *priv)
1726  {
1727  	int idx;
1728  
1729  	if (netif_carrier_ok(priv->dev))
1730  		netif_carrier_off(priv->dev);
1731  
1732  	if (!gve_get_napi_enabled(priv))
1733  		return;
1734  
1735  	/* Disable napi to prevent more work from coming in */
1736  	for (idx = 0; idx < gve_num_tx_queues(priv); idx++) {
1737  		int ntfy_idx = gve_tx_idx_to_ntfy(priv, idx);
1738  		struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
1739  
1740  		napi_disable(&block->napi);
1741  	}
1742  	for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) {
1743  		int ntfy_idx = gve_rx_idx_to_ntfy(priv, idx);
1744  		struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
1745  
1746  		napi_disable(&block->napi);
1747  	}
1748  
1749  	/* Stop tx queues */
1750  	netif_tx_disable(priv->dev);
1751  
1752  	gve_clear_napi_enabled(priv);
1753  	gve_clear_report_stats(priv);
1754  }
1755  
gve_turnup(struct gve_priv * priv)1756  static void gve_turnup(struct gve_priv *priv)
1757  {
1758  	int idx;
1759  
1760  	/* Start the tx queues */
1761  	netif_tx_start_all_queues(priv->dev);
1762  
1763  	/* Enable napi and unmask interrupts for all queues */
1764  	for (idx = 0; idx < gve_num_tx_queues(priv); idx++) {
1765  		int ntfy_idx = gve_tx_idx_to_ntfy(priv, idx);
1766  		struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
1767  
1768  		napi_enable(&block->napi);
1769  		if (gve_is_gqi(priv)) {
1770  			iowrite32be(0, gve_irq_doorbell(priv, block));
1771  		} else {
1772  			gve_set_itr_coalesce_usecs_dqo(priv, block,
1773  						       priv->tx_coalesce_usecs);
1774  		}
1775  	}
1776  	for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) {
1777  		int ntfy_idx = gve_rx_idx_to_ntfy(priv, idx);
1778  		struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
1779  
1780  		napi_enable(&block->napi);
1781  		if (gve_is_gqi(priv)) {
1782  			iowrite32be(0, gve_irq_doorbell(priv, block));
1783  		} else {
1784  			gve_set_itr_coalesce_usecs_dqo(priv, block,
1785  						       priv->rx_coalesce_usecs);
1786  		}
1787  	}
1788  
1789  	gve_set_napi_enabled(priv);
1790  }
1791  
gve_tx_timeout(struct net_device * dev,unsigned int txqueue)1792  static void gve_tx_timeout(struct net_device *dev, unsigned int txqueue)
1793  {
1794  	struct gve_notify_block *block;
1795  	struct gve_tx_ring *tx = NULL;
1796  	struct gve_priv *priv;
1797  	u32 last_nic_done;
1798  	u32 current_time;
1799  	u32 ntfy_idx;
1800  
1801  	netdev_info(dev, "Timeout on tx queue, %d", txqueue);
1802  	priv = netdev_priv(dev);
1803  	if (txqueue > priv->tx_cfg.num_queues)
1804  		goto reset;
1805  
1806  	ntfy_idx = gve_tx_idx_to_ntfy(priv, txqueue);
1807  	if (ntfy_idx >= priv->num_ntfy_blks)
1808  		goto reset;
1809  
1810  	block = &priv->ntfy_blocks[ntfy_idx];
1811  	tx = block->tx;
1812  
1813  	current_time = jiffies_to_msecs(jiffies);
1814  	if (tx->last_kick_msec + MIN_TX_TIMEOUT_GAP > current_time)
1815  		goto reset;
1816  
1817  	/* Check to see if there are missed completions, which will allow us to
1818  	 * kick the queue.
1819  	 */
1820  	last_nic_done = gve_tx_load_event_counter(priv, tx);
1821  	if (last_nic_done - tx->done) {
1822  		netdev_info(dev, "Kicking queue %d", txqueue);
1823  		iowrite32be(GVE_IRQ_MASK, gve_irq_doorbell(priv, block));
1824  		napi_schedule(&block->napi);
1825  		tx->last_kick_msec = current_time;
1826  		goto out;
1827  	} // Else reset.
1828  
1829  reset:
1830  	gve_schedule_reset(priv);
1831  
1832  out:
1833  	if (tx)
1834  		tx->queue_timeout++;
1835  	priv->tx_timeo_cnt++;
1836  }
1837  
gve_set_features(struct net_device * netdev,netdev_features_t features)1838  static int gve_set_features(struct net_device *netdev,
1839  			    netdev_features_t features)
1840  {
1841  	const netdev_features_t orig_features = netdev->features;
1842  	struct gve_priv *priv = netdev_priv(netdev);
1843  	int err;
1844  
1845  	if ((netdev->features & NETIF_F_LRO) != (features & NETIF_F_LRO)) {
1846  		netdev->features ^= NETIF_F_LRO;
1847  		if (netif_carrier_ok(netdev)) {
1848  			/* To make this process as simple as possible we
1849  			 * teardown the device, set the new configuration,
1850  			 * and then bring the device up again.
1851  			 */
1852  			err = gve_close(netdev);
1853  			/* We have already tried to reset in close, just fail
1854  			 * at this point.
1855  			 */
1856  			if (err)
1857  				goto err;
1858  
1859  			err = gve_open(netdev);
1860  			if (err)
1861  				goto err;
1862  		}
1863  	}
1864  
1865  	return 0;
1866  err:
1867  	/* Reverts the change on error. */
1868  	netdev->features = orig_features;
1869  	netif_err(priv, drv, netdev,
1870  		  "Set features failed! !!! DISABLING ALL QUEUES !!!\n");
1871  	return err;
1872  }
1873  
1874  static const struct net_device_ops gve_netdev_ops = {
1875  	.ndo_start_xmit		=	gve_start_xmit,
1876  	.ndo_open		=	gve_open,
1877  	.ndo_stop		=	gve_close,
1878  	.ndo_get_stats64	=	gve_get_stats,
1879  	.ndo_tx_timeout         =       gve_tx_timeout,
1880  	.ndo_set_features	=	gve_set_features,
1881  	.ndo_bpf		=	gve_xdp,
1882  	.ndo_xdp_xmit		=	gve_xdp_xmit,
1883  	.ndo_xsk_wakeup		=	gve_xsk_wakeup,
1884  };
1885  
gve_handle_status(struct gve_priv * priv,u32 status)1886  static void gve_handle_status(struct gve_priv *priv, u32 status)
1887  {
1888  	if (GVE_DEVICE_STATUS_RESET_MASK & status) {
1889  		dev_info(&priv->pdev->dev, "Device requested reset.\n");
1890  		gve_set_do_reset(priv);
1891  	}
1892  	if (GVE_DEVICE_STATUS_REPORT_STATS_MASK & status) {
1893  		priv->stats_report_trigger_cnt++;
1894  		gve_set_do_report_stats(priv);
1895  	}
1896  }
1897  
gve_handle_reset(struct gve_priv * priv)1898  static void gve_handle_reset(struct gve_priv *priv)
1899  {
1900  	/* A service task will be scheduled at the end of probe to catch any
1901  	 * resets that need to happen, and we don't want to reset until
1902  	 * probe is done.
1903  	 */
1904  	if (gve_get_probe_in_progress(priv))
1905  		return;
1906  
1907  	if (gve_get_do_reset(priv)) {
1908  		rtnl_lock();
1909  		gve_reset(priv, false);
1910  		rtnl_unlock();
1911  	}
1912  }
1913  
gve_handle_report_stats(struct gve_priv * priv)1914  void gve_handle_report_stats(struct gve_priv *priv)
1915  {
1916  	struct stats *stats = priv->stats_report->stats;
1917  	int idx, stats_idx = 0;
1918  	unsigned int start = 0;
1919  	u64 tx_bytes;
1920  
1921  	if (!gve_get_report_stats(priv))
1922  		return;
1923  
1924  	be64_add_cpu(&priv->stats_report->written_count, 1);
1925  	/* tx stats */
1926  	if (priv->tx) {
1927  		for (idx = 0; idx < gve_num_tx_queues(priv); idx++) {
1928  			u32 last_completion = 0;
1929  			u32 tx_frames = 0;
1930  
1931  			/* DQO doesn't currently support these metrics. */
1932  			if (gve_is_gqi(priv)) {
1933  				last_completion = priv->tx[idx].done;
1934  				tx_frames = priv->tx[idx].req;
1935  			}
1936  
1937  			do {
1938  				start = u64_stats_fetch_begin(&priv->tx[idx].statss);
1939  				tx_bytes = priv->tx[idx].bytes_done;
1940  			} while (u64_stats_fetch_retry(&priv->tx[idx].statss, start));
1941  			stats[stats_idx++] = (struct stats) {
1942  				.stat_name = cpu_to_be32(TX_WAKE_CNT),
1943  				.value = cpu_to_be64(priv->tx[idx].wake_queue),
1944  				.queue_id = cpu_to_be32(idx),
1945  			};
1946  			stats[stats_idx++] = (struct stats) {
1947  				.stat_name = cpu_to_be32(TX_STOP_CNT),
1948  				.value = cpu_to_be64(priv->tx[idx].stop_queue),
1949  				.queue_id = cpu_to_be32(idx),
1950  			};
1951  			stats[stats_idx++] = (struct stats) {
1952  				.stat_name = cpu_to_be32(TX_FRAMES_SENT),
1953  				.value = cpu_to_be64(tx_frames),
1954  				.queue_id = cpu_to_be32(idx),
1955  			};
1956  			stats[stats_idx++] = (struct stats) {
1957  				.stat_name = cpu_to_be32(TX_BYTES_SENT),
1958  				.value = cpu_to_be64(tx_bytes),
1959  				.queue_id = cpu_to_be32(idx),
1960  			};
1961  			stats[stats_idx++] = (struct stats) {
1962  				.stat_name = cpu_to_be32(TX_LAST_COMPLETION_PROCESSED),
1963  				.value = cpu_to_be64(last_completion),
1964  				.queue_id = cpu_to_be32(idx),
1965  			};
1966  			stats[stats_idx++] = (struct stats) {
1967  				.stat_name = cpu_to_be32(TX_TIMEOUT_CNT),
1968  				.value = cpu_to_be64(priv->tx[idx].queue_timeout),
1969  				.queue_id = cpu_to_be32(idx),
1970  			};
1971  		}
1972  	}
1973  	/* rx stats */
1974  	if (priv->rx) {
1975  		for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) {
1976  			stats[stats_idx++] = (struct stats) {
1977  				.stat_name = cpu_to_be32(RX_NEXT_EXPECTED_SEQUENCE),
1978  				.value = cpu_to_be64(priv->rx[idx].desc.seqno),
1979  				.queue_id = cpu_to_be32(idx),
1980  			};
1981  			stats[stats_idx++] = (struct stats) {
1982  				.stat_name = cpu_to_be32(RX_BUFFERS_POSTED),
1983  				.value = cpu_to_be64(priv->rx[0].fill_cnt),
1984  				.queue_id = cpu_to_be32(idx),
1985  			};
1986  		}
1987  	}
1988  }
1989  
1990  /* Handle NIC status register changes, reset requests and report stats */
gve_service_task(struct work_struct * work)1991  static void gve_service_task(struct work_struct *work)
1992  {
1993  	struct gve_priv *priv = container_of(work, struct gve_priv,
1994  					     service_task);
1995  	u32 status = ioread32be(&priv->reg_bar0->device_status);
1996  
1997  	gve_handle_status(priv, status);
1998  
1999  	gve_handle_reset(priv);
2000  	gve_handle_link_status(priv, GVE_DEVICE_STATUS_LINK_STATUS_MASK & status);
2001  }
2002  
gve_set_netdev_xdp_features(struct gve_priv * priv)2003  static void gve_set_netdev_xdp_features(struct gve_priv *priv)
2004  {
2005  	if (priv->queue_format == GVE_GQI_QPL_FORMAT) {
2006  		priv->dev->xdp_features = NETDEV_XDP_ACT_BASIC;
2007  		priv->dev->xdp_features |= NETDEV_XDP_ACT_REDIRECT;
2008  		priv->dev->xdp_features |= NETDEV_XDP_ACT_NDO_XMIT;
2009  		priv->dev->xdp_features |= NETDEV_XDP_ACT_XSK_ZEROCOPY;
2010  	} else {
2011  		priv->dev->xdp_features = 0;
2012  	}
2013  }
2014  
gve_init_priv(struct gve_priv * priv,bool skip_describe_device)2015  static int gve_init_priv(struct gve_priv *priv, bool skip_describe_device)
2016  {
2017  	int num_ntfy;
2018  	int err;
2019  
2020  	/* Set up the adminq */
2021  	err = gve_adminq_alloc(&priv->pdev->dev, priv);
2022  	if (err) {
2023  		dev_err(&priv->pdev->dev,
2024  			"Failed to alloc admin queue: err=%d\n", err);
2025  		return err;
2026  	}
2027  
2028  	err = gve_verify_driver_compatibility(priv);
2029  	if (err) {
2030  		dev_err(&priv->pdev->dev,
2031  			"Could not verify driver compatibility: err=%d\n", err);
2032  		goto err;
2033  	}
2034  
2035  	if (skip_describe_device)
2036  		goto setup_device;
2037  
2038  	priv->queue_format = GVE_QUEUE_FORMAT_UNSPECIFIED;
2039  	/* Get the initial information we need from the device */
2040  	err = gve_adminq_describe_device(priv);
2041  	if (err) {
2042  		dev_err(&priv->pdev->dev,
2043  			"Could not get device information: err=%d\n", err);
2044  		goto err;
2045  	}
2046  	priv->dev->mtu = priv->dev->max_mtu;
2047  	num_ntfy = pci_msix_vec_count(priv->pdev);
2048  	if (num_ntfy <= 0) {
2049  		dev_err(&priv->pdev->dev,
2050  			"could not count MSI-x vectors: err=%d\n", num_ntfy);
2051  		err = num_ntfy;
2052  		goto err;
2053  	} else if (num_ntfy < GVE_MIN_MSIX) {
2054  		dev_err(&priv->pdev->dev, "gve needs at least %d MSI-x vectors, but only has %d\n",
2055  			GVE_MIN_MSIX, num_ntfy);
2056  		err = -EINVAL;
2057  		goto err;
2058  	}
2059  
2060  	/* Big TCP is only supported on DQ*/
2061  	if (!gve_is_gqi(priv))
2062  		netif_set_tso_max_size(priv->dev, GVE_DQO_TX_MAX);
2063  
2064  	priv->num_registered_pages = 0;
2065  	priv->rx_copybreak = GVE_DEFAULT_RX_COPYBREAK;
2066  	/* gvnic has one Notification Block per MSI-x vector, except for the
2067  	 * management vector
2068  	 */
2069  	priv->num_ntfy_blks = (num_ntfy - 1) & ~0x1;
2070  	priv->mgmt_msix_idx = priv->num_ntfy_blks;
2071  
2072  	priv->tx_cfg.max_queues =
2073  		min_t(int, priv->tx_cfg.max_queues, priv->num_ntfy_blks / 2);
2074  	priv->rx_cfg.max_queues =
2075  		min_t(int, priv->rx_cfg.max_queues, priv->num_ntfy_blks / 2);
2076  
2077  	priv->tx_cfg.num_queues = priv->tx_cfg.max_queues;
2078  	priv->rx_cfg.num_queues = priv->rx_cfg.max_queues;
2079  	if (priv->default_num_queues > 0) {
2080  		priv->tx_cfg.num_queues = min_t(int, priv->default_num_queues,
2081  						priv->tx_cfg.num_queues);
2082  		priv->rx_cfg.num_queues = min_t(int, priv->default_num_queues,
2083  						priv->rx_cfg.num_queues);
2084  	}
2085  
2086  	dev_info(&priv->pdev->dev, "TX queues %d, RX queues %d\n",
2087  		 priv->tx_cfg.num_queues, priv->rx_cfg.num_queues);
2088  	dev_info(&priv->pdev->dev, "Max TX queues %d, Max RX queues %d\n",
2089  		 priv->tx_cfg.max_queues, priv->rx_cfg.max_queues);
2090  
2091  	if (!gve_is_gqi(priv)) {
2092  		priv->tx_coalesce_usecs = GVE_TX_IRQ_RATELIMIT_US_DQO;
2093  		priv->rx_coalesce_usecs = GVE_RX_IRQ_RATELIMIT_US_DQO;
2094  	}
2095  
2096  setup_device:
2097  	gve_set_netdev_xdp_features(priv);
2098  	err = gve_setup_device_resources(priv);
2099  	if (!err)
2100  		return 0;
2101  err:
2102  	gve_adminq_free(&priv->pdev->dev, priv);
2103  	return err;
2104  }
2105  
gve_teardown_priv_resources(struct gve_priv * priv)2106  static void gve_teardown_priv_resources(struct gve_priv *priv)
2107  {
2108  	gve_teardown_device_resources(priv);
2109  	gve_adminq_free(&priv->pdev->dev, priv);
2110  }
2111  
gve_trigger_reset(struct gve_priv * priv)2112  static void gve_trigger_reset(struct gve_priv *priv)
2113  {
2114  	/* Reset the device by releasing the AQ */
2115  	gve_adminq_release(priv);
2116  }
2117  
gve_reset_and_teardown(struct gve_priv * priv,bool was_up)2118  static void gve_reset_and_teardown(struct gve_priv *priv, bool was_up)
2119  {
2120  	gve_trigger_reset(priv);
2121  	/* With the reset having already happened, close cannot fail */
2122  	if (was_up)
2123  		gve_close(priv->dev);
2124  	gve_teardown_priv_resources(priv);
2125  }
2126  
gve_reset_recovery(struct gve_priv * priv,bool was_up)2127  static int gve_reset_recovery(struct gve_priv *priv, bool was_up)
2128  {
2129  	int err;
2130  
2131  	err = gve_init_priv(priv, true);
2132  	if (err)
2133  		goto err;
2134  	if (was_up) {
2135  		err = gve_open(priv->dev);
2136  		if (err)
2137  			goto err;
2138  	}
2139  	return 0;
2140  err:
2141  	dev_err(&priv->pdev->dev, "Reset failed! !!! DISABLING ALL QUEUES !!!\n");
2142  	gve_turndown(priv);
2143  	return err;
2144  }
2145  
gve_reset(struct gve_priv * priv,bool attempt_teardown)2146  int gve_reset(struct gve_priv *priv, bool attempt_teardown)
2147  {
2148  	bool was_up = netif_carrier_ok(priv->dev);
2149  	int err;
2150  
2151  	dev_info(&priv->pdev->dev, "Performing reset\n");
2152  	gve_clear_do_reset(priv);
2153  	gve_set_reset_in_progress(priv);
2154  	/* If we aren't attempting to teardown normally, just go turndown and
2155  	 * reset right away.
2156  	 */
2157  	if (!attempt_teardown) {
2158  		gve_turndown(priv);
2159  		gve_reset_and_teardown(priv, was_up);
2160  	} else {
2161  		/* Otherwise attempt to close normally */
2162  		if (was_up) {
2163  			err = gve_close(priv->dev);
2164  			/* If that fails reset as we did above */
2165  			if (err)
2166  				gve_reset_and_teardown(priv, was_up);
2167  		}
2168  		/* Clean up any remaining resources */
2169  		gve_teardown_priv_resources(priv);
2170  	}
2171  
2172  	/* Set it all back up */
2173  	err = gve_reset_recovery(priv, was_up);
2174  	gve_clear_reset_in_progress(priv);
2175  	priv->reset_cnt++;
2176  	priv->interface_up_cnt = 0;
2177  	priv->interface_down_cnt = 0;
2178  	priv->stats_report_trigger_cnt = 0;
2179  	return err;
2180  }
2181  
gve_write_version(u8 __iomem * driver_version_register)2182  static void gve_write_version(u8 __iomem *driver_version_register)
2183  {
2184  	const char *c = gve_version_prefix;
2185  
2186  	while (*c) {
2187  		writeb(*c, driver_version_register);
2188  		c++;
2189  	}
2190  
2191  	c = gve_version_str;
2192  	while (*c) {
2193  		writeb(*c, driver_version_register);
2194  		c++;
2195  	}
2196  	writeb('\n', driver_version_register);
2197  }
2198  
gve_probe(struct pci_dev * pdev,const struct pci_device_id * ent)2199  static int gve_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
2200  {
2201  	int max_tx_queues, max_rx_queues;
2202  	struct net_device *dev;
2203  	__be32 __iomem *db_bar;
2204  	struct gve_registers __iomem *reg_bar;
2205  	struct gve_priv *priv;
2206  	int err;
2207  
2208  	err = pci_enable_device(pdev);
2209  	if (err)
2210  		return err;
2211  
2212  	err = pci_request_regions(pdev, gve_driver_name);
2213  	if (err)
2214  		goto abort_with_enabled;
2215  
2216  	pci_set_master(pdev);
2217  
2218  	err = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
2219  	if (err) {
2220  		dev_err(&pdev->dev, "Failed to set dma mask: err=%d\n", err);
2221  		goto abort_with_pci_region;
2222  	}
2223  
2224  	reg_bar = pci_iomap(pdev, GVE_REGISTER_BAR, 0);
2225  	if (!reg_bar) {
2226  		dev_err(&pdev->dev, "Failed to map pci bar!\n");
2227  		err = -ENOMEM;
2228  		goto abort_with_pci_region;
2229  	}
2230  
2231  	db_bar = pci_iomap(pdev, GVE_DOORBELL_BAR, 0);
2232  	if (!db_bar) {
2233  		dev_err(&pdev->dev, "Failed to map doorbell bar!\n");
2234  		err = -ENOMEM;
2235  		goto abort_with_reg_bar;
2236  	}
2237  
2238  	gve_write_version(&reg_bar->driver_version);
2239  	/* Get max queues to alloc etherdev */
2240  	max_tx_queues = ioread32be(&reg_bar->max_tx_queues);
2241  	max_rx_queues = ioread32be(&reg_bar->max_rx_queues);
2242  	/* Alloc and setup the netdev and priv */
2243  	dev = alloc_etherdev_mqs(sizeof(*priv), max_tx_queues, max_rx_queues);
2244  	if (!dev) {
2245  		dev_err(&pdev->dev, "could not allocate netdev\n");
2246  		err = -ENOMEM;
2247  		goto abort_with_db_bar;
2248  	}
2249  	SET_NETDEV_DEV(dev, &pdev->dev);
2250  	pci_set_drvdata(pdev, dev);
2251  	dev->ethtool_ops = &gve_ethtool_ops;
2252  	dev->netdev_ops = &gve_netdev_ops;
2253  
2254  	/* Set default and supported features.
2255  	 *
2256  	 * Features might be set in other locations as well (such as
2257  	 * `gve_adminq_describe_device`).
2258  	 */
2259  	dev->hw_features = NETIF_F_HIGHDMA;
2260  	dev->hw_features |= NETIF_F_SG;
2261  	dev->hw_features |= NETIF_F_HW_CSUM;
2262  	dev->hw_features |= NETIF_F_TSO;
2263  	dev->hw_features |= NETIF_F_TSO6;
2264  	dev->hw_features |= NETIF_F_TSO_ECN;
2265  	dev->hw_features |= NETIF_F_RXCSUM;
2266  	dev->hw_features |= NETIF_F_RXHASH;
2267  	dev->features = dev->hw_features;
2268  	dev->watchdog_timeo = 5 * HZ;
2269  	dev->min_mtu = ETH_MIN_MTU;
2270  	netif_carrier_off(dev);
2271  
2272  	priv = netdev_priv(dev);
2273  	priv->dev = dev;
2274  	priv->pdev = pdev;
2275  	priv->msg_enable = DEFAULT_MSG_LEVEL;
2276  	priv->reg_bar0 = reg_bar;
2277  	priv->db_bar2 = db_bar;
2278  	priv->service_task_flags = 0x0;
2279  	priv->state_flags = 0x0;
2280  	priv->ethtool_flags = 0x0;
2281  
2282  	gve_set_probe_in_progress(priv);
2283  	priv->gve_wq = alloc_ordered_workqueue("gve", 0);
2284  	if (!priv->gve_wq) {
2285  		dev_err(&pdev->dev, "Could not allocate workqueue");
2286  		err = -ENOMEM;
2287  		goto abort_with_netdev;
2288  	}
2289  	INIT_WORK(&priv->service_task, gve_service_task);
2290  	INIT_WORK(&priv->stats_report_task, gve_stats_report_task);
2291  	priv->tx_cfg.max_queues = max_tx_queues;
2292  	priv->rx_cfg.max_queues = max_rx_queues;
2293  
2294  	err = gve_init_priv(priv, false);
2295  	if (err)
2296  		goto abort_with_wq;
2297  
2298  	err = register_netdev(dev);
2299  	if (err)
2300  		goto abort_with_gve_init;
2301  
2302  	dev_info(&pdev->dev, "GVE version %s\n", gve_version_str);
2303  	dev_info(&pdev->dev, "GVE queue format %d\n", (int)priv->queue_format);
2304  	gve_clear_probe_in_progress(priv);
2305  	queue_work(priv->gve_wq, &priv->service_task);
2306  	return 0;
2307  
2308  abort_with_gve_init:
2309  	gve_teardown_priv_resources(priv);
2310  
2311  abort_with_wq:
2312  	destroy_workqueue(priv->gve_wq);
2313  
2314  abort_with_netdev:
2315  	free_netdev(dev);
2316  
2317  abort_with_db_bar:
2318  	pci_iounmap(pdev, db_bar);
2319  
2320  abort_with_reg_bar:
2321  	pci_iounmap(pdev, reg_bar);
2322  
2323  abort_with_pci_region:
2324  	pci_release_regions(pdev);
2325  
2326  abort_with_enabled:
2327  	pci_disable_device(pdev);
2328  	return err;
2329  }
2330  
gve_remove(struct pci_dev * pdev)2331  static void gve_remove(struct pci_dev *pdev)
2332  {
2333  	struct net_device *netdev = pci_get_drvdata(pdev);
2334  	struct gve_priv *priv = netdev_priv(netdev);
2335  	__be32 __iomem *db_bar = priv->db_bar2;
2336  	void __iomem *reg_bar = priv->reg_bar0;
2337  
2338  	unregister_netdev(netdev);
2339  	gve_teardown_priv_resources(priv);
2340  	destroy_workqueue(priv->gve_wq);
2341  	free_netdev(netdev);
2342  	pci_iounmap(pdev, db_bar);
2343  	pci_iounmap(pdev, reg_bar);
2344  	pci_release_regions(pdev);
2345  	pci_disable_device(pdev);
2346  }
2347  
gve_shutdown(struct pci_dev * pdev)2348  static void gve_shutdown(struct pci_dev *pdev)
2349  {
2350  	struct net_device *netdev = pci_get_drvdata(pdev);
2351  	struct gve_priv *priv = netdev_priv(netdev);
2352  	bool was_up = netif_carrier_ok(priv->dev);
2353  
2354  	rtnl_lock();
2355  	if (was_up && gve_close(priv->dev)) {
2356  		/* If the dev was up, attempt to close, if close fails, reset */
2357  		gve_reset_and_teardown(priv, was_up);
2358  	} else {
2359  		/* If the dev wasn't up or close worked, finish tearing down */
2360  		gve_teardown_priv_resources(priv);
2361  	}
2362  	rtnl_unlock();
2363  }
2364  
2365  #ifdef CONFIG_PM
gve_suspend(struct pci_dev * pdev,pm_message_t state)2366  static int gve_suspend(struct pci_dev *pdev, pm_message_t state)
2367  {
2368  	struct net_device *netdev = pci_get_drvdata(pdev);
2369  	struct gve_priv *priv = netdev_priv(netdev);
2370  	bool was_up = netif_carrier_ok(priv->dev);
2371  
2372  	priv->suspend_cnt++;
2373  	rtnl_lock();
2374  	if (was_up && gve_close(priv->dev)) {
2375  		/* If the dev was up, attempt to close, if close fails, reset */
2376  		gve_reset_and_teardown(priv, was_up);
2377  	} else {
2378  		/* If the dev wasn't up or close worked, finish tearing down */
2379  		gve_teardown_priv_resources(priv);
2380  	}
2381  	priv->up_before_suspend = was_up;
2382  	rtnl_unlock();
2383  	return 0;
2384  }
2385  
gve_resume(struct pci_dev * pdev)2386  static int gve_resume(struct pci_dev *pdev)
2387  {
2388  	struct net_device *netdev = pci_get_drvdata(pdev);
2389  	struct gve_priv *priv = netdev_priv(netdev);
2390  	int err;
2391  
2392  	priv->resume_cnt++;
2393  	rtnl_lock();
2394  	err = gve_reset_recovery(priv, priv->up_before_suspend);
2395  	rtnl_unlock();
2396  	return err;
2397  }
2398  #endif /* CONFIG_PM */
2399  
2400  static const struct pci_device_id gve_id_table[] = {
2401  	{ PCI_DEVICE(PCI_VENDOR_ID_GOOGLE, PCI_DEV_ID_GVNIC) },
2402  	{ }
2403  };
2404  
2405  static struct pci_driver gve_driver = {
2406  	.name		= gve_driver_name,
2407  	.id_table	= gve_id_table,
2408  	.probe		= gve_probe,
2409  	.remove		= gve_remove,
2410  	.shutdown	= gve_shutdown,
2411  #ifdef CONFIG_PM
2412  	.suspend        = gve_suspend,
2413  	.resume         = gve_resume,
2414  #endif
2415  };
2416  
2417  module_pci_driver(gve_driver);
2418  
2419  MODULE_DEVICE_TABLE(pci, gve_id_table);
2420  MODULE_AUTHOR("Google, Inc.");
2421  MODULE_DESCRIPTION("Google Virtual NIC Driver");
2422  MODULE_LICENSE("Dual MIT/GPL");
2423  MODULE_VERSION(GVE_VERSION);
2424