1 // SPDX-License-Identifier: (GPL-2.0 OR MIT)
2 /* Google virtual Ethernet (gve) driver
3 *
4 * Copyright (C) 2015-2021 Google, Inc.
5 */
6
7 #include <linux/bpf.h>
8 #include <linux/cpumask.h>
9 #include <linux/etherdevice.h>
10 #include <linux/filter.h>
11 #include <linux/interrupt.h>
12 #include <linux/module.h>
13 #include <linux/pci.h>
14 #include <linux/sched.h>
15 #include <linux/timer.h>
16 #include <linux/workqueue.h>
17 #include <linux/utsname.h>
18 #include <linux/version.h>
19 #include <net/sch_generic.h>
20 #include <net/xdp_sock_drv.h>
21 #include "gve.h"
22 #include "gve_dqo.h"
23 #include "gve_adminq.h"
24 #include "gve_register.h"
25
26 #define GVE_DEFAULT_RX_COPYBREAK (256)
27
28 #define DEFAULT_MSG_LEVEL (NETIF_MSG_DRV | NETIF_MSG_LINK)
29 #define GVE_VERSION "1.0.0"
30 #define GVE_VERSION_PREFIX "GVE-"
31
32 // Minimum amount of time between queue kicks in msec (10 seconds)
33 #define MIN_TX_TIMEOUT_GAP (1000 * 10)
34
35 char gve_driver_name[] = "gve";
36 const char gve_version_str[] = GVE_VERSION;
37 static const char gve_version_prefix[] = GVE_VERSION_PREFIX;
38
gve_verify_driver_compatibility(struct gve_priv * priv)39 static int gve_verify_driver_compatibility(struct gve_priv *priv)
40 {
41 int err;
42 struct gve_driver_info *driver_info;
43 dma_addr_t driver_info_bus;
44
45 driver_info = dma_alloc_coherent(&priv->pdev->dev,
46 sizeof(struct gve_driver_info),
47 &driver_info_bus, GFP_KERNEL);
48 if (!driver_info)
49 return -ENOMEM;
50
51 *driver_info = (struct gve_driver_info) {
52 .os_type = 1, /* Linux */
53 .os_version_major = cpu_to_be32(LINUX_VERSION_MAJOR),
54 .os_version_minor = cpu_to_be32(LINUX_VERSION_SUBLEVEL),
55 .os_version_sub = cpu_to_be32(LINUX_VERSION_PATCHLEVEL),
56 .driver_capability_flags = {
57 cpu_to_be64(GVE_DRIVER_CAPABILITY_FLAGS1),
58 cpu_to_be64(GVE_DRIVER_CAPABILITY_FLAGS2),
59 cpu_to_be64(GVE_DRIVER_CAPABILITY_FLAGS3),
60 cpu_to_be64(GVE_DRIVER_CAPABILITY_FLAGS4),
61 },
62 };
63 strscpy(driver_info->os_version_str1, utsname()->release,
64 sizeof(driver_info->os_version_str1));
65 strscpy(driver_info->os_version_str2, utsname()->version,
66 sizeof(driver_info->os_version_str2));
67
68 err = gve_adminq_verify_driver_compatibility(priv,
69 sizeof(struct gve_driver_info),
70 driver_info_bus);
71
72 /* It's ok if the device doesn't support this */
73 if (err == -EOPNOTSUPP)
74 err = 0;
75
76 dma_free_coherent(&priv->pdev->dev,
77 sizeof(struct gve_driver_info),
78 driver_info, driver_info_bus);
79 return err;
80 }
81
gve_start_xmit(struct sk_buff * skb,struct net_device * dev)82 static netdev_tx_t gve_start_xmit(struct sk_buff *skb, struct net_device *dev)
83 {
84 struct gve_priv *priv = netdev_priv(dev);
85
86 if (gve_is_gqi(priv))
87 return gve_tx(skb, dev);
88 else
89 return gve_tx_dqo(skb, dev);
90 }
91
gve_get_stats(struct net_device * dev,struct rtnl_link_stats64 * s)92 static void gve_get_stats(struct net_device *dev, struct rtnl_link_stats64 *s)
93 {
94 struct gve_priv *priv = netdev_priv(dev);
95 unsigned int start;
96 u64 packets, bytes;
97 int num_tx_queues;
98 int ring;
99
100 num_tx_queues = gve_num_tx_queues(priv);
101 if (priv->rx) {
102 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) {
103 do {
104 start =
105 u64_stats_fetch_begin(&priv->rx[ring].statss);
106 packets = priv->rx[ring].rpackets;
107 bytes = priv->rx[ring].rbytes;
108 } while (u64_stats_fetch_retry(&priv->rx[ring].statss,
109 start));
110 s->rx_packets += packets;
111 s->rx_bytes += bytes;
112 }
113 }
114 if (priv->tx) {
115 for (ring = 0; ring < num_tx_queues; ring++) {
116 do {
117 start =
118 u64_stats_fetch_begin(&priv->tx[ring].statss);
119 packets = priv->tx[ring].pkt_done;
120 bytes = priv->tx[ring].bytes_done;
121 } while (u64_stats_fetch_retry(&priv->tx[ring].statss,
122 start));
123 s->tx_packets += packets;
124 s->tx_bytes += bytes;
125 }
126 }
127 }
128
gve_alloc_counter_array(struct gve_priv * priv)129 static int gve_alloc_counter_array(struct gve_priv *priv)
130 {
131 priv->counter_array =
132 dma_alloc_coherent(&priv->pdev->dev,
133 priv->num_event_counters *
134 sizeof(*priv->counter_array),
135 &priv->counter_array_bus, GFP_KERNEL);
136 if (!priv->counter_array)
137 return -ENOMEM;
138
139 return 0;
140 }
141
gve_free_counter_array(struct gve_priv * priv)142 static void gve_free_counter_array(struct gve_priv *priv)
143 {
144 if (!priv->counter_array)
145 return;
146
147 dma_free_coherent(&priv->pdev->dev,
148 priv->num_event_counters *
149 sizeof(*priv->counter_array),
150 priv->counter_array, priv->counter_array_bus);
151 priv->counter_array = NULL;
152 }
153
154 /* NIC requests to report stats */
gve_stats_report_task(struct work_struct * work)155 static void gve_stats_report_task(struct work_struct *work)
156 {
157 struct gve_priv *priv = container_of(work, struct gve_priv,
158 stats_report_task);
159 if (gve_get_do_report_stats(priv)) {
160 gve_handle_report_stats(priv);
161 gve_clear_do_report_stats(priv);
162 }
163 }
164
gve_stats_report_schedule(struct gve_priv * priv)165 static void gve_stats_report_schedule(struct gve_priv *priv)
166 {
167 if (!gve_get_probe_in_progress(priv) &&
168 !gve_get_reset_in_progress(priv)) {
169 gve_set_do_report_stats(priv);
170 queue_work(priv->gve_wq, &priv->stats_report_task);
171 }
172 }
173
gve_stats_report_timer(struct timer_list * t)174 static void gve_stats_report_timer(struct timer_list *t)
175 {
176 struct gve_priv *priv = from_timer(priv, t, stats_report_timer);
177
178 mod_timer(&priv->stats_report_timer,
179 round_jiffies(jiffies +
180 msecs_to_jiffies(priv->stats_report_timer_period)));
181 gve_stats_report_schedule(priv);
182 }
183
gve_alloc_stats_report(struct gve_priv * priv)184 static int gve_alloc_stats_report(struct gve_priv *priv)
185 {
186 int tx_stats_num, rx_stats_num;
187
188 tx_stats_num = (GVE_TX_STATS_REPORT_NUM + NIC_TX_STATS_REPORT_NUM) *
189 gve_num_tx_queues(priv);
190 rx_stats_num = (GVE_RX_STATS_REPORT_NUM + NIC_RX_STATS_REPORT_NUM) *
191 priv->rx_cfg.num_queues;
192 priv->stats_report_len = struct_size(priv->stats_report, stats,
193 tx_stats_num + rx_stats_num);
194 priv->stats_report =
195 dma_alloc_coherent(&priv->pdev->dev, priv->stats_report_len,
196 &priv->stats_report_bus, GFP_KERNEL);
197 if (!priv->stats_report)
198 return -ENOMEM;
199 /* Set up timer for the report-stats task */
200 timer_setup(&priv->stats_report_timer, gve_stats_report_timer, 0);
201 priv->stats_report_timer_period = GVE_STATS_REPORT_TIMER_PERIOD;
202 return 0;
203 }
204
gve_free_stats_report(struct gve_priv * priv)205 static void gve_free_stats_report(struct gve_priv *priv)
206 {
207 if (!priv->stats_report)
208 return;
209
210 del_timer_sync(&priv->stats_report_timer);
211 dma_free_coherent(&priv->pdev->dev, priv->stats_report_len,
212 priv->stats_report, priv->stats_report_bus);
213 priv->stats_report = NULL;
214 }
215
gve_mgmnt_intr(int irq,void * arg)216 static irqreturn_t gve_mgmnt_intr(int irq, void *arg)
217 {
218 struct gve_priv *priv = arg;
219
220 queue_work(priv->gve_wq, &priv->service_task);
221 return IRQ_HANDLED;
222 }
223
gve_intr(int irq,void * arg)224 static irqreturn_t gve_intr(int irq, void *arg)
225 {
226 struct gve_notify_block *block = arg;
227 struct gve_priv *priv = block->priv;
228
229 iowrite32be(GVE_IRQ_MASK, gve_irq_doorbell(priv, block));
230 napi_schedule_irqoff(&block->napi);
231 return IRQ_HANDLED;
232 }
233
gve_intr_dqo(int irq,void * arg)234 static irqreturn_t gve_intr_dqo(int irq, void *arg)
235 {
236 struct gve_notify_block *block = arg;
237
238 /* Interrupts are automatically masked */
239 napi_schedule_irqoff(&block->napi);
240 return IRQ_HANDLED;
241 }
242
gve_napi_poll(struct napi_struct * napi,int budget)243 static int gve_napi_poll(struct napi_struct *napi, int budget)
244 {
245 struct gve_notify_block *block;
246 __be32 __iomem *irq_doorbell;
247 bool reschedule = false;
248 struct gve_priv *priv;
249 int work_done = 0;
250
251 block = container_of(napi, struct gve_notify_block, napi);
252 priv = block->priv;
253
254 if (block->tx) {
255 if (block->tx->q_num < priv->tx_cfg.num_queues)
256 reschedule |= gve_tx_poll(block, budget);
257 else
258 reschedule |= gve_xdp_poll(block, budget);
259 }
260
261 if (block->rx) {
262 work_done = gve_rx_poll(block, budget);
263 reschedule |= work_done == budget;
264 }
265
266 if (reschedule)
267 return budget;
268
269 /* Complete processing - don't unmask irq if busy polling is enabled */
270 if (likely(napi_complete_done(napi, work_done))) {
271 irq_doorbell = gve_irq_doorbell(priv, block);
272 iowrite32be(GVE_IRQ_ACK | GVE_IRQ_EVENT, irq_doorbell);
273
274 /* Ensure IRQ ACK is visible before we check pending work.
275 * If queue had issued updates, it would be truly visible.
276 */
277 mb();
278
279 if (block->tx)
280 reschedule |= gve_tx_clean_pending(priv, block->tx);
281 if (block->rx)
282 reschedule |= gve_rx_work_pending(block->rx);
283
284 if (reschedule && napi_reschedule(napi))
285 iowrite32be(GVE_IRQ_MASK, irq_doorbell);
286 }
287 return work_done;
288 }
289
gve_napi_poll_dqo(struct napi_struct * napi,int budget)290 static int gve_napi_poll_dqo(struct napi_struct *napi, int budget)
291 {
292 struct gve_notify_block *block =
293 container_of(napi, struct gve_notify_block, napi);
294 struct gve_priv *priv = block->priv;
295 bool reschedule = false;
296 int work_done = 0;
297
298 if (block->tx)
299 reschedule |= gve_tx_poll_dqo(block, /*do_clean=*/true);
300
301 if (block->rx) {
302 work_done = gve_rx_poll_dqo(block, budget);
303 reschedule |= work_done == budget;
304 }
305
306 if (reschedule)
307 return budget;
308
309 if (likely(napi_complete_done(napi, work_done))) {
310 /* Enable interrupts again.
311 *
312 * We don't need to repoll afterwards because HW supports the
313 * PCI MSI-X PBA feature.
314 *
315 * Another interrupt would be triggered if a new event came in
316 * since the last one.
317 */
318 gve_write_irq_doorbell_dqo(priv, block,
319 GVE_ITR_NO_UPDATE_DQO | GVE_ITR_ENABLE_BIT_DQO);
320 }
321
322 return work_done;
323 }
324
gve_alloc_notify_blocks(struct gve_priv * priv)325 static int gve_alloc_notify_blocks(struct gve_priv *priv)
326 {
327 int num_vecs_requested = priv->num_ntfy_blks + 1;
328 unsigned int active_cpus;
329 int vecs_enabled;
330 int i, j;
331 int err;
332
333 priv->msix_vectors = kvcalloc(num_vecs_requested,
334 sizeof(*priv->msix_vectors), GFP_KERNEL);
335 if (!priv->msix_vectors)
336 return -ENOMEM;
337 for (i = 0; i < num_vecs_requested; i++)
338 priv->msix_vectors[i].entry = i;
339 vecs_enabled = pci_enable_msix_range(priv->pdev, priv->msix_vectors,
340 GVE_MIN_MSIX, num_vecs_requested);
341 if (vecs_enabled < 0) {
342 dev_err(&priv->pdev->dev, "Could not enable min msix %d/%d\n",
343 GVE_MIN_MSIX, vecs_enabled);
344 err = vecs_enabled;
345 goto abort_with_msix_vectors;
346 }
347 if (vecs_enabled != num_vecs_requested) {
348 int new_num_ntfy_blks = (vecs_enabled - 1) & ~0x1;
349 int vecs_per_type = new_num_ntfy_blks / 2;
350 int vecs_left = new_num_ntfy_blks % 2;
351
352 priv->num_ntfy_blks = new_num_ntfy_blks;
353 priv->mgmt_msix_idx = priv->num_ntfy_blks;
354 priv->tx_cfg.max_queues = min_t(int, priv->tx_cfg.max_queues,
355 vecs_per_type);
356 priv->rx_cfg.max_queues = min_t(int, priv->rx_cfg.max_queues,
357 vecs_per_type + vecs_left);
358 dev_err(&priv->pdev->dev,
359 "Could not enable desired msix, only enabled %d, adjusting tx max queues to %d, and rx max queues to %d\n",
360 vecs_enabled, priv->tx_cfg.max_queues,
361 priv->rx_cfg.max_queues);
362 if (priv->tx_cfg.num_queues > priv->tx_cfg.max_queues)
363 priv->tx_cfg.num_queues = priv->tx_cfg.max_queues;
364 if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues)
365 priv->rx_cfg.num_queues = priv->rx_cfg.max_queues;
366 }
367 /* Half the notification blocks go to TX and half to RX */
368 active_cpus = min_t(int, priv->num_ntfy_blks / 2, num_online_cpus());
369
370 /* Setup Management Vector - the last vector */
371 snprintf(priv->mgmt_msix_name, sizeof(priv->mgmt_msix_name), "gve-mgmnt@pci:%s",
372 pci_name(priv->pdev));
373 err = request_irq(priv->msix_vectors[priv->mgmt_msix_idx].vector,
374 gve_mgmnt_intr, 0, priv->mgmt_msix_name, priv);
375 if (err) {
376 dev_err(&priv->pdev->dev, "Did not receive management vector.\n");
377 goto abort_with_msix_enabled;
378 }
379 priv->irq_db_indices =
380 dma_alloc_coherent(&priv->pdev->dev,
381 priv->num_ntfy_blks *
382 sizeof(*priv->irq_db_indices),
383 &priv->irq_db_indices_bus, GFP_KERNEL);
384 if (!priv->irq_db_indices) {
385 err = -ENOMEM;
386 goto abort_with_mgmt_vector;
387 }
388
389 priv->ntfy_blocks = kvzalloc(priv->num_ntfy_blks *
390 sizeof(*priv->ntfy_blocks), GFP_KERNEL);
391 if (!priv->ntfy_blocks) {
392 err = -ENOMEM;
393 goto abort_with_irq_db_indices;
394 }
395
396 /* Setup the other blocks - the first n-1 vectors */
397 for (i = 0; i < priv->num_ntfy_blks; i++) {
398 struct gve_notify_block *block = &priv->ntfy_blocks[i];
399 int msix_idx = i;
400
401 snprintf(block->name, sizeof(block->name), "gve-ntfy-blk%d@pci:%s",
402 i, pci_name(priv->pdev));
403 block->priv = priv;
404 err = request_irq(priv->msix_vectors[msix_idx].vector,
405 gve_is_gqi(priv) ? gve_intr : gve_intr_dqo,
406 0, block->name, block);
407 if (err) {
408 dev_err(&priv->pdev->dev,
409 "Failed to receive msix vector %d\n", i);
410 goto abort_with_some_ntfy_blocks;
411 }
412 irq_set_affinity_hint(priv->msix_vectors[msix_idx].vector,
413 get_cpu_mask(i % active_cpus));
414 block->irq_db_index = &priv->irq_db_indices[i].index;
415 }
416 return 0;
417 abort_with_some_ntfy_blocks:
418 for (j = 0; j < i; j++) {
419 struct gve_notify_block *block = &priv->ntfy_blocks[j];
420 int msix_idx = j;
421
422 irq_set_affinity_hint(priv->msix_vectors[msix_idx].vector,
423 NULL);
424 free_irq(priv->msix_vectors[msix_idx].vector, block);
425 }
426 kvfree(priv->ntfy_blocks);
427 priv->ntfy_blocks = NULL;
428 abort_with_irq_db_indices:
429 dma_free_coherent(&priv->pdev->dev, priv->num_ntfy_blks *
430 sizeof(*priv->irq_db_indices),
431 priv->irq_db_indices, priv->irq_db_indices_bus);
432 priv->irq_db_indices = NULL;
433 abort_with_mgmt_vector:
434 free_irq(priv->msix_vectors[priv->mgmt_msix_idx].vector, priv);
435 abort_with_msix_enabled:
436 pci_disable_msix(priv->pdev);
437 abort_with_msix_vectors:
438 kvfree(priv->msix_vectors);
439 priv->msix_vectors = NULL;
440 return err;
441 }
442
gve_free_notify_blocks(struct gve_priv * priv)443 static void gve_free_notify_blocks(struct gve_priv *priv)
444 {
445 int i;
446
447 if (!priv->msix_vectors)
448 return;
449
450 /* Free the irqs */
451 for (i = 0; i < priv->num_ntfy_blks; i++) {
452 struct gve_notify_block *block = &priv->ntfy_blocks[i];
453 int msix_idx = i;
454
455 irq_set_affinity_hint(priv->msix_vectors[msix_idx].vector,
456 NULL);
457 free_irq(priv->msix_vectors[msix_idx].vector, block);
458 }
459 free_irq(priv->msix_vectors[priv->mgmt_msix_idx].vector, priv);
460 kvfree(priv->ntfy_blocks);
461 priv->ntfy_blocks = NULL;
462 dma_free_coherent(&priv->pdev->dev, priv->num_ntfy_blks *
463 sizeof(*priv->irq_db_indices),
464 priv->irq_db_indices, priv->irq_db_indices_bus);
465 priv->irq_db_indices = NULL;
466 pci_disable_msix(priv->pdev);
467 kvfree(priv->msix_vectors);
468 priv->msix_vectors = NULL;
469 }
470
gve_setup_device_resources(struct gve_priv * priv)471 static int gve_setup_device_resources(struct gve_priv *priv)
472 {
473 int err;
474
475 err = gve_alloc_counter_array(priv);
476 if (err)
477 return err;
478 err = gve_alloc_notify_blocks(priv);
479 if (err)
480 goto abort_with_counter;
481 err = gve_alloc_stats_report(priv);
482 if (err)
483 goto abort_with_ntfy_blocks;
484 err = gve_adminq_configure_device_resources(priv,
485 priv->counter_array_bus,
486 priv->num_event_counters,
487 priv->irq_db_indices_bus,
488 priv->num_ntfy_blks);
489 if (unlikely(err)) {
490 dev_err(&priv->pdev->dev,
491 "could not setup device_resources: err=%d\n", err);
492 err = -ENXIO;
493 goto abort_with_stats_report;
494 }
495
496 if (!gve_is_gqi(priv)) {
497 priv->ptype_lut_dqo = kvzalloc(sizeof(*priv->ptype_lut_dqo),
498 GFP_KERNEL);
499 if (!priv->ptype_lut_dqo) {
500 err = -ENOMEM;
501 goto abort_with_stats_report;
502 }
503 err = gve_adminq_get_ptype_map_dqo(priv, priv->ptype_lut_dqo);
504 if (err) {
505 dev_err(&priv->pdev->dev,
506 "Failed to get ptype map: err=%d\n", err);
507 goto abort_with_ptype_lut;
508 }
509 }
510
511 err = gve_adminq_report_stats(priv, priv->stats_report_len,
512 priv->stats_report_bus,
513 GVE_STATS_REPORT_TIMER_PERIOD);
514 if (err)
515 dev_err(&priv->pdev->dev,
516 "Failed to report stats: err=%d\n", err);
517 gve_set_device_resources_ok(priv);
518 return 0;
519
520 abort_with_ptype_lut:
521 kvfree(priv->ptype_lut_dqo);
522 priv->ptype_lut_dqo = NULL;
523 abort_with_stats_report:
524 gve_free_stats_report(priv);
525 abort_with_ntfy_blocks:
526 gve_free_notify_blocks(priv);
527 abort_with_counter:
528 gve_free_counter_array(priv);
529
530 return err;
531 }
532
533 static void gve_trigger_reset(struct gve_priv *priv);
534
gve_teardown_device_resources(struct gve_priv * priv)535 static void gve_teardown_device_resources(struct gve_priv *priv)
536 {
537 int err;
538
539 /* Tell device its resources are being freed */
540 if (gve_get_device_resources_ok(priv)) {
541 /* detach the stats report */
542 err = gve_adminq_report_stats(priv, 0, 0x0, GVE_STATS_REPORT_TIMER_PERIOD);
543 if (err) {
544 dev_err(&priv->pdev->dev,
545 "Failed to detach stats report: err=%d\n", err);
546 gve_trigger_reset(priv);
547 }
548 err = gve_adminq_deconfigure_device_resources(priv);
549 if (err) {
550 dev_err(&priv->pdev->dev,
551 "Could not deconfigure device resources: err=%d\n",
552 err);
553 gve_trigger_reset(priv);
554 }
555 }
556
557 kvfree(priv->ptype_lut_dqo);
558 priv->ptype_lut_dqo = NULL;
559
560 gve_free_counter_array(priv);
561 gve_free_notify_blocks(priv);
562 gve_free_stats_report(priv);
563 gve_clear_device_resources_ok(priv);
564 }
565
gve_add_napi(struct gve_priv * priv,int ntfy_idx,int (* gve_poll)(struct napi_struct *,int))566 static void gve_add_napi(struct gve_priv *priv, int ntfy_idx,
567 int (*gve_poll)(struct napi_struct *, int))
568 {
569 struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
570
571 netif_napi_add(priv->dev, &block->napi, gve_poll);
572 }
573
gve_remove_napi(struct gve_priv * priv,int ntfy_idx)574 static void gve_remove_napi(struct gve_priv *priv, int ntfy_idx)
575 {
576 struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
577
578 netif_napi_del(&block->napi);
579 }
580
gve_register_xdp_qpls(struct gve_priv * priv)581 static int gve_register_xdp_qpls(struct gve_priv *priv)
582 {
583 int start_id;
584 int err;
585 int i;
586
587 start_id = gve_tx_qpl_id(priv, gve_xdp_tx_start_queue_id(priv));
588 for (i = start_id; i < start_id + gve_num_xdp_qpls(priv); i++) {
589 err = gve_adminq_register_page_list(priv, &priv->qpls[i]);
590 if (err) {
591 netif_err(priv, drv, priv->dev,
592 "failed to register queue page list %d\n",
593 priv->qpls[i].id);
594 /* This failure will trigger a reset - no need to clean
595 * up
596 */
597 return err;
598 }
599 }
600 return 0;
601 }
602
gve_register_qpls(struct gve_priv * priv)603 static int gve_register_qpls(struct gve_priv *priv)
604 {
605 int start_id;
606 int err;
607 int i;
608
609 start_id = gve_tx_start_qpl_id(priv);
610 for (i = start_id; i < start_id + gve_num_tx_qpls(priv); i++) {
611 err = gve_adminq_register_page_list(priv, &priv->qpls[i]);
612 if (err) {
613 netif_err(priv, drv, priv->dev,
614 "failed to register queue page list %d\n",
615 priv->qpls[i].id);
616 /* This failure will trigger a reset - no need to clean
617 * up
618 */
619 return err;
620 }
621 }
622
623 start_id = gve_rx_start_qpl_id(priv);
624 for (i = start_id; i < start_id + gve_num_rx_qpls(priv); i++) {
625 err = gve_adminq_register_page_list(priv, &priv->qpls[i]);
626 if (err) {
627 netif_err(priv, drv, priv->dev,
628 "failed to register queue page list %d\n",
629 priv->qpls[i].id);
630 /* This failure will trigger a reset - no need to clean
631 * up
632 */
633 return err;
634 }
635 }
636 return 0;
637 }
638
gve_unregister_xdp_qpls(struct gve_priv * priv)639 static int gve_unregister_xdp_qpls(struct gve_priv *priv)
640 {
641 int start_id;
642 int err;
643 int i;
644
645 start_id = gve_tx_qpl_id(priv, gve_xdp_tx_start_queue_id(priv));
646 for (i = start_id; i < start_id + gve_num_xdp_qpls(priv); i++) {
647 err = gve_adminq_unregister_page_list(priv, priv->qpls[i].id);
648 /* This failure will trigger a reset - no need to clean up */
649 if (err) {
650 netif_err(priv, drv, priv->dev,
651 "Failed to unregister queue page list %d\n",
652 priv->qpls[i].id);
653 return err;
654 }
655 }
656 return 0;
657 }
658
gve_unregister_qpls(struct gve_priv * priv)659 static int gve_unregister_qpls(struct gve_priv *priv)
660 {
661 int start_id;
662 int err;
663 int i;
664
665 start_id = gve_tx_start_qpl_id(priv);
666 for (i = start_id; i < start_id + gve_num_tx_qpls(priv); i++) {
667 err = gve_adminq_unregister_page_list(priv, priv->qpls[i].id);
668 /* This failure will trigger a reset - no need to clean up */
669 if (err) {
670 netif_err(priv, drv, priv->dev,
671 "Failed to unregister queue page list %d\n",
672 priv->qpls[i].id);
673 return err;
674 }
675 }
676
677 start_id = gve_rx_start_qpl_id(priv);
678 for (i = start_id; i < start_id + gve_num_rx_qpls(priv); i++) {
679 err = gve_adminq_unregister_page_list(priv, priv->qpls[i].id);
680 /* This failure will trigger a reset - no need to clean up */
681 if (err) {
682 netif_err(priv, drv, priv->dev,
683 "Failed to unregister queue page list %d\n",
684 priv->qpls[i].id);
685 return err;
686 }
687 }
688 return 0;
689 }
690
gve_create_xdp_rings(struct gve_priv * priv)691 static int gve_create_xdp_rings(struct gve_priv *priv)
692 {
693 int err;
694
695 err = gve_adminq_create_tx_queues(priv,
696 gve_xdp_tx_start_queue_id(priv),
697 priv->num_xdp_queues);
698 if (err) {
699 netif_err(priv, drv, priv->dev, "failed to create %d XDP tx queues\n",
700 priv->num_xdp_queues);
701 /* This failure will trigger a reset - no need to clean
702 * up
703 */
704 return err;
705 }
706 netif_dbg(priv, drv, priv->dev, "created %d XDP tx queues\n",
707 priv->num_xdp_queues);
708
709 return 0;
710 }
711
gve_create_rings(struct gve_priv * priv)712 static int gve_create_rings(struct gve_priv *priv)
713 {
714 int num_tx_queues = gve_num_tx_queues(priv);
715 int err;
716 int i;
717
718 err = gve_adminq_create_tx_queues(priv, 0, num_tx_queues);
719 if (err) {
720 netif_err(priv, drv, priv->dev, "failed to create %d tx queues\n",
721 num_tx_queues);
722 /* This failure will trigger a reset - no need to clean
723 * up
724 */
725 return err;
726 }
727 netif_dbg(priv, drv, priv->dev, "created %d tx queues\n",
728 num_tx_queues);
729
730 err = gve_adminq_create_rx_queues(priv, priv->rx_cfg.num_queues);
731 if (err) {
732 netif_err(priv, drv, priv->dev, "failed to create %d rx queues\n",
733 priv->rx_cfg.num_queues);
734 /* This failure will trigger a reset - no need to clean
735 * up
736 */
737 return err;
738 }
739 netif_dbg(priv, drv, priv->dev, "created %d rx queues\n",
740 priv->rx_cfg.num_queues);
741
742 if (gve_is_gqi(priv)) {
743 /* Rx data ring has been prefilled with packet buffers at queue
744 * allocation time.
745 *
746 * Write the doorbell to provide descriptor slots and packet
747 * buffers to the NIC.
748 */
749 for (i = 0; i < priv->rx_cfg.num_queues; i++)
750 gve_rx_write_doorbell(priv, &priv->rx[i]);
751 } else {
752 for (i = 0; i < priv->rx_cfg.num_queues; i++) {
753 /* Post buffers and ring doorbell. */
754 gve_rx_post_buffers_dqo(&priv->rx[i]);
755 }
756 }
757
758 return 0;
759 }
760
add_napi_init_xdp_sync_stats(struct gve_priv * priv,int (* napi_poll)(struct napi_struct * napi,int budget))761 static void add_napi_init_xdp_sync_stats(struct gve_priv *priv,
762 int (*napi_poll)(struct napi_struct *napi,
763 int budget))
764 {
765 int start_id = gve_xdp_tx_start_queue_id(priv);
766 int i;
767
768 /* Add xdp tx napi & init sync stats*/
769 for (i = start_id; i < start_id + priv->num_xdp_queues; i++) {
770 int ntfy_idx = gve_tx_idx_to_ntfy(priv, i);
771
772 u64_stats_init(&priv->tx[i].statss);
773 priv->tx[i].ntfy_id = ntfy_idx;
774 gve_add_napi(priv, ntfy_idx, napi_poll);
775 }
776 }
777
add_napi_init_sync_stats(struct gve_priv * priv,int (* napi_poll)(struct napi_struct * napi,int budget))778 static void add_napi_init_sync_stats(struct gve_priv *priv,
779 int (*napi_poll)(struct napi_struct *napi,
780 int budget))
781 {
782 int i;
783
784 /* Add tx napi & init sync stats*/
785 for (i = 0; i < gve_num_tx_queues(priv); i++) {
786 int ntfy_idx = gve_tx_idx_to_ntfy(priv, i);
787
788 u64_stats_init(&priv->tx[i].statss);
789 priv->tx[i].ntfy_id = ntfy_idx;
790 gve_add_napi(priv, ntfy_idx, napi_poll);
791 }
792 /* Add rx napi & init sync stats*/
793 for (i = 0; i < priv->rx_cfg.num_queues; i++) {
794 int ntfy_idx = gve_rx_idx_to_ntfy(priv, i);
795
796 u64_stats_init(&priv->rx[i].statss);
797 priv->rx[i].ntfy_id = ntfy_idx;
798 gve_add_napi(priv, ntfy_idx, napi_poll);
799 }
800 }
801
gve_tx_free_rings(struct gve_priv * priv,int start_id,int num_rings)802 static void gve_tx_free_rings(struct gve_priv *priv, int start_id, int num_rings)
803 {
804 if (gve_is_gqi(priv)) {
805 gve_tx_free_rings_gqi(priv, start_id, num_rings);
806 } else {
807 gve_tx_free_rings_dqo(priv);
808 }
809 }
810
gve_alloc_xdp_rings(struct gve_priv * priv)811 static int gve_alloc_xdp_rings(struct gve_priv *priv)
812 {
813 int start_id;
814 int err = 0;
815
816 if (!priv->num_xdp_queues)
817 return 0;
818
819 start_id = gve_xdp_tx_start_queue_id(priv);
820 err = gve_tx_alloc_rings(priv, start_id, priv->num_xdp_queues);
821 if (err)
822 return err;
823 add_napi_init_xdp_sync_stats(priv, gve_napi_poll);
824
825 return 0;
826 }
827
gve_alloc_rings(struct gve_priv * priv)828 static int gve_alloc_rings(struct gve_priv *priv)
829 {
830 int err;
831
832 /* Setup tx rings */
833 priv->tx = kvcalloc(priv->tx_cfg.max_queues, sizeof(*priv->tx),
834 GFP_KERNEL);
835 if (!priv->tx)
836 return -ENOMEM;
837
838 if (gve_is_gqi(priv))
839 err = gve_tx_alloc_rings(priv, 0, gve_num_tx_queues(priv));
840 else
841 err = gve_tx_alloc_rings_dqo(priv);
842 if (err)
843 goto free_tx;
844
845 /* Setup rx rings */
846 priv->rx = kvcalloc(priv->rx_cfg.max_queues, sizeof(*priv->rx),
847 GFP_KERNEL);
848 if (!priv->rx) {
849 err = -ENOMEM;
850 goto free_tx_queue;
851 }
852
853 if (gve_is_gqi(priv))
854 err = gve_rx_alloc_rings(priv);
855 else
856 err = gve_rx_alloc_rings_dqo(priv);
857 if (err)
858 goto free_rx;
859
860 if (gve_is_gqi(priv))
861 add_napi_init_sync_stats(priv, gve_napi_poll);
862 else
863 add_napi_init_sync_stats(priv, gve_napi_poll_dqo);
864
865 return 0;
866
867 free_rx:
868 kvfree(priv->rx);
869 priv->rx = NULL;
870 free_tx_queue:
871 gve_tx_free_rings(priv, 0, gve_num_tx_queues(priv));
872 free_tx:
873 kvfree(priv->tx);
874 priv->tx = NULL;
875 return err;
876 }
877
gve_destroy_xdp_rings(struct gve_priv * priv)878 static int gve_destroy_xdp_rings(struct gve_priv *priv)
879 {
880 int start_id;
881 int err;
882
883 start_id = gve_xdp_tx_start_queue_id(priv);
884 err = gve_adminq_destroy_tx_queues(priv,
885 start_id,
886 priv->num_xdp_queues);
887 if (err) {
888 netif_err(priv, drv, priv->dev,
889 "failed to destroy XDP queues\n");
890 /* This failure will trigger a reset - no need to clean up */
891 return err;
892 }
893 netif_dbg(priv, drv, priv->dev, "destroyed XDP queues\n");
894
895 return 0;
896 }
897
gve_destroy_rings(struct gve_priv * priv)898 static int gve_destroy_rings(struct gve_priv *priv)
899 {
900 int num_tx_queues = gve_num_tx_queues(priv);
901 int err;
902
903 err = gve_adminq_destroy_tx_queues(priv, 0, num_tx_queues);
904 if (err) {
905 netif_err(priv, drv, priv->dev,
906 "failed to destroy tx queues\n");
907 /* This failure will trigger a reset - no need to clean up */
908 return err;
909 }
910 netif_dbg(priv, drv, priv->dev, "destroyed tx queues\n");
911 err = gve_adminq_destroy_rx_queues(priv, priv->rx_cfg.num_queues);
912 if (err) {
913 netif_err(priv, drv, priv->dev,
914 "failed to destroy rx queues\n");
915 /* This failure will trigger a reset - no need to clean up */
916 return err;
917 }
918 netif_dbg(priv, drv, priv->dev, "destroyed rx queues\n");
919 return 0;
920 }
921
gve_rx_free_rings(struct gve_priv * priv)922 static void gve_rx_free_rings(struct gve_priv *priv)
923 {
924 if (gve_is_gqi(priv))
925 gve_rx_free_rings_gqi(priv);
926 else
927 gve_rx_free_rings_dqo(priv);
928 }
929
gve_free_xdp_rings(struct gve_priv * priv)930 static void gve_free_xdp_rings(struct gve_priv *priv)
931 {
932 int ntfy_idx, start_id;
933 int i;
934
935 start_id = gve_xdp_tx_start_queue_id(priv);
936 if (priv->tx) {
937 for (i = start_id; i < start_id + priv->num_xdp_queues; i++) {
938 ntfy_idx = gve_tx_idx_to_ntfy(priv, i);
939 gve_remove_napi(priv, ntfy_idx);
940 }
941 gve_tx_free_rings(priv, start_id, priv->num_xdp_queues);
942 }
943 }
944
gve_free_rings(struct gve_priv * priv)945 static void gve_free_rings(struct gve_priv *priv)
946 {
947 int num_tx_queues = gve_num_tx_queues(priv);
948 int ntfy_idx;
949 int i;
950
951 if (priv->tx) {
952 for (i = 0; i < num_tx_queues; i++) {
953 ntfy_idx = gve_tx_idx_to_ntfy(priv, i);
954 gve_remove_napi(priv, ntfy_idx);
955 }
956 gve_tx_free_rings(priv, 0, num_tx_queues);
957 kvfree(priv->tx);
958 priv->tx = NULL;
959 }
960 if (priv->rx) {
961 for (i = 0; i < priv->rx_cfg.num_queues; i++) {
962 ntfy_idx = gve_rx_idx_to_ntfy(priv, i);
963 gve_remove_napi(priv, ntfy_idx);
964 }
965 gve_rx_free_rings(priv);
966 kvfree(priv->rx);
967 priv->rx = NULL;
968 }
969 }
970
gve_alloc_page(struct gve_priv * priv,struct device * dev,struct page ** page,dma_addr_t * dma,enum dma_data_direction dir,gfp_t gfp_flags)971 int gve_alloc_page(struct gve_priv *priv, struct device *dev,
972 struct page **page, dma_addr_t *dma,
973 enum dma_data_direction dir, gfp_t gfp_flags)
974 {
975 *page = alloc_page(gfp_flags);
976 if (!*page) {
977 priv->page_alloc_fail++;
978 return -ENOMEM;
979 }
980 *dma = dma_map_page(dev, *page, 0, PAGE_SIZE, dir);
981 if (dma_mapping_error(dev, *dma)) {
982 priv->dma_mapping_error++;
983 put_page(*page);
984 return -ENOMEM;
985 }
986 return 0;
987 }
988
gve_alloc_queue_page_list(struct gve_priv * priv,u32 id,int pages)989 static int gve_alloc_queue_page_list(struct gve_priv *priv, u32 id,
990 int pages)
991 {
992 struct gve_queue_page_list *qpl = &priv->qpls[id];
993 int err;
994 int i;
995
996 if (pages + priv->num_registered_pages > priv->max_registered_pages) {
997 netif_err(priv, drv, priv->dev,
998 "Reached max number of registered pages %llu > %llu\n",
999 pages + priv->num_registered_pages,
1000 priv->max_registered_pages);
1001 return -EINVAL;
1002 }
1003
1004 qpl->id = id;
1005 qpl->num_entries = 0;
1006 qpl->pages = kvcalloc(pages, sizeof(*qpl->pages), GFP_KERNEL);
1007 /* caller handles clean up */
1008 if (!qpl->pages)
1009 return -ENOMEM;
1010 qpl->page_buses = kvcalloc(pages, sizeof(*qpl->page_buses), GFP_KERNEL);
1011 /* caller handles clean up */
1012 if (!qpl->page_buses)
1013 return -ENOMEM;
1014
1015 for (i = 0; i < pages; i++) {
1016 err = gve_alloc_page(priv, &priv->pdev->dev, &qpl->pages[i],
1017 &qpl->page_buses[i],
1018 gve_qpl_dma_dir(priv, id), GFP_KERNEL);
1019 /* caller handles clean up */
1020 if (err)
1021 return -ENOMEM;
1022 qpl->num_entries++;
1023 }
1024 priv->num_registered_pages += pages;
1025
1026 return 0;
1027 }
1028
gve_free_page(struct device * dev,struct page * page,dma_addr_t dma,enum dma_data_direction dir)1029 void gve_free_page(struct device *dev, struct page *page, dma_addr_t dma,
1030 enum dma_data_direction dir)
1031 {
1032 if (!dma_mapping_error(dev, dma))
1033 dma_unmap_page(dev, dma, PAGE_SIZE, dir);
1034 if (page)
1035 put_page(page);
1036 }
1037
gve_free_queue_page_list(struct gve_priv * priv,u32 id)1038 static void gve_free_queue_page_list(struct gve_priv *priv, u32 id)
1039 {
1040 struct gve_queue_page_list *qpl = &priv->qpls[id];
1041 int i;
1042
1043 if (!qpl->pages)
1044 return;
1045 if (!qpl->page_buses)
1046 goto free_pages;
1047
1048 for (i = 0; i < qpl->num_entries; i++)
1049 gve_free_page(&priv->pdev->dev, qpl->pages[i],
1050 qpl->page_buses[i], gve_qpl_dma_dir(priv, id));
1051
1052 kvfree(qpl->page_buses);
1053 qpl->page_buses = NULL;
1054 free_pages:
1055 kvfree(qpl->pages);
1056 qpl->pages = NULL;
1057 priv->num_registered_pages -= qpl->num_entries;
1058 }
1059
gve_alloc_xdp_qpls(struct gve_priv * priv)1060 static int gve_alloc_xdp_qpls(struct gve_priv *priv)
1061 {
1062 int start_id;
1063 int i, j;
1064 int err;
1065
1066 start_id = gve_tx_qpl_id(priv, gve_xdp_tx_start_queue_id(priv));
1067 for (i = start_id; i < start_id + gve_num_xdp_qpls(priv); i++) {
1068 err = gve_alloc_queue_page_list(priv, i,
1069 priv->tx_pages_per_qpl);
1070 if (err)
1071 goto free_qpls;
1072 }
1073
1074 return 0;
1075
1076 free_qpls:
1077 for (j = start_id; j <= i; j++)
1078 gve_free_queue_page_list(priv, j);
1079 return err;
1080 }
1081
gve_alloc_qpls(struct gve_priv * priv)1082 static int gve_alloc_qpls(struct gve_priv *priv)
1083 {
1084 int max_queues = priv->tx_cfg.max_queues + priv->rx_cfg.max_queues;
1085 int page_count;
1086 int start_id;
1087 int i, j;
1088 int err;
1089
1090 if (!gve_is_qpl(priv))
1091 return 0;
1092
1093 priv->qpls = kvcalloc(max_queues, sizeof(*priv->qpls), GFP_KERNEL);
1094 if (!priv->qpls)
1095 return -ENOMEM;
1096
1097 start_id = gve_tx_start_qpl_id(priv);
1098 page_count = priv->tx_pages_per_qpl;
1099 for (i = start_id; i < start_id + gve_num_tx_qpls(priv); i++) {
1100 err = gve_alloc_queue_page_list(priv, i,
1101 page_count);
1102 if (err)
1103 goto free_qpls;
1104 }
1105
1106 start_id = gve_rx_start_qpl_id(priv);
1107
1108 /* For GQI_QPL number of pages allocated have 1:1 relationship with
1109 * number of descriptors. For DQO, number of pages required are
1110 * more than descriptors (because of out of order completions).
1111 */
1112 page_count = priv->queue_format == GVE_GQI_QPL_FORMAT ?
1113 priv->rx_data_slot_cnt : priv->rx_pages_per_qpl;
1114 for (i = start_id; i < start_id + gve_num_rx_qpls(priv); i++) {
1115 err = gve_alloc_queue_page_list(priv, i,
1116 page_count);
1117 if (err)
1118 goto free_qpls;
1119 }
1120
1121 priv->qpl_cfg.qpl_map_size = BITS_TO_LONGS(max_queues) *
1122 sizeof(unsigned long) * BITS_PER_BYTE;
1123 priv->qpl_cfg.qpl_id_map = kvcalloc(BITS_TO_LONGS(max_queues),
1124 sizeof(unsigned long), GFP_KERNEL);
1125 if (!priv->qpl_cfg.qpl_id_map) {
1126 err = -ENOMEM;
1127 goto free_qpls;
1128 }
1129
1130 return 0;
1131
1132 free_qpls:
1133 for (j = 0; j <= i; j++)
1134 gve_free_queue_page_list(priv, j);
1135 kvfree(priv->qpls);
1136 priv->qpls = NULL;
1137 return err;
1138 }
1139
gve_free_xdp_qpls(struct gve_priv * priv)1140 static void gve_free_xdp_qpls(struct gve_priv *priv)
1141 {
1142 int start_id;
1143 int i;
1144
1145 start_id = gve_tx_qpl_id(priv, gve_xdp_tx_start_queue_id(priv));
1146 for (i = start_id; i < start_id + gve_num_xdp_qpls(priv); i++)
1147 gve_free_queue_page_list(priv, i);
1148 }
1149
gve_free_qpls(struct gve_priv * priv)1150 static void gve_free_qpls(struct gve_priv *priv)
1151 {
1152 int max_queues = priv->tx_cfg.max_queues + priv->rx_cfg.max_queues;
1153 int i;
1154
1155 if (!priv->qpls)
1156 return;
1157
1158 kvfree(priv->qpl_cfg.qpl_id_map);
1159 priv->qpl_cfg.qpl_id_map = NULL;
1160
1161 for (i = 0; i < max_queues; i++)
1162 gve_free_queue_page_list(priv, i);
1163
1164 kvfree(priv->qpls);
1165 priv->qpls = NULL;
1166 }
1167
1168 /* Use this to schedule a reset when the device is capable of continuing
1169 * to handle other requests in its current state. If it is not, do a reset
1170 * in thread instead.
1171 */
gve_schedule_reset(struct gve_priv * priv)1172 void gve_schedule_reset(struct gve_priv *priv)
1173 {
1174 gve_set_do_reset(priv);
1175 queue_work(priv->gve_wq, &priv->service_task);
1176 }
1177
1178 static void gve_reset_and_teardown(struct gve_priv *priv, bool was_up);
1179 static int gve_reset_recovery(struct gve_priv *priv, bool was_up);
1180 static void gve_turndown(struct gve_priv *priv);
1181 static void gve_turnup(struct gve_priv *priv);
1182
gve_reg_xdp_info(struct gve_priv * priv,struct net_device * dev)1183 static int gve_reg_xdp_info(struct gve_priv *priv, struct net_device *dev)
1184 {
1185 struct napi_struct *napi;
1186 struct gve_rx_ring *rx;
1187 int err = 0;
1188 int i, j;
1189 u32 tx_qid;
1190
1191 if (!priv->num_xdp_queues)
1192 return 0;
1193
1194 for (i = 0; i < priv->rx_cfg.num_queues; i++) {
1195 rx = &priv->rx[i];
1196 napi = &priv->ntfy_blocks[rx->ntfy_id].napi;
1197
1198 err = xdp_rxq_info_reg(&rx->xdp_rxq, dev, i,
1199 napi->napi_id);
1200 if (err)
1201 goto err;
1202 err = xdp_rxq_info_reg_mem_model(&rx->xdp_rxq,
1203 MEM_TYPE_PAGE_SHARED, NULL);
1204 if (err)
1205 goto err;
1206 rx->xsk_pool = xsk_get_pool_from_qid(dev, i);
1207 if (rx->xsk_pool) {
1208 err = xdp_rxq_info_reg(&rx->xsk_rxq, dev, i,
1209 napi->napi_id);
1210 if (err)
1211 goto err;
1212 err = xdp_rxq_info_reg_mem_model(&rx->xsk_rxq,
1213 MEM_TYPE_XSK_BUFF_POOL, NULL);
1214 if (err)
1215 goto err;
1216 xsk_pool_set_rxq_info(rx->xsk_pool,
1217 &rx->xsk_rxq);
1218 }
1219 }
1220
1221 for (i = 0; i < priv->num_xdp_queues; i++) {
1222 tx_qid = gve_xdp_tx_queue_id(priv, i);
1223 priv->tx[tx_qid].xsk_pool = xsk_get_pool_from_qid(dev, i);
1224 }
1225 return 0;
1226
1227 err:
1228 for (j = i; j >= 0; j--) {
1229 rx = &priv->rx[j];
1230 if (xdp_rxq_info_is_reg(&rx->xdp_rxq))
1231 xdp_rxq_info_unreg(&rx->xdp_rxq);
1232 if (xdp_rxq_info_is_reg(&rx->xsk_rxq))
1233 xdp_rxq_info_unreg(&rx->xsk_rxq);
1234 }
1235 return err;
1236 }
1237
gve_unreg_xdp_info(struct gve_priv * priv)1238 static void gve_unreg_xdp_info(struct gve_priv *priv)
1239 {
1240 int i, tx_qid;
1241
1242 if (!priv->num_xdp_queues)
1243 return;
1244
1245 for (i = 0; i < priv->rx_cfg.num_queues; i++) {
1246 struct gve_rx_ring *rx = &priv->rx[i];
1247
1248 xdp_rxq_info_unreg(&rx->xdp_rxq);
1249 if (rx->xsk_pool) {
1250 xdp_rxq_info_unreg(&rx->xsk_rxq);
1251 rx->xsk_pool = NULL;
1252 }
1253 }
1254
1255 for (i = 0; i < priv->num_xdp_queues; i++) {
1256 tx_qid = gve_xdp_tx_queue_id(priv, i);
1257 priv->tx[tx_qid].xsk_pool = NULL;
1258 }
1259 }
1260
gve_drain_page_cache(struct gve_priv * priv)1261 static void gve_drain_page_cache(struct gve_priv *priv)
1262 {
1263 struct page_frag_cache *nc;
1264 int i;
1265
1266 for (i = 0; i < priv->rx_cfg.num_queues; i++) {
1267 nc = &priv->rx[i].page_cache;
1268 if (nc->va) {
1269 __page_frag_cache_drain(virt_to_page(nc->va),
1270 nc->pagecnt_bias);
1271 nc->va = NULL;
1272 }
1273 }
1274 }
1275
gve_open(struct net_device * dev)1276 static int gve_open(struct net_device *dev)
1277 {
1278 struct gve_priv *priv = netdev_priv(dev);
1279 int err;
1280
1281 if (priv->xdp_prog)
1282 priv->num_xdp_queues = priv->rx_cfg.num_queues;
1283 else
1284 priv->num_xdp_queues = 0;
1285
1286 err = gve_alloc_qpls(priv);
1287 if (err)
1288 return err;
1289
1290 err = gve_alloc_rings(priv);
1291 if (err)
1292 goto free_qpls;
1293
1294 err = netif_set_real_num_tx_queues(dev, priv->tx_cfg.num_queues);
1295 if (err)
1296 goto free_rings;
1297 err = netif_set_real_num_rx_queues(dev, priv->rx_cfg.num_queues);
1298 if (err)
1299 goto free_rings;
1300
1301 err = gve_reg_xdp_info(priv, dev);
1302 if (err)
1303 goto free_rings;
1304
1305 err = gve_register_qpls(priv);
1306 if (err)
1307 goto reset;
1308
1309 if (!gve_is_gqi(priv)) {
1310 /* Hard code this for now. This may be tuned in the future for
1311 * performance.
1312 */
1313 priv->data_buffer_size_dqo = GVE_RX_BUFFER_SIZE_DQO;
1314 }
1315 err = gve_create_rings(priv);
1316 if (err)
1317 goto reset;
1318
1319 gve_set_device_rings_ok(priv);
1320
1321 if (gve_get_report_stats(priv))
1322 mod_timer(&priv->stats_report_timer,
1323 round_jiffies(jiffies +
1324 msecs_to_jiffies(priv->stats_report_timer_period)));
1325
1326 gve_turnup(priv);
1327 queue_work(priv->gve_wq, &priv->service_task);
1328 priv->interface_up_cnt++;
1329 return 0;
1330
1331 free_rings:
1332 gve_free_rings(priv);
1333 free_qpls:
1334 gve_free_qpls(priv);
1335 return err;
1336
1337 reset:
1338 /* This must have been called from a reset due to the rtnl lock
1339 * so just return at this point.
1340 */
1341 if (gve_get_reset_in_progress(priv))
1342 return err;
1343 /* Otherwise reset before returning */
1344 gve_reset_and_teardown(priv, true);
1345 /* if this fails there is nothing we can do so just ignore the return */
1346 gve_reset_recovery(priv, false);
1347 /* return the original error */
1348 return err;
1349 }
1350
gve_close(struct net_device * dev)1351 static int gve_close(struct net_device *dev)
1352 {
1353 struct gve_priv *priv = netdev_priv(dev);
1354 int err;
1355
1356 netif_carrier_off(dev);
1357 if (gve_get_device_rings_ok(priv)) {
1358 gve_turndown(priv);
1359 gve_drain_page_cache(priv);
1360 err = gve_destroy_rings(priv);
1361 if (err)
1362 goto err;
1363 err = gve_unregister_qpls(priv);
1364 if (err)
1365 goto err;
1366 gve_clear_device_rings_ok(priv);
1367 }
1368 del_timer_sync(&priv->stats_report_timer);
1369
1370 gve_unreg_xdp_info(priv);
1371 gve_free_rings(priv);
1372 gve_free_qpls(priv);
1373 priv->interface_down_cnt++;
1374 return 0;
1375
1376 err:
1377 /* This must have been called from a reset due to the rtnl lock
1378 * so just return at this point.
1379 */
1380 if (gve_get_reset_in_progress(priv))
1381 return err;
1382 /* Otherwise reset before returning */
1383 gve_reset_and_teardown(priv, true);
1384 return gve_reset_recovery(priv, false);
1385 }
1386
gve_remove_xdp_queues(struct gve_priv * priv)1387 static int gve_remove_xdp_queues(struct gve_priv *priv)
1388 {
1389 int err;
1390
1391 err = gve_destroy_xdp_rings(priv);
1392 if (err)
1393 return err;
1394
1395 err = gve_unregister_xdp_qpls(priv);
1396 if (err)
1397 return err;
1398
1399 gve_unreg_xdp_info(priv);
1400 gve_free_xdp_rings(priv);
1401 gve_free_xdp_qpls(priv);
1402 priv->num_xdp_queues = 0;
1403 return 0;
1404 }
1405
gve_add_xdp_queues(struct gve_priv * priv)1406 static int gve_add_xdp_queues(struct gve_priv *priv)
1407 {
1408 int err;
1409
1410 priv->num_xdp_queues = priv->tx_cfg.num_queues;
1411
1412 err = gve_alloc_xdp_qpls(priv);
1413 if (err)
1414 goto err;
1415
1416 err = gve_alloc_xdp_rings(priv);
1417 if (err)
1418 goto free_xdp_qpls;
1419
1420 err = gve_reg_xdp_info(priv, priv->dev);
1421 if (err)
1422 goto free_xdp_rings;
1423
1424 err = gve_register_xdp_qpls(priv);
1425 if (err)
1426 goto free_xdp_rings;
1427
1428 err = gve_create_xdp_rings(priv);
1429 if (err)
1430 goto free_xdp_rings;
1431
1432 return 0;
1433
1434 free_xdp_rings:
1435 gve_free_xdp_rings(priv);
1436 free_xdp_qpls:
1437 gve_free_xdp_qpls(priv);
1438 err:
1439 priv->num_xdp_queues = 0;
1440 return err;
1441 }
1442
gve_handle_link_status(struct gve_priv * priv,bool link_status)1443 static void gve_handle_link_status(struct gve_priv *priv, bool link_status)
1444 {
1445 if (!gve_get_napi_enabled(priv))
1446 return;
1447
1448 if (link_status == netif_carrier_ok(priv->dev))
1449 return;
1450
1451 if (link_status) {
1452 netdev_info(priv->dev, "Device link is up.\n");
1453 netif_carrier_on(priv->dev);
1454 } else {
1455 netdev_info(priv->dev, "Device link is down.\n");
1456 netif_carrier_off(priv->dev);
1457 }
1458 }
1459
gve_set_xdp(struct gve_priv * priv,struct bpf_prog * prog,struct netlink_ext_ack * extack)1460 static int gve_set_xdp(struct gve_priv *priv, struct bpf_prog *prog,
1461 struct netlink_ext_ack *extack)
1462 {
1463 struct bpf_prog *old_prog;
1464 int err = 0;
1465 u32 status;
1466
1467 old_prog = READ_ONCE(priv->xdp_prog);
1468 if (!netif_carrier_ok(priv->dev)) {
1469 WRITE_ONCE(priv->xdp_prog, prog);
1470 if (old_prog)
1471 bpf_prog_put(old_prog);
1472 return 0;
1473 }
1474
1475 gve_turndown(priv);
1476 if (!old_prog && prog) {
1477 // Allocate XDP TX queues if an XDP program is
1478 // being installed
1479 err = gve_add_xdp_queues(priv);
1480 if (err)
1481 goto out;
1482 } else if (old_prog && !prog) {
1483 // Remove XDP TX queues if an XDP program is
1484 // being uninstalled
1485 err = gve_remove_xdp_queues(priv);
1486 if (err)
1487 goto out;
1488 }
1489 WRITE_ONCE(priv->xdp_prog, prog);
1490 if (old_prog)
1491 bpf_prog_put(old_prog);
1492
1493 out:
1494 gve_turnup(priv);
1495 status = ioread32be(&priv->reg_bar0->device_status);
1496 gve_handle_link_status(priv, GVE_DEVICE_STATUS_LINK_STATUS_MASK & status);
1497 return err;
1498 }
1499
gve_xsk_pool_enable(struct net_device * dev,struct xsk_buff_pool * pool,u16 qid)1500 static int gve_xsk_pool_enable(struct net_device *dev,
1501 struct xsk_buff_pool *pool,
1502 u16 qid)
1503 {
1504 struct gve_priv *priv = netdev_priv(dev);
1505 struct napi_struct *napi;
1506 struct gve_rx_ring *rx;
1507 int tx_qid;
1508 int err;
1509
1510 if (qid >= priv->rx_cfg.num_queues) {
1511 dev_err(&priv->pdev->dev, "xsk pool invalid qid %d", qid);
1512 return -EINVAL;
1513 }
1514 if (xsk_pool_get_rx_frame_size(pool) <
1515 priv->dev->max_mtu + sizeof(struct ethhdr)) {
1516 dev_err(&priv->pdev->dev, "xsk pool frame_len too small");
1517 return -EINVAL;
1518 }
1519
1520 err = xsk_pool_dma_map(pool, &priv->pdev->dev,
1521 DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING);
1522 if (err)
1523 return err;
1524
1525 /* If XDP prog is not installed, return */
1526 if (!priv->xdp_prog)
1527 return 0;
1528
1529 rx = &priv->rx[qid];
1530 napi = &priv->ntfy_blocks[rx->ntfy_id].napi;
1531 err = xdp_rxq_info_reg(&rx->xsk_rxq, dev, qid, napi->napi_id);
1532 if (err)
1533 goto err;
1534
1535 err = xdp_rxq_info_reg_mem_model(&rx->xsk_rxq,
1536 MEM_TYPE_XSK_BUFF_POOL, NULL);
1537 if (err)
1538 goto err;
1539
1540 xsk_pool_set_rxq_info(pool, &rx->xsk_rxq);
1541 rx->xsk_pool = pool;
1542
1543 tx_qid = gve_xdp_tx_queue_id(priv, qid);
1544 priv->tx[tx_qid].xsk_pool = pool;
1545
1546 return 0;
1547 err:
1548 if (xdp_rxq_info_is_reg(&rx->xsk_rxq))
1549 xdp_rxq_info_unreg(&rx->xsk_rxq);
1550
1551 xsk_pool_dma_unmap(pool,
1552 DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING);
1553 return err;
1554 }
1555
gve_xsk_pool_disable(struct net_device * dev,u16 qid)1556 static int gve_xsk_pool_disable(struct net_device *dev,
1557 u16 qid)
1558 {
1559 struct gve_priv *priv = netdev_priv(dev);
1560 struct napi_struct *napi_rx;
1561 struct napi_struct *napi_tx;
1562 struct xsk_buff_pool *pool;
1563 int tx_qid;
1564
1565 pool = xsk_get_pool_from_qid(dev, qid);
1566 if (!pool)
1567 return -EINVAL;
1568 if (qid >= priv->rx_cfg.num_queues)
1569 return -EINVAL;
1570
1571 /* If XDP prog is not installed, unmap DMA and return */
1572 if (!priv->xdp_prog)
1573 goto done;
1574
1575 tx_qid = gve_xdp_tx_queue_id(priv, qid);
1576 if (!netif_running(dev)) {
1577 priv->rx[qid].xsk_pool = NULL;
1578 xdp_rxq_info_unreg(&priv->rx[qid].xsk_rxq);
1579 priv->tx[tx_qid].xsk_pool = NULL;
1580 goto done;
1581 }
1582
1583 napi_rx = &priv->ntfy_blocks[priv->rx[qid].ntfy_id].napi;
1584 napi_disable(napi_rx); /* make sure current rx poll is done */
1585
1586 napi_tx = &priv->ntfy_blocks[priv->tx[tx_qid].ntfy_id].napi;
1587 napi_disable(napi_tx); /* make sure current tx poll is done */
1588
1589 priv->rx[qid].xsk_pool = NULL;
1590 xdp_rxq_info_unreg(&priv->rx[qid].xsk_rxq);
1591 priv->tx[tx_qid].xsk_pool = NULL;
1592 smp_mb(); /* Make sure it is visible to the workers on datapath */
1593
1594 napi_enable(napi_rx);
1595 if (gve_rx_work_pending(&priv->rx[qid]))
1596 napi_schedule(napi_rx);
1597
1598 napi_enable(napi_tx);
1599 if (gve_tx_clean_pending(priv, &priv->tx[tx_qid]))
1600 napi_schedule(napi_tx);
1601
1602 done:
1603 xsk_pool_dma_unmap(pool,
1604 DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING);
1605 return 0;
1606 }
1607
gve_xsk_wakeup(struct net_device * dev,u32 queue_id,u32 flags)1608 static int gve_xsk_wakeup(struct net_device *dev, u32 queue_id, u32 flags)
1609 {
1610 struct gve_priv *priv = netdev_priv(dev);
1611 int tx_queue_id = gve_xdp_tx_queue_id(priv, queue_id);
1612
1613 if (queue_id >= priv->rx_cfg.num_queues || !priv->xdp_prog)
1614 return -EINVAL;
1615
1616 if (flags & XDP_WAKEUP_TX) {
1617 struct gve_tx_ring *tx = &priv->tx[tx_queue_id];
1618 struct napi_struct *napi =
1619 &priv->ntfy_blocks[tx->ntfy_id].napi;
1620
1621 if (!napi_if_scheduled_mark_missed(napi)) {
1622 /* Call local_bh_enable to trigger SoftIRQ processing */
1623 local_bh_disable();
1624 napi_schedule(napi);
1625 local_bh_enable();
1626 }
1627
1628 tx->xdp_xsk_wakeup++;
1629 }
1630
1631 return 0;
1632 }
1633
verify_xdp_configuration(struct net_device * dev)1634 static int verify_xdp_configuration(struct net_device *dev)
1635 {
1636 struct gve_priv *priv = netdev_priv(dev);
1637
1638 if (dev->features & NETIF_F_LRO) {
1639 netdev_warn(dev, "XDP is not supported when LRO is on.\n");
1640 return -EOPNOTSUPP;
1641 }
1642
1643 if (priv->queue_format != GVE_GQI_QPL_FORMAT) {
1644 netdev_warn(dev, "XDP is not supported in mode %d.\n",
1645 priv->queue_format);
1646 return -EOPNOTSUPP;
1647 }
1648
1649 if (dev->mtu > (PAGE_SIZE / 2) - sizeof(struct ethhdr) - GVE_RX_PAD) {
1650 netdev_warn(dev, "XDP is not supported for mtu %d.\n",
1651 dev->mtu);
1652 return -EOPNOTSUPP;
1653 }
1654
1655 if (priv->rx_cfg.num_queues != priv->tx_cfg.num_queues ||
1656 (2 * priv->tx_cfg.num_queues > priv->tx_cfg.max_queues)) {
1657 netdev_warn(dev, "XDP load failed: The number of configured RX queues %d should be equal to the number of configured TX queues %d and the number of configured RX/TX queues should be less than or equal to half the maximum number of RX/TX queues %d",
1658 priv->rx_cfg.num_queues,
1659 priv->tx_cfg.num_queues,
1660 priv->tx_cfg.max_queues);
1661 return -EINVAL;
1662 }
1663 return 0;
1664 }
1665
gve_xdp(struct net_device * dev,struct netdev_bpf * xdp)1666 static int gve_xdp(struct net_device *dev, struct netdev_bpf *xdp)
1667 {
1668 struct gve_priv *priv = netdev_priv(dev);
1669 int err;
1670
1671 err = verify_xdp_configuration(dev);
1672 if (err)
1673 return err;
1674 switch (xdp->command) {
1675 case XDP_SETUP_PROG:
1676 return gve_set_xdp(priv, xdp->prog, xdp->extack);
1677 case XDP_SETUP_XSK_POOL:
1678 if (xdp->xsk.pool)
1679 return gve_xsk_pool_enable(dev, xdp->xsk.pool, xdp->xsk.queue_id);
1680 else
1681 return gve_xsk_pool_disable(dev, xdp->xsk.queue_id);
1682 default:
1683 return -EINVAL;
1684 }
1685 }
1686
gve_adjust_queues(struct gve_priv * priv,struct gve_queue_config new_rx_config,struct gve_queue_config new_tx_config)1687 int gve_adjust_queues(struct gve_priv *priv,
1688 struct gve_queue_config new_rx_config,
1689 struct gve_queue_config new_tx_config)
1690 {
1691 int err;
1692
1693 if (netif_carrier_ok(priv->dev)) {
1694 /* To make this process as simple as possible we teardown the
1695 * device, set the new configuration, and then bring the device
1696 * up again.
1697 */
1698 err = gve_close(priv->dev);
1699 /* we have already tried to reset in close,
1700 * just fail at this point
1701 */
1702 if (err)
1703 return err;
1704 priv->tx_cfg = new_tx_config;
1705 priv->rx_cfg = new_rx_config;
1706
1707 err = gve_open(priv->dev);
1708 if (err)
1709 goto err;
1710
1711 return 0;
1712 }
1713 /* Set the config for the next up. */
1714 priv->tx_cfg = new_tx_config;
1715 priv->rx_cfg = new_rx_config;
1716
1717 return 0;
1718 err:
1719 netif_err(priv, drv, priv->dev,
1720 "Adjust queues failed! !!! DISABLING ALL QUEUES !!!\n");
1721 gve_turndown(priv);
1722 return err;
1723 }
1724
gve_turndown(struct gve_priv * priv)1725 static void gve_turndown(struct gve_priv *priv)
1726 {
1727 int idx;
1728
1729 if (netif_carrier_ok(priv->dev))
1730 netif_carrier_off(priv->dev);
1731
1732 if (!gve_get_napi_enabled(priv))
1733 return;
1734
1735 /* Disable napi to prevent more work from coming in */
1736 for (idx = 0; idx < gve_num_tx_queues(priv); idx++) {
1737 int ntfy_idx = gve_tx_idx_to_ntfy(priv, idx);
1738 struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
1739
1740 napi_disable(&block->napi);
1741 }
1742 for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) {
1743 int ntfy_idx = gve_rx_idx_to_ntfy(priv, idx);
1744 struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
1745
1746 napi_disable(&block->napi);
1747 }
1748
1749 /* Stop tx queues */
1750 netif_tx_disable(priv->dev);
1751
1752 gve_clear_napi_enabled(priv);
1753 gve_clear_report_stats(priv);
1754 }
1755
gve_turnup(struct gve_priv * priv)1756 static void gve_turnup(struct gve_priv *priv)
1757 {
1758 int idx;
1759
1760 /* Start the tx queues */
1761 netif_tx_start_all_queues(priv->dev);
1762
1763 /* Enable napi and unmask interrupts for all queues */
1764 for (idx = 0; idx < gve_num_tx_queues(priv); idx++) {
1765 int ntfy_idx = gve_tx_idx_to_ntfy(priv, idx);
1766 struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
1767
1768 napi_enable(&block->napi);
1769 if (gve_is_gqi(priv)) {
1770 iowrite32be(0, gve_irq_doorbell(priv, block));
1771 } else {
1772 gve_set_itr_coalesce_usecs_dqo(priv, block,
1773 priv->tx_coalesce_usecs);
1774 }
1775 }
1776 for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) {
1777 int ntfy_idx = gve_rx_idx_to_ntfy(priv, idx);
1778 struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
1779
1780 napi_enable(&block->napi);
1781 if (gve_is_gqi(priv)) {
1782 iowrite32be(0, gve_irq_doorbell(priv, block));
1783 } else {
1784 gve_set_itr_coalesce_usecs_dqo(priv, block,
1785 priv->rx_coalesce_usecs);
1786 }
1787 }
1788
1789 gve_set_napi_enabled(priv);
1790 }
1791
gve_tx_timeout(struct net_device * dev,unsigned int txqueue)1792 static void gve_tx_timeout(struct net_device *dev, unsigned int txqueue)
1793 {
1794 struct gve_notify_block *block;
1795 struct gve_tx_ring *tx = NULL;
1796 struct gve_priv *priv;
1797 u32 last_nic_done;
1798 u32 current_time;
1799 u32 ntfy_idx;
1800
1801 netdev_info(dev, "Timeout on tx queue, %d", txqueue);
1802 priv = netdev_priv(dev);
1803 if (txqueue > priv->tx_cfg.num_queues)
1804 goto reset;
1805
1806 ntfy_idx = gve_tx_idx_to_ntfy(priv, txqueue);
1807 if (ntfy_idx >= priv->num_ntfy_blks)
1808 goto reset;
1809
1810 block = &priv->ntfy_blocks[ntfy_idx];
1811 tx = block->tx;
1812
1813 current_time = jiffies_to_msecs(jiffies);
1814 if (tx->last_kick_msec + MIN_TX_TIMEOUT_GAP > current_time)
1815 goto reset;
1816
1817 /* Check to see if there are missed completions, which will allow us to
1818 * kick the queue.
1819 */
1820 last_nic_done = gve_tx_load_event_counter(priv, tx);
1821 if (last_nic_done - tx->done) {
1822 netdev_info(dev, "Kicking queue %d", txqueue);
1823 iowrite32be(GVE_IRQ_MASK, gve_irq_doorbell(priv, block));
1824 napi_schedule(&block->napi);
1825 tx->last_kick_msec = current_time;
1826 goto out;
1827 } // Else reset.
1828
1829 reset:
1830 gve_schedule_reset(priv);
1831
1832 out:
1833 if (tx)
1834 tx->queue_timeout++;
1835 priv->tx_timeo_cnt++;
1836 }
1837
gve_set_features(struct net_device * netdev,netdev_features_t features)1838 static int gve_set_features(struct net_device *netdev,
1839 netdev_features_t features)
1840 {
1841 const netdev_features_t orig_features = netdev->features;
1842 struct gve_priv *priv = netdev_priv(netdev);
1843 int err;
1844
1845 if ((netdev->features & NETIF_F_LRO) != (features & NETIF_F_LRO)) {
1846 netdev->features ^= NETIF_F_LRO;
1847 if (netif_carrier_ok(netdev)) {
1848 /* To make this process as simple as possible we
1849 * teardown the device, set the new configuration,
1850 * and then bring the device up again.
1851 */
1852 err = gve_close(netdev);
1853 /* We have already tried to reset in close, just fail
1854 * at this point.
1855 */
1856 if (err)
1857 goto err;
1858
1859 err = gve_open(netdev);
1860 if (err)
1861 goto err;
1862 }
1863 }
1864
1865 return 0;
1866 err:
1867 /* Reverts the change on error. */
1868 netdev->features = orig_features;
1869 netif_err(priv, drv, netdev,
1870 "Set features failed! !!! DISABLING ALL QUEUES !!!\n");
1871 return err;
1872 }
1873
1874 static const struct net_device_ops gve_netdev_ops = {
1875 .ndo_start_xmit = gve_start_xmit,
1876 .ndo_open = gve_open,
1877 .ndo_stop = gve_close,
1878 .ndo_get_stats64 = gve_get_stats,
1879 .ndo_tx_timeout = gve_tx_timeout,
1880 .ndo_set_features = gve_set_features,
1881 .ndo_bpf = gve_xdp,
1882 .ndo_xdp_xmit = gve_xdp_xmit,
1883 .ndo_xsk_wakeup = gve_xsk_wakeup,
1884 };
1885
gve_handle_status(struct gve_priv * priv,u32 status)1886 static void gve_handle_status(struct gve_priv *priv, u32 status)
1887 {
1888 if (GVE_DEVICE_STATUS_RESET_MASK & status) {
1889 dev_info(&priv->pdev->dev, "Device requested reset.\n");
1890 gve_set_do_reset(priv);
1891 }
1892 if (GVE_DEVICE_STATUS_REPORT_STATS_MASK & status) {
1893 priv->stats_report_trigger_cnt++;
1894 gve_set_do_report_stats(priv);
1895 }
1896 }
1897
gve_handle_reset(struct gve_priv * priv)1898 static void gve_handle_reset(struct gve_priv *priv)
1899 {
1900 /* A service task will be scheduled at the end of probe to catch any
1901 * resets that need to happen, and we don't want to reset until
1902 * probe is done.
1903 */
1904 if (gve_get_probe_in_progress(priv))
1905 return;
1906
1907 if (gve_get_do_reset(priv)) {
1908 rtnl_lock();
1909 gve_reset(priv, false);
1910 rtnl_unlock();
1911 }
1912 }
1913
gve_handle_report_stats(struct gve_priv * priv)1914 void gve_handle_report_stats(struct gve_priv *priv)
1915 {
1916 struct stats *stats = priv->stats_report->stats;
1917 int idx, stats_idx = 0;
1918 unsigned int start = 0;
1919 u64 tx_bytes;
1920
1921 if (!gve_get_report_stats(priv))
1922 return;
1923
1924 be64_add_cpu(&priv->stats_report->written_count, 1);
1925 /* tx stats */
1926 if (priv->tx) {
1927 for (idx = 0; idx < gve_num_tx_queues(priv); idx++) {
1928 u32 last_completion = 0;
1929 u32 tx_frames = 0;
1930
1931 /* DQO doesn't currently support these metrics. */
1932 if (gve_is_gqi(priv)) {
1933 last_completion = priv->tx[idx].done;
1934 tx_frames = priv->tx[idx].req;
1935 }
1936
1937 do {
1938 start = u64_stats_fetch_begin(&priv->tx[idx].statss);
1939 tx_bytes = priv->tx[idx].bytes_done;
1940 } while (u64_stats_fetch_retry(&priv->tx[idx].statss, start));
1941 stats[stats_idx++] = (struct stats) {
1942 .stat_name = cpu_to_be32(TX_WAKE_CNT),
1943 .value = cpu_to_be64(priv->tx[idx].wake_queue),
1944 .queue_id = cpu_to_be32(idx),
1945 };
1946 stats[stats_idx++] = (struct stats) {
1947 .stat_name = cpu_to_be32(TX_STOP_CNT),
1948 .value = cpu_to_be64(priv->tx[idx].stop_queue),
1949 .queue_id = cpu_to_be32(idx),
1950 };
1951 stats[stats_idx++] = (struct stats) {
1952 .stat_name = cpu_to_be32(TX_FRAMES_SENT),
1953 .value = cpu_to_be64(tx_frames),
1954 .queue_id = cpu_to_be32(idx),
1955 };
1956 stats[stats_idx++] = (struct stats) {
1957 .stat_name = cpu_to_be32(TX_BYTES_SENT),
1958 .value = cpu_to_be64(tx_bytes),
1959 .queue_id = cpu_to_be32(idx),
1960 };
1961 stats[stats_idx++] = (struct stats) {
1962 .stat_name = cpu_to_be32(TX_LAST_COMPLETION_PROCESSED),
1963 .value = cpu_to_be64(last_completion),
1964 .queue_id = cpu_to_be32(idx),
1965 };
1966 stats[stats_idx++] = (struct stats) {
1967 .stat_name = cpu_to_be32(TX_TIMEOUT_CNT),
1968 .value = cpu_to_be64(priv->tx[idx].queue_timeout),
1969 .queue_id = cpu_to_be32(idx),
1970 };
1971 }
1972 }
1973 /* rx stats */
1974 if (priv->rx) {
1975 for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) {
1976 stats[stats_idx++] = (struct stats) {
1977 .stat_name = cpu_to_be32(RX_NEXT_EXPECTED_SEQUENCE),
1978 .value = cpu_to_be64(priv->rx[idx].desc.seqno),
1979 .queue_id = cpu_to_be32(idx),
1980 };
1981 stats[stats_idx++] = (struct stats) {
1982 .stat_name = cpu_to_be32(RX_BUFFERS_POSTED),
1983 .value = cpu_to_be64(priv->rx[0].fill_cnt),
1984 .queue_id = cpu_to_be32(idx),
1985 };
1986 }
1987 }
1988 }
1989
1990 /* Handle NIC status register changes, reset requests and report stats */
gve_service_task(struct work_struct * work)1991 static void gve_service_task(struct work_struct *work)
1992 {
1993 struct gve_priv *priv = container_of(work, struct gve_priv,
1994 service_task);
1995 u32 status = ioread32be(&priv->reg_bar0->device_status);
1996
1997 gve_handle_status(priv, status);
1998
1999 gve_handle_reset(priv);
2000 gve_handle_link_status(priv, GVE_DEVICE_STATUS_LINK_STATUS_MASK & status);
2001 }
2002
gve_set_netdev_xdp_features(struct gve_priv * priv)2003 static void gve_set_netdev_xdp_features(struct gve_priv *priv)
2004 {
2005 if (priv->queue_format == GVE_GQI_QPL_FORMAT) {
2006 priv->dev->xdp_features = NETDEV_XDP_ACT_BASIC;
2007 priv->dev->xdp_features |= NETDEV_XDP_ACT_REDIRECT;
2008 priv->dev->xdp_features |= NETDEV_XDP_ACT_NDO_XMIT;
2009 priv->dev->xdp_features |= NETDEV_XDP_ACT_XSK_ZEROCOPY;
2010 } else {
2011 priv->dev->xdp_features = 0;
2012 }
2013 }
2014
gve_init_priv(struct gve_priv * priv,bool skip_describe_device)2015 static int gve_init_priv(struct gve_priv *priv, bool skip_describe_device)
2016 {
2017 int num_ntfy;
2018 int err;
2019
2020 /* Set up the adminq */
2021 err = gve_adminq_alloc(&priv->pdev->dev, priv);
2022 if (err) {
2023 dev_err(&priv->pdev->dev,
2024 "Failed to alloc admin queue: err=%d\n", err);
2025 return err;
2026 }
2027
2028 err = gve_verify_driver_compatibility(priv);
2029 if (err) {
2030 dev_err(&priv->pdev->dev,
2031 "Could not verify driver compatibility: err=%d\n", err);
2032 goto err;
2033 }
2034
2035 if (skip_describe_device)
2036 goto setup_device;
2037
2038 priv->queue_format = GVE_QUEUE_FORMAT_UNSPECIFIED;
2039 /* Get the initial information we need from the device */
2040 err = gve_adminq_describe_device(priv);
2041 if (err) {
2042 dev_err(&priv->pdev->dev,
2043 "Could not get device information: err=%d\n", err);
2044 goto err;
2045 }
2046 priv->dev->mtu = priv->dev->max_mtu;
2047 num_ntfy = pci_msix_vec_count(priv->pdev);
2048 if (num_ntfy <= 0) {
2049 dev_err(&priv->pdev->dev,
2050 "could not count MSI-x vectors: err=%d\n", num_ntfy);
2051 err = num_ntfy;
2052 goto err;
2053 } else if (num_ntfy < GVE_MIN_MSIX) {
2054 dev_err(&priv->pdev->dev, "gve needs at least %d MSI-x vectors, but only has %d\n",
2055 GVE_MIN_MSIX, num_ntfy);
2056 err = -EINVAL;
2057 goto err;
2058 }
2059
2060 /* Big TCP is only supported on DQ*/
2061 if (!gve_is_gqi(priv))
2062 netif_set_tso_max_size(priv->dev, GVE_DQO_TX_MAX);
2063
2064 priv->num_registered_pages = 0;
2065 priv->rx_copybreak = GVE_DEFAULT_RX_COPYBREAK;
2066 /* gvnic has one Notification Block per MSI-x vector, except for the
2067 * management vector
2068 */
2069 priv->num_ntfy_blks = (num_ntfy - 1) & ~0x1;
2070 priv->mgmt_msix_idx = priv->num_ntfy_blks;
2071
2072 priv->tx_cfg.max_queues =
2073 min_t(int, priv->tx_cfg.max_queues, priv->num_ntfy_blks / 2);
2074 priv->rx_cfg.max_queues =
2075 min_t(int, priv->rx_cfg.max_queues, priv->num_ntfy_blks / 2);
2076
2077 priv->tx_cfg.num_queues = priv->tx_cfg.max_queues;
2078 priv->rx_cfg.num_queues = priv->rx_cfg.max_queues;
2079 if (priv->default_num_queues > 0) {
2080 priv->tx_cfg.num_queues = min_t(int, priv->default_num_queues,
2081 priv->tx_cfg.num_queues);
2082 priv->rx_cfg.num_queues = min_t(int, priv->default_num_queues,
2083 priv->rx_cfg.num_queues);
2084 }
2085
2086 dev_info(&priv->pdev->dev, "TX queues %d, RX queues %d\n",
2087 priv->tx_cfg.num_queues, priv->rx_cfg.num_queues);
2088 dev_info(&priv->pdev->dev, "Max TX queues %d, Max RX queues %d\n",
2089 priv->tx_cfg.max_queues, priv->rx_cfg.max_queues);
2090
2091 if (!gve_is_gqi(priv)) {
2092 priv->tx_coalesce_usecs = GVE_TX_IRQ_RATELIMIT_US_DQO;
2093 priv->rx_coalesce_usecs = GVE_RX_IRQ_RATELIMIT_US_DQO;
2094 }
2095
2096 setup_device:
2097 gve_set_netdev_xdp_features(priv);
2098 err = gve_setup_device_resources(priv);
2099 if (!err)
2100 return 0;
2101 err:
2102 gve_adminq_free(&priv->pdev->dev, priv);
2103 return err;
2104 }
2105
gve_teardown_priv_resources(struct gve_priv * priv)2106 static void gve_teardown_priv_resources(struct gve_priv *priv)
2107 {
2108 gve_teardown_device_resources(priv);
2109 gve_adminq_free(&priv->pdev->dev, priv);
2110 }
2111
gve_trigger_reset(struct gve_priv * priv)2112 static void gve_trigger_reset(struct gve_priv *priv)
2113 {
2114 /* Reset the device by releasing the AQ */
2115 gve_adminq_release(priv);
2116 }
2117
gve_reset_and_teardown(struct gve_priv * priv,bool was_up)2118 static void gve_reset_and_teardown(struct gve_priv *priv, bool was_up)
2119 {
2120 gve_trigger_reset(priv);
2121 /* With the reset having already happened, close cannot fail */
2122 if (was_up)
2123 gve_close(priv->dev);
2124 gve_teardown_priv_resources(priv);
2125 }
2126
gve_reset_recovery(struct gve_priv * priv,bool was_up)2127 static int gve_reset_recovery(struct gve_priv *priv, bool was_up)
2128 {
2129 int err;
2130
2131 err = gve_init_priv(priv, true);
2132 if (err)
2133 goto err;
2134 if (was_up) {
2135 err = gve_open(priv->dev);
2136 if (err)
2137 goto err;
2138 }
2139 return 0;
2140 err:
2141 dev_err(&priv->pdev->dev, "Reset failed! !!! DISABLING ALL QUEUES !!!\n");
2142 gve_turndown(priv);
2143 return err;
2144 }
2145
gve_reset(struct gve_priv * priv,bool attempt_teardown)2146 int gve_reset(struct gve_priv *priv, bool attempt_teardown)
2147 {
2148 bool was_up = netif_carrier_ok(priv->dev);
2149 int err;
2150
2151 dev_info(&priv->pdev->dev, "Performing reset\n");
2152 gve_clear_do_reset(priv);
2153 gve_set_reset_in_progress(priv);
2154 /* If we aren't attempting to teardown normally, just go turndown and
2155 * reset right away.
2156 */
2157 if (!attempt_teardown) {
2158 gve_turndown(priv);
2159 gve_reset_and_teardown(priv, was_up);
2160 } else {
2161 /* Otherwise attempt to close normally */
2162 if (was_up) {
2163 err = gve_close(priv->dev);
2164 /* If that fails reset as we did above */
2165 if (err)
2166 gve_reset_and_teardown(priv, was_up);
2167 }
2168 /* Clean up any remaining resources */
2169 gve_teardown_priv_resources(priv);
2170 }
2171
2172 /* Set it all back up */
2173 err = gve_reset_recovery(priv, was_up);
2174 gve_clear_reset_in_progress(priv);
2175 priv->reset_cnt++;
2176 priv->interface_up_cnt = 0;
2177 priv->interface_down_cnt = 0;
2178 priv->stats_report_trigger_cnt = 0;
2179 return err;
2180 }
2181
gve_write_version(u8 __iomem * driver_version_register)2182 static void gve_write_version(u8 __iomem *driver_version_register)
2183 {
2184 const char *c = gve_version_prefix;
2185
2186 while (*c) {
2187 writeb(*c, driver_version_register);
2188 c++;
2189 }
2190
2191 c = gve_version_str;
2192 while (*c) {
2193 writeb(*c, driver_version_register);
2194 c++;
2195 }
2196 writeb('\n', driver_version_register);
2197 }
2198
gve_probe(struct pci_dev * pdev,const struct pci_device_id * ent)2199 static int gve_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
2200 {
2201 int max_tx_queues, max_rx_queues;
2202 struct net_device *dev;
2203 __be32 __iomem *db_bar;
2204 struct gve_registers __iomem *reg_bar;
2205 struct gve_priv *priv;
2206 int err;
2207
2208 err = pci_enable_device(pdev);
2209 if (err)
2210 return err;
2211
2212 err = pci_request_regions(pdev, gve_driver_name);
2213 if (err)
2214 goto abort_with_enabled;
2215
2216 pci_set_master(pdev);
2217
2218 err = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
2219 if (err) {
2220 dev_err(&pdev->dev, "Failed to set dma mask: err=%d\n", err);
2221 goto abort_with_pci_region;
2222 }
2223
2224 reg_bar = pci_iomap(pdev, GVE_REGISTER_BAR, 0);
2225 if (!reg_bar) {
2226 dev_err(&pdev->dev, "Failed to map pci bar!\n");
2227 err = -ENOMEM;
2228 goto abort_with_pci_region;
2229 }
2230
2231 db_bar = pci_iomap(pdev, GVE_DOORBELL_BAR, 0);
2232 if (!db_bar) {
2233 dev_err(&pdev->dev, "Failed to map doorbell bar!\n");
2234 err = -ENOMEM;
2235 goto abort_with_reg_bar;
2236 }
2237
2238 gve_write_version(®_bar->driver_version);
2239 /* Get max queues to alloc etherdev */
2240 max_tx_queues = ioread32be(®_bar->max_tx_queues);
2241 max_rx_queues = ioread32be(®_bar->max_rx_queues);
2242 /* Alloc and setup the netdev and priv */
2243 dev = alloc_etherdev_mqs(sizeof(*priv), max_tx_queues, max_rx_queues);
2244 if (!dev) {
2245 dev_err(&pdev->dev, "could not allocate netdev\n");
2246 err = -ENOMEM;
2247 goto abort_with_db_bar;
2248 }
2249 SET_NETDEV_DEV(dev, &pdev->dev);
2250 pci_set_drvdata(pdev, dev);
2251 dev->ethtool_ops = &gve_ethtool_ops;
2252 dev->netdev_ops = &gve_netdev_ops;
2253
2254 /* Set default and supported features.
2255 *
2256 * Features might be set in other locations as well (such as
2257 * `gve_adminq_describe_device`).
2258 */
2259 dev->hw_features = NETIF_F_HIGHDMA;
2260 dev->hw_features |= NETIF_F_SG;
2261 dev->hw_features |= NETIF_F_HW_CSUM;
2262 dev->hw_features |= NETIF_F_TSO;
2263 dev->hw_features |= NETIF_F_TSO6;
2264 dev->hw_features |= NETIF_F_TSO_ECN;
2265 dev->hw_features |= NETIF_F_RXCSUM;
2266 dev->hw_features |= NETIF_F_RXHASH;
2267 dev->features = dev->hw_features;
2268 dev->watchdog_timeo = 5 * HZ;
2269 dev->min_mtu = ETH_MIN_MTU;
2270 netif_carrier_off(dev);
2271
2272 priv = netdev_priv(dev);
2273 priv->dev = dev;
2274 priv->pdev = pdev;
2275 priv->msg_enable = DEFAULT_MSG_LEVEL;
2276 priv->reg_bar0 = reg_bar;
2277 priv->db_bar2 = db_bar;
2278 priv->service_task_flags = 0x0;
2279 priv->state_flags = 0x0;
2280 priv->ethtool_flags = 0x0;
2281
2282 gve_set_probe_in_progress(priv);
2283 priv->gve_wq = alloc_ordered_workqueue("gve", 0);
2284 if (!priv->gve_wq) {
2285 dev_err(&pdev->dev, "Could not allocate workqueue");
2286 err = -ENOMEM;
2287 goto abort_with_netdev;
2288 }
2289 INIT_WORK(&priv->service_task, gve_service_task);
2290 INIT_WORK(&priv->stats_report_task, gve_stats_report_task);
2291 priv->tx_cfg.max_queues = max_tx_queues;
2292 priv->rx_cfg.max_queues = max_rx_queues;
2293
2294 err = gve_init_priv(priv, false);
2295 if (err)
2296 goto abort_with_wq;
2297
2298 err = register_netdev(dev);
2299 if (err)
2300 goto abort_with_gve_init;
2301
2302 dev_info(&pdev->dev, "GVE version %s\n", gve_version_str);
2303 dev_info(&pdev->dev, "GVE queue format %d\n", (int)priv->queue_format);
2304 gve_clear_probe_in_progress(priv);
2305 queue_work(priv->gve_wq, &priv->service_task);
2306 return 0;
2307
2308 abort_with_gve_init:
2309 gve_teardown_priv_resources(priv);
2310
2311 abort_with_wq:
2312 destroy_workqueue(priv->gve_wq);
2313
2314 abort_with_netdev:
2315 free_netdev(dev);
2316
2317 abort_with_db_bar:
2318 pci_iounmap(pdev, db_bar);
2319
2320 abort_with_reg_bar:
2321 pci_iounmap(pdev, reg_bar);
2322
2323 abort_with_pci_region:
2324 pci_release_regions(pdev);
2325
2326 abort_with_enabled:
2327 pci_disable_device(pdev);
2328 return err;
2329 }
2330
gve_remove(struct pci_dev * pdev)2331 static void gve_remove(struct pci_dev *pdev)
2332 {
2333 struct net_device *netdev = pci_get_drvdata(pdev);
2334 struct gve_priv *priv = netdev_priv(netdev);
2335 __be32 __iomem *db_bar = priv->db_bar2;
2336 void __iomem *reg_bar = priv->reg_bar0;
2337
2338 unregister_netdev(netdev);
2339 gve_teardown_priv_resources(priv);
2340 destroy_workqueue(priv->gve_wq);
2341 free_netdev(netdev);
2342 pci_iounmap(pdev, db_bar);
2343 pci_iounmap(pdev, reg_bar);
2344 pci_release_regions(pdev);
2345 pci_disable_device(pdev);
2346 }
2347
gve_shutdown(struct pci_dev * pdev)2348 static void gve_shutdown(struct pci_dev *pdev)
2349 {
2350 struct net_device *netdev = pci_get_drvdata(pdev);
2351 struct gve_priv *priv = netdev_priv(netdev);
2352 bool was_up = netif_carrier_ok(priv->dev);
2353
2354 rtnl_lock();
2355 if (was_up && gve_close(priv->dev)) {
2356 /* If the dev was up, attempt to close, if close fails, reset */
2357 gve_reset_and_teardown(priv, was_up);
2358 } else {
2359 /* If the dev wasn't up or close worked, finish tearing down */
2360 gve_teardown_priv_resources(priv);
2361 }
2362 rtnl_unlock();
2363 }
2364
2365 #ifdef CONFIG_PM
gve_suspend(struct pci_dev * pdev,pm_message_t state)2366 static int gve_suspend(struct pci_dev *pdev, pm_message_t state)
2367 {
2368 struct net_device *netdev = pci_get_drvdata(pdev);
2369 struct gve_priv *priv = netdev_priv(netdev);
2370 bool was_up = netif_carrier_ok(priv->dev);
2371
2372 priv->suspend_cnt++;
2373 rtnl_lock();
2374 if (was_up && gve_close(priv->dev)) {
2375 /* If the dev was up, attempt to close, if close fails, reset */
2376 gve_reset_and_teardown(priv, was_up);
2377 } else {
2378 /* If the dev wasn't up or close worked, finish tearing down */
2379 gve_teardown_priv_resources(priv);
2380 }
2381 priv->up_before_suspend = was_up;
2382 rtnl_unlock();
2383 return 0;
2384 }
2385
gve_resume(struct pci_dev * pdev)2386 static int gve_resume(struct pci_dev *pdev)
2387 {
2388 struct net_device *netdev = pci_get_drvdata(pdev);
2389 struct gve_priv *priv = netdev_priv(netdev);
2390 int err;
2391
2392 priv->resume_cnt++;
2393 rtnl_lock();
2394 err = gve_reset_recovery(priv, priv->up_before_suspend);
2395 rtnl_unlock();
2396 return err;
2397 }
2398 #endif /* CONFIG_PM */
2399
2400 static const struct pci_device_id gve_id_table[] = {
2401 { PCI_DEVICE(PCI_VENDOR_ID_GOOGLE, PCI_DEV_ID_GVNIC) },
2402 { }
2403 };
2404
2405 static struct pci_driver gve_driver = {
2406 .name = gve_driver_name,
2407 .id_table = gve_id_table,
2408 .probe = gve_probe,
2409 .remove = gve_remove,
2410 .shutdown = gve_shutdown,
2411 #ifdef CONFIG_PM
2412 .suspend = gve_suspend,
2413 .resume = gve_resume,
2414 #endif
2415 };
2416
2417 module_pci_driver(gve_driver);
2418
2419 MODULE_DEVICE_TABLE(pci, gve_id_table);
2420 MODULE_AUTHOR("Google, Inc.");
2421 MODULE_DESCRIPTION("Google Virtual NIC Driver");
2422 MODULE_LICENSE("Dual MIT/GPL");
2423 MODULE_VERSION(GVE_VERSION);
2424