1 /*
2  * Copyright (C) 2015-2017 Netronome Systems, Inc.
3  *
4  * This software is dual licensed under the GNU General License Version 2,
5  * June 1991 as shown in the file COPYING in the top-level directory of this
6  * source tree or the BSD 2-Clause License provided below.  You have the
7  * option to license this software under the complete terms of either license.
8  *
9  * The BSD 2-Clause License:
10  *
11  *     Redistribution and use in source and binary forms, with or
12  *     without modification, are permitted provided that the following
13  *     conditions are met:
14  *
15  *      1. Redistributions of source code must retain the above
16  *         copyright notice, this list of conditions and the following
17  *         disclaimer.
18  *
19  *      2. Redistributions in binary form must reproduce the above
20  *         copyright notice, this list of conditions and the following
21  *         disclaimer in the documentation and/or other materials
22  *         provided with the distribution.
23  *
24  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
31  * SOFTWARE.
32  */
33 
34 /*
35  * nfp_netvf_main.c
36  * Netronome virtual function network device driver: Main entry point
37  * Author: Jason McMullan <jason.mcmullan@netronome.com>
38  *         Rolf Neugebauer <rolf.neugebauer@netronome.com>
39  */
40 
41 #include <linux/module.h>
42 #include <linux/kernel.h>
43 #include <linux/init.h>
44 #include <linux/etherdevice.h>
45 
46 #include "nfp_net_ctrl.h"
47 #include "nfp_net.h"
48 #include "nfp_main.h"
49 
50 /**
51  * struct nfp_net_vf - NFP VF-specific device structure
52  * @nn:		NFP Net structure for this device
53  * @irq_entries: Pre-allocated array of MSI-X entries
54  * @q_bar:	Pointer to mapped QC memory (NULL if TX/RX mapped directly)
55  * @ddir:	Per-device debugfs directory
56  */
57 struct nfp_net_vf {
58 	struct nfp_net *nn;
59 
60 	struct msix_entry irq_entries[NFP_NET_NON_Q_VECTORS +
61 				      NFP_NET_MAX_TX_RINGS];
62 	u8 __iomem *q_bar;
63 
64 	struct dentry *ddir;
65 };
66 
67 static const char nfp_net_driver_name[] = "nfp_netvf";
68 
69 #define PCI_DEVICE_NFP6000VF		0x6003
70 static const struct pci_device_id nfp_netvf_pci_device_ids[] = {
71 	{ PCI_VENDOR_ID_NETRONOME, PCI_DEVICE_NFP6000VF,
72 	  PCI_VENDOR_ID_NETRONOME, PCI_ANY_ID,
73 	  PCI_ANY_ID, 0,
74 	},
75 	{ 0, } /* Required last entry. */
76 };
77 MODULE_DEVICE_TABLE(pci, nfp_netvf_pci_device_ids);
78 
nfp_netvf_get_mac_addr(struct nfp_net * nn)79 static void nfp_netvf_get_mac_addr(struct nfp_net *nn)
80 {
81 	u8 mac_addr[ETH_ALEN];
82 
83 	put_unaligned_be32(nn_readl(nn, NFP_NET_CFG_MACADDR + 0), &mac_addr[0]);
84 	put_unaligned_be16(nn_readw(nn, NFP_NET_CFG_MACADDR + 6), &mac_addr[4]);
85 
86 	if (!is_valid_ether_addr(mac_addr)) {
87 		eth_hw_addr_random(nn->dp.netdev);
88 		return;
89 	}
90 
91 	ether_addr_copy(nn->dp.netdev->dev_addr, mac_addr);
92 	ether_addr_copy(nn->dp.netdev->perm_addr, mac_addr);
93 }
94 
nfp_netvf_pci_probe(struct pci_dev * pdev,const struct pci_device_id * pci_id)95 static int nfp_netvf_pci_probe(struct pci_dev *pdev,
96 			       const struct pci_device_id *pci_id)
97 {
98 	struct nfp_net_fw_version fw_ver;
99 	int max_tx_rings, max_rx_rings;
100 	u32 tx_bar_off, rx_bar_off;
101 	u32 tx_bar_sz, rx_bar_sz;
102 	int tx_bar_no, rx_bar_no;
103 	struct nfp_net_vf *vf;
104 	unsigned int num_irqs;
105 	u8 __iomem *ctrl_bar;
106 	struct nfp_net *nn;
107 	u32 startq;
108 	int stride;
109 	int err;
110 
111 	vf = kzalloc(sizeof(*vf), GFP_KERNEL);
112 	if (!vf)
113 		return -ENOMEM;
114 	pci_set_drvdata(pdev, vf);
115 
116 	err = pci_enable_device_mem(pdev);
117 	if (err)
118 		goto err_free_vf;
119 
120 	err = pci_request_regions(pdev, nfp_net_driver_name);
121 	if (err) {
122 		dev_err(&pdev->dev, "Unable to allocate device memory.\n");
123 		goto err_pci_disable;
124 	}
125 
126 	pci_set_master(pdev);
127 
128 	err = dma_set_mask_and_coherent(&pdev->dev,
129 					DMA_BIT_MASK(NFP_NET_MAX_DMA_BITS));
130 	if (err)
131 		goto err_pci_regions;
132 
133 	/* Map the Control BAR.
134 	 *
135 	 * Irrespective of the advertised BAR size we only map the
136 	 * first NFP_NET_CFG_BAR_SZ of the BAR.  This keeps the code
137 	 * the identical for PF and VF drivers.
138 	 */
139 	ctrl_bar = ioremap_nocache(pci_resource_start(pdev, NFP_NET_CTRL_BAR),
140 				   NFP_NET_CFG_BAR_SZ);
141 	if (!ctrl_bar) {
142 		dev_err(&pdev->dev,
143 			"Failed to map resource %d\n", NFP_NET_CTRL_BAR);
144 		err = -EIO;
145 		goto err_pci_regions;
146 	}
147 
148 	nfp_net_get_fw_version(&fw_ver, ctrl_bar);
149 	if (fw_ver.resv || fw_ver.class != NFP_NET_CFG_VERSION_CLASS_GENERIC) {
150 		dev_err(&pdev->dev, "Unknown Firmware ABI %d.%d.%d.%d\n",
151 			fw_ver.resv, fw_ver.class, fw_ver.major, fw_ver.minor);
152 		err = -EINVAL;
153 		goto err_ctrl_unmap;
154 	}
155 
156 	/* Determine stride */
157 	if (nfp_net_fw_ver_eq(&fw_ver, 0, 0, 0, 1)) {
158 		stride = 2;
159 		tx_bar_no = NFP_NET_Q0_BAR;
160 		rx_bar_no = NFP_NET_Q1_BAR;
161 		dev_warn(&pdev->dev, "OBSOLETE Firmware detected - VF isolation not available\n");
162 	} else {
163 		switch (fw_ver.major) {
164 		case 1 ... 5:
165 			stride = 4;
166 			tx_bar_no = NFP_NET_Q0_BAR;
167 			rx_bar_no = tx_bar_no;
168 			break;
169 		default:
170 			dev_err(&pdev->dev, "Unsupported Firmware ABI %d.%d.%d.%d\n",
171 				fw_ver.resv, fw_ver.class,
172 				fw_ver.major, fw_ver.minor);
173 			err = -EINVAL;
174 			goto err_ctrl_unmap;
175 		}
176 	}
177 
178 	/* Find out how many rings are supported */
179 	max_tx_rings = readl(ctrl_bar + NFP_NET_CFG_MAX_TXRINGS);
180 	max_rx_rings = readl(ctrl_bar + NFP_NET_CFG_MAX_RXRINGS);
181 
182 	tx_bar_sz = NFP_QCP_QUEUE_ADDR_SZ * max_tx_rings * stride;
183 	rx_bar_sz = NFP_QCP_QUEUE_ADDR_SZ * max_rx_rings * stride;
184 
185 	/* Sanity checks */
186 	if (tx_bar_sz > pci_resource_len(pdev, tx_bar_no)) {
187 		dev_err(&pdev->dev,
188 			"TX BAR too small for number of TX rings. Adjusting\n");
189 		tx_bar_sz = pci_resource_len(pdev, tx_bar_no);
190 		max_tx_rings = (tx_bar_sz / NFP_QCP_QUEUE_ADDR_SZ) / 2;
191 	}
192 	if (rx_bar_sz > pci_resource_len(pdev, rx_bar_no)) {
193 		dev_err(&pdev->dev,
194 			"RX BAR too small for number of RX rings. Adjusting\n");
195 		rx_bar_sz = pci_resource_len(pdev, rx_bar_no);
196 		max_rx_rings = (rx_bar_sz / NFP_QCP_QUEUE_ADDR_SZ) / 2;
197 	}
198 
199 	startq = readl(ctrl_bar + NFP_NET_CFG_START_TXQ);
200 	tx_bar_off = NFP_PCIE_QUEUE(startq);
201 	startq = readl(ctrl_bar + NFP_NET_CFG_START_RXQ);
202 	rx_bar_off = NFP_PCIE_QUEUE(startq);
203 
204 	/* Allocate and initialise the netdev */
205 	nn = nfp_net_alloc(pdev, true, max_tx_rings, max_rx_rings);
206 	if (IS_ERR(nn)) {
207 		err = PTR_ERR(nn);
208 		goto err_ctrl_unmap;
209 	}
210 	vf->nn = nn;
211 
212 	nn->fw_ver = fw_ver;
213 	nn->dp.ctrl_bar = ctrl_bar;
214 	nn->dp.is_vf = 1;
215 	nn->stride_tx = stride;
216 	nn->stride_rx = stride;
217 
218 	if (rx_bar_no == tx_bar_no) {
219 		u32 bar_off, bar_sz;
220 		resource_size_t map_addr;
221 
222 		/* Make a single overlapping BAR mapping */
223 		if (tx_bar_off < rx_bar_off)
224 			bar_off = tx_bar_off;
225 		else
226 			bar_off = rx_bar_off;
227 
228 		if ((tx_bar_off + tx_bar_sz) > (rx_bar_off + rx_bar_sz))
229 			bar_sz = (tx_bar_off + tx_bar_sz) - bar_off;
230 		else
231 			bar_sz = (rx_bar_off + rx_bar_sz) - bar_off;
232 
233 		map_addr = pci_resource_start(pdev, tx_bar_no) + bar_off;
234 		vf->q_bar = ioremap_nocache(map_addr, bar_sz);
235 		if (!vf->q_bar) {
236 			nn_err(nn, "Failed to map resource %d\n", tx_bar_no);
237 			err = -EIO;
238 			goto err_netdev_free;
239 		}
240 
241 		/* TX queues */
242 		nn->tx_bar = vf->q_bar + (tx_bar_off - bar_off);
243 		/* RX queues */
244 		nn->rx_bar = vf->q_bar + (rx_bar_off - bar_off);
245 	} else {
246 		resource_size_t map_addr;
247 
248 		/* TX queues */
249 		map_addr = pci_resource_start(pdev, tx_bar_no) + tx_bar_off;
250 		nn->tx_bar = ioremap_nocache(map_addr, tx_bar_sz);
251 		if (!nn->tx_bar) {
252 			nn_err(nn, "Failed to map resource %d\n", tx_bar_no);
253 			err = -EIO;
254 			goto err_netdev_free;
255 		}
256 
257 		/* RX queues */
258 		map_addr = pci_resource_start(pdev, rx_bar_no) + rx_bar_off;
259 		nn->rx_bar = ioremap_nocache(map_addr, rx_bar_sz);
260 		if (!nn->rx_bar) {
261 			nn_err(nn, "Failed to map resource %d\n", rx_bar_no);
262 			err = -EIO;
263 			goto err_unmap_tx;
264 		}
265 	}
266 
267 	nfp_netvf_get_mac_addr(nn);
268 
269 	num_irqs = nfp_net_irqs_alloc(pdev, vf->irq_entries,
270 				      NFP_NET_MIN_VNIC_IRQS,
271 				      NFP_NET_NON_Q_VECTORS +
272 				      nn->dp.num_r_vecs);
273 	if (!num_irqs) {
274 		nn_warn(nn, "Unable to allocate MSI-X Vectors. Exiting\n");
275 		err = -EIO;
276 		goto err_unmap_rx;
277 	}
278 	nfp_net_irqs_assign(nn, vf->irq_entries, num_irqs);
279 
280 	err = nfp_net_init(nn);
281 	if (err)
282 		goto err_irqs_disable;
283 
284 	nfp_net_info(nn);
285 	vf->ddir = nfp_net_debugfs_device_add(pdev);
286 	nfp_net_debugfs_vnic_add(nn, vf->ddir);
287 
288 	return 0;
289 
290 err_irqs_disable:
291 	nfp_net_irqs_disable(pdev);
292 err_unmap_rx:
293 	if (!vf->q_bar)
294 		iounmap(nn->rx_bar);
295 err_unmap_tx:
296 	if (!vf->q_bar)
297 		iounmap(nn->tx_bar);
298 	else
299 		iounmap(vf->q_bar);
300 err_netdev_free:
301 	nfp_net_free(nn);
302 err_ctrl_unmap:
303 	iounmap(ctrl_bar);
304 err_pci_regions:
305 	pci_release_regions(pdev);
306 err_pci_disable:
307 	pci_disable_device(pdev);
308 err_free_vf:
309 	pci_set_drvdata(pdev, NULL);
310 	kfree(vf);
311 	return err;
312 }
313 
nfp_netvf_pci_remove(struct pci_dev * pdev)314 static void nfp_netvf_pci_remove(struct pci_dev *pdev)
315 {
316 	struct nfp_net_vf *vf = pci_get_drvdata(pdev);
317 	struct nfp_net *nn = vf->nn;
318 
319 	/* Note, the order is slightly different from above as we need
320 	 * to keep the nn pointer around till we have freed everything.
321 	 */
322 	nfp_net_debugfs_dir_clean(&nn->debugfs_dir);
323 	nfp_net_debugfs_dir_clean(&vf->ddir);
324 
325 	nfp_net_clean(nn);
326 
327 	nfp_net_irqs_disable(pdev);
328 
329 	if (!vf->q_bar) {
330 		iounmap(nn->rx_bar);
331 		iounmap(nn->tx_bar);
332 	} else {
333 		iounmap(vf->q_bar);
334 	}
335 	iounmap(nn->dp.ctrl_bar);
336 
337 	nfp_net_free(nn);
338 
339 	pci_release_regions(pdev);
340 	pci_disable_device(pdev);
341 
342 	pci_set_drvdata(pdev, NULL);
343 	kfree(vf);
344 }
345 
346 struct pci_driver nfp_netvf_pci_driver = {
347 	.name        = nfp_net_driver_name,
348 	.id_table    = nfp_netvf_pci_device_ids,
349 	.probe       = nfp_netvf_pci_probe,
350 	.remove      = nfp_netvf_pci_remove,
351 };
352