1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright(c) 1999 - 2018 Intel Corporation. */
3 
4 #include "vf.h"
5 #include "ixgbevf.h"
6 
7 /* On Hyper-V, to reset, we need to read from this offset
8  * from the PCI config space. This is the mechanism used on
9  * Hyper-V to support PF/VF communication.
10  */
11 #define IXGBE_HV_RESET_OFFSET           0x201
12 
ixgbevf_write_msg_read_ack(struct ixgbe_hw * hw,u32 * msg,u32 * retmsg,u16 size)13 static inline s32 ixgbevf_write_msg_read_ack(struct ixgbe_hw *hw, u32 *msg,
14 					     u32 *retmsg, u16 size)
15 {
16 	struct ixgbe_mbx_info *mbx = &hw->mbx;
17 	s32 retval = mbx->ops.write_posted(hw, msg, size);
18 
19 	if (retval)
20 		return retval;
21 
22 	return mbx->ops.read_posted(hw, retmsg, size);
23 }
24 
25 /**
26  *  ixgbevf_start_hw_vf - Prepare hardware for Tx/Rx
27  *  @hw: pointer to hardware structure
28  *
29  *  Starts the hardware by filling the bus info structure and media type, clears
30  *  all on chip counters, initializes receive address registers, multicast
31  *  table, VLAN filter table, calls routine to set up link and flow control
32  *  settings, and leaves transmit and receive units disabled and uninitialized
33  **/
ixgbevf_start_hw_vf(struct ixgbe_hw * hw)34 static s32 ixgbevf_start_hw_vf(struct ixgbe_hw *hw)
35 {
36 	/* Clear adapter stopped flag */
37 	hw->adapter_stopped = false;
38 
39 	return 0;
40 }
41 
42 /**
43  *  ixgbevf_init_hw_vf - virtual function hardware initialization
44  *  @hw: pointer to hardware structure
45  *
46  *  Initialize the hardware by resetting the hardware and then starting
47  *  the hardware
48  **/
ixgbevf_init_hw_vf(struct ixgbe_hw * hw)49 static s32 ixgbevf_init_hw_vf(struct ixgbe_hw *hw)
50 {
51 	s32 status = hw->mac.ops.start_hw(hw);
52 
53 	hw->mac.ops.get_mac_addr(hw, hw->mac.addr);
54 
55 	return status;
56 }
57 
58 /**
59  *  ixgbevf_reset_hw_vf - Performs hardware reset
60  *  @hw: pointer to hardware structure
61  *
62  *  Resets the hardware by resetting the transmit and receive units, masks and
63  *  clears all interrupts.
64  **/
ixgbevf_reset_hw_vf(struct ixgbe_hw * hw)65 static s32 ixgbevf_reset_hw_vf(struct ixgbe_hw *hw)
66 {
67 	struct ixgbe_mbx_info *mbx = &hw->mbx;
68 	u32 timeout = IXGBE_VF_INIT_TIMEOUT;
69 	s32 ret_val = IXGBE_ERR_INVALID_MAC_ADDR;
70 	u32 msgbuf[IXGBE_VF_PERMADDR_MSG_LEN];
71 	u8 *addr = (u8 *)(&msgbuf[1]);
72 
73 	/* Call adapter stop to disable tx/rx and clear interrupts */
74 	hw->mac.ops.stop_adapter(hw);
75 
76 	/* reset the api version */
77 	hw->api_version = ixgbe_mbox_api_10;
78 
79 	IXGBE_WRITE_REG(hw, IXGBE_VFCTRL, IXGBE_CTRL_RST);
80 	IXGBE_WRITE_FLUSH(hw);
81 
82 	/* we cannot reset while the RSTI / RSTD bits are asserted */
83 	while (!mbx->ops.check_for_rst(hw) && timeout) {
84 		timeout--;
85 		udelay(5);
86 	}
87 
88 	if (!timeout)
89 		return IXGBE_ERR_RESET_FAILED;
90 
91 	/* mailbox timeout can now become active */
92 	mbx->timeout = IXGBE_VF_MBX_INIT_TIMEOUT;
93 
94 	msgbuf[0] = IXGBE_VF_RESET;
95 	mbx->ops.write_posted(hw, msgbuf, 1);
96 
97 	mdelay(10);
98 
99 	/* set our "perm_addr" based on info provided by PF
100 	 * also set up the mc_filter_type which is piggy backed
101 	 * on the mac address in word 3
102 	 */
103 	ret_val = mbx->ops.read_posted(hw, msgbuf, IXGBE_VF_PERMADDR_MSG_LEN);
104 	if (ret_val)
105 		return ret_val;
106 
107 	/* New versions of the PF may NACK the reset return message
108 	 * to indicate that no MAC address has yet been assigned for
109 	 * the VF.
110 	 */
111 	if (msgbuf[0] != (IXGBE_VF_RESET | IXGBE_VT_MSGTYPE_ACK) &&
112 	    msgbuf[0] != (IXGBE_VF_RESET | IXGBE_VT_MSGTYPE_NACK))
113 		return IXGBE_ERR_INVALID_MAC_ADDR;
114 
115 	if (msgbuf[0] == (IXGBE_VF_RESET | IXGBE_VT_MSGTYPE_ACK))
116 		ether_addr_copy(hw->mac.perm_addr, addr);
117 
118 	hw->mac.mc_filter_type = msgbuf[IXGBE_VF_MC_TYPE_WORD];
119 
120 	return 0;
121 }
122 
123 /**
124  * Hyper-V variant; the VF/PF communication is through the PCI
125  * config space.
126  * @hw: pointer to private hardware struct
127  */
ixgbevf_hv_reset_hw_vf(struct ixgbe_hw * hw)128 static s32 ixgbevf_hv_reset_hw_vf(struct ixgbe_hw *hw)
129 {
130 #if IS_ENABLED(CONFIG_PCI_MMCONFIG)
131 	struct ixgbevf_adapter *adapter = hw->back;
132 	int i;
133 
134 	for (i = 0; i < 6; i++)
135 		pci_read_config_byte(adapter->pdev,
136 				     (i + IXGBE_HV_RESET_OFFSET),
137 				     &hw->mac.perm_addr[i]);
138 	return 0;
139 #else
140 	pr_err("PCI_MMCONFIG needs to be enabled for Hyper-V\n");
141 	return -EOPNOTSUPP;
142 #endif
143 }
144 
145 /**
146  *  ixgbevf_stop_hw_vf - Generic stop Tx/Rx units
147  *  @hw: pointer to hardware structure
148  *
149  *  Sets the adapter_stopped flag within ixgbe_hw struct. Clears interrupts,
150  *  disables transmit and receive units. The adapter_stopped flag is used by
151  *  the shared code and drivers to determine if the adapter is in a stopped
152  *  state and should not touch the hardware.
153  **/
ixgbevf_stop_hw_vf(struct ixgbe_hw * hw)154 static s32 ixgbevf_stop_hw_vf(struct ixgbe_hw *hw)
155 {
156 	u32 number_of_queues;
157 	u32 reg_val;
158 	u16 i;
159 
160 	/* Set the adapter_stopped flag so other driver functions stop touching
161 	 * the hardware
162 	 */
163 	hw->adapter_stopped = true;
164 
165 	/* Disable the receive unit by stopped each queue */
166 	number_of_queues = hw->mac.max_rx_queues;
167 	for (i = 0; i < number_of_queues; i++) {
168 		reg_val = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(i));
169 		if (reg_val & IXGBE_RXDCTL_ENABLE) {
170 			reg_val &= ~IXGBE_RXDCTL_ENABLE;
171 			IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(i), reg_val);
172 		}
173 	}
174 
175 	IXGBE_WRITE_FLUSH(hw);
176 
177 	/* Clear interrupt mask to stop from interrupts being generated */
178 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMC, IXGBE_VF_IRQ_CLEAR_MASK);
179 
180 	/* Clear any pending interrupts */
181 	IXGBE_READ_REG(hw, IXGBE_VTEICR);
182 
183 	/* Disable the transmit unit.  Each queue must be disabled. */
184 	number_of_queues = hw->mac.max_tx_queues;
185 	for (i = 0; i < number_of_queues; i++) {
186 		reg_val = IXGBE_READ_REG(hw, IXGBE_VFTXDCTL(i));
187 		if (reg_val & IXGBE_TXDCTL_ENABLE) {
188 			reg_val &= ~IXGBE_TXDCTL_ENABLE;
189 			IXGBE_WRITE_REG(hw, IXGBE_VFTXDCTL(i), reg_val);
190 		}
191 	}
192 
193 	return 0;
194 }
195 
196 /**
197  *  ixgbevf_mta_vector - Determines bit-vector in multicast table to set
198  *  @hw: pointer to hardware structure
199  *  @mc_addr: the multicast address
200  *
201  *  Extracts the 12 bits, from a multicast address, to determine which
202  *  bit-vector to set in the multicast table. The hardware uses 12 bits, from
203  *  incoming Rx multicast addresses, to determine the bit-vector to check in
204  *  the MTA. Which of the 4 combination, of 12-bits, the hardware uses is set
205  *  by the MO field of the MCSTCTRL. The MO field is set during initialization
206  *  to mc_filter_type.
207  **/
ixgbevf_mta_vector(struct ixgbe_hw * hw,u8 * mc_addr)208 static s32 ixgbevf_mta_vector(struct ixgbe_hw *hw, u8 *mc_addr)
209 {
210 	u32 vector = 0;
211 
212 	switch (hw->mac.mc_filter_type) {
213 	case 0:   /* use bits [47:36] of the address */
214 		vector = ((mc_addr[4] >> 4) | (((u16)mc_addr[5]) << 4));
215 		break;
216 	case 1:   /* use bits [46:35] of the address */
217 		vector = ((mc_addr[4] >> 3) | (((u16)mc_addr[5]) << 5));
218 		break;
219 	case 2:   /* use bits [45:34] of the address */
220 		vector = ((mc_addr[4] >> 2) | (((u16)mc_addr[5]) << 6));
221 		break;
222 	case 3:   /* use bits [43:32] of the address */
223 		vector = ((mc_addr[4]) | (((u16)mc_addr[5]) << 8));
224 		break;
225 	default:  /* Invalid mc_filter_type */
226 		break;
227 	}
228 
229 	/* vector can only be 12-bits or boundary will be exceeded */
230 	vector &= 0xFFF;
231 	return vector;
232 }
233 
234 /**
235  *  ixgbevf_get_mac_addr_vf - Read device MAC address
236  *  @hw: pointer to the HW structure
237  *  @mac_addr: pointer to storage for retrieved MAC address
238  **/
ixgbevf_get_mac_addr_vf(struct ixgbe_hw * hw,u8 * mac_addr)239 static s32 ixgbevf_get_mac_addr_vf(struct ixgbe_hw *hw, u8 *mac_addr)
240 {
241 	ether_addr_copy(mac_addr, hw->mac.perm_addr);
242 
243 	return 0;
244 }
245 
ixgbevf_set_uc_addr_vf(struct ixgbe_hw * hw,u32 index,u8 * addr)246 static s32 ixgbevf_set_uc_addr_vf(struct ixgbe_hw *hw, u32 index, u8 *addr)
247 {
248 	u32 msgbuf[3], msgbuf_chk;
249 	u8 *msg_addr = (u8 *)(&msgbuf[1]);
250 	s32 ret_val;
251 
252 	memset(msgbuf, 0, sizeof(msgbuf));
253 	/* If index is one then this is the start of a new list and needs
254 	 * indication to the PF so it can do it's own list management.
255 	 * If it is zero then that tells the PF to just clear all of
256 	 * this VF's macvlans and there is no new list.
257 	 */
258 	msgbuf[0] |= index << IXGBE_VT_MSGINFO_SHIFT;
259 	msgbuf[0] |= IXGBE_VF_SET_MACVLAN;
260 	msgbuf_chk = msgbuf[0];
261 
262 	if (addr)
263 		ether_addr_copy(msg_addr, addr);
264 
265 	ret_val = ixgbevf_write_msg_read_ack(hw, msgbuf, msgbuf,
266 					     ARRAY_SIZE(msgbuf));
267 	if (!ret_val) {
268 		msgbuf[0] &= ~IXGBE_VT_MSGTYPE_CTS;
269 
270 		if (msgbuf[0] == (msgbuf_chk | IXGBE_VT_MSGTYPE_NACK))
271 			return -ENOMEM;
272 	}
273 
274 	return ret_val;
275 }
276 
ixgbevf_hv_set_uc_addr_vf(struct ixgbe_hw * hw,u32 index,u8 * addr)277 static s32 ixgbevf_hv_set_uc_addr_vf(struct ixgbe_hw *hw, u32 index, u8 *addr)
278 {
279 	return -EOPNOTSUPP;
280 }
281 
282 /**
283  * ixgbevf_get_reta_locked - get the RSS redirection table (RETA) contents.
284  * @hw: pointer to hardware structure
285  * @reta: buffer to fill with RETA contents.
286  * @num_rx_queues: Number of Rx queues configured for this port
287  *
288  * The "reta" buffer should be big enough to contain 32 registers.
289  *
290  * Returns: 0 on success.
291  *          if API doesn't support this operation - (-EOPNOTSUPP).
292  */
ixgbevf_get_reta_locked(struct ixgbe_hw * hw,u32 * reta,int num_rx_queues)293 int ixgbevf_get_reta_locked(struct ixgbe_hw *hw, u32 *reta, int num_rx_queues)
294 {
295 	int err, i, j;
296 	u32 msgbuf[IXGBE_VFMAILBOX_SIZE];
297 	u32 *hw_reta = &msgbuf[1];
298 	u32 mask = 0;
299 
300 	/* We have to use a mailbox for 82599 and x540 devices only.
301 	 * For these devices RETA has 128 entries.
302 	 * Also these VFs support up to 4 RSS queues. Therefore PF will compress
303 	 * 16 RETA entries in each DWORD giving 2 bits to each entry.
304 	 */
305 	int dwords = IXGBEVF_82599_RETA_SIZE / 16;
306 
307 	/* We support the RSS querying for 82599 and x540 devices only.
308 	 * Thus return an error if API doesn't support RETA querying or querying
309 	 * is not supported for this device type.
310 	 */
311 	switch (hw->api_version) {
312 	case ixgbe_mbox_api_14:
313 	case ixgbe_mbox_api_13:
314 	case ixgbe_mbox_api_12:
315 		if (hw->mac.type < ixgbe_mac_X550_vf)
316 			break;
317 		fallthrough;
318 	default:
319 		return -EOPNOTSUPP;
320 	}
321 
322 	msgbuf[0] = IXGBE_VF_GET_RETA;
323 
324 	err = hw->mbx.ops.write_posted(hw, msgbuf, 1);
325 
326 	if (err)
327 		return err;
328 
329 	err = hw->mbx.ops.read_posted(hw, msgbuf, dwords + 1);
330 
331 	if (err)
332 		return err;
333 
334 	msgbuf[0] &= ~IXGBE_VT_MSGTYPE_CTS;
335 
336 	/* If the operation has been refused by a PF return -EPERM */
337 	if (msgbuf[0] == (IXGBE_VF_GET_RETA | IXGBE_VT_MSGTYPE_NACK))
338 		return -EPERM;
339 
340 	/* If we didn't get an ACK there must have been
341 	 * some sort of mailbox error so we should treat it
342 	 * as such.
343 	 */
344 	if (msgbuf[0] != (IXGBE_VF_GET_RETA | IXGBE_VT_MSGTYPE_ACK))
345 		return IXGBE_ERR_MBX;
346 
347 	/* ixgbevf doesn't support more than 2 queues at the moment */
348 	if (num_rx_queues > 1)
349 		mask = 0x1;
350 
351 	for (i = 0; i < dwords; i++)
352 		for (j = 0; j < 16; j++)
353 			reta[i * 16 + j] = (hw_reta[i] >> (2 * j)) & mask;
354 
355 	return 0;
356 }
357 
358 /**
359  * ixgbevf_get_rss_key_locked - get the RSS Random Key
360  * @hw: pointer to the HW structure
361  * @rss_key: buffer to fill with RSS Hash Key contents.
362  *
363  * The "rss_key" buffer should be big enough to contain 10 registers.
364  *
365  * Returns: 0 on success.
366  *          if API doesn't support this operation - (-EOPNOTSUPP).
367  */
ixgbevf_get_rss_key_locked(struct ixgbe_hw * hw,u8 * rss_key)368 int ixgbevf_get_rss_key_locked(struct ixgbe_hw *hw, u8 *rss_key)
369 {
370 	int err;
371 	u32 msgbuf[IXGBE_VFMAILBOX_SIZE];
372 
373 	/* We currently support the RSS Random Key retrieval for 82599 and x540
374 	 * devices only.
375 	 *
376 	 * Thus return an error if API doesn't support RSS Random Key retrieval
377 	 * or if the operation is not supported for this device type.
378 	 */
379 	switch (hw->api_version) {
380 	case ixgbe_mbox_api_14:
381 	case ixgbe_mbox_api_13:
382 	case ixgbe_mbox_api_12:
383 		if (hw->mac.type < ixgbe_mac_X550_vf)
384 			break;
385 		fallthrough;
386 	default:
387 		return -EOPNOTSUPP;
388 	}
389 
390 	msgbuf[0] = IXGBE_VF_GET_RSS_KEY;
391 	err = hw->mbx.ops.write_posted(hw, msgbuf, 1);
392 
393 	if (err)
394 		return err;
395 
396 	err = hw->mbx.ops.read_posted(hw, msgbuf, 11);
397 
398 	if (err)
399 		return err;
400 
401 	msgbuf[0] &= ~IXGBE_VT_MSGTYPE_CTS;
402 
403 	/* If the operation has been refused by a PF return -EPERM */
404 	if (msgbuf[0] == (IXGBE_VF_GET_RSS_KEY | IXGBE_VT_MSGTYPE_NACK))
405 		return -EPERM;
406 
407 	/* If we didn't get an ACK there must have been
408 	 * some sort of mailbox error so we should treat it
409 	 * as such.
410 	 */
411 	if (msgbuf[0] != (IXGBE_VF_GET_RSS_KEY | IXGBE_VT_MSGTYPE_ACK))
412 		return IXGBE_ERR_MBX;
413 
414 	memcpy(rss_key, msgbuf + 1, IXGBEVF_RSS_HASH_KEY_SIZE);
415 
416 	return 0;
417 }
418 
419 /**
420  *  ixgbevf_set_rar_vf - set device MAC address
421  *  @hw: pointer to hardware structure
422  *  @index: Receive address register to write
423  *  @addr: Address to put into receive address register
424  *  @vmdq: Unused in this implementation
425  **/
ixgbevf_set_rar_vf(struct ixgbe_hw * hw,u32 index,u8 * addr,u32 vmdq)426 static s32 ixgbevf_set_rar_vf(struct ixgbe_hw *hw, u32 index, u8 *addr,
427 			      u32 vmdq)
428 {
429 	u32 msgbuf[3];
430 	u8 *msg_addr = (u8 *)(&msgbuf[1]);
431 	s32 ret_val;
432 
433 	memset(msgbuf, 0, sizeof(msgbuf));
434 	msgbuf[0] = IXGBE_VF_SET_MAC_ADDR;
435 	ether_addr_copy(msg_addr, addr);
436 
437 	ret_val = ixgbevf_write_msg_read_ack(hw, msgbuf, msgbuf,
438 					     ARRAY_SIZE(msgbuf));
439 	msgbuf[0] &= ~IXGBE_VT_MSGTYPE_CTS;
440 
441 	/* if nacked the address was rejected, use "perm_addr" */
442 	if (!ret_val &&
443 	    (msgbuf[0] == (IXGBE_VF_SET_MAC_ADDR | IXGBE_VT_MSGTYPE_NACK))) {
444 		ixgbevf_get_mac_addr_vf(hw, hw->mac.addr);
445 		return IXGBE_ERR_MBX;
446 	}
447 
448 	return ret_val;
449 }
450 
451 /**
452  *  ixgbevf_hv_set_rar_vf - set device MAC address Hyper-V variant
453  *  @hw: pointer to hardware structure
454  *  @index: Receive address register to write
455  *  @addr: Address to put into receive address register
456  *  @vmdq: Unused in this implementation
457  *
458  * We don't really allow setting the device MAC address. However,
459  * if the address being set is the permanent MAC address we will
460  * permit that.
461  **/
ixgbevf_hv_set_rar_vf(struct ixgbe_hw * hw,u32 index,u8 * addr,u32 vmdq)462 static s32 ixgbevf_hv_set_rar_vf(struct ixgbe_hw *hw, u32 index, u8 *addr,
463 				 u32 vmdq)
464 {
465 	if (ether_addr_equal(addr, hw->mac.perm_addr))
466 		return 0;
467 
468 	return -EOPNOTSUPP;
469 }
470 
471 /**
472  *  ixgbevf_update_mc_addr_list_vf - Update Multicast addresses
473  *  @hw: pointer to the HW structure
474  *  @netdev: pointer to net device structure
475  *
476  *  Updates the Multicast Table Array.
477  **/
ixgbevf_update_mc_addr_list_vf(struct ixgbe_hw * hw,struct net_device * netdev)478 static s32 ixgbevf_update_mc_addr_list_vf(struct ixgbe_hw *hw,
479 					  struct net_device *netdev)
480 {
481 	struct netdev_hw_addr *ha;
482 	u32 msgbuf[IXGBE_VFMAILBOX_SIZE];
483 	u16 *vector_list = (u16 *)&msgbuf[1];
484 	u32 cnt, i;
485 
486 	/* Each entry in the list uses 1 16 bit word.  We have 30
487 	 * 16 bit words available in our HW msg buffer (minus 1 for the
488 	 * msg type).  That's 30 hash values if we pack 'em right.  If
489 	 * there are more than 30 MC addresses to add then punt the
490 	 * extras for now and then add code to handle more than 30 later.
491 	 * It would be unusual for a server to request that many multi-cast
492 	 * addresses except for in large enterprise network environments.
493 	 */
494 
495 	cnt = netdev_mc_count(netdev);
496 	if (cnt > 30)
497 		cnt = 30;
498 	msgbuf[0] = IXGBE_VF_SET_MULTICAST;
499 	msgbuf[0] |= cnt << IXGBE_VT_MSGINFO_SHIFT;
500 
501 	i = 0;
502 	netdev_for_each_mc_addr(ha, netdev) {
503 		if (i == cnt)
504 			break;
505 		if (is_link_local_ether_addr(ha->addr))
506 			continue;
507 
508 		vector_list[i++] = ixgbevf_mta_vector(hw, ha->addr);
509 	}
510 
511 	return ixgbevf_write_msg_read_ack(hw, msgbuf, msgbuf,
512 			IXGBE_VFMAILBOX_SIZE);
513 }
514 
515 /**
516  * Hyper-V variant - just a stub.
517  * @hw: unused
518  * @netdev: unused
519  */
ixgbevf_hv_update_mc_addr_list_vf(struct ixgbe_hw * hw,struct net_device * netdev)520 static s32 ixgbevf_hv_update_mc_addr_list_vf(struct ixgbe_hw *hw,
521 					     struct net_device *netdev)
522 {
523 	return -EOPNOTSUPP;
524 }
525 
526 /**
527  *  ixgbevf_update_xcast_mode - Update Multicast mode
528  *  @hw: pointer to the HW structure
529  *  @xcast_mode: new multicast mode
530  *
531  *  Updates the Multicast Mode of VF.
532  **/
ixgbevf_update_xcast_mode(struct ixgbe_hw * hw,int xcast_mode)533 static s32 ixgbevf_update_xcast_mode(struct ixgbe_hw *hw, int xcast_mode)
534 {
535 	u32 msgbuf[2];
536 	s32 err;
537 
538 	switch (hw->api_version) {
539 	case ixgbe_mbox_api_12:
540 		/* promisc introduced in 1.3 version */
541 		if (xcast_mode == IXGBEVF_XCAST_MODE_PROMISC)
542 			return -EOPNOTSUPP;
543 		fallthrough;
544 	case ixgbe_mbox_api_14:
545 	case ixgbe_mbox_api_13:
546 		break;
547 	default:
548 		return -EOPNOTSUPP;
549 	}
550 
551 	msgbuf[0] = IXGBE_VF_UPDATE_XCAST_MODE;
552 	msgbuf[1] = xcast_mode;
553 
554 	err = ixgbevf_write_msg_read_ack(hw, msgbuf, msgbuf,
555 					 ARRAY_SIZE(msgbuf));
556 	if (err)
557 		return err;
558 
559 	msgbuf[0] &= ~IXGBE_VT_MSGTYPE_CTS;
560 	if (msgbuf[0] == (IXGBE_VF_UPDATE_XCAST_MODE | IXGBE_VT_MSGTYPE_NACK))
561 		return -EPERM;
562 
563 	return 0;
564 }
565 
566 /**
567  * Hyper-V variant - just a stub.
568  * @hw: unused
569  * @xcast_mode: unused
570  */
ixgbevf_hv_update_xcast_mode(struct ixgbe_hw * hw,int xcast_mode)571 static s32 ixgbevf_hv_update_xcast_mode(struct ixgbe_hw *hw, int xcast_mode)
572 {
573 	return -EOPNOTSUPP;
574 }
575 
576 /**
577  *  ixgbevf_set_vfta_vf - Set/Unset VLAN filter table address
578  *  @hw: pointer to the HW structure
579  *  @vlan: 12 bit VLAN ID
580  *  @vind: unused by VF drivers
581  *  @vlan_on: if true then set bit, else clear bit
582  **/
ixgbevf_set_vfta_vf(struct ixgbe_hw * hw,u32 vlan,u32 vind,bool vlan_on)583 static s32 ixgbevf_set_vfta_vf(struct ixgbe_hw *hw, u32 vlan, u32 vind,
584 			       bool vlan_on)
585 {
586 	u32 msgbuf[2];
587 	s32 err;
588 
589 	msgbuf[0] = IXGBE_VF_SET_VLAN;
590 	msgbuf[1] = vlan;
591 	/* Setting the 8 bit field MSG INFO to TRUE indicates "add" */
592 	msgbuf[0] |= vlan_on << IXGBE_VT_MSGINFO_SHIFT;
593 
594 	err = ixgbevf_write_msg_read_ack(hw, msgbuf, msgbuf,
595 					 ARRAY_SIZE(msgbuf));
596 	if (err)
597 		goto mbx_err;
598 
599 	/* remove extra bits from the message */
600 	msgbuf[0] &= ~IXGBE_VT_MSGTYPE_CTS;
601 	msgbuf[0] &= ~(0xFF << IXGBE_VT_MSGINFO_SHIFT);
602 
603 	if (msgbuf[0] != (IXGBE_VF_SET_VLAN | IXGBE_VT_MSGTYPE_ACK))
604 		err = IXGBE_ERR_INVALID_ARGUMENT;
605 
606 mbx_err:
607 	return err;
608 }
609 
610 /**
611  * Hyper-V variant - just a stub.
612  * @hw: unused
613  * @vlan: unused
614  * @vind: unused
615  * @vlan_on: unused
616  */
ixgbevf_hv_set_vfta_vf(struct ixgbe_hw * hw,u32 vlan,u32 vind,bool vlan_on)617 static s32 ixgbevf_hv_set_vfta_vf(struct ixgbe_hw *hw, u32 vlan, u32 vind,
618 				  bool vlan_on)
619 {
620 	return -EOPNOTSUPP;
621 }
622 
623 /**
624  *  ixgbevf_setup_mac_link_vf - Setup MAC link settings
625  *  @hw: pointer to hardware structure
626  *  @speed: Unused in this implementation
627  *  @autoneg: Unused in this implementation
628  *  @autoneg_wait_to_complete: Unused in this implementation
629  *
630  *  Do nothing and return success.  VF drivers are not allowed to change
631  *  global settings.  Maintained for driver compatibility.
632  **/
ixgbevf_setup_mac_link_vf(struct ixgbe_hw * hw,ixgbe_link_speed speed,bool autoneg,bool autoneg_wait_to_complete)633 static s32 ixgbevf_setup_mac_link_vf(struct ixgbe_hw *hw,
634 				     ixgbe_link_speed speed, bool autoneg,
635 				     bool autoneg_wait_to_complete)
636 {
637 	return 0;
638 }
639 
640 /**
641  *  ixgbevf_check_mac_link_vf - Get link/speed status
642  *  @hw: pointer to hardware structure
643  *  @speed: pointer to link speed
644  *  @link_up: true is link is up, false otherwise
645  *  @autoneg_wait_to_complete: unused
646  *
647  *  Reads the links register to determine if link is up and the current speed
648  **/
ixgbevf_check_mac_link_vf(struct ixgbe_hw * hw,ixgbe_link_speed * speed,bool * link_up,bool autoneg_wait_to_complete)649 static s32 ixgbevf_check_mac_link_vf(struct ixgbe_hw *hw,
650 				     ixgbe_link_speed *speed,
651 				     bool *link_up,
652 				     bool autoneg_wait_to_complete)
653 {
654 	struct ixgbe_mbx_info *mbx = &hw->mbx;
655 	struct ixgbe_mac_info *mac = &hw->mac;
656 	s32 ret_val = 0;
657 	u32 links_reg;
658 	u32 in_msg = 0;
659 
660 	/* If we were hit with a reset drop the link */
661 	if (!mbx->ops.check_for_rst(hw) || !mbx->timeout)
662 		mac->get_link_status = true;
663 
664 	if (!mac->get_link_status)
665 		goto out;
666 
667 	/* if link status is down no point in checking to see if pf is up */
668 	links_reg = IXGBE_READ_REG(hw, IXGBE_VFLINKS);
669 	if (!(links_reg & IXGBE_LINKS_UP))
670 		goto out;
671 
672 	/* for SFP+ modules and DA cables on 82599 it can take up to 500usecs
673 	 * before the link status is correct
674 	 */
675 	if (mac->type == ixgbe_mac_82599_vf) {
676 		int i;
677 
678 		for (i = 0; i < 5; i++) {
679 			udelay(100);
680 			links_reg = IXGBE_READ_REG(hw, IXGBE_VFLINKS);
681 
682 			if (!(links_reg & IXGBE_LINKS_UP))
683 				goto out;
684 		}
685 	}
686 
687 	switch (links_reg & IXGBE_LINKS_SPEED_82599) {
688 	case IXGBE_LINKS_SPEED_10G_82599:
689 		*speed = IXGBE_LINK_SPEED_10GB_FULL;
690 		break;
691 	case IXGBE_LINKS_SPEED_1G_82599:
692 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
693 		break;
694 	case IXGBE_LINKS_SPEED_100_82599:
695 		*speed = IXGBE_LINK_SPEED_100_FULL;
696 		break;
697 	}
698 
699 	/* if the read failed it could just be a mailbox collision, best wait
700 	 * until we are called again and don't report an error
701 	 */
702 	if (mbx->ops.read(hw, &in_msg, 1))
703 		goto out;
704 
705 	if (!(in_msg & IXGBE_VT_MSGTYPE_CTS)) {
706 		/* msg is not CTS and is NACK we must have lost CTS status */
707 		if (in_msg & IXGBE_VT_MSGTYPE_NACK)
708 			ret_val = -1;
709 		goto out;
710 	}
711 
712 	/* the pf is talking, if we timed out in the past we reinit */
713 	if (!mbx->timeout) {
714 		ret_val = -1;
715 		goto out;
716 	}
717 
718 	/* if we passed all the tests above then the link is up and we no
719 	 * longer need to check for link
720 	 */
721 	mac->get_link_status = false;
722 
723 out:
724 	*link_up = !mac->get_link_status;
725 	return ret_val;
726 }
727 
728 /**
729  * Hyper-V variant; there is no mailbox communication.
730  * @hw: pointer to private hardware struct
731  * @speed: pointer to link speed
732  * @link_up: true is link is up, false otherwise
733  * @autoneg_wait_to_complete: unused
734  */
ixgbevf_hv_check_mac_link_vf(struct ixgbe_hw * hw,ixgbe_link_speed * speed,bool * link_up,bool autoneg_wait_to_complete)735 static s32 ixgbevf_hv_check_mac_link_vf(struct ixgbe_hw *hw,
736 					ixgbe_link_speed *speed,
737 					bool *link_up,
738 					bool autoneg_wait_to_complete)
739 {
740 	struct ixgbe_mbx_info *mbx = &hw->mbx;
741 	struct ixgbe_mac_info *mac = &hw->mac;
742 	u32 links_reg;
743 
744 	/* If we were hit with a reset drop the link */
745 	if (!mbx->ops.check_for_rst(hw) || !mbx->timeout)
746 		mac->get_link_status = true;
747 
748 	if (!mac->get_link_status)
749 		goto out;
750 
751 	/* if link status is down no point in checking to see if pf is up */
752 	links_reg = IXGBE_READ_REG(hw, IXGBE_VFLINKS);
753 	if (!(links_reg & IXGBE_LINKS_UP))
754 		goto out;
755 
756 	/* for SFP+ modules and DA cables on 82599 it can take up to 500usecs
757 	 * before the link status is correct
758 	 */
759 	if (mac->type == ixgbe_mac_82599_vf) {
760 		int i;
761 
762 		for (i = 0; i < 5; i++) {
763 			udelay(100);
764 			links_reg = IXGBE_READ_REG(hw, IXGBE_VFLINKS);
765 
766 			if (!(links_reg & IXGBE_LINKS_UP))
767 				goto out;
768 		}
769 	}
770 
771 	switch (links_reg & IXGBE_LINKS_SPEED_82599) {
772 	case IXGBE_LINKS_SPEED_10G_82599:
773 		*speed = IXGBE_LINK_SPEED_10GB_FULL;
774 		break;
775 	case IXGBE_LINKS_SPEED_1G_82599:
776 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
777 		break;
778 	case IXGBE_LINKS_SPEED_100_82599:
779 		*speed = IXGBE_LINK_SPEED_100_FULL;
780 		break;
781 	}
782 
783 	/* if we passed all the tests above then the link is up and we no
784 	 * longer need to check for link
785 	 */
786 	mac->get_link_status = false;
787 
788 out:
789 	*link_up = !mac->get_link_status;
790 	return 0;
791 }
792 
793 /**
794  *  ixgbevf_set_rlpml_vf - Set the maximum receive packet length
795  *  @hw: pointer to the HW structure
796  *  @max_size: value to assign to max frame size
797  **/
ixgbevf_set_rlpml_vf(struct ixgbe_hw * hw,u16 max_size)798 static s32 ixgbevf_set_rlpml_vf(struct ixgbe_hw *hw, u16 max_size)
799 {
800 	u32 msgbuf[2];
801 	s32 ret_val;
802 
803 	msgbuf[0] = IXGBE_VF_SET_LPE;
804 	msgbuf[1] = max_size;
805 
806 	ret_val = ixgbevf_write_msg_read_ack(hw, msgbuf, msgbuf,
807 					     ARRAY_SIZE(msgbuf));
808 	if (ret_val)
809 		return ret_val;
810 	if ((msgbuf[0] & IXGBE_VF_SET_LPE) &&
811 	    (msgbuf[0] & IXGBE_VT_MSGTYPE_NACK))
812 		return IXGBE_ERR_MBX;
813 
814 	return 0;
815 }
816 
817 /**
818  * ixgbevf_hv_set_rlpml_vf - Set the maximum receive packet length
819  * @hw: pointer to the HW structure
820  * @max_size: value to assign to max frame size
821  * Hyper-V variant.
822  **/
ixgbevf_hv_set_rlpml_vf(struct ixgbe_hw * hw,u16 max_size)823 static s32 ixgbevf_hv_set_rlpml_vf(struct ixgbe_hw *hw, u16 max_size)
824 {
825 	u32 reg;
826 
827 	/* If we are on Hyper-V, we implement this functionality
828 	 * differently.
829 	 */
830 	reg =  IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(0));
831 	/* CRC == 4 */
832 	reg |= ((max_size + 4) | IXGBE_RXDCTL_RLPML_EN);
833 	IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(0), reg);
834 
835 	return 0;
836 }
837 
838 /**
839  *  ixgbevf_negotiate_api_version_vf - Negotiate supported API version
840  *  @hw: pointer to the HW structure
841  *  @api: integer containing requested API version
842  **/
ixgbevf_negotiate_api_version_vf(struct ixgbe_hw * hw,int api)843 static int ixgbevf_negotiate_api_version_vf(struct ixgbe_hw *hw, int api)
844 {
845 	int err;
846 	u32 msg[3];
847 
848 	/* Negotiate the mailbox API version */
849 	msg[0] = IXGBE_VF_API_NEGOTIATE;
850 	msg[1] = api;
851 	msg[2] = 0;
852 
853 	err = ixgbevf_write_msg_read_ack(hw, msg, msg, ARRAY_SIZE(msg));
854 	if (!err) {
855 		msg[0] &= ~IXGBE_VT_MSGTYPE_CTS;
856 
857 		/* Store value and return 0 on success */
858 		if (msg[0] == (IXGBE_VF_API_NEGOTIATE | IXGBE_VT_MSGTYPE_ACK)) {
859 			hw->api_version = api;
860 			return 0;
861 		}
862 
863 		err = IXGBE_ERR_INVALID_ARGUMENT;
864 	}
865 
866 	return err;
867 }
868 
869 /**
870  *  ixgbevf_hv_negotiate_api_version_vf - Negotiate supported API version
871  *  @hw: pointer to the HW structure
872  *  @api: integer containing requested API version
873  *  Hyper-V version - only ixgbe_mbox_api_10 supported.
874  **/
ixgbevf_hv_negotiate_api_version_vf(struct ixgbe_hw * hw,int api)875 static int ixgbevf_hv_negotiate_api_version_vf(struct ixgbe_hw *hw, int api)
876 {
877 	/* Hyper-V only supports api version ixgbe_mbox_api_10 */
878 	if (api != ixgbe_mbox_api_10)
879 		return IXGBE_ERR_INVALID_ARGUMENT;
880 
881 	return 0;
882 }
883 
ixgbevf_get_queues(struct ixgbe_hw * hw,unsigned int * num_tcs,unsigned int * default_tc)884 int ixgbevf_get_queues(struct ixgbe_hw *hw, unsigned int *num_tcs,
885 		       unsigned int *default_tc)
886 {
887 	int err;
888 	u32 msg[5];
889 
890 	/* do nothing if API doesn't support ixgbevf_get_queues */
891 	switch (hw->api_version) {
892 	case ixgbe_mbox_api_11:
893 	case ixgbe_mbox_api_12:
894 	case ixgbe_mbox_api_13:
895 	case ixgbe_mbox_api_14:
896 		break;
897 	default:
898 		return 0;
899 	}
900 
901 	/* Fetch queue configuration from the PF */
902 	msg[0] = IXGBE_VF_GET_QUEUE;
903 	msg[1] = msg[2] = msg[3] = msg[4] = 0;
904 
905 	err = ixgbevf_write_msg_read_ack(hw, msg, msg, ARRAY_SIZE(msg));
906 	if (!err) {
907 		msg[0] &= ~IXGBE_VT_MSGTYPE_CTS;
908 
909 		/* if we we didn't get an ACK there must have been
910 		 * some sort of mailbox error so we should treat it
911 		 * as such
912 		 */
913 		if (msg[0] != (IXGBE_VF_GET_QUEUE | IXGBE_VT_MSGTYPE_ACK))
914 			return IXGBE_ERR_MBX;
915 
916 		/* record and validate values from message */
917 		hw->mac.max_tx_queues = msg[IXGBE_VF_TX_QUEUES];
918 		if (hw->mac.max_tx_queues == 0 ||
919 		    hw->mac.max_tx_queues > IXGBE_VF_MAX_TX_QUEUES)
920 			hw->mac.max_tx_queues = IXGBE_VF_MAX_TX_QUEUES;
921 
922 		hw->mac.max_rx_queues = msg[IXGBE_VF_RX_QUEUES];
923 		if (hw->mac.max_rx_queues == 0 ||
924 		    hw->mac.max_rx_queues > IXGBE_VF_MAX_RX_QUEUES)
925 			hw->mac.max_rx_queues = IXGBE_VF_MAX_RX_QUEUES;
926 
927 		*num_tcs = msg[IXGBE_VF_TRANS_VLAN];
928 		/* in case of unknown state assume we cannot tag frames */
929 		if (*num_tcs > hw->mac.max_rx_queues)
930 			*num_tcs = 1;
931 
932 		*default_tc = msg[IXGBE_VF_DEF_QUEUE];
933 		/* default to queue 0 on out-of-bounds queue number */
934 		if (*default_tc >= hw->mac.max_tx_queues)
935 			*default_tc = 0;
936 	}
937 
938 	return err;
939 }
940 
941 static const struct ixgbe_mac_operations ixgbevf_mac_ops = {
942 	.init_hw		= ixgbevf_init_hw_vf,
943 	.reset_hw		= ixgbevf_reset_hw_vf,
944 	.start_hw		= ixgbevf_start_hw_vf,
945 	.get_mac_addr		= ixgbevf_get_mac_addr_vf,
946 	.stop_adapter		= ixgbevf_stop_hw_vf,
947 	.setup_link		= ixgbevf_setup_mac_link_vf,
948 	.check_link		= ixgbevf_check_mac_link_vf,
949 	.negotiate_api_version	= ixgbevf_negotiate_api_version_vf,
950 	.set_rar		= ixgbevf_set_rar_vf,
951 	.update_mc_addr_list	= ixgbevf_update_mc_addr_list_vf,
952 	.update_xcast_mode	= ixgbevf_update_xcast_mode,
953 	.set_uc_addr		= ixgbevf_set_uc_addr_vf,
954 	.set_vfta		= ixgbevf_set_vfta_vf,
955 	.set_rlpml		= ixgbevf_set_rlpml_vf,
956 };
957 
958 static const struct ixgbe_mac_operations ixgbevf_hv_mac_ops = {
959 	.init_hw		= ixgbevf_init_hw_vf,
960 	.reset_hw		= ixgbevf_hv_reset_hw_vf,
961 	.start_hw		= ixgbevf_start_hw_vf,
962 	.get_mac_addr		= ixgbevf_get_mac_addr_vf,
963 	.stop_adapter		= ixgbevf_stop_hw_vf,
964 	.setup_link		= ixgbevf_setup_mac_link_vf,
965 	.check_link		= ixgbevf_hv_check_mac_link_vf,
966 	.negotiate_api_version	= ixgbevf_hv_negotiate_api_version_vf,
967 	.set_rar		= ixgbevf_hv_set_rar_vf,
968 	.update_mc_addr_list	= ixgbevf_hv_update_mc_addr_list_vf,
969 	.update_xcast_mode	= ixgbevf_hv_update_xcast_mode,
970 	.set_uc_addr		= ixgbevf_hv_set_uc_addr_vf,
971 	.set_vfta		= ixgbevf_hv_set_vfta_vf,
972 	.set_rlpml		= ixgbevf_hv_set_rlpml_vf,
973 };
974 
975 const struct ixgbevf_info ixgbevf_82599_vf_info = {
976 	.mac = ixgbe_mac_82599_vf,
977 	.mac_ops = &ixgbevf_mac_ops,
978 };
979 
980 const struct ixgbevf_info ixgbevf_82599_vf_hv_info = {
981 	.mac = ixgbe_mac_82599_vf,
982 	.mac_ops = &ixgbevf_hv_mac_ops,
983 };
984 
985 const struct ixgbevf_info ixgbevf_X540_vf_info = {
986 	.mac = ixgbe_mac_X540_vf,
987 	.mac_ops = &ixgbevf_mac_ops,
988 };
989 
990 const struct ixgbevf_info ixgbevf_X540_vf_hv_info = {
991 	.mac = ixgbe_mac_X540_vf,
992 	.mac_ops = &ixgbevf_hv_mac_ops,
993 };
994 
995 const struct ixgbevf_info ixgbevf_X550_vf_info = {
996 	.mac = ixgbe_mac_X550_vf,
997 	.mac_ops = &ixgbevf_mac_ops,
998 };
999 
1000 const struct ixgbevf_info ixgbevf_X550_vf_hv_info = {
1001 	.mac = ixgbe_mac_X550_vf,
1002 	.mac_ops = &ixgbevf_hv_mac_ops,
1003 };
1004 
1005 const struct ixgbevf_info ixgbevf_X550EM_x_vf_info = {
1006 	.mac = ixgbe_mac_X550EM_x_vf,
1007 	.mac_ops = &ixgbevf_mac_ops,
1008 };
1009 
1010 const struct ixgbevf_info ixgbevf_X550EM_x_vf_hv_info = {
1011 	.mac = ixgbe_mac_X550EM_x_vf,
1012 	.mac_ops = &ixgbevf_hv_mac_ops,
1013 };
1014 
1015 const struct ixgbevf_info ixgbevf_x550em_a_vf_info = {
1016 	.mac = ixgbe_mac_x550em_a_vf,
1017 	.mac_ops = &ixgbevf_mac_ops,
1018 };
1019