1 /*
2 * Copyright (c) 2009, Microsoft Corporation.
3 *
4 * This program is free software; you can redistribute it and/or modify it
5 * under the terms and conditions of the GNU General Public License,
6 * version 2, as published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope it will be useful, but WITHOUT
9 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
10 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
11 * more details.
12 *
13 * You should have received a copy of the GNU General Public License along with
14 * this program; if not, see <http://www.gnu.org/licenses/>.
15 *
16 * Authors:
17 * Haiyang Zhang <haiyangz@microsoft.com>
18 * Hank Janssen <hjanssen@microsoft.com>
19 */
20 #include <linux/kernel.h>
21 #include <linux/sched.h>
22 #include <linux/wait.h>
23 #include <linux/highmem.h>
24 #include <linux/slab.h>
25 #include <linux/io.h>
26 #include <linux/if_ether.h>
27 #include <linux/netdevice.h>
28 #include <linux/if_vlan.h>
29 #include <linux/nls.h>
30 #include <linux/vmalloc.h>
31 #include <linux/rtnetlink.h>
32 #include <linux/ucs2_string.h>
33
34 #include "hyperv_net.h"
35 #include "netvsc_trace.h"
36
37 static void rndis_set_multicast(struct work_struct *w);
38
39 #define RNDIS_EXT_LEN PAGE_SIZE
40 struct rndis_request {
41 struct list_head list_ent;
42 struct completion wait_event;
43
44 struct rndis_message response_msg;
45 /*
46 * The buffer for extended info after the RNDIS response message. It's
47 * referenced based on the data offset in the RNDIS message. Its size
48 * is enough for current needs, and should be sufficient for the near
49 * future.
50 */
51 u8 response_ext[RNDIS_EXT_LEN];
52
53 /* Simplify allocation by having a netvsc packet inline */
54 struct hv_netvsc_packet pkt;
55
56 struct rndis_message request_msg;
57 /*
58 * The buffer for the extended info after the RNDIS request message.
59 * It is referenced and sized in a similar way as response_ext.
60 */
61 u8 request_ext[RNDIS_EXT_LEN];
62 };
63
64 static const u8 netvsc_hash_key[NETVSC_HASH_KEYLEN] = {
65 0x6d, 0x5a, 0x56, 0xda, 0x25, 0x5b, 0x0e, 0xc2,
66 0x41, 0x67, 0x25, 0x3d, 0x43, 0xa3, 0x8f, 0xb0,
67 0xd0, 0xca, 0x2b, 0xcb, 0xae, 0x7b, 0x30, 0xb4,
68 0x77, 0xcb, 0x2d, 0xa3, 0x80, 0x30, 0xf2, 0x0c,
69 0x6a, 0x42, 0xb7, 0x3b, 0xbe, 0xac, 0x01, 0xfa
70 };
71
get_rndis_device(void)72 static struct rndis_device *get_rndis_device(void)
73 {
74 struct rndis_device *device;
75
76 device = kzalloc(sizeof(struct rndis_device), GFP_KERNEL);
77 if (!device)
78 return NULL;
79
80 spin_lock_init(&device->request_lock);
81
82 INIT_LIST_HEAD(&device->req_list);
83 INIT_WORK(&device->mcast_work, rndis_set_multicast);
84
85 device->state = RNDIS_DEV_UNINITIALIZED;
86
87 return device;
88 }
89
get_rndis_request(struct rndis_device * dev,u32 msg_type,u32 msg_len)90 static struct rndis_request *get_rndis_request(struct rndis_device *dev,
91 u32 msg_type,
92 u32 msg_len)
93 {
94 struct rndis_request *request;
95 struct rndis_message *rndis_msg;
96 struct rndis_set_request *set;
97 unsigned long flags;
98
99 request = kzalloc(sizeof(struct rndis_request), GFP_KERNEL);
100 if (!request)
101 return NULL;
102
103 init_completion(&request->wait_event);
104
105 rndis_msg = &request->request_msg;
106 rndis_msg->ndis_msg_type = msg_type;
107 rndis_msg->msg_len = msg_len;
108
109 request->pkt.q_idx = 0;
110
111 /*
112 * Set the request id. This field is always after the rndis header for
113 * request/response packet types so we just used the SetRequest as a
114 * template
115 */
116 set = &rndis_msg->msg.set_req;
117 set->req_id = atomic_inc_return(&dev->new_req_id);
118
119 /* Add to the request list */
120 spin_lock_irqsave(&dev->request_lock, flags);
121 list_add_tail(&request->list_ent, &dev->req_list);
122 spin_unlock_irqrestore(&dev->request_lock, flags);
123
124 return request;
125 }
126
put_rndis_request(struct rndis_device * dev,struct rndis_request * req)127 static void put_rndis_request(struct rndis_device *dev,
128 struct rndis_request *req)
129 {
130 unsigned long flags;
131
132 spin_lock_irqsave(&dev->request_lock, flags);
133 list_del(&req->list_ent);
134 spin_unlock_irqrestore(&dev->request_lock, flags);
135
136 kfree(req);
137 }
138
dump_rndis_message(struct net_device * netdev,const struct rndis_message * rndis_msg)139 static void dump_rndis_message(struct net_device *netdev,
140 const struct rndis_message *rndis_msg)
141 {
142 switch (rndis_msg->ndis_msg_type) {
143 case RNDIS_MSG_PACKET:
144 netdev_dbg(netdev, "RNDIS_MSG_PACKET (len %u, "
145 "data offset %u data len %u, # oob %u, "
146 "oob offset %u, oob len %u, pkt offset %u, "
147 "pkt len %u\n",
148 rndis_msg->msg_len,
149 rndis_msg->msg.pkt.data_offset,
150 rndis_msg->msg.pkt.data_len,
151 rndis_msg->msg.pkt.num_oob_data_elements,
152 rndis_msg->msg.pkt.oob_data_offset,
153 rndis_msg->msg.pkt.oob_data_len,
154 rndis_msg->msg.pkt.per_pkt_info_offset,
155 rndis_msg->msg.pkt.per_pkt_info_len);
156 break;
157
158 case RNDIS_MSG_INIT_C:
159 netdev_dbg(netdev, "RNDIS_MSG_INIT_C "
160 "(len %u, id 0x%x, status 0x%x, major %d, minor %d, "
161 "device flags %d, max xfer size 0x%x, max pkts %u, "
162 "pkt aligned %u)\n",
163 rndis_msg->msg_len,
164 rndis_msg->msg.init_complete.req_id,
165 rndis_msg->msg.init_complete.status,
166 rndis_msg->msg.init_complete.major_ver,
167 rndis_msg->msg.init_complete.minor_ver,
168 rndis_msg->msg.init_complete.dev_flags,
169 rndis_msg->msg.init_complete.max_xfer_size,
170 rndis_msg->msg.init_complete.
171 max_pkt_per_msg,
172 rndis_msg->msg.init_complete.
173 pkt_alignment_factor);
174 break;
175
176 case RNDIS_MSG_QUERY_C:
177 netdev_dbg(netdev, "RNDIS_MSG_QUERY_C "
178 "(len %u, id 0x%x, status 0x%x, buf len %u, "
179 "buf offset %u)\n",
180 rndis_msg->msg_len,
181 rndis_msg->msg.query_complete.req_id,
182 rndis_msg->msg.query_complete.status,
183 rndis_msg->msg.query_complete.
184 info_buflen,
185 rndis_msg->msg.query_complete.
186 info_buf_offset);
187 break;
188
189 case RNDIS_MSG_SET_C:
190 netdev_dbg(netdev,
191 "RNDIS_MSG_SET_C (len %u, id 0x%x, status 0x%x)\n",
192 rndis_msg->msg_len,
193 rndis_msg->msg.set_complete.req_id,
194 rndis_msg->msg.set_complete.status);
195 break;
196
197 case RNDIS_MSG_INDICATE:
198 netdev_dbg(netdev, "RNDIS_MSG_INDICATE "
199 "(len %u, status 0x%x, buf len %u, buf offset %u)\n",
200 rndis_msg->msg_len,
201 rndis_msg->msg.indicate_status.status,
202 rndis_msg->msg.indicate_status.status_buflen,
203 rndis_msg->msg.indicate_status.status_buf_offset);
204 break;
205
206 default:
207 netdev_dbg(netdev, "0x%x (len %u)\n",
208 rndis_msg->ndis_msg_type,
209 rndis_msg->msg_len);
210 break;
211 }
212 }
213
rndis_filter_send_request(struct rndis_device * dev,struct rndis_request * req)214 static int rndis_filter_send_request(struct rndis_device *dev,
215 struct rndis_request *req)
216 {
217 struct hv_netvsc_packet *packet;
218 struct hv_page_buffer page_buf[2];
219 struct hv_page_buffer *pb = page_buf;
220 int ret;
221
222 /* Setup the packet to send it */
223 packet = &req->pkt;
224
225 packet->total_data_buflen = req->request_msg.msg_len;
226 packet->page_buf_cnt = 1;
227
228 pb[0].pfn = virt_to_phys(&req->request_msg) >>
229 PAGE_SHIFT;
230 pb[0].len = req->request_msg.msg_len;
231 pb[0].offset =
232 (unsigned long)&req->request_msg & (PAGE_SIZE - 1);
233
234 /* Add one page_buf when request_msg crossing page boundary */
235 if (pb[0].offset + pb[0].len > PAGE_SIZE) {
236 packet->page_buf_cnt++;
237 pb[0].len = PAGE_SIZE -
238 pb[0].offset;
239 pb[1].pfn = virt_to_phys((void *)&req->request_msg
240 + pb[0].len) >> PAGE_SHIFT;
241 pb[1].offset = 0;
242 pb[1].len = req->request_msg.msg_len -
243 pb[0].len;
244 }
245
246 trace_rndis_send(dev->ndev, 0, &req->request_msg);
247
248 rcu_read_lock_bh();
249 ret = netvsc_send(dev->ndev, packet, NULL, pb, NULL);
250 rcu_read_unlock_bh();
251
252 return ret;
253 }
254
rndis_set_link_state(struct rndis_device * rdev,struct rndis_request * request)255 static void rndis_set_link_state(struct rndis_device *rdev,
256 struct rndis_request *request)
257 {
258 u32 link_status;
259 struct rndis_query_complete *query_complete;
260
261 query_complete = &request->response_msg.msg.query_complete;
262
263 if (query_complete->status == RNDIS_STATUS_SUCCESS &&
264 query_complete->info_buflen == sizeof(u32)) {
265 memcpy(&link_status, (void *)((unsigned long)query_complete +
266 query_complete->info_buf_offset), sizeof(u32));
267 rdev->link_state = link_status != 0;
268 }
269 }
270
rndis_filter_receive_response(struct net_device * ndev,struct netvsc_device * nvdev,const struct rndis_message * resp)271 static void rndis_filter_receive_response(struct net_device *ndev,
272 struct netvsc_device *nvdev,
273 const struct rndis_message *resp)
274 {
275 struct rndis_device *dev = nvdev->extension;
276 struct rndis_request *request = NULL;
277 bool found = false;
278 unsigned long flags;
279
280 /* This should never happen, it means control message
281 * response received after device removed.
282 */
283 if (dev->state == RNDIS_DEV_UNINITIALIZED) {
284 netdev_err(ndev,
285 "got rndis message uninitialized\n");
286 return;
287 }
288
289 spin_lock_irqsave(&dev->request_lock, flags);
290 list_for_each_entry(request, &dev->req_list, list_ent) {
291 /*
292 * All request/response message contains RequestId as the 1st
293 * field
294 */
295 if (request->request_msg.msg.init_req.req_id
296 == resp->msg.init_complete.req_id) {
297 found = true;
298 break;
299 }
300 }
301 spin_unlock_irqrestore(&dev->request_lock, flags);
302
303 if (found) {
304 if (resp->msg_len <=
305 sizeof(struct rndis_message) + RNDIS_EXT_LEN) {
306 memcpy(&request->response_msg, resp,
307 resp->msg_len);
308 if (request->request_msg.ndis_msg_type ==
309 RNDIS_MSG_QUERY && request->request_msg.msg.
310 query_req.oid == RNDIS_OID_GEN_MEDIA_CONNECT_STATUS)
311 rndis_set_link_state(dev, request);
312 } else {
313 netdev_err(ndev,
314 "rndis response buffer overflow "
315 "detected (size %u max %zu)\n",
316 resp->msg_len,
317 sizeof(struct rndis_message));
318
319 if (resp->ndis_msg_type ==
320 RNDIS_MSG_RESET_C) {
321 /* does not have a request id field */
322 request->response_msg.msg.reset_complete.
323 status = RNDIS_STATUS_BUFFER_OVERFLOW;
324 } else {
325 request->response_msg.msg.
326 init_complete.status =
327 RNDIS_STATUS_BUFFER_OVERFLOW;
328 }
329 }
330
331 complete(&request->wait_event);
332 } else {
333 netdev_err(ndev,
334 "no rndis request found for this response "
335 "(id 0x%x res type 0x%x)\n",
336 resp->msg.init_complete.req_id,
337 resp->ndis_msg_type);
338 }
339 }
340
341 /*
342 * Get the Per-Packet-Info with the specified type
343 * return NULL if not found.
344 */
rndis_get_ppi(struct rndis_packet * rpkt,u32 type)345 static inline void *rndis_get_ppi(struct rndis_packet *rpkt, u32 type)
346 {
347 struct rndis_per_packet_info *ppi;
348 int len;
349
350 if (rpkt->per_pkt_info_offset == 0)
351 return NULL;
352
353 ppi = (struct rndis_per_packet_info *)((ulong)rpkt +
354 rpkt->per_pkt_info_offset);
355 len = rpkt->per_pkt_info_len;
356
357 while (len > 0) {
358 if (ppi->type == type)
359 return (void *)((ulong)ppi + ppi->ppi_offset);
360 len -= ppi->size;
361 ppi = (struct rndis_per_packet_info *)((ulong)ppi + ppi->size);
362 }
363
364 return NULL;
365 }
366
rndis_filter_receive_data(struct net_device * ndev,struct netvsc_device * nvdev,struct vmbus_channel * channel,struct rndis_message * msg,u32 data_buflen)367 static int rndis_filter_receive_data(struct net_device *ndev,
368 struct netvsc_device *nvdev,
369 struct vmbus_channel *channel,
370 struct rndis_message *msg,
371 u32 data_buflen)
372 {
373 struct rndis_packet *rndis_pkt = &msg->msg.pkt;
374 const struct ndis_tcp_ip_checksum_info *csum_info;
375 const struct ndis_pkt_8021q_info *vlan;
376 u32 data_offset;
377 void *data;
378
379 /* Remove the rndis header and pass it back up the stack */
380 data_offset = RNDIS_HEADER_SIZE + rndis_pkt->data_offset;
381
382 data_buflen -= data_offset;
383
384 /*
385 * Make sure we got a valid RNDIS message, now total_data_buflen
386 * should be the data packet size plus the trailer padding size
387 */
388 if (unlikely(data_buflen < rndis_pkt->data_len)) {
389 netdev_err(ndev, "rndis message buffer "
390 "overflow detected (got %u, min %u)"
391 "...dropping this message!\n",
392 data_buflen, rndis_pkt->data_len);
393 return NVSP_STAT_FAIL;
394 }
395
396 vlan = rndis_get_ppi(rndis_pkt, IEEE_8021Q_INFO);
397
398 csum_info = rndis_get_ppi(rndis_pkt, TCPIP_CHKSUM_PKTINFO);
399
400 data = (void *)msg + data_offset;
401
402 /*
403 * Remove the rndis trailer padding from rndis packet message
404 * rndis_pkt->data_len tell us the real data length, we only copy
405 * the data packet to the stack, without the rndis trailer padding
406 */
407 return netvsc_recv_callback(ndev, nvdev, channel,
408 data, rndis_pkt->data_len,
409 csum_info, vlan);
410 }
411
rndis_filter_receive(struct net_device * ndev,struct netvsc_device * net_dev,struct vmbus_channel * channel,void * data,u32 buflen)412 int rndis_filter_receive(struct net_device *ndev,
413 struct netvsc_device *net_dev,
414 struct vmbus_channel *channel,
415 void *data, u32 buflen)
416 {
417 struct net_device_context *net_device_ctx = netdev_priv(ndev);
418 struct rndis_message *rndis_msg = data;
419
420 if (netif_msg_rx_status(net_device_ctx))
421 dump_rndis_message(ndev, rndis_msg);
422
423 switch (rndis_msg->ndis_msg_type) {
424 case RNDIS_MSG_PACKET:
425 return rndis_filter_receive_data(ndev, net_dev, channel,
426 rndis_msg, buflen);
427 case RNDIS_MSG_INIT_C:
428 case RNDIS_MSG_QUERY_C:
429 case RNDIS_MSG_SET_C:
430 /* completion msgs */
431 rndis_filter_receive_response(ndev, net_dev, rndis_msg);
432 break;
433
434 case RNDIS_MSG_INDICATE:
435 /* notification msgs */
436 netvsc_linkstatus_callback(ndev, rndis_msg);
437 break;
438 default:
439 netdev_err(ndev,
440 "unhandled rndis message (type %u len %u)\n",
441 rndis_msg->ndis_msg_type,
442 rndis_msg->msg_len);
443 return NVSP_STAT_FAIL;
444 }
445
446 return NVSP_STAT_SUCCESS;
447 }
448
rndis_filter_query_device(struct rndis_device * dev,struct netvsc_device * nvdev,u32 oid,void * result,u32 * result_size)449 static int rndis_filter_query_device(struct rndis_device *dev,
450 struct netvsc_device *nvdev,
451 u32 oid, void *result, u32 *result_size)
452 {
453 struct rndis_request *request;
454 u32 inresult_size = *result_size;
455 struct rndis_query_request *query;
456 struct rndis_query_complete *query_complete;
457 int ret = 0;
458
459 if (!result)
460 return -EINVAL;
461
462 *result_size = 0;
463 request = get_rndis_request(dev, RNDIS_MSG_QUERY,
464 RNDIS_MESSAGE_SIZE(struct rndis_query_request));
465 if (!request) {
466 ret = -ENOMEM;
467 goto cleanup;
468 }
469
470 /* Setup the rndis query */
471 query = &request->request_msg.msg.query_req;
472 query->oid = oid;
473 query->info_buf_offset = sizeof(struct rndis_query_request);
474 query->info_buflen = 0;
475 query->dev_vc_handle = 0;
476
477 if (oid == OID_TCP_OFFLOAD_HARDWARE_CAPABILITIES) {
478 struct ndis_offload *hwcaps;
479 u32 nvsp_version = nvdev->nvsp_version;
480 u8 ndis_rev;
481 size_t size;
482
483 if (nvsp_version >= NVSP_PROTOCOL_VERSION_5) {
484 ndis_rev = NDIS_OFFLOAD_PARAMETERS_REVISION_3;
485 size = NDIS_OFFLOAD_SIZE;
486 } else if (nvsp_version >= NVSP_PROTOCOL_VERSION_4) {
487 ndis_rev = NDIS_OFFLOAD_PARAMETERS_REVISION_2;
488 size = NDIS_OFFLOAD_SIZE_6_1;
489 } else {
490 ndis_rev = NDIS_OFFLOAD_PARAMETERS_REVISION_1;
491 size = NDIS_OFFLOAD_SIZE_6_0;
492 }
493
494 request->request_msg.msg_len += size;
495 query->info_buflen = size;
496 hwcaps = (struct ndis_offload *)
497 ((unsigned long)query + query->info_buf_offset);
498
499 hwcaps->header.type = NDIS_OBJECT_TYPE_OFFLOAD;
500 hwcaps->header.revision = ndis_rev;
501 hwcaps->header.size = size;
502
503 } else if (oid == OID_GEN_RECEIVE_SCALE_CAPABILITIES) {
504 struct ndis_recv_scale_cap *cap;
505
506 request->request_msg.msg_len +=
507 sizeof(struct ndis_recv_scale_cap);
508 query->info_buflen = sizeof(struct ndis_recv_scale_cap);
509 cap = (struct ndis_recv_scale_cap *)((unsigned long)query +
510 query->info_buf_offset);
511 cap->hdr.type = NDIS_OBJECT_TYPE_RSS_CAPABILITIES;
512 cap->hdr.rev = NDIS_RECEIVE_SCALE_CAPABILITIES_REVISION_2;
513 cap->hdr.size = sizeof(struct ndis_recv_scale_cap);
514 }
515
516 ret = rndis_filter_send_request(dev, request);
517 if (ret != 0)
518 goto cleanup;
519
520 wait_for_completion(&request->wait_event);
521
522 /* Copy the response back */
523 query_complete = &request->response_msg.msg.query_complete;
524
525 if (query_complete->info_buflen > inresult_size) {
526 ret = -1;
527 goto cleanup;
528 }
529
530 memcpy(result,
531 (void *)((unsigned long)query_complete +
532 query_complete->info_buf_offset),
533 query_complete->info_buflen);
534
535 *result_size = query_complete->info_buflen;
536
537 cleanup:
538 if (request)
539 put_rndis_request(dev, request);
540
541 return ret;
542 }
543
544 /* Get the hardware offload capabilities */
545 static int
rndis_query_hwcaps(struct rndis_device * dev,struct netvsc_device * net_device,struct ndis_offload * caps)546 rndis_query_hwcaps(struct rndis_device *dev, struct netvsc_device *net_device,
547 struct ndis_offload *caps)
548 {
549 u32 caps_len = sizeof(*caps);
550 int ret;
551
552 memset(caps, 0, sizeof(*caps));
553
554 ret = rndis_filter_query_device(dev, net_device,
555 OID_TCP_OFFLOAD_HARDWARE_CAPABILITIES,
556 caps, &caps_len);
557 if (ret)
558 return ret;
559
560 if (caps->header.type != NDIS_OBJECT_TYPE_OFFLOAD) {
561 netdev_warn(dev->ndev, "invalid NDIS objtype %#x\n",
562 caps->header.type);
563 return -EINVAL;
564 }
565
566 if (caps->header.revision < NDIS_OFFLOAD_PARAMETERS_REVISION_1) {
567 netdev_warn(dev->ndev, "invalid NDIS objrev %x\n",
568 caps->header.revision);
569 return -EINVAL;
570 }
571
572 if (caps->header.size > caps_len ||
573 caps->header.size < NDIS_OFFLOAD_SIZE_6_0) {
574 netdev_warn(dev->ndev,
575 "invalid NDIS objsize %u, data size %u\n",
576 caps->header.size, caps_len);
577 return -EINVAL;
578 }
579
580 return 0;
581 }
582
rndis_filter_query_device_mac(struct rndis_device * dev,struct netvsc_device * net_device)583 static int rndis_filter_query_device_mac(struct rndis_device *dev,
584 struct netvsc_device *net_device)
585 {
586 u32 size = ETH_ALEN;
587
588 return rndis_filter_query_device(dev, net_device,
589 RNDIS_OID_802_3_PERMANENT_ADDRESS,
590 dev->hw_mac_adr, &size);
591 }
592
593 #define NWADR_STR "NetworkAddress"
594 #define NWADR_STRLEN 14
595
rndis_filter_set_device_mac(struct netvsc_device * nvdev,const char * mac)596 int rndis_filter_set_device_mac(struct netvsc_device *nvdev,
597 const char *mac)
598 {
599 struct rndis_device *rdev = nvdev->extension;
600 struct rndis_request *request;
601 struct rndis_set_request *set;
602 struct rndis_config_parameter_info *cpi;
603 wchar_t *cfg_nwadr, *cfg_mac;
604 struct rndis_set_complete *set_complete;
605 char macstr[2*ETH_ALEN+1];
606 u32 extlen = sizeof(struct rndis_config_parameter_info) +
607 2*NWADR_STRLEN + 4*ETH_ALEN;
608 int ret;
609
610 request = get_rndis_request(rdev, RNDIS_MSG_SET,
611 RNDIS_MESSAGE_SIZE(struct rndis_set_request) + extlen);
612 if (!request)
613 return -ENOMEM;
614
615 set = &request->request_msg.msg.set_req;
616 set->oid = RNDIS_OID_GEN_RNDIS_CONFIG_PARAMETER;
617 set->info_buflen = extlen;
618 set->info_buf_offset = sizeof(struct rndis_set_request);
619 set->dev_vc_handle = 0;
620
621 cpi = (struct rndis_config_parameter_info *)((ulong)set +
622 set->info_buf_offset);
623 cpi->parameter_name_offset =
624 sizeof(struct rndis_config_parameter_info);
625 /* Multiply by 2 because host needs 2 bytes (utf16) for each char */
626 cpi->parameter_name_length = 2*NWADR_STRLEN;
627 cpi->parameter_type = RNDIS_CONFIG_PARAM_TYPE_STRING;
628 cpi->parameter_value_offset =
629 cpi->parameter_name_offset + cpi->parameter_name_length;
630 /* Multiply by 4 because each MAC byte displayed as 2 utf16 chars */
631 cpi->parameter_value_length = 4*ETH_ALEN;
632
633 cfg_nwadr = (wchar_t *)((ulong)cpi + cpi->parameter_name_offset);
634 cfg_mac = (wchar_t *)((ulong)cpi + cpi->parameter_value_offset);
635 ret = utf8s_to_utf16s(NWADR_STR, NWADR_STRLEN, UTF16_HOST_ENDIAN,
636 cfg_nwadr, NWADR_STRLEN);
637 if (ret < 0)
638 goto cleanup;
639 snprintf(macstr, 2*ETH_ALEN+1, "%pm", mac);
640 ret = utf8s_to_utf16s(macstr, 2*ETH_ALEN, UTF16_HOST_ENDIAN,
641 cfg_mac, 2*ETH_ALEN);
642 if (ret < 0)
643 goto cleanup;
644
645 ret = rndis_filter_send_request(rdev, request);
646 if (ret != 0)
647 goto cleanup;
648
649 wait_for_completion(&request->wait_event);
650
651 set_complete = &request->response_msg.msg.set_complete;
652 if (set_complete->status != RNDIS_STATUS_SUCCESS)
653 ret = -EIO;
654
655 cleanup:
656 put_rndis_request(rdev, request);
657 return ret;
658 }
659
660 static int
rndis_filter_set_offload_params(struct net_device * ndev,struct netvsc_device * nvdev,struct ndis_offload_params * req_offloads)661 rndis_filter_set_offload_params(struct net_device *ndev,
662 struct netvsc_device *nvdev,
663 struct ndis_offload_params *req_offloads)
664 {
665 struct rndis_device *rdev = nvdev->extension;
666 struct rndis_request *request;
667 struct rndis_set_request *set;
668 struct ndis_offload_params *offload_params;
669 struct rndis_set_complete *set_complete;
670 u32 extlen = sizeof(struct ndis_offload_params);
671 int ret;
672 u32 vsp_version = nvdev->nvsp_version;
673
674 if (vsp_version <= NVSP_PROTOCOL_VERSION_4) {
675 extlen = VERSION_4_OFFLOAD_SIZE;
676 /* On NVSP_PROTOCOL_VERSION_4 and below, we do not support
677 * UDP checksum offload.
678 */
679 req_offloads->udp_ip_v4_csum = 0;
680 req_offloads->udp_ip_v6_csum = 0;
681 }
682
683 request = get_rndis_request(rdev, RNDIS_MSG_SET,
684 RNDIS_MESSAGE_SIZE(struct rndis_set_request) + extlen);
685 if (!request)
686 return -ENOMEM;
687
688 set = &request->request_msg.msg.set_req;
689 set->oid = OID_TCP_OFFLOAD_PARAMETERS;
690 set->info_buflen = extlen;
691 set->info_buf_offset = sizeof(struct rndis_set_request);
692 set->dev_vc_handle = 0;
693
694 offload_params = (struct ndis_offload_params *)((ulong)set +
695 set->info_buf_offset);
696 *offload_params = *req_offloads;
697 offload_params->header.type = NDIS_OBJECT_TYPE_DEFAULT;
698 offload_params->header.revision = NDIS_OFFLOAD_PARAMETERS_REVISION_3;
699 offload_params->header.size = extlen;
700
701 ret = rndis_filter_send_request(rdev, request);
702 if (ret != 0)
703 goto cleanup;
704
705 wait_for_completion(&request->wait_event);
706 set_complete = &request->response_msg.msg.set_complete;
707 if (set_complete->status != RNDIS_STATUS_SUCCESS) {
708 netdev_err(ndev, "Fail to set offload on host side:0x%x\n",
709 set_complete->status);
710 ret = -EINVAL;
711 }
712
713 cleanup:
714 put_rndis_request(rdev, request);
715 return ret;
716 }
717
rndis_filter_set_rss_param(struct rndis_device * rdev,const u8 * rss_key)718 int rndis_filter_set_rss_param(struct rndis_device *rdev,
719 const u8 *rss_key)
720 {
721 struct net_device *ndev = rdev->ndev;
722 struct rndis_request *request;
723 struct rndis_set_request *set;
724 struct rndis_set_complete *set_complete;
725 u32 extlen = sizeof(struct ndis_recv_scale_param) +
726 4 * ITAB_NUM + NETVSC_HASH_KEYLEN;
727 struct ndis_recv_scale_param *rssp;
728 u32 *itab;
729 u8 *keyp;
730 int i, ret;
731
732 request = get_rndis_request(
733 rdev, RNDIS_MSG_SET,
734 RNDIS_MESSAGE_SIZE(struct rndis_set_request) + extlen);
735 if (!request)
736 return -ENOMEM;
737
738 set = &request->request_msg.msg.set_req;
739 set->oid = OID_GEN_RECEIVE_SCALE_PARAMETERS;
740 set->info_buflen = extlen;
741 set->info_buf_offset = sizeof(struct rndis_set_request);
742 set->dev_vc_handle = 0;
743
744 rssp = (struct ndis_recv_scale_param *)(set + 1);
745 rssp->hdr.type = NDIS_OBJECT_TYPE_RSS_PARAMETERS;
746 rssp->hdr.rev = NDIS_RECEIVE_SCALE_PARAMETERS_REVISION_2;
747 rssp->hdr.size = sizeof(struct ndis_recv_scale_param);
748 rssp->flag = 0;
749 rssp->hashinfo = NDIS_HASH_FUNC_TOEPLITZ | NDIS_HASH_IPV4 |
750 NDIS_HASH_TCP_IPV4 | NDIS_HASH_IPV6 |
751 NDIS_HASH_TCP_IPV6;
752 rssp->indirect_tabsize = 4*ITAB_NUM;
753 rssp->indirect_taboffset = sizeof(struct ndis_recv_scale_param);
754 rssp->hashkey_size = NETVSC_HASH_KEYLEN;
755 rssp->hashkey_offset = rssp->indirect_taboffset +
756 rssp->indirect_tabsize;
757
758 /* Set indirection table entries */
759 itab = (u32 *)(rssp + 1);
760 for (i = 0; i < ITAB_NUM; i++)
761 itab[i] = rdev->rx_table[i];
762
763 /* Set hask key values */
764 keyp = (u8 *)((unsigned long)rssp + rssp->hashkey_offset);
765 memcpy(keyp, rss_key, NETVSC_HASH_KEYLEN);
766
767 ret = rndis_filter_send_request(rdev, request);
768 if (ret != 0)
769 goto cleanup;
770
771 wait_for_completion(&request->wait_event);
772 set_complete = &request->response_msg.msg.set_complete;
773 if (set_complete->status == RNDIS_STATUS_SUCCESS)
774 memcpy(rdev->rss_key, rss_key, NETVSC_HASH_KEYLEN);
775 else {
776 netdev_err(ndev, "Fail to set RSS parameters:0x%x\n",
777 set_complete->status);
778 ret = -EINVAL;
779 }
780
781 cleanup:
782 put_rndis_request(rdev, request);
783 return ret;
784 }
785
rndis_filter_query_device_link_status(struct rndis_device * dev,struct netvsc_device * net_device)786 static int rndis_filter_query_device_link_status(struct rndis_device *dev,
787 struct netvsc_device *net_device)
788 {
789 u32 size = sizeof(u32);
790 u32 link_status;
791
792 return rndis_filter_query_device(dev, net_device,
793 RNDIS_OID_GEN_MEDIA_CONNECT_STATUS,
794 &link_status, &size);
795 }
796
rndis_filter_query_link_speed(struct rndis_device * dev,struct netvsc_device * net_device)797 static int rndis_filter_query_link_speed(struct rndis_device *dev,
798 struct netvsc_device *net_device)
799 {
800 u32 size = sizeof(u32);
801 u32 link_speed;
802 struct net_device_context *ndc;
803 int ret;
804
805 ret = rndis_filter_query_device(dev, net_device,
806 RNDIS_OID_GEN_LINK_SPEED,
807 &link_speed, &size);
808
809 if (!ret) {
810 ndc = netdev_priv(dev->ndev);
811
812 /* The link speed reported from host is in 100bps unit, so
813 * we convert it to Mbps here.
814 */
815 ndc->speed = link_speed / 10000;
816 }
817
818 return ret;
819 }
820
rndis_filter_set_packet_filter(struct rndis_device * dev,u32 new_filter)821 static int rndis_filter_set_packet_filter(struct rndis_device *dev,
822 u32 new_filter)
823 {
824 struct rndis_request *request;
825 struct rndis_set_request *set;
826 int ret;
827
828 if (dev->filter == new_filter)
829 return 0;
830
831 request = get_rndis_request(dev, RNDIS_MSG_SET,
832 RNDIS_MESSAGE_SIZE(struct rndis_set_request) +
833 sizeof(u32));
834 if (!request)
835 return -ENOMEM;
836
837 /* Setup the rndis set */
838 set = &request->request_msg.msg.set_req;
839 set->oid = RNDIS_OID_GEN_CURRENT_PACKET_FILTER;
840 set->info_buflen = sizeof(u32);
841 set->info_buf_offset = sizeof(struct rndis_set_request);
842
843 memcpy((void *)(unsigned long)set + sizeof(struct rndis_set_request),
844 &new_filter, sizeof(u32));
845
846 ret = rndis_filter_send_request(dev, request);
847 if (ret == 0) {
848 wait_for_completion(&request->wait_event);
849 dev->filter = new_filter;
850 }
851
852 put_rndis_request(dev, request);
853
854 return ret;
855 }
856
rndis_set_multicast(struct work_struct * w)857 static void rndis_set_multicast(struct work_struct *w)
858 {
859 struct rndis_device *rdev
860 = container_of(w, struct rndis_device, mcast_work);
861 u32 filter = NDIS_PACKET_TYPE_DIRECTED;
862 unsigned int flags = rdev->ndev->flags;
863
864 if (flags & IFF_PROMISC) {
865 filter = NDIS_PACKET_TYPE_PROMISCUOUS;
866 } else {
867 if (!netdev_mc_empty(rdev->ndev) || (flags & IFF_ALLMULTI))
868 filter |= NDIS_PACKET_TYPE_ALL_MULTICAST;
869 if (flags & IFF_BROADCAST)
870 filter |= NDIS_PACKET_TYPE_BROADCAST;
871 }
872
873 rndis_filter_set_packet_filter(rdev, filter);
874 }
875
rndis_filter_update(struct netvsc_device * nvdev)876 void rndis_filter_update(struct netvsc_device *nvdev)
877 {
878 struct rndis_device *rdev = nvdev->extension;
879
880 schedule_work(&rdev->mcast_work);
881 }
882
rndis_filter_init_device(struct rndis_device * dev,struct netvsc_device * nvdev)883 static int rndis_filter_init_device(struct rndis_device *dev,
884 struct netvsc_device *nvdev)
885 {
886 struct rndis_request *request;
887 struct rndis_initialize_request *init;
888 struct rndis_initialize_complete *init_complete;
889 u32 status;
890 int ret;
891
892 request = get_rndis_request(dev, RNDIS_MSG_INIT,
893 RNDIS_MESSAGE_SIZE(struct rndis_initialize_request));
894 if (!request) {
895 ret = -ENOMEM;
896 goto cleanup;
897 }
898
899 /* Setup the rndis set */
900 init = &request->request_msg.msg.init_req;
901 init->major_ver = RNDIS_MAJOR_VERSION;
902 init->minor_ver = RNDIS_MINOR_VERSION;
903 init->max_xfer_size = 0x4000;
904
905 dev->state = RNDIS_DEV_INITIALIZING;
906
907 ret = rndis_filter_send_request(dev, request);
908 if (ret != 0) {
909 dev->state = RNDIS_DEV_UNINITIALIZED;
910 goto cleanup;
911 }
912
913 wait_for_completion(&request->wait_event);
914
915 init_complete = &request->response_msg.msg.init_complete;
916 status = init_complete->status;
917 if (status == RNDIS_STATUS_SUCCESS) {
918 dev->state = RNDIS_DEV_INITIALIZED;
919 nvdev->max_pkt = init_complete->max_pkt_per_msg;
920 nvdev->pkt_align = 1 << init_complete->pkt_alignment_factor;
921 ret = 0;
922 } else {
923 dev->state = RNDIS_DEV_UNINITIALIZED;
924 ret = -EINVAL;
925 }
926
927 cleanup:
928 if (request)
929 put_rndis_request(dev, request);
930
931 return ret;
932 }
933
netvsc_device_idle(const struct netvsc_device * nvdev)934 static bool netvsc_device_idle(const struct netvsc_device *nvdev)
935 {
936 int i;
937
938 for (i = 0; i < nvdev->num_chn; i++) {
939 const struct netvsc_channel *nvchan = &nvdev->chan_table[i];
940
941 if (nvchan->mrc.first != nvchan->mrc.next)
942 return false;
943
944 if (atomic_read(&nvchan->queue_sends) > 0)
945 return false;
946 }
947
948 return true;
949 }
950
rndis_filter_halt_device(struct netvsc_device * nvdev,struct rndis_device * dev)951 static void rndis_filter_halt_device(struct netvsc_device *nvdev,
952 struct rndis_device *dev)
953 {
954 struct rndis_request *request;
955 struct rndis_halt_request *halt;
956
957 /* Attempt to do a rndis device halt */
958 request = get_rndis_request(dev, RNDIS_MSG_HALT,
959 RNDIS_MESSAGE_SIZE(struct rndis_halt_request));
960 if (!request)
961 goto cleanup;
962
963 /* Setup the rndis set */
964 halt = &request->request_msg.msg.halt_req;
965 halt->req_id = atomic_inc_return(&dev->new_req_id);
966
967 /* Ignore return since this msg is optional. */
968 rndis_filter_send_request(dev, request);
969
970 dev->state = RNDIS_DEV_UNINITIALIZED;
971
972 cleanup:
973 nvdev->destroy = true;
974
975 /* Force flag to be ordered before waiting */
976 wmb();
977
978 /* Wait for all send completions */
979 wait_event(nvdev->wait_drain, netvsc_device_idle(nvdev));
980
981 if (request)
982 put_rndis_request(dev, request);
983 }
984
rndis_filter_open_device(struct rndis_device * dev)985 static int rndis_filter_open_device(struct rndis_device *dev)
986 {
987 int ret;
988
989 if (dev->state != RNDIS_DEV_INITIALIZED)
990 return 0;
991
992 ret = rndis_filter_set_packet_filter(dev,
993 NDIS_PACKET_TYPE_BROADCAST |
994 NDIS_PACKET_TYPE_ALL_MULTICAST |
995 NDIS_PACKET_TYPE_DIRECTED);
996 if (ret == 0)
997 dev->state = RNDIS_DEV_DATAINITIALIZED;
998
999 return ret;
1000 }
1001
rndis_filter_close_device(struct rndis_device * dev)1002 static int rndis_filter_close_device(struct rndis_device *dev)
1003 {
1004 int ret;
1005
1006 if (dev->state != RNDIS_DEV_DATAINITIALIZED)
1007 return 0;
1008
1009 /* Make sure rndis_set_multicast doesn't re-enable filter! */
1010 cancel_work_sync(&dev->mcast_work);
1011
1012 ret = rndis_filter_set_packet_filter(dev, 0);
1013 if (ret == -ENODEV)
1014 ret = 0;
1015
1016 if (ret == 0)
1017 dev->state = RNDIS_DEV_INITIALIZED;
1018
1019 return ret;
1020 }
1021
netvsc_sc_open(struct vmbus_channel * new_sc)1022 static void netvsc_sc_open(struct vmbus_channel *new_sc)
1023 {
1024 struct net_device *ndev =
1025 hv_get_drvdata(new_sc->primary_channel->device_obj);
1026 struct net_device_context *ndev_ctx = netdev_priv(ndev);
1027 struct netvsc_device *nvscdev;
1028 u16 chn_index = new_sc->offermsg.offer.sub_channel_index;
1029 struct netvsc_channel *nvchan;
1030 int ret;
1031
1032 /* This is safe because this callback only happens when
1033 * new device is being setup and waiting on the channel_init_wait.
1034 */
1035 nvscdev = rcu_dereference_raw(ndev_ctx->nvdev);
1036 if (!nvscdev || chn_index >= nvscdev->num_chn)
1037 return;
1038
1039 nvchan = nvscdev->chan_table + chn_index;
1040
1041 /* Because the device uses NAPI, all the interrupt batching and
1042 * control is done via Net softirq, not the channel handling
1043 */
1044 set_channel_read_mode(new_sc, HV_CALL_ISR);
1045
1046 /* Set the channel before opening.*/
1047 nvchan->channel = new_sc;
1048
1049 ret = vmbus_open(new_sc, netvsc_ring_bytes,
1050 netvsc_ring_bytes, NULL, 0,
1051 netvsc_channel_cb, nvchan);
1052 if (ret == 0)
1053 napi_enable(&nvchan->napi);
1054 else
1055 netdev_notice(ndev, "sub channel open failed: %d\n", ret);
1056
1057 if (atomic_inc_return(&nvscdev->open_chn) == nvscdev->num_chn)
1058 wake_up(&nvscdev->subchan_open);
1059 }
1060
1061 /* Open sub-channels after completing the handling of the device probe.
1062 * This breaks overlap of processing the host message for the
1063 * new primary channel with the initialization of sub-channels.
1064 */
rndis_set_subchannel(struct net_device * ndev,struct netvsc_device * nvdev)1065 int rndis_set_subchannel(struct net_device *ndev, struct netvsc_device *nvdev)
1066 {
1067 struct nvsp_message *init_packet = &nvdev->channel_init_pkt;
1068 struct net_device_context *ndev_ctx = netdev_priv(ndev);
1069 struct hv_device *hv_dev = ndev_ctx->device_ctx;
1070 struct rndis_device *rdev = nvdev->extension;
1071 int i, ret;
1072
1073 ASSERT_RTNL();
1074
1075 memset(init_packet, 0, sizeof(struct nvsp_message));
1076 init_packet->hdr.msg_type = NVSP_MSG5_TYPE_SUBCHANNEL;
1077 init_packet->msg.v5_msg.subchn_req.op = NVSP_SUBCHANNEL_ALLOCATE;
1078 init_packet->msg.v5_msg.subchn_req.num_subchannels =
1079 nvdev->num_chn - 1;
1080 trace_nvsp_send(ndev, init_packet);
1081
1082 ret = vmbus_sendpacket(hv_dev->channel, init_packet,
1083 sizeof(struct nvsp_message),
1084 (unsigned long)init_packet,
1085 VM_PKT_DATA_INBAND,
1086 VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED);
1087 if (ret) {
1088 netdev_err(ndev, "sub channel allocate send failed: %d\n", ret);
1089 return ret;
1090 }
1091
1092 wait_for_completion(&nvdev->channel_init_wait);
1093 if (init_packet->msg.v5_msg.subchn_comp.status != NVSP_STAT_SUCCESS) {
1094 netdev_err(ndev, "sub channel request failed\n");
1095 return -EIO;
1096 }
1097
1098 nvdev->num_chn = 1 +
1099 init_packet->msg.v5_msg.subchn_comp.num_subchannels;
1100
1101 /* wait for all sub channels to open */
1102 wait_event(nvdev->subchan_open,
1103 atomic_read(&nvdev->open_chn) == nvdev->num_chn);
1104
1105 /* ignore failues from setting rss parameters, still have channels */
1106 rndis_filter_set_rss_param(rdev, netvsc_hash_key);
1107
1108 netif_set_real_num_tx_queues(ndev, nvdev->num_chn);
1109 netif_set_real_num_rx_queues(ndev, nvdev->num_chn);
1110
1111 for (i = 0; i < VRSS_SEND_TAB_SIZE; i++)
1112 ndev_ctx->tx_table[i] = i % nvdev->num_chn;
1113
1114 return 0;
1115 }
1116
rndis_netdev_set_hwcaps(struct rndis_device * rndis_device,struct netvsc_device * nvdev)1117 static int rndis_netdev_set_hwcaps(struct rndis_device *rndis_device,
1118 struct netvsc_device *nvdev)
1119 {
1120 struct net_device *net = rndis_device->ndev;
1121 struct net_device_context *net_device_ctx = netdev_priv(net);
1122 struct ndis_offload hwcaps;
1123 struct ndis_offload_params offloads;
1124 unsigned int gso_max_size = GSO_MAX_SIZE;
1125 int ret;
1126
1127 /* Find HW offload capabilities */
1128 ret = rndis_query_hwcaps(rndis_device, nvdev, &hwcaps);
1129 if (ret != 0)
1130 return ret;
1131
1132 /* A value of zero means "no change"; now turn on what we want. */
1133 memset(&offloads, 0, sizeof(struct ndis_offload_params));
1134
1135 /* Linux does not care about IP checksum, always does in kernel */
1136 offloads.ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_DISABLED;
1137
1138 /* Reset previously set hw_features flags */
1139 net->hw_features &= ~NETVSC_SUPPORTED_HW_FEATURES;
1140 net_device_ctx->tx_checksum_mask = 0;
1141
1142 /* Compute tx offload settings based on hw capabilities */
1143 net->hw_features |= NETIF_F_RXCSUM;
1144
1145 if ((hwcaps.csum.ip4_txcsum & NDIS_TXCSUM_ALL_TCP4) == NDIS_TXCSUM_ALL_TCP4) {
1146 /* Can checksum TCP */
1147 net->hw_features |= NETIF_F_IP_CSUM;
1148 net_device_ctx->tx_checksum_mask |= TRANSPORT_INFO_IPV4_TCP;
1149
1150 offloads.tcp_ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED;
1151
1152 if (hwcaps.lsov2.ip4_encap & NDIS_OFFLOAD_ENCAP_8023) {
1153 offloads.lso_v2_ipv4 = NDIS_OFFLOAD_PARAMETERS_LSOV2_ENABLED;
1154 net->hw_features |= NETIF_F_TSO;
1155
1156 if (hwcaps.lsov2.ip4_maxsz < gso_max_size)
1157 gso_max_size = hwcaps.lsov2.ip4_maxsz;
1158 }
1159
1160 if (hwcaps.csum.ip4_txcsum & NDIS_TXCSUM_CAP_UDP4) {
1161 offloads.udp_ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED;
1162 net_device_ctx->tx_checksum_mask |= TRANSPORT_INFO_IPV4_UDP;
1163 }
1164 }
1165
1166 if ((hwcaps.csum.ip6_txcsum & NDIS_TXCSUM_ALL_TCP6) == NDIS_TXCSUM_ALL_TCP6) {
1167 net->hw_features |= NETIF_F_IPV6_CSUM;
1168
1169 offloads.tcp_ip_v6_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED;
1170 net_device_ctx->tx_checksum_mask |= TRANSPORT_INFO_IPV6_TCP;
1171
1172 if ((hwcaps.lsov2.ip6_encap & NDIS_OFFLOAD_ENCAP_8023) &&
1173 (hwcaps.lsov2.ip6_opts & NDIS_LSOV2_CAP_IP6) == NDIS_LSOV2_CAP_IP6) {
1174 offloads.lso_v2_ipv6 = NDIS_OFFLOAD_PARAMETERS_LSOV2_ENABLED;
1175 net->hw_features |= NETIF_F_TSO6;
1176
1177 if (hwcaps.lsov2.ip6_maxsz < gso_max_size)
1178 gso_max_size = hwcaps.lsov2.ip6_maxsz;
1179 }
1180
1181 if (hwcaps.csum.ip6_txcsum & NDIS_TXCSUM_CAP_UDP6) {
1182 offloads.udp_ip_v6_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED;
1183 net_device_ctx->tx_checksum_mask |= TRANSPORT_INFO_IPV6_UDP;
1184 }
1185 }
1186
1187 /* In case some hw_features disappeared we need to remove them from
1188 * net->features list as they're no longer supported.
1189 */
1190 net->features &= ~NETVSC_SUPPORTED_HW_FEATURES | net->hw_features;
1191
1192 netif_set_gso_max_size(net, gso_max_size);
1193
1194 ret = rndis_filter_set_offload_params(net, nvdev, &offloads);
1195
1196 return ret;
1197 }
1198
rndis_get_friendly_name(struct net_device * net,struct rndis_device * rndis_device,struct netvsc_device * net_device)1199 static void rndis_get_friendly_name(struct net_device *net,
1200 struct rndis_device *rndis_device,
1201 struct netvsc_device *net_device)
1202 {
1203 ucs2_char_t wname[256];
1204 unsigned long len;
1205 u8 ifalias[256];
1206 u32 size;
1207
1208 size = sizeof(wname);
1209 if (rndis_filter_query_device(rndis_device, net_device,
1210 RNDIS_OID_GEN_FRIENDLY_NAME,
1211 wname, &size) != 0)
1212 return; /* ignore if host does not support */
1213
1214 if (size == 0)
1215 return; /* name not set */
1216
1217 /* Convert Windows Unicode string to UTF-8 */
1218 len = ucs2_as_utf8(ifalias, wname, sizeof(ifalias));
1219
1220 /* ignore the default value from host */
1221 if (strcmp(ifalias, "Network Adapter") != 0)
1222 dev_set_alias(net, ifalias, len);
1223 }
1224
rndis_filter_device_add(struct hv_device * dev,struct netvsc_device_info * device_info)1225 struct netvsc_device *rndis_filter_device_add(struct hv_device *dev,
1226 struct netvsc_device_info *device_info)
1227 {
1228 struct net_device *net = hv_get_drvdata(dev);
1229 struct netvsc_device *net_device;
1230 struct rndis_device *rndis_device;
1231 struct ndis_recv_scale_cap rsscap;
1232 u32 rsscap_size = sizeof(struct ndis_recv_scale_cap);
1233 u32 mtu, size;
1234 u32 num_possible_rss_qs;
1235 int i, ret;
1236
1237 rndis_device = get_rndis_device();
1238 if (!rndis_device)
1239 return ERR_PTR(-ENODEV);
1240
1241 /* Let the inner driver handle this first to create the netvsc channel
1242 * NOTE! Once the channel is created, we may get a receive callback
1243 * (RndisFilterOnReceive()) before this call is completed
1244 */
1245 net_device = netvsc_device_add(dev, device_info);
1246 if (IS_ERR(net_device)) {
1247 kfree(rndis_device);
1248 return net_device;
1249 }
1250
1251 /* Initialize the rndis device */
1252 net_device->max_chn = 1;
1253 net_device->num_chn = 1;
1254
1255 net_device->extension = rndis_device;
1256 rndis_device->ndev = net;
1257
1258 /* Send the rndis initialization message */
1259 ret = rndis_filter_init_device(rndis_device, net_device);
1260 if (ret != 0)
1261 goto err_dev_remv;
1262
1263 /* Get the MTU from the host */
1264 size = sizeof(u32);
1265 ret = rndis_filter_query_device(rndis_device, net_device,
1266 RNDIS_OID_GEN_MAXIMUM_FRAME_SIZE,
1267 &mtu, &size);
1268 if (ret == 0 && size == sizeof(u32) && mtu < net->mtu)
1269 net->mtu = mtu;
1270
1271 /* Get the mac address */
1272 ret = rndis_filter_query_device_mac(rndis_device, net_device);
1273 if (ret != 0)
1274 goto err_dev_remv;
1275
1276 memcpy(device_info->mac_adr, rndis_device->hw_mac_adr, ETH_ALEN);
1277
1278 /* Get friendly name as ifalias*/
1279 if (!net->ifalias)
1280 rndis_get_friendly_name(net, rndis_device, net_device);
1281
1282 /* Query and set hardware capabilities */
1283 ret = rndis_netdev_set_hwcaps(rndis_device, net_device);
1284 if (ret != 0)
1285 goto err_dev_remv;
1286
1287 rndis_filter_query_device_link_status(rndis_device, net_device);
1288
1289 netdev_dbg(net, "Device MAC %pM link state %s\n",
1290 rndis_device->hw_mac_adr,
1291 rndis_device->link_state ? "down" : "up");
1292
1293 if (net_device->nvsp_version < NVSP_PROTOCOL_VERSION_5)
1294 goto out;
1295
1296 rndis_filter_query_link_speed(rndis_device, net_device);
1297
1298 /* vRSS setup */
1299 memset(&rsscap, 0, rsscap_size);
1300 ret = rndis_filter_query_device(rndis_device, net_device,
1301 OID_GEN_RECEIVE_SCALE_CAPABILITIES,
1302 &rsscap, &rsscap_size);
1303 if (ret || rsscap.num_recv_que < 2)
1304 goto out;
1305
1306 /* This guarantees that num_possible_rss_qs <= num_online_cpus */
1307 num_possible_rss_qs = min_t(u32, num_online_cpus(),
1308 rsscap.num_recv_que);
1309
1310 net_device->max_chn = min_t(u32, VRSS_CHANNEL_MAX, num_possible_rss_qs);
1311
1312 /* We will use the given number of channels if available. */
1313 net_device->num_chn = min(net_device->max_chn, device_info->num_chn);
1314
1315 for (i = 0; i < ITAB_NUM; i++)
1316 rndis_device->rx_table[i] = ethtool_rxfh_indir_default(
1317 i, net_device->num_chn);
1318
1319 atomic_set(&net_device->open_chn, 1);
1320 vmbus_set_sc_create_callback(dev->channel, netvsc_sc_open);
1321
1322 for (i = 1; i < net_device->num_chn; i++) {
1323 ret = netvsc_alloc_recv_comp_ring(net_device, i);
1324 if (ret) {
1325 while (--i != 0)
1326 vfree(net_device->chan_table[i].mrc.slots);
1327 goto out;
1328 }
1329 }
1330
1331 for (i = 1; i < net_device->num_chn; i++)
1332 netif_napi_add(net, &net_device->chan_table[i].napi,
1333 netvsc_poll, NAPI_POLL_WEIGHT);
1334
1335 return net_device;
1336
1337 out:
1338 /* setting up multiple channels failed */
1339 net_device->max_chn = 1;
1340 net_device->num_chn = 1;
1341 return net_device;
1342
1343 err_dev_remv:
1344 rndis_filter_device_remove(dev, net_device);
1345 return ERR_PTR(ret);
1346 }
1347
rndis_filter_device_remove(struct hv_device * dev,struct netvsc_device * net_dev)1348 void rndis_filter_device_remove(struct hv_device *dev,
1349 struct netvsc_device *net_dev)
1350 {
1351 struct rndis_device *rndis_dev = net_dev->extension;
1352
1353 /* Halt and release the rndis device */
1354 rndis_filter_halt_device(net_dev, rndis_dev);
1355
1356 net_dev->extension = NULL;
1357
1358 netvsc_device_remove(dev);
1359 }
1360
rndis_filter_open(struct netvsc_device * nvdev)1361 int rndis_filter_open(struct netvsc_device *nvdev)
1362 {
1363 if (!nvdev)
1364 return -EINVAL;
1365
1366 return rndis_filter_open_device(nvdev->extension);
1367 }
1368
rndis_filter_close(struct netvsc_device * nvdev)1369 int rndis_filter_close(struct netvsc_device *nvdev)
1370 {
1371 if (!nvdev)
1372 return -EINVAL;
1373
1374 return rndis_filter_close_device(nvdev->extension);
1375 }
1376