1 // SPDX-License-Identifier: LGPL-2.1
2 /*
3 *
4 * Copyright (C) International Business Machines Corp., 2002,2011
5 * Author(s): Steve French (sfrench@us.ibm.com)
6 *
7 */
8 #include <linux/fs.h>
9 #include <linux/net.h>
10 #include <linux/string.h>
11 #include <linux/sched/mm.h>
12 #include <linux/sched/signal.h>
13 #include <linux/list.h>
14 #include <linux/wait.h>
15 #include <linux/slab.h>
16 #include <linux/pagemap.h>
17 #include <linux/ctype.h>
18 #include <linux/utsname.h>
19 #include <linux/mempool.h>
20 #include <linux/delay.h>
21 #include <linux/completion.h>
22 #include <linux/kthread.h>
23 #include <linux/pagevec.h>
24 #include <linux/freezer.h>
25 #include <linux/namei.h>
26 #include <linux/uuid.h>
27 #include <linux/uaccess.h>
28 #include <asm/processor.h>
29 #include <linux/inet.h>
30 #include <linux/module.h>
31 #include <keys/user-type.h>
32 #include <net/ipv6.h>
33 #include <linux/parser.h>
34 #include <linux/bvec.h>
35 #include "cifspdu.h"
36 #include "cifsglob.h"
37 #include "cifsproto.h"
38 #include "cifs_unicode.h"
39 #include "cifs_debug.h"
40 #include "cifs_fs_sb.h"
41 #include "ntlmssp.h"
42 #include "nterr.h"
43 #include "rfc1002pdu.h"
44 #include "fscache.h"
45 #include "smb2proto.h"
46 #include "smbdirect.h"
47 #include "dns_resolve.h"
48 #ifdef CONFIG_CIFS_DFS_UPCALL
49 #include "dfs_cache.h"
50 #endif
51 #include "fs_context.h"
52 #include "cifs_swn.h"
53
54 extern mempool_t *cifs_req_poolp;
55 extern bool disable_legacy_dialects;
56
57 /* FIXME: should these be tunable? */
58 #define TLINK_ERROR_EXPIRE (1 * HZ)
59 #define TLINK_IDLE_EXPIRE (600 * HZ)
60
61 /* Drop the connection to not overload the server */
62 #define NUM_STATUS_IO_TIMEOUT 5
63
64 struct mount_ctx {
65 struct cifs_sb_info *cifs_sb;
66 struct smb3_fs_context *fs_ctx;
67 unsigned int xid;
68 struct TCP_Server_Info *server;
69 struct cifs_ses *ses;
70 struct cifs_tcon *tcon;
71 #ifdef CONFIG_CIFS_DFS_UPCALL
72 struct cifs_ses *root_ses;
73 uuid_t mount_id;
74 char *origin_fullpath, *leaf_fullpath;
75 #endif
76 };
77
78 static int ip_connect(struct TCP_Server_Info *server);
79 static int generic_ip_connect(struct TCP_Server_Info *server);
80 static void tlink_rb_insert(struct rb_root *root, struct tcon_link *new_tlink);
81 static void cifs_prune_tlinks(struct work_struct *work);
82
83 /*
84 * Resolve hostname and set ip addr in tcp ses. Useful for hostnames that may
85 * get their ip addresses changed at some point.
86 *
87 * This should be called with server->srv_mutex held.
88 */
reconn_set_ipaddr_from_hostname(struct TCP_Server_Info * server)89 static int reconn_set_ipaddr_from_hostname(struct TCP_Server_Info *server)
90 {
91 int rc;
92 int len;
93 char *unc, *ipaddr = NULL;
94 time64_t expiry, now;
95 unsigned long ttl = SMB_DNS_RESOLVE_INTERVAL_DEFAULT;
96
97 if (!server->hostname)
98 return -EINVAL;
99
100 /* if server hostname isn't populated, there's nothing to do here */
101 if (server->hostname[0] == '\0')
102 return 0;
103
104 len = strlen(server->hostname) + 3;
105
106 unc = kmalloc(len, GFP_KERNEL);
107 if (!unc) {
108 cifs_dbg(FYI, "%s: failed to create UNC path\n", __func__);
109 return -ENOMEM;
110 }
111 scnprintf(unc, len, "\\\\%s", server->hostname);
112
113 rc = dns_resolve_server_name_to_ip(unc, &ipaddr, &expiry);
114 kfree(unc);
115
116 if (rc < 0) {
117 cifs_dbg(FYI, "%s: failed to resolve server part of %s to IP: %d\n",
118 __func__, server->hostname, rc);
119 goto requeue_resolve;
120 }
121
122 spin_lock(&server->srv_lock);
123 rc = cifs_convert_address((struct sockaddr *)&server->dstaddr, ipaddr,
124 strlen(ipaddr));
125 spin_unlock(&server->srv_lock);
126 kfree(ipaddr);
127
128 /* rc == 1 means success here */
129 if (rc) {
130 now = ktime_get_real_seconds();
131 if (expiry && expiry > now)
132 /*
133 * To make sure we don't use the cached entry, retry 1s
134 * after expiry.
135 */
136 ttl = max_t(unsigned long, expiry - now, SMB_DNS_RESOLVE_INTERVAL_MIN) + 1;
137 }
138 rc = !rc ? -1 : 0;
139
140 requeue_resolve:
141 cifs_dbg(FYI, "%s: next dns resolution scheduled for %lu seconds in the future\n",
142 __func__, ttl);
143 mod_delayed_work(cifsiod_wq, &server->resolve, (ttl * HZ));
144
145 return rc;
146 }
147
smb2_query_server_interfaces(struct work_struct * work)148 static void smb2_query_server_interfaces(struct work_struct *work)
149 {
150 int rc;
151 struct cifs_tcon *tcon = container_of(work,
152 struct cifs_tcon,
153 query_interfaces.work);
154
155 /*
156 * query server network interfaces, in case they change
157 */
158 rc = SMB3_request_interfaces(0, tcon, false);
159 if (rc) {
160 cifs_dbg(FYI, "%s: failed to query server interfaces: %d\n",
161 __func__, rc);
162 }
163
164 queue_delayed_work(cifsiod_wq, &tcon->query_interfaces,
165 (SMB_INTERFACE_POLL_INTERVAL * HZ));
166 }
167
cifs_resolve_server(struct work_struct * work)168 static void cifs_resolve_server(struct work_struct *work)
169 {
170 int rc;
171 struct TCP_Server_Info *server = container_of(work,
172 struct TCP_Server_Info, resolve.work);
173
174 cifs_server_lock(server);
175
176 /*
177 * Resolve the hostname again to make sure that IP address is up-to-date.
178 */
179 rc = reconn_set_ipaddr_from_hostname(server);
180 if (rc) {
181 cifs_dbg(FYI, "%s: failed to resolve hostname: %d\n",
182 __func__, rc);
183 }
184
185 cifs_server_unlock(server);
186 }
187
188 /*
189 * Update the tcpStatus for the server.
190 * This is used to signal the cifsd thread to call cifs_reconnect
191 * ONLY cifsd thread should call cifs_reconnect. For any other
192 * thread, use this function
193 *
194 * @server: the tcp ses for which reconnect is needed
195 * @all_channels: if this needs to be done for all channels
196 */
197 void
cifs_signal_cifsd_for_reconnect(struct TCP_Server_Info * server,bool all_channels)198 cifs_signal_cifsd_for_reconnect(struct TCP_Server_Info *server,
199 bool all_channels)
200 {
201 struct TCP_Server_Info *pserver;
202 struct cifs_ses *ses;
203 int i;
204
205 /* If server is a channel, select the primary channel */
206 pserver = CIFS_SERVER_IS_CHAN(server) ? server->primary_server : server;
207
208 spin_lock(&pserver->srv_lock);
209 if (!all_channels) {
210 pserver->tcpStatus = CifsNeedReconnect;
211 spin_unlock(&pserver->srv_lock);
212 return;
213 }
214 spin_unlock(&pserver->srv_lock);
215
216 spin_lock(&cifs_tcp_ses_lock);
217 list_for_each_entry(ses, &pserver->smb_ses_list, smb_ses_list) {
218 spin_lock(&ses->chan_lock);
219 for (i = 0; i < ses->chan_count; i++) {
220 spin_lock(&ses->chans[i].server->srv_lock);
221 ses->chans[i].server->tcpStatus = CifsNeedReconnect;
222 spin_unlock(&ses->chans[i].server->srv_lock);
223 }
224 spin_unlock(&ses->chan_lock);
225 }
226 spin_unlock(&cifs_tcp_ses_lock);
227 }
228
229 /*
230 * Mark all sessions and tcons for reconnect.
231 * IMPORTANT: make sure that this gets called only from
232 * cifsd thread. For any other thread, use
233 * cifs_signal_cifsd_for_reconnect
234 *
235 * @server: the tcp ses for which reconnect is needed
236 * @server needs to be previously set to CifsNeedReconnect.
237 * @mark_smb_session: whether even sessions need to be marked
238 */
239 void
cifs_mark_tcp_ses_conns_for_reconnect(struct TCP_Server_Info * server,bool mark_smb_session)240 cifs_mark_tcp_ses_conns_for_reconnect(struct TCP_Server_Info *server,
241 bool mark_smb_session)
242 {
243 struct TCP_Server_Info *pserver;
244 struct cifs_ses *ses, *nses;
245 struct cifs_tcon *tcon;
246
247 /*
248 * before reconnecting the tcp session, mark the smb session (uid) and the tid bad so they
249 * are not used until reconnected.
250 */
251 cifs_dbg(FYI, "%s: marking necessary sessions and tcons for reconnect\n", __func__);
252
253 /* If server is a channel, select the primary channel */
254 pserver = CIFS_SERVER_IS_CHAN(server) ? server->primary_server : server;
255
256
257 spin_lock(&cifs_tcp_ses_lock);
258 list_for_each_entry_safe(ses, nses, &pserver->smb_ses_list, smb_ses_list) {
259 /* check if iface is still active */
260 if (!cifs_chan_is_iface_active(ses, server))
261 cifs_chan_update_iface(ses, server);
262
263 spin_lock(&ses->chan_lock);
264 if (!mark_smb_session && cifs_chan_needs_reconnect(ses, server))
265 goto next_session;
266
267 if (mark_smb_session)
268 CIFS_SET_ALL_CHANS_NEED_RECONNECT(ses);
269 else
270 cifs_chan_set_need_reconnect(ses, server);
271
272 /* If all channels need reconnect, then tcon needs reconnect */
273 if (!mark_smb_session && !CIFS_ALL_CHANS_NEED_RECONNECT(ses))
274 goto next_session;
275
276 ses->ses_status = SES_NEED_RECON;
277
278 list_for_each_entry(tcon, &ses->tcon_list, tcon_list) {
279 tcon->need_reconnect = true;
280 tcon->status = TID_NEED_RECON;
281 }
282 if (ses->tcon_ipc)
283 ses->tcon_ipc->need_reconnect = true;
284
285 next_session:
286 spin_unlock(&ses->chan_lock);
287 }
288 spin_unlock(&cifs_tcp_ses_lock);
289 }
290
291 static void
cifs_abort_connection(struct TCP_Server_Info * server)292 cifs_abort_connection(struct TCP_Server_Info *server)
293 {
294 struct mid_q_entry *mid, *nmid;
295 struct list_head retry_list;
296
297 server->maxBuf = 0;
298 server->max_read = 0;
299
300 /* do not want to be sending data on a socket we are freeing */
301 cifs_dbg(FYI, "%s: tearing down socket\n", __func__);
302 cifs_server_lock(server);
303 if (server->ssocket) {
304 cifs_dbg(FYI, "State: 0x%x Flags: 0x%lx\n", server->ssocket->state,
305 server->ssocket->flags);
306 kernel_sock_shutdown(server->ssocket, SHUT_WR);
307 cifs_dbg(FYI, "Post shutdown state: 0x%x Flags: 0x%lx\n", server->ssocket->state,
308 server->ssocket->flags);
309 sock_release(server->ssocket);
310 server->ssocket = NULL;
311 }
312 server->sequence_number = 0;
313 server->session_estab = false;
314 kfree_sensitive(server->session_key.response);
315 server->session_key.response = NULL;
316 server->session_key.len = 0;
317 server->lstrp = jiffies;
318
319 /* mark submitted MIDs for retry and issue callback */
320 INIT_LIST_HEAD(&retry_list);
321 cifs_dbg(FYI, "%s: moving mids to private list\n", __func__);
322 spin_lock(&server->mid_lock);
323 list_for_each_entry_safe(mid, nmid, &server->pending_mid_q, qhead) {
324 kref_get(&mid->refcount);
325 if (mid->mid_state == MID_REQUEST_SUBMITTED)
326 mid->mid_state = MID_RETRY_NEEDED;
327 list_move(&mid->qhead, &retry_list);
328 mid->mid_flags |= MID_DELETED;
329 }
330 spin_unlock(&server->mid_lock);
331 cifs_server_unlock(server);
332
333 cifs_dbg(FYI, "%s: issuing mid callbacks\n", __func__);
334 list_for_each_entry_safe(mid, nmid, &retry_list, qhead) {
335 list_del_init(&mid->qhead);
336 mid->callback(mid);
337 release_mid(mid);
338 }
339
340 if (cifs_rdma_enabled(server)) {
341 cifs_server_lock(server);
342 smbd_destroy(server);
343 cifs_server_unlock(server);
344 }
345 }
346
cifs_tcp_ses_needs_reconnect(struct TCP_Server_Info * server,int num_targets)347 static bool cifs_tcp_ses_needs_reconnect(struct TCP_Server_Info *server, int num_targets)
348 {
349 spin_lock(&server->srv_lock);
350 server->nr_targets = num_targets;
351 if (server->tcpStatus == CifsExiting) {
352 /* the demux thread will exit normally next time through the loop */
353 spin_unlock(&server->srv_lock);
354 wake_up(&server->response_q);
355 return false;
356 }
357
358 cifs_dbg(FYI, "Mark tcp session as need reconnect\n");
359 trace_smb3_reconnect(server->CurrentMid, server->conn_id,
360 server->hostname);
361 server->tcpStatus = CifsNeedReconnect;
362
363 spin_unlock(&server->srv_lock);
364 return true;
365 }
366
367 /*
368 * cifs tcp session reconnection
369 *
370 * mark tcp session as reconnecting so temporarily locked
371 * mark all smb sessions as reconnecting for tcp session
372 * reconnect tcp session
373 * wake up waiters on reconnection? - (not needed currently)
374 *
375 * if mark_smb_session is passed as true, unconditionally mark
376 * the smb session (and tcon) for reconnect as well. This value
377 * doesn't really matter for non-multichannel scenario.
378 *
379 */
__cifs_reconnect(struct TCP_Server_Info * server,bool mark_smb_session)380 static int __cifs_reconnect(struct TCP_Server_Info *server,
381 bool mark_smb_session)
382 {
383 int rc = 0;
384
385 if (!cifs_tcp_ses_needs_reconnect(server, 1))
386 return 0;
387
388 cifs_mark_tcp_ses_conns_for_reconnect(server, mark_smb_session);
389
390 cifs_abort_connection(server);
391
392 do {
393 try_to_freeze();
394 cifs_server_lock(server);
395
396 if (!cifs_swn_set_server_dstaddr(server)) {
397 /* resolve the hostname again to make sure that IP address is up-to-date */
398 rc = reconn_set_ipaddr_from_hostname(server);
399 cifs_dbg(FYI, "%s: reconn_set_ipaddr_from_hostname: rc=%d\n", __func__, rc);
400 }
401
402 if (cifs_rdma_enabled(server))
403 rc = smbd_reconnect(server);
404 else
405 rc = generic_ip_connect(server);
406 if (rc) {
407 cifs_server_unlock(server);
408 cifs_dbg(FYI, "%s: reconnect error %d\n", __func__, rc);
409 msleep(3000);
410 } else {
411 atomic_inc(&tcpSesReconnectCount);
412 set_credits(server, 1);
413 spin_lock(&server->srv_lock);
414 if (server->tcpStatus != CifsExiting)
415 server->tcpStatus = CifsNeedNegotiate;
416 spin_unlock(&server->srv_lock);
417 cifs_swn_reset_server_dstaddr(server);
418 cifs_server_unlock(server);
419 mod_delayed_work(cifsiod_wq, &server->reconnect, 0);
420 }
421 } while (server->tcpStatus == CifsNeedReconnect);
422
423 spin_lock(&server->srv_lock);
424 if (server->tcpStatus == CifsNeedNegotiate)
425 mod_delayed_work(cifsiod_wq, &server->echo, 0);
426 spin_unlock(&server->srv_lock);
427
428 wake_up(&server->response_q);
429 return rc;
430 }
431
432 #ifdef CONFIG_CIFS_DFS_UPCALL
__reconnect_target_unlocked(struct TCP_Server_Info * server,const char * target)433 static int __reconnect_target_unlocked(struct TCP_Server_Info *server, const char *target)
434 {
435 int rc;
436 char *hostname;
437
438 if (!cifs_swn_set_server_dstaddr(server)) {
439 if (server->hostname != target) {
440 hostname = extract_hostname(target);
441 if (!IS_ERR(hostname)) {
442 kfree(server->hostname);
443 server->hostname = hostname;
444 } else {
445 cifs_dbg(FYI, "%s: couldn't extract hostname or address from dfs target: %ld\n",
446 __func__, PTR_ERR(hostname));
447 cifs_dbg(FYI, "%s: default to last target server: %s\n", __func__,
448 server->hostname);
449 }
450 }
451 /* resolve the hostname again to make sure that IP address is up-to-date. */
452 rc = reconn_set_ipaddr_from_hostname(server);
453 cifs_dbg(FYI, "%s: reconn_set_ipaddr_from_hostname: rc=%d\n", __func__, rc);
454 }
455 /* Reconnect the socket */
456 if (cifs_rdma_enabled(server))
457 rc = smbd_reconnect(server);
458 else
459 rc = generic_ip_connect(server);
460
461 return rc;
462 }
463
reconnect_target_unlocked(struct TCP_Server_Info * server,struct dfs_cache_tgt_list * tl,struct dfs_cache_tgt_iterator ** target_hint)464 static int reconnect_target_unlocked(struct TCP_Server_Info *server, struct dfs_cache_tgt_list *tl,
465 struct dfs_cache_tgt_iterator **target_hint)
466 {
467 int rc;
468 struct dfs_cache_tgt_iterator *tit;
469
470 *target_hint = NULL;
471
472 /* If dfs target list is empty, then reconnect to last server */
473 tit = dfs_cache_get_tgt_iterator(tl);
474 if (!tit)
475 return __reconnect_target_unlocked(server, server->hostname);
476
477 /* Otherwise, try every dfs target in @tl */
478 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) {
479 rc = __reconnect_target_unlocked(server, dfs_cache_get_tgt_name(tit));
480 if (!rc) {
481 *target_hint = tit;
482 break;
483 }
484 }
485 return rc;
486 }
487
reconnect_dfs_server(struct TCP_Server_Info * server)488 static int reconnect_dfs_server(struct TCP_Server_Info *server)
489 {
490 int rc = 0;
491 const char *refpath = server->current_fullpath + 1;
492 struct dfs_cache_tgt_list tl = DFS_CACHE_TGT_LIST_INIT(tl);
493 struct dfs_cache_tgt_iterator *target_hint = NULL;
494 int num_targets = 0;
495
496 /*
497 * Determine the number of dfs targets the referral path in @cifs_sb resolves to.
498 *
499 * smb2_reconnect() needs to know how long it should wait based upon the number of dfs
500 * targets (server->nr_targets). It's also possible that the cached referral was cleared
501 * through /proc/fs/cifs/dfscache or the target list is empty due to server settings after
502 * refreshing the referral, so, in this case, default it to 1.
503 */
504 if (!dfs_cache_noreq_find(refpath, NULL, &tl))
505 num_targets = dfs_cache_get_nr_tgts(&tl);
506 if (!num_targets)
507 num_targets = 1;
508
509 if (!cifs_tcp_ses_needs_reconnect(server, num_targets))
510 return 0;
511
512 /*
513 * Unconditionally mark all sessions & tcons for reconnect as we might be connecting to a
514 * different server or share during failover. It could be improved by adding some logic to
515 * only do that in case it connects to a different server or share, though.
516 */
517 cifs_mark_tcp_ses_conns_for_reconnect(server, true);
518
519 cifs_abort_connection(server);
520
521 do {
522 try_to_freeze();
523 cifs_server_lock(server);
524
525 rc = reconnect_target_unlocked(server, &tl, &target_hint);
526 if (rc) {
527 /* Failed to reconnect socket */
528 cifs_server_unlock(server);
529 cifs_dbg(FYI, "%s: reconnect error %d\n", __func__, rc);
530 msleep(3000);
531 continue;
532 }
533 /*
534 * Socket was created. Update tcp session status to CifsNeedNegotiate so that a
535 * process waiting for reconnect will know it needs to re-establish session and tcon
536 * through the reconnected target server.
537 */
538 atomic_inc(&tcpSesReconnectCount);
539 set_credits(server, 1);
540 spin_lock(&server->srv_lock);
541 if (server->tcpStatus != CifsExiting)
542 server->tcpStatus = CifsNeedNegotiate;
543 spin_unlock(&server->srv_lock);
544 cifs_swn_reset_server_dstaddr(server);
545 cifs_server_unlock(server);
546 mod_delayed_work(cifsiod_wq, &server->reconnect, 0);
547 } while (server->tcpStatus == CifsNeedReconnect);
548
549 if (target_hint)
550 dfs_cache_noreq_update_tgthint(refpath, target_hint);
551
552 dfs_cache_free_tgts(&tl);
553
554 /* Need to set up echo worker again once connection has been established */
555 spin_lock(&server->srv_lock);
556 if (server->tcpStatus == CifsNeedNegotiate)
557 mod_delayed_work(cifsiod_wq, &server->echo, 0);
558 spin_unlock(&server->srv_lock);
559
560 wake_up(&server->response_q);
561 return rc;
562 }
563
cifs_reconnect(struct TCP_Server_Info * server,bool mark_smb_session)564 int cifs_reconnect(struct TCP_Server_Info *server, bool mark_smb_session)
565 {
566 /* If tcp session is not an dfs connection, then reconnect to last target server */
567 spin_lock(&server->srv_lock);
568 if (!server->is_dfs_conn) {
569 spin_unlock(&server->srv_lock);
570 return __cifs_reconnect(server, mark_smb_session);
571 }
572 spin_unlock(&server->srv_lock);
573
574 mutex_lock(&server->refpath_lock);
575 if (!server->origin_fullpath || !server->leaf_fullpath) {
576 mutex_unlock(&server->refpath_lock);
577 return __cifs_reconnect(server, mark_smb_session);
578 }
579 mutex_unlock(&server->refpath_lock);
580
581 return reconnect_dfs_server(server);
582 }
583 #else
cifs_reconnect(struct TCP_Server_Info * server,bool mark_smb_session)584 int cifs_reconnect(struct TCP_Server_Info *server, bool mark_smb_session)
585 {
586 return __cifs_reconnect(server, mark_smb_session);
587 }
588 #endif
589
590 static void
cifs_echo_request(struct work_struct * work)591 cifs_echo_request(struct work_struct *work)
592 {
593 int rc;
594 struct TCP_Server_Info *server = container_of(work,
595 struct TCP_Server_Info, echo.work);
596
597 /*
598 * We cannot send an echo if it is disabled.
599 * Also, no need to ping if we got a response recently.
600 */
601
602 if (server->tcpStatus == CifsNeedReconnect ||
603 server->tcpStatus == CifsExiting ||
604 server->tcpStatus == CifsNew ||
605 (server->ops->can_echo && !server->ops->can_echo(server)) ||
606 time_before(jiffies, server->lstrp + server->echo_interval - HZ))
607 goto requeue_echo;
608
609 rc = server->ops->echo ? server->ops->echo(server) : -ENOSYS;
610 if (rc)
611 cifs_dbg(FYI, "Unable to send echo request to server: %s\n",
612 server->hostname);
613
614 /* Check witness registrations */
615 cifs_swn_check();
616
617 requeue_echo:
618 queue_delayed_work(cifsiod_wq, &server->echo, server->echo_interval);
619 }
620
621 static bool
allocate_buffers(struct TCP_Server_Info * server)622 allocate_buffers(struct TCP_Server_Info *server)
623 {
624 if (!server->bigbuf) {
625 server->bigbuf = (char *)cifs_buf_get();
626 if (!server->bigbuf) {
627 cifs_server_dbg(VFS, "No memory for large SMB response\n");
628 msleep(3000);
629 /* retry will check if exiting */
630 return false;
631 }
632 } else if (server->large_buf) {
633 /* we are reusing a dirty large buf, clear its start */
634 memset(server->bigbuf, 0, HEADER_SIZE(server));
635 }
636
637 if (!server->smallbuf) {
638 server->smallbuf = (char *)cifs_small_buf_get();
639 if (!server->smallbuf) {
640 cifs_server_dbg(VFS, "No memory for SMB response\n");
641 msleep(1000);
642 /* retry will check if exiting */
643 return false;
644 }
645 /* beginning of smb buffer is cleared in our buf_get */
646 } else {
647 /* if existing small buf clear beginning */
648 memset(server->smallbuf, 0, HEADER_SIZE(server));
649 }
650
651 return true;
652 }
653
654 static bool
server_unresponsive(struct TCP_Server_Info * server)655 server_unresponsive(struct TCP_Server_Info *server)
656 {
657 /*
658 * We need to wait 3 echo intervals to make sure we handle such
659 * situations right:
660 * 1s client sends a normal SMB request
661 * 2s client gets a response
662 * 30s echo workqueue job pops, and decides we got a response recently
663 * and don't need to send another
664 * ...
665 * 65s kernel_recvmsg times out, and we see that we haven't gotten
666 * a response in >60s.
667 */
668 spin_lock(&server->srv_lock);
669 if ((server->tcpStatus == CifsGood ||
670 server->tcpStatus == CifsNeedNegotiate) &&
671 (!server->ops->can_echo || server->ops->can_echo(server)) &&
672 time_after(jiffies, server->lstrp + 3 * server->echo_interval)) {
673 spin_unlock(&server->srv_lock);
674 cifs_server_dbg(VFS, "has not responded in %lu seconds. Reconnecting...\n",
675 (3 * server->echo_interval) / HZ);
676 cifs_reconnect(server, false);
677 return true;
678 }
679 spin_unlock(&server->srv_lock);
680
681 return false;
682 }
683
684 static inline bool
zero_credits(struct TCP_Server_Info * server)685 zero_credits(struct TCP_Server_Info *server)
686 {
687 int val;
688
689 spin_lock(&server->req_lock);
690 val = server->credits + server->echo_credits + server->oplock_credits;
691 if (server->in_flight == 0 && val == 0) {
692 spin_unlock(&server->req_lock);
693 return true;
694 }
695 spin_unlock(&server->req_lock);
696 return false;
697 }
698
699 static int
cifs_readv_from_socket(struct TCP_Server_Info * server,struct msghdr * smb_msg)700 cifs_readv_from_socket(struct TCP_Server_Info *server, struct msghdr *smb_msg)
701 {
702 int length = 0;
703 int total_read;
704
705 for (total_read = 0; msg_data_left(smb_msg); total_read += length) {
706 try_to_freeze();
707
708 /* reconnect if no credits and no requests in flight */
709 if (zero_credits(server)) {
710 cifs_reconnect(server, false);
711 return -ECONNABORTED;
712 }
713
714 if (server_unresponsive(server))
715 return -ECONNABORTED;
716 if (cifs_rdma_enabled(server) && server->smbd_conn)
717 length = smbd_recv(server->smbd_conn, smb_msg);
718 else
719 length = sock_recvmsg(server->ssocket, smb_msg, 0);
720
721 spin_lock(&server->srv_lock);
722 if (server->tcpStatus == CifsExiting) {
723 spin_unlock(&server->srv_lock);
724 return -ESHUTDOWN;
725 }
726
727 if (server->tcpStatus == CifsNeedReconnect) {
728 spin_unlock(&server->srv_lock);
729 cifs_reconnect(server, false);
730 return -ECONNABORTED;
731 }
732 spin_unlock(&server->srv_lock);
733
734 if (length == -ERESTARTSYS ||
735 length == -EAGAIN ||
736 length == -EINTR) {
737 /*
738 * Minimum sleep to prevent looping, allowing socket
739 * to clear and app threads to set tcpStatus
740 * CifsNeedReconnect if server hung.
741 */
742 usleep_range(1000, 2000);
743 length = 0;
744 continue;
745 }
746
747 if (length <= 0) {
748 cifs_dbg(FYI, "Received no data or error: %d\n", length);
749 cifs_reconnect(server, false);
750 return -ECONNABORTED;
751 }
752 }
753 return total_read;
754 }
755
756 int
cifs_read_from_socket(struct TCP_Server_Info * server,char * buf,unsigned int to_read)757 cifs_read_from_socket(struct TCP_Server_Info *server, char *buf,
758 unsigned int to_read)
759 {
760 struct msghdr smb_msg = {};
761 struct kvec iov = {.iov_base = buf, .iov_len = to_read};
762 iov_iter_kvec(&smb_msg.msg_iter, READ, &iov, 1, to_read);
763
764 return cifs_readv_from_socket(server, &smb_msg);
765 }
766
767 ssize_t
cifs_discard_from_socket(struct TCP_Server_Info * server,size_t to_read)768 cifs_discard_from_socket(struct TCP_Server_Info *server, size_t to_read)
769 {
770 struct msghdr smb_msg = {};
771
772 /*
773 * iov_iter_discard already sets smb_msg.type and count and iov_offset
774 * and cifs_readv_from_socket sets msg_control and msg_controllen
775 * so little to initialize in struct msghdr
776 */
777 iov_iter_discard(&smb_msg.msg_iter, READ, to_read);
778
779 return cifs_readv_from_socket(server, &smb_msg);
780 }
781
782 int
cifs_read_page_from_socket(struct TCP_Server_Info * server,struct page * page,unsigned int page_offset,unsigned int to_read)783 cifs_read_page_from_socket(struct TCP_Server_Info *server, struct page *page,
784 unsigned int page_offset, unsigned int to_read)
785 {
786 struct msghdr smb_msg = {};
787 struct bio_vec bv = {
788 .bv_page = page, .bv_len = to_read, .bv_offset = page_offset};
789 iov_iter_bvec(&smb_msg.msg_iter, READ, &bv, 1, to_read);
790 return cifs_readv_from_socket(server, &smb_msg);
791 }
792
793 static bool
is_smb_response(struct TCP_Server_Info * server,unsigned char type)794 is_smb_response(struct TCP_Server_Info *server, unsigned char type)
795 {
796 /*
797 * The first byte big endian of the length field,
798 * is actually not part of the length but the type
799 * with the most common, zero, as regular data.
800 */
801 switch (type) {
802 case RFC1002_SESSION_MESSAGE:
803 /* Regular SMB response */
804 return true;
805 case RFC1002_SESSION_KEEP_ALIVE:
806 cifs_dbg(FYI, "RFC 1002 session keep alive\n");
807 break;
808 case RFC1002_POSITIVE_SESSION_RESPONSE:
809 cifs_dbg(FYI, "RFC 1002 positive session response\n");
810 break;
811 case RFC1002_NEGATIVE_SESSION_RESPONSE:
812 /*
813 * We get this from Windows 98 instead of an error on
814 * SMB negprot response.
815 */
816 cifs_dbg(FYI, "RFC 1002 negative session response\n");
817 /* give server a second to clean up */
818 msleep(1000);
819 /*
820 * Always try 445 first on reconnect since we get NACK
821 * on some if we ever connected to port 139 (the NACK
822 * is since we do not begin with RFC1001 session
823 * initialize frame).
824 */
825 cifs_set_port((struct sockaddr *)&server->dstaddr, CIFS_PORT);
826 cifs_reconnect(server, true);
827 break;
828 default:
829 cifs_server_dbg(VFS, "RFC 1002 unknown response type 0x%x\n", type);
830 cifs_reconnect(server, true);
831 }
832
833 return false;
834 }
835
836 void
dequeue_mid(struct mid_q_entry * mid,bool malformed)837 dequeue_mid(struct mid_q_entry *mid, bool malformed)
838 {
839 #ifdef CONFIG_CIFS_STATS2
840 mid->when_received = jiffies;
841 #endif
842 spin_lock(&mid->server->mid_lock);
843 if (!malformed)
844 mid->mid_state = MID_RESPONSE_RECEIVED;
845 else
846 mid->mid_state = MID_RESPONSE_MALFORMED;
847 /*
848 * Trying to handle/dequeue a mid after the send_recv()
849 * function has finished processing it is a bug.
850 */
851 if (mid->mid_flags & MID_DELETED) {
852 spin_unlock(&mid->server->mid_lock);
853 pr_warn_once("trying to dequeue a deleted mid\n");
854 } else {
855 list_del_init(&mid->qhead);
856 mid->mid_flags |= MID_DELETED;
857 spin_unlock(&mid->server->mid_lock);
858 }
859 }
860
861 static unsigned int
smb2_get_credits_from_hdr(char * buffer,struct TCP_Server_Info * server)862 smb2_get_credits_from_hdr(char *buffer, struct TCP_Server_Info *server)
863 {
864 struct smb2_hdr *shdr = (struct smb2_hdr *)buffer;
865
866 /*
867 * SMB1 does not use credits.
868 */
869 if (is_smb1(server))
870 return 0;
871
872 return le16_to_cpu(shdr->CreditRequest);
873 }
874
875 static void
handle_mid(struct mid_q_entry * mid,struct TCP_Server_Info * server,char * buf,int malformed)876 handle_mid(struct mid_q_entry *mid, struct TCP_Server_Info *server,
877 char *buf, int malformed)
878 {
879 if (server->ops->check_trans2 &&
880 server->ops->check_trans2(mid, server, buf, malformed))
881 return;
882 mid->credits_received = smb2_get_credits_from_hdr(buf, server);
883 mid->resp_buf = buf;
884 mid->large_buf = server->large_buf;
885 /* Was previous buf put in mpx struct for multi-rsp? */
886 if (!mid->multiRsp) {
887 /* smb buffer will be freed by user thread */
888 if (server->large_buf)
889 server->bigbuf = NULL;
890 else
891 server->smallbuf = NULL;
892 }
893 dequeue_mid(mid, malformed);
894 }
895
896 int
cifs_enable_signing(struct TCP_Server_Info * server,bool mnt_sign_required)897 cifs_enable_signing(struct TCP_Server_Info *server, bool mnt_sign_required)
898 {
899 bool srv_sign_required = server->sec_mode & server->vals->signing_required;
900 bool srv_sign_enabled = server->sec_mode & server->vals->signing_enabled;
901 bool mnt_sign_enabled;
902
903 /*
904 * Is signing required by mnt options? If not then check
905 * global_secflags to see if it is there.
906 */
907 if (!mnt_sign_required)
908 mnt_sign_required = ((global_secflags & CIFSSEC_MUST_SIGN) ==
909 CIFSSEC_MUST_SIGN);
910
911 /*
912 * If signing is required then it's automatically enabled too,
913 * otherwise, check to see if the secflags allow it.
914 */
915 mnt_sign_enabled = mnt_sign_required ? mnt_sign_required :
916 (global_secflags & CIFSSEC_MAY_SIGN);
917
918 /* If server requires signing, does client allow it? */
919 if (srv_sign_required) {
920 if (!mnt_sign_enabled) {
921 cifs_dbg(VFS, "Server requires signing, but it's disabled in SecurityFlags!\n");
922 return -EOPNOTSUPP;
923 }
924 server->sign = true;
925 }
926
927 /* If client requires signing, does server allow it? */
928 if (mnt_sign_required) {
929 if (!srv_sign_enabled) {
930 cifs_dbg(VFS, "Server does not support signing!\n");
931 return -EOPNOTSUPP;
932 }
933 server->sign = true;
934 }
935
936 if (cifs_rdma_enabled(server) && server->sign)
937 cifs_dbg(VFS, "Signing is enabled, and RDMA read/write will be disabled\n");
938
939 return 0;
940 }
941
942
clean_demultiplex_info(struct TCP_Server_Info * server)943 static void clean_demultiplex_info(struct TCP_Server_Info *server)
944 {
945 int length;
946
947 /* take it off the list, if it's not already */
948 spin_lock(&server->srv_lock);
949 list_del_init(&server->tcp_ses_list);
950 spin_unlock(&server->srv_lock);
951
952 cancel_delayed_work_sync(&server->echo);
953 cancel_delayed_work_sync(&server->resolve);
954
955 spin_lock(&server->srv_lock);
956 server->tcpStatus = CifsExiting;
957 spin_unlock(&server->srv_lock);
958 wake_up_all(&server->response_q);
959
960 /* check if we have blocked requests that need to free */
961 spin_lock(&server->req_lock);
962 if (server->credits <= 0)
963 server->credits = 1;
964 spin_unlock(&server->req_lock);
965 /*
966 * Although there should not be any requests blocked on this queue it
967 * can not hurt to be paranoid and try to wake up requests that may
968 * haven been blocked when more than 50 at time were on the wire to the
969 * same server - they now will see the session is in exit state and get
970 * out of SendReceive.
971 */
972 wake_up_all(&server->request_q);
973 /* give those requests time to exit */
974 msleep(125);
975 if (cifs_rdma_enabled(server))
976 smbd_destroy(server);
977 if (server->ssocket) {
978 sock_release(server->ssocket);
979 server->ssocket = NULL;
980 }
981
982 if (!list_empty(&server->pending_mid_q)) {
983 struct list_head dispose_list;
984 struct mid_q_entry *mid_entry;
985 struct list_head *tmp, *tmp2;
986
987 INIT_LIST_HEAD(&dispose_list);
988 spin_lock(&server->mid_lock);
989 list_for_each_safe(tmp, tmp2, &server->pending_mid_q) {
990 mid_entry = list_entry(tmp, struct mid_q_entry, qhead);
991 cifs_dbg(FYI, "Clearing mid %llu\n", mid_entry->mid);
992 kref_get(&mid_entry->refcount);
993 mid_entry->mid_state = MID_SHUTDOWN;
994 list_move(&mid_entry->qhead, &dispose_list);
995 mid_entry->mid_flags |= MID_DELETED;
996 }
997 spin_unlock(&server->mid_lock);
998
999 /* now walk dispose list and issue callbacks */
1000 list_for_each_safe(tmp, tmp2, &dispose_list) {
1001 mid_entry = list_entry(tmp, struct mid_q_entry, qhead);
1002 cifs_dbg(FYI, "Callback mid %llu\n", mid_entry->mid);
1003 list_del_init(&mid_entry->qhead);
1004 mid_entry->callback(mid_entry);
1005 release_mid(mid_entry);
1006 }
1007 /* 1/8th of sec is more than enough time for them to exit */
1008 msleep(125);
1009 }
1010
1011 if (!list_empty(&server->pending_mid_q)) {
1012 /*
1013 * mpx threads have not exited yet give them at least the smb
1014 * send timeout time for long ops.
1015 *
1016 * Due to delays on oplock break requests, we need to wait at
1017 * least 45 seconds before giving up on a request getting a
1018 * response and going ahead and killing cifsd.
1019 */
1020 cifs_dbg(FYI, "Wait for exit from demultiplex thread\n");
1021 msleep(46000);
1022 /*
1023 * If threads still have not exited they are probably never
1024 * coming home not much else we can do but free the memory.
1025 */
1026 }
1027
1028 #ifdef CONFIG_CIFS_DFS_UPCALL
1029 kfree(server->origin_fullpath);
1030 kfree(server->leaf_fullpath);
1031 #endif
1032 kfree(server);
1033
1034 length = atomic_dec_return(&tcpSesAllocCount);
1035 if (length > 0)
1036 mempool_resize(cifs_req_poolp, length + cifs_min_rcv);
1037 }
1038
1039 static int
standard_receive3(struct TCP_Server_Info * server,struct mid_q_entry * mid)1040 standard_receive3(struct TCP_Server_Info *server, struct mid_q_entry *mid)
1041 {
1042 int length;
1043 char *buf = server->smallbuf;
1044 unsigned int pdu_length = server->pdu_size;
1045
1046 /* make sure this will fit in a large buffer */
1047 if (pdu_length > CIFSMaxBufSize + MAX_HEADER_SIZE(server) -
1048 HEADER_PREAMBLE_SIZE(server)) {
1049 cifs_server_dbg(VFS, "SMB response too long (%u bytes)\n", pdu_length);
1050 cifs_reconnect(server, true);
1051 return -ECONNABORTED;
1052 }
1053
1054 /* switch to large buffer if too big for a small one */
1055 if (pdu_length > MAX_CIFS_SMALL_BUFFER_SIZE - 4) {
1056 server->large_buf = true;
1057 memcpy(server->bigbuf, buf, server->total_read);
1058 buf = server->bigbuf;
1059 }
1060
1061 /* now read the rest */
1062 length = cifs_read_from_socket(server, buf + HEADER_SIZE(server) - 1,
1063 pdu_length - MID_HEADER_SIZE(server));
1064
1065 if (length < 0)
1066 return length;
1067 server->total_read += length;
1068
1069 dump_smb(buf, server->total_read);
1070
1071 return cifs_handle_standard(server, mid);
1072 }
1073
1074 int
cifs_handle_standard(struct TCP_Server_Info * server,struct mid_q_entry * mid)1075 cifs_handle_standard(struct TCP_Server_Info *server, struct mid_q_entry *mid)
1076 {
1077 char *buf = server->large_buf ? server->bigbuf : server->smallbuf;
1078 int rc;
1079
1080 /*
1081 * We know that we received enough to get to the MID as we
1082 * checked the pdu_length earlier. Now check to see
1083 * if the rest of the header is OK.
1084 *
1085 * 48 bytes is enough to display the header and a little bit
1086 * into the payload for debugging purposes.
1087 */
1088 rc = server->ops->check_message(buf, server->total_read, server);
1089 if (rc)
1090 cifs_dump_mem("Bad SMB: ", buf,
1091 min_t(unsigned int, server->total_read, 48));
1092
1093 if (server->ops->is_session_expired &&
1094 server->ops->is_session_expired(buf)) {
1095 cifs_reconnect(server, true);
1096 return -1;
1097 }
1098
1099 if (server->ops->is_status_pending &&
1100 server->ops->is_status_pending(buf, server))
1101 return -1;
1102
1103 if (!mid)
1104 return rc;
1105
1106 handle_mid(mid, server, buf, rc);
1107 return 0;
1108 }
1109
1110 static void
smb2_add_credits_from_hdr(char * buffer,struct TCP_Server_Info * server)1111 smb2_add_credits_from_hdr(char *buffer, struct TCP_Server_Info *server)
1112 {
1113 struct smb2_hdr *shdr = (struct smb2_hdr *)buffer;
1114 int scredits, in_flight;
1115
1116 /*
1117 * SMB1 does not use credits.
1118 */
1119 if (is_smb1(server))
1120 return;
1121
1122 if (shdr->CreditRequest) {
1123 spin_lock(&server->req_lock);
1124 server->credits += le16_to_cpu(shdr->CreditRequest);
1125 scredits = server->credits;
1126 in_flight = server->in_flight;
1127 spin_unlock(&server->req_lock);
1128 wake_up(&server->request_q);
1129
1130 trace_smb3_hdr_credits(server->CurrentMid,
1131 server->conn_id, server->hostname, scredits,
1132 le16_to_cpu(shdr->CreditRequest), in_flight);
1133 cifs_server_dbg(FYI, "%s: added %u credits total=%d\n",
1134 __func__, le16_to_cpu(shdr->CreditRequest),
1135 scredits);
1136 }
1137 }
1138
1139
1140 static int
cifs_demultiplex_thread(void * p)1141 cifs_demultiplex_thread(void *p)
1142 {
1143 int i, num_mids, length;
1144 struct TCP_Server_Info *server = p;
1145 unsigned int pdu_length;
1146 unsigned int next_offset;
1147 char *buf = NULL;
1148 struct task_struct *task_to_wake = NULL;
1149 struct mid_q_entry *mids[MAX_COMPOUND];
1150 char *bufs[MAX_COMPOUND];
1151 unsigned int noreclaim_flag, num_io_timeout = 0;
1152
1153 noreclaim_flag = memalloc_noreclaim_save();
1154 cifs_dbg(FYI, "Demultiplex PID: %d\n", task_pid_nr(current));
1155
1156 length = atomic_inc_return(&tcpSesAllocCount);
1157 if (length > 1)
1158 mempool_resize(cifs_req_poolp, length + cifs_min_rcv);
1159
1160 set_freezable();
1161 allow_kernel_signal(SIGKILL);
1162 while (server->tcpStatus != CifsExiting) {
1163 if (try_to_freeze())
1164 continue;
1165
1166 if (!allocate_buffers(server))
1167 continue;
1168
1169 server->large_buf = false;
1170 buf = server->smallbuf;
1171 pdu_length = 4; /* enough to get RFC1001 header */
1172
1173 length = cifs_read_from_socket(server, buf, pdu_length);
1174 if (length < 0)
1175 continue;
1176
1177 if (is_smb1(server))
1178 server->total_read = length;
1179 else
1180 server->total_read = 0;
1181
1182 /*
1183 * The right amount was read from socket - 4 bytes,
1184 * so we can now interpret the length field.
1185 */
1186 pdu_length = get_rfc1002_length(buf);
1187
1188 cifs_dbg(FYI, "RFC1002 header 0x%x\n", pdu_length);
1189 if (!is_smb_response(server, buf[0]))
1190 continue;
1191 next_pdu:
1192 server->pdu_size = pdu_length;
1193
1194 /* make sure we have enough to get to the MID */
1195 if (server->pdu_size < MID_HEADER_SIZE(server)) {
1196 cifs_server_dbg(VFS, "SMB response too short (%u bytes)\n",
1197 server->pdu_size);
1198 cifs_reconnect(server, true);
1199 continue;
1200 }
1201
1202 /* read down to the MID */
1203 length = cifs_read_from_socket(server,
1204 buf + HEADER_PREAMBLE_SIZE(server),
1205 MID_HEADER_SIZE(server));
1206 if (length < 0)
1207 continue;
1208 server->total_read += length;
1209
1210 if (server->ops->next_header) {
1211 next_offset = server->ops->next_header(buf);
1212 if (next_offset)
1213 server->pdu_size = next_offset;
1214 }
1215
1216 memset(mids, 0, sizeof(mids));
1217 memset(bufs, 0, sizeof(bufs));
1218 num_mids = 0;
1219
1220 if (server->ops->is_transform_hdr &&
1221 server->ops->receive_transform &&
1222 server->ops->is_transform_hdr(buf)) {
1223 length = server->ops->receive_transform(server,
1224 mids,
1225 bufs,
1226 &num_mids);
1227 } else {
1228 mids[0] = server->ops->find_mid(server, buf);
1229 bufs[0] = buf;
1230 num_mids = 1;
1231
1232 if (!mids[0] || !mids[0]->receive)
1233 length = standard_receive3(server, mids[0]);
1234 else
1235 length = mids[0]->receive(server, mids[0]);
1236 }
1237
1238 if (length < 0) {
1239 for (i = 0; i < num_mids; i++)
1240 if (mids[i])
1241 release_mid(mids[i]);
1242 continue;
1243 }
1244
1245 if (server->ops->is_status_io_timeout &&
1246 server->ops->is_status_io_timeout(buf)) {
1247 num_io_timeout++;
1248 if (num_io_timeout > NUM_STATUS_IO_TIMEOUT) {
1249 cifs_reconnect(server, false);
1250 num_io_timeout = 0;
1251 continue;
1252 }
1253 }
1254
1255 server->lstrp = jiffies;
1256
1257 for (i = 0; i < num_mids; i++) {
1258 if (mids[i] != NULL) {
1259 mids[i]->resp_buf_size = server->pdu_size;
1260
1261 if (bufs[i] && server->ops->is_network_name_deleted)
1262 server->ops->is_network_name_deleted(bufs[i],
1263 server);
1264
1265 if (!mids[i]->multiRsp || mids[i]->multiEnd)
1266 mids[i]->callback(mids[i]);
1267
1268 release_mid(mids[i]);
1269 } else if (server->ops->is_oplock_break &&
1270 server->ops->is_oplock_break(bufs[i],
1271 server)) {
1272 smb2_add_credits_from_hdr(bufs[i], server);
1273 cifs_dbg(FYI, "Received oplock break\n");
1274 } else {
1275 cifs_server_dbg(VFS, "No task to wake, unknown frame received! NumMids %d\n",
1276 atomic_read(&mid_count));
1277 cifs_dump_mem("Received Data is: ", bufs[i],
1278 HEADER_SIZE(server));
1279 smb2_add_credits_from_hdr(bufs[i], server);
1280 #ifdef CONFIG_CIFS_DEBUG2
1281 if (server->ops->dump_detail)
1282 server->ops->dump_detail(bufs[i],
1283 server);
1284 cifs_dump_mids(server);
1285 #endif /* CIFS_DEBUG2 */
1286 }
1287 }
1288
1289 if (pdu_length > server->pdu_size) {
1290 if (!allocate_buffers(server))
1291 continue;
1292 pdu_length -= server->pdu_size;
1293 server->total_read = 0;
1294 server->large_buf = false;
1295 buf = server->smallbuf;
1296 goto next_pdu;
1297 }
1298 } /* end while !EXITING */
1299
1300 /* buffer usually freed in free_mid - need to free it here on exit */
1301 cifs_buf_release(server->bigbuf);
1302 if (server->smallbuf) /* no sense logging a debug message if NULL */
1303 cifs_small_buf_release(server->smallbuf);
1304
1305 task_to_wake = xchg(&server->tsk, NULL);
1306 clean_demultiplex_info(server);
1307
1308 /* if server->tsk was NULL then wait for a signal before exiting */
1309 if (!task_to_wake) {
1310 set_current_state(TASK_INTERRUPTIBLE);
1311 while (!signal_pending(current)) {
1312 schedule();
1313 set_current_state(TASK_INTERRUPTIBLE);
1314 }
1315 set_current_state(TASK_RUNNING);
1316 }
1317
1318 memalloc_noreclaim_restore(noreclaim_flag);
1319 module_put_and_kthread_exit(0);
1320 }
1321
1322 /*
1323 * Returns true if srcaddr isn't specified and rhs isn't specified, or
1324 * if srcaddr is specified and matches the IP address of the rhs argument
1325 */
1326 bool
cifs_match_ipaddr(struct sockaddr * srcaddr,struct sockaddr * rhs)1327 cifs_match_ipaddr(struct sockaddr *srcaddr, struct sockaddr *rhs)
1328 {
1329 switch (srcaddr->sa_family) {
1330 case AF_UNSPEC:
1331 return (rhs->sa_family == AF_UNSPEC);
1332 case AF_INET: {
1333 struct sockaddr_in *saddr4 = (struct sockaddr_in *)srcaddr;
1334 struct sockaddr_in *vaddr4 = (struct sockaddr_in *)rhs;
1335 return (saddr4->sin_addr.s_addr == vaddr4->sin_addr.s_addr);
1336 }
1337 case AF_INET6: {
1338 struct sockaddr_in6 *saddr6 = (struct sockaddr_in6 *)srcaddr;
1339 struct sockaddr_in6 *vaddr6 = (struct sockaddr_in6 *)rhs;
1340 return ipv6_addr_equal(&saddr6->sin6_addr, &vaddr6->sin6_addr);
1341 }
1342 default:
1343 WARN_ON(1);
1344 return false; /* don't expect to be here */
1345 }
1346 }
1347
1348 /*
1349 * If no port is specified in addr structure, we try to match with 445 port
1350 * and if it fails - with 139 ports. It should be called only if address
1351 * families of server and addr are equal.
1352 */
1353 static bool
match_port(struct TCP_Server_Info * server,struct sockaddr * addr)1354 match_port(struct TCP_Server_Info *server, struct sockaddr *addr)
1355 {
1356 __be16 port, *sport;
1357
1358 /* SMBDirect manages its own ports, don't match it here */
1359 if (server->rdma)
1360 return true;
1361
1362 switch (addr->sa_family) {
1363 case AF_INET:
1364 sport = &((struct sockaddr_in *) &server->dstaddr)->sin_port;
1365 port = ((struct sockaddr_in *) addr)->sin_port;
1366 break;
1367 case AF_INET6:
1368 sport = &((struct sockaddr_in6 *) &server->dstaddr)->sin6_port;
1369 port = ((struct sockaddr_in6 *) addr)->sin6_port;
1370 break;
1371 default:
1372 WARN_ON(1);
1373 return false;
1374 }
1375
1376 if (!port) {
1377 port = htons(CIFS_PORT);
1378 if (port == *sport)
1379 return true;
1380
1381 port = htons(RFC1001_PORT);
1382 }
1383
1384 return port == *sport;
1385 }
1386
1387 static bool
match_address(struct TCP_Server_Info * server,struct sockaddr * addr,struct sockaddr * srcaddr)1388 match_address(struct TCP_Server_Info *server, struct sockaddr *addr,
1389 struct sockaddr *srcaddr)
1390 {
1391 switch (addr->sa_family) {
1392 case AF_INET: {
1393 struct sockaddr_in *addr4 = (struct sockaddr_in *)addr;
1394 struct sockaddr_in *srv_addr4 =
1395 (struct sockaddr_in *)&server->dstaddr;
1396
1397 if (addr4->sin_addr.s_addr != srv_addr4->sin_addr.s_addr)
1398 return false;
1399 break;
1400 }
1401 case AF_INET6: {
1402 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)addr;
1403 struct sockaddr_in6 *srv_addr6 =
1404 (struct sockaddr_in6 *)&server->dstaddr;
1405
1406 if (!ipv6_addr_equal(&addr6->sin6_addr,
1407 &srv_addr6->sin6_addr))
1408 return false;
1409 if (addr6->sin6_scope_id != srv_addr6->sin6_scope_id)
1410 return false;
1411 break;
1412 }
1413 default:
1414 WARN_ON(1);
1415 return false; /* don't expect to be here */
1416 }
1417
1418 if (!cifs_match_ipaddr(srcaddr, (struct sockaddr *)&server->srcaddr))
1419 return false;
1420
1421 return true;
1422 }
1423
1424 static bool
match_security(struct TCP_Server_Info * server,struct smb3_fs_context * ctx)1425 match_security(struct TCP_Server_Info *server, struct smb3_fs_context *ctx)
1426 {
1427 /*
1428 * The select_sectype function should either return the ctx->sectype
1429 * that was specified, or "Unspecified" if that sectype was not
1430 * compatible with the given NEGOTIATE request.
1431 */
1432 if (server->ops->select_sectype(server, ctx->sectype)
1433 == Unspecified)
1434 return false;
1435
1436 /*
1437 * Now check if signing mode is acceptable. No need to check
1438 * global_secflags at this point since if MUST_SIGN is set then
1439 * the server->sign had better be too.
1440 */
1441 if (ctx->sign && !server->sign)
1442 return false;
1443
1444 return true;
1445 }
1446
1447 /* this function must be called with srv_lock held */
match_server(struct TCP_Server_Info * server,struct smb3_fs_context * ctx)1448 static int match_server(struct TCP_Server_Info *server, struct smb3_fs_context *ctx)
1449 {
1450 struct sockaddr *addr = (struct sockaddr *)&ctx->dstaddr;
1451
1452 if (ctx->nosharesock)
1453 return 0;
1454
1455 /* this server does not share socket */
1456 if (server->nosharesock)
1457 return 0;
1458
1459 /* If multidialect negotiation see if existing sessions match one */
1460 if (strcmp(ctx->vals->version_string, SMB3ANY_VERSION_STRING) == 0) {
1461 if (server->vals->protocol_id < SMB30_PROT_ID)
1462 return 0;
1463 } else if (strcmp(ctx->vals->version_string,
1464 SMBDEFAULT_VERSION_STRING) == 0) {
1465 if (server->vals->protocol_id < SMB21_PROT_ID)
1466 return 0;
1467 } else if ((server->vals != ctx->vals) || (server->ops != ctx->ops))
1468 return 0;
1469
1470 if (!net_eq(cifs_net_ns(server), current->nsproxy->net_ns))
1471 return 0;
1472
1473 if (strcasecmp(server->hostname, ctx->server_hostname))
1474 return 0;
1475
1476 if (!match_address(server, addr,
1477 (struct sockaddr *)&ctx->srcaddr))
1478 return 0;
1479
1480 if (!match_port(server, addr))
1481 return 0;
1482
1483 if (!match_security(server, ctx))
1484 return 0;
1485
1486 if (server->echo_interval != ctx->echo_interval * HZ)
1487 return 0;
1488
1489 if (server->rdma != ctx->rdma)
1490 return 0;
1491
1492 if (server->ignore_signature != ctx->ignore_signature)
1493 return 0;
1494
1495 if (server->min_offload != ctx->min_offload)
1496 return 0;
1497
1498 return 1;
1499 }
1500
1501 struct TCP_Server_Info *
cifs_find_tcp_session(struct smb3_fs_context * ctx)1502 cifs_find_tcp_session(struct smb3_fs_context *ctx)
1503 {
1504 struct TCP_Server_Info *server;
1505
1506 spin_lock(&cifs_tcp_ses_lock);
1507 list_for_each_entry(server, &cifs_tcp_ses_list, tcp_ses_list) {
1508 spin_lock(&server->srv_lock);
1509 #ifdef CONFIG_CIFS_DFS_UPCALL
1510 /*
1511 * DFS failover implementation in cifs_reconnect() requires unique tcp sessions for
1512 * DFS connections to do failover properly, so avoid sharing them with regular
1513 * shares or even links that may connect to same server but having completely
1514 * different failover targets.
1515 */
1516 if (server->is_dfs_conn) {
1517 spin_unlock(&server->srv_lock);
1518 continue;
1519 }
1520 #endif
1521 /*
1522 * Skip ses channels since they're only handled in lower layers
1523 * (e.g. cifs_send_recv).
1524 */
1525 if (CIFS_SERVER_IS_CHAN(server) || !match_server(server, ctx)) {
1526 spin_unlock(&server->srv_lock);
1527 continue;
1528 }
1529 spin_unlock(&server->srv_lock);
1530
1531 ++server->srv_count;
1532 spin_unlock(&cifs_tcp_ses_lock);
1533 cifs_dbg(FYI, "Existing tcp session with server found\n");
1534 return server;
1535 }
1536 spin_unlock(&cifs_tcp_ses_lock);
1537 return NULL;
1538 }
1539
1540 void
cifs_put_tcp_session(struct TCP_Server_Info * server,int from_reconnect)1541 cifs_put_tcp_session(struct TCP_Server_Info *server, int from_reconnect)
1542 {
1543 struct task_struct *task;
1544
1545 spin_lock(&cifs_tcp_ses_lock);
1546 if (--server->srv_count > 0) {
1547 spin_unlock(&cifs_tcp_ses_lock);
1548 return;
1549 }
1550
1551 /* srv_count can never go negative */
1552 WARN_ON(server->srv_count < 0);
1553
1554 put_net(cifs_net_ns(server));
1555
1556 list_del_init(&server->tcp_ses_list);
1557 spin_unlock(&cifs_tcp_ses_lock);
1558
1559 /* For secondary channels, we pick up ref-count on the primary server */
1560 if (CIFS_SERVER_IS_CHAN(server))
1561 cifs_put_tcp_session(server->primary_server, from_reconnect);
1562
1563 cancel_delayed_work_sync(&server->echo);
1564 cancel_delayed_work_sync(&server->resolve);
1565
1566 if (from_reconnect)
1567 /*
1568 * Avoid deadlock here: reconnect work calls
1569 * cifs_put_tcp_session() at its end. Need to be sure
1570 * that reconnect work does nothing with server pointer after
1571 * that step.
1572 */
1573 cancel_delayed_work(&server->reconnect);
1574 else
1575 cancel_delayed_work_sync(&server->reconnect);
1576
1577 spin_lock(&server->srv_lock);
1578 server->tcpStatus = CifsExiting;
1579 spin_unlock(&server->srv_lock);
1580
1581 cifs_crypto_secmech_release(server);
1582
1583 kfree_sensitive(server->session_key.response);
1584 server->session_key.response = NULL;
1585 server->session_key.len = 0;
1586 kfree(server->hostname);
1587 server->hostname = NULL;
1588
1589 task = xchg(&server->tsk, NULL);
1590 if (task)
1591 send_sig(SIGKILL, task, 1);
1592 }
1593
1594 struct TCP_Server_Info *
cifs_get_tcp_session(struct smb3_fs_context * ctx,struct TCP_Server_Info * primary_server)1595 cifs_get_tcp_session(struct smb3_fs_context *ctx,
1596 struct TCP_Server_Info *primary_server)
1597 {
1598 struct TCP_Server_Info *tcp_ses = NULL;
1599 int rc;
1600
1601 cifs_dbg(FYI, "UNC: %s\n", ctx->UNC);
1602
1603 /* see if we already have a matching tcp_ses */
1604 tcp_ses = cifs_find_tcp_session(ctx);
1605 if (tcp_ses)
1606 return tcp_ses;
1607
1608 tcp_ses = kzalloc(sizeof(struct TCP_Server_Info), GFP_KERNEL);
1609 if (!tcp_ses) {
1610 rc = -ENOMEM;
1611 goto out_err;
1612 }
1613
1614 tcp_ses->hostname = kstrdup(ctx->server_hostname, GFP_KERNEL);
1615 if (!tcp_ses->hostname) {
1616 rc = -ENOMEM;
1617 goto out_err;
1618 }
1619
1620 if (ctx->nosharesock)
1621 tcp_ses->nosharesock = true;
1622
1623 tcp_ses->ops = ctx->ops;
1624 tcp_ses->vals = ctx->vals;
1625 cifs_set_net_ns(tcp_ses, get_net(current->nsproxy->net_ns));
1626
1627 tcp_ses->conn_id = atomic_inc_return(&tcpSesNextId);
1628 tcp_ses->noblockcnt = ctx->rootfs;
1629 tcp_ses->noblocksnd = ctx->noblocksnd || ctx->rootfs;
1630 tcp_ses->noautotune = ctx->noautotune;
1631 tcp_ses->tcp_nodelay = ctx->sockopt_tcp_nodelay;
1632 tcp_ses->rdma = ctx->rdma;
1633 tcp_ses->in_flight = 0;
1634 tcp_ses->max_in_flight = 0;
1635 tcp_ses->credits = 1;
1636 if (primary_server) {
1637 spin_lock(&cifs_tcp_ses_lock);
1638 ++primary_server->srv_count;
1639 spin_unlock(&cifs_tcp_ses_lock);
1640 tcp_ses->primary_server = primary_server;
1641 }
1642 init_waitqueue_head(&tcp_ses->response_q);
1643 init_waitqueue_head(&tcp_ses->request_q);
1644 INIT_LIST_HEAD(&tcp_ses->pending_mid_q);
1645 mutex_init(&tcp_ses->_srv_mutex);
1646 memcpy(tcp_ses->workstation_RFC1001_name,
1647 ctx->source_rfc1001_name, RFC1001_NAME_LEN_WITH_NULL);
1648 memcpy(tcp_ses->server_RFC1001_name,
1649 ctx->target_rfc1001_name, RFC1001_NAME_LEN_WITH_NULL);
1650 tcp_ses->session_estab = false;
1651 tcp_ses->sequence_number = 0;
1652 tcp_ses->reconnect_instance = 1;
1653 tcp_ses->lstrp = jiffies;
1654 tcp_ses->compress_algorithm = cpu_to_le16(ctx->compression);
1655 spin_lock_init(&tcp_ses->req_lock);
1656 spin_lock_init(&tcp_ses->srv_lock);
1657 spin_lock_init(&tcp_ses->mid_lock);
1658 INIT_LIST_HEAD(&tcp_ses->tcp_ses_list);
1659 INIT_LIST_HEAD(&tcp_ses->smb_ses_list);
1660 INIT_DELAYED_WORK(&tcp_ses->echo, cifs_echo_request);
1661 INIT_DELAYED_WORK(&tcp_ses->resolve, cifs_resolve_server);
1662 INIT_DELAYED_WORK(&tcp_ses->reconnect, smb2_reconnect_server);
1663 mutex_init(&tcp_ses->reconnect_mutex);
1664 #ifdef CONFIG_CIFS_DFS_UPCALL
1665 mutex_init(&tcp_ses->refpath_lock);
1666 #endif
1667 memcpy(&tcp_ses->srcaddr, &ctx->srcaddr,
1668 sizeof(tcp_ses->srcaddr));
1669 memcpy(&tcp_ses->dstaddr, &ctx->dstaddr,
1670 sizeof(tcp_ses->dstaddr));
1671 if (ctx->use_client_guid)
1672 memcpy(tcp_ses->client_guid, ctx->client_guid,
1673 SMB2_CLIENT_GUID_SIZE);
1674 else
1675 generate_random_uuid(tcp_ses->client_guid);
1676 /*
1677 * at this point we are the only ones with the pointer
1678 * to the struct since the kernel thread not created yet
1679 * no need to spinlock this init of tcpStatus or srv_count
1680 */
1681 tcp_ses->tcpStatus = CifsNew;
1682 ++tcp_ses->srv_count;
1683
1684 if (ctx->echo_interval >= SMB_ECHO_INTERVAL_MIN &&
1685 ctx->echo_interval <= SMB_ECHO_INTERVAL_MAX)
1686 tcp_ses->echo_interval = ctx->echo_interval * HZ;
1687 else
1688 tcp_ses->echo_interval = SMB_ECHO_INTERVAL_DEFAULT * HZ;
1689 if (tcp_ses->rdma) {
1690 #ifndef CONFIG_CIFS_SMB_DIRECT
1691 cifs_dbg(VFS, "CONFIG_CIFS_SMB_DIRECT is not enabled\n");
1692 rc = -ENOENT;
1693 goto out_err_crypto_release;
1694 #endif
1695 tcp_ses->smbd_conn = smbd_get_connection(
1696 tcp_ses, (struct sockaddr *)&ctx->dstaddr);
1697 if (tcp_ses->smbd_conn) {
1698 cifs_dbg(VFS, "RDMA transport established\n");
1699 rc = 0;
1700 goto smbd_connected;
1701 } else {
1702 rc = -ENOENT;
1703 goto out_err_crypto_release;
1704 }
1705 }
1706 rc = ip_connect(tcp_ses);
1707 if (rc < 0) {
1708 cifs_dbg(VFS, "Error connecting to socket. Aborting operation.\n");
1709 goto out_err_crypto_release;
1710 }
1711 smbd_connected:
1712 /*
1713 * since we're in a cifs function already, we know that
1714 * this will succeed. No need for try_module_get().
1715 */
1716 __module_get(THIS_MODULE);
1717 tcp_ses->tsk = kthread_run(cifs_demultiplex_thread,
1718 tcp_ses, "cifsd");
1719 if (IS_ERR(tcp_ses->tsk)) {
1720 rc = PTR_ERR(tcp_ses->tsk);
1721 cifs_dbg(VFS, "error %d create cifsd thread\n", rc);
1722 module_put(THIS_MODULE);
1723 goto out_err_crypto_release;
1724 }
1725 tcp_ses->min_offload = ctx->min_offload;
1726 /*
1727 * at this point we are the only ones with the pointer
1728 * to the struct since the kernel thread not created yet
1729 * no need to spinlock this update of tcpStatus
1730 */
1731 spin_lock(&tcp_ses->srv_lock);
1732 tcp_ses->tcpStatus = CifsNeedNegotiate;
1733 spin_unlock(&tcp_ses->srv_lock);
1734
1735 if ((ctx->max_credits < 20) || (ctx->max_credits > 60000))
1736 tcp_ses->max_credits = SMB2_MAX_CREDITS_AVAILABLE;
1737 else
1738 tcp_ses->max_credits = ctx->max_credits;
1739
1740 tcp_ses->nr_targets = 1;
1741 tcp_ses->ignore_signature = ctx->ignore_signature;
1742 /* thread spawned, put it on the list */
1743 spin_lock(&cifs_tcp_ses_lock);
1744 list_add(&tcp_ses->tcp_ses_list, &cifs_tcp_ses_list);
1745 spin_unlock(&cifs_tcp_ses_lock);
1746
1747 /* queue echo request delayed work */
1748 queue_delayed_work(cifsiod_wq, &tcp_ses->echo, tcp_ses->echo_interval);
1749
1750 /* queue dns resolution delayed work */
1751 cifs_dbg(FYI, "%s: next dns resolution scheduled for %d seconds in the future\n",
1752 __func__, SMB_DNS_RESOLVE_INTERVAL_DEFAULT);
1753
1754 queue_delayed_work(cifsiod_wq, &tcp_ses->resolve, (SMB_DNS_RESOLVE_INTERVAL_DEFAULT * HZ));
1755
1756 return tcp_ses;
1757
1758 out_err_crypto_release:
1759 cifs_crypto_secmech_release(tcp_ses);
1760
1761 put_net(cifs_net_ns(tcp_ses));
1762
1763 out_err:
1764 if (tcp_ses) {
1765 if (CIFS_SERVER_IS_CHAN(tcp_ses))
1766 cifs_put_tcp_session(tcp_ses->primary_server, false);
1767 kfree(tcp_ses->hostname);
1768 if (tcp_ses->ssocket)
1769 sock_release(tcp_ses->ssocket);
1770 kfree(tcp_ses);
1771 }
1772 return ERR_PTR(rc);
1773 }
1774
1775 /* this function must be called with ses_lock held */
match_session(struct cifs_ses * ses,struct smb3_fs_context * ctx)1776 static int match_session(struct cifs_ses *ses, struct smb3_fs_context *ctx)
1777 {
1778 if (ctx->sectype != Unspecified &&
1779 ctx->sectype != ses->sectype)
1780 return 0;
1781
1782 /*
1783 * If an existing session is limited to less channels than
1784 * requested, it should not be reused
1785 */
1786 spin_lock(&ses->chan_lock);
1787 if (ses->chan_max < ctx->max_channels) {
1788 spin_unlock(&ses->chan_lock);
1789 return 0;
1790 }
1791 spin_unlock(&ses->chan_lock);
1792
1793 switch (ses->sectype) {
1794 case Kerberos:
1795 if (!uid_eq(ctx->cred_uid, ses->cred_uid))
1796 return 0;
1797 break;
1798 default:
1799 /* NULL username means anonymous session */
1800 if (ses->user_name == NULL) {
1801 if (!ctx->nullauth)
1802 return 0;
1803 break;
1804 }
1805
1806 /* anything else takes username/password */
1807 if (strncmp(ses->user_name,
1808 ctx->username ? ctx->username : "",
1809 CIFS_MAX_USERNAME_LEN))
1810 return 0;
1811 if ((ctx->username && strlen(ctx->username) != 0) &&
1812 ses->password != NULL &&
1813 strncmp(ses->password,
1814 ctx->password ? ctx->password : "",
1815 CIFS_MAX_PASSWORD_LEN))
1816 return 0;
1817 }
1818 return 1;
1819 }
1820
1821 /**
1822 * cifs_setup_ipc - helper to setup the IPC tcon for the session
1823 * @ses: smb session to issue the request on
1824 * @ctx: the superblock configuration context to use for building the
1825 * new tree connection for the IPC (interprocess communication RPC)
1826 *
1827 * A new IPC connection is made and stored in the session
1828 * tcon_ipc. The IPC tcon has the same lifetime as the session.
1829 */
1830 static int
cifs_setup_ipc(struct cifs_ses * ses,struct smb3_fs_context * ctx)1831 cifs_setup_ipc(struct cifs_ses *ses, struct smb3_fs_context *ctx)
1832 {
1833 int rc = 0, xid;
1834 struct cifs_tcon *tcon;
1835 char unc[SERVER_NAME_LENGTH + sizeof("//x/IPC$")] = {0};
1836 bool seal = false;
1837 struct TCP_Server_Info *server = ses->server;
1838
1839 /*
1840 * If the mount request that resulted in the creation of the
1841 * session requires encryption, force IPC to be encrypted too.
1842 */
1843 if (ctx->seal) {
1844 if (server->capabilities & SMB2_GLOBAL_CAP_ENCRYPTION)
1845 seal = true;
1846 else {
1847 cifs_server_dbg(VFS,
1848 "IPC: server doesn't support encryption\n");
1849 return -EOPNOTSUPP;
1850 }
1851 }
1852
1853 tcon = tconInfoAlloc();
1854 if (tcon == NULL)
1855 return -ENOMEM;
1856
1857 scnprintf(unc, sizeof(unc), "\\\\%s\\IPC$", server->hostname);
1858
1859 xid = get_xid();
1860 tcon->ses = ses;
1861 tcon->ipc = true;
1862 tcon->seal = seal;
1863 rc = server->ops->tree_connect(xid, ses, unc, tcon, ctx->local_nls);
1864 free_xid(xid);
1865
1866 if (rc) {
1867 cifs_server_dbg(VFS, "failed to connect to IPC (rc=%d)\n", rc);
1868 tconInfoFree(tcon);
1869 goto out;
1870 }
1871
1872 cifs_dbg(FYI, "IPC tcon rc=%d ipc tid=0x%x\n", rc, tcon->tid);
1873
1874 ses->tcon_ipc = tcon;
1875 out:
1876 return rc;
1877 }
1878
1879 /**
1880 * cifs_free_ipc - helper to release the session IPC tcon
1881 * @ses: smb session to unmount the IPC from
1882 *
1883 * Needs to be called everytime a session is destroyed.
1884 *
1885 * On session close, the IPC is closed and the server must release all tcons of the session.
1886 * No need to send a tree disconnect here.
1887 *
1888 * Besides, it will make the server to not close durable and resilient files on session close, as
1889 * specified in MS-SMB2 3.3.5.6 Receiving an SMB2 LOGOFF Request.
1890 */
1891 static int
cifs_free_ipc(struct cifs_ses * ses)1892 cifs_free_ipc(struct cifs_ses *ses)
1893 {
1894 struct cifs_tcon *tcon = ses->tcon_ipc;
1895
1896 if (tcon == NULL)
1897 return 0;
1898
1899 tconInfoFree(tcon);
1900 ses->tcon_ipc = NULL;
1901 return 0;
1902 }
1903
1904 static struct cifs_ses *
cifs_find_smb_ses(struct TCP_Server_Info * server,struct smb3_fs_context * ctx)1905 cifs_find_smb_ses(struct TCP_Server_Info *server, struct smb3_fs_context *ctx)
1906 {
1907 struct cifs_ses *ses;
1908
1909 spin_lock(&cifs_tcp_ses_lock);
1910 list_for_each_entry(ses, &server->smb_ses_list, smb_ses_list) {
1911 spin_lock(&ses->ses_lock);
1912 if (ses->ses_status == SES_EXITING) {
1913 spin_unlock(&ses->ses_lock);
1914 continue;
1915 }
1916 if (!match_session(ses, ctx)) {
1917 spin_unlock(&ses->ses_lock);
1918 continue;
1919 }
1920 spin_unlock(&ses->ses_lock);
1921
1922 ++ses->ses_count;
1923 spin_unlock(&cifs_tcp_ses_lock);
1924 return ses;
1925 }
1926 spin_unlock(&cifs_tcp_ses_lock);
1927 return NULL;
1928 }
1929
cifs_put_smb_ses(struct cifs_ses * ses)1930 void cifs_put_smb_ses(struct cifs_ses *ses)
1931 {
1932 unsigned int rc, xid;
1933 unsigned int chan_count;
1934 struct TCP_Server_Info *server = ses->server;
1935
1936 spin_lock(&ses->ses_lock);
1937 if (ses->ses_status == SES_EXITING) {
1938 spin_unlock(&ses->ses_lock);
1939 return;
1940 }
1941 spin_unlock(&ses->ses_lock);
1942
1943 cifs_dbg(FYI, "%s: ses_count=%d\n", __func__, ses->ses_count);
1944 cifs_dbg(FYI,
1945 "%s: ses ipc: %s\n", __func__, ses->tcon_ipc ? ses->tcon_ipc->tree_name : "NONE");
1946
1947 spin_lock(&cifs_tcp_ses_lock);
1948 if (--ses->ses_count > 0) {
1949 spin_unlock(&cifs_tcp_ses_lock);
1950 return;
1951 }
1952 spin_unlock(&cifs_tcp_ses_lock);
1953
1954 /* ses_count can never go negative */
1955 WARN_ON(ses->ses_count < 0);
1956
1957 if (ses->ses_status == SES_GOOD)
1958 ses->ses_status = SES_EXITING;
1959
1960 cifs_free_ipc(ses);
1961
1962 if (ses->ses_status == SES_EXITING && server->ops->logoff) {
1963 xid = get_xid();
1964 rc = server->ops->logoff(xid, ses);
1965 if (rc)
1966 cifs_server_dbg(VFS, "%s: Session Logoff failure rc=%d\n",
1967 __func__, rc);
1968 _free_xid(xid);
1969 }
1970
1971 spin_lock(&cifs_tcp_ses_lock);
1972 list_del_init(&ses->smb_ses_list);
1973 spin_unlock(&cifs_tcp_ses_lock);
1974
1975 chan_count = ses->chan_count;
1976
1977 /* close any extra channels */
1978 if (chan_count > 1) {
1979 int i;
1980
1981 for (i = 1; i < chan_count; i++) {
1982 if (ses->chans[i].iface) {
1983 kref_put(&ses->chans[i].iface->refcount, release_iface);
1984 ses->chans[i].iface = NULL;
1985 }
1986 cifs_put_tcp_session(ses->chans[i].server, 0);
1987 ses->chans[i].server = NULL;
1988 }
1989 }
1990
1991 sesInfoFree(ses);
1992 cifs_put_tcp_session(server, 0);
1993 }
1994
1995 #ifdef CONFIG_KEYS
1996
1997 /* strlen("cifs:a:") + CIFS_MAX_DOMAINNAME_LEN + 1 */
1998 #define CIFSCREDS_DESC_SIZE (7 + CIFS_MAX_DOMAINNAME_LEN + 1)
1999
2000 /* Populate username and pw fields from keyring if possible */
2001 static int
cifs_set_cifscreds(struct smb3_fs_context * ctx,struct cifs_ses * ses)2002 cifs_set_cifscreds(struct smb3_fs_context *ctx, struct cifs_ses *ses)
2003 {
2004 int rc = 0;
2005 int is_domain = 0;
2006 const char *delim, *payload;
2007 char *desc;
2008 ssize_t len;
2009 struct key *key;
2010 struct TCP_Server_Info *server = ses->server;
2011 struct sockaddr_in *sa;
2012 struct sockaddr_in6 *sa6;
2013 const struct user_key_payload *upayload;
2014
2015 desc = kmalloc(CIFSCREDS_DESC_SIZE, GFP_KERNEL);
2016 if (!desc)
2017 return -ENOMEM;
2018
2019 /* try to find an address key first */
2020 switch (server->dstaddr.ss_family) {
2021 case AF_INET:
2022 sa = (struct sockaddr_in *)&server->dstaddr;
2023 sprintf(desc, "cifs:a:%pI4", &sa->sin_addr.s_addr);
2024 break;
2025 case AF_INET6:
2026 sa6 = (struct sockaddr_in6 *)&server->dstaddr;
2027 sprintf(desc, "cifs:a:%pI6c", &sa6->sin6_addr.s6_addr);
2028 break;
2029 default:
2030 cifs_dbg(FYI, "Bad ss_family (%hu)\n",
2031 server->dstaddr.ss_family);
2032 rc = -EINVAL;
2033 goto out_err;
2034 }
2035
2036 cifs_dbg(FYI, "%s: desc=%s\n", __func__, desc);
2037 key = request_key(&key_type_logon, desc, "");
2038 if (IS_ERR(key)) {
2039 if (!ses->domainName) {
2040 cifs_dbg(FYI, "domainName is NULL\n");
2041 rc = PTR_ERR(key);
2042 goto out_err;
2043 }
2044
2045 /* didn't work, try to find a domain key */
2046 sprintf(desc, "cifs:d:%s", ses->domainName);
2047 cifs_dbg(FYI, "%s: desc=%s\n", __func__, desc);
2048 key = request_key(&key_type_logon, desc, "");
2049 if (IS_ERR(key)) {
2050 rc = PTR_ERR(key);
2051 goto out_err;
2052 }
2053 is_domain = 1;
2054 }
2055
2056 down_read(&key->sem);
2057 upayload = user_key_payload_locked(key);
2058 if (IS_ERR_OR_NULL(upayload)) {
2059 rc = upayload ? PTR_ERR(upayload) : -EINVAL;
2060 goto out_key_put;
2061 }
2062
2063 /* find first : in payload */
2064 payload = upayload->data;
2065 delim = strnchr(payload, upayload->datalen, ':');
2066 cifs_dbg(FYI, "payload=%s\n", payload);
2067 if (!delim) {
2068 cifs_dbg(FYI, "Unable to find ':' in payload (datalen=%d)\n",
2069 upayload->datalen);
2070 rc = -EINVAL;
2071 goto out_key_put;
2072 }
2073
2074 len = delim - payload;
2075 if (len > CIFS_MAX_USERNAME_LEN || len <= 0) {
2076 cifs_dbg(FYI, "Bad value from username search (len=%zd)\n",
2077 len);
2078 rc = -EINVAL;
2079 goto out_key_put;
2080 }
2081
2082 ctx->username = kstrndup(payload, len, GFP_KERNEL);
2083 if (!ctx->username) {
2084 cifs_dbg(FYI, "Unable to allocate %zd bytes for username\n",
2085 len);
2086 rc = -ENOMEM;
2087 goto out_key_put;
2088 }
2089 cifs_dbg(FYI, "%s: username=%s\n", __func__, ctx->username);
2090
2091 len = key->datalen - (len + 1);
2092 if (len > CIFS_MAX_PASSWORD_LEN || len <= 0) {
2093 cifs_dbg(FYI, "Bad len for password search (len=%zd)\n", len);
2094 rc = -EINVAL;
2095 kfree(ctx->username);
2096 ctx->username = NULL;
2097 goto out_key_put;
2098 }
2099
2100 ++delim;
2101 ctx->password = kstrndup(delim, len, GFP_KERNEL);
2102 if (!ctx->password) {
2103 cifs_dbg(FYI, "Unable to allocate %zd bytes for password\n",
2104 len);
2105 rc = -ENOMEM;
2106 kfree(ctx->username);
2107 ctx->username = NULL;
2108 goto out_key_put;
2109 }
2110
2111 /*
2112 * If we have a domain key then we must set the domainName in the
2113 * for the request.
2114 */
2115 if (is_domain && ses->domainName) {
2116 ctx->domainname = kstrdup(ses->domainName, GFP_KERNEL);
2117 if (!ctx->domainname) {
2118 cifs_dbg(FYI, "Unable to allocate %zd bytes for domain\n",
2119 len);
2120 rc = -ENOMEM;
2121 kfree(ctx->username);
2122 ctx->username = NULL;
2123 kfree_sensitive(ctx->password);
2124 ctx->password = NULL;
2125 goto out_key_put;
2126 }
2127 }
2128
2129 strscpy(ctx->workstation_name, ses->workstation_name, sizeof(ctx->workstation_name));
2130
2131 out_key_put:
2132 up_read(&key->sem);
2133 key_put(key);
2134 out_err:
2135 kfree(desc);
2136 cifs_dbg(FYI, "%s: returning %d\n", __func__, rc);
2137 return rc;
2138 }
2139 #else /* ! CONFIG_KEYS */
2140 static inline int
cifs_set_cifscreds(struct smb3_fs_context * ctx,struct cifs_ses * ses)2141 cifs_set_cifscreds(struct smb3_fs_context *ctx __attribute__((unused)),
2142 struct cifs_ses *ses __attribute__((unused)))
2143 {
2144 return -ENOSYS;
2145 }
2146 #endif /* CONFIG_KEYS */
2147
2148 /**
2149 * cifs_get_smb_ses - get a session matching @ctx data from @server
2150 * @server: server to setup the session to
2151 * @ctx: superblock configuration context to use to setup the session
2152 *
2153 * This function assumes it is being called from cifs_mount() where we
2154 * already got a server reference (server refcount +1). See
2155 * cifs_get_tcon() for refcount explanations.
2156 */
2157 struct cifs_ses *
cifs_get_smb_ses(struct TCP_Server_Info * server,struct smb3_fs_context * ctx)2158 cifs_get_smb_ses(struct TCP_Server_Info *server, struct smb3_fs_context *ctx)
2159 {
2160 int rc = -ENOMEM;
2161 unsigned int xid;
2162 struct cifs_ses *ses;
2163 struct sockaddr_in *addr = (struct sockaddr_in *)&server->dstaddr;
2164 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&server->dstaddr;
2165
2166 xid = get_xid();
2167
2168 ses = cifs_find_smb_ses(server, ctx);
2169 if (ses) {
2170 cifs_dbg(FYI, "Existing smb sess found (status=%d)\n",
2171 ses->ses_status);
2172
2173 spin_lock(&ses->chan_lock);
2174 if (cifs_chan_needs_reconnect(ses, server)) {
2175 spin_unlock(&ses->chan_lock);
2176 cifs_dbg(FYI, "Session needs reconnect\n");
2177
2178 mutex_lock(&ses->session_mutex);
2179 rc = cifs_negotiate_protocol(xid, ses, server);
2180 if (rc) {
2181 mutex_unlock(&ses->session_mutex);
2182 /* problem -- put our ses reference */
2183 cifs_put_smb_ses(ses);
2184 free_xid(xid);
2185 return ERR_PTR(rc);
2186 }
2187
2188 rc = cifs_setup_session(xid, ses, server,
2189 ctx->local_nls);
2190 if (rc) {
2191 mutex_unlock(&ses->session_mutex);
2192 /* problem -- put our reference */
2193 cifs_put_smb_ses(ses);
2194 free_xid(xid);
2195 return ERR_PTR(rc);
2196 }
2197 mutex_unlock(&ses->session_mutex);
2198
2199 spin_lock(&ses->chan_lock);
2200 }
2201 spin_unlock(&ses->chan_lock);
2202
2203 /* existing SMB ses has a server reference already */
2204 cifs_put_tcp_session(server, 0);
2205 free_xid(xid);
2206 return ses;
2207 }
2208
2209 cifs_dbg(FYI, "Existing smb sess not found\n");
2210 ses = sesInfoAlloc();
2211 if (ses == NULL)
2212 goto get_ses_fail;
2213
2214 /* new SMB session uses our server ref */
2215 ses->server = server;
2216 if (server->dstaddr.ss_family == AF_INET6)
2217 sprintf(ses->ip_addr, "%pI6", &addr6->sin6_addr);
2218 else
2219 sprintf(ses->ip_addr, "%pI4", &addr->sin_addr);
2220
2221 if (ctx->username) {
2222 ses->user_name = kstrdup(ctx->username, GFP_KERNEL);
2223 if (!ses->user_name)
2224 goto get_ses_fail;
2225 }
2226
2227 /* ctx->password freed at unmount */
2228 if (ctx->password) {
2229 ses->password = kstrdup(ctx->password, GFP_KERNEL);
2230 if (!ses->password)
2231 goto get_ses_fail;
2232 }
2233 if (ctx->domainname) {
2234 ses->domainName = kstrdup(ctx->domainname, GFP_KERNEL);
2235 if (!ses->domainName)
2236 goto get_ses_fail;
2237 }
2238
2239 strscpy(ses->workstation_name, ctx->workstation_name, sizeof(ses->workstation_name));
2240
2241 if (ctx->domainauto)
2242 ses->domainAuto = ctx->domainauto;
2243 ses->cred_uid = ctx->cred_uid;
2244 ses->linux_uid = ctx->linux_uid;
2245
2246 ses->sectype = ctx->sectype;
2247 ses->sign = ctx->sign;
2248
2249 /* add server as first channel */
2250 spin_lock(&ses->chan_lock);
2251 ses->chans[0].server = server;
2252 ses->chan_count = 1;
2253 ses->chan_max = ctx->multichannel ? ctx->max_channels:1;
2254 ses->chans_need_reconnect = 1;
2255 spin_unlock(&ses->chan_lock);
2256
2257 mutex_lock(&ses->session_mutex);
2258 rc = cifs_negotiate_protocol(xid, ses, server);
2259 if (!rc)
2260 rc = cifs_setup_session(xid, ses, server, ctx->local_nls);
2261 mutex_unlock(&ses->session_mutex);
2262
2263 /* each channel uses a different signing key */
2264 spin_lock(&ses->chan_lock);
2265 memcpy(ses->chans[0].signkey, ses->smb3signingkey,
2266 sizeof(ses->smb3signingkey));
2267 spin_unlock(&ses->chan_lock);
2268
2269 if (rc)
2270 goto get_ses_fail;
2271
2272 /*
2273 * success, put it on the list and add it as first channel
2274 * note: the session becomes active soon after this. So you'll
2275 * need to lock before changing something in the session.
2276 */
2277 spin_lock(&cifs_tcp_ses_lock);
2278 list_add(&ses->smb_ses_list, &server->smb_ses_list);
2279 spin_unlock(&cifs_tcp_ses_lock);
2280
2281 free_xid(xid);
2282
2283 cifs_setup_ipc(ses, ctx);
2284
2285 return ses;
2286
2287 get_ses_fail:
2288 sesInfoFree(ses);
2289 free_xid(xid);
2290 return ERR_PTR(rc);
2291 }
2292
2293 /* this function must be called with tc_lock held */
match_tcon(struct cifs_tcon * tcon,struct smb3_fs_context * ctx)2294 static int match_tcon(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
2295 {
2296 if (tcon->status == TID_EXITING)
2297 return 0;
2298 if (strncmp(tcon->tree_name, ctx->UNC, MAX_TREE_SIZE))
2299 return 0;
2300 if (tcon->seal != ctx->seal)
2301 return 0;
2302 if (tcon->snapshot_time != ctx->snapshot_time)
2303 return 0;
2304 if (tcon->handle_timeout != ctx->handle_timeout)
2305 return 0;
2306 if (tcon->no_lease != ctx->no_lease)
2307 return 0;
2308 if (tcon->nodelete != ctx->nodelete)
2309 return 0;
2310 return 1;
2311 }
2312
2313 static struct cifs_tcon *
cifs_find_tcon(struct cifs_ses * ses,struct smb3_fs_context * ctx)2314 cifs_find_tcon(struct cifs_ses *ses, struct smb3_fs_context *ctx)
2315 {
2316 struct cifs_tcon *tcon;
2317
2318 spin_lock(&cifs_tcp_ses_lock);
2319 list_for_each_entry(tcon, &ses->tcon_list, tcon_list) {
2320 spin_lock(&tcon->tc_lock);
2321 if (!match_tcon(tcon, ctx)) {
2322 spin_unlock(&tcon->tc_lock);
2323 continue;
2324 }
2325 ++tcon->tc_count;
2326 spin_unlock(&tcon->tc_lock);
2327 spin_unlock(&cifs_tcp_ses_lock);
2328 return tcon;
2329 }
2330 spin_unlock(&cifs_tcp_ses_lock);
2331 return NULL;
2332 }
2333
2334 void
cifs_put_tcon(struct cifs_tcon * tcon)2335 cifs_put_tcon(struct cifs_tcon *tcon)
2336 {
2337 unsigned int xid;
2338 struct cifs_ses *ses;
2339
2340 /*
2341 * IPC tcon share the lifetime of their session and are
2342 * destroyed in the session put function
2343 */
2344 if (tcon == NULL || tcon->ipc)
2345 return;
2346
2347 ses = tcon->ses;
2348 cifs_dbg(FYI, "%s: tc_count=%d\n", __func__, tcon->tc_count);
2349 spin_lock(&cifs_tcp_ses_lock);
2350 spin_lock(&tcon->tc_lock);
2351 if (--tcon->tc_count > 0) {
2352 spin_unlock(&tcon->tc_lock);
2353 spin_unlock(&cifs_tcp_ses_lock);
2354 return;
2355 }
2356
2357 /* tc_count can never go negative */
2358 WARN_ON(tcon->tc_count < 0);
2359
2360 list_del_init(&tcon->tcon_list);
2361 spin_unlock(&tcon->tc_lock);
2362 spin_unlock(&cifs_tcp_ses_lock);
2363
2364 /* cancel polling of interfaces */
2365 cancel_delayed_work_sync(&tcon->query_interfaces);
2366
2367 if (tcon->use_witness) {
2368 int rc;
2369
2370 rc = cifs_swn_unregister(tcon);
2371 if (rc < 0) {
2372 cifs_dbg(VFS, "%s: Failed to unregister for witness notifications: %d\n",
2373 __func__, rc);
2374 }
2375 }
2376
2377 xid = get_xid();
2378 if (ses->server->ops->tree_disconnect)
2379 ses->server->ops->tree_disconnect(xid, tcon);
2380 _free_xid(xid);
2381
2382 cifs_fscache_release_super_cookie(tcon);
2383 tconInfoFree(tcon);
2384 cifs_put_smb_ses(ses);
2385 }
2386
2387 /**
2388 * cifs_get_tcon - get a tcon matching @ctx data from @ses
2389 * @ses: smb session to issue the request on
2390 * @ctx: the superblock configuration context to use for building the
2391 *
2392 * - tcon refcount is the number of mount points using the tcon.
2393 * - ses refcount is the number of tcon using the session.
2394 *
2395 * 1. This function assumes it is being called from cifs_mount() where
2396 * we already got a session reference (ses refcount +1).
2397 *
2398 * 2. Since we're in the context of adding a mount point, the end
2399 * result should be either:
2400 *
2401 * a) a new tcon already allocated with refcount=1 (1 mount point) and
2402 * its session refcount incremented (1 new tcon). This +1 was
2403 * already done in (1).
2404 *
2405 * b) an existing tcon with refcount+1 (add a mount point to it) and
2406 * identical ses refcount (no new tcon). Because of (1) we need to
2407 * decrement the ses refcount.
2408 */
2409 static struct cifs_tcon *
cifs_get_tcon(struct cifs_ses * ses,struct smb3_fs_context * ctx)2410 cifs_get_tcon(struct cifs_ses *ses, struct smb3_fs_context *ctx)
2411 {
2412 int rc, xid;
2413 struct cifs_tcon *tcon;
2414
2415 tcon = cifs_find_tcon(ses, ctx);
2416 if (tcon) {
2417 /*
2418 * tcon has refcount already incremented but we need to
2419 * decrement extra ses reference gotten by caller (case b)
2420 */
2421 cifs_dbg(FYI, "Found match on UNC path\n");
2422 cifs_put_smb_ses(ses);
2423 return tcon;
2424 }
2425
2426 if (!ses->server->ops->tree_connect) {
2427 rc = -ENOSYS;
2428 goto out_fail;
2429 }
2430
2431 tcon = tconInfoAlloc();
2432 if (tcon == NULL) {
2433 rc = -ENOMEM;
2434 goto out_fail;
2435 }
2436
2437 if (ctx->snapshot_time) {
2438 if (ses->server->vals->protocol_id == 0) {
2439 cifs_dbg(VFS,
2440 "Use SMB2 or later for snapshot mount option\n");
2441 rc = -EOPNOTSUPP;
2442 goto out_fail;
2443 } else
2444 tcon->snapshot_time = ctx->snapshot_time;
2445 }
2446
2447 if (ctx->handle_timeout) {
2448 if (ses->server->vals->protocol_id == 0) {
2449 cifs_dbg(VFS,
2450 "Use SMB2.1 or later for handle timeout option\n");
2451 rc = -EOPNOTSUPP;
2452 goto out_fail;
2453 } else
2454 tcon->handle_timeout = ctx->handle_timeout;
2455 }
2456
2457 tcon->ses = ses;
2458 if (ctx->password) {
2459 tcon->password = kstrdup(ctx->password, GFP_KERNEL);
2460 if (!tcon->password) {
2461 rc = -ENOMEM;
2462 goto out_fail;
2463 }
2464 }
2465
2466 if (ctx->seal) {
2467 if (ses->server->vals->protocol_id == 0) {
2468 cifs_dbg(VFS,
2469 "SMB3 or later required for encryption\n");
2470 rc = -EOPNOTSUPP;
2471 goto out_fail;
2472 } else if (tcon->ses->server->capabilities &
2473 SMB2_GLOBAL_CAP_ENCRYPTION)
2474 tcon->seal = true;
2475 else {
2476 cifs_dbg(VFS, "Encryption is not supported on share\n");
2477 rc = -EOPNOTSUPP;
2478 goto out_fail;
2479 }
2480 }
2481
2482 if (ctx->linux_ext) {
2483 if (ses->server->posix_ext_supported) {
2484 tcon->posix_extensions = true;
2485 pr_warn_once("SMB3.11 POSIX Extensions are experimental\n");
2486 } else if ((ses->server->vals->protocol_id == SMB311_PROT_ID) ||
2487 (strcmp(ses->server->vals->version_string,
2488 SMB3ANY_VERSION_STRING) == 0) ||
2489 (strcmp(ses->server->vals->version_string,
2490 SMBDEFAULT_VERSION_STRING) == 0)) {
2491 cifs_dbg(VFS, "Server does not support mounting with posix SMB3.11 extensions\n");
2492 rc = -EOPNOTSUPP;
2493 goto out_fail;
2494 } else {
2495 cifs_dbg(VFS, "Check vers= mount option. SMB3.11 "
2496 "disabled but required for POSIX extensions\n");
2497 rc = -EOPNOTSUPP;
2498 goto out_fail;
2499 }
2500 }
2501
2502 xid = get_xid();
2503 rc = ses->server->ops->tree_connect(xid, ses, ctx->UNC, tcon,
2504 ctx->local_nls);
2505 free_xid(xid);
2506 cifs_dbg(FYI, "Tcon rc = %d\n", rc);
2507 if (rc)
2508 goto out_fail;
2509
2510 tcon->use_persistent = false;
2511 /* check if SMB2 or later, CIFS does not support persistent handles */
2512 if (ctx->persistent) {
2513 if (ses->server->vals->protocol_id == 0) {
2514 cifs_dbg(VFS,
2515 "SMB3 or later required for persistent handles\n");
2516 rc = -EOPNOTSUPP;
2517 goto out_fail;
2518 } else if (ses->server->capabilities &
2519 SMB2_GLOBAL_CAP_PERSISTENT_HANDLES)
2520 tcon->use_persistent = true;
2521 else /* persistent handles requested but not supported */ {
2522 cifs_dbg(VFS,
2523 "Persistent handles not supported on share\n");
2524 rc = -EOPNOTSUPP;
2525 goto out_fail;
2526 }
2527 } else if ((tcon->capabilities & SMB2_SHARE_CAP_CONTINUOUS_AVAILABILITY)
2528 && (ses->server->capabilities & SMB2_GLOBAL_CAP_PERSISTENT_HANDLES)
2529 && (ctx->nopersistent == false)) {
2530 cifs_dbg(FYI, "enabling persistent handles\n");
2531 tcon->use_persistent = true;
2532 } else if (ctx->resilient) {
2533 if (ses->server->vals->protocol_id == 0) {
2534 cifs_dbg(VFS,
2535 "SMB2.1 or later required for resilient handles\n");
2536 rc = -EOPNOTSUPP;
2537 goto out_fail;
2538 }
2539 tcon->use_resilient = true;
2540 }
2541
2542 tcon->use_witness = false;
2543 if (IS_ENABLED(CONFIG_CIFS_SWN_UPCALL) && ctx->witness) {
2544 if (ses->server->vals->protocol_id >= SMB30_PROT_ID) {
2545 if (tcon->capabilities & SMB2_SHARE_CAP_CLUSTER) {
2546 /*
2547 * Set witness in use flag in first place
2548 * to retry registration in the echo task
2549 */
2550 tcon->use_witness = true;
2551 /* And try to register immediately */
2552 rc = cifs_swn_register(tcon);
2553 if (rc < 0) {
2554 cifs_dbg(VFS, "Failed to register for witness notifications: %d\n", rc);
2555 goto out_fail;
2556 }
2557 } else {
2558 /* TODO: try to extend for non-cluster uses (eg multichannel) */
2559 cifs_dbg(VFS, "witness requested on mount but no CLUSTER capability on share\n");
2560 rc = -EOPNOTSUPP;
2561 goto out_fail;
2562 }
2563 } else {
2564 cifs_dbg(VFS, "SMB3 or later required for witness option\n");
2565 rc = -EOPNOTSUPP;
2566 goto out_fail;
2567 }
2568 }
2569
2570 /* If the user really knows what they are doing they can override */
2571 if (tcon->share_flags & SMB2_SHAREFLAG_NO_CACHING) {
2572 if (ctx->cache_ro)
2573 cifs_dbg(VFS, "cache=ro requested on mount but NO_CACHING flag set on share\n");
2574 else if (ctx->cache_rw)
2575 cifs_dbg(VFS, "cache=singleclient requested on mount but NO_CACHING flag set on share\n");
2576 }
2577
2578 if (ctx->no_lease) {
2579 if (ses->server->vals->protocol_id == 0) {
2580 cifs_dbg(VFS,
2581 "SMB2 or later required for nolease option\n");
2582 rc = -EOPNOTSUPP;
2583 goto out_fail;
2584 } else
2585 tcon->no_lease = ctx->no_lease;
2586 }
2587
2588 /*
2589 * We can have only one retry value for a connection to a share so for
2590 * resources mounted more than once to the same server share the last
2591 * value passed in for the retry flag is used.
2592 */
2593 tcon->retry = ctx->retry;
2594 tcon->nocase = ctx->nocase;
2595 tcon->broken_sparse_sup = ctx->no_sparse;
2596 if (ses->server->capabilities & SMB2_GLOBAL_CAP_DIRECTORY_LEASING)
2597 tcon->nohandlecache = ctx->nohandlecache;
2598 else
2599 tcon->nohandlecache = true;
2600 tcon->nodelete = ctx->nodelete;
2601 tcon->local_lease = ctx->local_lease;
2602 INIT_LIST_HEAD(&tcon->pending_opens);
2603
2604 /* schedule query interfaces poll */
2605 INIT_DELAYED_WORK(&tcon->query_interfaces,
2606 smb2_query_server_interfaces);
2607 queue_delayed_work(cifsiod_wq, &tcon->query_interfaces,
2608 (SMB_INTERFACE_POLL_INTERVAL * HZ));
2609
2610 spin_lock(&cifs_tcp_ses_lock);
2611 list_add(&tcon->tcon_list, &ses->tcon_list);
2612 spin_unlock(&cifs_tcp_ses_lock);
2613
2614 return tcon;
2615
2616 out_fail:
2617 tconInfoFree(tcon);
2618 return ERR_PTR(rc);
2619 }
2620
2621 void
cifs_put_tlink(struct tcon_link * tlink)2622 cifs_put_tlink(struct tcon_link *tlink)
2623 {
2624 if (!tlink || IS_ERR(tlink))
2625 return;
2626
2627 if (!atomic_dec_and_test(&tlink->tl_count) ||
2628 test_bit(TCON_LINK_IN_TREE, &tlink->tl_flags)) {
2629 tlink->tl_time = jiffies;
2630 return;
2631 }
2632
2633 if (!IS_ERR(tlink_tcon(tlink)))
2634 cifs_put_tcon(tlink_tcon(tlink));
2635 kfree(tlink);
2636 return;
2637 }
2638
2639 static int
compare_mount_options(struct super_block * sb,struct cifs_mnt_data * mnt_data)2640 compare_mount_options(struct super_block *sb, struct cifs_mnt_data *mnt_data)
2641 {
2642 struct cifs_sb_info *old = CIFS_SB(sb);
2643 struct cifs_sb_info *new = mnt_data->cifs_sb;
2644 unsigned int oldflags = old->mnt_cifs_flags & CIFS_MOUNT_MASK;
2645 unsigned int newflags = new->mnt_cifs_flags & CIFS_MOUNT_MASK;
2646
2647 if ((sb->s_flags & CIFS_MS_MASK) != (mnt_data->flags & CIFS_MS_MASK))
2648 return 0;
2649
2650 if (old->mnt_cifs_serverino_autodisabled)
2651 newflags &= ~CIFS_MOUNT_SERVER_INUM;
2652
2653 if (oldflags != newflags)
2654 return 0;
2655
2656 /*
2657 * We want to share sb only if we don't specify an r/wsize or
2658 * specified r/wsize is greater than or equal to existing one.
2659 */
2660 if (new->ctx->wsize && new->ctx->wsize < old->ctx->wsize)
2661 return 0;
2662
2663 if (new->ctx->rsize && new->ctx->rsize < old->ctx->rsize)
2664 return 0;
2665
2666 if (!uid_eq(old->ctx->linux_uid, new->ctx->linux_uid) ||
2667 !gid_eq(old->ctx->linux_gid, new->ctx->linux_gid))
2668 return 0;
2669
2670 if (old->ctx->file_mode != new->ctx->file_mode ||
2671 old->ctx->dir_mode != new->ctx->dir_mode)
2672 return 0;
2673
2674 if (strcmp(old->local_nls->charset, new->local_nls->charset))
2675 return 0;
2676
2677 if (old->ctx->acregmax != new->ctx->acregmax)
2678 return 0;
2679 if (old->ctx->acdirmax != new->ctx->acdirmax)
2680 return 0;
2681 if (old->ctx->closetimeo != new->ctx->closetimeo)
2682 return 0;
2683
2684 return 1;
2685 }
2686
2687 static int
match_prepath(struct super_block * sb,struct cifs_mnt_data * mnt_data)2688 match_prepath(struct super_block *sb, struct cifs_mnt_data *mnt_data)
2689 {
2690 struct cifs_sb_info *old = CIFS_SB(sb);
2691 struct cifs_sb_info *new = mnt_data->cifs_sb;
2692 bool old_set = (old->mnt_cifs_flags & CIFS_MOUNT_USE_PREFIX_PATH) &&
2693 old->prepath;
2694 bool new_set = (new->mnt_cifs_flags & CIFS_MOUNT_USE_PREFIX_PATH) &&
2695 new->prepath;
2696
2697 if (old_set && new_set && !strcmp(new->prepath, old->prepath))
2698 return 1;
2699 else if (!old_set && !new_set)
2700 return 1;
2701
2702 return 0;
2703 }
2704
2705 int
cifs_match_super(struct super_block * sb,void * data)2706 cifs_match_super(struct super_block *sb, void *data)
2707 {
2708 struct cifs_mnt_data *mnt_data = data;
2709 struct smb3_fs_context *ctx;
2710 struct cifs_sb_info *cifs_sb;
2711 struct TCP_Server_Info *tcp_srv;
2712 struct cifs_ses *ses;
2713 struct cifs_tcon *tcon;
2714 struct tcon_link *tlink;
2715 int rc = 0;
2716
2717 spin_lock(&cifs_tcp_ses_lock);
2718 cifs_sb = CIFS_SB(sb);
2719 tlink = cifs_get_tlink(cifs_sb_master_tlink(cifs_sb));
2720 if (tlink == NULL) {
2721 /* can not match superblock if tlink were ever null */
2722 spin_unlock(&cifs_tcp_ses_lock);
2723 return 0;
2724 }
2725 tcon = tlink_tcon(tlink);
2726 ses = tcon->ses;
2727 tcp_srv = ses->server;
2728
2729 ctx = mnt_data->ctx;
2730
2731 spin_lock(&tcp_srv->srv_lock);
2732 spin_lock(&ses->ses_lock);
2733 spin_lock(&tcon->tc_lock);
2734 if (!match_server(tcp_srv, ctx) ||
2735 !match_session(ses, ctx) ||
2736 !match_tcon(tcon, ctx) ||
2737 !match_prepath(sb, mnt_data)) {
2738 rc = 0;
2739 goto out;
2740 }
2741
2742 rc = compare_mount_options(sb, mnt_data);
2743 out:
2744 spin_unlock(&tcon->tc_lock);
2745 spin_unlock(&ses->ses_lock);
2746 spin_unlock(&tcp_srv->srv_lock);
2747
2748 spin_unlock(&cifs_tcp_ses_lock);
2749 cifs_put_tlink(tlink);
2750 return rc;
2751 }
2752
2753 #ifdef CONFIG_DEBUG_LOCK_ALLOC
2754 static struct lock_class_key cifs_key[2];
2755 static struct lock_class_key cifs_slock_key[2];
2756
2757 static inline void
cifs_reclassify_socket4(struct socket * sock)2758 cifs_reclassify_socket4(struct socket *sock)
2759 {
2760 struct sock *sk = sock->sk;
2761 BUG_ON(!sock_allow_reclassification(sk));
2762 sock_lock_init_class_and_name(sk, "slock-AF_INET-CIFS",
2763 &cifs_slock_key[0], "sk_lock-AF_INET-CIFS", &cifs_key[0]);
2764 }
2765
2766 static inline void
cifs_reclassify_socket6(struct socket * sock)2767 cifs_reclassify_socket6(struct socket *sock)
2768 {
2769 struct sock *sk = sock->sk;
2770 BUG_ON(!sock_allow_reclassification(sk));
2771 sock_lock_init_class_and_name(sk, "slock-AF_INET6-CIFS",
2772 &cifs_slock_key[1], "sk_lock-AF_INET6-CIFS", &cifs_key[1]);
2773 }
2774 #else
2775 static inline void
cifs_reclassify_socket4(struct socket * sock)2776 cifs_reclassify_socket4(struct socket *sock)
2777 {
2778 }
2779
2780 static inline void
cifs_reclassify_socket6(struct socket * sock)2781 cifs_reclassify_socket6(struct socket *sock)
2782 {
2783 }
2784 #endif
2785
2786 /* See RFC1001 section 14 on representation of Netbios names */
rfc1002mangle(char * target,char * source,unsigned int length)2787 static void rfc1002mangle(char *target, char *source, unsigned int length)
2788 {
2789 unsigned int i, j;
2790
2791 for (i = 0, j = 0; i < (length); i++) {
2792 /* mask a nibble at a time and encode */
2793 target[j] = 'A' + (0x0F & (source[i] >> 4));
2794 target[j+1] = 'A' + (0x0F & source[i]);
2795 j += 2;
2796 }
2797
2798 }
2799
2800 static int
bind_socket(struct TCP_Server_Info * server)2801 bind_socket(struct TCP_Server_Info *server)
2802 {
2803 int rc = 0;
2804 if (server->srcaddr.ss_family != AF_UNSPEC) {
2805 /* Bind to the specified local IP address */
2806 struct socket *socket = server->ssocket;
2807 rc = socket->ops->bind(socket,
2808 (struct sockaddr *) &server->srcaddr,
2809 sizeof(server->srcaddr));
2810 if (rc < 0) {
2811 struct sockaddr_in *saddr4;
2812 struct sockaddr_in6 *saddr6;
2813 saddr4 = (struct sockaddr_in *)&server->srcaddr;
2814 saddr6 = (struct sockaddr_in6 *)&server->srcaddr;
2815 if (saddr6->sin6_family == AF_INET6)
2816 cifs_server_dbg(VFS, "Failed to bind to: %pI6c, error: %d\n",
2817 &saddr6->sin6_addr, rc);
2818 else
2819 cifs_server_dbg(VFS, "Failed to bind to: %pI4, error: %d\n",
2820 &saddr4->sin_addr.s_addr, rc);
2821 }
2822 }
2823 return rc;
2824 }
2825
2826 static int
ip_rfc1001_connect(struct TCP_Server_Info * server)2827 ip_rfc1001_connect(struct TCP_Server_Info *server)
2828 {
2829 int rc = 0;
2830 /*
2831 * some servers require RFC1001 sessinit before sending
2832 * negprot - BB check reconnection in case where second
2833 * sessinit is sent but no second negprot
2834 */
2835 struct rfc1002_session_packet *ses_init_buf;
2836 unsigned int req_noscope_len;
2837 struct smb_hdr *smb_buf;
2838
2839 ses_init_buf = kzalloc(sizeof(struct rfc1002_session_packet),
2840 GFP_KERNEL);
2841
2842 if (ses_init_buf) {
2843 ses_init_buf->trailer.session_req.called_len = 32;
2844
2845 if (server->server_RFC1001_name[0] != 0)
2846 rfc1002mangle(ses_init_buf->trailer.
2847 session_req.called_name,
2848 server->server_RFC1001_name,
2849 RFC1001_NAME_LEN_WITH_NULL);
2850 else
2851 rfc1002mangle(ses_init_buf->trailer.
2852 session_req.called_name,
2853 DEFAULT_CIFS_CALLED_NAME,
2854 RFC1001_NAME_LEN_WITH_NULL);
2855
2856 ses_init_buf->trailer.session_req.calling_len = 32;
2857
2858 /*
2859 * calling name ends in null (byte 16) from old smb
2860 * convention.
2861 */
2862 if (server->workstation_RFC1001_name[0] != 0)
2863 rfc1002mangle(ses_init_buf->trailer.
2864 session_req.calling_name,
2865 server->workstation_RFC1001_name,
2866 RFC1001_NAME_LEN_WITH_NULL);
2867 else
2868 rfc1002mangle(ses_init_buf->trailer.
2869 session_req.calling_name,
2870 "LINUX_CIFS_CLNT",
2871 RFC1001_NAME_LEN_WITH_NULL);
2872
2873 ses_init_buf->trailer.session_req.scope1 = 0;
2874 ses_init_buf->trailer.session_req.scope2 = 0;
2875 smb_buf = (struct smb_hdr *)ses_init_buf;
2876
2877 /* sizeof RFC1002_SESSION_REQUEST with no scopes */
2878 req_noscope_len = sizeof(struct rfc1002_session_packet) - 2;
2879
2880 /* == cpu_to_be32(0x81000044) */
2881 smb_buf->smb_buf_length =
2882 cpu_to_be32((RFC1002_SESSION_REQUEST << 24) | req_noscope_len);
2883 rc = smb_send(server, smb_buf, 0x44);
2884 kfree(ses_init_buf);
2885 /*
2886 * RFC1001 layer in at least one server
2887 * requires very short break before negprot
2888 * presumably because not expecting negprot
2889 * to follow so fast. This is a simple
2890 * solution that works without
2891 * complicating the code and causes no
2892 * significant slowing down on mount
2893 * for everyone else
2894 */
2895 usleep_range(1000, 2000);
2896 }
2897 /*
2898 * else the negprot may still work without this
2899 * even though malloc failed
2900 */
2901
2902 return rc;
2903 }
2904
2905 static int
generic_ip_connect(struct TCP_Server_Info * server)2906 generic_ip_connect(struct TCP_Server_Info *server)
2907 {
2908 int rc = 0;
2909 __be16 sport;
2910 int slen, sfamily;
2911 struct socket *socket = server->ssocket;
2912 struct sockaddr *saddr;
2913
2914 saddr = (struct sockaddr *) &server->dstaddr;
2915
2916 if (server->dstaddr.ss_family == AF_INET6) {
2917 struct sockaddr_in6 *ipv6 = (struct sockaddr_in6 *)&server->dstaddr;
2918
2919 sport = ipv6->sin6_port;
2920 slen = sizeof(struct sockaddr_in6);
2921 sfamily = AF_INET6;
2922 cifs_dbg(FYI, "%s: connecting to [%pI6]:%d\n", __func__, &ipv6->sin6_addr,
2923 ntohs(sport));
2924 } else {
2925 struct sockaddr_in *ipv4 = (struct sockaddr_in *)&server->dstaddr;
2926
2927 sport = ipv4->sin_port;
2928 slen = sizeof(struct sockaddr_in);
2929 sfamily = AF_INET;
2930 cifs_dbg(FYI, "%s: connecting to %pI4:%d\n", __func__, &ipv4->sin_addr,
2931 ntohs(sport));
2932 }
2933
2934 if (socket == NULL) {
2935 rc = __sock_create(cifs_net_ns(server), sfamily, SOCK_STREAM,
2936 IPPROTO_TCP, &socket, 1);
2937 if (rc < 0) {
2938 cifs_server_dbg(VFS, "Error %d creating socket\n", rc);
2939 server->ssocket = NULL;
2940 return rc;
2941 }
2942
2943 /* BB other socket options to set KEEPALIVE, NODELAY? */
2944 cifs_dbg(FYI, "Socket created\n");
2945 server->ssocket = socket;
2946 socket->sk->sk_allocation = GFP_NOFS;
2947 if (sfamily == AF_INET6)
2948 cifs_reclassify_socket6(socket);
2949 else
2950 cifs_reclassify_socket4(socket);
2951 }
2952
2953 rc = bind_socket(server);
2954 if (rc < 0)
2955 return rc;
2956
2957 /*
2958 * Eventually check for other socket options to change from
2959 * the default. sock_setsockopt not used because it expects
2960 * user space buffer
2961 */
2962 socket->sk->sk_rcvtimeo = 7 * HZ;
2963 socket->sk->sk_sndtimeo = 5 * HZ;
2964
2965 /* make the bufsizes depend on wsize/rsize and max requests */
2966 if (server->noautotune) {
2967 if (socket->sk->sk_sndbuf < (200 * 1024))
2968 socket->sk->sk_sndbuf = 200 * 1024;
2969 if (socket->sk->sk_rcvbuf < (140 * 1024))
2970 socket->sk->sk_rcvbuf = 140 * 1024;
2971 }
2972
2973 if (server->tcp_nodelay)
2974 tcp_sock_set_nodelay(socket->sk);
2975
2976 cifs_dbg(FYI, "sndbuf %d rcvbuf %d rcvtimeo 0x%lx\n",
2977 socket->sk->sk_sndbuf,
2978 socket->sk->sk_rcvbuf, socket->sk->sk_rcvtimeo);
2979
2980 rc = socket->ops->connect(socket, saddr, slen,
2981 server->noblockcnt ? O_NONBLOCK : 0);
2982 /*
2983 * When mounting SMB root file systems, we do not want to block in
2984 * connect. Otherwise bail out and then let cifs_reconnect() perform
2985 * reconnect failover - if possible.
2986 */
2987 if (server->noblockcnt && rc == -EINPROGRESS)
2988 rc = 0;
2989 if (rc < 0) {
2990 cifs_dbg(FYI, "Error %d connecting to server\n", rc);
2991 trace_smb3_connect_err(server->hostname, server->conn_id, &server->dstaddr, rc);
2992 sock_release(socket);
2993 server->ssocket = NULL;
2994 return rc;
2995 }
2996 trace_smb3_connect_done(server->hostname, server->conn_id, &server->dstaddr);
2997 if (sport == htons(RFC1001_PORT))
2998 rc = ip_rfc1001_connect(server);
2999
3000 return rc;
3001 }
3002
3003 static int
ip_connect(struct TCP_Server_Info * server)3004 ip_connect(struct TCP_Server_Info *server)
3005 {
3006 __be16 *sport;
3007 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&server->dstaddr;
3008 struct sockaddr_in *addr = (struct sockaddr_in *)&server->dstaddr;
3009
3010 if (server->dstaddr.ss_family == AF_INET6)
3011 sport = &addr6->sin6_port;
3012 else
3013 sport = &addr->sin_port;
3014
3015 if (*sport == 0) {
3016 int rc;
3017
3018 /* try with 445 port at first */
3019 *sport = htons(CIFS_PORT);
3020
3021 rc = generic_ip_connect(server);
3022 if (rc >= 0)
3023 return rc;
3024
3025 /* if it failed, try with 139 port */
3026 *sport = htons(RFC1001_PORT);
3027 }
3028
3029 return generic_ip_connect(server);
3030 }
3031
3032 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
reset_cifs_unix_caps(unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb,struct smb3_fs_context * ctx)3033 void reset_cifs_unix_caps(unsigned int xid, struct cifs_tcon *tcon,
3034 struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx)
3035 {
3036 /*
3037 * If we are reconnecting then should we check to see if
3038 * any requested capabilities changed locally e.g. via
3039 * remount but we can not do much about it here
3040 * if they have (even if we could detect it by the following)
3041 * Perhaps we could add a backpointer to array of sb from tcon
3042 * or if we change to make all sb to same share the same
3043 * sb as NFS - then we only have one backpointer to sb.
3044 * What if we wanted to mount the server share twice once with
3045 * and once without posixacls or posix paths?
3046 */
3047 __u64 saved_cap = le64_to_cpu(tcon->fsUnixInfo.Capability);
3048
3049 if (ctx && ctx->no_linux_ext) {
3050 tcon->fsUnixInfo.Capability = 0;
3051 tcon->unix_ext = 0; /* Unix Extensions disabled */
3052 cifs_dbg(FYI, "Linux protocol extensions disabled\n");
3053 return;
3054 } else if (ctx)
3055 tcon->unix_ext = 1; /* Unix Extensions supported */
3056
3057 if (!tcon->unix_ext) {
3058 cifs_dbg(FYI, "Unix extensions disabled so not set on reconnect\n");
3059 return;
3060 }
3061
3062 if (!CIFSSMBQFSUnixInfo(xid, tcon)) {
3063 __u64 cap = le64_to_cpu(tcon->fsUnixInfo.Capability);
3064 cifs_dbg(FYI, "unix caps which server supports %lld\n", cap);
3065 /*
3066 * check for reconnect case in which we do not
3067 * want to change the mount behavior if we can avoid it
3068 */
3069 if (ctx == NULL) {
3070 /*
3071 * turn off POSIX ACL and PATHNAMES if not set
3072 * originally at mount time
3073 */
3074 if ((saved_cap & CIFS_UNIX_POSIX_ACL_CAP) == 0)
3075 cap &= ~CIFS_UNIX_POSIX_ACL_CAP;
3076 if ((saved_cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) == 0) {
3077 if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP)
3078 cifs_dbg(VFS, "POSIXPATH support change\n");
3079 cap &= ~CIFS_UNIX_POSIX_PATHNAMES_CAP;
3080 } else if ((cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) == 0) {
3081 cifs_dbg(VFS, "possible reconnect error\n");
3082 cifs_dbg(VFS, "server disabled POSIX path support\n");
3083 }
3084 }
3085
3086 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP)
3087 cifs_dbg(VFS, "per-share encryption not supported yet\n");
3088
3089 cap &= CIFS_UNIX_CAP_MASK;
3090 if (ctx && ctx->no_psx_acl)
3091 cap &= ~CIFS_UNIX_POSIX_ACL_CAP;
3092 else if (CIFS_UNIX_POSIX_ACL_CAP & cap) {
3093 cifs_dbg(FYI, "negotiated posix acl support\n");
3094 if (cifs_sb)
3095 cifs_sb->mnt_cifs_flags |=
3096 CIFS_MOUNT_POSIXACL;
3097 }
3098
3099 if (ctx && ctx->posix_paths == 0)
3100 cap &= ~CIFS_UNIX_POSIX_PATHNAMES_CAP;
3101 else if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) {
3102 cifs_dbg(FYI, "negotiate posix pathnames\n");
3103 if (cifs_sb)
3104 cifs_sb->mnt_cifs_flags |=
3105 CIFS_MOUNT_POSIX_PATHS;
3106 }
3107
3108 cifs_dbg(FYI, "Negotiate caps 0x%x\n", (int)cap);
3109 #ifdef CONFIG_CIFS_DEBUG2
3110 if (cap & CIFS_UNIX_FCNTL_CAP)
3111 cifs_dbg(FYI, "FCNTL cap\n");
3112 if (cap & CIFS_UNIX_EXTATTR_CAP)
3113 cifs_dbg(FYI, "EXTATTR cap\n");
3114 if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP)
3115 cifs_dbg(FYI, "POSIX path cap\n");
3116 if (cap & CIFS_UNIX_XATTR_CAP)
3117 cifs_dbg(FYI, "XATTR cap\n");
3118 if (cap & CIFS_UNIX_POSIX_ACL_CAP)
3119 cifs_dbg(FYI, "POSIX ACL cap\n");
3120 if (cap & CIFS_UNIX_LARGE_READ_CAP)
3121 cifs_dbg(FYI, "very large read cap\n");
3122 if (cap & CIFS_UNIX_LARGE_WRITE_CAP)
3123 cifs_dbg(FYI, "very large write cap\n");
3124 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_CAP)
3125 cifs_dbg(FYI, "transport encryption cap\n");
3126 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP)
3127 cifs_dbg(FYI, "mandatory transport encryption cap\n");
3128 #endif /* CIFS_DEBUG2 */
3129 if (CIFSSMBSetFSUnixInfo(xid, tcon, cap)) {
3130 if (ctx == NULL)
3131 cifs_dbg(FYI, "resetting capabilities failed\n");
3132 else
3133 cifs_dbg(VFS, "Negotiating Unix capabilities with the server failed. Consider mounting with the Unix Extensions disabled if problems are found by specifying the nounix mount option.\n");
3134
3135 }
3136 }
3137 }
3138 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
3139
cifs_setup_cifs_sb(struct cifs_sb_info * cifs_sb)3140 int cifs_setup_cifs_sb(struct cifs_sb_info *cifs_sb)
3141 {
3142 struct smb3_fs_context *ctx = cifs_sb->ctx;
3143
3144 INIT_DELAYED_WORK(&cifs_sb->prune_tlinks, cifs_prune_tlinks);
3145
3146 spin_lock_init(&cifs_sb->tlink_tree_lock);
3147 cifs_sb->tlink_tree = RB_ROOT;
3148
3149 cifs_dbg(FYI, "file mode: %04ho dir mode: %04ho\n",
3150 ctx->file_mode, ctx->dir_mode);
3151
3152 /* this is needed for ASCII cp to Unicode converts */
3153 if (ctx->iocharset == NULL) {
3154 /* load_nls_default cannot return null */
3155 cifs_sb->local_nls = load_nls_default();
3156 } else {
3157 cifs_sb->local_nls = load_nls(ctx->iocharset);
3158 if (cifs_sb->local_nls == NULL) {
3159 cifs_dbg(VFS, "CIFS mount error: iocharset %s not found\n",
3160 ctx->iocharset);
3161 return -ELIBACC;
3162 }
3163 }
3164 ctx->local_nls = cifs_sb->local_nls;
3165
3166 smb3_update_mnt_flags(cifs_sb);
3167
3168 if (ctx->direct_io)
3169 cifs_dbg(FYI, "mounting share using direct i/o\n");
3170 if (ctx->cache_ro) {
3171 cifs_dbg(VFS, "mounting share with read only caching. Ensure that the share will not be modified while in use.\n");
3172 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_RO_CACHE;
3173 } else if (ctx->cache_rw) {
3174 cifs_dbg(VFS, "mounting share in single client RW caching mode. Ensure that no other systems will be accessing the share.\n");
3175 cifs_sb->mnt_cifs_flags |= (CIFS_MOUNT_RO_CACHE |
3176 CIFS_MOUNT_RW_CACHE);
3177 }
3178
3179 if ((ctx->cifs_acl) && (ctx->dynperm))
3180 cifs_dbg(VFS, "mount option dynperm ignored if cifsacl mount option supported\n");
3181
3182 if (ctx->prepath) {
3183 cifs_sb->prepath = kstrdup(ctx->prepath, GFP_KERNEL);
3184 if (cifs_sb->prepath == NULL)
3185 return -ENOMEM;
3186 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH;
3187 }
3188
3189 return 0;
3190 }
3191
3192 /* Release all succeed connections */
mount_put_conns(struct mount_ctx * mnt_ctx)3193 static inline void mount_put_conns(struct mount_ctx *mnt_ctx)
3194 {
3195 int rc = 0;
3196
3197 if (mnt_ctx->tcon)
3198 cifs_put_tcon(mnt_ctx->tcon);
3199 else if (mnt_ctx->ses)
3200 cifs_put_smb_ses(mnt_ctx->ses);
3201 else if (mnt_ctx->server)
3202 cifs_put_tcp_session(mnt_ctx->server, 0);
3203 mnt_ctx->cifs_sb->mnt_cifs_flags &= ~CIFS_MOUNT_POSIX_PATHS;
3204 free_xid(mnt_ctx->xid);
3205 }
3206
3207 /* Get connections for tcp, ses and tcon */
mount_get_conns(struct mount_ctx * mnt_ctx)3208 static int mount_get_conns(struct mount_ctx *mnt_ctx)
3209 {
3210 int rc = 0;
3211 struct TCP_Server_Info *server = NULL;
3212 struct cifs_ses *ses = NULL;
3213 struct cifs_tcon *tcon = NULL;
3214 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx;
3215 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb;
3216 unsigned int xid;
3217
3218 xid = get_xid();
3219
3220 /* get a reference to a tcp session */
3221 server = cifs_get_tcp_session(ctx, NULL);
3222 if (IS_ERR(server)) {
3223 rc = PTR_ERR(server);
3224 server = NULL;
3225 goto out;
3226 }
3227
3228 /* get a reference to a SMB session */
3229 ses = cifs_get_smb_ses(server, ctx);
3230 if (IS_ERR(ses)) {
3231 rc = PTR_ERR(ses);
3232 ses = NULL;
3233 goto out;
3234 }
3235
3236 if ((ctx->persistent == true) && (!(ses->server->capabilities &
3237 SMB2_GLOBAL_CAP_PERSISTENT_HANDLES))) {
3238 cifs_server_dbg(VFS, "persistent handles not supported by server\n");
3239 rc = -EOPNOTSUPP;
3240 goto out;
3241 }
3242
3243 /* search for existing tcon to this server share */
3244 tcon = cifs_get_tcon(ses, ctx);
3245 if (IS_ERR(tcon)) {
3246 rc = PTR_ERR(tcon);
3247 tcon = NULL;
3248 goto out;
3249 }
3250
3251 /* if new SMB3.11 POSIX extensions are supported do not remap / and \ */
3252 if (tcon->posix_extensions)
3253 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_POSIX_PATHS;
3254
3255 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
3256 /* tell server which Unix caps we support */
3257 if (cap_unix(tcon->ses)) {
3258 /*
3259 * reset of caps checks mount to see if unix extensions disabled
3260 * for just this mount.
3261 */
3262 reset_cifs_unix_caps(xid, tcon, cifs_sb, ctx);
3263 spin_lock(&tcon->ses->server->srv_lock);
3264 if ((tcon->ses->server->tcpStatus == CifsNeedReconnect) &&
3265 (le64_to_cpu(tcon->fsUnixInfo.Capability) &
3266 CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP)) {
3267 spin_unlock(&tcon->ses->server->srv_lock);
3268 rc = -EACCES;
3269 goto out;
3270 }
3271 spin_unlock(&tcon->ses->server->srv_lock);
3272 } else
3273 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
3274 tcon->unix_ext = 0; /* server does not support them */
3275
3276 /* do not care if a following call succeed - informational */
3277 if (!tcon->pipe && server->ops->qfs_tcon) {
3278 server->ops->qfs_tcon(xid, tcon, cifs_sb);
3279 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RO_CACHE) {
3280 if (tcon->fsDevInfo.DeviceCharacteristics &
3281 cpu_to_le32(FILE_READ_ONLY_DEVICE))
3282 cifs_dbg(VFS, "mounted to read only share\n");
3283 else if ((cifs_sb->mnt_cifs_flags &
3284 CIFS_MOUNT_RW_CACHE) == 0)
3285 cifs_dbg(VFS, "read only mount of RW share\n");
3286 /* no need to log a RW mount of a typical RW share */
3287 }
3288 }
3289
3290 /*
3291 * Clamp the rsize/wsize mount arguments if they are too big for the server
3292 * and set the rsize/wsize to the negotiated values if not passed in by
3293 * the user on mount
3294 */
3295 if ((cifs_sb->ctx->wsize == 0) ||
3296 (cifs_sb->ctx->wsize > server->ops->negotiate_wsize(tcon, ctx)))
3297 cifs_sb->ctx->wsize = server->ops->negotiate_wsize(tcon, ctx);
3298 if ((cifs_sb->ctx->rsize == 0) ||
3299 (cifs_sb->ctx->rsize > server->ops->negotiate_rsize(tcon, ctx)))
3300 cifs_sb->ctx->rsize = server->ops->negotiate_rsize(tcon, ctx);
3301
3302 /*
3303 * The cookie is initialized from volume info returned above.
3304 * Inside cifs_fscache_get_super_cookie it checks
3305 * that we do not get super cookie twice.
3306 */
3307 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_FSCACHE)
3308 cifs_fscache_get_super_cookie(tcon);
3309
3310 out:
3311 mnt_ctx->server = server;
3312 mnt_ctx->ses = ses;
3313 mnt_ctx->tcon = tcon;
3314 mnt_ctx->xid = xid;
3315
3316 return rc;
3317 }
3318
mount_setup_tlink(struct cifs_sb_info * cifs_sb,struct cifs_ses * ses,struct cifs_tcon * tcon)3319 static int mount_setup_tlink(struct cifs_sb_info *cifs_sb, struct cifs_ses *ses,
3320 struct cifs_tcon *tcon)
3321 {
3322 struct tcon_link *tlink;
3323
3324 /* hang the tcon off of the superblock */
3325 tlink = kzalloc(sizeof(*tlink), GFP_KERNEL);
3326 if (tlink == NULL)
3327 return -ENOMEM;
3328
3329 tlink->tl_uid = ses->linux_uid;
3330 tlink->tl_tcon = tcon;
3331 tlink->tl_time = jiffies;
3332 set_bit(TCON_LINK_MASTER, &tlink->tl_flags);
3333 set_bit(TCON_LINK_IN_TREE, &tlink->tl_flags);
3334
3335 cifs_sb->master_tlink = tlink;
3336 spin_lock(&cifs_sb->tlink_tree_lock);
3337 tlink_rb_insert(&cifs_sb->tlink_tree, tlink);
3338 spin_unlock(&cifs_sb->tlink_tree_lock);
3339
3340 queue_delayed_work(cifsiod_wq, &cifs_sb->prune_tlinks,
3341 TLINK_IDLE_EXPIRE);
3342 return 0;
3343 }
3344
3345 #ifdef CONFIG_CIFS_DFS_UPCALL
3346 /* Get unique dfs connections */
mount_get_dfs_conns(struct mount_ctx * mnt_ctx)3347 static int mount_get_dfs_conns(struct mount_ctx *mnt_ctx)
3348 {
3349 int rc;
3350
3351 mnt_ctx->fs_ctx->nosharesock = true;
3352 rc = mount_get_conns(mnt_ctx);
3353 if (mnt_ctx->server) {
3354 cifs_dbg(FYI, "%s: marking tcp session as a dfs connection\n", __func__);
3355 spin_lock(&mnt_ctx->server->srv_lock);
3356 mnt_ctx->server->is_dfs_conn = true;
3357 spin_unlock(&mnt_ctx->server->srv_lock);
3358 }
3359 return rc;
3360 }
3361
3362 /*
3363 * cifs_build_path_to_root returns full path to root when we do not have an
3364 * existing connection (tcon)
3365 */
3366 static char *
build_unc_path_to_root(const struct smb3_fs_context * ctx,const struct cifs_sb_info * cifs_sb,bool useppath)3367 build_unc_path_to_root(const struct smb3_fs_context *ctx,
3368 const struct cifs_sb_info *cifs_sb, bool useppath)
3369 {
3370 char *full_path, *pos;
3371 unsigned int pplen = useppath && ctx->prepath ?
3372 strlen(ctx->prepath) + 1 : 0;
3373 unsigned int unc_len = strnlen(ctx->UNC, MAX_TREE_SIZE + 1);
3374
3375 if (unc_len > MAX_TREE_SIZE)
3376 return ERR_PTR(-EINVAL);
3377
3378 full_path = kmalloc(unc_len + pplen + 1, GFP_KERNEL);
3379 if (full_path == NULL)
3380 return ERR_PTR(-ENOMEM);
3381
3382 memcpy(full_path, ctx->UNC, unc_len);
3383 pos = full_path + unc_len;
3384
3385 if (pplen) {
3386 *pos = CIFS_DIR_SEP(cifs_sb);
3387 memcpy(pos + 1, ctx->prepath, pplen);
3388 pos += pplen;
3389 }
3390
3391 *pos = '\0'; /* add trailing null */
3392 convert_delimiter(full_path, CIFS_DIR_SEP(cifs_sb));
3393 cifs_dbg(FYI, "%s: full_path=%s\n", __func__, full_path);
3394 return full_path;
3395 }
3396
3397 /*
3398 * expand_dfs_referral - Update cifs_sb from dfs referral path
3399 *
3400 * cifs_sb->ctx->mount_options will be (re-)allocated to a string containing updated options for the
3401 * submount. Otherwise it will be left untouched.
3402 */
expand_dfs_referral(struct mount_ctx * mnt_ctx,const char * full_path,struct dfs_info3_param * referral)3403 static int expand_dfs_referral(struct mount_ctx *mnt_ctx, const char *full_path,
3404 struct dfs_info3_param *referral)
3405 {
3406 int rc;
3407 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb;
3408 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx;
3409 char *fake_devname = NULL, *mdata = NULL;
3410
3411 mdata = cifs_compose_mount_options(cifs_sb->ctx->mount_options, full_path + 1, referral,
3412 &fake_devname);
3413 if (IS_ERR(mdata)) {
3414 rc = PTR_ERR(mdata);
3415 mdata = NULL;
3416 } else {
3417 /*
3418 * We can not clear out the whole structure since we no longer have an explicit
3419 * function to parse a mount-string. Instead we need to clear out the individual
3420 * fields that are no longer valid.
3421 */
3422 kfree(ctx->prepath);
3423 ctx->prepath = NULL;
3424 rc = cifs_setup_volume_info(ctx, mdata, fake_devname);
3425 }
3426 kfree(fake_devname);
3427 kfree(cifs_sb->ctx->mount_options);
3428 cifs_sb->ctx->mount_options = mdata;
3429
3430 return rc;
3431 }
3432 #endif
3433
3434 /* TODO: all callers to this are broken. We are not parsing mount_options here
3435 * we should pass a clone of the original context?
3436 */
3437 int
cifs_setup_volume_info(struct smb3_fs_context * ctx,const char * mntopts,const char * devname)3438 cifs_setup_volume_info(struct smb3_fs_context *ctx, const char *mntopts, const char *devname)
3439 {
3440 int rc;
3441
3442 if (devname) {
3443 cifs_dbg(FYI, "%s: devname=%s\n", __func__, devname);
3444 rc = smb3_parse_devname(devname, ctx);
3445 if (rc) {
3446 cifs_dbg(VFS, "%s: failed to parse %s: %d\n", __func__, devname, rc);
3447 return rc;
3448 }
3449 }
3450
3451 if (mntopts) {
3452 char *ip;
3453
3454 rc = smb3_parse_opt(mntopts, "ip", &ip);
3455 if (rc) {
3456 cifs_dbg(VFS, "%s: failed to parse ip options: %d\n", __func__, rc);
3457 return rc;
3458 }
3459
3460 rc = cifs_convert_address((struct sockaddr *)&ctx->dstaddr, ip, strlen(ip));
3461 kfree(ip);
3462 if (!rc) {
3463 cifs_dbg(VFS, "%s: failed to convert ip address\n", __func__);
3464 return -EINVAL;
3465 }
3466 }
3467
3468 if (ctx->nullauth) {
3469 cifs_dbg(FYI, "Anonymous login\n");
3470 kfree(ctx->username);
3471 ctx->username = NULL;
3472 } else if (ctx->username) {
3473 /* BB fixme parse for domain name here */
3474 cifs_dbg(FYI, "Username: %s\n", ctx->username);
3475 } else {
3476 cifs_dbg(VFS, "No username specified\n");
3477 /* In userspace mount helper we can get user name from alternate
3478 locations such as env variables and files on disk */
3479 return -EINVAL;
3480 }
3481
3482 return 0;
3483 }
3484
3485 static int
cifs_are_all_path_components_accessible(struct TCP_Server_Info * server,unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb,char * full_path,int added_treename)3486 cifs_are_all_path_components_accessible(struct TCP_Server_Info *server,
3487 unsigned int xid,
3488 struct cifs_tcon *tcon,
3489 struct cifs_sb_info *cifs_sb,
3490 char *full_path,
3491 int added_treename)
3492 {
3493 int rc;
3494 char *s;
3495 char sep, tmp;
3496 int skip = added_treename ? 1 : 0;
3497
3498 sep = CIFS_DIR_SEP(cifs_sb);
3499 s = full_path;
3500
3501 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb, "");
3502 while (rc == 0) {
3503 /* skip separators */
3504 while (*s == sep)
3505 s++;
3506 if (!*s)
3507 break;
3508 /* next separator */
3509 while (*s && *s != sep)
3510 s++;
3511 /*
3512 * if the treename is added, we then have to skip the first
3513 * part within the separators
3514 */
3515 if (skip) {
3516 skip = 0;
3517 continue;
3518 }
3519 /*
3520 * temporarily null-terminate the path at the end of
3521 * the current component
3522 */
3523 tmp = *s;
3524 *s = 0;
3525 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb,
3526 full_path);
3527 *s = tmp;
3528 }
3529 return rc;
3530 }
3531
3532 /*
3533 * Check if path is remote (i.e. a DFS share).
3534 *
3535 * Return -EREMOTE if it is, otherwise 0 or -errno.
3536 */
is_path_remote(struct mount_ctx * mnt_ctx)3537 static int is_path_remote(struct mount_ctx *mnt_ctx)
3538 {
3539 int rc;
3540 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb;
3541 struct TCP_Server_Info *server = mnt_ctx->server;
3542 unsigned int xid = mnt_ctx->xid;
3543 struct cifs_tcon *tcon = mnt_ctx->tcon;
3544 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx;
3545 char *full_path;
3546 #ifdef CONFIG_CIFS_DFS_UPCALL
3547 bool nodfs = cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_DFS;
3548 #endif
3549
3550 if (!server->ops->is_path_accessible)
3551 return -EOPNOTSUPP;
3552
3553 /*
3554 * cifs_build_path_to_root works only when we have a valid tcon
3555 */
3556 full_path = cifs_build_path_to_root(ctx, cifs_sb, tcon,
3557 tcon->Flags & SMB_SHARE_IS_IN_DFS);
3558 if (full_path == NULL)
3559 return -ENOMEM;
3560
3561 cifs_dbg(FYI, "%s: full_path: %s\n", __func__, full_path);
3562
3563 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb,
3564 full_path);
3565 #ifdef CONFIG_CIFS_DFS_UPCALL
3566 if (nodfs) {
3567 if (rc == -EREMOTE)
3568 rc = -EOPNOTSUPP;
3569 goto out;
3570 }
3571
3572 /* path *might* exist with non-ASCII characters in DFS root
3573 * try again with full path (only if nodfs is not set) */
3574 if (rc == -ENOENT && is_tcon_dfs(tcon))
3575 rc = cifs_dfs_query_info_nonascii_quirk(xid, tcon, cifs_sb,
3576 full_path);
3577 #endif
3578 if (rc != 0 && rc != -EREMOTE)
3579 goto out;
3580
3581 if (rc != -EREMOTE) {
3582 rc = cifs_are_all_path_components_accessible(server, xid, tcon,
3583 cifs_sb, full_path, tcon->Flags & SMB_SHARE_IS_IN_DFS);
3584 if (rc != 0) {
3585 cifs_server_dbg(VFS, "cannot query dirs between root and final path, enabling CIFS_MOUNT_USE_PREFIX_PATH\n");
3586 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH;
3587 rc = 0;
3588 }
3589 }
3590
3591 out:
3592 kfree(full_path);
3593 return rc;
3594 }
3595
3596 #ifdef CONFIG_CIFS_DFS_UPCALL
set_root_ses(struct mount_ctx * mnt_ctx)3597 static void set_root_ses(struct mount_ctx *mnt_ctx)
3598 {
3599 if (mnt_ctx->ses) {
3600 spin_lock(&cifs_tcp_ses_lock);
3601 mnt_ctx->ses->ses_count++;
3602 spin_unlock(&cifs_tcp_ses_lock);
3603 dfs_cache_add_refsrv_session(&mnt_ctx->mount_id, mnt_ctx->ses);
3604 }
3605 mnt_ctx->root_ses = mnt_ctx->ses;
3606 }
3607
is_dfs_mount(struct mount_ctx * mnt_ctx,bool * isdfs,struct dfs_cache_tgt_list * root_tl)3608 static int is_dfs_mount(struct mount_ctx *mnt_ctx, bool *isdfs, struct dfs_cache_tgt_list *root_tl)
3609 {
3610 int rc;
3611 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb;
3612 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx;
3613
3614 *isdfs = true;
3615
3616 rc = mount_get_conns(mnt_ctx);
3617 /*
3618 * If called with 'nodfs' mount option, then skip DFS resolving. Otherwise unconditionally
3619 * try to get an DFS referral (even cached) to determine whether it is an DFS mount.
3620 *
3621 * Skip prefix path to provide support for DFS referrals from w2k8 servers which don't seem
3622 * to respond with PATH_NOT_COVERED to requests that include the prefix.
3623 */
3624 if ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_DFS) ||
3625 dfs_cache_find(mnt_ctx->xid, mnt_ctx->ses, cifs_sb->local_nls, cifs_remap(cifs_sb),
3626 ctx->UNC + 1, NULL, root_tl)) {
3627 if (rc)
3628 return rc;
3629 /* Check if it is fully accessible and then mount it */
3630 rc = is_path_remote(mnt_ctx);
3631 if (!rc)
3632 *isdfs = false;
3633 else if (rc != -EREMOTE)
3634 return rc;
3635 }
3636 return 0;
3637 }
3638
connect_dfs_target(struct mount_ctx * mnt_ctx,const char * full_path,const char * ref_path,struct dfs_cache_tgt_iterator * tit)3639 static int connect_dfs_target(struct mount_ctx *mnt_ctx, const char *full_path,
3640 const char *ref_path, struct dfs_cache_tgt_iterator *tit)
3641 {
3642 int rc;
3643 struct dfs_info3_param ref = {};
3644 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb;
3645 char *oldmnt = cifs_sb->ctx->mount_options;
3646
3647 cifs_dbg(FYI, "%s: full_path=%s ref_path=%s target=%s\n", __func__, full_path, ref_path,
3648 dfs_cache_get_tgt_name(tit));
3649
3650 rc = dfs_cache_get_tgt_referral(ref_path, tit, &ref);
3651 if (rc)
3652 goto out;
3653
3654 rc = expand_dfs_referral(mnt_ctx, full_path, &ref);
3655 if (rc)
3656 goto out;
3657
3658 /* Connect to new target only if we were redirected (e.g. mount options changed) */
3659 if (oldmnt != cifs_sb->ctx->mount_options) {
3660 mount_put_conns(mnt_ctx);
3661 rc = mount_get_dfs_conns(mnt_ctx);
3662 }
3663 if (!rc) {
3664 if (cifs_is_referral_server(mnt_ctx->tcon, &ref))
3665 set_root_ses(mnt_ctx);
3666 rc = dfs_cache_update_tgthint(mnt_ctx->xid, mnt_ctx->root_ses, cifs_sb->local_nls,
3667 cifs_remap(cifs_sb), ref_path, tit);
3668 }
3669
3670 out:
3671 free_dfs_info_param(&ref);
3672 return rc;
3673 }
3674
connect_dfs_root(struct mount_ctx * mnt_ctx,struct dfs_cache_tgt_list * root_tl)3675 static int connect_dfs_root(struct mount_ctx *mnt_ctx, struct dfs_cache_tgt_list *root_tl)
3676 {
3677 int rc;
3678 char *full_path;
3679 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb;
3680 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx;
3681 struct dfs_cache_tgt_iterator *tit;
3682
3683 /* Put initial connections as they might be shared with other mounts. We need unique dfs
3684 * connections per mount to properly failover, so mount_get_dfs_conns() must be used from
3685 * now on.
3686 */
3687 mount_put_conns(mnt_ctx);
3688 mount_get_dfs_conns(mnt_ctx);
3689 set_root_ses(mnt_ctx);
3690
3691 full_path = build_unc_path_to_root(ctx, cifs_sb, true);
3692 if (IS_ERR(full_path))
3693 return PTR_ERR(full_path);
3694
3695 mnt_ctx->origin_fullpath = dfs_cache_canonical_path(ctx->UNC, cifs_sb->local_nls,
3696 cifs_remap(cifs_sb));
3697 if (IS_ERR(mnt_ctx->origin_fullpath)) {
3698 rc = PTR_ERR(mnt_ctx->origin_fullpath);
3699 mnt_ctx->origin_fullpath = NULL;
3700 goto out;
3701 }
3702
3703 /* Try all dfs root targets */
3704 for (rc = -ENOENT, tit = dfs_cache_get_tgt_iterator(root_tl);
3705 tit; tit = dfs_cache_get_next_tgt(root_tl, tit)) {
3706 rc = connect_dfs_target(mnt_ctx, full_path, mnt_ctx->origin_fullpath + 1, tit);
3707 if (!rc) {
3708 mnt_ctx->leaf_fullpath = kstrdup(mnt_ctx->origin_fullpath, GFP_KERNEL);
3709 if (!mnt_ctx->leaf_fullpath)
3710 rc = -ENOMEM;
3711 break;
3712 }
3713 }
3714
3715 out:
3716 kfree(full_path);
3717 return rc;
3718 }
3719
__follow_dfs_link(struct mount_ctx * mnt_ctx)3720 static int __follow_dfs_link(struct mount_ctx *mnt_ctx)
3721 {
3722 int rc;
3723 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb;
3724 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx;
3725 char *full_path;
3726 struct dfs_cache_tgt_list tl = DFS_CACHE_TGT_LIST_INIT(tl);
3727 struct dfs_cache_tgt_iterator *tit;
3728
3729 full_path = build_unc_path_to_root(ctx, cifs_sb, true);
3730 if (IS_ERR(full_path))
3731 return PTR_ERR(full_path);
3732
3733 kfree(mnt_ctx->leaf_fullpath);
3734 mnt_ctx->leaf_fullpath = dfs_cache_canonical_path(full_path, cifs_sb->local_nls,
3735 cifs_remap(cifs_sb));
3736 if (IS_ERR(mnt_ctx->leaf_fullpath)) {
3737 rc = PTR_ERR(mnt_ctx->leaf_fullpath);
3738 mnt_ctx->leaf_fullpath = NULL;
3739 goto out;
3740 }
3741
3742 /* Get referral from dfs link */
3743 rc = dfs_cache_find(mnt_ctx->xid, mnt_ctx->root_ses, cifs_sb->local_nls,
3744 cifs_remap(cifs_sb), mnt_ctx->leaf_fullpath + 1, NULL, &tl);
3745 if (rc)
3746 goto out;
3747
3748 /* Try all dfs link targets. If an I/O fails from currently connected DFS target with an
3749 * error other than STATUS_PATH_NOT_COVERED (-EREMOTE), then retry it from other targets as
3750 * specified in MS-DFSC "3.1.5.2 I/O Operation to Target Fails with an Error Other Than
3751 * STATUS_PATH_NOT_COVERED."
3752 */
3753 for (rc = -ENOENT, tit = dfs_cache_get_tgt_iterator(&tl);
3754 tit; tit = dfs_cache_get_next_tgt(&tl, tit)) {
3755 rc = connect_dfs_target(mnt_ctx, full_path, mnt_ctx->leaf_fullpath + 1, tit);
3756 if (!rc) {
3757 rc = is_path_remote(mnt_ctx);
3758 if (!rc || rc == -EREMOTE)
3759 break;
3760 }
3761 }
3762
3763 out:
3764 kfree(full_path);
3765 dfs_cache_free_tgts(&tl);
3766 return rc;
3767 }
3768
follow_dfs_link(struct mount_ctx * mnt_ctx)3769 static int follow_dfs_link(struct mount_ctx *mnt_ctx)
3770 {
3771 int rc;
3772 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb;
3773 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx;
3774 char *full_path;
3775 int num_links = 0;
3776
3777 full_path = build_unc_path_to_root(ctx, cifs_sb, true);
3778 if (IS_ERR(full_path))
3779 return PTR_ERR(full_path);
3780
3781 kfree(mnt_ctx->origin_fullpath);
3782 mnt_ctx->origin_fullpath = dfs_cache_canonical_path(full_path, cifs_sb->local_nls,
3783 cifs_remap(cifs_sb));
3784 kfree(full_path);
3785
3786 if (IS_ERR(mnt_ctx->origin_fullpath)) {
3787 rc = PTR_ERR(mnt_ctx->origin_fullpath);
3788 mnt_ctx->origin_fullpath = NULL;
3789 return rc;
3790 }
3791
3792 do {
3793 rc = __follow_dfs_link(mnt_ctx);
3794 if (!rc || rc != -EREMOTE)
3795 break;
3796 } while (rc = -ELOOP, ++num_links < MAX_NESTED_LINKS);
3797
3798 return rc;
3799 }
3800
3801 /* Set up DFS referral paths for failover */
setup_server_referral_paths(struct mount_ctx * mnt_ctx)3802 static void setup_server_referral_paths(struct mount_ctx *mnt_ctx)
3803 {
3804 struct TCP_Server_Info *server = mnt_ctx->server;
3805
3806 mutex_lock(&server->refpath_lock);
3807 server->origin_fullpath = mnt_ctx->origin_fullpath;
3808 server->leaf_fullpath = mnt_ctx->leaf_fullpath;
3809 server->current_fullpath = mnt_ctx->leaf_fullpath;
3810 mutex_unlock(&server->refpath_lock);
3811 mnt_ctx->origin_fullpath = mnt_ctx->leaf_fullpath = NULL;
3812 }
3813
cifs_mount(struct cifs_sb_info * cifs_sb,struct smb3_fs_context * ctx)3814 int cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx)
3815 {
3816 int rc;
3817 struct mount_ctx mnt_ctx = { .cifs_sb = cifs_sb, .fs_ctx = ctx, };
3818 struct dfs_cache_tgt_list tl = DFS_CACHE_TGT_LIST_INIT(tl);
3819 bool isdfs;
3820
3821 rc = is_dfs_mount(&mnt_ctx, &isdfs, &tl);
3822 if (rc)
3823 goto error;
3824 if (!isdfs)
3825 goto out;
3826
3827 /* proceed as DFS mount */
3828 uuid_gen(&mnt_ctx.mount_id);
3829 rc = connect_dfs_root(&mnt_ctx, &tl);
3830 dfs_cache_free_tgts(&tl);
3831
3832 if (rc)
3833 goto error;
3834
3835 rc = is_path_remote(&mnt_ctx);
3836 if (rc)
3837 rc = follow_dfs_link(&mnt_ctx);
3838 if (rc)
3839 goto error;
3840
3841 setup_server_referral_paths(&mnt_ctx);
3842 /*
3843 * After reconnecting to a different server, unique ids won't match anymore, so we disable
3844 * serverino. This prevents dentry revalidation to think the dentry are stale (ESTALE).
3845 */
3846 cifs_autodisable_serverino(cifs_sb);
3847 /*
3848 * Force the use of prefix path to support failover on DFS paths that resolve to targets
3849 * that have different prefix paths.
3850 */
3851 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH;
3852 kfree(cifs_sb->prepath);
3853 cifs_sb->prepath = ctx->prepath;
3854 ctx->prepath = NULL;
3855 uuid_copy(&cifs_sb->dfs_mount_id, &mnt_ctx.mount_id);
3856
3857 out:
3858 cifs_try_adding_channels(cifs_sb, mnt_ctx.ses);
3859 rc = mount_setup_tlink(cifs_sb, mnt_ctx.ses, mnt_ctx.tcon);
3860 if (rc)
3861 goto error;
3862
3863 free_xid(mnt_ctx.xid);
3864 return rc;
3865
3866 error:
3867 dfs_cache_put_refsrv_sessions(&mnt_ctx.mount_id);
3868 kfree(mnt_ctx.origin_fullpath);
3869 kfree(mnt_ctx.leaf_fullpath);
3870 mount_put_conns(&mnt_ctx);
3871 return rc;
3872 }
3873 #else
cifs_mount(struct cifs_sb_info * cifs_sb,struct smb3_fs_context * ctx)3874 int cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx)
3875 {
3876 int rc = 0;
3877 struct mount_ctx mnt_ctx = { .cifs_sb = cifs_sb, .fs_ctx = ctx, };
3878
3879 rc = mount_get_conns(&mnt_ctx);
3880 if (rc)
3881 goto error;
3882
3883 if (mnt_ctx.tcon) {
3884 rc = is_path_remote(&mnt_ctx);
3885 if (rc == -EREMOTE)
3886 rc = -EOPNOTSUPP;
3887 if (rc)
3888 goto error;
3889 }
3890
3891 rc = mount_setup_tlink(cifs_sb, mnt_ctx.ses, mnt_ctx.tcon);
3892 if (rc)
3893 goto error;
3894
3895 free_xid(mnt_ctx.xid);
3896 return rc;
3897
3898 error:
3899 mount_put_conns(&mnt_ctx);
3900 return rc;
3901 }
3902 #endif
3903
3904 /*
3905 * Issue a TREE_CONNECT request.
3906 */
3907 int
CIFSTCon(const unsigned int xid,struct cifs_ses * ses,const char * tree,struct cifs_tcon * tcon,const struct nls_table * nls_codepage)3908 CIFSTCon(const unsigned int xid, struct cifs_ses *ses,
3909 const char *tree, struct cifs_tcon *tcon,
3910 const struct nls_table *nls_codepage)
3911 {
3912 struct smb_hdr *smb_buffer;
3913 struct smb_hdr *smb_buffer_response;
3914 TCONX_REQ *pSMB;
3915 TCONX_RSP *pSMBr;
3916 unsigned char *bcc_ptr;
3917 int rc = 0;
3918 int length;
3919 __u16 bytes_left, count;
3920
3921 if (ses == NULL)
3922 return -EIO;
3923
3924 smb_buffer = cifs_buf_get();
3925 if (smb_buffer == NULL)
3926 return -ENOMEM;
3927
3928 smb_buffer_response = smb_buffer;
3929
3930 header_assemble(smb_buffer, SMB_COM_TREE_CONNECT_ANDX,
3931 NULL /*no tid */ , 4 /*wct */ );
3932
3933 smb_buffer->Mid = get_next_mid(ses->server);
3934 smb_buffer->Uid = ses->Suid;
3935 pSMB = (TCONX_REQ *) smb_buffer;
3936 pSMBr = (TCONX_RSP *) smb_buffer_response;
3937
3938 pSMB->AndXCommand = 0xFF;
3939 pSMB->Flags = cpu_to_le16(TCON_EXTENDED_SECINFO);
3940 bcc_ptr = &pSMB->Password[0];
3941
3942 pSMB->PasswordLength = cpu_to_le16(1); /* minimum */
3943 *bcc_ptr = 0; /* password is null byte */
3944 bcc_ptr++; /* skip password */
3945 /* already aligned so no need to do it below */
3946
3947 if (ses->server->sign)
3948 smb_buffer->Flags2 |= SMBFLG2_SECURITY_SIGNATURE;
3949
3950 if (ses->capabilities & CAP_STATUS32) {
3951 smb_buffer->Flags2 |= SMBFLG2_ERR_STATUS;
3952 }
3953 if (ses->capabilities & CAP_DFS) {
3954 smb_buffer->Flags2 |= SMBFLG2_DFS;
3955 }
3956 if (ses->capabilities & CAP_UNICODE) {
3957 smb_buffer->Flags2 |= SMBFLG2_UNICODE;
3958 length =
3959 cifs_strtoUTF16((__le16 *) bcc_ptr, tree,
3960 6 /* max utf8 char length in bytes */ *
3961 (/* server len*/ + 256 /* share len */), nls_codepage);
3962 bcc_ptr += 2 * length; /* convert num 16 bit words to bytes */
3963 bcc_ptr += 2; /* skip trailing null */
3964 } else { /* ASCII */
3965 strcpy(bcc_ptr, tree);
3966 bcc_ptr += strlen(tree) + 1;
3967 }
3968 strcpy(bcc_ptr, "?????");
3969 bcc_ptr += strlen("?????");
3970 bcc_ptr += 1;
3971 count = bcc_ptr - &pSMB->Password[0];
3972 be32_add_cpu(&pSMB->hdr.smb_buf_length, count);
3973 pSMB->ByteCount = cpu_to_le16(count);
3974
3975 rc = SendReceive(xid, ses, smb_buffer, smb_buffer_response, &length,
3976 0);
3977
3978 /* above now done in SendReceive */
3979 if (rc == 0) {
3980 bool is_unicode;
3981
3982 tcon->tid = smb_buffer_response->Tid;
3983 bcc_ptr = pByteArea(smb_buffer_response);
3984 bytes_left = get_bcc(smb_buffer_response);
3985 length = strnlen(bcc_ptr, bytes_left - 2);
3986 if (smb_buffer->Flags2 & SMBFLG2_UNICODE)
3987 is_unicode = true;
3988 else
3989 is_unicode = false;
3990
3991
3992 /* skip service field (NB: this field is always ASCII) */
3993 if (length == 3) {
3994 if ((bcc_ptr[0] == 'I') && (bcc_ptr[1] == 'P') &&
3995 (bcc_ptr[2] == 'C')) {
3996 cifs_dbg(FYI, "IPC connection\n");
3997 tcon->ipc = true;
3998 tcon->pipe = true;
3999 }
4000 } else if (length == 2) {
4001 if ((bcc_ptr[0] == 'A') && (bcc_ptr[1] == ':')) {
4002 /* the most common case */
4003 cifs_dbg(FYI, "disk share connection\n");
4004 }
4005 }
4006 bcc_ptr += length + 1;
4007 bytes_left -= (length + 1);
4008 strscpy(tcon->tree_name, tree, sizeof(tcon->tree_name));
4009
4010 /* mostly informational -- no need to fail on error here */
4011 kfree(tcon->nativeFileSystem);
4012 tcon->nativeFileSystem = cifs_strndup_from_utf16(bcc_ptr,
4013 bytes_left, is_unicode,
4014 nls_codepage);
4015
4016 cifs_dbg(FYI, "nativeFileSystem=%s\n", tcon->nativeFileSystem);
4017
4018 if ((smb_buffer_response->WordCount == 3) ||
4019 (smb_buffer_response->WordCount == 7))
4020 /* field is in same location */
4021 tcon->Flags = le16_to_cpu(pSMBr->OptionalSupport);
4022 else
4023 tcon->Flags = 0;
4024 cifs_dbg(FYI, "Tcon flags: 0x%x\n", tcon->Flags);
4025 }
4026
4027 cifs_buf_release(smb_buffer);
4028 return rc;
4029 }
4030
delayed_free(struct rcu_head * p)4031 static void delayed_free(struct rcu_head *p)
4032 {
4033 struct cifs_sb_info *cifs_sb = container_of(p, struct cifs_sb_info, rcu);
4034
4035 unload_nls(cifs_sb->local_nls);
4036 smb3_cleanup_fs_context(cifs_sb->ctx);
4037 kfree(cifs_sb);
4038 }
4039
4040 void
cifs_umount(struct cifs_sb_info * cifs_sb)4041 cifs_umount(struct cifs_sb_info *cifs_sb)
4042 {
4043 struct rb_root *root = &cifs_sb->tlink_tree;
4044 struct rb_node *node;
4045 struct tcon_link *tlink;
4046
4047 cancel_delayed_work_sync(&cifs_sb->prune_tlinks);
4048
4049 spin_lock(&cifs_sb->tlink_tree_lock);
4050 while ((node = rb_first(root))) {
4051 tlink = rb_entry(node, struct tcon_link, tl_rbnode);
4052 cifs_get_tlink(tlink);
4053 clear_bit(TCON_LINK_IN_TREE, &tlink->tl_flags);
4054 rb_erase(node, root);
4055
4056 spin_unlock(&cifs_sb->tlink_tree_lock);
4057 cifs_put_tlink(tlink);
4058 spin_lock(&cifs_sb->tlink_tree_lock);
4059 }
4060 spin_unlock(&cifs_sb->tlink_tree_lock);
4061
4062 kfree(cifs_sb->prepath);
4063 #ifdef CONFIG_CIFS_DFS_UPCALL
4064 dfs_cache_put_refsrv_sessions(&cifs_sb->dfs_mount_id);
4065 #endif
4066 call_rcu(&cifs_sb->rcu, delayed_free);
4067 }
4068
4069 int
cifs_negotiate_protocol(const unsigned int xid,struct cifs_ses * ses,struct TCP_Server_Info * server)4070 cifs_negotiate_protocol(const unsigned int xid, struct cifs_ses *ses,
4071 struct TCP_Server_Info *server)
4072 {
4073 int rc = 0;
4074
4075 if (!server->ops->need_neg || !server->ops->negotiate)
4076 return -ENOSYS;
4077
4078 /* only send once per connect */
4079 spin_lock(&server->srv_lock);
4080 if (!server->ops->need_neg(server) ||
4081 server->tcpStatus != CifsNeedNegotiate) {
4082 spin_unlock(&server->srv_lock);
4083 return 0;
4084 }
4085 server->tcpStatus = CifsInNegotiate;
4086 spin_unlock(&server->srv_lock);
4087
4088 rc = server->ops->negotiate(xid, ses, server);
4089 if (rc == 0) {
4090 spin_lock(&server->srv_lock);
4091 if (server->tcpStatus == CifsInNegotiate)
4092 server->tcpStatus = CifsGood;
4093 else
4094 rc = -EHOSTDOWN;
4095 spin_unlock(&server->srv_lock);
4096 } else {
4097 spin_lock(&server->srv_lock);
4098 if (server->tcpStatus == CifsInNegotiate)
4099 server->tcpStatus = CifsNeedNegotiate;
4100 spin_unlock(&server->srv_lock);
4101 }
4102
4103 return rc;
4104 }
4105
4106 int
cifs_setup_session(const unsigned int xid,struct cifs_ses * ses,struct TCP_Server_Info * server,struct nls_table * nls_info)4107 cifs_setup_session(const unsigned int xid, struct cifs_ses *ses,
4108 struct TCP_Server_Info *server,
4109 struct nls_table *nls_info)
4110 {
4111 int rc = -ENOSYS;
4112 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&server->dstaddr;
4113 struct sockaddr_in *addr = (struct sockaddr_in *)&server->dstaddr;
4114 bool is_binding = false;
4115
4116 spin_lock(&ses->ses_lock);
4117 if (server->dstaddr.ss_family == AF_INET6)
4118 scnprintf(ses->ip_addr, sizeof(ses->ip_addr), "%pI6", &addr6->sin6_addr);
4119 else
4120 scnprintf(ses->ip_addr, sizeof(ses->ip_addr), "%pI4", &addr->sin_addr);
4121
4122 if (ses->ses_status != SES_GOOD &&
4123 ses->ses_status != SES_NEW &&
4124 ses->ses_status != SES_NEED_RECON) {
4125 spin_unlock(&ses->ses_lock);
4126 return 0;
4127 }
4128
4129 /* only send once per connect */
4130 spin_lock(&ses->chan_lock);
4131 if (CIFS_ALL_CHANS_GOOD(ses) ||
4132 cifs_chan_in_reconnect(ses, server)) {
4133 spin_unlock(&ses->chan_lock);
4134 spin_unlock(&ses->ses_lock);
4135 return 0;
4136 }
4137 is_binding = !CIFS_ALL_CHANS_NEED_RECONNECT(ses);
4138 cifs_chan_set_in_reconnect(ses, server);
4139 spin_unlock(&ses->chan_lock);
4140
4141 if (!is_binding)
4142 ses->ses_status = SES_IN_SETUP;
4143 spin_unlock(&ses->ses_lock);
4144
4145 if (!is_binding) {
4146 ses->capabilities = server->capabilities;
4147 if (!linuxExtEnabled)
4148 ses->capabilities &= (~server->vals->cap_unix);
4149
4150 if (ses->auth_key.response) {
4151 cifs_dbg(FYI, "Free previous auth_key.response = %p\n",
4152 ses->auth_key.response);
4153 kfree_sensitive(ses->auth_key.response);
4154 ses->auth_key.response = NULL;
4155 ses->auth_key.len = 0;
4156 }
4157 }
4158
4159 cifs_dbg(FYI, "Security Mode: 0x%x Capabilities: 0x%x TimeAdjust: %d\n",
4160 server->sec_mode, server->capabilities, server->timeAdj);
4161
4162 if (server->ops->sess_setup)
4163 rc = server->ops->sess_setup(xid, ses, server, nls_info);
4164
4165 if (rc) {
4166 cifs_server_dbg(VFS, "Send error in SessSetup = %d\n", rc);
4167 spin_lock(&ses->ses_lock);
4168 if (ses->ses_status == SES_IN_SETUP)
4169 ses->ses_status = SES_NEED_RECON;
4170 spin_lock(&ses->chan_lock);
4171 cifs_chan_clear_in_reconnect(ses, server);
4172 spin_unlock(&ses->chan_lock);
4173 spin_unlock(&ses->ses_lock);
4174 } else {
4175 spin_lock(&ses->ses_lock);
4176 if (ses->ses_status == SES_IN_SETUP)
4177 ses->ses_status = SES_GOOD;
4178 spin_lock(&ses->chan_lock);
4179 cifs_chan_clear_in_reconnect(ses, server);
4180 cifs_chan_clear_need_reconnect(ses, server);
4181 spin_unlock(&ses->chan_lock);
4182 spin_unlock(&ses->ses_lock);
4183 }
4184
4185 return rc;
4186 }
4187
4188 static int
cifs_set_vol_auth(struct smb3_fs_context * ctx,struct cifs_ses * ses)4189 cifs_set_vol_auth(struct smb3_fs_context *ctx, struct cifs_ses *ses)
4190 {
4191 ctx->sectype = ses->sectype;
4192
4193 /* krb5 is special, since we don't need username or pw */
4194 if (ctx->sectype == Kerberos)
4195 return 0;
4196
4197 return cifs_set_cifscreds(ctx, ses);
4198 }
4199
4200 static struct cifs_tcon *
cifs_construct_tcon(struct cifs_sb_info * cifs_sb,kuid_t fsuid)4201 cifs_construct_tcon(struct cifs_sb_info *cifs_sb, kuid_t fsuid)
4202 {
4203 int rc;
4204 struct cifs_tcon *master_tcon = cifs_sb_master_tcon(cifs_sb);
4205 struct cifs_ses *ses;
4206 struct cifs_tcon *tcon = NULL;
4207 struct smb3_fs_context *ctx;
4208
4209 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
4210 if (ctx == NULL)
4211 return ERR_PTR(-ENOMEM);
4212
4213 ctx->local_nls = cifs_sb->local_nls;
4214 ctx->linux_uid = fsuid;
4215 ctx->cred_uid = fsuid;
4216 ctx->UNC = master_tcon->tree_name;
4217 ctx->retry = master_tcon->retry;
4218 ctx->nocase = master_tcon->nocase;
4219 ctx->nohandlecache = master_tcon->nohandlecache;
4220 ctx->local_lease = master_tcon->local_lease;
4221 ctx->no_lease = master_tcon->no_lease;
4222 ctx->resilient = master_tcon->use_resilient;
4223 ctx->persistent = master_tcon->use_persistent;
4224 ctx->handle_timeout = master_tcon->handle_timeout;
4225 ctx->no_linux_ext = !master_tcon->unix_ext;
4226 ctx->linux_ext = master_tcon->posix_extensions;
4227 ctx->sectype = master_tcon->ses->sectype;
4228 ctx->sign = master_tcon->ses->sign;
4229 ctx->seal = master_tcon->seal;
4230 ctx->witness = master_tcon->use_witness;
4231
4232 rc = cifs_set_vol_auth(ctx, master_tcon->ses);
4233 if (rc) {
4234 tcon = ERR_PTR(rc);
4235 goto out;
4236 }
4237
4238 /* get a reference for the same TCP session */
4239 spin_lock(&cifs_tcp_ses_lock);
4240 ++master_tcon->ses->server->srv_count;
4241 spin_unlock(&cifs_tcp_ses_lock);
4242
4243 ses = cifs_get_smb_ses(master_tcon->ses->server, ctx);
4244 if (IS_ERR(ses)) {
4245 tcon = (struct cifs_tcon *)ses;
4246 cifs_put_tcp_session(master_tcon->ses->server, 0);
4247 goto out;
4248 }
4249
4250 tcon = cifs_get_tcon(ses, ctx);
4251 if (IS_ERR(tcon)) {
4252 cifs_put_smb_ses(ses);
4253 goto out;
4254 }
4255
4256 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
4257 if (cap_unix(ses))
4258 reset_cifs_unix_caps(0, tcon, NULL, ctx);
4259 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
4260
4261 out:
4262 kfree(ctx->username);
4263 kfree_sensitive(ctx->password);
4264 kfree(ctx);
4265
4266 return tcon;
4267 }
4268
4269 struct cifs_tcon *
cifs_sb_master_tcon(struct cifs_sb_info * cifs_sb)4270 cifs_sb_master_tcon(struct cifs_sb_info *cifs_sb)
4271 {
4272 return tlink_tcon(cifs_sb_master_tlink(cifs_sb));
4273 }
4274
4275 /* find and return a tlink with given uid */
4276 static struct tcon_link *
tlink_rb_search(struct rb_root * root,kuid_t uid)4277 tlink_rb_search(struct rb_root *root, kuid_t uid)
4278 {
4279 struct rb_node *node = root->rb_node;
4280 struct tcon_link *tlink;
4281
4282 while (node) {
4283 tlink = rb_entry(node, struct tcon_link, tl_rbnode);
4284
4285 if (uid_gt(tlink->tl_uid, uid))
4286 node = node->rb_left;
4287 else if (uid_lt(tlink->tl_uid, uid))
4288 node = node->rb_right;
4289 else
4290 return tlink;
4291 }
4292 return NULL;
4293 }
4294
4295 /* insert a tcon_link into the tree */
4296 static void
tlink_rb_insert(struct rb_root * root,struct tcon_link * new_tlink)4297 tlink_rb_insert(struct rb_root *root, struct tcon_link *new_tlink)
4298 {
4299 struct rb_node **new = &(root->rb_node), *parent = NULL;
4300 struct tcon_link *tlink;
4301
4302 while (*new) {
4303 tlink = rb_entry(*new, struct tcon_link, tl_rbnode);
4304 parent = *new;
4305
4306 if (uid_gt(tlink->tl_uid, new_tlink->tl_uid))
4307 new = &((*new)->rb_left);
4308 else
4309 new = &((*new)->rb_right);
4310 }
4311
4312 rb_link_node(&new_tlink->tl_rbnode, parent, new);
4313 rb_insert_color(&new_tlink->tl_rbnode, root);
4314 }
4315
4316 /*
4317 * Find or construct an appropriate tcon given a cifs_sb and the fsuid of the
4318 * current task.
4319 *
4320 * If the superblock doesn't refer to a multiuser mount, then just return
4321 * the master tcon for the mount.
4322 *
4323 * First, search the rbtree for an existing tcon for this fsuid. If one
4324 * exists, then check to see if it's pending construction. If it is then wait
4325 * for construction to complete. Once it's no longer pending, check to see if
4326 * it failed and either return an error or retry construction, depending on
4327 * the timeout.
4328 *
4329 * If one doesn't exist then insert a new tcon_link struct into the tree and
4330 * try to construct a new one.
4331 */
4332 struct tcon_link *
cifs_sb_tlink(struct cifs_sb_info * cifs_sb)4333 cifs_sb_tlink(struct cifs_sb_info *cifs_sb)
4334 {
4335 int ret;
4336 kuid_t fsuid = current_fsuid();
4337 struct tcon_link *tlink, *newtlink;
4338
4339 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
4340 return cifs_get_tlink(cifs_sb_master_tlink(cifs_sb));
4341
4342 spin_lock(&cifs_sb->tlink_tree_lock);
4343 tlink = tlink_rb_search(&cifs_sb->tlink_tree, fsuid);
4344 if (tlink)
4345 cifs_get_tlink(tlink);
4346 spin_unlock(&cifs_sb->tlink_tree_lock);
4347
4348 if (tlink == NULL) {
4349 newtlink = kzalloc(sizeof(*tlink), GFP_KERNEL);
4350 if (newtlink == NULL)
4351 return ERR_PTR(-ENOMEM);
4352 newtlink->tl_uid = fsuid;
4353 newtlink->tl_tcon = ERR_PTR(-EACCES);
4354 set_bit(TCON_LINK_PENDING, &newtlink->tl_flags);
4355 set_bit(TCON_LINK_IN_TREE, &newtlink->tl_flags);
4356 cifs_get_tlink(newtlink);
4357
4358 spin_lock(&cifs_sb->tlink_tree_lock);
4359 /* was one inserted after previous search? */
4360 tlink = tlink_rb_search(&cifs_sb->tlink_tree, fsuid);
4361 if (tlink) {
4362 cifs_get_tlink(tlink);
4363 spin_unlock(&cifs_sb->tlink_tree_lock);
4364 kfree(newtlink);
4365 goto wait_for_construction;
4366 }
4367 tlink = newtlink;
4368 tlink_rb_insert(&cifs_sb->tlink_tree, tlink);
4369 spin_unlock(&cifs_sb->tlink_tree_lock);
4370 } else {
4371 wait_for_construction:
4372 ret = wait_on_bit(&tlink->tl_flags, TCON_LINK_PENDING,
4373 TASK_INTERRUPTIBLE);
4374 if (ret) {
4375 cifs_put_tlink(tlink);
4376 return ERR_PTR(-ERESTARTSYS);
4377 }
4378
4379 /* if it's good, return it */
4380 if (!IS_ERR(tlink->tl_tcon))
4381 return tlink;
4382
4383 /* return error if we tried this already recently */
4384 if (time_before(jiffies, tlink->tl_time + TLINK_ERROR_EXPIRE)) {
4385 cifs_put_tlink(tlink);
4386 return ERR_PTR(-EACCES);
4387 }
4388
4389 if (test_and_set_bit(TCON_LINK_PENDING, &tlink->tl_flags))
4390 goto wait_for_construction;
4391 }
4392
4393 tlink->tl_tcon = cifs_construct_tcon(cifs_sb, fsuid);
4394 clear_bit(TCON_LINK_PENDING, &tlink->tl_flags);
4395 wake_up_bit(&tlink->tl_flags, TCON_LINK_PENDING);
4396
4397 if (IS_ERR(tlink->tl_tcon)) {
4398 cifs_put_tlink(tlink);
4399 return ERR_PTR(-EACCES);
4400 }
4401
4402 return tlink;
4403 }
4404
4405 /*
4406 * periodic workqueue job that scans tcon_tree for a superblock and closes
4407 * out tcons.
4408 */
4409 static void
cifs_prune_tlinks(struct work_struct * work)4410 cifs_prune_tlinks(struct work_struct *work)
4411 {
4412 struct cifs_sb_info *cifs_sb = container_of(work, struct cifs_sb_info,
4413 prune_tlinks.work);
4414 struct rb_root *root = &cifs_sb->tlink_tree;
4415 struct rb_node *node;
4416 struct rb_node *tmp;
4417 struct tcon_link *tlink;
4418
4419 /*
4420 * Because we drop the spinlock in the loop in order to put the tlink
4421 * it's not guarded against removal of links from the tree. The only
4422 * places that remove entries from the tree are this function and
4423 * umounts. Because this function is non-reentrant and is canceled
4424 * before umount can proceed, this is safe.
4425 */
4426 spin_lock(&cifs_sb->tlink_tree_lock);
4427 node = rb_first(root);
4428 while (node != NULL) {
4429 tmp = node;
4430 node = rb_next(tmp);
4431 tlink = rb_entry(tmp, struct tcon_link, tl_rbnode);
4432
4433 if (test_bit(TCON_LINK_MASTER, &tlink->tl_flags) ||
4434 atomic_read(&tlink->tl_count) != 0 ||
4435 time_after(tlink->tl_time + TLINK_IDLE_EXPIRE, jiffies))
4436 continue;
4437
4438 cifs_get_tlink(tlink);
4439 clear_bit(TCON_LINK_IN_TREE, &tlink->tl_flags);
4440 rb_erase(tmp, root);
4441
4442 spin_unlock(&cifs_sb->tlink_tree_lock);
4443 cifs_put_tlink(tlink);
4444 spin_lock(&cifs_sb->tlink_tree_lock);
4445 }
4446 spin_unlock(&cifs_sb->tlink_tree_lock);
4447
4448 queue_delayed_work(cifsiod_wq, &cifs_sb->prune_tlinks,
4449 TLINK_IDLE_EXPIRE);
4450 }
4451
4452 #ifdef CONFIG_CIFS_DFS_UPCALL
4453 /* Update dfs referral path of superblock */
update_server_fullpath(struct TCP_Server_Info * server,struct cifs_sb_info * cifs_sb,const char * target)4454 static int update_server_fullpath(struct TCP_Server_Info *server, struct cifs_sb_info *cifs_sb,
4455 const char *target)
4456 {
4457 int rc = 0;
4458 size_t len = strlen(target);
4459 char *refpath, *npath;
4460
4461 if (unlikely(len < 2 || *target != '\\'))
4462 return -EINVAL;
4463
4464 if (target[1] == '\\') {
4465 len += 1;
4466 refpath = kmalloc(len, GFP_KERNEL);
4467 if (!refpath)
4468 return -ENOMEM;
4469
4470 scnprintf(refpath, len, "%s", target);
4471 } else {
4472 len += sizeof("\\");
4473 refpath = kmalloc(len, GFP_KERNEL);
4474 if (!refpath)
4475 return -ENOMEM;
4476
4477 scnprintf(refpath, len, "\\%s", target);
4478 }
4479
4480 npath = dfs_cache_canonical_path(refpath, cifs_sb->local_nls, cifs_remap(cifs_sb));
4481 kfree(refpath);
4482
4483 if (IS_ERR(npath)) {
4484 rc = PTR_ERR(npath);
4485 } else {
4486 mutex_lock(&server->refpath_lock);
4487 kfree(server->leaf_fullpath);
4488 server->leaf_fullpath = npath;
4489 mutex_unlock(&server->refpath_lock);
4490 server->current_fullpath = server->leaf_fullpath;
4491 }
4492 return rc;
4493 }
4494
target_share_matches_server(struct TCP_Server_Info * server,const char * tcp_host,size_t tcp_host_len,char * share,bool * target_match)4495 static int target_share_matches_server(struct TCP_Server_Info *server, const char *tcp_host,
4496 size_t tcp_host_len, char *share, bool *target_match)
4497 {
4498 int rc = 0;
4499 const char *dfs_host;
4500 size_t dfs_host_len;
4501
4502 *target_match = true;
4503 extract_unc_hostname(share, &dfs_host, &dfs_host_len);
4504
4505 /* Check if hostnames or addresses match */
4506 if (dfs_host_len != tcp_host_len || strncasecmp(dfs_host, tcp_host, dfs_host_len) != 0) {
4507 cifs_dbg(FYI, "%s: %.*s doesn't match %.*s\n", __func__, (int)dfs_host_len,
4508 dfs_host, (int)tcp_host_len, tcp_host);
4509 rc = match_target_ip(server, dfs_host, dfs_host_len, target_match);
4510 if (rc)
4511 cifs_dbg(VFS, "%s: failed to match target ip: %d\n", __func__, rc);
4512 }
4513 return rc;
4514 }
4515
__tree_connect_dfs_target(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb,char * tree,bool islink,struct dfs_cache_tgt_list * tl)4516 static int __tree_connect_dfs_target(const unsigned int xid, struct cifs_tcon *tcon,
4517 struct cifs_sb_info *cifs_sb, char *tree, bool islink,
4518 struct dfs_cache_tgt_list *tl)
4519 {
4520 int rc;
4521 struct TCP_Server_Info *server = tcon->ses->server;
4522 const struct smb_version_operations *ops = server->ops;
4523 struct cifs_tcon *ipc = tcon->ses->tcon_ipc;
4524 char *share = NULL, *prefix = NULL;
4525 const char *tcp_host;
4526 size_t tcp_host_len;
4527 struct dfs_cache_tgt_iterator *tit;
4528 bool target_match;
4529
4530 extract_unc_hostname(server->hostname, &tcp_host, &tcp_host_len);
4531
4532 tit = dfs_cache_get_tgt_iterator(tl);
4533 if (!tit) {
4534 rc = -ENOENT;
4535 goto out;
4536 }
4537
4538 /* Try to tree connect to all dfs targets */
4539 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) {
4540 const char *target = dfs_cache_get_tgt_name(tit);
4541 struct dfs_cache_tgt_list ntl = DFS_CACHE_TGT_LIST_INIT(ntl);
4542
4543 kfree(share);
4544 kfree(prefix);
4545 share = prefix = NULL;
4546
4547 /* Check if share matches with tcp ses */
4548 rc = dfs_cache_get_tgt_share(server->current_fullpath + 1, tit, &share, &prefix);
4549 if (rc) {
4550 cifs_dbg(VFS, "%s: failed to parse target share: %d\n", __func__, rc);
4551 break;
4552 }
4553
4554 rc = target_share_matches_server(server, tcp_host, tcp_host_len, share,
4555 &target_match);
4556 if (rc)
4557 break;
4558 if (!target_match) {
4559 rc = -EHOSTUNREACH;
4560 continue;
4561 }
4562
4563 if (ipc->need_reconnect) {
4564 scnprintf(tree, MAX_TREE_SIZE, "\\\\%s\\IPC$", server->hostname);
4565 rc = ops->tree_connect(xid, ipc->ses, tree, ipc, cifs_sb->local_nls);
4566 if (rc)
4567 break;
4568 }
4569
4570 scnprintf(tree, MAX_TREE_SIZE, "\\%s", share);
4571 if (!islink) {
4572 rc = ops->tree_connect(xid, tcon->ses, tree, tcon, cifs_sb->local_nls);
4573 break;
4574 }
4575 /*
4576 * If no dfs referrals were returned from link target, then just do a TREE_CONNECT
4577 * to it. Otherwise, cache the dfs referral and then mark current tcp ses for
4578 * reconnect so either the demultiplex thread or the echo worker will reconnect to
4579 * newly resolved target.
4580 */
4581 if (dfs_cache_find(xid, tcon->ses, cifs_sb->local_nls, cifs_remap(cifs_sb), target,
4582 NULL, &ntl)) {
4583 rc = ops->tree_connect(xid, tcon->ses, tree, tcon, cifs_sb->local_nls);
4584 if (rc)
4585 continue;
4586 rc = dfs_cache_noreq_update_tgthint(server->current_fullpath + 1, tit);
4587 if (!rc)
4588 rc = cifs_update_super_prepath(cifs_sb, prefix);
4589 } else {
4590 /* Target is another dfs share */
4591 rc = update_server_fullpath(server, cifs_sb, target);
4592 dfs_cache_free_tgts(tl);
4593
4594 if (!rc) {
4595 rc = -EREMOTE;
4596 list_replace_init(&ntl.tl_list, &tl->tl_list);
4597 } else
4598 dfs_cache_free_tgts(&ntl);
4599 }
4600 break;
4601 }
4602
4603 out:
4604 kfree(share);
4605 kfree(prefix);
4606
4607 return rc;
4608 }
4609
tree_connect_dfs_target(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb,char * tree,bool islink,struct dfs_cache_tgt_list * tl)4610 static int tree_connect_dfs_target(const unsigned int xid, struct cifs_tcon *tcon,
4611 struct cifs_sb_info *cifs_sb, char *tree, bool islink,
4612 struct dfs_cache_tgt_list *tl)
4613 {
4614 int rc;
4615 int num_links = 0;
4616 struct TCP_Server_Info *server = tcon->ses->server;
4617
4618 do {
4619 rc = __tree_connect_dfs_target(xid, tcon, cifs_sb, tree, islink, tl);
4620 if (!rc || rc != -EREMOTE)
4621 break;
4622 } while (rc = -ELOOP, ++num_links < MAX_NESTED_LINKS);
4623 /*
4624 * If we couldn't tree connect to any targets from last referral path, then retry from
4625 * original referral path.
4626 */
4627 if (rc && server->current_fullpath != server->origin_fullpath) {
4628 server->current_fullpath = server->origin_fullpath;
4629 cifs_signal_cifsd_for_reconnect(server, true);
4630 }
4631
4632 dfs_cache_free_tgts(tl);
4633 return rc;
4634 }
4635
cifs_tree_connect(const unsigned int xid,struct cifs_tcon * tcon,const struct nls_table * nlsc)4636 int cifs_tree_connect(const unsigned int xid, struct cifs_tcon *tcon, const struct nls_table *nlsc)
4637 {
4638 int rc;
4639 struct TCP_Server_Info *server = tcon->ses->server;
4640 const struct smb_version_operations *ops = server->ops;
4641 struct super_block *sb = NULL;
4642 struct cifs_sb_info *cifs_sb;
4643 struct dfs_cache_tgt_list tl = DFS_CACHE_TGT_LIST_INIT(tl);
4644 char *tree;
4645 struct dfs_info3_param ref = {0};
4646
4647 /* only send once per connect */
4648 spin_lock(&tcon->tc_lock);
4649 if (tcon->ses->ses_status != SES_GOOD ||
4650 (tcon->status != TID_NEW &&
4651 tcon->status != TID_NEED_TCON)) {
4652 spin_unlock(&tcon->tc_lock);
4653 return 0;
4654 }
4655 tcon->status = TID_IN_TCON;
4656 spin_unlock(&tcon->tc_lock);
4657
4658 tree = kzalloc(MAX_TREE_SIZE, GFP_KERNEL);
4659 if (!tree) {
4660 rc = -ENOMEM;
4661 goto out;
4662 }
4663
4664 if (tcon->ipc) {
4665 scnprintf(tree, MAX_TREE_SIZE, "\\\\%s\\IPC$", server->hostname);
4666 rc = ops->tree_connect(xid, tcon->ses, tree, tcon, nlsc);
4667 goto out;
4668 }
4669
4670 sb = cifs_get_tcp_super(server);
4671 if (IS_ERR(sb)) {
4672 rc = PTR_ERR(sb);
4673 cifs_dbg(VFS, "%s: could not find superblock: %d\n", __func__, rc);
4674 goto out;
4675 }
4676
4677 cifs_sb = CIFS_SB(sb);
4678
4679 /* If it is not dfs or there was no cached dfs referral, then reconnect to same share */
4680 if (!server->current_fullpath ||
4681 dfs_cache_noreq_find(server->current_fullpath + 1, &ref, &tl)) {
4682 rc = ops->tree_connect(xid, tcon->ses, tcon->tree_name, tcon, cifs_sb->local_nls);
4683 goto out;
4684 }
4685
4686 rc = tree_connect_dfs_target(xid, tcon, cifs_sb, tree, ref.server_type == DFS_TYPE_LINK,
4687 &tl);
4688 free_dfs_info_param(&ref);
4689
4690 out:
4691 kfree(tree);
4692 cifs_put_tcp_super(sb);
4693
4694 if (rc) {
4695 spin_lock(&tcon->tc_lock);
4696 if (tcon->status == TID_IN_TCON)
4697 tcon->status = TID_NEED_TCON;
4698 spin_unlock(&tcon->tc_lock);
4699 } else {
4700 spin_lock(&tcon->tc_lock);
4701 if (tcon->status == TID_IN_TCON)
4702 tcon->status = TID_GOOD;
4703 spin_unlock(&tcon->tc_lock);
4704 tcon->need_reconnect = false;
4705 }
4706
4707 return rc;
4708 }
4709 #else
cifs_tree_connect(const unsigned int xid,struct cifs_tcon * tcon,const struct nls_table * nlsc)4710 int cifs_tree_connect(const unsigned int xid, struct cifs_tcon *tcon, const struct nls_table *nlsc)
4711 {
4712 int rc;
4713 const struct smb_version_operations *ops = tcon->ses->server->ops;
4714
4715 /* only send once per connect */
4716 spin_lock(&tcon->tc_lock);
4717 if (tcon->ses->ses_status != SES_GOOD ||
4718 (tcon->status != TID_NEW &&
4719 tcon->status != TID_NEED_TCON)) {
4720 spin_unlock(&tcon->tc_lock);
4721 return 0;
4722 }
4723 tcon->status = TID_IN_TCON;
4724 spin_unlock(&tcon->tc_lock);
4725
4726 rc = ops->tree_connect(xid, tcon->ses, tcon->tree_name, tcon, nlsc);
4727 if (rc) {
4728 spin_lock(&tcon->tc_lock);
4729 if (tcon->status == TID_IN_TCON)
4730 tcon->status = TID_NEED_TCON;
4731 spin_unlock(&tcon->tc_lock);
4732 } else {
4733 spin_lock(&tcon->tc_lock);
4734 if (tcon->status == TID_IN_TCON)
4735 tcon->status = TID_GOOD;
4736 tcon->need_reconnect = false;
4737 spin_unlock(&tcon->tc_lock);
4738 }
4739
4740 return rc;
4741 }
4742 #endif
4743