1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * zfcp device driver
4  *
5  * Implementation of FSF commands.
6  *
7  * Copyright IBM Corp. 2002, 2023
8  */
9 
10 #define KMSG_COMPONENT "zfcp"
11 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
12 
13 #include <linux/blktrace_api.h>
14 #include <linux/jiffies.h>
15 #include <linux/types.h>
16 #include <linux/slab.h>
17 #include <scsi/fc/fc_els.h>
18 #include "zfcp_ext.h"
19 #include "zfcp_fc.h"
20 #include "zfcp_dbf.h"
21 #include "zfcp_qdio.h"
22 #include "zfcp_reqlist.h"
23 #include "zfcp_diag.h"
24 
25 /* timeout for FSF requests sent during scsi_eh: abort or FCP TMF */
26 #define ZFCP_FSF_SCSI_ER_TIMEOUT (10*HZ)
27 /* timeout for: exchange config/port data outside ERP, or open/close WKA port */
28 #define ZFCP_FSF_REQUEST_TIMEOUT (60*HZ)
29 
30 struct kmem_cache *zfcp_fsf_qtcb_cache;
31 
32 static bool ber_stop = true;
33 module_param(ber_stop, bool, 0600);
34 MODULE_PARM_DESC(ber_stop,
35 		 "Shuts down FCP devices for FCP channels that report a bit-error count in excess of its threshold (default on)");
36 
zfcp_fsf_request_timeout_handler(struct timer_list * t)37 static void zfcp_fsf_request_timeout_handler(struct timer_list *t)
38 {
39 	struct zfcp_fsf_req *fsf_req = from_timer(fsf_req, t, timer);
40 	struct zfcp_adapter *adapter = fsf_req->adapter;
41 
42 	zfcp_qdio_siosl(adapter);
43 	zfcp_erp_adapter_reopen(adapter, ZFCP_STATUS_COMMON_ERP_FAILED,
44 				"fsrth_1");
45 }
46 
zfcp_fsf_start_timer(struct zfcp_fsf_req * fsf_req,unsigned long timeout)47 static void zfcp_fsf_start_timer(struct zfcp_fsf_req *fsf_req,
48 				 unsigned long timeout)
49 {
50 	fsf_req->timer.function = zfcp_fsf_request_timeout_handler;
51 	fsf_req->timer.expires = jiffies + timeout;
52 	add_timer(&fsf_req->timer);
53 }
54 
zfcp_fsf_start_erp_timer(struct zfcp_fsf_req * fsf_req)55 static void zfcp_fsf_start_erp_timer(struct zfcp_fsf_req *fsf_req)
56 {
57 	BUG_ON(!fsf_req->erp_action);
58 	fsf_req->timer.function = zfcp_erp_timeout_handler;
59 	fsf_req->timer.expires = jiffies + 30 * HZ;
60 	add_timer(&fsf_req->timer);
61 }
62 
63 /* association between FSF command and FSF QTCB type */
64 static u32 fsf_qtcb_type[] = {
65 	[FSF_QTCB_FCP_CMND] =             FSF_IO_COMMAND,
66 	[FSF_QTCB_ABORT_FCP_CMND] =       FSF_SUPPORT_COMMAND,
67 	[FSF_QTCB_OPEN_PORT_WITH_DID] =   FSF_SUPPORT_COMMAND,
68 	[FSF_QTCB_OPEN_LUN] =             FSF_SUPPORT_COMMAND,
69 	[FSF_QTCB_CLOSE_LUN] =            FSF_SUPPORT_COMMAND,
70 	[FSF_QTCB_CLOSE_PORT] =           FSF_SUPPORT_COMMAND,
71 	[FSF_QTCB_CLOSE_PHYSICAL_PORT] =  FSF_SUPPORT_COMMAND,
72 	[FSF_QTCB_SEND_ELS] =             FSF_SUPPORT_COMMAND,
73 	[FSF_QTCB_SEND_GENERIC] =         FSF_SUPPORT_COMMAND,
74 	[FSF_QTCB_EXCHANGE_CONFIG_DATA] = FSF_CONFIG_COMMAND,
75 	[FSF_QTCB_EXCHANGE_PORT_DATA] =   FSF_PORT_COMMAND,
76 	[FSF_QTCB_DOWNLOAD_CONTROL_FILE] = FSF_SUPPORT_COMMAND,
77 	[FSF_QTCB_UPLOAD_CONTROL_FILE] =  FSF_SUPPORT_COMMAND
78 };
79 
zfcp_fsf_class_not_supp(struct zfcp_fsf_req * req)80 static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req)
81 {
82 	dev_err(&req->adapter->ccw_device->dev, "FCP device not "
83 		"operational because of an unsupported FC class\n");
84 	zfcp_erp_adapter_shutdown(req->adapter, 0, "fscns_1");
85 	req->status |= ZFCP_STATUS_FSFREQ_ERROR;
86 }
87 
88 /**
89  * zfcp_fsf_req_free - free memory used by fsf request
90  * @req: pointer to struct zfcp_fsf_req
91  */
zfcp_fsf_req_free(struct zfcp_fsf_req * req)92 void zfcp_fsf_req_free(struct zfcp_fsf_req *req)
93 {
94 	if (likely(req->pool)) {
95 		if (likely(!zfcp_fsf_req_is_status_read_buffer(req)))
96 			mempool_free(req->qtcb, req->adapter->pool.qtcb_pool);
97 		mempool_free(req, req->pool);
98 		return;
99 	}
100 
101 	if (likely(!zfcp_fsf_req_is_status_read_buffer(req)))
102 		kmem_cache_free(zfcp_fsf_qtcb_cache, req->qtcb);
103 	kfree(req);
104 }
105 
zfcp_fsf_status_read_port_closed(struct zfcp_fsf_req * req)106 static void zfcp_fsf_status_read_port_closed(struct zfcp_fsf_req *req)
107 {
108 	unsigned long flags;
109 	struct fsf_status_read_buffer *sr_buf = req->data;
110 	struct zfcp_adapter *adapter = req->adapter;
111 	struct zfcp_port *port;
112 	int d_id = ntoh24(sr_buf->d_id);
113 
114 	read_lock_irqsave(&adapter->port_list_lock, flags);
115 	list_for_each_entry(port, &adapter->port_list, list)
116 		if (port->d_id == d_id) {
117 			zfcp_erp_port_reopen(port, 0, "fssrpc1");
118 			break;
119 		}
120 	read_unlock_irqrestore(&adapter->port_list_lock, flags);
121 }
122 
zfcp_fsf_fc_host_link_down(struct zfcp_adapter * adapter)123 void zfcp_fsf_fc_host_link_down(struct zfcp_adapter *adapter)
124 {
125 	struct Scsi_Host *shost = adapter->scsi_host;
126 
127 	adapter->hydra_version = 0;
128 	adapter->peer_wwpn = 0;
129 	adapter->peer_wwnn = 0;
130 	adapter->peer_d_id = 0;
131 
132 	/* if there is no shost yet, we have nothing to zero-out */
133 	if (shost == NULL)
134 		return;
135 
136 	fc_host_port_id(shost) = 0;
137 	fc_host_fabric_name(shost) = 0;
138 	fc_host_speed(shost) = FC_PORTSPEED_UNKNOWN;
139 	fc_host_port_type(shost) = FC_PORTTYPE_UNKNOWN;
140 	snprintf(fc_host_model(shost), FC_SYMBOLIC_NAME_SIZE, "0x%04x", 0);
141 	memset(fc_host_active_fc4s(shost), 0, FC_FC4_LIST_SIZE);
142 }
143 
zfcp_fsf_link_down_info_eval(struct zfcp_fsf_req * req,struct fsf_link_down_info * link_down)144 static void zfcp_fsf_link_down_info_eval(struct zfcp_fsf_req *req,
145 					 struct fsf_link_down_info *link_down)
146 {
147 	struct zfcp_adapter *adapter = req->adapter;
148 
149 	if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED)
150 		return;
151 
152 	atomic_or(ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED, &adapter->status);
153 
154 	zfcp_scsi_schedule_rports_block(adapter);
155 
156 	zfcp_fsf_fc_host_link_down(adapter);
157 
158 	if (!link_down)
159 		goto out;
160 
161 	switch (link_down->error_code) {
162 	case FSF_PSQ_LINK_NO_LIGHT:
163 		dev_warn(&req->adapter->ccw_device->dev,
164 			 "There is no light signal from the local "
165 			 "fibre channel cable\n");
166 		break;
167 	case FSF_PSQ_LINK_WRAP_PLUG:
168 		dev_warn(&req->adapter->ccw_device->dev,
169 			 "There is a wrap plug instead of a fibre "
170 			 "channel cable\n");
171 		break;
172 	case FSF_PSQ_LINK_NO_FCP:
173 		dev_warn(&req->adapter->ccw_device->dev,
174 			 "The adjacent fibre channel node does not "
175 			 "support FCP\n");
176 		break;
177 	case FSF_PSQ_LINK_FIRMWARE_UPDATE:
178 		dev_warn(&req->adapter->ccw_device->dev,
179 			 "The FCP device is suspended because of a "
180 			 "firmware update\n");
181 		break;
182 	case FSF_PSQ_LINK_INVALID_WWPN:
183 		dev_warn(&req->adapter->ccw_device->dev,
184 			 "The FCP device detected a WWPN that is "
185 			 "duplicate or not valid\n");
186 		break;
187 	case FSF_PSQ_LINK_NO_NPIV_SUPPORT:
188 		dev_warn(&req->adapter->ccw_device->dev,
189 			 "The fibre channel fabric does not support NPIV\n");
190 		break;
191 	case FSF_PSQ_LINK_NO_FCP_RESOURCES:
192 		dev_warn(&req->adapter->ccw_device->dev,
193 			 "The FCP adapter cannot support more NPIV ports\n");
194 		break;
195 	case FSF_PSQ_LINK_NO_FABRIC_RESOURCES:
196 		dev_warn(&req->adapter->ccw_device->dev,
197 			 "The adjacent switch cannot support "
198 			 "more NPIV ports\n");
199 		break;
200 	case FSF_PSQ_LINK_FABRIC_LOGIN_UNABLE:
201 		dev_warn(&req->adapter->ccw_device->dev,
202 			 "The FCP adapter could not log in to the "
203 			 "fibre channel fabric\n");
204 		break;
205 	case FSF_PSQ_LINK_WWPN_ASSIGNMENT_CORRUPTED:
206 		dev_warn(&req->adapter->ccw_device->dev,
207 			 "The WWPN assignment file on the FCP adapter "
208 			 "has been damaged\n");
209 		break;
210 	case FSF_PSQ_LINK_MODE_TABLE_CURRUPTED:
211 		dev_warn(&req->adapter->ccw_device->dev,
212 			 "The mode table on the FCP adapter "
213 			 "has been damaged\n");
214 		break;
215 	case FSF_PSQ_LINK_NO_WWPN_ASSIGNMENT:
216 		dev_warn(&req->adapter->ccw_device->dev,
217 			 "All NPIV ports on the FCP adapter have "
218 			 "been assigned\n");
219 		break;
220 	default:
221 		dev_warn(&req->adapter->ccw_device->dev,
222 			 "The link between the FCP adapter and "
223 			 "the FC fabric is down\n");
224 	}
225 out:
226 	zfcp_erp_set_adapter_status(adapter, ZFCP_STATUS_COMMON_ERP_FAILED);
227 }
228 
zfcp_fsf_status_read_link_down(struct zfcp_fsf_req * req)229 static void zfcp_fsf_status_read_link_down(struct zfcp_fsf_req *req)
230 {
231 	struct fsf_status_read_buffer *sr_buf = req->data;
232 	struct fsf_link_down_info *ldi =
233 		(struct fsf_link_down_info *) &sr_buf->payload;
234 
235 	switch (sr_buf->status_subtype) {
236 	case FSF_STATUS_READ_SUB_NO_PHYSICAL_LINK:
237 	case FSF_STATUS_READ_SUB_FDISC_FAILED:
238 		zfcp_fsf_link_down_info_eval(req, ldi);
239 		break;
240 	case FSF_STATUS_READ_SUB_FIRMWARE_UPDATE:
241 		zfcp_fsf_link_down_info_eval(req, NULL);
242 	}
243 }
244 
245 static void
zfcp_fsf_status_read_version_change(struct zfcp_adapter * adapter,struct fsf_status_read_buffer * sr_buf)246 zfcp_fsf_status_read_version_change(struct zfcp_adapter *adapter,
247 				    struct fsf_status_read_buffer *sr_buf)
248 {
249 	if (sr_buf->status_subtype == FSF_STATUS_READ_SUB_LIC_CHANGE) {
250 		u32 version = sr_buf->payload.version_change.current_version;
251 
252 		WRITE_ONCE(adapter->fsf_lic_version, version);
253 		snprintf(fc_host_firmware_version(adapter->scsi_host),
254 			 FC_VERSION_STRING_SIZE, "%#08x", version);
255 	}
256 }
257 
zfcp_fsf_status_read_handler(struct zfcp_fsf_req * req)258 static void zfcp_fsf_status_read_handler(struct zfcp_fsf_req *req)
259 {
260 	struct zfcp_adapter *adapter = req->adapter;
261 	struct fsf_status_read_buffer *sr_buf = req->data;
262 
263 	if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) {
264 		zfcp_dbf_hba_fsf_uss("fssrh_1", req);
265 		mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data);
266 		zfcp_fsf_req_free(req);
267 		return;
268 	}
269 
270 	zfcp_dbf_hba_fsf_uss("fssrh_4", req);
271 
272 	switch (sr_buf->status_type) {
273 	case FSF_STATUS_READ_PORT_CLOSED:
274 		zfcp_fsf_status_read_port_closed(req);
275 		break;
276 	case FSF_STATUS_READ_INCOMING_ELS:
277 		zfcp_fc_incoming_els(req);
278 		break;
279 	case FSF_STATUS_READ_SENSE_DATA_AVAIL:
280 		break;
281 	case FSF_STATUS_READ_BIT_ERROR_THRESHOLD:
282 		zfcp_dbf_hba_bit_err("fssrh_3", req);
283 		if (ber_stop) {
284 			dev_warn(&adapter->ccw_device->dev,
285 				 "All paths over this FCP device are disused because of excessive bit errors\n");
286 			zfcp_erp_adapter_shutdown(adapter, 0, "fssrh_b");
287 		} else {
288 			dev_warn(&adapter->ccw_device->dev,
289 				 "The error threshold for checksum statistics has been exceeded\n");
290 		}
291 		break;
292 	case FSF_STATUS_READ_LINK_DOWN:
293 		zfcp_fsf_status_read_link_down(req);
294 		zfcp_fc_enqueue_event(adapter, FCH_EVT_LINKDOWN, 0);
295 		break;
296 	case FSF_STATUS_READ_LINK_UP:
297 		dev_info(&adapter->ccw_device->dev,
298 			 "The local link has been restored\n");
299 		/* All ports should be marked as ready to run again */
300 		zfcp_erp_set_adapter_status(adapter,
301 					    ZFCP_STATUS_COMMON_RUNNING);
302 		zfcp_erp_adapter_reopen(adapter,
303 					ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED |
304 					ZFCP_STATUS_COMMON_ERP_FAILED,
305 					"fssrh_2");
306 		zfcp_fc_enqueue_event(adapter, FCH_EVT_LINKUP, 0);
307 
308 		break;
309 	case FSF_STATUS_READ_NOTIFICATION_LOST:
310 		if (sr_buf->status_subtype & FSF_STATUS_READ_SUB_INCOMING_ELS)
311 			zfcp_fc_conditional_port_scan(adapter);
312 		if (sr_buf->status_subtype & FSF_STATUS_READ_SUB_VERSION_CHANGE)
313 			queue_work(adapter->work_queue,
314 				   &adapter->version_change_lost_work);
315 		break;
316 	case FSF_STATUS_READ_FEATURE_UPDATE_ALERT:
317 		adapter->adapter_features = sr_buf->payload.word[0];
318 		break;
319 	case FSF_STATUS_READ_VERSION_CHANGE:
320 		zfcp_fsf_status_read_version_change(adapter, sr_buf);
321 		break;
322 	}
323 
324 	mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data);
325 	zfcp_fsf_req_free(req);
326 
327 	atomic_inc(&adapter->stat_miss);
328 	queue_work(adapter->work_queue, &adapter->stat_work);
329 }
330 
zfcp_fsf_fsfstatus_qual_eval(struct zfcp_fsf_req * req)331 static void zfcp_fsf_fsfstatus_qual_eval(struct zfcp_fsf_req *req)
332 {
333 	switch (req->qtcb->header.fsf_status_qual.word[0]) {
334 	case FSF_SQ_FCP_RSP_AVAILABLE:
335 	case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
336 	case FSF_SQ_NO_RETRY_POSSIBLE:
337 	case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
338 		return;
339 	case FSF_SQ_COMMAND_ABORTED:
340 		break;
341 	case FSF_SQ_NO_RECOM:
342 		dev_err(&req->adapter->ccw_device->dev,
343 			"The FCP adapter reported a problem "
344 			"that cannot be recovered\n");
345 		zfcp_qdio_siosl(req->adapter);
346 		zfcp_erp_adapter_shutdown(req->adapter, 0, "fsfsqe1");
347 		break;
348 	}
349 	/* all non-return stats set FSFREQ_ERROR*/
350 	req->status |= ZFCP_STATUS_FSFREQ_ERROR;
351 }
352 
zfcp_fsf_fsfstatus_eval(struct zfcp_fsf_req * req)353 static void zfcp_fsf_fsfstatus_eval(struct zfcp_fsf_req *req)
354 {
355 	if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR))
356 		return;
357 
358 	switch (req->qtcb->header.fsf_status) {
359 	case FSF_UNKNOWN_COMMAND:
360 		dev_err(&req->adapter->ccw_device->dev,
361 			"The FCP adapter does not recognize the command 0x%x\n",
362 			req->qtcb->header.fsf_command);
363 		zfcp_erp_adapter_shutdown(req->adapter, 0, "fsfse_1");
364 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
365 		break;
366 	case FSF_ADAPTER_STATUS_AVAILABLE:
367 		zfcp_fsf_fsfstatus_qual_eval(req);
368 		break;
369 	}
370 }
371 
zfcp_fsf_protstatus_eval(struct zfcp_fsf_req * req)372 static void zfcp_fsf_protstatus_eval(struct zfcp_fsf_req *req)
373 {
374 	struct zfcp_adapter *adapter = req->adapter;
375 	struct fsf_qtcb *qtcb = req->qtcb;
376 	union fsf_prot_status_qual *psq = &qtcb->prefix.prot_status_qual;
377 
378 	zfcp_dbf_hba_fsf_response(req);
379 
380 	if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) {
381 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
382 		return;
383 	}
384 
385 	switch (qtcb->prefix.prot_status) {
386 	case FSF_PROT_GOOD:
387 	case FSF_PROT_FSF_STATUS_PRESENTED:
388 		return;
389 	case FSF_PROT_QTCB_VERSION_ERROR:
390 		dev_err(&adapter->ccw_device->dev,
391 			"QTCB version 0x%x not supported by FCP adapter "
392 			"(0x%x to 0x%x)\n", FSF_QTCB_CURRENT_VERSION,
393 			psq->word[0], psq->word[1]);
394 		zfcp_erp_adapter_shutdown(adapter, 0, "fspse_1");
395 		break;
396 	case FSF_PROT_ERROR_STATE:
397 	case FSF_PROT_SEQ_NUMB_ERROR:
398 		zfcp_erp_adapter_reopen(adapter, 0, "fspse_2");
399 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
400 		break;
401 	case FSF_PROT_UNSUPP_QTCB_TYPE:
402 		dev_err(&adapter->ccw_device->dev,
403 			"The QTCB type is not supported by the FCP adapter\n");
404 		zfcp_erp_adapter_shutdown(adapter, 0, "fspse_3");
405 		break;
406 	case FSF_PROT_HOST_CONNECTION_INITIALIZING:
407 		atomic_or(ZFCP_STATUS_ADAPTER_HOST_CON_INIT,
408 				&adapter->status);
409 		break;
410 	case FSF_PROT_DUPLICATE_REQUEST_ID:
411 		dev_err(&adapter->ccw_device->dev,
412 			"0x%Lx is an ambiguous request identifier\n",
413 			(unsigned long long)qtcb->bottom.support.req_handle);
414 		zfcp_erp_adapter_shutdown(adapter, 0, "fspse_4");
415 		break;
416 	case FSF_PROT_LINK_DOWN:
417 		zfcp_fsf_link_down_info_eval(req, &psq->link_down_info);
418 		/* go through reopen to flush pending requests */
419 		zfcp_erp_adapter_reopen(adapter, 0, "fspse_6");
420 		break;
421 	case FSF_PROT_REEST_QUEUE:
422 		/* All ports should be marked as ready to run again */
423 		zfcp_erp_set_adapter_status(adapter,
424 					    ZFCP_STATUS_COMMON_RUNNING);
425 		zfcp_erp_adapter_reopen(adapter,
426 					ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED |
427 					ZFCP_STATUS_COMMON_ERP_FAILED,
428 					"fspse_8");
429 		break;
430 	default:
431 		dev_err(&adapter->ccw_device->dev,
432 			"0x%x is not a valid transfer protocol status\n",
433 			qtcb->prefix.prot_status);
434 		zfcp_qdio_siosl(adapter);
435 		zfcp_erp_adapter_shutdown(adapter, 0, "fspse_9");
436 	}
437 	req->status |= ZFCP_STATUS_FSFREQ_ERROR;
438 }
439 
440 /**
441  * zfcp_fsf_req_complete - process completion of a FSF request
442  * @req: The FSF request that has been completed.
443  *
444  * When a request has been completed either from the FCP adapter,
445  * or it has been dismissed due to a queue shutdown, this function
446  * is called to process the completion status and trigger further
447  * events related to the FSF request.
448  * Caller must ensure that the request has been removed from
449  * adapter->req_list, to protect against concurrent modification
450  * by zfcp_erp_strategy_check_fsfreq().
451  */
zfcp_fsf_req_complete(struct zfcp_fsf_req * req)452 static void zfcp_fsf_req_complete(struct zfcp_fsf_req *req)
453 {
454 	struct zfcp_erp_action *erp_action;
455 
456 	if (unlikely(zfcp_fsf_req_is_status_read_buffer(req))) {
457 		zfcp_fsf_status_read_handler(req);
458 		return;
459 	}
460 
461 	del_timer_sync(&req->timer);
462 	zfcp_fsf_protstatus_eval(req);
463 	zfcp_fsf_fsfstatus_eval(req);
464 	req->handler(req);
465 
466 	erp_action = req->erp_action;
467 	if (erp_action)
468 		zfcp_erp_notify(erp_action, 0);
469 
470 	if (likely(req->status & ZFCP_STATUS_FSFREQ_CLEANUP))
471 		zfcp_fsf_req_free(req);
472 	else
473 		complete(&req->completion);
474 }
475 
476 /**
477  * zfcp_fsf_req_dismiss_all - dismiss all fsf requests
478  * @adapter: pointer to struct zfcp_adapter
479  *
480  * Never ever call this without shutting down the adapter first.
481  * Otherwise the adapter would continue using and corrupting s390 storage.
482  * Included BUG_ON() call to ensure this is done.
483  * ERP is supposed to be the only user of this function.
484  */
zfcp_fsf_req_dismiss_all(struct zfcp_adapter * adapter)485 void zfcp_fsf_req_dismiss_all(struct zfcp_adapter *adapter)
486 {
487 	struct zfcp_fsf_req *req, *tmp;
488 	LIST_HEAD(remove_queue);
489 
490 	BUG_ON(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP);
491 	zfcp_reqlist_move(adapter->req_list, &remove_queue);
492 
493 	list_for_each_entry_safe(req, tmp, &remove_queue, list) {
494 		list_del(&req->list);
495 		req->status |= ZFCP_STATUS_FSFREQ_DISMISSED;
496 		zfcp_fsf_req_complete(req);
497 	}
498 }
499 
500 #define ZFCP_FSF_PORTSPEED_1GBIT	(1 <<  0)
501 #define ZFCP_FSF_PORTSPEED_2GBIT	(1 <<  1)
502 #define ZFCP_FSF_PORTSPEED_4GBIT	(1 <<  2)
503 #define ZFCP_FSF_PORTSPEED_10GBIT	(1 <<  3)
504 #define ZFCP_FSF_PORTSPEED_8GBIT	(1 <<  4)
505 #define ZFCP_FSF_PORTSPEED_16GBIT	(1 <<  5)
506 #define ZFCP_FSF_PORTSPEED_32GBIT	(1 <<  6)
507 #define ZFCP_FSF_PORTSPEED_64GBIT	(1 <<  7)
508 #define ZFCP_FSF_PORTSPEED_128GBIT	(1 <<  8)
509 #define ZFCP_FSF_PORTSPEED_NOT_NEGOTIATED (1 << 15)
510 
zfcp_fsf_convert_portspeed(u32 fsf_speed)511 u32 zfcp_fsf_convert_portspeed(u32 fsf_speed)
512 {
513 	u32 fdmi_speed = 0;
514 	if (fsf_speed & ZFCP_FSF_PORTSPEED_1GBIT)
515 		fdmi_speed |= FC_PORTSPEED_1GBIT;
516 	if (fsf_speed & ZFCP_FSF_PORTSPEED_2GBIT)
517 		fdmi_speed |= FC_PORTSPEED_2GBIT;
518 	if (fsf_speed & ZFCP_FSF_PORTSPEED_4GBIT)
519 		fdmi_speed |= FC_PORTSPEED_4GBIT;
520 	if (fsf_speed & ZFCP_FSF_PORTSPEED_10GBIT)
521 		fdmi_speed |= FC_PORTSPEED_10GBIT;
522 	if (fsf_speed & ZFCP_FSF_PORTSPEED_8GBIT)
523 		fdmi_speed |= FC_PORTSPEED_8GBIT;
524 	if (fsf_speed & ZFCP_FSF_PORTSPEED_16GBIT)
525 		fdmi_speed |= FC_PORTSPEED_16GBIT;
526 	if (fsf_speed & ZFCP_FSF_PORTSPEED_32GBIT)
527 		fdmi_speed |= FC_PORTSPEED_32GBIT;
528 	if (fsf_speed & ZFCP_FSF_PORTSPEED_64GBIT)
529 		fdmi_speed |= FC_PORTSPEED_64GBIT;
530 	if (fsf_speed & ZFCP_FSF_PORTSPEED_128GBIT)
531 		fdmi_speed |= FC_PORTSPEED_128GBIT;
532 	if (fsf_speed & ZFCP_FSF_PORTSPEED_NOT_NEGOTIATED)
533 		fdmi_speed |= FC_PORTSPEED_NOT_NEGOTIATED;
534 	return fdmi_speed;
535 }
536 
zfcp_fsf_exchange_config_evaluate(struct zfcp_fsf_req * req)537 static int zfcp_fsf_exchange_config_evaluate(struct zfcp_fsf_req *req)
538 {
539 	struct fsf_qtcb_bottom_config *bottom = &req->qtcb->bottom.config;
540 	struct zfcp_adapter *adapter = req->adapter;
541 	struct fc_els_flogi *plogi;
542 
543 	/* adjust pointers for missing command code */
544 	plogi = (struct fc_els_flogi *) ((u8 *)&bottom->plogi_payload
545 					- sizeof(u32));
546 
547 	if (req->data)
548 		memcpy(req->data, bottom, sizeof(*bottom));
549 
550 	adapter->timer_ticks = bottom->timer_interval & ZFCP_FSF_TIMER_INT_MASK;
551 	adapter->stat_read_buf_num = max(bottom->status_read_buf_num,
552 					 (u16)FSF_STATUS_READS_RECOM);
553 
554 	/* no error return above here, otherwise must fix call chains */
555 	/* do not evaluate invalid fields */
556 	if (req->qtcb->header.fsf_status == FSF_EXCHANGE_CONFIG_DATA_INCOMPLETE)
557 		return 0;
558 
559 	adapter->hydra_version = bottom->adapter_type;
560 
561 	switch (bottom->fc_topology) {
562 	case FSF_TOPO_P2P:
563 		adapter->peer_d_id = ntoh24(bottom->peer_d_id);
564 		adapter->peer_wwpn = be64_to_cpu(plogi->fl_wwpn);
565 		adapter->peer_wwnn = be64_to_cpu(plogi->fl_wwnn);
566 		break;
567 	case FSF_TOPO_FABRIC:
568 		break;
569 	case FSF_TOPO_AL:
570 	default:
571 		dev_err(&adapter->ccw_device->dev,
572 			"Unknown or unsupported arbitrated loop "
573 			"fibre channel topology detected\n");
574 		zfcp_erp_adapter_shutdown(adapter, 0, "fsece_1");
575 		return -EIO;
576 	}
577 
578 	return 0;
579 }
580 
zfcp_fsf_exchange_config_data_handler(struct zfcp_fsf_req * req)581 static void zfcp_fsf_exchange_config_data_handler(struct zfcp_fsf_req *req)
582 {
583 	struct zfcp_adapter *adapter = req->adapter;
584 	struct zfcp_diag_header *const diag_hdr =
585 		&adapter->diagnostics->config_data.header;
586 	struct fsf_qtcb *qtcb = req->qtcb;
587 	struct fsf_qtcb_bottom_config *bottom = &qtcb->bottom.config;
588 
589 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
590 		return;
591 
592 	adapter->fsf_lic_version = bottom->lic_version;
593 	adapter->adapter_features = bottom->adapter_features;
594 	adapter->connection_features = bottom->connection_features;
595 	adapter->peer_wwpn = 0;
596 	adapter->peer_wwnn = 0;
597 	adapter->peer_d_id = 0;
598 
599 	switch (qtcb->header.fsf_status) {
600 	case FSF_GOOD:
601 		/*
602 		 * usually we wait with an update till the cache is too old,
603 		 * but because we have the data available, update it anyway
604 		 */
605 		zfcp_diag_update_xdata(diag_hdr, bottom, false);
606 
607 		zfcp_scsi_shost_update_config_data(adapter, bottom, false);
608 		if (zfcp_fsf_exchange_config_evaluate(req))
609 			return;
610 
611 		if (bottom->max_qtcb_size < sizeof(struct fsf_qtcb)) {
612 			dev_err(&adapter->ccw_device->dev,
613 				"FCP adapter maximum QTCB size (%d bytes) "
614 				"is too small\n",
615 				bottom->max_qtcb_size);
616 			zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh1");
617 			return;
618 		}
619 		atomic_or(ZFCP_STATUS_ADAPTER_XCONFIG_OK,
620 				&adapter->status);
621 		break;
622 	case FSF_EXCHANGE_CONFIG_DATA_INCOMPLETE:
623 		zfcp_diag_update_xdata(diag_hdr, bottom, true);
624 		req->status |= ZFCP_STATUS_FSFREQ_XDATAINCOMPLETE;
625 
626 		/* avoids adapter shutdown to be able to recognize
627 		 * events such as LINK UP */
628 		atomic_or(ZFCP_STATUS_ADAPTER_XCONFIG_OK,
629 				&adapter->status);
630 		zfcp_fsf_link_down_info_eval(req,
631 			&qtcb->header.fsf_status_qual.link_down_info);
632 
633 		zfcp_scsi_shost_update_config_data(adapter, bottom, true);
634 		if (zfcp_fsf_exchange_config_evaluate(req))
635 			return;
636 		break;
637 	default:
638 		zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh3");
639 		return;
640 	}
641 
642 	if (adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT)
643 		adapter->hardware_version = bottom->hardware_version;
644 
645 	if (FSF_QTCB_CURRENT_VERSION < bottom->low_qtcb_version) {
646 		dev_err(&adapter->ccw_device->dev,
647 			"The FCP adapter only supports newer "
648 			"control block versions\n");
649 		zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh4");
650 		return;
651 	}
652 	if (FSF_QTCB_CURRENT_VERSION > bottom->high_qtcb_version) {
653 		dev_err(&adapter->ccw_device->dev,
654 			"The FCP adapter only supports older "
655 			"control block versions\n");
656 		zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh5");
657 	}
658 }
659 
660 /*
661  * Mapping of FC Endpoint Security flag masks to mnemonics
662  *
663  * NOTE: Update macro ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH when making any
664  *       changes.
665  */
666 static const struct {
667 	u32	mask;
668 	char	*name;
669 } zfcp_fsf_fc_security_mnemonics[] = {
670 	{ FSF_FC_SECURITY_AUTH,		"Authentication" },
671 	{ FSF_FC_SECURITY_ENC_FCSP2 |
672 	  FSF_FC_SECURITY_ENC_ERAS,	"Encryption" },
673 };
674 
675 /* maximum strlen(zfcp_fsf_fc_security_mnemonics[...].name) + 1 */
676 #define ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH 15
677 
678 /**
679  * zfcp_fsf_scnprint_fc_security() - translate FC Endpoint Security flags into
680  *                                   mnemonics and place in a buffer
681  * @buf        : the buffer to place the translated FC Endpoint Security flag(s)
682  *               into
683  * @size       : the size of the buffer, including the trailing null space
684  * @fc_security: one or more FC Endpoint Security flags, or zero
685  * @fmt        : specifies whether a list or a single item is to be put into the
686  *               buffer
687  *
688  * The Fibre Channel (FC) Endpoint Security flags are translated into mnemonics.
689  * If the FC Endpoint Security flags are zero "none" is placed into the buffer.
690  *
691  * With ZFCP_FSF_PRINT_FMT_LIST the mnemonics are placed as a list separated by
692  * a comma followed by a space into the buffer. If one or more FC Endpoint
693  * Security flags cannot be translated into a mnemonic, as they are undefined
694  * in zfcp_fsf_fc_security_mnemonics, their bitwise ORed value in hexadecimal
695  * representation is placed into the buffer.
696  *
697  * With ZFCP_FSF_PRINT_FMT_SINGLEITEM only one single mnemonic is placed into
698  * the buffer. If the FC Endpoint Security flag cannot be translated, as it is
699  * undefined in zfcp_fsf_fc_security_mnemonics, its value in hexadecimal
700  * representation is placed into the buffer. If more than one FC Endpoint
701  * Security flag was specified, their value in hexadecimal representation is
702  * placed into the buffer. The macro ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH
703  * can be used to define a buffer that is large enough to hold one mnemonic.
704  *
705  * Return: The number of characters written into buf not including the trailing
706  *         '\0'. If size is == 0 the function returns 0.
707  */
zfcp_fsf_scnprint_fc_security(char * buf,size_t size,u32 fc_security,enum zfcp_fsf_print_fmt fmt)708 ssize_t zfcp_fsf_scnprint_fc_security(char *buf, size_t size, u32 fc_security,
709 				      enum zfcp_fsf_print_fmt fmt)
710 {
711 	const char *prefix = "";
712 	ssize_t len = 0;
713 	int i;
714 
715 	if (fc_security == 0)
716 		return scnprintf(buf, size, "none");
717 	if (fmt == ZFCP_FSF_PRINT_FMT_SINGLEITEM && hweight32(fc_security) != 1)
718 		return scnprintf(buf, size, "0x%08x", fc_security);
719 
720 	for (i = 0; i < ARRAY_SIZE(zfcp_fsf_fc_security_mnemonics); i++) {
721 		if (!(fc_security & zfcp_fsf_fc_security_mnemonics[i].mask))
722 			continue;
723 
724 		len += scnprintf(buf + len, size - len, "%s%s", prefix,
725 				 zfcp_fsf_fc_security_mnemonics[i].name);
726 		prefix = ", ";
727 		fc_security &= ~zfcp_fsf_fc_security_mnemonics[i].mask;
728 	}
729 
730 	if (fc_security != 0)
731 		len += scnprintf(buf + len, size - len, "%s0x%08x",
732 				 prefix, fc_security);
733 
734 	return len;
735 }
736 
zfcp_fsf_dbf_adapter_fc_security(struct zfcp_adapter * adapter,struct zfcp_fsf_req * req)737 static void zfcp_fsf_dbf_adapter_fc_security(struct zfcp_adapter *adapter,
738 					     struct zfcp_fsf_req *req)
739 {
740 	if (adapter->fc_security_algorithms ==
741 	    adapter->fc_security_algorithms_old) {
742 		/* no change, no trace */
743 		return;
744 	}
745 
746 	zfcp_dbf_hba_fsf_fces("fsfcesa", req, ZFCP_DBF_INVALID_WWPN,
747 			      adapter->fc_security_algorithms_old,
748 			      adapter->fc_security_algorithms);
749 
750 	adapter->fc_security_algorithms_old = adapter->fc_security_algorithms;
751 }
752 
zfcp_fsf_exchange_port_evaluate(struct zfcp_fsf_req * req)753 static void zfcp_fsf_exchange_port_evaluate(struct zfcp_fsf_req *req)
754 {
755 	struct zfcp_adapter *adapter = req->adapter;
756 	struct fsf_qtcb_bottom_port *bottom = &req->qtcb->bottom.port;
757 
758 	if (req->data)
759 		memcpy(req->data, bottom, sizeof(*bottom));
760 
761 	if (adapter->adapter_features & FSF_FEATURE_FC_SECURITY)
762 		adapter->fc_security_algorithms =
763 			bottom->fc_security_algorithms;
764 	else
765 		adapter->fc_security_algorithms = 0;
766 	zfcp_fsf_dbf_adapter_fc_security(adapter, req);
767 }
768 
zfcp_fsf_exchange_port_data_handler(struct zfcp_fsf_req * req)769 static void zfcp_fsf_exchange_port_data_handler(struct zfcp_fsf_req *req)
770 {
771 	struct zfcp_diag_header *const diag_hdr =
772 		&req->adapter->diagnostics->port_data.header;
773 	struct fsf_qtcb *qtcb = req->qtcb;
774 	struct fsf_qtcb_bottom_port *bottom = &qtcb->bottom.port;
775 
776 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
777 		return;
778 
779 	switch (qtcb->header.fsf_status) {
780 	case FSF_GOOD:
781 		/*
782 		 * usually we wait with an update till the cache is too old,
783 		 * but because we have the data available, update it anyway
784 		 */
785 		zfcp_diag_update_xdata(diag_hdr, bottom, false);
786 
787 		zfcp_scsi_shost_update_port_data(req->adapter, bottom);
788 		zfcp_fsf_exchange_port_evaluate(req);
789 		break;
790 	case FSF_EXCHANGE_CONFIG_DATA_INCOMPLETE:
791 		zfcp_diag_update_xdata(diag_hdr, bottom, true);
792 		req->status |= ZFCP_STATUS_FSFREQ_XDATAINCOMPLETE;
793 
794 		zfcp_fsf_link_down_info_eval(req,
795 			&qtcb->header.fsf_status_qual.link_down_info);
796 
797 		zfcp_scsi_shost_update_port_data(req->adapter, bottom);
798 		zfcp_fsf_exchange_port_evaluate(req);
799 		break;
800 	}
801 }
802 
zfcp_fsf_alloc(mempool_t * pool)803 static struct zfcp_fsf_req *zfcp_fsf_alloc(mempool_t *pool)
804 {
805 	struct zfcp_fsf_req *req;
806 
807 	if (likely(pool))
808 		req = mempool_alloc(pool, GFP_ATOMIC);
809 	else
810 		req = kmalloc(sizeof(*req), GFP_ATOMIC);
811 
812 	if (unlikely(!req))
813 		return NULL;
814 
815 	memset(req, 0, sizeof(*req));
816 	req->pool = pool;
817 	return req;
818 }
819 
zfcp_fsf_qtcb_alloc(mempool_t * pool)820 static struct fsf_qtcb *zfcp_fsf_qtcb_alloc(mempool_t *pool)
821 {
822 	struct fsf_qtcb *qtcb;
823 
824 	if (likely(pool))
825 		qtcb = mempool_alloc(pool, GFP_ATOMIC);
826 	else
827 		qtcb = kmem_cache_alloc(zfcp_fsf_qtcb_cache, GFP_ATOMIC);
828 
829 	if (unlikely(!qtcb))
830 		return NULL;
831 
832 	memset(qtcb, 0, sizeof(*qtcb));
833 	return qtcb;
834 }
835 
zfcp_fsf_req_create(struct zfcp_qdio * qdio,u32 fsf_cmd,u8 sbtype,mempool_t * pool)836 static struct zfcp_fsf_req *zfcp_fsf_req_create(struct zfcp_qdio *qdio,
837 						u32 fsf_cmd, u8 sbtype,
838 						mempool_t *pool)
839 {
840 	struct zfcp_adapter *adapter = qdio->adapter;
841 	struct zfcp_fsf_req *req = zfcp_fsf_alloc(pool);
842 
843 	if (unlikely(!req))
844 		return ERR_PTR(-ENOMEM);
845 
846 	if (adapter->req_no == 0)
847 		adapter->req_no++;
848 
849 	timer_setup(&req->timer, NULL, 0);
850 	init_completion(&req->completion);
851 
852 	req->adapter = adapter;
853 	req->req_id = adapter->req_no;
854 
855 	if (likely(fsf_cmd != FSF_QTCB_UNSOLICITED_STATUS)) {
856 		if (likely(pool))
857 			req->qtcb = zfcp_fsf_qtcb_alloc(
858 				adapter->pool.qtcb_pool);
859 		else
860 			req->qtcb = zfcp_fsf_qtcb_alloc(NULL);
861 
862 		if (unlikely(!req->qtcb)) {
863 			zfcp_fsf_req_free(req);
864 			return ERR_PTR(-ENOMEM);
865 		}
866 
867 		req->qtcb->prefix.req_seq_no = adapter->fsf_req_seq_no;
868 		req->qtcb->prefix.req_id = req->req_id;
869 		req->qtcb->prefix.ulp_info = 26;
870 		req->qtcb->prefix.qtcb_type = fsf_qtcb_type[fsf_cmd];
871 		req->qtcb->prefix.qtcb_version = FSF_QTCB_CURRENT_VERSION;
872 		req->qtcb->header.req_handle = req->req_id;
873 		req->qtcb->header.fsf_command = fsf_cmd;
874 	}
875 
876 	zfcp_qdio_req_init(adapter->qdio, &req->qdio_req, req->req_id, sbtype,
877 			   req->qtcb, sizeof(struct fsf_qtcb));
878 
879 	return req;
880 }
881 
zfcp_fsf_req_send(struct zfcp_fsf_req * req)882 static int zfcp_fsf_req_send(struct zfcp_fsf_req *req)
883 {
884 	const bool is_srb = zfcp_fsf_req_is_status_read_buffer(req);
885 	struct zfcp_adapter *adapter = req->adapter;
886 	struct zfcp_qdio *qdio = adapter->qdio;
887 	u64 req_id = req->req_id;
888 
889 	zfcp_reqlist_add(adapter->req_list, req);
890 
891 	req->qdio_req.qdio_outb_usage = atomic_read(&qdio->req_q_free);
892 	req->issued = get_tod_clock();
893 	if (zfcp_qdio_send(qdio, &req->qdio_req)) {
894 		del_timer_sync(&req->timer);
895 
896 		/* lookup request again, list might have changed */
897 		if (zfcp_reqlist_find_rm(adapter->req_list, req_id) == NULL)
898 			zfcp_dbf_hba_fsf_reqid("fsrsrmf", 1, adapter, req_id);
899 
900 		zfcp_erp_adapter_reopen(adapter, 0, "fsrs__1");
901 		return -EIO;
902 	}
903 
904 	/*
905 	 * NOTE: DO NOT TOUCH ASYNC req PAST THIS POINT.
906 	 *	 ONLY TOUCH SYNC req AGAIN ON req->completion.
907 	 *
908 	 * The request might complete and be freed concurrently at any point
909 	 * now. This is not protected by the QDIO-lock (req_q_lock). So any
910 	 * uncontrolled access after this might result in an use-after-free bug.
911 	 * Only if the request doesn't have ZFCP_STATUS_FSFREQ_CLEANUP set, and
912 	 * when it is completed via req->completion, is it safe to use req
913 	 * again.
914 	 */
915 
916 	/* Don't increase for unsolicited status */
917 	if (!is_srb)
918 		adapter->fsf_req_seq_no++;
919 	adapter->req_no++;
920 
921 	return 0;
922 }
923 
924 /**
925  * zfcp_fsf_status_read - send status read request
926  * @qdio: pointer to struct zfcp_qdio
927  * Returns: 0 on success, ERROR otherwise
928  */
zfcp_fsf_status_read(struct zfcp_qdio * qdio)929 int zfcp_fsf_status_read(struct zfcp_qdio *qdio)
930 {
931 	struct zfcp_adapter *adapter = qdio->adapter;
932 	struct zfcp_fsf_req *req;
933 	struct fsf_status_read_buffer *sr_buf;
934 	struct page *page;
935 	int retval = -EIO;
936 
937 	spin_lock_irq(&qdio->req_q_lock);
938 	if (zfcp_qdio_sbal_get(qdio))
939 		goto out;
940 
941 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_UNSOLICITED_STATUS,
942 				  SBAL_SFLAGS0_TYPE_STATUS,
943 				  adapter->pool.status_read_req);
944 	if (IS_ERR(req)) {
945 		retval = PTR_ERR(req);
946 		goto out;
947 	}
948 
949 	page = mempool_alloc(adapter->pool.sr_data, GFP_ATOMIC);
950 	if (!page) {
951 		retval = -ENOMEM;
952 		goto failed_buf;
953 	}
954 	sr_buf = page_address(page);
955 	memset(sr_buf, 0, sizeof(*sr_buf));
956 	req->data = sr_buf;
957 
958 	zfcp_qdio_fill_next(qdio, &req->qdio_req, sr_buf, sizeof(*sr_buf));
959 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
960 
961 	retval = zfcp_fsf_req_send(req);
962 	if (retval)
963 		goto failed_req_send;
964 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
965 
966 	goto out;
967 
968 failed_req_send:
969 	req->data = NULL;
970 	mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data);
971 failed_buf:
972 	zfcp_dbf_hba_fsf_uss("fssr__1", req);
973 	zfcp_fsf_req_free(req);
974 out:
975 	spin_unlock_irq(&qdio->req_q_lock);
976 	return retval;
977 }
978 
zfcp_fsf_abort_fcp_command_handler(struct zfcp_fsf_req * req)979 static void zfcp_fsf_abort_fcp_command_handler(struct zfcp_fsf_req *req)
980 {
981 	struct scsi_device *sdev = req->data;
982 	struct zfcp_scsi_dev *zfcp_sdev;
983 	union fsf_status_qual *fsq = &req->qtcb->header.fsf_status_qual;
984 
985 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
986 		return;
987 
988 	zfcp_sdev = sdev_to_zfcp(sdev);
989 
990 	switch (req->qtcb->header.fsf_status) {
991 	case FSF_PORT_HANDLE_NOT_VALID:
992 		if (fsq->word[0] == fsq->word[1]) {
993 			zfcp_erp_adapter_reopen(zfcp_sdev->port->adapter, 0,
994 						"fsafch1");
995 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
996 		}
997 		break;
998 	case FSF_LUN_HANDLE_NOT_VALID:
999 		if (fsq->word[0] == fsq->word[1]) {
1000 			zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fsafch2");
1001 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1002 		}
1003 		break;
1004 	case FSF_FCP_COMMAND_DOES_NOT_EXIST:
1005 		req->status |= ZFCP_STATUS_FSFREQ_ABORTNOTNEEDED;
1006 		break;
1007 	case FSF_PORT_BOXED:
1008 		zfcp_erp_set_port_status(zfcp_sdev->port,
1009 					 ZFCP_STATUS_COMMON_ACCESS_BOXED);
1010 		zfcp_erp_port_reopen(zfcp_sdev->port,
1011 				     ZFCP_STATUS_COMMON_ERP_FAILED, "fsafch3");
1012 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1013 		break;
1014 	case FSF_LUN_BOXED:
1015 		zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ACCESS_BOXED);
1016 		zfcp_erp_lun_reopen(sdev, ZFCP_STATUS_COMMON_ERP_FAILED,
1017 				    "fsafch4");
1018 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1019                 break;
1020 	case FSF_ADAPTER_STATUS_AVAILABLE:
1021 		switch (fsq->word[0]) {
1022 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
1023 			zfcp_fc_test_link(zfcp_sdev->port);
1024 			fallthrough;
1025 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
1026 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1027 			break;
1028 		}
1029 		break;
1030 	case FSF_GOOD:
1031 		req->status |= ZFCP_STATUS_FSFREQ_ABORTSUCCEEDED;
1032 		break;
1033 	}
1034 }
1035 
1036 /**
1037  * zfcp_fsf_abort_fcp_cmnd - abort running SCSI command
1038  * @scmnd: The SCSI command to abort
1039  * Returns: pointer to struct zfcp_fsf_req
1040  */
1041 
zfcp_fsf_abort_fcp_cmnd(struct scsi_cmnd * scmnd)1042 struct zfcp_fsf_req *zfcp_fsf_abort_fcp_cmnd(struct scsi_cmnd *scmnd)
1043 {
1044 	struct zfcp_fsf_req *req = NULL;
1045 	struct scsi_device *sdev = scmnd->device;
1046 	struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev);
1047 	struct zfcp_qdio *qdio = zfcp_sdev->port->adapter->qdio;
1048 	u64 old_req_id = (u64) scmnd->host_scribble;
1049 
1050 	spin_lock_irq(&qdio->req_q_lock);
1051 	if (zfcp_qdio_sbal_get(qdio))
1052 		goto out;
1053 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_ABORT_FCP_CMND,
1054 				  SBAL_SFLAGS0_TYPE_READ,
1055 				  qdio->adapter->pool.scsi_abort);
1056 	if (IS_ERR(req)) {
1057 		req = NULL;
1058 		goto out;
1059 	}
1060 
1061 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
1062 		       ZFCP_STATUS_COMMON_UNBLOCKED)))
1063 		goto out_error_free;
1064 
1065 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1066 
1067 	req->data = sdev;
1068 	req->handler = zfcp_fsf_abort_fcp_command_handler;
1069 	req->qtcb->header.lun_handle = zfcp_sdev->lun_handle;
1070 	req->qtcb->header.port_handle = zfcp_sdev->port->handle;
1071 	req->qtcb->bottom.support.req_handle = old_req_id;
1072 
1073 	zfcp_fsf_start_timer(req, ZFCP_FSF_SCSI_ER_TIMEOUT);
1074 	if (!zfcp_fsf_req_send(req)) {
1075 		/* NOTE: DO NOT TOUCH req, UNTIL IT COMPLETES! */
1076 		goto out;
1077 	}
1078 
1079 out_error_free:
1080 	zfcp_fsf_req_free(req);
1081 	req = NULL;
1082 out:
1083 	spin_unlock_irq(&qdio->req_q_lock);
1084 	return req;
1085 }
1086 
zfcp_fsf_send_ct_handler(struct zfcp_fsf_req * req)1087 static void zfcp_fsf_send_ct_handler(struct zfcp_fsf_req *req)
1088 {
1089 	struct zfcp_adapter *adapter = req->adapter;
1090 	struct zfcp_fsf_ct_els *ct = req->data;
1091 	struct fsf_qtcb_header *header = &req->qtcb->header;
1092 
1093 	ct->status = -EINVAL;
1094 
1095 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
1096 		goto skip_fsfstatus;
1097 
1098 	switch (header->fsf_status) {
1099         case FSF_GOOD:
1100 		ct->status = 0;
1101 		zfcp_dbf_san_res("fsscth2", req);
1102 		break;
1103         case FSF_SERVICE_CLASS_NOT_SUPPORTED:
1104 		zfcp_fsf_class_not_supp(req);
1105 		break;
1106         case FSF_ADAPTER_STATUS_AVAILABLE:
1107                 switch (header->fsf_status_qual.word[0]){
1108                 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
1109                 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
1110 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1111 			break;
1112                 }
1113                 break;
1114         case FSF_PORT_BOXED:
1115 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1116 		break;
1117 	case FSF_PORT_HANDLE_NOT_VALID:
1118 		zfcp_erp_adapter_reopen(adapter, 0, "fsscth1");
1119 		fallthrough;
1120 	case FSF_GENERIC_COMMAND_REJECTED:
1121 	case FSF_PAYLOAD_SIZE_MISMATCH:
1122 	case FSF_REQUEST_SIZE_TOO_LARGE:
1123 	case FSF_RESPONSE_SIZE_TOO_LARGE:
1124 	case FSF_SBAL_MISMATCH:
1125 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1126 		break;
1127 	}
1128 
1129 skip_fsfstatus:
1130 	if (ct->handler)
1131 		ct->handler(ct->handler_data);
1132 }
1133 
zfcp_fsf_setup_ct_els_unchained(struct zfcp_qdio * qdio,struct zfcp_qdio_req * q_req,struct scatterlist * sg_req,struct scatterlist * sg_resp)1134 static void zfcp_fsf_setup_ct_els_unchained(struct zfcp_qdio *qdio,
1135 					    struct zfcp_qdio_req *q_req,
1136 					    struct scatterlist *sg_req,
1137 					    struct scatterlist *sg_resp)
1138 {
1139 	zfcp_qdio_fill_next(qdio, q_req, sg_virt(sg_req), sg_req->length);
1140 	zfcp_qdio_fill_next(qdio, q_req, sg_virt(sg_resp), sg_resp->length);
1141 	zfcp_qdio_set_sbale_last(qdio, q_req);
1142 }
1143 
zfcp_fsf_setup_ct_els_sbals(struct zfcp_fsf_req * req,struct scatterlist * sg_req,struct scatterlist * sg_resp)1144 static int zfcp_fsf_setup_ct_els_sbals(struct zfcp_fsf_req *req,
1145 				       struct scatterlist *sg_req,
1146 				       struct scatterlist *sg_resp)
1147 {
1148 	struct zfcp_adapter *adapter = req->adapter;
1149 	struct zfcp_qdio *qdio = adapter->qdio;
1150 	struct fsf_qtcb *qtcb = req->qtcb;
1151 	u32 feat = adapter->adapter_features;
1152 
1153 	if (zfcp_adapter_multi_buffer_active(adapter)) {
1154 		if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_req))
1155 			return -EIO;
1156 		qtcb->bottom.support.req_buf_length =
1157 			zfcp_qdio_real_bytes(sg_req);
1158 		if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_resp))
1159 			return -EIO;
1160 		qtcb->bottom.support.resp_buf_length =
1161 			zfcp_qdio_real_bytes(sg_resp);
1162 
1163 		zfcp_qdio_set_data_div(qdio, &req->qdio_req, sg_nents(sg_req));
1164 		zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1165 		zfcp_qdio_set_scount(qdio, &req->qdio_req);
1166 		return 0;
1167 	}
1168 
1169 	/* use single, unchained SBAL if it can hold the request */
1170 	if (zfcp_qdio_sg_one_sbale(sg_req) && zfcp_qdio_sg_one_sbale(sg_resp)) {
1171 		zfcp_fsf_setup_ct_els_unchained(qdio, &req->qdio_req,
1172 						sg_req, sg_resp);
1173 		return 0;
1174 	}
1175 
1176 	if (!(feat & FSF_FEATURE_ELS_CT_CHAINED_SBALS))
1177 		return -EOPNOTSUPP;
1178 
1179 	if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_req))
1180 		return -EIO;
1181 
1182 	qtcb->bottom.support.req_buf_length = zfcp_qdio_real_bytes(sg_req);
1183 
1184 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1185 	zfcp_qdio_skip_to_last_sbale(qdio, &req->qdio_req);
1186 
1187 	if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_resp))
1188 		return -EIO;
1189 
1190 	qtcb->bottom.support.resp_buf_length = zfcp_qdio_real_bytes(sg_resp);
1191 
1192 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1193 
1194 	return 0;
1195 }
1196 
zfcp_fsf_setup_ct_els(struct zfcp_fsf_req * req,struct scatterlist * sg_req,struct scatterlist * sg_resp,unsigned int timeout)1197 static int zfcp_fsf_setup_ct_els(struct zfcp_fsf_req *req,
1198 				 struct scatterlist *sg_req,
1199 				 struct scatterlist *sg_resp,
1200 				 unsigned int timeout)
1201 {
1202 	int ret;
1203 
1204 	ret = zfcp_fsf_setup_ct_els_sbals(req, sg_req, sg_resp);
1205 	if (ret)
1206 		return ret;
1207 
1208 	/* common settings for ct/gs and els requests */
1209 	if (timeout > 255)
1210 		timeout = 255; /* max value accepted by hardware */
1211 	req->qtcb->bottom.support.service_class = FSF_CLASS_3;
1212 	req->qtcb->bottom.support.timeout = timeout;
1213 	zfcp_fsf_start_timer(req, (timeout + 10) * HZ);
1214 
1215 	return 0;
1216 }
1217 
1218 /**
1219  * zfcp_fsf_send_ct - initiate a Generic Service request (FC-GS)
1220  * @wka_port: pointer to zfcp WKA port to send CT/GS to
1221  * @ct: pointer to struct zfcp_send_ct with data for request
1222  * @pool: if non-null this mempool is used to allocate struct zfcp_fsf_req
1223  * @timeout: timeout that hardware should use, and a later software timeout
1224  */
zfcp_fsf_send_ct(struct zfcp_fc_wka_port * wka_port,struct zfcp_fsf_ct_els * ct,mempool_t * pool,unsigned int timeout)1225 int zfcp_fsf_send_ct(struct zfcp_fc_wka_port *wka_port,
1226 		     struct zfcp_fsf_ct_els *ct, mempool_t *pool,
1227 		     unsigned int timeout)
1228 {
1229 	struct zfcp_qdio *qdio = wka_port->adapter->qdio;
1230 	struct zfcp_fsf_req *req;
1231 	int ret = -EIO;
1232 
1233 	spin_lock_irq(&qdio->req_q_lock);
1234 	if (zfcp_qdio_sbal_get(qdio))
1235 		goto out;
1236 
1237 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_SEND_GENERIC,
1238 				  SBAL_SFLAGS0_TYPE_WRITE_READ, pool);
1239 
1240 	if (IS_ERR(req)) {
1241 		ret = PTR_ERR(req);
1242 		goto out;
1243 	}
1244 
1245 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1246 	ret = zfcp_fsf_setup_ct_els(req, ct->req, ct->resp, timeout);
1247 	if (ret)
1248 		goto failed_send;
1249 
1250 	req->handler = zfcp_fsf_send_ct_handler;
1251 	req->qtcb->header.port_handle = wka_port->handle;
1252 	ct->d_id = wka_port->d_id;
1253 	req->data = ct;
1254 
1255 	zfcp_dbf_san_req("fssct_1", req, wka_port->d_id);
1256 
1257 	ret = zfcp_fsf_req_send(req);
1258 	if (ret)
1259 		goto failed_send;
1260 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1261 
1262 	goto out;
1263 
1264 failed_send:
1265 	zfcp_fsf_req_free(req);
1266 out:
1267 	spin_unlock_irq(&qdio->req_q_lock);
1268 	return ret;
1269 }
1270 
zfcp_fsf_send_els_handler(struct zfcp_fsf_req * req)1271 static void zfcp_fsf_send_els_handler(struct zfcp_fsf_req *req)
1272 {
1273 	struct zfcp_fsf_ct_els *send_els = req->data;
1274 	struct fsf_qtcb_header *header = &req->qtcb->header;
1275 
1276 	send_els->status = -EINVAL;
1277 
1278 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
1279 		goto skip_fsfstatus;
1280 
1281 	switch (header->fsf_status) {
1282 	case FSF_GOOD:
1283 		send_els->status = 0;
1284 		zfcp_dbf_san_res("fsselh1", req);
1285 		break;
1286 	case FSF_SERVICE_CLASS_NOT_SUPPORTED:
1287 		zfcp_fsf_class_not_supp(req);
1288 		break;
1289 	case FSF_ADAPTER_STATUS_AVAILABLE:
1290 		switch (header->fsf_status_qual.word[0]){
1291 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
1292 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
1293 		case FSF_SQ_RETRY_IF_POSSIBLE:
1294 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1295 			break;
1296 		}
1297 		break;
1298 	case FSF_ELS_COMMAND_REJECTED:
1299 	case FSF_PAYLOAD_SIZE_MISMATCH:
1300 	case FSF_REQUEST_SIZE_TOO_LARGE:
1301 	case FSF_RESPONSE_SIZE_TOO_LARGE:
1302 		break;
1303 	case FSF_SBAL_MISMATCH:
1304 		/* should never occur, avoided in zfcp_fsf_send_els */
1305 		fallthrough;
1306 	default:
1307 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1308 		break;
1309 	}
1310 skip_fsfstatus:
1311 	if (send_els->handler)
1312 		send_els->handler(send_els->handler_data);
1313 }
1314 
1315 /**
1316  * zfcp_fsf_send_els - initiate an ELS command (FC-FS)
1317  * @adapter: pointer to zfcp adapter
1318  * @d_id: N_Port_ID to send ELS to
1319  * @els: pointer to struct zfcp_send_els with data for the command
1320  * @timeout: timeout that hardware should use, and a later software timeout
1321  */
zfcp_fsf_send_els(struct zfcp_adapter * adapter,u32 d_id,struct zfcp_fsf_ct_els * els,unsigned int timeout)1322 int zfcp_fsf_send_els(struct zfcp_adapter *adapter, u32 d_id,
1323 		      struct zfcp_fsf_ct_els *els, unsigned int timeout)
1324 {
1325 	struct zfcp_fsf_req *req;
1326 	struct zfcp_qdio *qdio = adapter->qdio;
1327 	int ret = -EIO;
1328 
1329 	spin_lock_irq(&qdio->req_q_lock);
1330 	if (zfcp_qdio_sbal_get(qdio))
1331 		goto out;
1332 
1333 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_SEND_ELS,
1334 				  SBAL_SFLAGS0_TYPE_WRITE_READ, NULL);
1335 
1336 	if (IS_ERR(req)) {
1337 		ret = PTR_ERR(req);
1338 		goto out;
1339 	}
1340 
1341 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1342 
1343 	if (!zfcp_adapter_multi_buffer_active(adapter))
1344 		zfcp_qdio_sbal_limit(qdio, &req->qdio_req, 2);
1345 
1346 	ret = zfcp_fsf_setup_ct_els(req, els->req, els->resp, timeout);
1347 
1348 	if (ret)
1349 		goto failed_send;
1350 
1351 	hton24(req->qtcb->bottom.support.d_id, d_id);
1352 	req->handler = zfcp_fsf_send_els_handler;
1353 	els->d_id = d_id;
1354 	req->data = els;
1355 
1356 	zfcp_dbf_san_req("fssels1", req, d_id);
1357 
1358 	ret = zfcp_fsf_req_send(req);
1359 	if (ret)
1360 		goto failed_send;
1361 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1362 
1363 	goto out;
1364 
1365 failed_send:
1366 	zfcp_fsf_req_free(req);
1367 out:
1368 	spin_unlock_irq(&qdio->req_q_lock);
1369 	return ret;
1370 }
1371 
zfcp_fsf_exchange_config_data(struct zfcp_erp_action * erp_action)1372 int zfcp_fsf_exchange_config_data(struct zfcp_erp_action *erp_action)
1373 {
1374 	struct zfcp_fsf_req *req;
1375 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
1376 	int retval = -EIO;
1377 
1378 	spin_lock_irq(&qdio->req_q_lock);
1379 	if (zfcp_qdio_sbal_get(qdio))
1380 		goto out;
1381 
1382 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_CONFIG_DATA,
1383 				  SBAL_SFLAGS0_TYPE_READ,
1384 				  qdio->adapter->pool.erp_req);
1385 
1386 	if (IS_ERR(req)) {
1387 		retval = PTR_ERR(req);
1388 		goto out;
1389 	}
1390 
1391 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1392 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1393 
1394 	req->qtcb->bottom.config.feature_selection =
1395 			FSF_FEATURE_NOTIFICATION_LOST |
1396 			FSF_FEATURE_UPDATE_ALERT |
1397 			FSF_FEATURE_REQUEST_SFP_DATA |
1398 			FSF_FEATURE_FC_SECURITY;
1399 	req->erp_action = erp_action;
1400 	req->handler = zfcp_fsf_exchange_config_data_handler;
1401 	erp_action->fsf_req_id = req->req_id;
1402 
1403 	zfcp_fsf_start_erp_timer(req);
1404 	retval = zfcp_fsf_req_send(req);
1405 	if (retval) {
1406 		zfcp_fsf_req_free(req);
1407 		erp_action->fsf_req_id = 0;
1408 	}
1409 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1410 out:
1411 	spin_unlock_irq(&qdio->req_q_lock);
1412 	return retval;
1413 }
1414 
1415 
1416 /**
1417  * zfcp_fsf_exchange_config_data_sync() - Request information about FCP channel.
1418  * @qdio: pointer to the QDIO-Queue to use for sending the command.
1419  * @data: pointer to the QTCB-Bottom for storing the result of the command,
1420  *	  might be %NULL.
1421  *
1422  * Returns:
1423  * * 0		- Exchange Config Data was successful, @data is complete
1424  * * -EIO	- Exchange Config Data was not successful, @data is invalid
1425  * * -EAGAIN	- @data contains incomplete data
1426  * * -ENOMEM	- Some memory allocation failed along the way
1427  */
zfcp_fsf_exchange_config_data_sync(struct zfcp_qdio * qdio,struct fsf_qtcb_bottom_config * data)1428 int zfcp_fsf_exchange_config_data_sync(struct zfcp_qdio *qdio,
1429 				       struct fsf_qtcb_bottom_config *data)
1430 {
1431 	struct zfcp_fsf_req *req = NULL;
1432 	int retval = -EIO;
1433 
1434 	spin_lock_irq(&qdio->req_q_lock);
1435 	if (zfcp_qdio_sbal_get(qdio))
1436 		goto out_unlock;
1437 
1438 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_CONFIG_DATA,
1439 				  SBAL_SFLAGS0_TYPE_READ, NULL);
1440 
1441 	if (IS_ERR(req)) {
1442 		retval = PTR_ERR(req);
1443 		goto out_unlock;
1444 	}
1445 
1446 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1447 	req->handler = zfcp_fsf_exchange_config_data_handler;
1448 
1449 	req->qtcb->bottom.config.feature_selection =
1450 			FSF_FEATURE_NOTIFICATION_LOST |
1451 			FSF_FEATURE_UPDATE_ALERT |
1452 			FSF_FEATURE_REQUEST_SFP_DATA |
1453 			FSF_FEATURE_FC_SECURITY;
1454 
1455 	if (data)
1456 		req->data = data;
1457 
1458 	zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT);
1459 	retval = zfcp_fsf_req_send(req);
1460 	spin_unlock_irq(&qdio->req_q_lock);
1461 
1462 	if (!retval) {
1463 		/* NOTE: ONLY TOUCH SYNC req AGAIN ON req->completion. */
1464 		wait_for_completion(&req->completion);
1465 
1466 		if (req->status &
1467 		    (ZFCP_STATUS_FSFREQ_ERROR | ZFCP_STATUS_FSFREQ_DISMISSED))
1468 			retval = -EIO;
1469 		else if (req->status & ZFCP_STATUS_FSFREQ_XDATAINCOMPLETE)
1470 			retval = -EAGAIN;
1471 	}
1472 
1473 	zfcp_fsf_req_free(req);
1474 	return retval;
1475 
1476 out_unlock:
1477 	spin_unlock_irq(&qdio->req_q_lock);
1478 	return retval;
1479 }
1480 
1481 /**
1482  * zfcp_fsf_exchange_port_data - request information about local port
1483  * @erp_action: ERP action for the adapter for which port data is requested
1484  * Returns: 0 on success, error otherwise
1485  */
zfcp_fsf_exchange_port_data(struct zfcp_erp_action * erp_action)1486 int zfcp_fsf_exchange_port_data(struct zfcp_erp_action *erp_action)
1487 {
1488 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
1489 	struct zfcp_fsf_req *req;
1490 	int retval = -EIO;
1491 
1492 	if (!(qdio->adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT))
1493 		return -EOPNOTSUPP;
1494 
1495 	spin_lock_irq(&qdio->req_q_lock);
1496 	if (zfcp_qdio_sbal_get(qdio))
1497 		goto out;
1498 
1499 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_PORT_DATA,
1500 				  SBAL_SFLAGS0_TYPE_READ,
1501 				  qdio->adapter->pool.erp_req);
1502 
1503 	if (IS_ERR(req)) {
1504 		retval = PTR_ERR(req);
1505 		goto out;
1506 	}
1507 
1508 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1509 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1510 
1511 	req->handler = zfcp_fsf_exchange_port_data_handler;
1512 	req->erp_action = erp_action;
1513 	erp_action->fsf_req_id = req->req_id;
1514 
1515 	zfcp_fsf_start_erp_timer(req);
1516 	retval = zfcp_fsf_req_send(req);
1517 	if (retval) {
1518 		zfcp_fsf_req_free(req);
1519 		erp_action->fsf_req_id = 0;
1520 	}
1521 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1522 out:
1523 	spin_unlock_irq(&qdio->req_q_lock);
1524 	return retval;
1525 }
1526 
1527 /**
1528  * zfcp_fsf_exchange_port_data_sync() - Request information about local port.
1529  * @qdio: pointer to the QDIO-Queue to use for sending the command.
1530  * @data: pointer to the QTCB-Bottom for storing the result of the command,
1531  *	  might be %NULL.
1532  *
1533  * Returns:
1534  * * 0		- Exchange Port Data was successful, @data is complete
1535  * * -EIO	- Exchange Port Data was not successful, @data is invalid
1536  * * -EAGAIN	- @data contains incomplete data
1537  * * -ENOMEM	- Some memory allocation failed along the way
1538  * * -EOPNOTSUPP	- This operation is not supported
1539  */
zfcp_fsf_exchange_port_data_sync(struct zfcp_qdio * qdio,struct fsf_qtcb_bottom_port * data)1540 int zfcp_fsf_exchange_port_data_sync(struct zfcp_qdio *qdio,
1541 				     struct fsf_qtcb_bottom_port *data)
1542 {
1543 	struct zfcp_fsf_req *req = NULL;
1544 	int retval = -EIO;
1545 
1546 	if (!(qdio->adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT))
1547 		return -EOPNOTSUPP;
1548 
1549 	spin_lock_irq(&qdio->req_q_lock);
1550 	if (zfcp_qdio_sbal_get(qdio))
1551 		goto out_unlock;
1552 
1553 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_PORT_DATA,
1554 				  SBAL_SFLAGS0_TYPE_READ, NULL);
1555 
1556 	if (IS_ERR(req)) {
1557 		retval = PTR_ERR(req);
1558 		goto out_unlock;
1559 	}
1560 
1561 	if (data)
1562 		req->data = data;
1563 
1564 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1565 
1566 	req->handler = zfcp_fsf_exchange_port_data_handler;
1567 	zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT);
1568 	retval = zfcp_fsf_req_send(req);
1569 	spin_unlock_irq(&qdio->req_q_lock);
1570 
1571 	if (!retval) {
1572 		/* NOTE: ONLY TOUCH SYNC req AGAIN ON req->completion. */
1573 		wait_for_completion(&req->completion);
1574 
1575 		if (req->status &
1576 		    (ZFCP_STATUS_FSFREQ_ERROR | ZFCP_STATUS_FSFREQ_DISMISSED))
1577 			retval = -EIO;
1578 		else if (req->status & ZFCP_STATUS_FSFREQ_XDATAINCOMPLETE)
1579 			retval = -EAGAIN;
1580 	}
1581 
1582 	zfcp_fsf_req_free(req);
1583 	return retval;
1584 
1585 out_unlock:
1586 	spin_unlock_irq(&qdio->req_q_lock);
1587 	return retval;
1588 }
1589 
zfcp_fsf_log_port_fc_security(struct zfcp_port * port,struct zfcp_fsf_req * req)1590 static void zfcp_fsf_log_port_fc_security(struct zfcp_port *port,
1591 					  struct zfcp_fsf_req *req)
1592 {
1593 	char mnemonic_old[ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH];
1594 	char mnemonic_new[ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH];
1595 
1596 	if (port->connection_info == port->connection_info_old) {
1597 		/* no change, no log nor trace */
1598 		return;
1599 	}
1600 
1601 	zfcp_dbf_hba_fsf_fces("fsfcesp", req, port->wwpn,
1602 			      port->connection_info_old,
1603 			      port->connection_info);
1604 
1605 	zfcp_fsf_scnprint_fc_security(mnemonic_old, sizeof(mnemonic_old),
1606 				      port->connection_info_old,
1607 				      ZFCP_FSF_PRINT_FMT_SINGLEITEM);
1608 	zfcp_fsf_scnprint_fc_security(mnemonic_new, sizeof(mnemonic_new),
1609 				      port->connection_info,
1610 				      ZFCP_FSF_PRINT_FMT_SINGLEITEM);
1611 
1612 	if (strncmp(mnemonic_old, mnemonic_new,
1613 		    ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH) == 0) {
1614 		/* no change in string representation, no log */
1615 		goto out;
1616 	}
1617 
1618 	if (port->connection_info_old == 0) {
1619 		/* activation */
1620 		dev_info(&port->adapter->ccw_device->dev,
1621 			 "FC Endpoint Security of connection to remote port 0x%16llx enabled: %s\n",
1622 			 port->wwpn, mnemonic_new);
1623 	} else if (port->connection_info == 0) {
1624 		/* deactivation */
1625 		dev_warn(&port->adapter->ccw_device->dev,
1626 			 "FC Endpoint Security of connection to remote port 0x%16llx disabled: was %s\n",
1627 			 port->wwpn, mnemonic_old);
1628 	} else {
1629 		/* change */
1630 		dev_warn(&port->adapter->ccw_device->dev,
1631 			 "FC Endpoint Security of connection to remote port 0x%16llx changed: from %s to %s\n",
1632 			 port->wwpn, mnemonic_old, mnemonic_new);
1633 	}
1634 
1635 out:
1636 	port->connection_info_old = port->connection_info;
1637 }
1638 
zfcp_fsf_log_security_error(const struct device * dev,u32 fsf_sqw0,u64 wwpn)1639 static void zfcp_fsf_log_security_error(const struct device *dev, u32 fsf_sqw0,
1640 					u64 wwpn)
1641 {
1642 	switch (fsf_sqw0) {
1643 
1644 	/*
1645 	 * Open Port command error codes
1646 	 */
1647 
1648 	case FSF_SQ_SECURITY_REQUIRED:
1649 		dev_warn_ratelimited(dev,
1650 				     "FC Endpoint Security error: FC security is required but not supported or configured on remote port 0x%016llx\n",
1651 				     wwpn);
1652 		break;
1653 	case FSF_SQ_SECURITY_TIMEOUT:
1654 		dev_warn_ratelimited(dev,
1655 				     "FC Endpoint Security error: a timeout prevented opening remote port 0x%016llx\n",
1656 				     wwpn);
1657 		break;
1658 	case FSF_SQ_SECURITY_KM_UNAVAILABLE:
1659 		dev_warn_ratelimited(dev,
1660 				     "FC Endpoint Security error: opening remote port 0x%016llx failed because local and external key manager cannot communicate\n",
1661 				     wwpn);
1662 		break;
1663 	case FSF_SQ_SECURITY_RKM_UNAVAILABLE:
1664 		dev_warn_ratelimited(dev,
1665 				     "FC Endpoint Security error: opening remote port 0x%016llx failed because it cannot communicate with the external key manager\n",
1666 				     wwpn);
1667 		break;
1668 	case FSF_SQ_SECURITY_AUTH_FAILURE:
1669 		dev_warn_ratelimited(dev,
1670 				     "FC Endpoint Security error: the device could not verify the identity of remote port 0x%016llx\n",
1671 				     wwpn);
1672 		break;
1673 
1674 	/*
1675 	 * Send FCP command error codes
1676 	 */
1677 
1678 	case FSF_SQ_SECURITY_ENC_FAILURE:
1679 		dev_warn_ratelimited(dev,
1680 				     "FC Endpoint Security error: FC connection to remote port 0x%016llx closed because encryption broke down\n",
1681 				     wwpn);
1682 		break;
1683 
1684 	/*
1685 	 * Unknown error codes
1686 	 */
1687 
1688 	default:
1689 		dev_warn_ratelimited(dev,
1690 				     "FC Endpoint Security error: the device issued an unknown error code 0x%08x related to the FC connection to remote port 0x%016llx\n",
1691 				     fsf_sqw0, wwpn);
1692 	}
1693 }
1694 
zfcp_fsf_open_port_handler(struct zfcp_fsf_req * req)1695 static void zfcp_fsf_open_port_handler(struct zfcp_fsf_req *req)
1696 {
1697 	struct zfcp_adapter *adapter = req->adapter;
1698 	struct zfcp_port *port = req->data;
1699 	struct fsf_qtcb_header *header = &req->qtcb->header;
1700 	struct fsf_qtcb_bottom_support *bottom = &req->qtcb->bottom.support;
1701 	struct fc_els_flogi *plogi;
1702 
1703 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
1704 		goto out;
1705 
1706 	switch (header->fsf_status) {
1707 	case FSF_PORT_ALREADY_OPEN:
1708 		break;
1709 	case FSF_MAXIMUM_NUMBER_OF_PORTS_EXCEEDED:
1710 		dev_warn(&adapter->ccw_device->dev,
1711 			 "Not enough FCP adapter resources to open "
1712 			 "remote port 0x%016Lx\n",
1713 			 (unsigned long long)port->wwpn);
1714 		zfcp_erp_set_port_status(port,
1715 					 ZFCP_STATUS_COMMON_ERP_FAILED);
1716 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1717 		break;
1718 	case FSF_SECURITY_ERROR:
1719 		zfcp_fsf_log_security_error(&req->adapter->ccw_device->dev,
1720 					    header->fsf_status_qual.word[0],
1721 					    port->wwpn);
1722 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1723 		break;
1724 	case FSF_ADAPTER_STATUS_AVAILABLE:
1725 		switch (header->fsf_status_qual.word[0]) {
1726 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
1727 			/* no zfcp_fc_test_link() with failed open port */
1728 			fallthrough;
1729 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
1730 		case FSF_SQ_NO_RETRY_POSSIBLE:
1731 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1732 			break;
1733 		}
1734 		break;
1735 	case FSF_GOOD:
1736 		port->handle = header->port_handle;
1737 		if (adapter->adapter_features & FSF_FEATURE_FC_SECURITY)
1738 			port->connection_info = bottom->connection_info;
1739 		else
1740 			port->connection_info = 0;
1741 		zfcp_fsf_log_port_fc_security(port, req);
1742 		atomic_or(ZFCP_STATUS_COMMON_OPEN |
1743 				ZFCP_STATUS_PORT_PHYS_OPEN, &port->status);
1744 		atomic_andnot(ZFCP_STATUS_COMMON_ACCESS_BOXED,
1745 		                  &port->status);
1746 		/* check whether D_ID has changed during open */
1747 		/*
1748 		 * FIXME: This check is not airtight, as the FCP channel does
1749 		 * not monitor closures of target port connections caused on
1750 		 * the remote side. Thus, they might miss out on invalidating
1751 		 * locally cached WWPNs (and other N_Port parameters) of gone
1752 		 * target ports. So, our heroic attempt to make things safe
1753 		 * could be undermined by 'open port' response data tagged with
1754 		 * obsolete WWPNs. Another reason to monitor potential
1755 		 * connection closures ourself at least (by interpreting
1756 		 * incoming ELS' and unsolicited status). It just crosses my
1757 		 * mind that one should be able to cross-check by means of
1758 		 * another GID_PN straight after a port has been opened.
1759 		 * Alternately, an ADISC/PDISC ELS should suffice, as well.
1760 		 */
1761 		plogi = (struct fc_els_flogi *) bottom->els;
1762 		if (bottom->els1_length >= FSF_PLOGI_MIN_LEN)
1763 			zfcp_fc_plogi_evaluate(port, plogi);
1764 		break;
1765 	case FSF_UNKNOWN_OP_SUBTYPE:
1766 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1767 		break;
1768 	}
1769 
1770 out:
1771 	put_device(&port->dev);
1772 }
1773 
1774 /**
1775  * zfcp_fsf_open_port - create and send open port request
1776  * @erp_action: pointer to struct zfcp_erp_action
1777  * Returns: 0 on success, error otherwise
1778  */
zfcp_fsf_open_port(struct zfcp_erp_action * erp_action)1779 int zfcp_fsf_open_port(struct zfcp_erp_action *erp_action)
1780 {
1781 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
1782 	struct zfcp_port *port = erp_action->port;
1783 	struct zfcp_fsf_req *req;
1784 	int retval = -EIO;
1785 
1786 	spin_lock_irq(&qdio->req_q_lock);
1787 	if (zfcp_qdio_sbal_get(qdio))
1788 		goto out;
1789 
1790 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_PORT_WITH_DID,
1791 				  SBAL_SFLAGS0_TYPE_READ,
1792 				  qdio->adapter->pool.erp_req);
1793 
1794 	if (IS_ERR(req)) {
1795 		retval = PTR_ERR(req);
1796 		goto out;
1797 	}
1798 
1799 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1800 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1801 
1802 	req->handler = zfcp_fsf_open_port_handler;
1803 	hton24(req->qtcb->bottom.support.d_id, port->d_id);
1804 	req->data = port;
1805 	req->erp_action = erp_action;
1806 	erp_action->fsf_req_id = req->req_id;
1807 	get_device(&port->dev);
1808 
1809 	zfcp_fsf_start_erp_timer(req);
1810 	retval = zfcp_fsf_req_send(req);
1811 	if (retval) {
1812 		zfcp_fsf_req_free(req);
1813 		erp_action->fsf_req_id = 0;
1814 		put_device(&port->dev);
1815 	}
1816 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1817 out:
1818 	spin_unlock_irq(&qdio->req_q_lock);
1819 	return retval;
1820 }
1821 
zfcp_fsf_close_port_handler(struct zfcp_fsf_req * req)1822 static void zfcp_fsf_close_port_handler(struct zfcp_fsf_req *req)
1823 {
1824 	struct zfcp_port *port = req->data;
1825 
1826 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
1827 		return;
1828 
1829 	switch (req->qtcb->header.fsf_status) {
1830 	case FSF_PORT_HANDLE_NOT_VALID:
1831 		zfcp_erp_adapter_reopen(port->adapter, 0, "fscph_1");
1832 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1833 		break;
1834 	case FSF_ADAPTER_STATUS_AVAILABLE:
1835 		break;
1836 	case FSF_GOOD:
1837 		zfcp_erp_clear_port_status(port, ZFCP_STATUS_COMMON_OPEN);
1838 		break;
1839 	}
1840 }
1841 
1842 /**
1843  * zfcp_fsf_close_port - create and send close port request
1844  * @erp_action: pointer to struct zfcp_erp_action
1845  * Returns: 0 on success, error otherwise
1846  */
zfcp_fsf_close_port(struct zfcp_erp_action * erp_action)1847 int zfcp_fsf_close_port(struct zfcp_erp_action *erp_action)
1848 {
1849 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
1850 	struct zfcp_fsf_req *req;
1851 	int retval = -EIO;
1852 
1853 	spin_lock_irq(&qdio->req_q_lock);
1854 	if (zfcp_qdio_sbal_get(qdio))
1855 		goto out;
1856 
1857 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PORT,
1858 				  SBAL_SFLAGS0_TYPE_READ,
1859 				  qdio->adapter->pool.erp_req);
1860 
1861 	if (IS_ERR(req)) {
1862 		retval = PTR_ERR(req);
1863 		goto out;
1864 	}
1865 
1866 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1867 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1868 
1869 	req->handler = zfcp_fsf_close_port_handler;
1870 	req->data = erp_action->port;
1871 	req->erp_action = erp_action;
1872 	req->qtcb->header.port_handle = erp_action->port->handle;
1873 	erp_action->fsf_req_id = req->req_id;
1874 
1875 	zfcp_fsf_start_erp_timer(req);
1876 	retval = zfcp_fsf_req_send(req);
1877 	if (retval) {
1878 		zfcp_fsf_req_free(req);
1879 		erp_action->fsf_req_id = 0;
1880 	}
1881 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1882 out:
1883 	spin_unlock_irq(&qdio->req_q_lock);
1884 	return retval;
1885 }
1886 
zfcp_fsf_open_wka_port_handler(struct zfcp_fsf_req * req)1887 static void zfcp_fsf_open_wka_port_handler(struct zfcp_fsf_req *req)
1888 {
1889 	struct zfcp_fc_wka_port *wka_port = req->data;
1890 	struct fsf_qtcb_header *header = &req->qtcb->header;
1891 
1892 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR) {
1893 		wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE;
1894 		goto out;
1895 	}
1896 
1897 	switch (header->fsf_status) {
1898 	case FSF_MAXIMUM_NUMBER_OF_PORTS_EXCEEDED:
1899 		dev_warn(&req->adapter->ccw_device->dev,
1900 			 "Opening WKA port 0x%x failed\n", wka_port->d_id);
1901 		fallthrough;
1902 	case FSF_ADAPTER_STATUS_AVAILABLE:
1903 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1904 		wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE;
1905 		break;
1906 	case FSF_GOOD:
1907 		wka_port->handle = header->port_handle;
1908 		fallthrough;
1909 	case FSF_PORT_ALREADY_OPEN:
1910 		wka_port->status = ZFCP_FC_WKA_PORT_ONLINE;
1911 	}
1912 out:
1913 	wake_up(&wka_port->opened);
1914 }
1915 
1916 /**
1917  * zfcp_fsf_open_wka_port - create and send open wka-port request
1918  * @wka_port: pointer to struct zfcp_fc_wka_port
1919  * Returns: 0 on success, error otherwise
1920  */
zfcp_fsf_open_wka_port(struct zfcp_fc_wka_port * wka_port)1921 int zfcp_fsf_open_wka_port(struct zfcp_fc_wka_port *wka_port)
1922 {
1923 	struct zfcp_qdio *qdio = wka_port->adapter->qdio;
1924 	struct zfcp_fsf_req *req;
1925 	u64 req_id = 0;
1926 	int retval = -EIO;
1927 
1928 	spin_lock_irq(&qdio->req_q_lock);
1929 	if (zfcp_qdio_sbal_get(qdio))
1930 		goto out;
1931 
1932 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_PORT_WITH_DID,
1933 				  SBAL_SFLAGS0_TYPE_READ,
1934 				  qdio->adapter->pool.erp_req);
1935 
1936 	if (IS_ERR(req)) {
1937 		retval = PTR_ERR(req);
1938 		goto out;
1939 	}
1940 
1941 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1942 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1943 
1944 	req->handler = zfcp_fsf_open_wka_port_handler;
1945 	hton24(req->qtcb->bottom.support.d_id, wka_port->d_id);
1946 	req->data = wka_port;
1947 
1948 	req_id = req->req_id;
1949 
1950 	zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT);
1951 	retval = zfcp_fsf_req_send(req);
1952 	if (retval)
1953 		zfcp_fsf_req_free(req);
1954 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1955 out:
1956 	spin_unlock_irq(&qdio->req_q_lock);
1957 	if (!retval)
1958 		zfcp_dbf_rec_run_wka("fsowp_1", wka_port, req_id);
1959 	return retval;
1960 }
1961 
zfcp_fsf_close_wka_port_handler(struct zfcp_fsf_req * req)1962 static void zfcp_fsf_close_wka_port_handler(struct zfcp_fsf_req *req)
1963 {
1964 	struct zfcp_fc_wka_port *wka_port = req->data;
1965 
1966 	if (req->qtcb->header.fsf_status == FSF_PORT_HANDLE_NOT_VALID) {
1967 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1968 		zfcp_erp_adapter_reopen(wka_port->adapter, 0, "fscwph1");
1969 	}
1970 
1971 	wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE;
1972 	wake_up(&wka_port->closed);
1973 }
1974 
1975 /**
1976  * zfcp_fsf_close_wka_port - create and send close wka port request
1977  * @wka_port: WKA port to open
1978  * Returns: 0 on success, error otherwise
1979  */
zfcp_fsf_close_wka_port(struct zfcp_fc_wka_port * wka_port)1980 int zfcp_fsf_close_wka_port(struct zfcp_fc_wka_port *wka_port)
1981 {
1982 	struct zfcp_qdio *qdio = wka_port->adapter->qdio;
1983 	struct zfcp_fsf_req *req;
1984 	u64 req_id = 0;
1985 	int retval = -EIO;
1986 
1987 	spin_lock_irq(&qdio->req_q_lock);
1988 	if (zfcp_qdio_sbal_get(qdio))
1989 		goto out;
1990 
1991 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PORT,
1992 				  SBAL_SFLAGS0_TYPE_READ,
1993 				  qdio->adapter->pool.erp_req);
1994 
1995 	if (IS_ERR(req)) {
1996 		retval = PTR_ERR(req);
1997 		goto out;
1998 	}
1999 
2000 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2001 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2002 
2003 	req->handler = zfcp_fsf_close_wka_port_handler;
2004 	req->data = wka_port;
2005 	req->qtcb->header.port_handle = wka_port->handle;
2006 
2007 	req_id = req->req_id;
2008 
2009 	zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT);
2010 	retval = zfcp_fsf_req_send(req);
2011 	if (retval)
2012 		zfcp_fsf_req_free(req);
2013 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2014 out:
2015 	spin_unlock_irq(&qdio->req_q_lock);
2016 	if (!retval)
2017 		zfcp_dbf_rec_run_wka("fscwp_1", wka_port, req_id);
2018 	return retval;
2019 }
2020 
zfcp_fsf_close_physical_port_handler(struct zfcp_fsf_req * req)2021 static void zfcp_fsf_close_physical_port_handler(struct zfcp_fsf_req *req)
2022 {
2023 	struct zfcp_port *port = req->data;
2024 	struct fsf_qtcb_header *header = &req->qtcb->header;
2025 	struct scsi_device *sdev;
2026 
2027 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
2028 		return;
2029 
2030 	switch (header->fsf_status) {
2031 	case FSF_PORT_HANDLE_NOT_VALID:
2032 		zfcp_erp_adapter_reopen(port->adapter, 0, "fscpph1");
2033 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2034 		break;
2035 	case FSF_PORT_BOXED:
2036 		/* can't use generic zfcp_erp_modify_port_status because
2037 		 * ZFCP_STATUS_COMMON_OPEN must not be reset for the port */
2038 		atomic_andnot(ZFCP_STATUS_PORT_PHYS_OPEN, &port->status);
2039 		shost_for_each_device(sdev, port->adapter->scsi_host)
2040 			if (sdev_to_zfcp(sdev)->port == port)
2041 				atomic_andnot(ZFCP_STATUS_COMMON_OPEN,
2042 						  &sdev_to_zfcp(sdev)->status);
2043 		zfcp_erp_set_port_status(port, ZFCP_STATUS_COMMON_ACCESS_BOXED);
2044 		zfcp_erp_port_reopen(port, ZFCP_STATUS_COMMON_ERP_FAILED,
2045 				     "fscpph2");
2046 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2047 		break;
2048 	case FSF_ADAPTER_STATUS_AVAILABLE:
2049 		switch (header->fsf_status_qual.word[0]) {
2050 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
2051 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
2052 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2053 			break;
2054 		}
2055 		break;
2056 	case FSF_GOOD:
2057 		/* can't use generic zfcp_erp_modify_port_status because
2058 		 * ZFCP_STATUS_COMMON_OPEN must not be reset for the port
2059 		 */
2060 		atomic_andnot(ZFCP_STATUS_PORT_PHYS_OPEN, &port->status);
2061 		shost_for_each_device(sdev, port->adapter->scsi_host)
2062 			if (sdev_to_zfcp(sdev)->port == port)
2063 				atomic_andnot(ZFCP_STATUS_COMMON_OPEN,
2064 						  &sdev_to_zfcp(sdev)->status);
2065 		break;
2066 	}
2067 }
2068 
2069 /**
2070  * zfcp_fsf_close_physical_port - close physical port
2071  * @erp_action: pointer to struct zfcp_erp_action
2072  * Returns: 0 on success
2073  */
zfcp_fsf_close_physical_port(struct zfcp_erp_action * erp_action)2074 int zfcp_fsf_close_physical_port(struct zfcp_erp_action *erp_action)
2075 {
2076 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
2077 	struct zfcp_fsf_req *req;
2078 	int retval = -EIO;
2079 
2080 	spin_lock_irq(&qdio->req_q_lock);
2081 	if (zfcp_qdio_sbal_get(qdio))
2082 		goto out;
2083 
2084 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PHYSICAL_PORT,
2085 				  SBAL_SFLAGS0_TYPE_READ,
2086 				  qdio->adapter->pool.erp_req);
2087 
2088 	if (IS_ERR(req)) {
2089 		retval = PTR_ERR(req);
2090 		goto out;
2091 	}
2092 
2093 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2094 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2095 
2096 	req->data = erp_action->port;
2097 	req->qtcb->header.port_handle = erp_action->port->handle;
2098 	req->erp_action = erp_action;
2099 	req->handler = zfcp_fsf_close_physical_port_handler;
2100 	erp_action->fsf_req_id = req->req_id;
2101 
2102 	zfcp_fsf_start_erp_timer(req);
2103 	retval = zfcp_fsf_req_send(req);
2104 	if (retval) {
2105 		zfcp_fsf_req_free(req);
2106 		erp_action->fsf_req_id = 0;
2107 	}
2108 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2109 out:
2110 	spin_unlock_irq(&qdio->req_q_lock);
2111 	return retval;
2112 }
2113 
zfcp_fsf_open_lun_handler(struct zfcp_fsf_req * req)2114 static void zfcp_fsf_open_lun_handler(struct zfcp_fsf_req *req)
2115 {
2116 	struct zfcp_adapter *adapter = req->adapter;
2117 	struct scsi_device *sdev = req->data;
2118 	struct zfcp_scsi_dev *zfcp_sdev;
2119 	struct fsf_qtcb_header *header = &req->qtcb->header;
2120 	union fsf_status_qual *qual = &header->fsf_status_qual;
2121 
2122 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
2123 		return;
2124 
2125 	zfcp_sdev = sdev_to_zfcp(sdev);
2126 
2127 	atomic_andnot(ZFCP_STATUS_COMMON_ACCESS_DENIED |
2128 			  ZFCP_STATUS_COMMON_ACCESS_BOXED,
2129 			  &zfcp_sdev->status);
2130 
2131 	switch (header->fsf_status) {
2132 
2133 	case FSF_PORT_HANDLE_NOT_VALID:
2134 		zfcp_erp_adapter_reopen(adapter, 0, "fsouh_1");
2135 		fallthrough;
2136 	case FSF_LUN_ALREADY_OPEN:
2137 		break;
2138 	case FSF_PORT_BOXED:
2139 		zfcp_erp_set_port_status(zfcp_sdev->port,
2140 					 ZFCP_STATUS_COMMON_ACCESS_BOXED);
2141 		zfcp_erp_port_reopen(zfcp_sdev->port,
2142 				     ZFCP_STATUS_COMMON_ERP_FAILED, "fsouh_2");
2143 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2144 		break;
2145 	case FSF_LUN_SHARING_VIOLATION:
2146 		if (qual->word[0])
2147 			dev_warn(&zfcp_sdev->port->adapter->ccw_device->dev,
2148 				 "LUN 0x%016Lx on port 0x%016Lx is already in "
2149 				 "use by CSS%d, MIF Image ID %x\n",
2150 				 zfcp_scsi_dev_lun(sdev),
2151 				 (unsigned long long)zfcp_sdev->port->wwpn,
2152 				 qual->fsf_queue_designator.cssid,
2153 				 qual->fsf_queue_designator.hla);
2154 		zfcp_erp_set_lun_status(sdev,
2155 					ZFCP_STATUS_COMMON_ERP_FAILED |
2156 					ZFCP_STATUS_COMMON_ACCESS_DENIED);
2157 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2158 		break;
2159 	case FSF_MAXIMUM_NUMBER_OF_LUNS_EXCEEDED:
2160 		dev_warn(&adapter->ccw_device->dev,
2161 			 "No handle is available for LUN "
2162 			 "0x%016Lx on port 0x%016Lx\n",
2163 			 (unsigned long long)zfcp_scsi_dev_lun(sdev),
2164 			 (unsigned long long)zfcp_sdev->port->wwpn);
2165 		zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ERP_FAILED);
2166 		fallthrough;
2167 	case FSF_INVALID_COMMAND_OPTION:
2168 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2169 		break;
2170 	case FSF_ADAPTER_STATUS_AVAILABLE:
2171 		switch (header->fsf_status_qual.word[0]) {
2172 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
2173 			zfcp_fc_test_link(zfcp_sdev->port);
2174 			fallthrough;
2175 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
2176 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2177 			break;
2178 		}
2179 		break;
2180 
2181 	case FSF_GOOD:
2182 		zfcp_sdev->lun_handle = header->lun_handle;
2183 		atomic_or(ZFCP_STATUS_COMMON_OPEN, &zfcp_sdev->status);
2184 		break;
2185 	}
2186 }
2187 
2188 /**
2189  * zfcp_fsf_open_lun - open LUN
2190  * @erp_action: pointer to struct zfcp_erp_action
2191  * Returns: 0 on success, error otherwise
2192  */
zfcp_fsf_open_lun(struct zfcp_erp_action * erp_action)2193 int zfcp_fsf_open_lun(struct zfcp_erp_action *erp_action)
2194 {
2195 	struct zfcp_adapter *adapter = erp_action->adapter;
2196 	struct zfcp_qdio *qdio = adapter->qdio;
2197 	struct zfcp_fsf_req *req;
2198 	int retval = -EIO;
2199 
2200 	spin_lock_irq(&qdio->req_q_lock);
2201 	if (zfcp_qdio_sbal_get(qdio))
2202 		goto out;
2203 
2204 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_LUN,
2205 				  SBAL_SFLAGS0_TYPE_READ,
2206 				  adapter->pool.erp_req);
2207 
2208 	if (IS_ERR(req)) {
2209 		retval = PTR_ERR(req);
2210 		goto out;
2211 	}
2212 
2213 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2214 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2215 
2216 	req->qtcb->header.port_handle = erp_action->port->handle;
2217 	req->qtcb->bottom.support.fcp_lun = zfcp_scsi_dev_lun(erp_action->sdev);
2218 	req->handler = zfcp_fsf_open_lun_handler;
2219 	req->data = erp_action->sdev;
2220 	req->erp_action = erp_action;
2221 	erp_action->fsf_req_id = req->req_id;
2222 
2223 	if (!(adapter->connection_features & FSF_FEATURE_NPIV_MODE))
2224 		req->qtcb->bottom.support.option = FSF_OPEN_LUN_SUPPRESS_BOXING;
2225 
2226 	zfcp_fsf_start_erp_timer(req);
2227 	retval = zfcp_fsf_req_send(req);
2228 	if (retval) {
2229 		zfcp_fsf_req_free(req);
2230 		erp_action->fsf_req_id = 0;
2231 	}
2232 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2233 out:
2234 	spin_unlock_irq(&qdio->req_q_lock);
2235 	return retval;
2236 }
2237 
zfcp_fsf_close_lun_handler(struct zfcp_fsf_req * req)2238 static void zfcp_fsf_close_lun_handler(struct zfcp_fsf_req *req)
2239 {
2240 	struct scsi_device *sdev = req->data;
2241 	struct zfcp_scsi_dev *zfcp_sdev;
2242 
2243 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
2244 		return;
2245 
2246 	zfcp_sdev = sdev_to_zfcp(sdev);
2247 
2248 	switch (req->qtcb->header.fsf_status) {
2249 	case FSF_PORT_HANDLE_NOT_VALID:
2250 		zfcp_erp_adapter_reopen(zfcp_sdev->port->adapter, 0, "fscuh_1");
2251 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2252 		break;
2253 	case FSF_LUN_HANDLE_NOT_VALID:
2254 		zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fscuh_2");
2255 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2256 		break;
2257 	case FSF_PORT_BOXED:
2258 		zfcp_erp_set_port_status(zfcp_sdev->port,
2259 					 ZFCP_STATUS_COMMON_ACCESS_BOXED);
2260 		zfcp_erp_port_reopen(zfcp_sdev->port,
2261 				     ZFCP_STATUS_COMMON_ERP_FAILED, "fscuh_3");
2262 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2263 		break;
2264 	case FSF_ADAPTER_STATUS_AVAILABLE:
2265 		switch (req->qtcb->header.fsf_status_qual.word[0]) {
2266 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
2267 			zfcp_fc_test_link(zfcp_sdev->port);
2268 			fallthrough;
2269 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
2270 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2271 			break;
2272 		}
2273 		break;
2274 	case FSF_GOOD:
2275 		atomic_andnot(ZFCP_STATUS_COMMON_OPEN, &zfcp_sdev->status);
2276 		break;
2277 	}
2278 }
2279 
2280 /**
2281  * zfcp_fsf_close_lun - close LUN
2282  * @erp_action: pointer to erp_action triggering the "close LUN"
2283  * Returns: 0 on success, error otherwise
2284  */
zfcp_fsf_close_lun(struct zfcp_erp_action * erp_action)2285 int zfcp_fsf_close_lun(struct zfcp_erp_action *erp_action)
2286 {
2287 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
2288 	struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(erp_action->sdev);
2289 	struct zfcp_fsf_req *req;
2290 	int retval = -EIO;
2291 
2292 	spin_lock_irq(&qdio->req_q_lock);
2293 	if (zfcp_qdio_sbal_get(qdio))
2294 		goto out;
2295 
2296 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_LUN,
2297 				  SBAL_SFLAGS0_TYPE_READ,
2298 				  qdio->adapter->pool.erp_req);
2299 
2300 	if (IS_ERR(req)) {
2301 		retval = PTR_ERR(req);
2302 		goto out;
2303 	}
2304 
2305 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2306 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2307 
2308 	req->qtcb->header.port_handle = erp_action->port->handle;
2309 	req->qtcb->header.lun_handle = zfcp_sdev->lun_handle;
2310 	req->handler = zfcp_fsf_close_lun_handler;
2311 	req->data = erp_action->sdev;
2312 	req->erp_action = erp_action;
2313 	erp_action->fsf_req_id = req->req_id;
2314 
2315 	zfcp_fsf_start_erp_timer(req);
2316 	retval = zfcp_fsf_req_send(req);
2317 	if (retval) {
2318 		zfcp_fsf_req_free(req);
2319 		erp_action->fsf_req_id = 0;
2320 	}
2321 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2322 out:
2323 	spin_unlock_irq(&qdio->req_q_lock);
2324 	return retval;
2325 }
2326 
zfcp_fsf_update_lat(struct zfcp_latency_record * lat_rec,u32 lat)2327 static void zfcp_fsf_update_lat(struct zfcp_latency_record *lat_rec, u32 lat)
2328 {
2329 	lat_rec->sum += lat;
2330 	lat_rec->min = min(lat_rec->min, lat);
2331 	lat_rec->max = max(lat_rec->max, lat);
2332 }
2333 
zfcp_fsf_req_trace(struct zfcp_fsf_req * req,struct scsi_cmnd * scsi)2334 static void zfcp_fsf_req_trace(struct zfcp_fsf_req *req, struct scsi_cmnd *scsi)
2335 {
2336 	struct fsf_qual_latency_info *lat_in;
2337 	struct zfcp_latency_cont *lat = NULL;
2338 	struct zfcp_scsi_dev *zfcp_sdev;
2339 	struct zfcp_blk_drv_data blktrc;
2340 	int ticks = req->adapter->timer_ticks;
2341 
2342 	lat_in = &req->qtcb->prefix.prot_status_qual.latency_info;
2343 
2344 	blktrc.flags = 0;
2345 	blktrc.magic = ZFCP_BLK_DRV_DATA_MAGIC;
2346 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
2347 		blktrc.flags |= ZFCP_BLK_REQ_ERROR;
2348 	blktrc.inb_usage = 0;
2349 	blktrc.outb_usage = req->qdio_req.qdio_outb_usage;
2350 
2351 	if (req->adapter->adapter_features & FSF_FEATURE_MEASUREMENT_DATA &&
2352 	    !(req->status & ZFCP_STATUS_FSFREQ_ERROR)) {
2353 		zfcp_sdev = sdev_to_zfcp(scsi->device);
2354 		blktrc.flags |= ZFCP_BLK_LAT_VALID;
2355 		blktrc.channel_lat = lat_in->channel_lat * ticks;
2356 		blktrc.fabric_lat = lat_in->fabric_lat * ticks;
2357 
2358 		switch (req->qtcb->bottom.io.data_direction) {
2359 		case FSF_DATADIR_DIF_READ_STRIP:
2360 		case FSF_DATADIR_DIF_READ_CONVERT:
2361 		case FSF_DATADIR_READ:
2362 			lat = &zfcp_sdev->latencies.read;
2363 			break;
2364 		case FSF_DATADIR_DIF_WRITE_INSERT:
2365 		case FSF_DATADIR_DIF_WRITE_CONVERT:
2366 		case FSF_DATADIR_WRITE:
2367 			lat = &zfcp_sdev->latencies.write;
2368 			break;
2369 		case FSF_DATADIR_CMND:
2370 			lat = &zfcp_sdev->latencies.cmd;
2371 			break;
2372 		}
2373 
2374 		if (lat) {
2375 			spin_lock(&zfcp_sdev->latencies.lock);
2376 			zfcp_fsf_update_lat(&lat->channel, lat_in->channel_lat);
2377 			zfcp_fsf_update_lat(&lat->fabric, lat_in->fabric_lat);
2378 			lat->counter++;
2379 			spin_unlock(&zfcp_sdev->latencies.lock);
2380 		}
2381 	}
2382 
2383 	blk_add_driver_data(scsi_cmd_to_rq(scsi), &blktrc, sizeof(blktrc));
2384 }
2385 
2386 /**
2387  * zfcp_fsf_fcp_handler_common() - FCP response handler common to I/O and TMF.
2388  * @req: Pointer to FSF request.
2389  * @sdev: Pointer to SCSI device as request context.
2390  */
zfcp_fsf_fcp_handler_common(struct zfcp_fsf_req * req,struct scsi_device * sdev)2391 static void zfcp_fsf_fcp_handler_common(struct zfcp_fsf_req *req,
2392 					struct scsi_device *sdev)
2393 {
2394 	struct zfcp_scsi_dev *zfcp_sdev;
2395 	struct fsf_qtcb_header *header = &req->qtcb->header;
2396 
2397 	if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR))
2398 		return;
2399 
2400 	zfcp_sdev = sdev_to_zfcp(sdev);
2401 
2402 	switch (header->fsf_status) {
2403 	case FSF_HANDLE_MISMATCH:
2404 	case FSF_PORT_HANDLE_NOT_VALID:
2405 		zfcp_erp_adapter_reopen(req->adapter, 0, "fssfch1");
2406 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2407 		break;
2408 	case FSF_FCPLUN_NOT_VALID:
2409 	case FSF_LUN_HANDLE_NOT_VALID:
2410 		zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fssfch2");
2411 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2412 		break;
2413 	case FSF_SERVICE_CLASS_NOT_SUPPORTED:
2414 		zfcp_fsf_class_not_supp(req);
2415 		break;
2416 	case FSF_DIRECTION_INDICATOR_NOT_VALID:
2417 		dev_err(&req->adapter->ccw_device->dev,
2418 			"Incorrect direction %d, LUN 0x%016Lx on port "
2419 			"0x%016Lx closed\n",
2420 			req->qtcb->bottom.io.data_direction,
2421 			(unsigned long long)zfcp_scsi_dev_lun(sdev),
2422 			(unsigned long long)zfcp_sdev->port->wwpn);
2423 		zfcp_erp_adapter_shutdown(req->adapter, 0, "fssfch3");
2424 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2425 		break;
2426 	case FSF_CMND_LENGTH_NOT_VALID:
2427 		dev_err(&req->adapter->ccw_device->dev,
2428 			"Incorrect FCP_CMND length %d, FCP device closed\n",
2429 			req->qtcb->bottom.io.fcp_cmnd_length);
2430 		zfcp_erp_adapter_shutdown(req->adapter, 0, "fssfch4");
2431 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2432 		break;
2433 	case FSF_PORT_BOXED:
2434 		zfcp_erp_set_port_status(zfcp_sdev->port,
2435 					 ZFCP_STATUS_COMMON_ACCESS_BOXED);
2436 		zfcp_erp_port_reopen(zfcp_sdev->port,
2437 				     ZFCP_STATUS_COMMON_ERP_FAILED, "fssfch5");
2438 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2439 		break;
2440 	case FSF_LUN_BOXED:
2441 		zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ACCESS_BOXED);
2442 		zfcp_erp_lun_reopen(sdev, ZFCP_STATUS_COMMON_ERP_FAILED,
2443 				    "fssfch6");
2444 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2445 		break;
2446 	case FSF_ADAPTER_STATUS_AVAILABLE:
2447 		if (header->fsf_status_qual.word[0] ==
2448 		    FSF_SQ_INVOKE_LINK_TEST_PROCEDURE)
2449 			zfcp_fc_test_link(zfcp_sdev->port);
2450 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2451 		break;
2452 	case FSF_SECURITY_ERROR:
2453 		zfcp_fsf_log_security_error(&req->adapter->ccw_device->dev,
2454 					    header->fsf_status_qual.word[0],
2455 					    zfcp_sdev->port->wwpn);
2456 		zfcp_erp_port_forced_reopen(zfcp_sdev->port, 0, "fssfch7");
2457 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2458 		break;
2459 	}
2460 }
2461 
zfcp_fsf_fcp_cmnd_handler(struct zfcp_fsf_req * req)2462 static void zfcp_fsf_fcp_cmnd_handler(struct zfcp_fsf_req *req)
2463 {
2464 	struct scsi_cmnd *scpnt;
2465 	struct fcp_resp_with_ext *fcp_rsp;
2466 	unsigned long flags;
2467 
2468 	read_lock_irqsave(&req->adapter->abort_lock, flags);
2469 
2470 	scpnt = req->data;
2471 	if (unlikely(!scpnt)) {
2472 		read_unlock_irqrestore(&req->adapter->abort_lock, flags);
2473 		return;
2474 	}
2475 
2476 	zfcp_fsf_fcp_handler_common(req, scpnt->device);
2477 
2478 	if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR)) {
2479 		set_host_byte(scpnt, DID_TRANSPORT_DISRUPTED);
2480 		goto skip_fsfstatus;
2481 	}
2482 
2483 	switch (req->qtcb->header.fsf_status) {
2484 	case FSF_INCONSISTENT_PROT_DATA:
2485 	case FSF_INVALID_PROT_PARM:
2486 		set_host_byte(scpnt, DID_ERROR);
2487 		goto skip_fsfstatus;
2488 	case FSF_BLOCK_GUARD_CHECK_FAILURE:
2489 		zfcp_scsi_dif_sense_error(scpnt, 0x1);
2490 		goto skip_fsfstatus;
2491 	case FSF_APP_TAG_CHECK_FAILURE:
2492 		zfcp_scsi_dif_sense_error(scpnt, 0x2);
2493 		goto skip_fsfstatus;
2494 	case FSF_REF_TAG_CHECK_FAILURE:
2495 		zfcp_scsi_dif_sense_error(scpnt, 0x3);
2496 		goto skip_fsfstatus;
2497 	}
2498 	BUILD_BUG_ON(sizeof(struct fcp_resp_with_ext) > FSF_FCP_RSP_SIZE);
2499 	fcp_rsp = &req->qtcb->bottom.io.fcp_rsp.iu;
2500 	zfcp_fc_eval_fcp_rsp(fcp_rsp, scpnt);
2501 
2502 skip_fsfstatus:
2503 	zfcp_fsf_req_trace(req, scpnt);
2504 	zfcp_dbf_scsi_result(scpnt, req);
2505 
2506 	scpnt->host_scribble = NULL;
2507 	scsi_done(scpnt);
2508 	/*
2509 	 * We must hold this lock until scsi_done has been called.
2510 	 * Otherwise we may call scsi_done after abort regarding this
2511 	 * command has completed.
2512 	 * Note: scsi_done must not block!
2513 	 */
2514 	read_unlock_irqrestore(&req->adapter->abort_lock, flags);
2515 }
2516 
zfcp_fsf_set_data_dir(struct scsi_cmnd * scsi_cmnd,u32 * data_dir)2517 static int zfcp_fsf_set_data_dir(struct scsi_cmnd *scsi_cmnd, u32 *data_dir)
2518 {
2519 	switch (scsi_get_prot_op(scsi_cmnd)) {
2520 	case SCSI_PROT_NORMAL:
2521 		switch (scsi_cmnd->sc_data_direction) {
2522 		case DMA_NONE:
2523 			*data_dir = FSF_DATADIR_CMND;
2524 			break;
2525 		case DMA_FROM_DEVICE:
2526 			*data_dir = FSF_DATADIR_READ;
2527 			break;
2528 		case DMA_TO_DEVICE:
2529 			*data_dir = FSF_DATADIR_WRITE;
2530 			break;
2531 		case DMA_BIDIRECTIONAL:
2532 			return -EINVAL;
2533 		}
2534 		break;
2535 
2536 	case SCSI_PROT_READ_STRIP:
2537 		*data_dir = FSF_DATADIR_DIF_READ_STRIP;
2538 		break;
2539 	case SCSI_PROT_WRITE_INSERT:
2540 		*data_dir = FSF_DATADIR_DIF_WRITE_INSERT;
2541 		break;
2542 	case SCSI_PROT_READ_PASS:
2543 		*data_dir = FSF_DATADIR_DIF_READ_CONVERT;
2544 		break;
2545 	case SCSI_PROT_WRITE_PASS:
2546 		*data_dir = FSF_DATADIR_DIF_WRITE_CONVERT;
2547 		break;
2548 	default:
2549 		return -EINVAL;
2550 	}
2551 
2552 	return 0;
2553 }
2554 
2555 /**
2556  * zfcp_fsf_fcp_cmnd - initiate an FCP command (for a SCSI command)
2557  * @scsi_cmnd: scsi command to be sent
2558  */
zfcp_fsf_fcp_cmnd(struct scsi_cmnd * scsi_cmnd)2559 int zfcp_fsf_fcp_cmnd(struct scsi_cmnd *scsi_cmnd)
2560 {
2561 	struct zfcp_fsf_req *req;
2562 	struct fcp_cmnd *fcp_cmnd;
2563 	u8 sbtype = SBAL_SFLAGS0_TYPE_READ;
2564 	int retval = -EIO;
2565 	struct scsi_device *sdev = scsi_cmnd->device;
2566 	struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev);
2567 	struct zfcp_adapter *adapter = zfcp_sdev->port->adapter;
2568 	struct zfcp_qdio *qdio = adapter->qdio;
2569 	struct fsf_qtcb_bottom_io *io;
2570 	unsigned long flags;
2571 
2572 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
2573 		       ZFCP_STATUS_COMMON_UNBLOCKED)))
2574 		return -EBUSY;
2575 
2576 	spin_lock_irqsave(&qdio->req_q_lock, flags);
2577 	if (atomic_read(&qdio->req_q_free) <= 0) {
2578 		atomic_inc(&qdio->req_q_full);
2579 		goto out;
2580 	}
2581 
2582 	if (scsi_cmnd->sc_data_direction == DMA_TO_DEVICE)
2583 		sbtype = SBAL_SFLAGS0_TYPE_WRITE;
2584 
2585 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_FCP_CMND,
2586 				  sbtype, adapter->pool.scsi_req);
2587 
2588 	if (IS_ERR(req)) {
2589 		retval = PTR_ERR(req);
2590 		goto out;
2591 	}
2592 
2593 	BUILD_BUG_ON(sizeof(scsi_cmnd->host_scribble) < sizeof(req->req_id));
2594 	scsi_cmnd->host_scribble = (unsigned char *) req->req_id;
2595 
2596 	io = &req->qtcb->bottom.io;
2597 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2598 	req->data = scsi_cmnd;
2599 	req->handler = zfcp_fsf_fcp_cmnd_handler;
2600 	req->qtcb->header.lun_handle = zfcp_sdev->lun_handle;
2601 	req->qtcb->header.port_handle = zfcp_sdev->port->handle;
2602 	io->service_class = FSF_CLASS_3;
2603 	io->fcp_cmnd_length = FCP_CMND_LEN;
2604 
2605 	if (scsi_get_prot_op(scsi_cmnd) != SCSI_PROT_NORMAL) {
2606 		io->data_block_length = scsi_prot_interval(scsi_cmnd);
2607 		io->ref_tag_value = scsi_prot_ref_tag(scsi_cmnd);
2608 	}
2609 
2610 	if (zfcp_fsf_set_data_dir(scsi_cmnd, &io->data_direction))
2611 		goto failed_scsi_cmnd;
2612 
2613 	BUILD_BUG_ON(sizeof(struct fcp_cmnd) > FSF_FCP_CMND_SIZE);
2614 	fcp_cmnd = &req->qtcb->bottom.io.fcp_cmnd.iu;
2615 	zfcp_fc_scsi_to_fcp(fcp_cmnd, scsi_cmnd);
2616 
2617 	if ((scsi_get_prot_op(scsi_cmnd) != SCSI_PROT_NORMAL) &&
2618 	    scsi_prot_sg_count(scsi_cmnd)) {
2619 		zfcp_qdio_set_data_div(qdio, &req->qdio_req,
2620 				       scsi_prot_sg_count(scsi_cmnd));
2621 		retval = zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req,
2622 						 scsi_prot_sglist(scsi_cmnd));
2623 		if (retval)
2624 			goto failed_scsi_cmnd;
2625 		io->prot_data_length = zfcp_qdio_real_bytes(
2626 						scsi_prot_sglist(scsi_cmnd));
2627 	}
2628 
2629 	retval = zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req,
2630 					 scsi_sglist(scsi_cmnd));
2631 	if (unlikely(retval))
2632 		goto failed_scsi_cmnd;
2633 
2634 	zfcp_qdio_set_sbale_last(adapter->qdio, &req->qdio_req);
2635 	if (zfcp_adapter_multi_buffer_active(adapter))
2636 		zfcp_qdio_set_scount(qdio, &req->qdio_req);
2637 
2638 	retval = zfcp_fsf_req_send(req);
2639 	if (unlikely(retval))
2640 		goto failed_scsi_cmnd;
2641 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2642 
2643 	goto out;
2644 
2645 failed_scsi_cmnd:
2646 	zfcp_fsf_req_free(req);
2647 	scsi_cmnd->host_scribble = NULL;
2648 out:
2649 	spin_unlock_irqrestore(&qdio->req_q_lock, flags);
2650 	return retval;
2651 }
2652 
zfcp_fsf_fcp_task_mgmt_handler(struct zfcp_fsf_req * req)2653 static void zfcp_fsf_fcp_task_mgmt_handler(struct zfcp_fsf_req *req)
2654 {
2655 	struct scsi_device *sdev = req->data;
2656 	struct fcp_resp_with_ext *fcp_rsp;
2657 	struct fcp_resp_rsp_info *rsp_info;
2658 
2659 	zfcp_fsf_fcp_handler_common(req, sdev);
2660 
2661 	fcp_rsp = &req->qtcb->bottom.io.fcp_rsp.iu;
2662 	rsp_info = (struct fcp_resp_rsp_info *) &fcp_rsp[1];
2663 
2664 	if ((rsp_info->rsp_code != FCP_TMF_CMPL) ||
2665 	     (req->status & ZFCP_STATUS_FSFREQ_ERROR))
2666 		req->status |= ZFCP_STATUS_FSFREQ_TMFUNCFAILED;
2667 }
2668 
2669 /**
2670  * zfcp_fsf_fcp_task_mgmt() - Send SCSI task management command (TMF).
2671  * @sdev: Pointer to SCSI device to send the task management command to.
2672  * @tm_flags: Unsigned byte for task management flags.
2673  *
2674  * Return: On success pointer to struct zfcp_fsf_req, %NULL otherwise.
2675  */
zfcp_fsf_fcp_task_mgmt(struct scsi_device * sdev,u8 tm_flags)2676 struct zfcp_fsf_req *zfcp_fsf_fcp_task_mgmt(struct scsi_device *sdev,
2677 					    u8 tm_flags)
2678 {
2679 	struct zfcp_fsf_req *req = NULL;
2680 	struct fcp_cmnd *fcp_cmnd;
2681 	struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev);
2682 	struct zfcp_qdio *qdio = zfcp_sdev->port->adapter->qdio;
2683 
2684 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
2685 		       ZFCP_STATUS_COMMON_UNBLOCKED)))
2686 		return NULL;
2687 
2688 	spin_lock_irq(&qdio->req_q_lock);
2689 	if (zfcp_qdio_sbal_get(qdio))
2690 		goto out;
2691 
2692 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_FCP_CMND,
2693 				  SBAL_SFLAGS0_TYPE_WRITE,
2694 				  qdio->adapter->pool.scsi_req);
2695 
2696 	if (IS_ERR(req)) {
2697 		req = NULL;
2698 		goto out;
2699 	}
2700 
2701 	req->data = sdev;
2702 
2703 	req->handler = zfcp_fsf_fcp_task_mgmt_handler;
2704 	req->qtcb->header.lun_handle = zfcp_sdev->lun_handle;
2705 	req->qtcb->header.port_handle = zfcp_sdev->port->handle;
2706 	req->qtcb->bottom.io.data_direction = FSF_DATADIR_CMND;
2707 	req->qtcb->bottom.io.service_class = FSF_CLASS_3;
2708 	req->qtcb->bottom.io.fcp_cmnd_length = FCP_CMND_LEN;
2709 
2710 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2711 
2712 	fcp_cmnd = &req->qtcb->bottom.io.fcp_cmnd.iu;
2713 	zfcp_fc_fcp_tm(fcp_cmnd, sdev, tm_flags);
2714 
2715 	zfcp_fsf_start_timer(req, ZFCP_FSF_SCSI_ER_TIMEOUT);
2716 	if (!zfcp_fsf_req_send(req)) {
2717 		/* NOTE: DO NOT TOUCH req, UNTIL IT COMPLETES! */
2718 		goto out;
2719 	}
2720 
2721 	zfcp_fsf_req_free(req);
2722 	req = NULL;
2723 out:
2724 	spin_unlock_irq(&qdio->req_q_lock);
2725 	return req;
2726 }
2727 
2728 /**
2729  * zfcp_fsf_reqid_check - validate req_id contained in SBAL returned by QDIO
2730  * @qdio: pointer to struct zfcp_qdio
2731  * @sbal_idx: response queue index of SBAL to be processed
2732  */
zfcp_fsf_reqid_check(struct zfcp_qdio * qdio,int sbal_idx)2733 void zfcp_fsf_reqid_check(struct zfcp_qdio *qdio, int sbal_idx)
2734 {
2735 	struct zfcp_adapter *adapter = qdio->adapter;
2736 	struct qdio_buffer *sbal = qdio->res_q[sbal_idx];
2737 	struct qdio_buffer_element *sbale;
2738 	struct zfcp_fsf_req *fsf_req;
2739 	u64 req_id;
2740 	int idx;
2741 
2742 	for (idx = 0; idx < QDIO_MAX_ELEMENTS_PER_BUFFER; idx++) {
2743 
2744 		sbale = &sbal->element[idx];
2745 		req_id = sbale->addr;
2746 		fsf_req = zfcp_reqlist_find_rm(adapter->req_list, req_id);
2747 
2748 		if (!fsf_req) {
2749 			/*
2750 			 * Unknown request means that we have potentially memory
2751 			 * corruption and must stop the machine immediately.
2752 			 */
2753 			zfcp_qdio_siosl(adapter);
2754 			panic("error: unknown req_id (%llx) on adapter %s.\n",
2755 			      req_id, dev_name(&adapter->ccw_device->dev));
2756 		}
2757 
2758 		zfcp_fsf_req_complete(fsf_req);
2759 
2760 		if (likely(sbale->eflags & SBAL_EFLAGS_LAST_ENTRY))
2761 			break;
2762 	}
2763 }
2764