1 /*
2 * Scsi Host Layer for MPT (Message Passing Technology) based controllers
3 *
4 * This code is based on drivers/scsi/mpt3sas/mpt3sas_scsih.c
5 * Copyright (C) 2012-2014 LSI Corporation
6 * Copyright (C) 2013-2014 Avago Technologies
7 * (mailto: MPT-FusionLinux.pdl@avagotech.com)
8 *
9 * This program is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU General Public License
11 * as published by the Free Software Foundation; either version 2
12 * of the License, or (at your option) any later version.
13 *
14 * This program is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 * GNU General Public License for more details.
18 *
19 * NO WARRANTY
20 * THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR
21 * CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT
22 * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
23 * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is
24 * solely responsible for determining the appropriateness of using and
25 * distributing the Program and assumes all risks associated with its
26 * exercise of rights under this Agreement, including but not limited to
27 * the risks and costs of program errors, damage to or loss of data,
28 * programs or equipment, and unavailability or interruption of operations.
29
30 * DISCLAIMER OF LIABILITY
31 * NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY
32 * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33 * DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND
34 * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
35 * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
36 * USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED
37 * HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES
38
39 * You should have received a copy of the GNU General Public License
40 * along with this program; if not, write to the Free Software
41 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301,
42 * USA.
43 */
44
45 #include <linux/module.h>
46 #include <linux/kernel.h>
47 #include <linux/init.h>
48 #include <linux/errno.h>
49 #include <linux/blkdev.h>
50 #include <linux/sched.h>
51 #include <linux/workqueue.h>
52 #include <linux/delay.h>
53 #include <linux/pci.h>
54 #include <linux/interrupt.h>
55 #include <linux/aer.h>
56 #include <linux/raid_class.h>
57 #include <asm/unaligned.h>
58
59 #include "mpt3sas_base.h"
60
61 #define RAID_CHANNEL 1
62
63 #define PCIE_CHANNEL 2
64
65 /* forward proto's */
66 static void _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc,
67 struct _sas_node *sas_expander);
68 static void _firmware_event_work(struct work_struct *work);
69
70 static void _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc,
71 struct _sas_device *sas_device);
72 static int _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle,
73 u8 retry_count, u8 is_pd);
74 static int _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle);
75 static void _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc,
76 struct _pcie_device *pcie_device);
77 static void
78 _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle);
79 static u8 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid);
80
81 /* global parameters */
82 LIST_HEAD(mpt3sas_ioc_list);
83 /* global ioc lock for list operations */
84 DEFINE_SPINLOCK(gioc_lock);
85
86 MODULE_AUTHOR(MPT3SAS_AUTHOR);
87 MODULE_DESCRIPTION(MPT3SAS_DESCRIPTION);
88 MODULE_LICENSE("GPL");
89 MODULE_VERSION(MPT3SAS_DRIVER_VERSION);
90 MODULE_ALIAS("mpt2sas");
91
92 /* local parameters */
93 static u8 scsi_io_cb_idx = -1;
94 static u8 tm_cb_idx = -1;
95 static u8 ctl_cb_idx = -1;
96 static u8 base_cb_idx = -1;
97 static u8 port_enable_cb_idx = -1;
98 static u8 transport_cb_idx = -1;
99 static u8 scsih_cb_idx = -1;
100 static u8 config_cb_idx = -1;
101 static int mpt2_ids;
102 static int mpt3_ids;
103
104 static u8 tm_tr_cb_idx = -1 ;
105 static u8 tm_tr_volume_cb_idx = -1 ;
106 static u8 tm_sas_control_cb_idx = -1;
107
108 /* command line options */
109 static u32 logging_level;
110 MODULE_PARM_DESC(logging_level,
111 " bits for enabling additional logging info (default=0)");
112
113
114 static ushort max_sectors = 0xFFFF;
115 module_param(max_sectors, ushort, 0444);
116 MODULE_PARM_DESC(max_sectors, "max sectors, range 64 to 32767 default=32767");
117
118
119 static int missing_delay[2] = {-1, -1};
120 module_param_array(missing_delay, int, NULL, 0444);
121 MODULE_PARM_DESC(missing_delay, " device missing delay , io missing delay");
122
123 /* scsi-mid layer global parmeter is max_report_luns, which is 511 */
124 #define MPT3SAS_MAX_LUN (16895)
125 static u64 max_lun = MPT3SAS_MAX_LUN;
126 module_param(max_lun, ullong, 0444);
127 MODULE_PARM_DESC(max_lun, " max lun, default=16895 ");
128
129 static ushort hbas_to_enumerate;
130 module_param(hbas_to_enumerate, ushort, 0444);
131 MODULE_PARM_DESC(hbas_to_enumerate,
132 " 0 - enumerates both SAS 2.0 & SAS 3.0 generation HBAs\n \
133 1 - enumerates only SAS 2.0 generation HBAs\n \
134 2 - enumerates only SAS 3.0 generation HBAs (default=0)");
135
136 /* diag_buffer_enable is bitwise
137 * bit 0 set = TRACE
138 * bit 1 set = SNAPSHOT
139 * bit 2 set = EXTENDED
140 *
141 * Either bit can be set, or both
142 */
143 static int diag_buffer_enable = -1;
144 module_param(diag_buffer_enable, int, 0444);
145 MODULE_PARM_DESC(diag_buffer_enable,
146 " post diag buffers (TRACE=1/SNAPSHOT=2/EXTENDED=4/default=0)");
147 static int disable_discovery = -1;
148 module_param(disable_discovery, int, 0444);
149 MODULE_PARM_DESC(disable_discovery, " disable discovery ");
150
151
152 /* permit overriding the host protection capabilities mask (EEDP/T10 PI) */
153 static int prot_mask = -1;
154 module_param(prot_mask, int, 0444);
155 MODULE_PARM_DESC(prot_mask, " host protection capabilities mask, def=7 ");
156
157 static bool enable_sdev_max_qd;
158 module_param(enable_sdev_max_qd, bool, 0444);
159 MODULE_PARM_DESC(enable_sdev_max_qd,
160 "Enable sdev max qd as can_queue, def=disabled(0)");
161
162 /* raid transport support */
163 static struct raid_template *mpt3sas_raid_template;
164 static struct raid_template *mpt2sas_raid_template;
165
166
167 /**
168 * struct sense_info - common structure for obtaining sense keys
169 * @skey: sense key
170 * @asc: additional sense code
171 * @ascq: additional sense code qualifier
172 */
173 struct sense_info {
174 u8 skey;
175 u8 asc;
176 u8 ascq;
177 };
178
179 #define MPT3SAS_PROCESS_TRIGGER_DIAG (0xFFFB)
180 #define MPT3SAS_TURN_ON_PFA_LED (0xFFFC)
181 #define MPT3SAS_PORT_ENABLE_COMPLETE (0xFFFD)
182 #define MPT3SAS_ABRT_TASK_SET (0xFFFE)
183 #define MPT3SAS_REMOVE_UNRESPONDING_DEVICES (0xFFFF)
184 /**
185 * struct fw_event_work - firmware event struct
186 * @list: link list framework
187 * @work: work object (ioc->fault_reset_work_q)
188 * @ioc: per adapter object
189 * @device_handle: device handle
190 * @VF_ID: virtual function id
191 * @VP_ID: virtual port id
192 * @ignore: flag meaning this event has been marked to ignore
193 * @event: firmware event MPI2_EVENT_XXX defined in mpi2_ioc.h
194 * @refcount: kref for this event
195 * @event_data: reply event data payload follows
196 *
197 * This object stored on ioc->fw_event_list.
198 */
199 struct fw_event_work {
200 struct list_head list;
201 struct work_struct work;
202
203 struct MPT3SAS_ADAPTER *ioc;
204 u16 device_handle;
205 u8 VF_ID;
206 u8 VP_ID;
207 u8 ignore;
208 u16 event;
209 struct kref refcount;
210 char event_data[0] __aligned(4);
211 };
212
fw_event_work_free(struct kref * r)213 static void fw_event_work_free(struct kref *r)
214 {
215 kfree(container_of(r, struct fw_event_work, refcount));
216 }
217
fw_event_work_get(struct fw_event_work * fw_work)218 static void fw_event_work_get(struct fw_event_work *fw_work)
219 {
220 kref_get(&fw_work->refcount);
221 }
222
fw_event_work_put(struct fw_event_work * fw_work)223 static void fw_event_work_put(struct fw_event_work *fw_work)
224 {
225 kref_put(&fw_work->refcount, fw_event_work_free);
226 }
227
alloc_fw_event_work(int len)228 static struct fw_event_work *alloc_fw_event_work(int len)
229 {
230 struct fw_event_work *fw_event;
231
232 fw_event = kzalloc(sizeof(*fw_event) + len, GFP_ATOMIC);
233 if (!fw_event)
234 return NULL;
235
236 kref_init(&fw_event->refcount);
237 return fw_event;
238 }
239
240 /**
241 * struct _scsi_io_transfer - scsi io transfer
242 * @handle: sas device handle (assigned by firmware)
243 * @is_raid: flag set for hidden raid components
244 * @dir: DMA_TO_DEVICE, DMA_FROM_DEVICE,
245 * @data_length: data transfer length
246 * @data_dma: dma pointer to data
247 * @sense: sense data
248 * @lun: lun number
249 * @cdb_length: cdb length
250 * @cdb: cdb contents
251 * @timeout: timeout for this command
252 * @VF_ID: virtual function id
253 * @VP_ID: virtual port id
254 * @valid_reply: flag set for reply message
255 * @sense_length: sense length
256 * @ioc_status: ioc status
257 * @scsi_state: scsi state
258 * @scsi_status: scsi staus
259 * @log_info: log information
260 * @transfer_length: data length transfer when there is a reply message
261 *
262 * Used for sending internal scsi commands to devices within this module.
263 * Refer to _scsi_send_scsi_io().
264 */
265 struct _scsi_io_transfer {
266 u16 handle;
267 u8 is_raid;
268 enum dma_data_direction dir;
269 u32 data_length;
270 dma_addr_t data_dma;
271 u8 sense[SCSI_SENSE_BUFFERSIZE];
272 u32 lun;
273 u8 cdb_length;
274 u8 cdb[32];
275 u8 timeout;
276 u8 VF_ID;
277 u8 VP_ID;
278 u8 valid_reply;
279 /* the following bits are only valid when 'valid_reply = 1' */
280 u32 sense_length;
281 u16 ioc_status;
282 u8 scsi_state;
283 u8 scsi_status;
284 u32 log_info;
285 u32 transfer_length;
286 };
287
288 /**
289 * _scsih_set_debug_level - global setting of ioc->logging_level.
290 * @val: ?
291 * @kp: ?
292 *
293 * Note: The logging levels are defined in mpt3sas_debug.h.
294 */
295 static int
_scsih_set_debug_level(const char * val,const struct kernel_param * kp)296 _scsih_set_debug_level(const char *val, const struct kernel_param *kp)
297 {
298 int ret = param_set_int(val, kp);
299 struct MPT3SAS_ADAPTER *ioc;
300
301 if (ret)
302 return ret;
303
304 pr_info("setting logging_level(0x%08x)\n", logging_level);
305 spin_lock(&gioc_lock);
306 list_for_each_entry(ioc, &mpt3sas_ioc_list, list)
307 ioc->logging_level = logging_level;
308 spin_unlock(&gioc_lock);
309 return 0;
310 }
311 module_param_call(logging_level, _scsih_set_debug_level, param_get_int,
312 &logging_level, 0644);
313
314 /**
315 * _scsih_srch_boot_sas_address - search based on sas_address
316 * @sas_address: sas address
317 * @boot_device: boot device object from bios page 2
318 *
319 * Return: 1 when there's a match, 0 means no match.
320 */
321 static inline int
_scsih_srch_boot_sas_address(u64 sas_address,Mpi2BootDeviceSasWwid_t * boot_device)322 _scsih_srch_boot_sas_address(u64 sas_address,
323 Mpi2BootDeviceSasWwid_t *boot_device)
324 {
325 return (sas_address == le64_to_cpu(boot_device->SASAddress)) ? 1 : 0;
326 }
327
328 /**
329 * _scsih_srch_boot_device_name - search based on device name
330 * @device_name: device name specified in INDENTIFY fram
331 * @boot_device: boot device object from bios page 2
332 *
333 * Return: 1 when there's a match, 0 means no match.
334 */
335 static inline int
_scsih_srch_boot_device_name(u64 device_name,Mpi2BootDeviceDeviceName_t * boot_device)336 _scsih_srch_boot_device_name(u64 device_name,
337 Mpi2BootDeviceDeviceName_t *boot_device)
338 {
339 return (device_name == le64_to_cpu(boot_device->DeviceName)) ? 1 : 0;
340 }
341
342 /**
343 * _scsih_srch_boot_encl_slot - search based on enclosure_logical_id/slot
344 * @enclosure_logical_id: enclosure logical id
345 * @slot_number: slot number
346 * @boot_device: boot device object from bios page 2
347 *
348 * Return: 1 when there's a match, 0 means no match.
349 */
350 static inline int
_scsih_srch_boot_encl_slot(u64 enclosure_logical_id,u16 slot_number,Mpi2BootDeviceEnclosureSlot_t * boot_device)351 _scsih_srch_boot_encl_slot(u64 enclosure_logical_id, u16 slot_number,
352 Mpi2BootDeviceEnclosureSlot_t *boot_device)
353 {
354 return (enclosure_logical_id == le64_to_cpu(boot_device->
355 EnclosureLogicalID) && slot_number == le16_to_cpu(boot_device->
356 SlotNumber)) ? 1 : 0;
357 }
358
359 /**
360 * _scsih_is_boot_device - search for matching boot device.
361 * @sas_address: sas address
362 * @device_name: device name specified in INDENTIFY fram
363 * @enclosure_logical_id: enclosure logical id
364 * @slot: slot number
365 * @form: specifies boot device form
366 * @boot_device: boot device object from bios page 2
367 *
368 * Return: 1 when there's a match, 0 means no match.
369 */
370 static int
_scsih_is_boot_device(u64 sas_address,u64 device_name,u64 enclosure_logical_id,u16 slot,u8 form,Mpi2BiosPage2BootDevice_t * boot_device)371 _scsih_is_boot_device(u64 sas_address, u64 device_name,
372 u64 enclosure_logical_id, u16 slot, u8 form,
373 Mpi2BiosPage2BootDevice_t *boot_device)
374 {
375 int rc = 0;
376
377 switch (form) {
378 case MPI2_BIOSPAGE2_FORM_SAS_WWID:
379 if (!sas_address)
380 break;
381 rc = _scsih_srch_boot_sas_address(
382 sas_address, &boot_device->SasWwid);
383 break;
384 case MPI2_BIOSPAGE2_FORM_ENCLOSURE_SLOT:
385 if (!enclosure_logical_id)
386 break;
387 rc = _scsih_srch_boot_encl_slot(
388 enclosure_logical_id,
389 slot, &boot_device->EnclosureSlot);
390 break;
391 case MPI2_BIOSPAGE2_FORM_DEVICE_NAME:
392 if (!device_name)
393 break;
394 rc = _scsih_srch_boot_device_name(
395 device_name, &boot_device->DeviceName);
396 break;
397 case MPI2_BIOSPAGE2_FORM_NO_DEVICE_SPECIFIED:
398 break;
399 }
400
401 return rc;
402 }
403
404 /**
405 * _scsih_get_sas_address - set the sas_address for given device handle
406 * @ioc: ?
407 * @handle: device handle
408 * @sas_address: sas address
409 *
410 * Return: 0 success, non-zero when failure
411 */
412 static int
_scsih_get_sas_address(struct MPT3SAS_ADAPTER * ioc,u16 handle,u64 * sas_address)413 _scsih_get_sas_address(struct MPT3SAS_ADAPTER *ioc, u16 handle,
414 u64 *sas_address)
415 {
416 Mpi2SasDevicePage0_t sas_device_pg0;
417 Mpi2ConfigReply_t mpi_reply;
418 u32 ioc_status;
419
420 *sas_address = 0;
421
422 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
423 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) {
424 ioc_err(ioc, "failure at %s:%d/%s()!\n",
425 __FILE__, __LINE__, __func__);
426 return -ENXIO;
427 }
428
429 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
430 if (ioc_status == MPI2_IOCSTATUS_SUCCESS) {
431 /* For HBA, vSES doesn't return HBA SAS address. Instead return
432 * vSES's sas address.
433 */
434 if ((handle <= ioc->sas_hba.num_phys) &&
435 (!(le32_to_cpu(sas_device_pg0.DeviceInfo) &
436 MPI2_SAS_DEVICE_INFO_SEP)))
437 *sas_address = ioc->sas_hba.sas_address;
438 else
439 *sas_address = le64_to_cpu(sas_device_pg0.SASAddress);
440 return 0;
441 }
442
443 /* we hit this because the given parent handle doesn't exist */
444 if (ioc_status == MPI2_IOCSTATUS_CONFIG_INVALID_PAGE)
445 return -ENXIO;
446
447 /* else error case */
448 ioc_err(ioc, "handle(0x%04x), ioc_status(0x%04x), failure at %s:%d/%s()!\n",
449 handle, ioc_status, __FILE__, __LINE__, __func__);
450 return -EIO;
451 }
452
453 /**
454 * _scsih_determine_boot_device - determine boot device.
455 * @ioc: per adapter object
456 * @device: sas_device or pcie_device object
457 * @channel: SAS or PCIe channel
458 *
459 * Determines whether this device should be first reported device to
460 * to scsi-ml or sas transport, this purpose is for persistent boot device.
461 * There are primary, alternate, and current entries in bios page 2. The order
462 * priority is primary, alternate, then current. This routine saves
463 * the corresponding device object.
464 * The saved data to be used later in _scsih_probe_boot_devices().
465 */
466 static void
_scsih_determine_boot_device(struct MPT3SAS_ADAPTER * ioc,void * device,u32 channel)467 _scsih_determine_boot_device(struct MPT3SAS_ADAPTER *ioc, void *device,
468 u32 channel)
469 {
470 struct _sas_device *sas_device;
471 struct _pcie_device *pcie_device;
472 struct _raid_device *raid_device;
473 u64 sas_address;
474 u64 device_name;
475 u64 enclosure_logical_id;
476 u16 slot;
477
478 /* only process this function when driver loads */
479 if (!ioc->is_driver_loading)
480 return;
481
482 /* no Bios, return immediately */
483 if (!ioc->bios_pg3.BiosVersion)
484 return;
485
486 if (channel == RAID_CHANNEL) {
487 raid_device = device;
488 sas_address = raid_device->wwid;
489 device_name = 0;
490 enclosure_logical_id = 0;
491 slot = 0;
492 } else if (channel == PCIE_CHANNEL) {
493 pcie_device = device;
494 sas_address = pcie_device->wwid;
495 device_name = 0;
496 enclosure_logical_id = 0;
497 slot = 0;
498 } else {
499 sas_device = device;
500 sas_address = sas_device->sas_address;
501 device_name = sas_device->device_name;
502 enclosure_logical_id = sas_device->enclosure_logical_id;
503 slot = sas_device->slot;
504 }
505
506 if (!ioc->req_boot_device.device) {
507 if (_scsih_is_boot_device(sas_address, device_name,
508 enclosure_logical_id, slot,
509 (ioc->bios_pg2.ReqBootDeviceForm &
510 MPI2_BIOSPAGE2_FORM_MASK),
511 &ioc->bios_pg2.RequestedBootDevice)) {
512 dinitprintk(ioc,
513 ioc_info(ioc, "%s: req_boot_device(0x%016llx)\n",
514 __func__, (u64)sas_address));
515 ioc->req_boot_device.device = device;
516 ioc->req_boot_device.channel = channel;
517 }
518 }
519
520 if (!ioc->req_alt_boot_device.device) {
521 if (_scsih_is_boot_device(sas_address, device_name,
522 enclosure_logical_id, slot,
523 (ioc->bios_pg2.ReqAltBootDeviceForm &
524 MPI2_BIOSPAGE2_FORM_MASK),
525 &ioc->bios_pg2.RequestedAltBootDevice)) {
526 dinitprintk(ioc,
527 ioc_info(ioc, "%s: req_alt_boot_device(0x%016llx)\n",
528 __func__, (u64)sas_address));
529 ioc->req_alt_boot_device.device = device;
530 ioc->req_alt_boot_device.channel = channel;
531 }
532 }
533
534 if (!ioc->current_boot_device.device) {
535 if (_scsih_is_boot_device(sas_address, device_name,
536 enclosure_logical_id, slot,
537 (ioc->bios_pg2.CurrentBootDeviceForm &
538 MPI2_BIOSPAGE2_FORM_MASK),
539 &ioc->bios_pg2.CurrentBootDevice)) {
540 dinitprintk(ioc,
541 ioc_info(ioc, "%s: current_boot_device(0x%016llx)\n",
542 __func__, (u64)sas_address));
543 ioc->current_boot_device.device = device;
544 ioc->current_boot_device.channel = channel;
545 }
546 }
547 }
548
549 static struct _sas_device *
__mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER * ioc,struct MPT3SAS_TARGET * tgt_priv)550 __mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc,
551 struct MPT3SAS_TARGET *tgt_priv)
552 {
553 struct _sas_device *ret;
554
555 assert_spin_locked(&ioc->sas_device_lock);
556
557 ret = tgt_priv->sas_dev;
558 if (ret)
559 sas_device_get(ret);
560
561 return ret;
562 }
563
564 static struct _sas_device *
mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER * ioc,struct MPT3SAS_TARGET * tgt_priv)565 mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc,
566 struct MPT3SAS_TARGET *tgt_priv)
567 {
568 struct _sas_device *ret;
569 unsigned long flags;
570
571 spin_lock_irqsave(&ioc->sas_device_lock, flags);
572 ret = __mpt3sas_get_sdev_from_target(ioc, tgt_priv);
573 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
574
575 return ret;
576 }
577
578 static struct _pcie_device *
__mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER * ioc,struct MPT3SAS_TARGET * tgt_priv)579 __mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc,
580 struct MPT3SAS_TARGET *tgt_priv)
581 {
582 struct _pcie_device *ret;
583
584 assert_spin_locked(&ioc->pcie_device_lock);
585
586 ret = tgt_priv->pcie_dev;
587 if (ret)
588 pcie_device_get(ret);
589
590 return ret;
591 }
592
593 /**
594 * mpt3sas_get_pdev_from_target - pcie device search
595 * @ioc: per adapter object
596 * @tgt_priv: starget private object
597 *
598 * Context: This function will acquire ioc->pcie_device_lock and will release
599 * before returning the pcie_device object.
600 *
601 * This searches for pcie_device from target, then return pcie_device object.
602 */
603 static struct _pcie_device *
mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER * ioc,struct MPT3SAS_TARGET * tgt_priv)604 mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc,
605 struct MPT3SAS_TARGET *tgt_priv)
606 {
607 struct _pcie_device *ret;
608 unsigned long flags;
609
610 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
611 ret = __mpt3sas_get_pdev_from_target(ioc, tgt_priv);
612 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
613
614 return ret;
615 }
616
617 struct _sas_device *
__mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER * ioc,u64 sas_address)618 __mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc,
619 u64 sas_address)
620 {
621 struct _sas_device *sas_device;
622
623 assert_spin_locked(&ioc->sas_device_lock);
624
625 list_for_each_entry(sas_device, &ioc->sas_device_list, list)
626 if (sas_device->sas_address == sas_address)
627 goto found_device;
628
629 list_for_each_entry(sas_device, &ioc->sas_device_init_list, list)
630 if (sas_device->sas_address == sas_address)
631 goto found_device;
632
633 return NULL;
634
635 found_device:
636 sas_device_get(sas_device);
637 return sas_device;
638 }
639
640 /**
641 * mpt3sas_get_sdev_by_addr - sas device search
642 * @ioc: per adapter object
643 * @sas_address: sas address
644 * Context: Calling function should acquire ioc->sas_device_lock
645 *
646 * This searches for sas_device based on sas_address, then return sas_device
647 * object.
648 */
649 struct _sas_device *
mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER * ioc,u64 sas_address)650 mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc,
651 u64 sas_address)
652 {
653 struct _sas_device *sas_device;
654 unsigned long flags;
655
656 spin_lock_irqsave(&ioc->sas_device_lock, flags);
657 sas_device = __mpt3sas_get_sdev_by_addr(ioc,
658 sas_address);
659 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
660
661 return sas_device;
662 }
663
664 static struct _sas_device *
__mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)665 __mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
666 {
667 struct _sas_device *sas_device;
668
669 assert_spin_locked(&ioc->sas_device_lock);
670
671 list_for_each_entry(sas_device, &ioc->sas_device_list, list)
672 if (sas_device->handle == handle)
673 goto found_device;
674
675 list_for_each_entry(sas_device, &ioc->sas_device_init_list, list)
676 if (sas_device->handle == handle)
677 goto found_device;
678
679 return NULL;
680
681 found_device:
682 sas_device_get(sas_device);
683 return sas_device;
684 }
685
686 /**
687 * mpt3sas_get_sdev_by_handle - sas device search
688 * @ioc: per adapter object
689 * @handle: sas device handle (assigned by firmware)
690 * Context: Calling function should acquire ioc->sas_device_lock
691 *
692 * This searches for sas_device based on sas_address, then return sas_device
693 * object.
694 */
695 struct _sas_device *
mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)696 mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
697 {
698 struct _sas_device *sas_device;
699 unsigned long flags;
700
701 spin_lock_irqsave(&ioc->sas_device_lock, flags);
702 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
703 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
704
705 return sas_device;
706 }
707
708 /**
709 * _scsih_display_enclosure_chassis_info - display device location info
710 * @ioc: per adapter object
711 * @sas_device: per sas device object
712 * @sdev: scsi device struct
713 * @starget: scsi target struct
714 */
715 static void
_scsih_display_enclosure_chassis_info(struct MPT3SAS_ADAPTER * ioc,struct _sas_device * sas_device,struct scsi_device * sdev,struct scsi_target * starget)716 _scsih_display_enclosure_chassis_info(struct MPT3SAS_ADAPTER *ioc,
717 struct _sas_device *sas_device, struct scsi_device *sdev,
718 struct scsi_target *starget)
719 {
720 if (sdev) {
721 if (sas_device->enclosure_handle != 0)
722 sdev_printk(KERN_INFO, sdev,
723 "enclosure logical id (0x%016llx), slot(%d) \n",
724 (unsigned long long)
725 sas_device->enclosure_logical_id,
726 sas_device->slot);
727 if (sas_device->connector_name[0] != '\0')
728 sdev_printk(KERN_INFO, sdev,
729 "enclosure level(0x%04x), connector name( %s)\n",
730 sas_device->enclosure_level,
731 sas_device->connector_name);
732 if (sas_device->is_chassis_slot_valid)
733 sdev_printk(KERN_INFO, sdev, "chassis slot(0x%04x)\n",
734 sas_device->chassis_slot);
735 } else if (starget) {
736 if (sas_device->enclosure_handle != 0)
737 starget_printk(KERN_INFO, starget,
738 "enclosure logical id(0x%016llx), slot(%d) \n",
739 (unsigned long long)
740 sas_device->enclosure_logical_id,
741 sas_device->slot);
742 if (sas_device->connector_name[0] != '\0')
743 starget_printk(KERN_INFO, starget,
744 "enclosure level(0x%04x), connector name( %s)\n",
745 sas_device->enclosure_level,
746 sas_device->connector_name);
747 if (sas_device->is_chassis_slot_valid)
748 starget_printk(KERN_INFO, starget,
749 "chassis slot(0x%04x)\n",
750 sas_device->chassis_slot);
751 } else {
752 if (sas_device->enclosure_handle != 0)
753 ioc_info(ioc, "enclosure logical id(0x%016llx), slot(%d)\n",
754 (u64)sas_device->enclosure_logical_id,
755 sas_device->slot);
756 if (sas_device->connector_name[0] != '\0')
757 ioc_info(ioc, "enclosure level(0x%04x), connector name( %s)\n",
758 sas_device->enclosure_level,
759 sas_device->connector_name);
760 if (sas_device->is_chassis_slot_valid)
761 ioc_info(ioc, "chassis slot(0x%04x)\n",
762 sas_device->chassis_slot);
763 }
764 }
765
766 /**
767 * _scsih_sas_device_remove - remove sas_device from list.
768 * @ioc: per adapter object
769 * @sas_device: the sas_device object
770 * Context: This function will acquire ioc->sas_device_lock.
771 *
772 * If sas_device is on the list, remove it and decrement its reference count.
773 */
774 static void
_scsih_sas_device_remove(struct MPT3SAS_ADAPTER * ioc,struct _sas_device * sas_device)775 _scsih_sas_device_remove(struct MPT3SAS_ADAPTER *ioc,
776 struct _sas_device *sas_device)
777 {
778 unsigned long flags;
779
780 if (!sas_device)
781 return;
782 ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n",
783 sas_device->handle, (u64)sas_device->sas_address);
784
785 _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL);
786
787 /*
788 * The lock serializes access to the list, but we still need to verify
789 * that nobody removed the entry while we were waiting on the lock.
790 */
791 spin_lock_irqsave(&ioc->sas_device_lock, flags);
792 if (!list_empty(&sas_device->list)) {
793 list_del_init(&sas_device->list);
794 sas_device_put(sas_device);
795 }
796 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
797 }
798
799 /**
800 * _scsih_device_remove_by_handle - removing device object by handle
801 * @ioc: per adapter object
802 * @handle: device handle
803 */
804 static void
_scsih_device_remove_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)805 _scsih_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
806 {
807 struct _sas_device *sas_device;
808 unsigned long flags;
809
810 if (ioc->shost_recovery)
811 return;
812
813 spin_lock_irqsave(&ioc->sas_device_lock, flags);
814 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
815 if (sas_device) {
816 list_del_init(&sas_device->list);
817 sas_device_put(sas_device);
818 }
819 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
820 if (sas_device) {
821 _scsih_remove_device(ioc, sas_device);
822 sas_device_put(sas_device);
823 }
824 }
825
826 /**
827 * mpt3sas_device_remove_by_sas_address - removing device object by sas address
828 * @ioc: per adapter object
829 * @sas_address: device sas_address
830 */
831 void
mpt3sas_device_remove_by_sas_address(struct MPT3SAS_ADAPTER * ioc,u64 sas_address)832 mpt3sas_device_remove_by_sas_address(struct MPT3SAS_ADAPTER *ioc,
833 u64 sas_address)
834 {
835 struct _sas_device *sas_device;
836 unsigned long flags;
837
838 if (ioc->shost_recovery)
839 return;
840
841 spin_lock_irqsave(&ioc->sas_device_lock, flags);
842 sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_address);
843 if (sas_device) {
844 list_del_init(&sas_device->list);
845 sas_device_put(sas_device);
846 }
847 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
848 if (sas_device) {
849 _scsih_remove_device(ioc, sas_device);
850 sas_device_put(sas_device);
851 }
852 }
853
854 /**
855 * _scsih_sas_device_add - insert sas_device to the list.
856 * @ioc: per adapter object
857 * @sas_device: the sas_device object
858 * Context: This function will acquire ioc->sas_device_lock.
859 *
860 * Adding new object to the ioc->sas_device_list.
861 */
862 static void
_scsih_sas_device_add(struct MPT3SAS_ADAPTER * ioc,struct _sas_device * sas_device)863 _scsih_sas_device_add(struct MPT3SAS_ADAPTER *ioc,
864 struct _sas_device *sas_device)
865 {
866 unsigned long flags;
867
868 dewtprintk(ioc,
869 ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n",
870 __func__, sas_device->handle,
871 (u64)sas_device->sas_address));
872
873 dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device,
874 NULL, NULL));
875
876 spin_lock_irqsave(&ioc->sas_device_lock, flags);
877 sas_device_get(sas_device);
878 list_add_tail(&sas_device->list, &ioc->sas_device_list);
879 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
880
881 if (ioc->hide_drives) {
882 clear_bit(sas_device->handle, ioc->pend_os_device_add);
883 return;
884 }
885
886 if (!mpt3sas_transport_port_add(ioc, sas_device->handle,
887 sas_device->sas_address_parent)) {
888 _scsih_sas_device_remove(ioc, sas_device);
889 } else if (!sas_device->starget) {
890 /*
891 * When asyn scanning is enabled, its not possible to remove
892 * devices while scanning is turned on due to an oops in
893 * scsi_sysfs_add_sdev()->add_device()->sysfs_addrm_start()
894 */
895 if (!ioc->is_driver_loading) {
896 mpt3sas_transport_port_remove(ioc,
897 sas_device->sas_address,
898 sas_device->sas_address_parent);
899 _scsih_sas_device_remove(ioc, sas_device);
900 }
901 } else
902 clear_bit(sas_device->handle, ioc->pend_os_device_add);
903 }
904
905 /**
906 * _scsih_sas_device_init_add - insert sas_device to the list.
907 * @ioc: per adapter object
908 * @sas_device: the sas_device object
909 * Context: This function will acquire ioc->sas_device_lock.
910 *
911 * Adding new object at driver load time to the ioc->sas_device_init_list.
912 */
913 static void
_scsih_sas_device_init_add(struct MPT3SAS_ADAPTER * ioc,struct _sas_device * sas_device)914 _scsih_sas_device_init_add(struct MPT3SAS_ADAPTER *ioc,
915 struct _sas_device *sas_device)
916 {
917 unsigned long flags;
918
919 dewtprintk(ioc,
920 ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n",
921 __func__, sas_device->handle,
922 (u64)sas_device->sas_address));
923
924 dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device,
925 NULL, NULL));
926
927 spin_lock_irqsave(&ioc->sas_device_lock, flags);
928 sas_device_get(sas_device);
929 list_add_tail(&sas_device->list, &ioc->sas_device_init_list);
930 _scsih_determine_boot_device(ioc, sas_device, 0);
931 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
932 }
933
934
935 static struct _pcie_device *
__mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER * ioc,u64 wwid)936 __mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
937 {
938 struct _pcie_device *pcie_device;
939
940 assert_spin_locked(&ioc->pcie_device_lock);
941
942 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list)
943 if (pcie_device->wwid == wwid)
944 goto found_device;
945
946 list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list)
947 if (pcie_device->wwid == wwid)
948 goto found_device;
949
950 return NULL;
951
952 found_device:
953 pcie_device_get(pcie_device);
954 return pcie_device;
955 }
956
957
958 /**
959 * mpt3sas_get_pdev_by_wwid - pcie device search
960 * @ioc: per adapter object
961 * @wwid: wwid
962 *
963 * Context: This function will acquire ioc->pcie_device_lock and will release
964 * before returning the pcie_device object.
965 *
966 * This searches for pcie_device based on wwid, then return pcie_device object.
967 */
968 static struct _pcie_device *
mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER * ioc,u64 wwid)969 mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
970 {
971 struct _pcie_device *pcie_device;
972 unsigned long flags;
973
974 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
975 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid);
976 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
977
978 return pcie_device;
979 }
980
981
982 static struct _pcie_device *
__mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER * ioc,int id,int channel)983 __mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER *ioc, int id,
984 int channel)
985 {
986 struct _pcie_device *pcie_device;
987
988 assert_spin_locked(&ioc->pcie_device_lock);
989
990 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list)
991 if (pcie_device->id == id && pcie_device->channel == channel)
992 goto found_device;
993
994 list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list)
995 if (pcie_device->id == id && pcie_device->channel == channel)
996 goto found_device;
997
998 return NULL;
999
1000 found_device:
1001 pcie_device_get(pcie_device);
1002 return pcie_device;
1003 }
1004
1005 static struct _pcie_device *
__mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)1006 __mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1007 {
1008 struct _pcie_device *pcie_device;
1009
1010 assert_spin_locked(&ioc->pcie_device_lock);
1011
1012 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list)
1013 if (pcie_device->handle == handle)
1014 goto found_device;
1015
1016 list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list)
1017 if (pcie_device->handle == handle)
1018 goto found_device;
1019
1020 return NULL;
1021
1022 found_device:
1023 pcie_device_get(pcie_device);
1024 return pcie_device;
1025 }
1026
1027
1028 /**
1029 * mpt3sas_get_pdev_by_handle - pcie device search
1030 * @ioc: per adapter object
1031 * @handle: Firmware device handle
1032 *
1033 * Context: This function will acquire ioc->pcie_device_lock and will release
1034 * before returning the pcie_device object.
1035 *
1036 * This searches for pcie_device based on handle, then return pcie_device
1037 * object.
1038 */
1039 struct _pcie_device *
mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)1040 mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1041 {
1042 struct _pcie_device *pcie_device;
1043 unsigned long flags;
1044
1045 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1046 pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle);
1047 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1048
1049 return pcie_device;
1050 }
1051
1052 /**
1053 * _scsih_pcie_device_remove - remove pcie_device from list.
1054 * @ioc: per adapter object
1055 * @pcie_device: the pcie_device object
1056 * Context: This function will acquire ioc->pcie_device_lock.
1057 *
1058 * If pcie_device is on the list, remove it and decrement its reference count.
1059 */
1060 static void
_scsih_pcie_device_remove(struct MPT3SAS_ADAPTER * ioc,struct _pcie_device * pcie_device)1061 _scsih_pcie_device_remove(struct MPT3SAS_ADAPTER *ioc,
1062 struct _pcie_device *pcie_device)
1063 {
1064 unsigned long flags;
1065 int was_on_pcie_device_list = 0;
1066
1067 if (!pcie_device)
1068 return;
1069 ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n",
1070 pcie_device->handle, (u64)pcie_device->wwid);
1071 if (pcie_device->enclosure_handle != 0)
1072 ioc_info(ioc, "removing enclosure logical id(0x%016llx), slot(%d)\n",
1073 (u64)pcie_device->enclosure_logical_id,
1074 pcie_device->slot);
1075 if (pcie_device->connector_name[0] != '\0')
1076 ioc_info(ioc, "removing enclosure level(0x%04x), connector name( %s)\n",
1077 pcie_device->enclosure_level,
1078 pcie_device->connector_name);
1079
1080 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1081 if (!list_empty(&pcie_device->list)) {
1082 list_del_init(&pcie_device->list);
1083 was_on_pcie_device_list = 1;
1084 }
1085 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1086 if (was_on_pcie_device_list) {
1087 kfree(pcie_device->serial_number);
1088 pcie_device_put(pcie_device);
1089 }
1090 }
1091
1092
1093 /**
1094 * _scsih_pcie_device_remove_by_handle - removing pcie device object by handle
1095 * @ioc: per adapter object
1096 * @handle: device handle
1097 */
1098 static void
_scsih_pcie_device_remove_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)1099 _scsih_pcie_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1100 {
1101 struct _pcie_device *pcie_device;
1102 unsigned long flags;
1103 int was_on_pcie_device_list = 0;
1104
1105 if (ioc->shost_recovery)
1106 return;
1107
1108 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1109 pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle);
1110 if (pcie_device) {
1111 if (!list_empty(&pcie_device->list)) {
1112 list_del_init(&pcie_device->list);
1113 was_on_pcie_device_list = 1;
1114 pcie_device_put(pcie_device);
1115 }
1116 }
1117 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1118 if (was_on_pcie_device_list) {
1119 _scsih_pcie_device_remove_from_sml(ioc, pcie_device);
1120 pcie_device_put(pcie_device);
1121 }
1122 }
1123
1124 /**
1125 * _scsih_pcie_device_add - add pcie_device object
1126 * @ioc: per adapter object
1127 * @pcie_device: pcie_device object
1128 *
1129 * This is added to the pcie_device_list link list.
1130 */
1131 static void
_scsih_pcie_device_add(struct MPT3SAS_ADAPTER * ioc,struct _pcie_device * pcie_device)1132 _scsih_pcie_device_add(struct MPT3SAS_ADAPTER *ioc,
1133 struct _pcie_device *pcie_device)
1134 {
1135 unsigned long flags;
1136
1137 dewtprintk(ioc,
1138 ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n",
1139 __func__,
1140 pcie_device->handle, (u64)pcie_device->wwid));
1141 if (pcie_device->enclosure_handle != 0)
1142 dewtprintk(ioc,
1143 ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n",
1144 __func__,
1145 (u64)pcie_device->enclosure_logical_id,
1146 pcie_device->slot));
1147 if (pcie_device->connector_name[0] != '\0')
1148 dewtprintk(ioc,
1149 ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n",
1150 __func__, pcie_device->enclosure_level,
1151 pcie_device->connector_name));
1152
1153 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1154 pcie_device_get(pcie_device);
1155 list_add_tail(&pcie_device->list, &ioc->pcie_device_list);
1156 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1157
1158 if (pcie_device->access_status ==
1159 MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) {
1160 clear_bit(pcie_device->handle, ioc->pend_os_device_add);
1161 return;
1162 }
1163 if (scsi_add_device(ioc->shost, PCIE_CHANNEL, pcie_device->id, 0)) {
1164 _scsih_pcie_device_remove(ioc, pcie_device);
1165 } else if (!pcie_device->starget) {
1166 if (!ioc->is_driver_loading) {
1167 /*TODO-- Need to find out whether this condition will occur or not*/
1168 clear_bit(pcie_device->handle, ioc->pend_os_device_add);
1169 }
1170 } else
1171 clear_bit(pcie_device->handle, ioc->pend_os_device_add);
1172 }
1173
1174 /*
1175 * _scsih_pcie_device_init_add - insert pcie_device to the init list.
1176 * @ioc: per adapter object
1177 * @pcie_device: the pcie_device object
1178 * Context: This function will acquire ioc->pcie_device_lock.
1179 *
1180 * Adding new object at driver load time to the ioc->pcie_device_init_list.
1181 */
1182 static void
_scsih_pcie_device_init_add(struct MPT3SAS_ADAPTER * ioc,struct _pcie_device * pcie_device)1183 _scsih_pcie_device_init_add(struct MPT3SAS_ADAPTER *ioc,
1184 struct _pcie_device *pcie_device)
1185 {
1186 unsigned long flags;
1187
1188 dewtprintk(ioc,
1189 ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n",
1190 __func__,
1191 pcie_device->handle, (u64)pcie_device->wwid));
1192 if (pcie_device->enclosure_handle != 0)
1193 dewtprintk(ioc,
1194 ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n",
1195 __func__,
1196 (u64)pcie_device->enclosure_logical_id,
1197 pcie_device->slot));
1198 if (pcie_device->connector_name[0] != '\0')
1199 dewtprintk(ioc,
1200 ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n",
1201 __func__, pcie_device->enclosure_level,
1202 pcie_device->connector_name));
1203
1204 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1205 pcie_device_get(pcie_device);
1206 list_add_tail(&pcie_device->list, &ioc->pcie_device_init_list);
1207 if (pcie_device->access_status !=
1208 MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED)
1209 _scsih_determine_boot_device(ioc, pcie_device, PCIE_CHANNEL);
1210 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1211 }
1212 /**
1213 * _scsih_raid_device_find_by_id - raid device search
1214 * @ioc: per adapter object
1215 * @id: sas device target id
1216 * @channel: sas device channel
1217 * Context: Calling function should acquire ioc->raid_device_lock
1218 *
1219 * This searches for raid_device based on target id, then return raid_device
1220 * object.
1221 */
1222 static struct _raid_device *
_scsih_raid_device_find_by_id(struct MPT3SAS_ADAPTER * ioc,int id,int channel)1223 _scsih_raid_device_find_by_id(struct MPT3SAS_ADAPTER *ioc, int id, int channel)
1224 {
1225 struct _raid_device *raid_device, *r;
1226
1227 r = NULL;
1228 list_for_each_entry(raid_device, &ioc->raid_device_list, list) {
1229 if (raid_device->id == id && raid_device->channel == channel) {
1230 r = raid_device;
1231 goto out;
1232 }
1233 }
1234
1235 out:
1236 return r;
1237 }
1238
1239 /**
1240 * mpt3sas_raid_device_find_by_handle - raid device search
1241 * @ioc: per adapter object
1242 * @handle: sas device handle (assigned by firmware)
1243 * Context: Calling function should acquire ioc->raid_device_lock
1244 *
1245 * This searches for raid_device based on handle, then return raid_device
1246 * object.
1247 */
1248 struct _raid_device *
mpt3sas_raid_device_find_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)1249 mpt3sas_raid_device_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1250 {
1251 struct _raid_device *raid_device, *r;
1252
1253 r = NULL;
1254 list_for_each_entry(raid_device, &ioc->raid_device_list, list) {
1255 if (raid_device->handle != handle)
1256 continue;
1257 r = raid_device;
1258 goto out;
1259 }
1260
1261 out:
1262 return r;
1263 }
1264
1265 /**
1266 * _scsih_raid_device_find_by_wwid - raid device search
1267 * @ioc: per adapter object
1268 * @wwid: ?
1269 * Context: Calling function should acquire ioc->raid_device_lock
1270 *
1271 * This searches for raid_device based on wwid, then return raid_device
1272 * object.
1273 */
1274 static struct _raid_device *
_scsih_raid_device_find_by_wwid(struct MPT3SAS_ADAPTER * ioc,u64 wwid)1275 _scsih_raid_device_find_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
1276 {
1277 struct _raid_device *raid_device, *r;
1278
1279 r = NULL;
1280 list_for_each_entry(raid_device, &ioc->raid_device_list, list) {
1281 if (raid_device->wwid != wwid)
1282 continue;
1283 r = raid_device;
1284 goto out;
1285 }
1286
1287 out:
1288 return r;
1289 }
1290
1291 /**
1292 * _scsih_raid_device_add - add raid_device object
1293 * @ioc: per adapter object
1294 * @raid_device: raid_device object
1295 *
1296 * This is added to the raid_device_list link list.
1297 */
1298 static void
_scsih_raid_device_add(struct MPT3SAS_ADAPTER * ioc,struct _raid_device * raid_device)1299 _scsih_raid_device_add(struct MPT3SAS_ADAPTER *ioc,
1300 struct _raid_device *raid_device)
1301 {
1302 unsigned long flags;
1303
1304 dewtprintk(ioc,
1305 ioc_info(ioc, "%s: handle(0x%04x), wwid(0x%016llx)\n",
1306 __func__,
1307 raid_device->handle, (u64)raid_device->wwid));
1308
1309 spin_lock_irqsave(&ioc->raid_device_lock, flags);
1310 list_add_tail(&raid_device->list, &ioc->raid_device_list);
1311 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1312 }
1313
1314 /**
1315 * _scsih_raid_device_remove - delete raid_device object
1316 * @ioc: per adapter object
1317 * @raid_device: raid_device object
1318 *
1319 */
1320 static void
_scsih_raid_device_remove(struct MPT3SAS_ADAPTER * ioc,struct _raid_device * raid_device)1321 _scsih_raid_device_remove(struct MPT3SAS_ADAPTER *ioc,
1322 struct _raid_device *raid_device)
1323 {
1324 unsigned long flags;
1325
1326 spin_lock_irqsave(&ioc->raid_device_lock, flags);
1327 list_del(&raid_device->list);
1328 kfree(raid_device);
1329 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1330 }
1331
1332 /**
1333 * mpt3sas_scsih_expander_find_by_handle - expander device search
1334 * @ioc: per adapter object
1335 * @handle: expander handle (assigned by firmware)
1336 * Context: Calling function should acquire ioc->sas_device_lock
1337 *
1338 * This searches for expander device based on handle, then returns the
1339 * sas_node object.
1340 */
1341 struct _sas_node *
mpt3sas_scsih_expander_find_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)1342 mpt3sas_scsih_expander_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1343 {
1344 struct _sas_node *sas_expander, *r;
1345
1346 r = NULL;
1347 list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) {
1348 if (sas_expander->handle != handle)
1349 continue;
1350 r = sas_expander;
1351 goto out;
1352 }
1353 out:
1354 return r;
1355 }
1356
1357 /**
1358 * mpt3sas_scsih_enclosure_find_by_handle - exclosure device search
1359 * @ioc: per adapter object
1360 * @handle: enclosure handle (assigned by firmware)
1361 * Context: Calling function should acquire ioc->sas_device_lock
1362 *
1363 * This searches for enclosure device based on handle, then returns the
1364 * enclosure object.
1365 */
1366 static struct _enclosure_node *
mpt3sas_scsih_enclosure_find_by_handle(struct MPT3SAS_ADAPTER * ioc,u16 handle)1367 mpt3sas_scsih_enclosure_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1368 {
1369 struct _enclosure_node *enclosure_dev, *r;
1370
1371 r = NULL;
1372 list_for_each_entry(enclosure_dev, &ioc->enclosure_list, list) {
1373 if (le16_to_cpu(enclosure_dev->pg0.EnclosureHandle) != handle)
1374 continue;
1375 r = enclosure_dev;
1376 goto out;
1377 }
1378 out:
1379 return r;
1380 }
1381 /**
1382 * mpt3sas_scsih_expander_find_by_sas_address - expander device search
1383 * @ioc: per adapter object
1384 * @sas_address: sas address
1385 * Context: Calling function should acquire ioc->sas_node_lock.
1386 *
1387 * This searches for expander device based on sas_address, then returns the
1388 * sas_node object.
1389 */
1390 struct _sas_node *
mpt3sas_scsih_expander_find_by_sas_address(struct MPT3SAS_ADAPTER * ioc,u64 sas_address)1391 mpt3sas_scsih_expander_find_by_sas_address(struct MPT3SAS_ADAPTER *ioc,
1392 u64 sas_address)
1393 {
1394 struct _sas_node *sas_expander, *r;
1395
1396 r = NULL;
1397 list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) {
1398 if (sas_expander->sas_address != sas_address)
1399 continue;
1400 r = sas_expander;
1401 goto out;
1402 }
1403 out:
1404 return r;
1405 }
1406
1407 /**
1408 * _scsih_expander_node_add - insert expander device to the list.
1409 * @ioc: per adapter object
1410 * @sas_expander: the sas_device object
1411 * Context: This function will acquire ioc->sas_node_lock.
1412 *
1413 * Adding new object to the ioc->sas_expander_list.
1414 */
1415 static void
_scsih_expander_node_add(struct MPT3SAS_ADAPTER * ioc,struct _sas_node * sas_expander)1416 _scsih_expander_node_add(struct MPT3SAS_ADAPTER *ioc,
1417 struct _sas_node *sas_expander)
1418 {
1419 unsigned long flags;
1420
1421 spin_lock_irqsave(&ioc->sas_node_lock, flags);
1422 list_add_tail(&sas_expander->list, &ioc->sas_expander_list);
1423 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
1424 }
1425
1426 /**
1427 * _scsih_is_end_device - determines if device is an end device
1428 * @device_info: bitfield providing information about the device.
1429 * Context: none
1430 *
1431 * Return: 1 if end device.
1432 */
1433 static int
_scsih_is_end_device(u32 device_info)1434 _scsih_is_end_device(u32 device_info)
1435 {
1436 if (device_info & MPI2_SAS_DEVICE_INFO_END_DEVICE &&
1437 ((device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) |
1438 (device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET) |
1439 (device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE)))
1440 return 1;
1441 else
1442 return 0;
1443 }
1444
1445 /**
1446 * _scsih_is_nvme_pciescsi_device - determines if
1447 * device is an pcie nvme/scsi device
1448 * @device_info: bitfield providing information about the device.
1449 * Context: none
1450 *
1451 * Returns 1 if device is pcie device type nvme/scsi.
1452 */
1453 static int
_scsih_is_nvme_pciescsi_device(u32 device_info)1454 _scsih_is_nvme_pciescsi_device(u32 device_info)
1455 {
1456 if (((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE)
1457 == MPI26_PCIE_DEVINFO_NVME) ||
1458 ((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE)
1459 == MPI26_PCIE_DEVINFO_SCSI))
1460 return 1;
1461 else
1462 return 0;
1463 }
1464
1465 /**
1466 * mpt3sas_scsih_scsi_lookup_get - returns scmd entry
1467 * @ioc: per adapter object
1468 * @smid: system request message index
1469 *
1470 * Return: the smid stored scmd pointer.
1471 * Then will dereference the stored scmd pointer.
1472 */
1473 struct scsi_cmnd *
mpt3sas_scsih_scsi_lookup_get(struct MPT3SAS_ADAPTER * ioc,u16 smid)1474 mpt3sas_scsih_scsi_lookup_get(struct MPT3SAS_ADAPTER *ioc, u16 smid)
1475 {
1476 struct scsi_cmnd *scmd = NULL;
1477 struct scsiio_tracker *st;
1478 Mpi25SCSIIORequest_t *mpi_request;
1479
1480 if (smid > 0 &&
1481 smid <= ioc->scsiio_depth - INTERNAL_SCSIIO_CMDS_COUNT) {
1482 u32 unique_tag = smid - 1;
1483
1484 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
1485
1486 /*
1487 * If SCSI IO request is outstanding at driver level then
1488 * DevHandle filed must be non-zero. If DevHandle is zero
1489 * then it means that this smid is free at driver level,
1490 * so return NULL.
1491 */
1492 if (!mpi_request->DevHandle)
1493 return scmd;
1494
1495 scmd = scsi_host_find_tag(ioc->shost, unique_tag);
1496 if (scmd) {
1497 st = scsi_cmd_priv(scmd);
1498 if (st->cb_idx == 0xFF || st->smid == 0)
1499 scmd = NULL;
1500 }
1501 }
1502 return scmd;
1503 }
1504
1505 /**
1506 * scsih_change_queue_depth - setting device queue depth
1507 * @sdev: scsi device struct
1508 * @qdepth: requested queue depth
1509 *
1510 * Return: queue depth.
1511 */
1512 static int
scsih_change_queue_depth(struct scsi_device * sdev,int qdepth)1513 scsih_change_queue_depth(struct scsi_device *sdev, int qdepth)
1514 {
1515 struct Scsi_Host *shost = sdev->host;
1516 int max_depth;
1517 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
1518 struct MPT3SAS_DEVICE *sas_device_priv_data;
1519 struct MPT3SAS_TARGET *sas_target_priv_data;
1520 struct _sas_device *sas_device;
1521 unsigned long flags;
1522
1523 max_depth = shost->can_queue;
1524
1525 /*
1526 * limit max device queue for SATA to 32 if enable_sdev_max_qd
1527 * is disabled.
1528 */
1529 if (ioc->enable_sdev_max_qd)
1530 goto not_sata;
1531
1532 sas_device_priv_data = sdev->hostdata;
1533 if (!sas_device_priv_data)
1534 goto not_sata;
1535 sas_target_priv_data = sas_device_priv_data->sas_target;
1536 if (!sas_target_priv_data)
1537 goto not_sata;
1538 if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME))
1539 goto not_sata;
1540
1541 spin_lock_irqsave(&ioc->sas_device_lock, flags);
1542 sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data);
1543 if (sas_device) {
1544 if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE)
1545 max_depth = MPT3SAS_SATA_QUEUE_DEPTH;
1546
1547 sas_device_put(sas_device);
1548 }
1549 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1550
1551 not_sata:
1552
1553 if (!sdev->tagged_supported)
1554 max_depth = 1;
1555 if (qdepth > max_depth)
1556 qdepth = max_depth;
1557 return scsi_change_queue_depth(sdev, qdepth);
1558 }
1559
1560 /**
1561 * mpt3sas_scsih_change_queue_depth - setting device queue depth
1562 * @sdev: scsi device struct
1563 * @qdepth: requested queue depth
1564 *
1565 * Returns nothing.
1566 */
1567 void
mpt3sas_scsih_change_queue_depth(struct scsi_device * sdev,int qdepth)1568 mpt3sas_scsih_change_queue_depth(struct scsi_device *sdev, int qdepth)
1569 {
1570 struct Scsi_Host *shost = sdev->host;
1571 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
1572
1573 if (ioc->enable_sdev_max_qd)
1574 qdepth = shost->can_queue;
1575
1576 scsih_change_queue_depth(sdev, qdepth);
1577 }
1578
1579 /**
1580 * scsih_target_alloc - target add routine
1581 * @starget: scsi target struct
1582 *
1583 * Return: 0 if ok. Any other return is assumed to be an error and
1584 * the device is ignored.
1585 */
1586 static int
scsih_target_alloc(struct scsi_target * starget)1587 scsih_target_alloc(struct scsi_target *starget)
1588 {
1589 struct Scsi_Host *shost = dev_to_shost(&starget->dev);
1590 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
1591 struct MPT3SAS_TARGET *sas_target_priv_data;
1592 struct _sas_device *sas_device;
1593 struct _raid_device *raid_device;
1594 struct _pcie_device *pcie_device;
1595 unsigned long flags;
1596 struct sas_rphy *rphy;
1597
1598 sas_target_priv_data = kzalloc(sizeof(*sas_target_priv_data),
1599 GFP_KERNEL);
1600 if (!sas_target_priv_data)
1601 return -ENOMEM;
1602
1603 starget->hostdata = sas_target_priv_data;
1604 sas_target_priv_data->starget = starget;
1605 sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE;
1606
1607 /* RAID volumes */
1608 if (starget->channel == RAID_CHANNEL) {
1609 spin_lock_irqsave(&ioc->raid_device_lock, flags);
1610 raid_device = _scsih_raid_device_find_by_id(ioc, starget->id,
1611 starget->channel);
1612 if (raid_device) {
1613 sas_target_priv_data->handle = raid_device->handle;
1614 sas_target_priv_data->sas_address = raid_device->wwid;
1615 sas_target_priv_data->flags |= MPT_TARGET_FLAGS_VOLUME;
1616 if (ioc->is_warpdrive)
1617 sas_target_priv_data->raid_device = raid_device;
1618 raid_device->starget = starget;
1619 }
1620 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1621 return 0;
1622 }
1623
1624 /* PCIe devices */
1625 if (starget->channel == PCIE_CHANNEL) {
1626 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1627 pcie_device = __mpt3sas_get_pdev_by_idchannel(ioc, starget->id,
1628 starget->channel);
1629 if (pcie_device) {
1630 sas_target_priv_data->handle = pcie_device->handle;
1631 sas_target_priv_data->sas_address = pcie_device->wwid;
1632 sas_target_priv_data->pcie_dev = pcie_device;
1633 pcie_device->starget = starget;
1634 pcie_device->id = starget->id;
1635 pcie_device->channel = starget->channel;
1636 sas_target_priv_data->flags |=
1637 MPT_TARGET_FLAGS_PCIE_DEVICE;
1638 if (pcie_device->fast_path)
1639 sas_target_priv_data->flags |=
1640 MPT_TARGET_FASTPATH_IO;
1641 }
1642 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1643 return 0;
1644 }
1645
1646 /* sas/sata devices */
1647 spin_lock_irqsave(&ioc->sas_device_lock, flags);
1648 rphy = dev_to_rphy(starget->dev.parent);
1649 sas_device = __mpt3sas_get_sdev_by_addr(ioc,
1650 rphy->identify.sas_address);
1651
1652 if (sas_device) {
1653 sas_target_priv_data->handle = sas_device->handle;
1654 sas_target_priv_data->sas_address = sas_device->sas_address;
1655 sas_target_priv_data->sas_dev = sas_device;
1656 sas_device->starget = starget;
1657 sas_device->id = starget->id;
1658 sas_device->channel = starget->channel;
1659 if (test_bit(sas_device->handle, ioc->pd_handles))
1660 sas_target_priv_data->flags |=
1661 MPT_TARGET_FLAGS_RAID_COMPONENT;
1662 if (sas_device->fast_path)
1663 sas_target_priv_data->flags |=
1664 MPT_TARGET_FASTPATH_IO;
1665 }
1666 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1667
1668 return 0;
1669 }
1670
1671 /**
1672 * scsih_target_destroy - target destroy routine
1673 * @starget: scsi target struct
1674 */
1675 static void
scsih_target_destroy(struct scsi_target * starget)1676 scsih_target_destroy(struct scsi_target *starget)
1677 {
1678 struct Scsi_Host *shost = dev_to_shost(&starget->dev);
1679 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
1680 struct MPT3SAS_TARGET *sas_target_priv_data;
1681 struct _sas_device *sas_device;
1682 struct _raid_device *raid_device;
1683 struct _pcie_device *pcie_device;
1684 unsigned long flags;
1685
1686 sas_target_priv_data = starget->hostdata;
1687 if (!sas_target_priv_data)
1688 return;
1689
1690 if (starget->channel == RAID_CHANNEL) {
1691 spin_lock_irqsave(&ioc->raid_device_lock, flags);
1692 raid_device = _scsih_raid_device_find_by_id(ioc, starget->id,
1693 starget->channel);
1694 if (raid_device) {
1695 raid_device->starget = NULL;
1696 raid_device->sdev = NULL;
1697 }
1698 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1699 goto out;
1700 }
1701
1702 if (starget->channel == PCIE_CHANNEL) {
1703 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1704 pcie_device = __mpt3sas_get_pdev_from_target(ioc,
1705 sas_target_priv_data);
1706 if (pcie_device && (pcie_device->starget == starget) &&
1707 (pcie_device->id == starget->id) &&
1708 (pcie_device->channel == starget->channel))
1709 pcie_device->starget = NULL;
1710
1711 if (pcie_device) {
1712 /*
1713 * Corresponding get() is in _scsih_target_alloc()
1714 */
1715 sas_target_priv_data->pcie_dev = NULL;
1716 pcie_device_put(pcie_device);
1717 pcie_device_put(pcie_device);
1718 }
1719 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1720 goto out;
1721 }
1722
1723 spin_lock_irqsave(&ioc->sas_device_lock, flags);
1724 sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data);
1725 if (sas_device && (sas_device->starget == starget) &&
1726 (sas_device->id == starget->id) &&
1727 (sas_device->channel == starget->channel))
1728 sas_device->starget = NULL;
1729
1730 if (sas_device) {
1731 /*
1732 * Corresponding get() is in _scsih_target_alloc()
1733 */
1734 sas_target_priv_data->sas_dev = NULL;
1735 sas_device_put(sas_device);
1736
1737 sas_device_put(sas_device);
1738 }
1739 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1740
1741 out:
1742 kfree(sas_target_priv_data);
1743 starget->hostdata = NULL;
1744 }
1745
1746 /**
1747 * scsih_slave_alloc - device add routine
1748 * @sdev: scsi device struct
1749 *
1750 * Return: 0 if ok. Any other return is assumed to be an error and
1751 * the device is ignored.
1752 */
1753 static int
scsih_slave_alloc(struct scsi_device * sdev)1754 scsih_slave_alloc(struct scsi_device *sdev)
1755 {
1756 struct Scsi_Host *shost;
1757 struct MPT3SAS_ADAPTER *ioc;
1758 struct MPT3SAS_TARGET *sas_target_priv_data;
1759 struct MPT3SAS_DEVICE *sas_device_priv_data;
1760 struct scsi_target *starget;
1761 struct _raid_device *raid_device;
1762 struct _sas_device *sas_device;
1763 struct _pcie_device *pcie_device;
1764 unsigned long flags;
1765
1766 sas_device_priv_data = kzalloc(sizeof(*sas_device_priv_data),
1767 GFP_KERNEL);
1768 if (!sas_device_priv_data)
1769 return -ENOMEM;
1770
1771 sas_device_priv_data->lun = sdev->lun;
1772 sas_device_priv_data->flags = MPT_DEVICE_FLAGS_INIT;
1773
1774 starget = scsi_target(sdev);
1775 sas_target_priv_data = starget->hostdata;
1776 sas_target_priv_data->num_luns++;
1777 sas_device_priv_data->sas_target = sas_target_priv_data;
1778 sdev->hostdata = sas_device_priv_data;
1779 if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT))
1780 sdev->no_uld_attach = 1;
1781
1782 shost = dev_to_shost(&starget->dev);
1783 ioc = shost_priv(shost);
1784 if (starget->channel == RAID_CHANNEL) {
1785 spin_lock_irqsave(&ioc->raid_device_lock, flags);
1786 raid_device = _scsih_raid_device_find_by_id(ioc,
1787 starget->id, starget->channel);
1788 if (raid_device)
1789 raid_device->sdev = sdev; /* raid is single lun */
1790 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1791 }
1792 if (starget->channel == PCIE_CHANNEL) {
1793 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1794 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc,
1795 sas_target_priv_data->sas_address);
1796 if (pcie_device && (pcie_device->starget == NULL)) {
1797 sdev_printk(KERN_INFO, sdev,
1798 "%s : pcie_device->starget set to starget @ %d\n",
1799 __func__, __LINE__);
1800 pcie_device->starget = starget;
1801 }
1802
1803 if (pcie_device)
1804 pcie_device_put(pcie_device);
1805 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1806
1807 } else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) {
1808 spin_lock_irqsave(&ioc->sas_device_lock, flags);
1809 sas_device = __mpt3sas_get_sdev_by_addr(ioc,
1810 sas_target_priv_data->sas_address);
1811 if (sas_device && (sas_device->starget == NULL)) {
1812 sdev_printk(KERN_INFO, sdev,
1813 "%s : sas_device->starget set to starget @ %d\n",
1814 __func__, __LINE__);
1815 sas_device->starget = starget;
1816 }
1817
1818 if (sas_device)
1819 sas_device_put(sas_device);
1820
1821 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1822 }
1823
1824 return 0;
1825 }
1826
1827 /**
1828 * scsih_slave_destroy - device destroy routine
1829 * @sdev: scsi device struct
1830 */
1831 static void
scsih_slave_destroy(struct scsi_device * sdev)1832 scsih_slave_destroy(struct scsi_device *sdev)
1833 {
1834 struct MPT3SAS_TARGET *sas_target_priv_data;
1835 struct scsi_target *starget;
1836 struct Scsi_Host *shost;
1837 struct MPT3SAS_ADAPTER *ioc;
1838 struct _sas_device *sas_device;
1839 struct _pcie_device *pcie_device;
1840 unsigned long flags;
1841
1842 if (!sdev->hostdata)
1843 return;
1844
1845 starget = scsi_target(sdev);
1846 sas_target_priv_data = starget->hostdata;
1847 sas_target_priv_data->num_luns--;
1848
1849 shost = dev_to_shost(&starget->dev);
1850 ioc = shost_priv(shost);
1851
1852 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) {
1853 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1854 pcie_device = __mpt3sas_get_pdev_from_target(ioc,
1855 sas_target_priv_data);
1856 if (pcie_device && !sas_target_priv_data->num_luns)
1857 pcie_device->starget = NULL;
1858
1859 if (pcie_device)
1860 pcie_device_put(pcie_device);
1861
1862 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1863
1864 } else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) {
1865 spin_lock_irqsave(&ioc->sas_device_lock, flags);
1866 sas_device = __mpt3sas_get_sdev_from_target(ioc,
1867 sas_target_priv_data);
1868 if (sas_device && !sas_target_priv_data->num_luns)
1869 sas_device->starget = NULL;
1870
1871 if (sas_device)
1872 sas_device_put(sas_device);
1873 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1874 }
1875
1876 kfree(sdev->hostdata);
1877 sdev->hostdata = NULL;
1878 }
1879
1880 /**
1881 * _scsih_display_sata_capabilities - sata capabilities
1882 * @ioc: per adapter object
1883 * @handle: device handle
1884 * @sdev: scsi device struct
1885 */
1886 static void
_scsih_display_sata_capabilities(struct MPT3SAS_ADAPTER * ioc,u16 handle,struct scsi_device * sdev)1887 _scsih_display_sata_capabilities(struct MPT3SAS_ADAPTER *ioc,
1888 u16 handle, struct scsi_device *sdev)
1889 {
1890 Mpi2ConfigReply_t mpi_reply;
1891 Mpi2SasDevicePage0_t sas_device_pg0;
1892 u32 ioc_status;
1893 u16 flags;
1894 u32 device_info;
1895
1896 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
1897 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) {
1898 ioc_err(ioc, "failure at %s:%d/%s()!\n",
1899 __FILE__, __LINE__, __func__);
1900 return;
1901 }
1902
1903 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
1904 MPI2_IOCSTATUS_MASK;
1905 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
1906 ioc_err(ioc, "failure at %s:%d/%s()!\n",
1907 __FILE__, __LINE__, __func__);
1908 return;
1909 }
1910
1911 flags = le16_to_cpu(sas_device_pg0.Flags);
1912 device_info = le32_to_cpu(sas_device_pg0.DeviceInfo);
1913
1914 sdev_printk(KERN_INFO, sdev,
1915 "atapi(%s), ncq(%s), asyn_notify(%s), smart(%s), fua(%s), "
1916 "sw_preserve(%s)\n",
1917 (device_info & MPI2_SAS_DEVICE_INFO_ATAPI_DEVICE) ? "y" : "n",
1918 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_NCQ_SUPPORTED) ? "y" : "n",
1919 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_ASYNCHRONOUS_NOTIFY) ? "y" :
1920 "n",
1921 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SMART_SUPPORTED) ? "y" : "n",
1922 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_FUA_SUPPORTED) ? "y" : "n",
1923 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SW_PRESERVE) ? "y" : "n");
1924 }
1925
1926 /*
1927 * raid transport support -
1928 * Enabled for SLES11 and newer, in older kernels the driver will panic when
1929 * unloading the driver followed by a load - I believe that the subroutine
1930 * raid_class_release() is not cleaning up properly.
1931 */
1932
1933 /**
1934 * scsih_is_raid - return boolean indicating device is raid volume
1935 * @dev: the device struct object
1936 */
1937 static int
scsih_is_raid(struct device * dev)1938 scsih_is_raid(struct device *dev)
1939 {
1940 struct scsi_device *sdev = to_scsi_device(dev);
1941 struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host);
1942
1943 if (ioc->is_warpdrive)
1944 return 0;
1945 return (sdev->channel == RAID_CHANNEL) ? 1 : 0;
1946 }
1947
1948 static int
scsih_is_nvme(struct device * dev)1949 scsih_is_nvme(struct device *dev)
1950 {
1951 struct scsi_device *sdev = to_scsi_device(dev);
1952
1953 return (sdev->channel == PCIE_CHANNEL) ? 1 : 0;
1954 }
1955
1956 /**
1957 * scsih_get_resync - get raid volume resync percent complete
1958 * @dev: the device struct object
1959 */
1960 static void
scsih_get_resync(struct device * dev)1961 scsih_get_resync(struct device *dev)
1962 {
1963 struct scsi_device *sdev = to_scsi_device(dev);
1964 struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host);
1965 static struct _raid_device *raid_device;
1966 unsigned long flags;
1967 Mpi2RaidVolPage0_t vol_pg0;
1968 Mpi2ConfigReply_t mpi_reply;
1969 u32 volume_status_flags;
1970 u8 percent_complete;
1971 u16 handle;
1972
1973 percent_complete = 0;
1974 handle = 0;
1975 if (ioc->is_warpdrive)
1976 goto out;
1977
1978 spin_lock_irqsave(&ioc->raid_device_lock, flags);
1979 raid_device = _scsih_raid_device_find_by_id(ioc, sdev->id,
1980 sdev->channel);
1981 if (raid_device) {
1982 handle = raid_device->handle;
1983 percent_complete = raid_device->percent_complete;
1984 }
1985 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1986
1987 if (!handle)
1988 goto out;
1989
1990 if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &vol_pg0,
1991 MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle,
1992 sizeof(Mpi2RaidVolPage0_t))) {
1993 ioc_err(ioc, "failure at %s:%d/%s()!\n",
1994 __FILE__, __LINE__, __func__);
1995 percent_complete = 0;
1996 goto out;
1997 }
1998
1999 volume_status_flags = le32_to_cpu(vol_pg0.VolumeStatusFlags);
2000 if (!(volume_status_flags &
2001 MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS))
2002 percent_complete = 0;
2003
2004 out:
2005
2006 switch (ioc->hba_mpi_version_belonged) {
2007 case MPI2_VERSION:
2008 raid_set_resync(mpt2sas_raid_template, dev, percent_complete);
2009 break;
2010 case MPI25_VERSION:
2011 case MPI26_VERSION:
2012 raid_set_resync(mpt3sas_raid_template, dev, percent_complete);
2013 break;
2014 }
2015 }
2016
2017 /**
2018 * scsih_get_state - get raid volume level
2019 * @dev: the device struct object
2020 */
2021 static void
scsih_get_state(struct device * dev)2022 scsih_get_state(struct device *dev)
2023 {
2024 struct scsi_device *sdev = to_scsi_device(dev);
2025 struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host);
2026 static struct _raid_device *raid_device;
2027 unsigned long flags;
2028 Mpi2RaidVolPage0_t vol_pg0;
2029 Mpi2ConfigReply_t mpi_reply;
2030 u32 volstate;
2031 enum raid_state state = RAID_STATE_UNKNOWN;
2032 u16 handle = 0;
2033
2034 spin_lock_irqsave(&ioc->raid_device_lock, flags);
2035 raid_device = _scsih_raid_device_find_by_id(ioc, sdev->id,
2036 sdev->channel);
2037 if (raid_device)
2038 handle = raid_device->handle;
2039 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
2040
2041 if (!raid_device)
2042 goto out;
2043
2044 if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &vol_pg0,
2045 MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle,
2046 sizeof(Mpi2RaidVolPage0_t))) {
2047 ioc_err(ioc, "failure at %s:%d/%s()!\n",
2048 __FILE__, __LINE__, __func__);
2049 goto out;
2050 }
2051
2052 volstate = le32_to_cpu(vol_pg0.VolumeStatusFlags);
2053 if (volstate & MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS) {
2054 state = RAID_STATE_RESYNCING;
2055 goto out;
2056 }
2057
2058 switch (vol_pg0.VolumeState) {
2059 case MPI2_RAID_VOL_STATE_OPTIMAL:
2060 case MPI2_RAID_VOL_STATE_ONLINE:
2061 state = RAID_STATE_ACTIVE;
2062 break;
2063 case MPI2_RAID_VOL_STATE_DEGRADED:
2064 state = RAID_STATE_DEGRADED;
2065 break;
2066 case MPI2_RAID_VOL_STATE_FAILED:
2067 case MPI2_RAID_VOL_STATE_MISSING:
2068 state = RAID_STATE_OFFLINE;
2069 break;
2070 }
2071 out:
2072 switch (ioc->hba_mpi_version_belonged) {
2073 case MPI2_VERSION:
2074 raid_set_state(mpt2sas_raid_template, dev, state);
2075 break;
2076 case MPI25_VERSION:
2077 case MPI26_VERSION:
2078 raid_set_state(mpt3sas_raid_template, dev, state);
2079 break;
2080 }
2081 }
2082
2083 /**
2084 * _scsih_set_level - set raid level
2085 * @ioc: ?
2086 * @sdev: scsi device struct
2087 * @volume_type: volume type
2088 */
2089 static void
_scsih_set_level(struct MPT3SAS_ADAPTER * ioc,struct scsi_device * sdev,u8 volume_type)2090 _scsih_set_level(struct MPT3SAS_ADAPTER *ioc,
2091 struct scsi_device *sdev, u8 volume_type)
2092 {
2093 enum raid_level level = RAID_LEVEL_UNKNOWN;
2094
2095 switch (volume_type) {
2096 case MPI2_RAID_VOL_TYPE_RAID0:
2097 level = RAID_LEVEL_0;
2098 break;
2099 case MPI2_RAID_VOL_TYPE_RAID10:
2100 level = RAID_LEVEL_10;
2101 break;
2102 case MPI2_RAID_VOL_TYPE_RAID1E:
2103 level = RAID_LEVEL_1E;
2104 break;
2105 case MPI2_RAID_VOL_TYPE_RAID1:
2106 level = RAID_LEVEL_1;
2107 break;
2108 }
2109
2110 switch (ioc->hba_mpi_version_belonged) {
2111 case MPI2_VERSION:
2112 raid_set_level(mpt2sas_raid_template,
2113 &sdev->sdev_gendev, level);
2114 break;
2115 case MPI25_VERSION:
2116 case MPI26_VERSION:
2117 raid_set_level(mpt3sas_raid_template,
2118 &sdev->sdev_gendev, level);
2119 break;
2120 }
2121 }
2122
2123
2124 /**
2125 * _scsih_get_volume_capabilities - volume capabilities
2126 * @ioc: per adapter object
2127 * @raid_device: the raid_device object
2128 *
2129 * Return: 0 for success, else 1
2130 */
2131 static int
_scsih_get_volume_capabilities(struct MPT3SAS_ADAPTER * ioc,struct _raid_device * raid_device)2132 _scsih_get_volume_capabilities(struct MPT3SAS_ADAPTER *ioc,
2133 struct _raid_device *raid_device)
2134 {
2135 Mpi2RaidVolPage0_t *vol_pg0;
2136 Mpi2RaidPhysDiskPage0_t pd_pg0;
2137 Mpi2SasDevicePage0_t sas_device_pg0;
2138 Mpi2ConfigReply_t mpi_reply;
2139 u16 sz;
2140 u8 num_pds;
2141
2142 if ((mpt3sas_config_get_number_pds(ioc, raid_device->handle,
2143 &num_pds)) || !num_pds) {
2144 dfailprintk(ioc,
2145 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2146 __FILE__, __LINE__, __func__));
2147 return 1;
2148 }
2149
2150 raid_device->num_pds = num_pds;
2151 sz = offsetof(Mpi2RaidVolPage0_t, PhysDisk) + (num_pds *
2152 sizeof(Mpi2RaidVol0PhysDisk_t));
2153 vol_pg0 = kzalloc(sz, GFP_KERNEL);
2154 if (!vol_pg0) {
2155 dfailprintk(ioc,
2156 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2157 __FILE__, __LINE__, __func__));
2158 return 1;
2159 }
2160
2161 if ((mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, vol_pg0,
2162 MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, raid_device->handle, sz))) {
2163 dfailprintk(ioc,
2164 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2165 __FILE__, __LINE__, __func__));
2166 kfree(vol_pg0);
2167 return 1;
2168 }
2169
2170 raid_device->volume_type = vol_pg0->VolumeType;
2171
2172 /* figure out what the underlying devices are by
2173 * obtaining the device_info bits for the 1st device
2174 */
2175 if (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply,
2176 &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_PHYSDISKNUM,
2177 vol_pg0->PhysDisk[0].PhysDiskNum))) {
2178 if (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
2179 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE,
2180 le16_to_cpu(pd_pg0.DevHandle)))) {
2181 raid_device->device_info =
2182 le32_to_cpu(sas_device_pg0.DeviceInfo);
2183 }
2184 }
2185
2186 kfree(vol_pg0);
2187 return 0;
2188 }
2189
2190 /**
2191 * _scsih_enable_tlr - setting TLR flags
2192 * @ioc: per adapter object
2193 * @sdev: scsi device struct
2194 *
2195 * Enabling Transaction Layer Retries for tape devices when
2196 * vpd page 0x90 is present
2197 *
2198 */
2199 static void
_scsih_enable_tlr(struct MPT3SAS_ADAPTER * ioc,struct scsi_device * sdev)2200 _scsih_enable_tlr(struct MPT3SAS_ADAPTER *ioc, struct scsi_device *sdev)
2201 {
2202
2203 /* only for TAPE */
2204 if (sdev->type != TYPE_TAPE)
2205 return;
2206
2207 if (!(ioc->facts.IOCCapabilities & MPI2_IOCFACTS_CAPABILITY_TLR))
2208 return;
2209
2210 sas_enable_tlr(sdev);
2211 sdev_printk(KERN_INFO, sdev, "TLR %s\n",
2212 sas_is_tlr_enabled(sdev) ? "Enabled" : "Disabled");
2213 return;
2214
2215 }
2216
2217 /**
2218 * scsih_slave_configure - device configure routine.
2219 * @sdev: scsi device struct
2220 *
2221 * Return: 0 if ok. Any other return is assumed to be an error and
2222 * the device is ignored.
2223 */
2224 static int
scsih_slave_configure(struct scsi_device * sdev)2225 scsih_slave_configure(struct scsi_device *sdev)
2226 {
2227 struct Scsi_Host *shost = sdev->host;
2228 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
2229 struct MPT3SAS_DEVICE *sas_device_priv_data;
2230 struct MPT3SAS_TARGET *sas_target_priv_data;
2231 struct _sas_device *sas_device;
2232 struct _pcie_device *pcie_device;
2233 struct _raid_device *raid_device;
2234 unsigned long flags;
2235 int qdepth;
2236 u8 ssp_target = 0;
2237 char *ds = "";
2238 char *r_level = "";
2239 u16 handle, volume_handle = 0;
2240 u64 volume_wwid = 0;
2241
2242 qdepth = 1;
2243 sas_device_priv_data = sdev->hostdata;
2244 sas_device_priv_data->configured_lun = 1;
2245 sas_device_priv_data->flags &= ~MPT_DEVICE_FLAGS_INIT;
2246 sas_target_priv_data = sas_device_priv_data->sas_target;
2247 handle = sas_target_priv_data->handle;
2248
2249 /* raid volume handling */
2250 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME) {
2251
2252 spin_lock_irqsave(&ioc->raid_device_lock, flags);
2253 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
2254 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
2255 if (!raid_device) {
2256 dfailprintk(ioc,
2257 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2258 __FILE__, __LINE__, __func__));
2259 return 1;
2260 }
2261
2262 if (_scsih_get_volume_capabilities(ioc, raid_device)) {
2263 dfailprintk(ioc,
2264 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2265 __FILE__, __LINE__, __func__));
2266 return 1;
2267 }
2268
2269 /*
2270 * WARPDRIVE: Initialize the required data for Direct IO
2271 */
2272 mpt3sas_init_warpdrive_properties(ioc, raid_device);
2273
2274 /* RAID Queue Depth Support
2275 * IS volume = underlying qdepth of drive type, either
2276 * MPT3SAS_SAS_QUEUE_DEPTH or MPT3SAS_SATA_QUEUE_DEPTH
2277 * IM/IME/R10 = 128 (MPT3SAS_RAID_QUEUE_DEPTH)
2278 */
2279 if (raid_device->device_info &
2280 MPI2_SAS_DEVICE_INFO_SSP_TARGET) {
2281 qdepth = MPT3SAS_SAS_QUEUE_DEPTH;
2282 ds = "SSP";
2283 } else {
2284 qdepth = MPT3SAS_SATA_QUEUE_DEPTH;
2285 if (raid_device->device_info &
2286 MPI2_SAS_DEVICE_INFO_SATA_DEVICE)
2287 ds = "SATA";
2288 else
2289 ds = "STP";
2290 }
2291
2292 switch (raid_device->volume_type) {
2293 case MPI2_RAID_VOL_TYPE_RAID0:
2294 r_level = "RAID0";
2295 break;
2296 case MPI2_RAID_VOL_TYPE_RAID1E:
2297 qdepth = MPT3SAS_RAID_QUEUE_DEPTH;
2298 if (ioc->manu_pg10.OEMIdentifier &&
2299 (le32_to_cpu(ioc->manu_pg10.GenericFlags0) &
2300 MFG10_GF0_R10_DISPLAY) &&
2301 !(raid_device->num_pds % 2))
2302 r_level = "RAID10";
2303 else
2304 r_level = "RAID1E";
2305 break;
2306 case MPI2_RAID_VOL_TYPE_RAID1:
2307 qdepth = MPT3SAS_RAID_QUEUE_DEPTH;
2308 r_level = "RAID1";
2309 break;
2310 case MPI2_RAID_VOL_TYPE_RAID10:
2311 qdepth = MPT3SAS_RAID_QUEUE_DEPTH;
2312 r_level = "RAID10";
2313 break;
2314 case MPI2_RAID_VOL_TYPE_UNKNOWN:
2315 default:
2316 qdepth = MPT3SAS_RAID_QUEUE_DEPTH;
2317 r_level = "RAIDX";
2318 break;
2319 }
2320
2321 if (!ioc->hide_ir_msg)
2322 sdev_printk(KERN_INFO, sdev,
2323 "%s: handle(0x%04x), wwid(0x%016llx),"
2324 " pd_count(%d), type(%s)\n",
2325 r_level, raid_device->handle,
2326 (unsigned long long)raid_device->wwid,
2327 raid_device->num_pds, ds);
2328
2329 if (shost->max_sectors > MPT3SAS_RAID_MAX_SECTORS) {
2330 blk_queue_max_hw_sectors(sdev->request_queue,
2331 MPT3SAS_RAID_MAX_SECTORS);
2332 sdev_printk(KERN_INFO, sdev,
2333 "Set queue's max_sector to: %u\n",
2334 MPT3SAS_RAID_MAX_SECTORS);
2335 }
2336
2337 mpt3sas_scsih_change_queue_depth(sdev, qdepth);
2338
2339 /* raid transport support */
2340 if (!ioc->is_warpdrive)
2341 _scsih_set_level(ioc, sdev, raid_device->volume_type);
2342 return 0;
2343 }
2344
2345 /* non-raid handling */
2346 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) {
2347 if (mpt3sas_config_get_volume_handle(ioc, handle,
2348 &volume_handle)) {
2349 dfailprintk(ioc,
2350 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2351 __FILE__, __LINE__, __func__));
2352 return 1;
2353 }
2354 if (volume_handle && mpt3sas_config_get_volume_wwid(ioc,
2355 volume_handle, &volume_wwid)) {
2356 dfailprintk(ioc,
2357 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2358 __FILE__, __LINE__, __func__));
2359 return 1;
2360 }
2361 }
2362
2363 /* PCIe handling */
2364 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) {
2365 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
2366 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc,
2367 sas_device_priv_data->sas_target->sas_address);
2368 if (!pcie_device) {
2369 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
2370 dfailprintk(ioc,
2371 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2372 __FILE__, __LINE__, __func__));
2373 return 1;
2374 }
2375
2376 qdepth = MPT3SAS_NVME_QUEUE_DEPTH;
2377 ds = "NVMe";
2378 sdev_printk(KERN_INFO, sdev,
2379 "%s: handle(0x%04x), wwid(0x%016llx), port(%d)\n",
2380 ds, handle, (unsigned long long)pcie_device->wwid,
2381 pcie_device->port_num);
2382 if (pcie_device->enclosure_handle != 0)
2383 sdev_printk(KERN_INFO, sdev,
2384 "%s: enclosure logical id(0x%016llx), slot(%d)\n",
2385 ds,
2386 (unsigned long long)pcie_device->enclosure_logical_id,
2387 pcie_device->slot);
2388 if (pcie_device->connector_name[0] != '\0')
2389 sdev_printk(KERN_INFO, sdev,
2390 "%s: enclosure level(0x%04x),"
2391 "connector name( %s)\n", ds,
2392 pcie_device->enclosure_level,
2393 pcie_device->connector_name);
2394
2395 if (pcie_device->nvme_mdts)
2396 blk_queue_max_hw_sectors(sdev->request_queue,
2397 pcie_device->nvme_mdts/512);
2398
2399 pcie_device_put(pcie_device);
2400 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
2401 mpt3sas_scsih_change_queue_depth(sdev, qdepth);
2402 /* Enable QUEUE_FLAG_NOMERGES flag, so that IOs won't be
2403 ** merged and can eliminate holes created during merging
2404 ** operation.
2405 **/
2406 blk_queue_flag_set(QUEUE_FLAG_NOMERGES,
2407 sdev->request_queue);
2408 blk_queue_virt_boundary(sdev->request_queue,
2409 ioc->page_size - 1);
2410 return 0;
2411 }
2412
2413 spin_lock_irqsave(&ioc->sas_device_lock, flags);
2414 sas_device = __mpt3sas_get_sdev_by_addr(ioc,
2415 sas_device_priv_data->sas_target->sas_address);
2416 if (!sas_device) {
2417 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
2418 dfailprintk(ioc,
2419 ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2420 __FILE__, __LINE__, __func__));
2421 return 1;
2422 }
2423
2424 sas_device->volume_handle = volume_handle;
2425 sas_device->volume_wwid = volume_wwid;
2426 if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) {
2427 qdepth = MPT3SAS_SAS_QUEUE_DEPTH;
2428 ssp_target = 1;
2429 if (sas_device->device_info &
2430 MPI2_SAS_DEVICE_INFO_SEP) {
2431 sdev_printk(KERN_WARNING, sdev,
2432 "set ignore_delay_remove for handle(0x%04x)\n",
2433 sas_device_priv_data->sas_target->handle);
2434 sas_device_priv_data->ignore_delay_remove = 1;
2435 ds = "SES";
2436 } else
2437 ds = "SSP";
2438 } else {
2439 qdepth = MPT3SAS_SATA_QUEUE_DEPTH;
2440 if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET)
2441 ds = "STP";
2442 else if (sas_device->device_info &
2443 MPI2_SAS_DEVICE_INFO_SATA_DEVICE)
2444 ds = "SATA";
2445 }
2446
2447 sdev_printk(KERN_INFO, sdev, "%s: handle(0x%04x), " \
2448 "sas_addr(0x%016llx), phy(%d), device_name(0x%016llx)\n",
2449 ds, handle, (unsigned long long)sas_device->sas_address,
2450 sas_device->phy, (unsigned long long)sas_device->device_name);
2451
2452 _scsih_display_enclosure_chassis_info(NULL, sas_device, sdev, NULL);
2453
2454 sas_device_put(sas_device);
2455 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
2456
2457 if (!ssp_target)
2458 _scsih_display_sata_capabilities(ioc, handle, sdev);
2459
2460
2461 mpt3sas_scsih_change_queue_depth(sdev, qdepth);
2462
2463 if (ssp_target) {
2464 sas_read_port_mode_page(sdev);
2465 _scsih_enable_tlr(ioc, sdev);
2466 }
2467
2468 return 0;
2469 }
2470
2471 /**
2472 * scsih_bios_param - fetch head, sector, cylinder info for a disk
2473 * @sdev: scsi device struct
2474 * @bdev: pointer to block device context
2475 * @capacity: device size (in 512 byte sectors)
2476 * @params: three element array to place output:
2477 * params[0] number of heads (max 255)
2478 * params[1] number of sectors (max 63)
2479 * params[2] number of cylinders
2480 */
2481 static int
scsih_bios_param(struct scsi_device * sdev,struct block_device * bdev,sector_t capacity,int params[])2482 scsih_bios_param(struct scsi_device *sdev, struct block_device *bdev,
2483 sector_t capacity, int params[])
2484 {
2485 int heads;
2486 int sectors;
2487 sector_t cylinders;
2488 ulong dummy;
2489
2490 heads = 64;
2491 sectors = 32;
2492
2493 dummy = heads * sectors;
2494 cylinders = capacity;
2495 sector_div(cylinders, dummy);
2496
2497 /*
2498 * Handle extended translation size for logical drives
2499 * > 1Gb
2500 */
2501 if ((ulong)capacity >= 0x200000) {
2502 heads = 255;
2503 sectors = 63;
2504 dummy = heads * sectors;
2505 cylinders = capacity;
2506 sector_div(cylinders, dummy);
2507 }
2508
2509 /* return result */
2510 params[0] = heads;
2511 params[1] = sectors;
2512 params[2] = cylinders;
2513
2514 return 0;
2515 }
2516
2517 /**
2518 * _scsih_response_code - translation of device response code
2519 * @ioc: per adapter object
2520 * @response_code: response code returned by the device
2521 */
2522 static void
_scsih_response_code(struct MPT3SAS_ADAPTER * ioc,u8 response_code)2523 _scsih_response_code(struct MPT3SAS_ADAPTER *ioc, u8 response_code)
2524 {
2525 char *desc;
2526
2527 switch (response_code) {
2528 case MPI2_SCSITASKMGMT_RSP_TM_COMPLETE:
2529 desc = "task management request completed";
2530 break;
2531 case MPI2_SCSITASKMGMT_RSP_INVALID_FRAME:
2532 desc = "invalid frame";
2533 break;
2534 case MPI2_SCSITASKMGMT_RSP_TM_NOT_SUPPORTED:
2535 desc = "task management request not supported";
2536 break;
2537 case MPI2_SCSITASKMGMT_RSP_TM_FAILED:
2538 desc = "task management request failed";
2539 break;
2540 case MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED:
2541 desc = "task management request succeeded";
2542 break;
2543 case MPI2_SCSITASKMGMT_RSP_TM_INVALID_LUN:
2544 desc = "invalid lun";
2545 break;
2546 case 0xA:
2547 desc = "overlapped tag attempted";
2548 break;
2549 case MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC:
2550 desc = "task queued, however not sent to target";
2551 break;
2552 default:
2553 desc = "unknown";
2554 break;
2555 }
2556 ioc_warn(ioc, "response_code(0x%01x): %s\n", response_code, desc);
2557 }
2558
2559 /**
2560 * _scsih_tm_done - tm completion routine
2561 * @ioc: per adapter object
2562 * @smid: system request message index
2563 * @msix_index: MSIX table index supplied by the OS
2564 * @reply: reply message frame(lower 32bit addr)
2565 * Context: none.
2566 *
2567 * The callback handler when using scsih_issue_tm.
2568 *
2569 * Return: 1 meaning mf should be freed from _base_interrupt
2570 * 0 means the mf is freed from this function.
2571 */
2572 static u8
_scsih_tm_done(struct MPT3SAS_ADAPTER * ioc,u16 smid,u8 msix_index,u32 reply)2573 _scsih_tm_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply)
2574 {
2575 MPI2DefaultReply_t *mpi_reply;
2576
2577 if (ioc->tm_cmds.status == MPT3_CMD_NOT_USED)
2578 return 1;
2579 if (ioc->tm_cmds.smid != smid)
2580 return 1;
2581 ioc->tm_cmds.status |= MPT3_CMD_COMPLETE;
2582 mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply);
2583 if (mpi_reply) {
2584 memcpy(ioc->tm_cmds.reply, mpi_reply, mpi_reply->MsgLength*4);
2585 ioc->tm_cmds.status |= MPT3_CMD_REPLY_VALID;
2586 }
2587 ioc->tm_cmds.status &= ~MPT3_CMD_PENDING;
2588 complete(&ioc->tm_cmds.done);
2589 return 1;
2590 }
2591
2592 /**
2593 * mpt3sas_scsih_set_tm_flag - set per target tm_busy
2594 * @ioc: per adapter object
2595 * @handle: device handle
2596 *
2597 * During taskmangement request, we need to freeze the device queue.
2598 */
2599 void
mpt3sas_scsih_set_tm_flag(struct MPT3SAS_ADAPTER * ioc,u16 handle)2600 mpt3sas_scsih_set_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle)
2601 {
2602 struct MPT3SAS_DEVICE *sas_device_priv_data;
2603 struct scsi_device *sdev;
2604 u8 skip = 0;
2605
2606 shost_for_each_device(sdev, ioc->shost) {
2607 if (skip)
2608 continue;
2609 sas_device_priv_data = sdev->hostdata;
2610 if (!sas_device_priv_data)
2611 continue;
2612 if (sas_device_priv_data->sas_target->handle == handle) {
2613 sas_device_priv_data->sas_target->tm_busy = 1;
2614 skip = 1;
2615 ioc->ignore_loginfos = 1;
2616 }
2617 }
2618 }
2619
2620 /**
2621 * mpt3sas_scsih_clear_tm_flag - clear per target tm_busy
2622 * @ioc: per adapter object
2623 * @handle: device handle
2624 *
2625 * During taskmangement request, we need to freeze the device queue.
2626 */
2627 void
mpt3sas_scsih_clear_tm_flag(struct MPT3SAS_ADAPTER * ioc,u16 handle)2628 mpt3sas_scsih_clear_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle)
2629 {
2630 struct MPT3SAS_DEVICE *sas_device_priv_data;
2631 struct scsi_device *sdev;
2632 u8 skip = 0;
2633
2634 shost_for_each_device(sdev, ioc->shost) {
2635 if (skip)
2636 continue;
2637 sas_device_priv_data = sdev->hostdata;
2638 if (!sas_device_priv_data)
2639 continue;
2640 if (sas_device_priv_data->sas_target->handle == handle) {
2641 sas_device_priv_data->sas_target->tm_busy = 0;
2642 skip = 1;
2643 ioc->ignore_loginfos = 0;
2644 }
2645 }
2646 }
2647
2648 /**
2649 * mpt3sas_scsih_issue_tm - main routine for sending tm requests
2650 * @ioc: per adapter struct
2651 * @handle: device handle
2652 * @lun: lun number
2653 * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h)
2654 * @smid_task: smid assigned to the task
2655 * @msix_task: MSIX table index supplied by the OS
2656 * @timeout: timeout in seconds
2657 * @tr_method: Target Reset Method
2658 * Context: user
2659 *
2660 * A generic API for sending task management requests to firmware.
2661 *
2662 * The callback index is set inside `ioc->tm_cb_idx`.
2663 * The caller is responsible to check for outstanding commands.
2664 *
2665 * Return: SUCCESS or FAILED.
2666 */
2667 int
mpt3sas_scsih_issue_tm(struct MPT3SAS_ADAPTER * ioc,u16 handle,u64 lun,u8 type,u16 smid_task,u16 msix_task,u8 timeout,u8 tr_method)2668 mpt3sas_scsih_issue_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle, u64 lun,
2669 u8 type, u16 smid_task, u16 msix_task, u8 timeout, u8 tr_method)
2670 {
2671 Mpi2SCSITaskManagementRequest_t *mpi_request;
2672 Mpi2SCSITaskManagementReply_t *mpi_reply;
2673 u16 smid = 0;
2674 u32 ioc_state;
2675 int rc;
2676
2677 lockdep_assert_held(&ioc->tm_cmds.mutex);
2678
2679 if (ioc->tm_cmds.status != MPT3_CMD_NOT_USED) {
2680 ioc_info(ioc, "%s: tm_cmd busy!!!\n", __func__);
2681 return FAILED;
2682 }
2683
2684 if (ioc->shost_recovery || ioc->remove_host ||
2685 ioc->pci_error_recovery) {
2686 ioc_info(ioc, "%s: host reset in progress!\n", __func__);
2687 return FAILED;
2688 }
2689
2690 ioc_state = mpt3sas_base_get_iocstate(ioc, 0);
2691 if (ioc_state & MPI2_DOORBELL_USED) {
2692 dhsprintk(ioc, ioc_info(ioc, "unexpected doorbell active!\n"));
2693 rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
2694 return (!rc) ? SUCCESS : FAILED;
2695 }
2696
2697 if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) {
2698 mpt3sas_base_fault_info(ioc, ioc_state &
2699 MPI2_DOORBELL_DATA_MASK);
2700 rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
2701 return (!rc) ? SUCCESS : FAILED;
2702 }
2703
2704 smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_cb_idx);
2705 if (!smid) {
2706 ioc_err(ioc, "%s: failed obtaining a smid\n", __func__);
2707 return FAILED;
2708 }
2709
2710 dtmprintk(ioc,
2711 ioc_info(ioc, "sending tm: handle(0x%04x), task_type(0x%02x), smid(%d), timeout(%d), tr_method(0x%x)\n",
2712 handle, type, smid_task, timeout, tr_method));
2713 ioc->tm_cmds.status = MPT3_CMD_PENDING;
2714 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
2715 ioc->tm_cmds.smid = smid;
2716 memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t));
2717 memset(ioc->tm_cmds.reply, 0, sizeof(Mpi2SCSITaskManagementReply_t));
2718 mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
2719 mpi_request->DevHandle = cpu_to_le16(handle);
2720 mpi_request->TaskType = type;
2721 mpi_request->MsgFlags = tr_method;
2722 mpi_request->TaskMID = cpu_to_le16(smid_task);
2723 int_to_scsilun(lun, (struct scsi_lun *)mpi_request->LUN);
2724 mpt3sas_scsih_set_tm_flag(ioc, handle);
2725 init_completion(&ioc->tm_cmds.done);
2726 ioc->put_smid_hi_priority(ioc, smid, msix_task);
2727 wait_for_completion_timeout(&ioc->tm_cmds.done, timeout*HZ);
2728 if (!(ioc->tm_cmds.status & MPT3_CMD_COMPLETE)) {
2729 if (mpt3sas_base_check_cmd_timeout(ioc,
2730 ioc->tm_cmds.status, mpi_request,
2731 sizeof(Mpi2SCSITaskManagementRequest_t)/4)) {
2732 rc = mpt3sas_base_hard_reset_handler(ioc,
2733 FORCE_BIG_HAMMER);
2734 rc = (!rc) ? SUCCESS : FAILED;
2735 goto out;
2736 }
2737 }
2738
2739 /* sync IRQs in case those were busy during flush. */
2740 mpt3sas_base_sync_reply_irqs(ioc);
2741
2742 if (ioc->tm_cmds.status & MPT3_CMD_REPLY_VALID) {
2743 mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT);
2744 mpi_reply = ioc->tm_cmds.reply;
2745 dtmprintk(ioc,
2746 ioc_info(ioc, "complete tm: ioc_status(0x%04x), loginfo(0x%08x), term_count(0x%08x)\n",
2747 le16_to_cpu(mpi_reply->IOCStatus),
2748 le32_to_cpu(mpi_reply->IOCLogInfo),
2749 le32_to_cpu(mpi_reply->TerminationCount)));
2750 if (ioc->logging_level & MPT_DEBUG_TM) {
2751 _scsih_response_code(ioc, mpi_reply->ResponseCode);
2752 if (mpi_reply->IOCStatus)
2753 _debug_dump_mf(mpi_request,
2754 sizeof(Mpi2SCSITaskManagementRequest_t)/4);
2755 }
2756 }
2757 rc = SUCCESS;
2758
2759 out:
2760 mpt3sas_scsih_clear_tm_flag(ioc, handle);
2761 ioc->tm_cmds.status = MPT3_CMD_NOT_USED;
2762 return rc;
2763 }
2764
mpt3sas_scsih_issue_locked_tm(struct MPT3SAS_ADAPTER * ioc,u16 handle,u64 lun,u8 type,u16 smid_task,u16 msix_task,u8 timeout,u8 tr_method)2765 int mpt3sas_scsih_issue_locked_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle,
2766 u64 lun, u8 type, u16 smid_task, u16 msix_task,
2767 u8 timeout, u8 tr_method)
2768 {
2769 int ret;
2770
2771 mutex_lock(&ioc->tm_cmds.mutex);
2772 ret = mpt3sas_scsih_issue_tm(ioc, handle, lun, type, smid_task,
2773 msix_task, timeout, tr_method);
2774 mutex_unlock(&ioc->tm_cmds.mutex);
2775
2776 return ret;
2777 }
2778
2779 /**
2780 * _scsih_tm_display_info - displays info about the device
2781 * @ioc: per adapter struct
2782 * @scmd: pointer to scsi command object
2783 *
2784 * Called by task management callback handlers.
2785 */
2786 static void
_scsih_tm_display_info(struct MPT3SAS_ADAPTER * ioc,struct scsi_cmnd * scmd)2787 _scsih_tm_display_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd)
2788 {
2789 struct scsi_target *starget = scmd->device->sdev_target;
2790 struct MPT3SAS_TARGET *priv_target = starget->hostdata;
2791 struct _sas_device *sas_device = NULL;
2792 struct _pcie_device *pcie_device = NULL;
2793 unsigned long flags;
2794 char *device_str = NULL;
2795
2796 if (!priv_target)
2797 return;
2798 if (ioc->hide_ir_msg)
2799 device_str = "WarpDrive";
2800 else
2801 device_str = "volume";
2802
2803 scsi_print_command(scmd);
2804 if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) {
2805 starget_printk(KERN_INFO, starget,
2806 "%s handle(0x%04x), %s wwid(0x%016llx)\n",
2807 device_str, priv_target->handle,
2808 device_str, (unsigned long long)priv_target->sas_address);
2809
2810 } else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) {
2811 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
2812 pcie_device = __mpt3sas_get_pdev_from_target(ioc, priv_target);
2813 if (pcie_device) {
2814 starget_printk(KERN_INFO, starget,
2815 "handle(0x%04x), wwid(0x%016llx), port(%d)\n",
2816 pcie_device->handle,
2817 (unsigned long long)pcie_device->wwid,
2818 pcie_device->port_num);
2819 if (pcie_device->enclosure_handle != 0)
2820 starget_printk(KERN_INFO, starget,
2821 "enclosure logical id(0x%016llx), slot(%d)\n",
2822 (unsigned long long)
2823 pcie_device->enclosure_logical_id,
2824 pcie_device->slot);
2825 if (pcie_device->connector_name[0] != '\0')
2826 starget_printk(KERN_INFO, starget,
2827 "enclosure level(0x%04x), connector name( %s)\n",
2828 pcie_device->enclosure_level,
2829 pcie_device->connector_name);
2830 pcie_device_put(pcie_device);
2831 }
2832 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
2833
2834 } else {
2835 spin_lock_irqsave(&ioc->sas_device_lock, flags);
2836 sas_device = __mpt3sas_get_sdev_from_target(ioc, priv_target);
2837 if (sas_device) {
2838 if (priv_target->flags &
2839 MPT_TARGET_FLAGS_RAID_COMPONENT) {
2840 starget_printk(KERN_INFO, starget,
2841 "volume handle(0x%04x), "
2842 "volume wwid(0x%016llx)\n",
2843 sas_device->volume_handle,
2844 (unsigned long long)sas_device->volume_wwid);
2845 }
2846 starget_printk(KERN_INFO, starget,
2847 "handle(0x%04x), sas_address(0x%016llx), phy(%d)\n",
2848 sas_device->handle,
2849 (unsigned long long)sas_device->sas_address,
2850 sas_device->phy);
2851
2852 _scsih_display_enclosure_chassis_info(NULL, sas_device,
2853 NULL, starget);
2854
2855 sas_device_put(sas_device);
2856 }
2857 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
2858 }
2859 }
2860
2861 /**
2862 * scsih_abort - eh threads main abort routine
2863 * @scmd: pointer to scsi command object
2864 *
2865 * Return: SUCCESS if command aborted else FAILED
2866 */
2867 static int
scsih_abort(struct scsi_cmnd * scmd)2868 scsih_abort(struct scsi_cmnd *scmd)
2869 {
2870 struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host);
2871 struct MPT3SAS_DEVICE *sas_device_priv_data;
2872 struct scsiio_tracker *st = scsi_cmd_priv(scmd);
2873 u16 handle;
2874 int r;
2875
2876 u8 timeout = 30;
2877 struct _pcie_device *pcie_device = NULL;
2878 sdev_printk(KERN_INFO, scmd->device,
2879 "attempting task abort! scmd(%p)\n", scmd);
2880 _scsih_tm_display_info(ioc, scmd);
2881
2882 sas_device_priv_data = scmd->device->hostdata;
2883 if (!sas_device_priv_data || !sas_device_priv_data->sas_target ||
2884 ioc->remove_host) {
2885 sdev_printk(KERN_INFO, scmd->device,
2886 "device been deleted! scmd(%p)\n", scmd);
2887 scmd->result = DID_NO_CONNECT << 16;
2888 scmd->scsi_done(scmd);
2889 r = SUCCESS;
2890 goto out;
2891 }
2892
2893 /* check for completed command */
2894 if (st == NULL || st->cb_idx == 0xFF) {
2895 scmd->result = DID_RESET << 16;
2896 r = SUCCESS;
2897 goto out;
2898 }
2899
2900 /* for hidden raid components and volumes this is not supported */
2901 if (sas_device_priv_data->sas_target->flags &
2902 MPT_TARGET_FLAGS_RAID_COMPONENT ||
2903 sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_VOLUME) {
2904 scmd->result = DID_RESET << 16;
2905 r = FAILED;
2906 goto out;
2907 }
2908
2909 mpt3sas_halt_firmware(ioc);
2910
2911 handle = sas_device_priv_data->sas_target->handle;
2912 pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle);
2913 if (pcie_device && (!ioc->tm_custom_handling) &&
2914 (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info))))
2915 timeout = ioc->nvme_abort_timeout;
2916 r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->lun,
2917 MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK,
2918 st->smid, st->msix_io, timeout, 0);
2919 /* Command must be cleared after abort */
2920 if (r == SUCCESS && st->cb_idx != 0xFF)
2921 r = FAILED;
2922 out:
2923 sdev_printk(KERN_INFO, scmd->device, "task abort: %s scmd(%p)\n",
2924 ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd);
2925 if (pcie_device)
2926 pcie_device_put(pcie_device);
2927 return r;
2928 }
2929
2930 /**
2931 * scsih_dev_reset - eh threads main device reset routine
2932 * @scmd: pointer to scsi command object
2933 *
2934 * Return: SUCCESS if command aborted else FAILED
2935 */
2936 static int
scsih_dev_reset(struct scsi_cmnd * scmd)2937 scsih_dev_reset(struct scsi_cmnd *scmd)
2938 {
2939 struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host);
2940 struct MPT3SAS_DEVICE *sas_device_priv_data;
2941 struct _sas_device *sas_device = NULL;
2942 struct _pcie_device *pcie_device = NULL;
2943 u16 handle;
2944 u8 tr_method = 0;
2945 u8 tr_timeout = 30;
2946 int r;
2947
2948 struct scsi_target *starget = scmd->device->sdev_target;
2949 struct MPT3SAS_TARGET *target_priv_data = starget->hostdata;
2950
2951 sdev_printk(KERN_INFO, scmd->device,
2952 "attempting device reset! scmd(%p)\n", scmd);
2953 _scsih_tm_display_info(ioc, scmd);
2954
2955 sas_device_priv_data = scmd->device->hostdata;
2956 if (!sas_device_priv_data || !sas_device_priv_data->sas_target ||
2957 ioc->remove_host) {
2958 sdev_printk(KERN_INFO, scmd->device,
2959 "device been deleted! scmd(%p)\n", scmd);
2960 scmd->result = DID_NO_CONNECT << 16;
2961 scmd->scsi_done(scmd);
2962 r = SUCCESS;
2963 goto out;
2964 }
2965
2966 /* for hidden raid components obtain the volume_handle */
2967 handle = 0;
2968 if (sas_device_priv_data->sas_target->flags &
2969 MPT_TARGET_FLAGS_RAID_COMPONENT) {
2970 sas_device = mpt3sas_get_sdev_from_target(ioc,
2971 target_priv_data);
2972 if (sas_device)
2973 handle = sas_device->volume_handle;
2974 } else
2975 handle = sas_device_priv_data->sas_target->handle;
2976
2977 if (!handle) {
2978 scmd->result = DID_RESET << 16;
2979 r = FAILED;
2980 goto out;
2981 }
2982
2983 pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle);
2984
2985 if (pcie_device && (!ioc->tm_custom_handling) &&
2986 (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) {
2987 tr_timeout = pcie_device->reset_timeout;
2988 tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE;
2989 } else
2990 tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
2991
2992 r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->lun,
2993 MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET, 0, 0,
2994 tr_timeout, tr_method);
2995 /* Check for busy commands after reset */
2996 if (r == SUCCESS && atomic_read(&scmd->device->device_busy))
2997 r = FAILED;
2998 out:
2999 sdev_printk(KERN_INFO, scmd->device, "device reset: %s scmd(%p)\n",
3000 ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd);
3001
3002 if (sas_device)
3003 sas_device_put(sas_device);
3004 if (pcie_device)
3005 pcie_device_put(pcie_device);
3006
3007 return r;
3008 }
3009
3010 /**
3011 * scsih_target_reset - eh threads main target reset routine
3012 * @scmd: pointer to scsi command object
3013 *
3014 * Return: SUCCESS if command aborted else FAILED
3015 */
3016 static int
scsih_target_reset(struct scsi_cmnd * scmd)3017 scsih_target_reset(struct scsi_cmnd *scmd)
3018 {
3019 struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host);
3020 struct MPT3SAS_DEVICE *sas_device_priv_data;
3021 struct _sas_device *sas_device = NULL;
3022 struct _pcie_device *pcie_device = NULL;
3023 u16 handle;
3024 u8 tr_method = 0;
3025 u8 tr_timeout = 30;
3026 int r;
3027 struct scsi_target *starget = scmd->device->sdev_target;
3028 struct MPT3SAS_TARGET *target_priv_data = starget->hostdata;
3029
3030 starget_printk(KERN_INFO, starget, "attempting target reset! scmd(%p)\n",
3031 scmd);
3032 _scsih_tm_display_info(ioc, scmd);
3033
3034 sas_device_priv_data = scmd->device->hostdata;
3035 if (!sas_device_priv_data || !sas_device_priv_data->sas_target ||
3036 ioc->remove_host) {
3037 starget_printk(KERN_INFO, starget, "target been deleted! scmd(%p)\n",
3038 scmd);
3039 scmd->result = DID_NO_CONNECT << 16;
3040 scmd->scsi_done(scmd);
3041 r = SUCCESS;
3042 goto out;
3043 }
3044
3045 /* for hidden raid components obtain the volume_handle */
3046 handle = 0;
3047 if (sas_device_priv_data->sas_target->flags &
3048 MPT_TARGET_FLAGS_RAID_COMPONENT) {
3049 sas_device = mpt3sas_get_sdev_from_target(ioc,
3050 target_priv_data);
3051 if (sas_device)
3052 handle = sas_device->volume_handle;
3053 } else
3054 handle = sas_device_priv_data->sas_target->handle;
3055
3056 if (!handle) {
3057 scmd->result = DID_RESET << 16;
3058 r = FAILED;
3059 goto out;
3060 }
3061
3062 pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle);
3063
3064 if (pcie_device && (!ioc->tm_custom_handling) &&
3065 (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) {
3066 tr_timeout = pcie_device->reset_timeout;
3067 tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE;
3068 } else
3069 tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
3070 r = mpt3sas_scsih_issue_locked_tm(ioc, handle, 0,
3071 MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET, 0, 0,
3072 tr_timeout, tr_method);
3073 /* Check for busy commands after reset */
3074 if (r == SUCCESS && atomic_read(&starget->target_busy))
3075 r = FAILED;
3076 out:
3077 starget_printk(KERN_INFO, starget, "target reset: %s scmd(%p)\n",
3078 ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd);
3079
3080 if (sas_device)
3081 sas_device_put(sas_device);
3082 if (pcie_device)
3083 pcie_device_put(pcie_device);
3084 return r;
3085 }
3086
3087
3088 /**
3089 * scsih_host_reset - eh threads main host reset routine
3090 * @scmd: pointer to scsi command object
3091 *
3092 * Return: SUCCESS if command aborted else FAILED
3093 */
3094 static int
scsih_host_reset(struct scsi_cmnd * scmd)3095 scsih_host_reset(struct scsi_cmnd *scmd)
3096 {
3097 struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host);
3098 int r, retval;
3099
3100 ioc_info(ioc, "attempting host reset! scmd(%p)\n", scmd);
3101 scsi_print_command(scmd);
3102
3103 if (ioc->is_driver_loading || ioc->remove_host) {
3104 ioc_info(ioc, "Blocking the host reset\n");
3105 r = FAILED;
3106 goto out;
3107 }
3108
3109 retval = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
3110 r = (retval < 0) ? FAILED : SUCCESS;
3111 out:
3112 ioc_info(ioc, "host reset: %s scmd(%p)\n",
3113 r == SUCCESS ? "SUCCESS" : "FAILED", scmd);
3114
3115 return r;
3116 }
3117
3118 /**
3119 * _scsih_fw_event_add - insert and queue up fw_event
3120 * @ioc: per adapter object
3121 * @fw_event: object describing the event
3122 * Context: This function will acquire ioc->fw_event_lock.
3123 *
3124 * This adds the firmware event object into link list, then queues it up to
3125 * be processed from user context.
3126 */
3127 static void
_scsih_fw_event_add(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)3128 _scsih_fw_event_add(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event)
3129 {
3130 unsigned long flags;
3131
3132 if (ioc->firmware_event_thread == NULL)
3133 return;
3134
3135 spin_lock_irqsave(&ioc->fw_event_lock, flags);
3136 fw_event_work_get(fw_event);
3137 INIT_LIST_HEAD(&fw_event->list);
3138 list_add_tail(&fw_event->list, &ioc->fw_event_list);
3139 INIT_WORK(&fw_event->work, _firmware_event_work);
3140 fw_event_work_get(fw_event);
3141 queue_work(ioc->firmware_event_thread, &fw_event->work);
3142 spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
3143 }
3144
3145 /**
3146 * _scsih_fw_event_del_from_list - delete fw_event from the list
3147 * @ioc: per adapter object
3148 * @fw_event: object describing the event
3149 * Context: This function will acquire ioc->fw_event_lock.
3150 *
3151 * If the fw_event is on the fw_event_list, remove it and do a put.
3152 */
3153 static void
_scsih_fw_event_del_from_list(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)3154 _scsih_fw_event_del_from_list(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work
3155 *fw_event)
3156 {
3157 unsigned long flags;
3158
3159 spin_lock_irqsave(&ioc->fw_event_lock, flags);
3160 if (!list_empty(&fw_event->list)) {
3161 list_del_init(&fw_event->list);
3162 fw_event_work_put(fw_event);
3163 }
3164 spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
3165 }
3166
3167
3168 /**
3169 * mpt3sas_send_trigger_data_event - send event for processing trigger data
3170 * @ioc: per adapter object
3171 * @event_data: trigger event data
3172 */
3173 void
mpt3sas_send_trigger_data_event(struct MPT3SAS_ADAPTER * ioc,struct SL_WH_TRIGGERS_EVENT_DATA_T * event_data)3174 mpt3sas_send_trigger_data_event(struct MPT3SAS_ADAPTER *ioc,
3175 struct SL_WH_TRIGGERS_EVENT_DATA_T *event_data)
3176 {
3177 struct fw_event_work *fw_event;
3178 u16 sz;
3179
3180 if (ioc->is_driver_loading)
3181 return;
3182 sz = sizeof(*event_data);
3183 fw_event = alloc_fw_event_work(sz);
3184 if (!fw_event)
3185 return;
3186 fw_event->event = MPT3SAS_PROCESS_TRIGGER_DIAG;
3187 fw_event->ioc = ioc;
3188 memcpy(fw_event->event_data, event_data, sizeof(*event_data));
3189 _scsih_fw_event_add(ioc, fw_event);
3190 fw_event_work_put(fw_event);
3191 }
3192
3193 /**
3194 * _scsih_error_recovery_delete_devices - remove devices not responding
3195 * @ioc: per adapter object
3196 */
3197 static void
_scsih_error_recovery_delete_devices(struct MPT3SAS_ADAPTER * ioc)3198 _scsih_error_recovery_delete_devices(struct MPT3SAS_ADAPTER *ioc)
3199 {
3200 struct fw_event_work *fw_event;
3201
3202 if (ioc->is_driver_loading)
3203 return;
3204 fw_event = alloc_fw_event_work(0);
3205 if (!fw_event)
3206 return;
3207 fw_event->event = MPT3SAS_REMOVE_UNRESPONDING_DEVICES;
3208 fw_event->ioc = ioc;
3209 _scsih_fw_event_add(ioc, fw_event);
3210 fw_event_work_put(fw_event);
3211 }
3212
3213 /**
3214 * mpt3sas_port_enable_complete - port enable completed (fake event)
3215 * @ioc: per adapter object
3216 */
3217 void
mpt3sas_port_enable_complete(struct MPT3SAS_ADAPTER * ioc)3218 mpt3sas_port_enable_complete(struct MPT3SAS_ADAPTER *ioc)
3219 {
3220 struct fw_event_work *fw_event;
3221
3222 fw_event = alloc_fw_event_work(0);
3223 if (!fw_event)
3224 return;
3225 fw_event->event = MPT3SAS_PORT_ENABLE_COMPLETE;
3226 fw_event->ioc = ioc;
3227 _scsih_fw_event_add(ioc, fw_event);
3228 fw_event_work_put(fw_event);
3229 }
3230
dequeue_next_fw_event(struct MPT3SAS_ADAPTER * ioc)3231 static struct fw_event_work *dequeue_next_fw_event(struct MPT3SAS_ADAPTER *ioc)
3232 {
3233 unsigned long flags;
3234 struct fw_event_work *fw_event = NULL;
3235
3236 spin_lock_irqsave(&ioc->fw_event_lock, flags);
3237 if (!list_empty(&ioc->fw_event_list)) {
3238 fw_event = list_first_entry(&ioc->fw_event_list,
3239 struct fw_event_work, list);
3240 list_del_init(&fw_event->list);
3241 }
3242 spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
3243
3244 return fw_event;
3245 }
3246
3247 /**
3248 * _scsih_fw_event_cleanup_queue - cleanup event queue
3249 * @ioc: per adapter object
3250 *
3251 * Walk the firmware event queue, either killing timers, or waiting
3252 * for outstanding events to complete
3253 */
3254 static void
_scsih_fw_event_cleanup_queue(struct MPT3SAS_ADAPTER * ioc)3255 _scsih_fw_event_cleanup_queue(struct MPT3SAS_ADAPTER *ioc)
3256 {
3257 struct fw_event_work *fw_event;
3258
3259 if (list_empty(&ioc->fw_event_list) ||
3260 !ioc->firmware_event_thread || in_interrupt())
3261 return;
3262
3263 while ((fw_event = dequeue_next_fw_event(ioc))) {
3264 /*
3265 * Wait on the fw_event to complete. If this returns 1, then
3266 * the event was never executed, and we need a put for the
3267 * reference the work had on the fw_event.
3268 *
3269 * If it did execute, we wait for it to finish, and the put will
3270 * happen from _firmware_event_work()
3271 */
3272 if (cancel_work_sync(&fw_event->work))
3273 fw_event_work_put(fw_event);
3274
3275 fw_event_work_put(fw_event);
3276 }
3277 }
3278
3279 /**
3280 * _scsih_internal_device_block - block the sdev device
3281 * @sdev: per device object
3282 * @sas_device_priv_data : per device driver private data
3283 *
3284 * make sure device is blocked without error, if not
3285 * print an error
3286 */
3287 static void
_scsih_internal_device_block(struct scsi_device * sdev,struct MPT3SAS_DEVICE * sas_device_priv_data)3288 _scsih_internal_device_block(struct scsi_device *sdev,
3289 struct MPT3SAS_DEVICE *sas_device_priv_data)
3290 {
3291 int r = 0;
3292
3293 sdev_printk(KERN_INFO, sdev, "device_block, handle(0x%04x)\n",
3294 sas_device_priv_data->sas_target->handle);
3295 sas_device_priv_data->block = 1;
3296
3297 r = scsi_internal_device_block_nowait(sdev);
3298 if (r == -EINVAL)
3299 sdev_printk(KERN_WARNING, sdev,
3300 "device_block failed with return(%d) for handle(0x%04x)\n",
3301 r, sas_device_priv_data->sas_target->handle);
3302 }
3303
3304 /**
3305 * _scsih_internal_device_unblock - unblock the sdev device
3306 * @sdev: per device object
3307 * @sas_device_priv_data : per device driver private data
3308 * make sure device is unblocked without error, if not retry
3309 * by blocking and then unblocking
3310 */
3311
3312 static void
_scsih_internal_device_unblock(struct scsi_device * sdev,struct MPT3SAS_DEVICE * sas_device_priv_data)3313 _scsih_internal_device_unblock(struct scsi_device *sdev,
3314 struct MPT3SAS_DEVICE *sas_device_priv_data)
3315 {
3316 int r = 0;
3317
3318 sdev_printk(KERN_WARNING, sdev, "device_unblock and setting to running, "
3319 "handle(0x%04x)\n", sas_device_priv_data->sas_target->handle);
3320 sas_device_priv_data->block = 0;
3321 r = scsi_internal_device_unblock_nowait(sdev, SDEV_RUNNING);
3322 if (r == -EINVAL) {
3323 /* The device has been set to SDEV_RUNNING by SD layer during
3324 * device addition but the request queue is still stopped by
3325 * our earlier block call. We need to perform a block again
3326 * to get the device to SDEV_BLOCK and then to SDEV_RUNNING */
3327
3328 sdev_printk(KERN_WARNING, sdev,
3329 "device_unblock failed with return(%d) for handle(0x%04x) "
3330 "performing a block followed by an unblock\n",
3331 r, sas_device_priv_data->sas_target->handle);
3332 sas_device_priv_data->block = 1;
3333 r = scsi_internal_device_block_nowait(sdev);
3334 if (r)
3335 sdev_printk(KERN_WARNING, sdev, "retried device_block "
3336 "failed with return(%d) for handle(0x%04x)\n",
3337 r, sas_device_priv_data->sas_target->handle);
3338
3339 sas_device_priv_data->block = 0;
3340 r = scsi_internal_device_unblock_nowait(sdev, SDEV_RUNNING);
3341 if (r)
3342 sdev_printk(KERN_WARNING, sdev, "retried device_unblock"
3343 " failed with return(%d) for handle(0x%04x)\n",
3344 r, sas_device_priv_data->sas_target->handle);
3345 }
3346 }
3347
3348 /**
3349 * _scsih_ublock_io_all_device - unblock every device
3350 * @ioc: per adapter object
3351 *
3352 * change the device state from block to running
3353 */
3354 static void
_scsih_ublock_io_all_device(struct MPT3SAS_ADAPTER * ioc)3355 _scsih_ublock_io_all_device(struct MPT3SAS_ADAPTER *ioc)
3356 {
3357 struct MPT3SAS_DEVICE *sas_device_priv_data;
3358 struct scsi_device *sdev;
3359
3360 shost_for_each_device(sdev, ioc->shost) {
3361 sas_device_priv_data = sdev->hostdata;
3362 if (!sas_device_priv_data)
3363 continue;
3364 if (!sas_device_priv_data->block)
3365 continue;
3366
3367 dewtprintk(ioc, sdev_printk(KERN_INFO, sdev,
3368 "device_running, handle(0x%04x)\n",
3369 sas_device_priv_data->sas_target->handle));
3370 _scsih_internal_device_unblock(sdev, sas_device_priv_data);
3371 }
3372 }
3373
3374
3375 /**
3376 * _scsih_ublock_io_device - prepare device to be deleted
3377 * @ioc: per adapter object
3378 * @sas_address: sas address
3379 *
3380 * unblock then put device in offline state
3381 */
3382 static void
_scsih_ublock_io_device(struct MPT3SAS_ADAPTER * ioc,u64 sas_address)3383 _scsih_ublock_io_device(struct MPT3SAS_ADAPTER *ioc, u64 sas_address)
3384 {
3385 struct MPT3SAS_DEVICE *sas_device_priv_data;
3386 struct scsi_device *sdev;
3387
3388 shost_for_each_device(sdev, ioc->shost) {
3389 sas_device_priv_data = sdev->hostdata;
3390 if (!sas_device_priv_data)
3391 continue;
3392 if (sas_device_priv_data->sas_target->sas_address
3393 != sas_address)
3394 continue;
3395 if (sas_device_priv_data->block)
3396 _scsih_internal_device_unblock(sdev,
3397 sas_device_priv_data);
3398 }
3399 }
3400
3401 /**
3402 * _scsih_block_io_all_device - set the device state to SDEV_BLOCK
3403 * @ioc: per adapter object
3404 *
3405 * During device pull we need to appropriately set the sdev state.
3406 */
3407 static void
_scsih_block_io_all_device(struct MPT3SAS_ADAPTER * ioc)3408 _scsih_block_io_all_device(struct MPT3SAS_ADAPTER *ioc)
3409 {
3410 struct MPT3SAS_DEVICE *sas_device_priv_data;
3411 struct scsi_device *sdev;
3412
3413 shost_for_each_device(sdev, ioc->shost) {
3414 sas_device_priv_data = sdev->hostdata;
3415 if (!sas_device_priv_data)
3416 continue;
3417 if (sas_device_priv_data->block)
3418 continue;
3419 if (sas_device_priv_data->ignore_delay_remove) {
3420 sdev_printk(KERN_INFO, sdev,
3421 "%s skip device_block for SES handle(0x%04x)\n",
3422 __func__, sas_device_priv_data->sas_target->handle);
3423 continue;
3424 }
3425 _scsih_internal_device_block(sdev, sas_device_priv_data);
3426 }
3427 }
3428
3429 /**
3430 * _scsih_block_io_device - set the device state to SDEV_BLOCK
3431 * @ioc: per adapter object
3432 * @handle: device handle
3433 *
3434 * During device pull we need to appropriately set the sdev state.
3435 */
3436 static void
_scsih_block_io_device(struct MPT3SAS_ADAPTER * ioc,u16 handle)3437 _scsih_block_io_device(struct MPT3SAS_ADAPTER *ioc, u16 handle)
3438 {
3439 struct MPT3SAS_DEVICE *sas_device_priv_data;
3440 struct scsi_device *sdev;
3441 struct _sas_device *sas_device;
3442
3443 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
3444
3445 shost_for_each_device(sdev, ioc->shost) {
3446 sas_device_priv_data = sdev->hostdata;
3447 if (!sas_device_priv_data)
3448 continue;
3449 if (sas_device_priv_data->sas_target->handle != handle)
3450 continue;
3451 if (sas_device_priv_data->block)
3452 continue;
3453 if (sas_device && sas_device->pend_sas_rphy_add)
3454 continue;
3455 if (sas_device_priv_data->ignore_delay_remove) {
3456 sdev_printk(KERN_INFO, sdev,
3457 "%s skip device_block for SES handle(0x%04x)\n",
3458 __func__, sas_device_priv_data->sas_target->handle);
3459 continue;
3460 }
3461 _scsih_internal_device_block(sdev, sas_device_priv_data);
3462 }
3463
3464 if (sas_device)
3465 sas_device_put(sas_device);
3466 }
3467
3468 /**
3469 * _scsih_block_io_to_children_attached_to_ex
3470 * @ioc: per adapter object
3471 * @sas_expander: the sas_device object
3472 *
3473 * This routine set sdev state to SDEV_BLOCK for all devices
3474 * attached to this expander. This function called when expander is
3475 * pulled.
3476 */
3477 static void
_scsih_block_io_to_children_attached_to_ex(struct MPT3SAS_ADAPTER * ioc,struct _sas_node * sas_expander)3478 _scsih_block_io_to_children_attached_to_ex(struct MPT3SAS_ADAPTER *ioc,
3479 struct _sas_node *sas_expander)
3480 {
3481 struct _sas_port *mpt3sas_port;
3482 struct _sas_device *sas_device;
3483 struct _sas_node *expander_sibling;
3484 unsigned long flags;
3485
3486 if (!sas_expander)
3487 return;
3488
3489 list_for_each_entry(mpt3sas_port,
3490 &sas_expander->sas_port_list, port_list) {
3491 if (mpt3sas_port->remote_identify.device_type ==
3492 SAS_END_DEVICE) {
3493 spin_lock_irqsave(&ioc->sas_device_lock, flags);
3494 sas_device = __mpt3sas_get_sdev_by_addr(ioc,
3495 mpt3sas_port->remote_identify.sas_address);
3496 if (sas_device) {
3497 set_bit(sas_device->handle,
3498 ioc->blocking_handles);
3499 sas_device_put(sas_device);
3500 }
3501 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
3502 }
3503 }
3504
3505 list_for_each_entry(mpt3sas_port,
3506 &sas_expander->sas_port_list, port_list) {
3507
3508 if (mpt3sas_port->remote_identify.device_type ==
3509 SAS_EDGE_EXPANDER_DEVICE ||
3510 mpt3sas_port->remote_identify.device_type ==
3511 SAS_FANOUT_EXPANDER_DEVICE) {
3512 expander_sibling =
3513 mpt3sas_scsih_expander_find_by_sas_address(
3514 ioc, mpt3sas_port->remote_identify.sas_address);
3515 _scsih_block_io_to_children_attached_to_ex(ioc,
3516 expander_sibling);
3517 }
3518 }
3519 }
3520
3521 /**
3522 * _scsih_block_io_to_children_attached_directly
3523 * @ioc: per adapter object
3524 * @event_data: topology change event data
3525 *
3526 * This routine set sdev state to SDEV_BLOCK for all devices
3527 * direct attached during device pull.
3528 */
3529 static void
_scsih_block_io_to_children_attached_directly(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataSasTopologyChangeList_t * event_data)3530 _scsih_block_io_to_children_attached_directly(struct MPT3SAS_ADAPTER *ioc,
3531 Mpi2EventDataSasTopologyChangeList_t *event_data)
3532 {
3533 int i;
3534 u16 handle;
3535 u16 reason_code;
3536
3537 for (i = 0; i < event_data->NumEntries; i++) {
3538 handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle);
3539 if (!handle)
3540 continue;
3541 reason_code = event_data->PHY[i].PhyStatus &
3542 MPI2_EVENT_SAS_TOPO_RC_MASK;
3543 if (reason_code == MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING)
3544 _scsih_block_io_device(ioc, handle);
3545 }
3546 }
3547
3548 /**
3549 * _scsih_block_io_to_pcie_children_attached_directly
3550 * @ioc: per adapter object
3551 * @event_data: topology change event data
3552 *
3553 * This routine set sdev state to SDEV_BLOCK for all devices
3554 * direct attached during device pull/reconnect.
3555 */
3556 static void
_scsih_block_io_to_pcie_children_attached_directly(struct MPT3SAS_ADAPTER * ioc,Mpi26EventDataPCIeTopologyChangeList_t * event_data)3557 _scsih_block_io_to_pcie_children_attached_directly(struct MPT3SAS_ADAPTER *ioc,
3558 Mpi26EventDataPCIeTopologyChangeList_t *event_data)
3559 {
3560 int i;
3561 u16 handle;
3562 u16 reason_code;
3563
3564 for (i = 0; i < event_data->NumEntries; i++) {
3565 handle =
3566 le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle);
3567 if (!handle)
3568 continue;
3569 reason_code = event_data->PortEntry[i].PortStatus;
3570 if (reason_code ==
3571 MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING)
3572 _scsih_block_io_device(ioc, handle);
3573 }
3574 }
3575 /**
3576 * _scsih_tm_tr_send - send task management request
3577 * @ioc: per adapter object
3578 * @handle: device handle
3579 * Context: interrupt time.
3580 *
3581 * This code is to initiate the device removal handshake protocol
3582 * with controller firmware. This function will issue target reset
3583 * using high priority request queue. It will send a sas iounit
3584 * control request (MPI2_SAS_OP_REMOVE_DEVICE) from this completion.
3585 *
3586 * This is designed to send muliple task management request at the same
3587 * time to the fifo. If the fifo is full, we will append the request,
3588 * and process it in a future completion.
3589 */
3590 static void
_scsih_tm_tr_send(struct MPT3SAS_ADAPTER * ioc,u16 handle)3591 _scsih_tm_tr_send(struct MPT3SAS_ADAPTER *ioc, u16 handle)
3592 {
3593 Mpi2SCSITaskManagementRequest_t *mpi_request;
3594 u16 smid;
3595 struct _sas_device *sas_device = NULL;
3596 struct _pcie_device *pcie_device = NULL;
3597 struct MPT3SAS_TARGET *sas_target_priv_data = NULL;
3598 u64 sas_address = 0;
3599 unsigned long flags;
3600 struct _tr_list *delayed_tr;
3601 u32 ioc_state;
3602 u8 tr_method = 0;
3603
3604 if (ioc->pci_error_recovery) {
3605 dewtprintk(ioc,
3606 ioc_info(ioc, "%s: host in pci error recovery: handle(0x%04x)\n",
3607 __func__, handle));
3608 return;
3609 }
3610 ioc_state = mpt3sas_base_get_iocstate(ioc, 1);
3611 if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) {
3612 dewtprintk(ioc,
3613 ioc_info(ioc, "%s: host is not operational: handle(0x%04x)\n",
3614 __func__, handle));
3615 return;
3616 }
3617
3618 /* if PD, then return */
3619 if (test_bit(handle, ioc->pd_handles))
3620 return;
3621
3622 clear_bit(handle, ioc->pend_os_device_add);
3623
3624 spin_lock_irqsave(&ioc->sas_device_lock, flags);
3625 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
3626 if (sas_device && sas_device->starget &&
3627 sas_device->starget->hostdata) {
3628 sas_target_priv_data = sas_device->starget->hostdata;
3629 sas_target_priv_data->deleted = 1;
3630 sas_address = sas_device->sas_address;
3631 }
3632 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
3633 if (!sas_device) {
3634 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
3635 pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle);
3636 if (pcie_device && pcie_device->starget &&
3637 pcie_device->starget->hostdata) {
3638 sas_target_priv_data = pcie_device->starget->hostdata;
3639 sas_target_priv_data->deleted = 1;
3640 sas_address = pcie_device->wwid;
3641 }
3642 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
3643 if (pcie_device && (!ioc->tm_custom_handling) &&
3644 (!(mpt3sas_scsih_is_pcie_scsi_device(
3645 pcie_device->device_info))))
3646 tr_method =
3647 MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE;
3648 else
3649 tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
3650 }
3651 if (sas_target_priv_data) {
3652 dewtprintk(ioc,
3653 ioc_info(ioc, "setting delete flag: handle(0x%04x), sas_addr(0x%016llx)\n",
3654 handle, (u64)sas_address));
3655 if (sas_device) {
3656 if (sas_device->enclosure_handle != 0)
3657 dewtprintk(ioc,
3658 ioc_info(ioc, "setting delete flag:enclosure logical id(0x%016llx), slot(%d)\n",
3659 (u64)sas_device->enclosure_logical_id,
3660 sas_device->slot));
3661 if (sas_device->connector_name[0] != '\0')
3662 dewtprintk(ioc,
3663 ioc_info(ioc, "setting delete flag: enclosure level(0x%04x), connector name( %s)\n",
3664 sas_device->enclosure_level,
3665 sas_device->connector_name));
3666 } else if (pcie_device) {
3667 if (pcie_device->enclosure_handle != 0)
3668 dewtprintk(ioc,
3669 ioc_info(ioc, "setting delete flag: logical id(0x%016llx), slot(%d)\n",
3670 (u64)pcie_device->enclosure_logical_id,
3671 pcie_device->slot));
3672 if (pcie_device->connector_name[0] != '\0')
3673 dewtprintk(ioc,
3674 ioc_info(ioc, "setting delete flag:, enclosure level(0x%04x), connector name( %s)\n",
3675 pcie_device->enclosure_level,
3676 pcie_device->connector_name));
3677 }
3678 _scsih_ublock_io_device(ioc, sas_address);
3679 sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE;
3680 }
3681
3682 smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_tr_cb_idx);
3683 if (!smid) {
3684 delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC);
3685 if (!delayed_tr)
3686 goto out;
3687 INIT_LIST_HEAD(&delayed_tr->list);
3688 delayed_tr->handle = handle;
3689 list_add_tail(&delayed_tr->list, &ioc->delayed_tr_list);
3690 dewtprintk(ioc,
3691 ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n",
3692 handle));
3693 goto out;
3694 }
3695
3696 dewtprintk(ioc,
3697 ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n",
3698 handle, smid, ioc->tm_tr_cb_idx));
3699 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
3700 memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t));
3701 mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
3702 mpi_request->DevHandle = cpu_to_le16(handle);
3703 mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET;
3704 mpi_request->MsgFlags = tr_method;
3705 set_bit(handle, ioc->device_remove_in_progress);
3706 ioc->put_smid_hi_priority(ioc, smid, 0);
3707 mpt3sas_trigger_master(ioc, MASTER_TRIGGER_DEVICE_REMOVAL);
3708
3709 out:
3710 if (sas_device)
3711 sas_device_put(sas_device);
3712 if (pcie_device)
3713 pcie_device_put(pcie_device);
3714 }
3715
3716 /**
3717 * _scsih_tm_tr_complete -
3718 * @ioc: per adapter object
3719 * @smid: system request message index
3720 * @msix_index: MSIX table index supplied by the OS
3721 * @reply: reply message frame(lower 32bit addr)
3722 * Context: interrupt time.
3723 *
3724 * This is the target reset completion routine.
3725 * This code is part of the code to initiate the device removal
3726 * handshake protocol with controller firmware.
3727 * It will send a sas iounit control request (MPI2_SAS_OP_REMOVE_DEVICE)
3728 *
3729 * Return: 1 meaning mf should be freed from _base_interrupt
3730 * 0 means the mf is freed from this function.
3731 */
3732 static u8
_scsih_tm_tr_complete(struct MPT3SAS_ADAPTER * ioc,u16 smid,u8 msix_index,u32 reply)3733 _scsih_tm_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index,
3734 u32 reply)
3735 {
3736 u16 handle;
3737 Mpi2SCSITaskManagementRequest_t *mpi_request_tm;
3738 Mpi2SCSITaskManagementReply_t *mpi_reply =
3739 mpt3sas_base_get_reply_virt_addr(ioc, reply);
3740 Mpi2SasIoUnitControlRequest_t *mpi_request;
3741 u16 smid_sas_ctrl;
3742 u32 ioc_state;
3743 struct _sc_list *delayed_sc;
3744
3745 if (ioc->pci_error_recovery) {
3746 dewtprintk(ioc,
3747 ioc_info(ioc, "%s: host in pci error recovery\n",
3748 __func__));
3749 return 1;
3750 }
3751 ioc_state = mpt3sas_base_get_iocstate(ioc, 1);
3752 if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) {
3753 dewtprintk(ioc,
3754 ioc_info(ioc, "%s: host is not operational\n",
3755 __func__));
3756 return 1;
3757 }
3758 if (unlikely(!mpi_reply)) {
3759 ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n",
3760 __FILE__, __LINE__, __func__);
3761 return 1;
3762 }
3763 mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid);
3764 handle = le16_to_cpu(mpi_request_tm->DevHandle);
3765 if (handle != le16_to_cpu(mpi_reply->DevHandle)) {
3766 dewtprintk(ioc,
3767 ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n",
3768 handle,
3769 le16_to_cpu(mpi_reply->DevHandle), smid));
3770 return 0;
3771 }
3772
3773 mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT);
3774 dewtprintk(ioc,
3775 ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n",
3776 handle, smid, le16_to_cpu(mpi_reply->IOCStatus),
3777 le32_to_cpu(mpi_reply->IOCLogInfo),
3778 le32_to_cpu(mpi_reply->TerminationCount)));
3779
3780 smid_sas_ctrl = mpt3sas_base_get_smid(ioc, ioc->tm_sas_control_cb_idx);
3781 if (!smid_sas_ctrl) {
3782 delayed_sc = kzalloc(sizeof(*delayed_sc), GFP_ATOMIC);
3783 if (!delayed_sc)
3784 return _scsih_check_for_pending_tm(ioc, smid);
3785 INIT_LIST_HEAD(&delayed_sc->list);
3786 delayed_sc->handle = le16_to_cpu(mpi_request_tm->DevHandle);
3787 list_add_tail(&delayed_sc->list, &ioc->delayed_sc_list);
3788 dewtprintk(ioc,
3789 ioc_info(ioc, "DELAYED:sc:handle(0x%04x), (open)\n",
3790 handle));
3791 return _scsih_check_for_pending_tm(ioc, smid);
3792 }
3793
3794 dewtprintk(ioc,
3795 ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n",
3796 handle, smid_sas_ctrl, ioc->tm_sas_control_cb_idx));
3797 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid_sas_ctrl);
3798 memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t));
3799 mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL;
3800 mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE;
3801 mpi_request->DevHandle = mpi_request_tm->DevHandle;
3802 ioc->put_smid_default(ioc, smid_sas_ctrl);
3803
3804 return _scsih_check_for_pending_tm(ioc, smid);
3805 }
3806
3807 /** _scsih_allow_scmd_to_device - check whether scmd needs to
3808 * issue to IOC or not.
3809 * @ioc: per adapter object
3810 * @scmd: pointer to scsi command object
3811 *
3812 * Returns true if scmd can be issued to IOC otherwise returns false.
3813 */
_scsih_allow_scmd_to_device(struct MPT3SAS_ADAPTER * ioc,struct scsi_cmnd * scmd)3814 inline bool _scsih_allow_scmd_to_device(struct MPT3SAS_ADAPTER *ioc,
3815 struct scsi_cmnd *scmd)
3816 {
3817
3818 if (ioc->pci_error_recovery)
3819 return false;
3820
3821 if (ioc->hba_mpi_version_belonged == MPI2_VERSION) {
3822 if (ioc->remove_host)
3823 return false;
3824
3825 return true;
3826 }
3827
3828 if (ioc->remove_host) {
3829
3830 switch (scmd->cmnd[0]) {
3831 case SYNCHRONIZE_CACHE:
3832 case START_STOP:
3833 return true;
3834 default:
3835 return false;
3836 }
3837 }
3838
3839 return true;
3840 }
3841
3842 /**
3843 * _scsih_sas_control_complete - completion routine
3844 * @ioc: per adapter object
3845 * @smid: system request message index
3846 * @msix_index: MSIX table index supplied by the OS
3847 * @reply: reply message frame(lower 32bit addr)
3848 * Context: interrupt time.
3849 *
3850 * This is the sas iounit control completion routine.
3851 * This code is part of the code to initiate the device removal
3852 * handshake protocol with controller firmware.
3853 *
3854 * Return: 1 meaning mf should be freed from _base_interrupt
3855 * 0 means the mf is freed from this function.
3856 */
3857 static u8
_scsih_sas_control_complete(struct MPT3SAS_ADAPTER * ioc,u16 smid,u8 msix_index,u32 reply)3858 _scsih_sas_control_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid,
3859 u8 msix_index, u32 reply)
3860 {
3861 Mpi2SasIoUnitControlReply_t *mpi_reply =
3862 mpt3sas_base_get_reply_virt_addr(ioc, reply);
3863
3864 if (likely(mpi_reply)) {
3865 dewtprintk(ioc,
3866 ioc_info(ioc, "sc_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x)\n",
3867 le16_to_cpu(mpi_reply->DevHandle), smid,
3868 le16_to_cpu(mpi_reply->IOCStatus),
3869 le32_to_cpu(mpi_reply->IOCLogInfo)));
3870 if (le16_to_cpu(mpi_reply->IOCStatus) ==
3871 MPI2_IOCSTATUS_SUCCESS) {
3872 clear_bit(le16_to_cpu(mpi_reply->DevHandle),
3873 ioc->device_remove_in_progress);
3874 }
3875 } else {
3876 ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n",
3877 __FILE__, __LINE__, __func__);
3878 }
3879 return mpt3sas_check_for_pending_internal_cmds(ioc, smid);
3880 }
3881
3882 /**
3883 * _scsih_tm_tr_volume_send - send target reset request for volumes
3884 * @ioc: per adapter object
3885 * @handle: device handle
3886 * Context: interrupt time.
3887 *
3888 * This is designed to send muliple task management request at the same
3889 * time to the fifo. If the fifo is full, we will append the request,
3890 * and process it in a future completion.
3891 */
3892 static void
_scsih_tm_tr_volume_send(struct MPT3SAS_ADAPTER * ioc,u16 handle)3893 _scsih_tm_tr_volume_send(struct MPT3SAS_ADAPTER *ioc, u16 handle)
3894 {
3895 Mpi2SCSITaskManagementRequest_t *mpi_request;
3896 u16 smid;
3897 struct _tr_list *delayed_tr;
3898
3899 if (ioc->pci_error_recovery) {
3900 dewtprintk(ioc,
3901 ioc_info(ioc, "%s: host reset in progress!\n",
3902 __func__));
3903 return;
3904 }
3905
3906 smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_tr_volume_cb_idx);
3907 if (!smid) {
3908 delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC);
3909 if (!delayed_tr)
3910 return;
3911 INIT_LIST_HEAD(&delayed_tr->list);
3912 delayed_tr->handle = handle;
3913 list_add_tail(&delayed_tr->list, &ioc->delayed_tr_volume_list);
3914 dewtprintk(ioc,
3915 ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n",
3916 handle));
3917 return;
3918 }
3919
3920 dewtprintk(ioc,
3921 ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n",
3922 handle, smid, ioc->tm_tr_volume_cb_idx));
3923 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
3924 memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t));
3925 mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
3926 mpi_request->DevHandle = cpu_to_le16(handle);
3927 mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET;
3928 ioc->put_smid_hi_priority(ioc, smid, 0);
3929 }
3930
3931 /**
3932 * _scsih_tm_volume_tr_complete - target reset completion
3933 * @ioc: per adapter object
3934 * @smid: system request message index
3935 * @msix_index: MSIX table index supplied by the OS
3936 * @reply: reply message frame(lower 32bit addr)
3937 * Context: interrupt time.
3938 *
3939 * Return: 1 meaning mf should be freed from _base_interrupt
3940 * 0 means the mf is freed from this function.
3941 */
3942 static u8
_scsih_tm_volume_tr_complete(struct MPT3SAS_ADAPTER * ioc,u16 smid,u8 msix_index,u32 reply)3943 _scsih_tm_volume_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid,
3944 u8 msix_index, u32 reply)
3945 {
3946 u16 handle;
3947 Mpi2SCSITaskManagementRequest_t *mpi_request_tm;
3948 Mpi2SCSITaskManagementReply_t *mpi_reply =
3949 mpt3sas_base_get_reply_virt_addr(ioc, reply);
3950
3951 if (ioc->shost_recovery || ioc->pci_error_recovery) {
3952 dewtprintk(ioc,
3953 ioc_info(ioc, "%s: host reset in progress!\n",
3954 __func__));
3955 return 1;
3956 }
3957 if (unlikely(!mpi_reply)) {
3958 ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n",
3959 __FILE__, __LINE__, __func__);
3960 return 1;
3961 }
3962
3963 mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid);
3964 handle = le16_to_cpu(mpi_request_tm->DevHandle);
3965 if (handle != le16_to_cpu(mpi_reply->DevHandle)) {
3966 dewtprintk(ioc,
3967 ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n",
3968 handle, le16_to_cpu(mpi_reply->DevHandle),
3969 smid));
3970 return 0;
3971 }
3972
3973 dewtprintk(ioc,
3974 ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n",
3975 handle, smid, le16_to_cpu(mpi_reply->IOCStatus),
3976 le32_to_cpu(mpi_reply->IOCLogInfo),
3977 le32_to_cpu(mpi_reply->TerminationCount)));
3978
3979 return _scsih_check_for_pending_tm(ioc, smid);
3980 }
3981
3982 /**
3983 * _scsih_issue_delayed_event_ack - issue delayed Event ACK messages
3984 * @ioc: per adapter object
3985 * @smid: system request message index
3986 * @event: Event ID
3987 * @event_context: used to track events uniquely
3988 *
3989 * Context - processed in interrupt context.
3990 */
3991 static void
_scsih_issue_delayed_event_ack(struct MPT3SAS_ADAPTER * ioc,u16 smid,U16 event,U32 event_context)3992 _scsih_issue_delayed_event_ack(struct MPT3SAS_ADAPTER *ioc, u16 smid, U16 event,
3993 U32 event_context)
3994 {
3995 Mpi2EventAckRequest_t *ack_request;
3996 int i = smid - ioc->internal_smid;
3997 unsigned long flags;
3998
3999 /* Without releasing the smid just update the
4000 * call back index and reuse the same smid for
4001 * processing this delayed request
4002 */
4003 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
4004 ioc->internal_lookup[i].cb_idx = ioc->base_cb_idx;
4005 spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags);
4006
4007 dewtprintk(ioc,
4008 ioc_info(ioc, "EVENT ACK: event(0x%04x), smid(%d), cb(%d)\n",
4009 le16_to_cpu(event), smid, ioc->base_cb_idx));
4010 ack_request = mpt3sas_base_get_msg_frame(ioc, smid);
4011 memset(ack_request, 0, sizeof(Mpi2EventAckRequest_t));
4012 ack_request->Function = MPI2_FUNCTION_EVENT_ACK;
4013 ack_request->Event = event;
4014 ack_request->EventContext = event_context;
4015 ack_request->VF_ID = 0; /* TODO */
4016 ack_request->VP_ID = 0;
4017 ioc->put_smid_default(ioc, smid);
4018 }
4019
4020 /**
4021 * _scsih_issue_delayed_sas_io_unit_ctrl - issue delayed
4022 * sas_io_unit_ctrl messages
4023 * @ioc: per adapter object
4024 * @smid: system request message index
4025 * @handle: device handle
4026 *
4027 * Context - processed in interrupt context.
4028 */
4029 static void
_scsih_issue_delayed_sas_io_unit_ctrl(struct MPT3SAS_ADAPTER * ioc,u16 smid,u16 handle)4030 _scsih_issue_delayed_sas_io_unit_ctrl(struct MPT3SAS_ADAPTER *ioc,
4031 u16 smid, u16 handle)
4032 {
4033 Mpi2SasIoUnitControlRequest_t *mpi_request;
4034 u32 ioc_state;
4035 int i = smid - ioc->internal_smid;
4036 unsigned long flags;
4037
4038 if (ioc->remove_host) {
4039 dewtprintk(ioc,
4040 ioc_info(ioc, "%s: host has been removed\n",
4041 __func__));
4042 return;
4043 } else if (ioc->pci_error_recovery) {
4044 dewtprintk(ioc,
4045 ioc_info(ioc, "%s: host in pci error recovery\n",
4046 __func__));
4047 return;
4048 }
4049 ioc_state = mpt3sas_base_get_iocstate(ioc, 1);
4050 if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) {
4051 dewtprintk(ioc,
4052 ioc_info(ioc, "%s: host is not operational\n",
4053 __func__));
4054 return;
4055 }
4056
4057 /* Without releasing the smid just update the
4058 * call back index and reuse the same smid for
4059 * processing this delayed request
4060 */
4061 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
4062 ioc->internal_lookup[i].cb_idx = ioc->tm_sas_control_cb_idx;
4063 spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags);
4064
4065 dewtprintk(ioc,
4066 ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n",
4067 handle, smid, ioc->tm_sas_control_cb_idx));
4068 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
4069 memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t));
4070 mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL;
4071 mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE;
4072 mpi_request->DevHandle = cpu_to_le16(handle);
4073 ioc->put_smid_default(ioc, smid);
4074 }
4075
4076 /**
4077 * _scsih_check_for_pending_internal_cmds - check for pending internal messages
4078 * @ioc: per adapter object
4079 * @smid: system request message index
4080 *
4081 * Context: Executed in interrupt context
4082 *
4083 * This will check delayed internal messages list, and process the
4084 * next request.
4085 *
4086 * Return: 1 meaning mf should be freed from _base_interrupt
4087 * 0 means the mf is freed from this function.
4088 */
4089 u8
mpt3sas_check_for_pending_internal_cmds(struct MPT3SAS_ADAPTER * ioc,u16 smid)4090 mpt3sas_check_for_pending_internal_cmds(struct MPT3SAS_ADAPTER *ioc, u16 smid)
4091 {
4092 struct _sc_list *delayed_sc;
4093 struct _event_ack_list *delayed_event_ack;
4094
4095 if (!list_empty(&ioc->delayed_event_ack_list)) {
4096 delayed_event_ack = list_entry(ioc->delayed_event_ack_list.next,
4097 struct _event_ack_list, list);
4098 _scsih_issue_delayed_event_ack(ioc, smid,
4099 delayed_event_ack->Event, delayed_event_ack->EventContext);
4100 list_del(&delayed_event_ack->list);
4101 kfree(delayed_event_ack);
4102 return 0;
4103 }
4104
4105 if (!list_empty(&ioc->delayed_sc_list)) {
4106 delayed_sc = list_entry(ioc->delayed_sc_list.next,
4107 struct _sc_list, list);
4108 _scsih_issue_delayed_sas_io_unit_ctrl(ioc, smid,
4109 delayed_sc->handle);
4110 list_del(&delayed_sc->list);
4111 kfree(delayed_sc);
4112 return 0;
4113 }
4114 return 1;
4115 }
4116
4117 /**
4118 * _scsih_check_for_pending_tm - check for pending task management
4119 * @ioc: per adapter object
4120 * @smid: system request message index
4121 *
4122 * This will check delayed target reset list, and feed the
4123 * next reqeust.
4124 *
4125 * Return: 1 meaning mf should be freed from _base_interrupt
4126 * 0 means the mf is freed from this function.
4127 */
4128 static u8
_scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER * ioc,u16 smid)4129 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid)
4130 {
4131 struct _tr_list *delayed_tr;
4132
4133 if (!list_empty(&ioc->delayed_tr_volume_list)) {
4134 delayed_tr = list_entry(ioc->delayed_tr_volume_list.next,
4135 struct _tr_list, list);
4136 mpt3sas_base_free_smid(ioc, smid);
4137 _scsih_tm_tr_volume_send(ioc, delayed_tr->handle);
4138 list_del(&delayed_tr->list);
4139 kfree(delayed_tr);
4140 return 0;
4141 }
4142
4143 if (!list_empty(&ioc->delayed_tr_list)) {
4144 delayed_tr = list_entry(ioc->delayed_tr_list.next,
4145 struct _tr_list, list);
4146 mpt3sas_base_free_smid(ioc, smid);
4147 _scsih_tm_tr_send(ioc, delayed_tr->handle);
4148 list_del(&delayed_tr->list);
4149 kfree(delayed_tr);
4150 return 0;
4151 }
4152
4153 return 1;
4154 }
4155
4156 /**
4157 * _scsih_check_topo_delete_events - sanity check on topo events
4158 * @ioc: per adapter object
4159 * @event_data: the event data payload
4160 *
4161 * This routine added to better handle cable breaker.
4162 *
4163 * This handles the case where driver receives multiple expander
4164 * add and delete events in a single shot. When there is a delete event
4165 * the routine will void any pending add events waiting in the event queue.
4166 */
4167 static void
_scsih_check_topo_delete_events(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataSasTopologyChangeList_t * event_data)4168 _scsih_check_topo_delete_events(struct MPT3SAS_ADAPTER *ioc,
4169 Mpi2EventDataSasTopologyChangeList_t *event_data)
4170 {
4171 struct fw_event_work *fw_event;
4172 Mpi2EventDataSasTopologyChangeList_t *local_event_data;
4173 u16 expander_handle;
4174 struct _sas_node *sas_expander;
4175 unsigned long flags;
4176 int i, reason_code;
4177 u16 handle;
4178
4179 for (i = 0 ; i < event_data->NumEntries; i++) {
4180 handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle);
4181 if (!handle)
4182 continue;
4183 reason_code = event_data->PHY[i].PhyStatus &
4184 MPI2_EVENT_SAS_TOPO_RC_MASK;
4185 if (reason_code == MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING)
4186 _scsih_tm_tr_send(ioc, handle);
4187 }
4188
4189 expander_handle = le16_to_cpu(event_data->ExpanderDevHandle);
4190 if (expander_handle < ioc->sas_hba.num_phys) {
4191 _scsih_block_io_to_children_attached_directly(ioc, event_data);
4192 return;
4193 }
4194 if (event_data->ExpStatus ==
4195 MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING) {
4196 /* put expander attached devices into blocking state */
4197 spin_lock_irqsave(&ioc->sas_node_lock, flags);
4198 sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc,
4199 expander_handle);
4200 _scsih_block_io_to_children_attached_to_ex(ioc, sas_expander);
4201 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
4202 do {
4203 handle = find_first_bit(ioc->blocking_handles,
4204 ioc->facts.MaxDevHandle);
4205 if (handle < ioc->facts.MaxDevHandle)
4206 _scsih_block_io_device(ioc, handle);
4207 } while (test_and_clear_bit(handle, ioc->blocking_handles));
4208 } else if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_RESPONDING)
4209 _scsih_block_io_to_children_attached_directly(ioc, event_data);
4210
4211 if (event_data->ExpStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING)
4212 return;
4213
4214 /* mark ignore flag for pending events */
4215 spin_lock_irqsave(&ioc->fw_event_lock, flags);
4216 list_for_each_entry(fw_event, &ioc->fw_event_list, list) {
4217 if (fw_event->event != MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST ||
4218 fw_event->ignore)
4219 continue;
4220 local_event_data = (Mpi2EventDataSasTopologyChangeList_t *)
4221 fw_event->event_data;
4222 if (local_event_data->ExpStatus ==
4223 MPI2_EVENT_SAS_TOPO_ES_ADDED ||
4224 local_event_data->ExpStatus ==
4225 MPI2_EVENT_SAS_TOPO_ES_RESPONDING) {
4226 if (le16_to_cpu(local_event_data->ExpanderDevHandle) ==
4227 expander_handle) {
4228 dewtprintk(ioc,
4229 ioc_info(ioc, "setting ignoring flag\n"));
4230 fw_event->ignore = 1;
4231 }
4232 }
4233 }
4234 spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
4235 }
4236
4237 /**
4238 * _scsih_check_pcie_topo_remove_events - sanity check on topo
4239 * events
4240 * @ioc: per adapter object
4241 * @event_data: the event data payload
4242 *
4243 * This handles the case where driver receives multiple switch
4244 * or device add and delete events in a single shot. When there
4245 * is a delete event the routine will void any pending add
4246 * events waiting in the event queue.
4247 */
4248 static void
_scsih_check_pcie_topo_remove_events(struct MPT3SAS_ADAPTER * ioc,Mpi26EventDataPCIeTopologyChangeList_t * event_data)4249 _scsih_check_pcie_topo_remove_events(struct MPT3SAS_ADAPTER *ioc,
4250 Mpi26EventDataPCIeTopologyChangeList_t *event_data)
4251 {
4252 struct fw_event_work *fw_event;
4253 Mpi26EventDataPCIeTopologyChangeList_t *local_event_data;
4254 unsigned long flags;
4255 int i, reason_code;
4256 u16 handle, switch_handle;
4257
4258 for (i = 0; i < event_data->NumEntries; i++) {
4259 handle =
4260 le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle);
4261 if (!handle)
4262 continue;
4263 reason_code = event_data->PortEntry[i].PortStatus;
4264 if (reason_code == MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING)
4265 _scsih_tm_tr_send(ioc, handle);
4266 }
4267
4268 switch_handle = le16_to_cpu(event_data->SwitchDevHandle);
4269 if (!switch_handle) {
4270 _scsih_block_io_to_pcie_children_attached_directly(
4271 ioc, event_data);
4272 return;
4273 }
4274 /* TODO We are not supporting cascaded PCIe Switch removal yet*/
4275 if ((event_data->SwitchStatus
4276 == MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING) ||
4277 (event_data->SwitchStatus ==
4278 MPI26_EVENT_PCIE_TOPO_SS_RESPONDING))
4279 _scsih_block_io_to_pcie_children_attached_directly(
4280 ioc, event_data);
4281
4282 if (event_data->SwitchStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING)
4283 return;
4284
4285 /* mark ignore flag for pending events */
4286 spin_lock_irqsave(&ioc->fw_event_lock, flags);
4287 list_for_each_entry(fw_event, &ioc->fw_event_list, list) {
4288 if (fw_event->event != MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST ||
4289 fw_event->ignore)
4290 continue;
4291 local_event_data =
4292 (Mpi26EventDataPCIeTopologyChangeList_t *)
4293 fw_event->event_data;
4294 if (local_event_data->SwitchStatus ==
4295 MPI2_EVENT_SAS_TOPO_ES_ADDED ||
4296 local_event_data->SwitchStatus ==
4297 MPI2_EVENT_SAS_TOPO_ES_RESPONDING) {
4298 if (le16_to_cpu(local_event_data->SwitchDevHandle) ==
4299 switch_handle) {
4300 dewtprintk(ioc,
4301 ioc_info(ioc, "setting ignoring flag for switch event\n"));
4302 fw_event->ignore = 1;
4303 }
4304 }
4305 }
4306 spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
4307 }
4308
4309 /**
4310 * _scsih_set_volume_delete_flag - setting volume delete flag
4311 * @ioc: per adapter object
4312 * @handle: device handle
4313 *
4314 * This returns nothing.
4315 */
4316 static void
_scsih_set_volume_delete_flag(struct MPT3SAS_ADAPTER * ioc,u16 handle)4317 _scsih_set_volume_delete_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle)
4318 {
4319 struct _raid_device *raid_device;
4320 struct MPT3SAS_TARGET *sas_target_priv_data;
4321 unsigned long flags;
4322
4323 spin_lock_irqsave(&ioc->raid_device_lock, flags);
4324 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
4325 if (raid_device && raid_device->starget &&
4326 raid_device->starget->hostdata) {
4327 sas_target_priv_data =
4328 raid_device->starget->hostdata;
4329 sas_target_priv_data->deleted = 1;
4330 dewtprintk(ioc,
4331 ioc_info(ioc, "setting delete flag: handle(0x%04x), wwid(0x%016llx)\n",
4332 handle, (u64)raid_device->wwid));
4333 }
4334 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
4335 }
4336
4337 /**
4338 * _scsih_set_volume_handle_for_tr - set handle for target reset to volume
4339 * @handle: input handle
4340 * @a: handle for volume a
4341 * @b: handle for volume b
4342 *
4343 * IR firmware only supports two raid volumes. The purpose of this
4344 * routine is to set the volume handle in either a or b. When the given
4345 * input handle is non-zero, or when a and b have not been set before.
4346 */
4347 static void
_scsih_set_volume_handle_for_tr(u16 handle,u16 * a,u16 * b)4348 _scsih_set_volume_handle_for_tr(u16 handle, u16 *a, u16 *b)
4349 {
4350 if (!handle || handle == *a || handle == *b)
4351 return;
4352 if (!*a)
4353 *a = handle;
4354 else if (!*b)
4355 *b = handle;
4356 }
4357
4358 /**
4359 * _scsih_check_ir_config_unhide_events - check for UNHIDE events
4360 * @ioc: per adapter object
4361 * @event_data: the event data payload
4362 * Context: interrupt time.
4363 *
4364 * This routine will send target reset to volume, followed by target
4365 * resets to the PDs. This is called when a PD has been removed, or
4366 * volume has been deleted or removed. When the target reset is sent
4367 * to volume, the PD target resets need to be queued to start upon
4368 * completion of the volume target reset.
4369 */
4370 static void
_scsih_check_ir_config_unhide_events(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataIrConfigChangeList_t * event_data)4371 _scsih_check_ir_config_unhide_events(struct MPT3SAS_ADAPTER *ioc,
4372 Mpi2EventDataIrConfigChangeList_t *event_data)
4373 {
4374 Mpi2EventIrConfigElement_t *element;
4375 int i;
4376 u16 handle, volume_handle, a, b;
4377 struct _tr_list *delayed_tr;
4378
4379 a = 0;
4380 b = 0;
4381
4382 if (ioc->is_warpdrive)
4383 return;
4384
4385 /* Volume Resets for Deleted or Removed */
4386 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
4387 for (i = 0; i < event_data->NumElements; i++, element++) {
4388 if (le32_to_cpu(event_data->Flags) &
4389 MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG)
4390 continue;
4391 if (element->ReasonCode ==
4392 MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED ||
4393 element->ReasonCode ==
4394 MPI2_EVENT_IR_CHANGE_RC_REMOVED) {
4395 volume_handle = le16_to_cpu(element->VolDevHandle);
4396 _scsih_set_volume_delete_flag(ioc, volume_handle);
4397 _scsih_set_volume_handle_for_tr(volume_handle, &a, &b);
4398 }
4399 }
4400
4401 /* Volume Resets for UNHIDE events */
4402 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
4403 for (i = 0; i < event_data->NumElements; i++, element++) {
4404 if (le32_to_cpu(event_data->Flags) &
4405 MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG)
4406 continue;
4407 if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_UNHIDE) {
4408 volume_handle = le16_to_cpu(element->VolDevHandle);
4409 _scsih_set_volume_handle_for_tr(volume_handle, &a, &b);
4410 }
4411 }
4412
4413 if (a)
4414 _scsih_tm_tr_volume_send(ioc, a);
4415 if (b)
4416 _scsih_tm_tr_volume_send(ioc, b);
4417
4418 /* PD target resets */
4419 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
4420 for (i = 0; i < event_data->NumElements; i++, element++) {
4421 if (element->ReasonCode != MPI2_EVENT_IR_CHANGE_RC_UNHIDE)
4422 continue;
4423 handle = le16_to_cpu(element->PhysDiskDevHandle);
4424 volume_handle = le16_to_cpu(element->VolDevHandle);
4425 clear_bit(handle, ioc->pd_handles);
4426 if (!volume_handle)
4427 _scsih_tm_tr_send(ioc, handle);
4428 else if (volume_handle == a || volume_handle == b) {
4429 delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC);
4430 BUG_ON(!delayed_tr);
4431 INIT_LIST_HEAD(&delayed_tr->list);
4432 delayed_tr->handle = handle;
4433 list_add_tail(&delayed_tr->list, &ioc->delayed_tr_list);
4434 dewtprintk(ioc,
4435 ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n",
4436 handle));
4437 } else
4438 _scsih_tm_tr_send(ioc, handle);
4439 }
4440 }
4441
4442
4443 /**
4444 * _scsih_check_volume_delete_events - set delete flag for volumes
4445 * @ioc: per adapter object
4446 * @event_data: the event data payload
4447 * Context: interrupt time.
4448 *
4449 * This will handle the case when the cable connected to entire volume is
4450 * pulled. We will take care of setting the deleted flag so normal IO will
4451 * not be sent.
4452 */
4453 static void
_scsih_check_volume_delete_events(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataIrVolume_t * event_data)4454 _scsih_check_volume_delete_events(struct MPT3SAS_ADAPTER *ioc,
4455 Mpi2EventDataIrVolume_t *event_data)
4456 {
4457 u32 state;
4458
4459 if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED)
4460 return;
4461 state = le32_to_cpu(event_data->NewValue);
4462 if (state == MPI2_RAID_VOL_STATE_MISSING || state ==
4463 MPI2_RAID_VOL_STATE_FAILED)
4464 _scsih_set_volume_delete_flag(ioc,
4465 le16_to_cpu(event_data->VolDevHandle));
4466 }
4467
4468 /**
4469 * _scsih_temp_threshold_events - display temperature threshold exceeded events
4470 * @ioc: per adapter object
4471 * @event_data: the temp threshold event data
4472 * Context: interrupt time.
4473 */
4474 static void
_scsih_temp_threshold_events(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataTemperature_t * event_data)4475 _scsih_temp_threshold_events(struct MPT3SAS_ADAPTER *ioc,
4476 Mpi2EventDataTemperature_t *event_data)
4477 {
4478 if (ioc->temp_sensors_count >= event_data->SensorNum) {
4479 ioc_err(ioc, "Temperature Threshold flags %s%s%s%s exceeded for Sensor: %d !!!\n",
4480 le16_to_cpu(event_data->Status) & 0x1 ? "0 " : " ",
4481 le16_to_cpu(event_data->Status) & 0x2 ? "1 " : " ",
4482 le16_to_cpu(event_data->Status) & 0x4 ? "2 " : " ",
4483 le16_to_cpu(event_data->Status) & 0x8 ? "3 " : " ",
4484 event_data->SensorNum);
4485 ioc_err(ioc, "Current Temp In Celsius: %d\n",
4486 event_data->CurrentTemperature);
4487 }
4488 }
4489
_scsih_set_satl_pending(struct scsi_cmnd * scmd,bool pending)4490 static int _scsih_set_satl_pending(struct scsi_cmnd *scmd, bool pending)
4491 {
4492 struct MPT3SAS_DEVICE *priv = scmd->device->hostdata;
4493
4494 if (scmd->cmnd[0] != ATA_12 && scmd->cmnd[0] != ATA_16)
4495 return 0;
4496
4497 if (pending)
4498 return test_and_set_bit(0, &priv->ata_command_pending);
4499
4500 clear_bit(0, &priv->ata_command_pending);
4501 return 0;
4502 }
4503
4504 /**
4505 * _scsih_flush_running_cmds - completing outstanding commands.
4506 * @ioc: per adapter object
4507 *
4508 * The flushing out of all pending scmd commands following host reset,
4509 * where all IO is dropped to the floor.
4510 */
4511 static void
_scsih_flush_running_cmds(struct MPT3SAS_ADAPTER * ioc)4512 _scsih_flush_running_cmds(struct MPT3SAS_ADAPTER *ioc)
4513 {
4514 struct scsi_cmnd *scmd;
4515 struct scsiio_tracker *st;
4516 u16 smid;
4517 int count = 0;
4518
4519 for (smid = 1; smid <= ioc->scsiio_depth; smid++) {
4520 scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid);
4521 if (!scmd)
4522 continue;
4523 count++;
4524 _scsih_set_satl_pending(scmd, false);
4525 st = scsi_cmd_priv(scmd);
4526 mpt3sas_base_clear_st(ioc, st);
4527 scsi_dma_unmap(scmd);
4528 if (ioc->pci_error_recovery || ioc->remove_host)
4529 scmd->result = DID_NO_CONNECT << 16;
4530 else
4531 scmd->result = DID_RESET << 16;
4532 scmd->scsi_done(scmd);
4533 }
4534 dtmprintk(ioc, ioc_info(ioc, "completing %d cmds\n", count));
4535 }
4536
4537 /**
4538 * _scsih_setup_eedp - setup MPI request for EEDP transfer
4539 * @ioc: per adapter object
4540 * @scmd: pointer to scsi command object
4541 * @mpi_request: pointer to the SCSI_IO request message frame
4542 *
4543 * Supporting protection 1 and 3.
4544 */
4545 static void
_scsih_setup_eedp(struct MPT3SAS_ADAPTER * ioc,struct scsi_cmnd * scmd,Mpi25SCSIIORequest_t * mpi_request)4546 _scsih_setup_eedp(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd,
4547 Mpi25SCSIIORequest_t *mpi_request)
4548 {
4549 u16 eedp_flags;
4550 unsigned char prot_op = scsi_get_prot_op(scmd);
4551 unsigned char prot_type = scsi_get_prot_type(scmd);
4552 Mpi25SCSIIORequest_t *mpi_request_3v =
4553 (Mpi25SCSIIORequest_t *)mpi_request;
4554
4555 if (prot_type == SCSI_PROT_DIF_TYPE0 || prot_op == SCSI_PROT_NORMAL)
4556 return;
4557
4558 if (prot_op == SCSI_PROT_READ_STRIP)
4559 eedp_flags = MPI2_SCSIIO_EEDPFLAGS_CHECK_REMOVE_OP;
4560 else if (prot_op == SCSI_PROT_WRITE_INSERT)
4561 eedp_flags = MPI2_SCSIIO_EEDPFLAGS_INSERT_OP;
4562 else
4563 return;
4564
4565 switch (prot_type) {
4566 case SCSI_PROT_DIF_TYPE1:
4567 case SCSI_PROT_DIF_TYPE2:
4568
4569 /*
4570 * enable ref/guard checking
4571 * auto increment ref tag
4572 */
4573 eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_INC_PRI_REFTAG |
4574 MPI2_SCSIIO_EEDPFLAGS_CHECK_REFTAG |
4575 MPI2_SCSIIO_EEDPFLAGS_CHECK_GUARD;
4576 mpi_request->CDB.EEDP32.PrimaryReferenceTag =
4577 cpu_to_be32(t10_pi_ref_tag(scmd->request));
4578 break;
4579
4580 case SCSI_PROT_DIF_TYPE3:
4581
4582 /*
4583 * enable guard checking
4584 */
4585 eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_CHECK_GUARD;
4586
4587 break;
4588 }
4589
4590 mpi_request_3v->EEDPBlockSize =
4591 cpu_to_le16(scmd->device->sector_size);
4592
4593 if (ioc->is_gen35_ioc)
4594 eedp_flags |= MPI25_SCSIIO_EEDPFLAGS_APPTAG_DISABLE_MODE;
4595 mpi_request->EEDPFlags = cpu_to_le16(eedp_flags);
4596 }
4597
4598 /**
4599 * _scsih_eedp_error_handling - return sense code for EEDP errors
4600 * @scmd: pointer to scsi command object
4601 * @ioc_status: ioc status
4602 */
4603 static void
_scsih_eedp_error_handling(struct scsi_cmnd * scmd,u16 ioc_status)4604 _scsih_eedp_error_handling(struct scsi_cmnd *scmd, u16 ioc_status)
4605 {
4606 u8 ascq;
4607
4608 switch (ioc_status) {
4609 case MPI2_IOCSTATUS_EEDP_GUARD_ERROR:
4610 ascq = 0x01;
4611 break;
4612 case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR:
4613 ascq = 0x02;
4614 break;
4615 case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR:
4616 ascq = 0x03;
4617 break;
4618 default:
4619 ascq = 0x00;
4620 break;
4621 }
4622 scsi_build_sense_buffer(0, scmd->sense_buffer, ILLEGAL_REQUEST, 0x10,
4623 ascq);
4624 scmd->result = DRIVER_SENSE << 24 | (DID_ABORT << 16) |
4625 SAM_STAT_CHECK_CONDITION;
4626 }
4627
4628 /**
4629 * scsih_qcmd - main scsi request entry point
4630 * @shost: SCSI host pointer
4631 * @scmd: pointer to scsi command object
4632 *
4633 * The callback index is set inside `ioc->scsi_io_cb_idx`.
4634 *
4635 * Return: 0 on success. If there's a failure, return either:
4636 * SCSI_MLQUEUE_DEVICE_BUSY if the device queue is full, or
4637 * SCSI_MLQUEUE_HOST_BUSY if the entire host queue is full
4638 */
4639 static int
scsih_qcmd(struct Scsi_Host * shost,struct scsi_cmnd * scmd)4640 scsih_qcmd(struct Scsi_Host *shost, struct scsi_cmnd *scmd)
4641 {
4642 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
4643 struct MPT3SAS_DEVICE *sas_device_priv_data;
4644 struct MPT3SAS_TARGET *sas_target_priv_data;
4645 struct _raid_device *raid_device;
4646 struct request *rq = scmd->request;
4647 int class;
4648 Mpi25SCSIIORequest_t *mpi_request;
4649 struct _pcie_device *pcie_device = NULL;
4650 u32 mpi_control;
4651 u16 smid;
4652 u16 handle;
4653
4654 if (ioc->logging_level & MPT_DEBUG_SCSI)
4655 scsi_print_command(scmd);
4656
4657 sas_device_priv_data = scmd->device->hostdata;
4658 if (!sas_device_priv_data || !sas_device_priv_data->sas_target) {
4659 scmd->result = DID_NO_CONNECT << 16;
4660 scmd->scsi_done(scmd);
4661 return 0;
4662 }
4663
4664 if (!(_scsih_allow_scmd_to_device(ioc, scmd))) {
4665 scmd->result = DID_NO_CONNECT << 16;
4666 scmd->scsi_done(scmd);
4667 return 0;
4668 }
4669
4670 sas_target_priv_data = sas_device_priv_data->sas_target;
4671
4672 /* invalid device handle */
4673 handle = sas_target_priv_data->handle;
4674 if (handle == MPT3SAS_INVALID_DEVICE_HANDLE) {
4675 scmd->result = DID_NO_CONNECT << 16;
4676 scmd->scsi_done(scmd);
4677 return 0;
4678 }
4679
4680
4681 if (ioc->shost_recovery || ioc->ioc_link_reset_in_progress) {
4682 /* host recovery or link resets sent via IOCTLs */
4683 return SCSI_MLQUEUE_HOST_BUSY;
4684 } else if (sas_target_priv_data->deleted) {
4685 /* device has been deleted */
4686 scmd->result = DID_NO_CONNECT << 16;
4687 scmd->scsi_done(scmd);
4688 return 0;
4689 } else if (sas_target_priv_data->tm_busy ||
4690 sas_device_priv_data->block) {
4691 /* device busy with task management */
4692 return SCSI_MLQUEUE_DEVICE_BUSY;
4693 }
4694
4695 /*
4696 * Bug work around for firmware SATL handling. The loop
4697 * is based on atomic operations and ensures consistency
4698 * since we're lockless at this point
4699 */
4700 do {
4701 if (test_bit(0, &sas_device_priv_data->ata_command_pending))
4702 return SCSI_MLQUEUE_DEVICE_BUSY;
4703 } while (_scsih_set_satl_pending(scmd, true));
4704
4705 if (scmd->sc_data_direction == DMA_FROM_DEVICE)
4706 mpi_control = MPI2_SCSIIO_CONTROL_READ;
4707 else if (scmd->sc_data_direction == DMA_TO_DEVICE)
4708 mpi_control = MPI2_SCSIIO_CONTROL_WRITE;
4709 else
4710 mpi_control = MPI2_SCSIIO_CONTROL_NODATATRANSFER;
4711
4712 /* set tags */
4713 mpi_control |= MPI2_SCSIIO_CONTROL_SIMPLEQ;
4714 /* NCQ Prio supported, make sure control indicated high priority */
4715 if (sas_device_priv_data->ncq_prio_enable) {
4716 class = IOPRIO_PRIO_CLASS(req_get_ioprio(rq));
4717 if (class == IOPRIO_CLASS_RT)
4718 mpi_control |= 1 << MPI2_SCSIIO_CONTROL_CMDPRI_SHIFT;
4719 }
4720 /* Make sure Device is not raid volume.
4721 * We do not expose raid functionality to upper layer for warpdrive.
4722 */
4723 if (((!ioc->is_warpdrive && !scsih_is_raid(&scmd->device->sdev_gendev))
4724 && !scsih_is_nvme(&scmd->device->sdev_gendev))
4725 && sas_is_tlr_enabled(scmd->device) && scmd->cmd_len != 32)
4726 mpi_control |= MPI2_SCSIIO_CONTROL_TLR_ON;
4727
4728 smid = mpt3sas_base_get_smid_scsiio(ioc, ioc->scsi_io_cb_idx, scmd);
4729 if (!smid) {
4730 ioc_err(ioc, "%s: failed obtaining a smid\n", __func__);
4731 _scsih_set_satl_pending(scmd, false);
4732 goto out;
4733 }
4734 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
4735 memset(mpi_request, 0, ioc->request_sz);
4736 _scsih_setup_eedp(ioc, scmd, mpi_request);
4737
4738 if (scmd->cmd_len == 32)
4739 mpi_control |= 4 << MPI2_SCSIIO_CONTROL_ADDCDBLEN_SHIFT;
4740 mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST;
4741 if (sas_device_priv_data->sas_target->flags &
4742 MPT_TARGET_FLAGS_RAID_COMPONENT)
4743 mpi_request->Function = MPI2_FUNCTION_RAID_SCSI_IO_PASSTHROUGH;
4744 else
4745 mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST;
4746 mpi_request->DevHandle = cpu_to_le16(handle);
4747 mpi_request->DataLength = cpu_to_le32(scsi_bufflen(scmd));
4748 mpi_request->Control = cpu_to_le32(mpi_control);
4749 mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len);
4750 mpi_request->MsgFlags = MPI2_SCSIIO_MSGFLAGS_SYSTEM_SENSE_ADDR;
4751 mpi_request->SenseBufferLength = SCSI_SENSE_BUFFERSIZE;
4752 mpi_request->SenseBufferLowAddress =
4753 mpt3sas_base_get_sense_buffer_dma(ioc, smid);
4754 mpi_request->SGLOffset0 = offsetof(Mpi25SCSIIORequest_t, SGL) / 4;
4755 int_to_scsilun(sas_device_priv_data->lun, (struct scsi_lun *)
4756 mpi_request->LUN);
4757 memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len);
4758
4759 if (mpi_request->DataLength) {
4760 pcie_device = sas_target_priv_data->pcie_dev;
4761 if (ioc->build_sg_scmd(ioc, scmd, smid, pcie_device)) {
4762 mpt3sas_base_free_smid(ioc, smid);
4763 _scsih_set_satl_pending(scmd, false);
4764 goto out;
4765 }
4766 } else
4767 ioc->build_zero_len_sge(ioc, &mpi_request->SGL);
4768
4769 raid_device = sas_target_priv_data->raid_device;
4770 if (raid_device && raid_device->direct_io_enabled)
4771 mpt3sas_setup_direct_io(ioc, scmd,
4772 raid_device, mpi_request);
4773
4774 if (likely(mpi_request->Function == MPI2_FUNCTION_SCSI_IO_REQUEST)) {
4775 if (sas_target_priv_data->flags & MPT_TARGET_FASTPATH_IO) {
4776 mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len |
4777 MPI25_SCSIIO_IOFLAGS_FAST_PATH);
4778 ioc->put_smid_fast_path(ioc, smid, handle);
4779 } else
4780 ioc->put_smid_scsi_io(ioc, smid,
4781 le16_to_cpu(mpi_request->DevHandle));
4782 } else
4783 ioc->put_smid_default(ioc, smid);
4784 return 0;
4785
4786 out:
4787 return SCSI_MLQUEUE_HOST_BUSY;
4788 }
4789
4790 /**
4791 * _scsih_normalize_sense - normalize descriptor and fixed format sense data
4792 * @sense_buffer: sense data returned by target
4793 * @data: normalized skey/asc/ascq
4794 */
4795 static void
_scsih_normalize_sense(char * sense_buffer,struct sense_info * data)4796 _scsih_normalize_sense(char *sense_buffer, struct sense_info *data)
4797 {
4798 if ((sense_buffer[0] & 0x7F) >= 0x72) {
4799 /* descriptor format */
4800 data->skey = sense_buffer[1] & 0x0F;
4801 data->asc = sense_buffer[2];
4802 data->ascq = sense_buffer[3];
4803 } else {
4804 /* fixed format */
4805 data->skey = sense_buffer[2] & 0x0F;
4806 data->asc = sense_buffer[12];
4807 data->ascq = sense_buffer[13];
4808 }
4809 }
4810
4811 /**
4812 * _scsih_scsi_ioc_info - translated non-succesfull SCSI_IO request
4813 * @ioc: per adapter object
4814 * @scmd: pointer to scsi command object
4815 * @mpi_reply: reply mf payload returned from firmware
4816 * @smid: ?
4817 *
4818 * scsi_status - SCSI Status code returned from target device
4819 * scsi_state - state info associated with SCSI_IO determined by ioc
4820 * ioc_status - ioc supplied status info
4821 */
4822 static void
_scsih_scsi_ioc_info(struct MPT3SAS_ADAPTER * ioc,struct scsi_cmnd * scmd,Mpi2SCSIIOReply_t * mpi_reply,u16 smid)4823 _scsih_scsi_ioc_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd,
4824 Mpi2SCSIIOReply_t *mpi_reply, u16 smid)
4825 {
4826 u32 response_info;
4827 u8 *response_bytes;
4828 u16 ioc_status = le16_to_cpu(mpi_reply->IOCStatus) &
4829 MPI2_IOCSTATUS_MASK;
4830 u8 scsi_state = mpi_reply->SCSIState;
4831 u8 scsi_status = mpi_reply->SCSIStatus;
4832 char *desc_ioc_state = NULL;
4833 char *desc_scsi_status = NULL;
4834 char *desc_scsi_state = ioc->tmp_string;
4835 u32 log_info = le32_to_cpu(mpi_reply->IOCLogInfo);
4836 struct _sas_device *sas_device = NULL;
4837 struct _pcie_device *pcie_device = NULL;
4838 struct scsi_target *starget = scmd->device->sdev_target;
4839 struct MPT3SAS_TARGET *priv_target = starget->hostdata;
4840 char *device_str = NULL;
4841
4842 if (!priv_target)
4843 return;
4844 if (ioc->hide_ir_msg)
4845 device_str = "WarpDrive";
4846 else
4847 device_str = "volume";
4848
4849 if (log_info == 0x31170000)
4850 return;
4851
4852 switch (ioc_status) {
4853 case MPI2_IOCSTATUS_SUCCESS:
4854 desc_ioc_state = "success";
4855 break;
4856 case MPI2_IOCSTATUS_INVALID_FUNCTION:
4857 desc_ioc_state = "invalid function";
4858 break;
4859 case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR:
4860 desc_ioc_state = "scsi recovered error";
4861 break;
4862 case MPI2_IOCSTATUS_SCSI_INVALID_DEVHANDLE:
4863 desc_ioc_state = "scsi invalid dev handle";
4864 break;
4865 case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE:
4866 desc_ioc_state = "scsi device not there";
4867 break;
4868 case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN:
4869 desc_ioc_state = "scsi data overrun";
4870 break;
4871 case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN:
4872 desc_ioc_state = "scsi data underrun";
4873 break;
4874 case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR:
4875 desc_ioc_state = "scsi io data error";
4876 break;
4877 case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR:
4878 desc_ioc_state = "scsi protocol error";
4879 break;
4880 case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED:
4881 desc_ioc_state = "scsi task terminated";
4882 break;
4883 case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH:
4884 desc_ioc_state = "scsi residual mismatch";
4885 break;
4886 case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED:
4887 desc_ioc_state = "scsi task mgmt failed";
4888 break;
4889 case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED:
4890 desc_ioc_state = "scsi ioc terminated";
4891 break;
4892 case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED:
4893 desc_ioc_state = "scsi ext terminated";
4894 break;
4895 case MPI2_IOCSTATUS_EEDP_GUARD_ERROR:
4896 desc_ioc_state = "eedp guard error";
4897 break;
4898 case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR:
4899 desc_ioc_state = "eedp ref tag error";
4900 break;
4901 case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR:
4902 desc_ioc_state = "eedp app tag error";
4903 break;
4904 case MPI2_IOCSTATUS_INSUFFICIENT_POWER:
4905 desc_ioc_state = "insufficient power";
4906 break;
4907 default:
4908 desc_ioc_state = "unknown";
4909 break;
4910 }
4911
4912 switch (scsi_status) {
4913 case MPI2_SCSI_STATUS_GOOD:
4914 desc_scsi_status = "good";
4915 break;
4916 case MPI2_SCSI_STATUS_CHECK_CONDITION:
4917 desc_scsi_status = "check condition";
4918 break;
4919 case MPI2_SCSI_STATUS_CONDITION_MET:
4920 desc_scsi_status = "condition met";
4921 break;
4922 case MPI2_SCSI_STATUS_BUSY:
4923 desc_scsi_status = "busy";
4924 break;
4925 case MPI2_SCSI_STATUS_INTERMEDIATE:
4926 desc_scsi_status = "intermediate";
4927 break;
4928 case MPI2_SCSI_STATUS_INTERMEDIATE_CONDMET:
4929 desc_scsi_status = "intermediate condmet";
4930 break;
4931 case MPI2_SCSI_STATUS_RESERVATION_CONFLICT:
4932 desc_scsi_status = "reservation conflict";
4933 break;
4934 case MPI2_SCSI_STATUS_COMMAND_TERMINATED:
4935 desc_scsi_status = "command terminated";
4936 break;
4937 case MPI2_SCSI_STATUS_TASK_SET_FULL:
4938 desc_scsi_status = "task set full";
4939 break;
4940 case MPI2_SCSI_STATUS_ACA_ACTIVE:
4941 desc_scsi_status = "aca active";
4942 break;
4943 case MPI2_SCSI_STATUS_TASK_ABORTED:
4944 desc_scsi_status = "task aborted";
4945 break;
4946 default:
4947 desc_scsi_status = "unknown";
4948 break;
4949 }
4950
4951 desc_scsi_state[0] = '\0';
4952 if (!scsi_state)
4953 desc_scsi_state = " ";
4954 if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID)
4955 strcat(desc_scsi_state, "response info ");
4956 if (scsi_state & MPI2_SCSI_STATE_TERMINATED)
4957 strcat(desc_scsi_state, "state terminated ");
4958 if (scsi_state & MPI2_SCSI_STATE_NO_SCSI_STATUS)
4959 strcat(desc_scsi_state, "no status ");
4960 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_FAILED)
4961 strcat(desc_scsi_state, "autosense failed ");
4962 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID)
4963 strcat(desc_scsi_state, "autosense valid ");
4964
4965 scsi_print_command(scmd);
4966
4967 if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) {
4968 ioc_warn(ioc, "\t%s wwid(0x%016llx)\n",
4969 device_str, (u64)priv_target->sas_address);
4970 } else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) {
4971 pcie_device = mpt3sas_get_pdev_from_target(ioc, priv_target);
4972 if (pcie_device) {
4973 ioc_info(ioc, "\twwid(0x%016llx), port(%d)\n",
4974 (u64)pcie_device->wwid, pcie_device->port_num);
4975 if (pcie_device->enclosure_handle != 0)
4976 ioc_info(ioc, "\tenclosure logical id(0x%016llx), slot(%d)\n",
4977 (u64)pcie_device->enclosure_logical_id,
4978 pcie_device->slot);
4979 if (pcie_device->connector_name[0])
4980 ioc_info(ioc, "\tenclosure level(0x%04x), connector name( %s)\n",
4981 pcie_device->enclosure_level,
4982 pcie_device->connector_name);
4983 pcie_device_put(pcie_device);
4984 }
4985 } else {
4986 sas_device = mpt3sas_get_sdev_from_target(ioc, priv_target);
4987 if (sas_device) {
4988 ioc_warn(ioc, "\tsas_address(0x%016llx), phy(%d)\n",
4989 (u64)sas_device->sas_address, sas_device->phy);
4990
4991 _scsih_display_enclosure_chassis_info(ioc, sas_device,
4992 NULL, NULL);
4993
4994 sas_device_put(sas_device);
4995 }
4996 }
4997
4998 ioc_warn(ioc, "\thandle(0x%04x), ioc_status(%s)(0x%04x), smid(%d)\n",
4999 le16_to_cpu(mpi_reply->DevHandle),
5000 desc_ioc_state, ioc_status, smid);
5001 ioc_warn(ioc, "\trequest_len(%d), underflow(%d), resid(%d)\n",
5002 scsi_bufflen(scmd), scmd->underflow, scsi_get_resid(scmd));
5003 ioc_warn(ioc, "\ttag(%d), transfer_count(%d), sc->result(0x%08x)\n",
5004 le16_to_cpu(mpi_reply->TaskTag),
5005 le32_to_cpu(mpi_reply->TransferCount), scmd->result);
5006 ioc_warn(ioc, "\tscsi_status(%s)(0x%02x), scsi_state(%s)(0x%02x)\n",
5007 desc_scsi_status, scsi_status, desc_scsi_state, scsi_state);
5008
5009 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) {
5010 struct sense_info data;
5011 _scsih_normalize_sense(scmd->sense_buffer, &data);
5012 ioc_warn(ioc, "\t[sense_key,asc,ascq]: [0x%02x,0x%02x,0x%02x], count(%d)\n",
5013 data.skey, data.asc, data.ascq,
5014 le32_to_cpu(mpi_reply->SenseCount));
5015 }
5016 if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) {
5017 response_info = le32_to_cpu(mpi_reply->ResponseInfo);
5018 response_bytes = (u8 *)&response_info;
5019 _scsih_response_code(ioc, response_bytes[0]);
5020 }
5021 }
5022
5023 /**
5024 * _scsih_turn_on_pfa_led - illuminate PFA LED
5025 * @ioc: per adapter object
5026 * @handle: device handle
5027 * Context: process
5028 */
5029 static void
_scsih_turn_on_pfa_led(struct MPT3SAS_ADAPTER * ioc,u16 handle)5030 _scsih_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle)
5031 {
5032 Mpi2SepReply_t mpi_reply;
5033 Mpi2SepRequest_t mpi_request;
5034 struct _sas_device *sas_device;
5035
5036 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
5037 if (!sas_device)
5038 return;
5039
5040 memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t));
5041 mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR;
5042 mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS;
5043 mpi_request.SlotStatus =
5044 cpu_to_le32(MPI2_SEP_REQ_SLOTSTATUS_PREDICTED_FAULT);
5045 mpi_request.DevHandle = cpu_to_le16(handle);
5046 mpi_request.Flags = MPI2_SEP_REQ_FLAGS_DEVHANDLE_ADDRESS;
5047 if ((mpt3sas_base_scsi_enclosure_processor(ioc, &mpi_reply,
5048 &mpi_request)) != 0) {
5049 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5050 __FILE__, __LINE__, __func__);
5051 goto out;
5052 }
5053 sas_device->pfa_led_on = 1;
5054
5055 if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) {
5056 dewtprintk(ioc,
5057 ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n",
5058 le16_to_cpu(mpi_reply.IOCStatus),
5059 le32_to_cpu(mpi_reply.IOCLogInfo)));
5060 goto out;
5061 }
5062 out:
5063 sas_device_put(sas_device);
5064 }
5065
5066 /**
5067 * _scsih_turn_off_pfa_led - turn off Fault LED
5068 * @ioc: per adapter object
5069 * @sas_device: sas device whose PFA LED has to turned off
5070 * Context: process
5071 */
5072 static void
_scsih_turn_off_pfa_led(struct MPT3SAS_ADAPTER * ioc,struct _sas_device * sas_device)5073 _scsih_turn_off_pfa_led(struct MPT3SAS_ADAPTER *ioc,
5074 struct _sas_device *sas_device)
5075 {
5076 Mpi2SepReply_t mpi_reply;
5077 Mpi2SepRequest_t mpi_request;
5078
5079 memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t));
5080 mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR;
5081 mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS;
5082 mpi_request.SlotStatus = 0;
5083 mpi_request.Slot = cpu_to_le16(sas_device->slot);
5084 mpi_request.DevHandle = 0;
5085 mpi_request.EnclosureHandle = cpu_to_le16(sas_device->enclosure_handle);
5086 mpi_request.Flags = MPI2_SEP_REQ_FLAGS_ENCLOSURE_SLOT_ADDRESS;
5087 if ((mpt3sas_base_scsi_enclosure_processor(ioc, &mpi_reply,
5088 &mpi_request)) != 0) {
5089 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5090 __FILE__, __LINE__, __func__);
5091 return;
5092 }
5093
5094 if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) {
5095 dewtprintk(ioc,
5096 ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n",
5097 le16_to_cpu(mpi_reply.IOCStatus),
5098 le32_to_cpu(mpi_reply.IOCLogInfo)));
5099 return;
5100 }
5101 }
5102
5103 /**
5104 * _scsih_send_event_to_turn_on_pfa_led - fire delayed event
5105 * @ioc: per adapter object
5106 * @handle: device handle
5107 * Context: interrupt.
5108 */
5109 static void
_scsih_send_event_to_turn_on_pfa_led(struct MPT3SAS_ADAPTER * ioc,u16 handle)5110 _scsih_send_event_to_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle)
5111 {
5112 struct fw_event_work *fw_event;
5113
5114 fw_event = alloc_fw_event_work(0);
5115 if (!fw_event)
5116 return;
5117 fw_event->event = MPT3SAS_TURN_ON_PFA_LED;
5118 fw_event->device_handle = handle;
5119 fw_event->ioc = ioc;
5120 _scsih_fw_event_add(ioc, fw_event);
5121 fw_event_work_put(fw_event);
5122 }
5123
5124 /**
5125 * _scsih_smart_predicted_fault - process smart errors
5126 * @ioc: per adapter object
5127 * @handle: device handle
5128 * Context: interrupt.
5129 */
5130 static void
_scsih_smart_predicted_fault(struct MPT3SAS_ADAPTER * ioc,u16 handle)5131 _scsih_smart_predicted_fault(struct MPT3SAS_ADAPTER *ioc, u16 handle)
5132 {
5133 struct scsi_target *starget;
5134 struct MPT3SAS_TARGET *sas_target_priv_data;
5135 Mpi2EventNotificationReply_t *event_reply;
5136 Mpi2EventDataSasDeviceStatusChange_t *event_data;
5137 struct _sas_device *sas_device;
5138 ssize_t sz;
5139 unsigned long flags;
5140
5141 /* only handle non-raid devices */
5142 spin_lock_irqsave(&ioc->sas_device_lock, flags);
5143 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
5144 if (!sas_device)
5145 goto out_unlock;
5146
5147 starget = sas_device->starget;
5148 sas_target_priv_data = starget->hostdata;
5149
5150 if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) ||
5151 ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)))
5152 goto out_unlock;
5153
5154 _scsih_display_enclosure_chassis_info(NULL, sas_device, NULL, starget);
5155
5156 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
5157
5158 if (ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM)
5159 _scsih_send_event_to_turn_on_pfa_led(ioc, handle);
5160
5161 /* insert into event log */
5162 sz = offsetof(Mpi2EventNotificationReply_t, EventData) +
5163 sizeof(Mpi2EventDataSasDeviceStatusChange_t);
5164 event_reply = kzalloc(sz, GFP_KERNEL);
5165 if (!event_reply) {
5166 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5167 __FILE__, __LINE__, __func__);
5168 goto out;
5169 }
5170
5171 event_reply->Function = MPI2_FUNCTION_EVENT_NOTIFICATION;
5172 event_reply->Event =
5173 cpu_to_le16(MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE);
5174 event_reply->MsgLength = sz/4;
5175 event_reply->EventDataLength =
5176 cpu_to_le16(sizeof(Mpi2EventDataSasDeviceStatusChange_t)/4);
5177 event_data = (Mpi2EventDataSasDeviceStatusChange_t *)
5178 event_reply->EventData;
5179 event_data->ReasonCode = MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA;
5180 event_data->ASC = 0x5D;
5181 event_data->DevHandle = cpu_to_le16(handle);
5182 event_data->SASAddress = cpu_to_le64(sas_target_priv_data->sas_address);
5183 mpt3sas_ctl_add_to_event_log(ioc, event_reply);
5184 kfree(event_reply);
5185 out:
5186 if (sas_device)
5187 sas_device_put(sas_device);
5188 return;
5189
5190 out_unlock:
5191 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
5192 goto out;
5193 }
5194
5195 /**
5196 * _scsih_io_done - scsi request callback
5197 * @ioc: per adapter object
5198 * @smid: system request message index
5199 * @msix_index: MSIX table index supplied by the OS
5200 * @reply: reply message frame(lower 32bit addr)
5201 *
5202 * Callback handler when using _scsih_qcmd.
5203 *
5204 * Return: 1 meaning mf should be freed from _base_interrupt
5205 * 0 means the mf is freed from this function.
5206 */
5207 static u8
_scsih_io_done(struct MPT3SAS_ADAPTER * ioc,u16 smid,u8 msix_index,u32 reply)5208 _scsih_io_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply)
5209 {
5210 Mpi25SCSIIORequest_t *mpi_request;
5211 Mpi2SCSIIOReply_t *mpi_reply;
5212 struct scsi_cmnd *scmd;
5213 struct scsiio_tracker *st;
5214 u16 ioc_status;
5215 u32 xfer_cnt;
5216 u8 scsi_state;
5217 u8 scsi_status;
5218 u32 log_info;
5219 struct MPT3SAS_DEVICE *sas_device_priv_data;
5220 u32 response_code = 0;
5221
5222 mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply);
5223
5224 scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid);
5225 if (scmd == NULL)
5226 return 1;
5227
5228 _scsih_set_satl_pending(scmd, false);
5229
5230 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
5231
5232 if (mpi_reply == NULL) {
5233 scmd->result = DID_OK << 16;
5234 goto out;
5235 }
5236
5237 sas_device_priv_data = scmd->device->hostdata;
5238 if (!sas_device_priv_data || !sas_device_priv_data->sas_target ||
5239 sas_device_priv_data->sas_target->deleted) {
5240 scmd->result = DID_NO_CONNECT << 16;
5241 goto out;
5242 }
5243 ioc_status = le16_to_cpu(mpi_reply->IOCStatus);
5244
5245 /*
5246 * WARPDRIVE: If direct_io is set then it is directIO,
5247 * the failed direct I/O should be redirected to volume
5248 */
5249 st = scsi_cmd_priv(scmd);
5250 if (st->direct_io &&
5251 ((ioc_status & MPI2_IOCSTATUS_MASK)
5252 != MPI2_IOCSTATUS_SCSI_TASK_TERMINATED)) {
5253 st->direct_io = 0;
5254 st->scmd = scmd;
5255 memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len);
5256 mpi_request->DevHandle =
5257 cpu_to_le16(sas_device_priv_data->sas_target->handle);
5258 ioc->put_smid_scsi_io(ioc, smid,
5259 sas_device_priv_data->sas_target->handle);
5260 return 0;
5261 }
5262 /* turning off TLR */
5263 scsi_state = mpi_reply->SCSIState;
5264 if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID)
5265 response_code =
5266 le32_to_cpu(mpi_reply->ResponseInfo) & 0xFF;
5267 if (!sas_device_priv_data->tlr_snoop_check) {
5268 sas_device_priv_data->tlr_snoop_check++;
5269 if ((!ioc->is_warpdrive &&
5270 !scsih_is_raid(&scmd->device->sdev_gendev) &&
5271 !scsih_is_nvme(&scmd->device->sdev_gendev))
5272 && sas_is_tlr_enabled(scmd->device) &&
5273 response_code == MPI2_SCSITASKMGMT_RSP_INVALID_FRAME) {
5274 sas_disable_tlr(scmd->device);
5275 sdev_printk(KERN_INFO, scmd->device, "TLR disabled\n");
5276 }
5277 }
5278
5279 xfer_cnt = le32_to_cpu(mpi_reply->TransferCount);
5280 scsi_set_resid(scmd, scsi_bufflen(scmd) - xfer_cnt);
5281 if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE)
5282 log_info = le32_to_cpu(mpi_reply->IOCLogInfo);
5283 else
5284 log_info = 0;
5285 ioc_status &= MPI2_IOCSTATUS_MASK;
5286 scsi_status = mpi_reply->SCSIStatus;
5287
5288 if (ioc_status == MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN && xfer_cnt == 0 &&
5289 (scsi_status == MPI2_SCSI_STATUS_BUSY ||
5290 scsi_status == MPI2_SCSI_STATUS_RESERVATION_CONFLICT ||
5291 scsi_status == MPI2_SCSI_STATUS_TASK_SET_FULL)) {
5292 ioc_status = MPI2_IOCSTATUS_SUCCESS;
5293 }
5294
5295 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) {
5296 struct sense_info data;
5297 const void *sense_data = mpt3sas_base_get_sense_buffer(ioc,
5298 smid);
5299 u32 sz = min_t(u32, SCSI_SENSE_BUFFERSIZE,
5300 le32_to_cpu(mpi_reply->SenseCount));
5301 memcpy(scmd->sense_buffer, sense_data, sz);
5302 _scsih_normalize_sense(scmd->sense_buffer, &data);
5303 /* failure prediction threshold exceeded */
5304 if (data.asc == 0x5D)
5305 _scsih_smart_predicted_fault(ioc,
5306 le16_to_cpu(mpi_reply->DevHandle));
5307 mpt3sas_trigger_scsi(ioc, data.skey, data.asc, data.ascq);
5308
5309 if ((ioc->logging_level & MPT_DEBUG_REPLY) &&
5310 ((scmd->sense_buffer[2] == UNIT_ATTENTION) ||
5311 (scmd->sense_buffer[2] == MEDIUM_ERROR) ||
5312 (scmd->sense_buffer[2] == HARDWARE_ERROR)))
5313 _scsih_scsi_ioc_info(ioc, scmd, mpi_reply, smid);
5314 }
5315 switch (ioc_status) {
5316 case MPI2_IOCSTATUS_BUSY:
5317 case MPI2_IOCSTATUS_INSUFFICIENT_RESOURCES:
5318 scmd->result = SAM_STAT_BUSY;
5319 break;
5320
5321 case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE:
5322 scmd->result = DID_NO_CONNECT << 16;
5323 break;
5324
5325 case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED:
5326 if (sas_device_priv_data->block) {
5327 scmd->result = DID_TRANSPORT_DISRUPTED << 16;
5328 goto out;
5329 }
5330 if (log_info == 0x31110630) {
5331 if (scmd->retries > 2) {
5332 scmd->result = DID_NO_CONNECT << 16;
5333 scsi_device_set_state(scmd->device,
5334 SDEV_OFFLINE);
5335 } else {
5336 scmd->result = DID_SOFT_ERROR << 16;
5337 scmd->device->expecting_cc_ua = 1;
5338 }
5339 break;
5340 } else if (log_info == VIRTUAL_IO_FAILED_RETRY) {
5341 scmd->result = DID_RESET << 16;
5342 break;
5343 } else if ((scmd->device->channel == RAID_CHANNEL) &&
5344 (scsi_state == (MPI2_SCSI_STATE_TERMINATED |
5345 MPI2_SCSI_STATE_NO_SCSI_STATUS))) {
5346 scmd->result = DID_RESET << 16;
5347 break;
5348 }
5349 scmd->result = DID_SOFT_ERROR << 16;
5350 break;
5351 case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED:
5352 case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED:
5353 scmd->result = DID_RESET << 16;
5354 break;
5355
5356 case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH:
5357 if ((xfer_cnt == 0) || (scmd->underflow > xfer_cnt))
5358 scmd->result = DID_SOFT_ERROR << 16;
5359 else
5360 scmd->result = (DID_OK << 16) | scsi_status;
5361 break;
5362
5363 case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN:
5364 scmd->result = (DID_OK << 16) | scsi_status;
5365
5366 if ((scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID))
5367 break;
5368
5369 if (xfer_cnt < scmd->underflow) {
5370 if (scsi_status == SAM_STAT_BUSY)
5371 scmd->result = SAM_STAT_BUSY;
5372 else
5373 scmd->result = DID_SOFT_ERROR << 16;
5374 } else if (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED |
5375 MPI2_SCSI_STATE_NO_SCSI_STATUS))
5376 scmd->result = DID_SOFT_ERROR << 16;
5377 else if (scsi_state & MPI2_SCSI_STATE_TERMINATED)
5378 scmd->result = DID_RESET << 16;
5379 else if (!xfer_cnt && scmd->cmnd[0] == REPORT_LUNS) {
5380 mpi_reply->SCSIState = MPI2_SCSI_STATE_AUTOSENSE_VALID;
5381 mpi_reply->SCSIStatus = SAM_STAT_CHECK_CONDITION;
5382 scmd->result = (DRIVER_SENSE << 24) |
5383 SAM_STAT_CHECK_CONDITION;
5384 scmd->sense_buffer[0] = 0x70;
5385 scmd->sense_buffer[2] = ILLEGAL_REQUEST;
5386 scmd->sense_buffer[12] = 0x20;
5387 scmd->sense_buffer[13] = 0;
5388 }
5389 break;
5390
5391 case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN:
5392 scsi_set_resid(scmd, 0);
5393 /* fall through */
5394 case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR:
5395 case MPI2_IOCSTATUS_SUCCESS:
5396 scmd->result = (DID_OK << 16) | scsi_status;
5397 if (response_code ==
5398 MPI2_SCSITASKMGMT_RSP_INVALID_FRAME ||
5399 (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED |
5400 MPI2_SCSI_STATE_NO_SCSI_STATUS)))
5401 scmd->result = DID_SOFT_ERROR << 16;
5402 else if (scsi_state & MPI2_SCSI_STATE_TERMINATED)
5403 scmd->result = DID_RESET << 16;
5404 break;
5405
5406 case MPI2_IOCSTATUS_EEDP_GUARD_ERROR:
5407 case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR:
5408 case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR:
5409 _scsih_eedp_error_handling(scmd, ioc_status);
5410 break;
5411
5412 case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR:
5413 case MPI2_IOCSTATUS_INVALID_FUNCTION:
5414 case MPI2_IOCSTATUS_INVALID_SGL:
5415 case MPI2_IOCSTATUS_INTERNAL_ERROR:
5416 case MPI2_IOCSTATUS_INVALID_FIELD:
5417 case MPI2_IOCSTATUS_INVALID_STATE:
5418 case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR:
5419 case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED:
5420 case MPI2_IOCSTATUS_INSUFFICIENT_POWER:
5421 default:
5422 scmd->result = DID_SOFT_ERROR << 16;
5423 break;
5424
5425 }
5426
5427 if (scmd->result && (ioc->logging_level & MPT_DEBUG_REPLY))
5428 _scsih_scsi_ioc_info(ioc , scmd, mpi_reply, smid);
5429
5430 out:
5431
5432 scsi_dma_unmap(scmd);
5433 mpt3sas_base_free_smid(ioc, smid);
5434 scmd->scsi_done(scmd);
5435 return 0;
5436 }
5437
5438 /**
5439 * _scsih_sas_host_refresh - refreshing sas host object contents
5440 * @ioc: per adapter object
5441 * Context: user
5442 *
5443 * During port enable, fw will send topology events for every device. Its
5444 * possible that the handles may change from the previous setting, so this
5445 * code keeping handles updating if changed.
5446 */
5447 static void
_scsih_sas_host_refresh(struct MPT3SAS_ADAPTER * ioc)5448 _scsih_sas_host_refresh(struct MPT3SAS_ADAPTER *ioc)
5449 {
5450 u16 sz;
5451 u16 ioc_status;
5452 int i;
5453 Mpi2ConfigReply_t mpi_reply;
5454 Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL;
5455 u16 attached_handle;
5456 u8 link_rate;
5457
5458 dtmprintk(ioc,
5459 ioc_info(ioc, "updating handles for sas_host(0x%016llx)\n",
5460 (u64)ioc->sas_hba.sas_address));
5461
5462 sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys
5463 * sizeof(Mpi2SasIOUnit0PhyData_t));
5464 sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL);
5465 if (!sas_iounit_pg0) {
5466 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5467 __FILE__, __LINE__, __func__);
5468 return;
5469 }
5470
5471 if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply,
5472 sas_iounit_pg0, sz)) != 0)
5473 goto out;
5474 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
5475 if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
5476 goto out;
5477 for (i = 0; i < ioc->sas_hba.num_phys ; i++) {
5478 link_rate = sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4;
5479 if (i == 0)
5480 ioc->sas_hba.handle = le16_to_cpu(sas_iounit_pg0->
5481 PhyData[0].ControllerDevHandle);
5482 ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle;
5483 attached_handle = le16_to_cpu(sas_iounit_pg0->PhyData[i].
5484 AttachedDevHandle);
5485 if (attached_handle && link_rate < MPI2_SAS_NEG_LINK_RATE_1_5)
5486 link_rate = MPI2_SAS_NEG_LINK_RATE_1_5;
5487 mpt3sas_transport_update_links(ioc, ioc->sas_hba.sas_address,
5488 attached_handle, i, link_rate);
5489 }
5490 out:
5491 kfree(sas_iounit_pg0);
5492 }
5493
5494 /**
5495 * _scsih_sas_host_add - create sas host object
5496 * @ioc: per adapter object
5497 *
5498 * Creating host side data object, stored in ioc->sas_hba
5499 */
5500 static void
_scsih_sas_host_add(struct MPT3SAS_ADAPTER * ioc)5501 _scsih_sas_host_add(struct MPT3SAS_ADAPTER *ioc)
5502 {
5503 int i;
5504 Mpi2ConfigReply_t mpi_reply;
5505 Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL;
5506 Mpi2SasIOUnitPage1_t *sas_iounit_pg1 = NULL;
5507 Mpi2SasPhyPage0_t phy_pg0;
5508 Mpi2SasDevicePage0_t sas_device_pg0;
5509 Mpi2SasEnclosurePage0_t enclosure_pg0;
5510 u16 ioc_status;
5511 u16 sz;
5512 u8 device_missing_delay;
5513 u8 num_phys;
5514
5515 mpt3sas_config_get_number_hba_phys(ioc, &num_phys);
5516 if (!num_phys) {
5517 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5518 __FILE__, __LINE__, __func__);
5519 return;
5520 }
5521 ioc->sas_hba.phy = kcalloc(num_phys,
5522 sizeof(struct _sas_phy), GFP_KERNEL);
5523 if (!ioc->sas_hba.phy) {
5524 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5525 __FILE__, __LINE__, __func__);
5526 goto out;
5527 }
5528 ioc->sas_hba.num_phys = num_phys;
5529
5530 /* sas_iounit page 0 */
5531 sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys *
5532 sizeof(Mpi2SasIOUnit0PhyData_t));
5533 sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL);
5534 if (!sas_iounit_pg0) {
5535 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5536 __FILE__, __LINE__, __func__);
5537 return;
5538 }
5539 if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply,
5540 sas_iounit_pg0, sz))) {
5541 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5542 __FILE__, __LINE__, __func__);
5543 goto out;
5544 }
5545 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
5546 MPI2_IOCSTATUS_MASK;
5547 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
5548 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5549 __FILE__, __LINE__, __func__);
5550 goto out;
5551 }
5552
5553 /* sas_iounit page 1 */
5554 sz = offsetof(Mpi2SasIOUnitPage1_t, PhyData) + (ioc->sas_hba.num_phys *
5555 sizeof(Mpi2SasIOUnit1PhyData_t));
5556 sas_iounit_pg1 = kzalloc(sz, GFP_KERNEL);
5557 if (!sas_iounit_pg1) {
5558 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5559 __FILE__, __LINE__, __func__);
5560 goto out;
5561 }
5562 if ((mpt3sas_config_get_sas_iounit_pg1(ioc, &mpi_reply,
5563 sas_iounit_pg1, sz))) {
5564 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5565 __FILE__, __LINE__, __func__);
5566 goto out;
5567 }
5568 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
5569 MPI2_IOCSTATUS_MASK;
5570 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
5571 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5572 __FILE__, __LINE__, __func__);
5573 goto out;
5574 }
5575
5576 ioc->io_missing_delay =
5577 sas_iounit_pg1->IODeviceMissingDelay;
5578 device_missing_delay =
5579 sas_iounit_pg1->ReportDeviceMissingDelay;
5580 if (device_missing_delay & MPI2_SASIOUNIT1_REPORT_MISSING_UNIT_16)
5581 ioc->device_missing_delay = (device_missing_delay &
5582 MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK) * 16;
5583 else
5584 ioc->device_missing_delay = device_missing_delay &
5585 MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK;
5586
5587 ioc->sas_hba.parent_dev = &ioc->shost->shost_gendev;
5588 for (i = 0; i < ioc->sas_hba.num_phys ; i++) {
5589 if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, &phy_pg0,
5590 i))) {
5591 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5592 __FILE__, __LINE__, __func__);
5593 goto out;
5594 }
5595 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
5596 MPI2_IOCSTATUS_MASK;
5597 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
5598 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5599 __FILE__, __LINE__, __func__);
5600 goto out;
5601 }
5602
5603 if (i == 0)
5604 ioc->sas_hba.handle = le16_to_cpu(sas_iounit_pg0->
5605 PhyData[0].ControllerDevHandle);
5606 ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle;
5607 ioc->sas_hba.phy[i].phy_id = i;
5608 mpt3sas_transport_add_host_phy(ioc, &ioc->sas_hba.phy[i],
5609 phy_pg0, ioc->sas_hba.parent_dev);
5610 }
5611 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
5612 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, ioc->sas_hba.handle))) {
5613 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5614 __FILE__, __LINE__, __func__);
5615 goto out;
5616 }
5617 ioc->sas_hba.enclosure_handle =
5618 le16_to_cpu(sas_device_pg0.EnclosureHandle);
5619 ioc->sas_hba.sas_address = le64_to_cpu(sas_device_pg0.SASAddress);
5620 ioc_info(ioc, "host_add: handle(0x%04x), sas_addr(0x%016llx), phys(%d)\n",
5621 ioc->sas_hba.handle,
5622 (u64)ioc->sas_hba.sas_address,
5623 ioc->sas_hba.num_phys);
5624
5625 if (ioc->sas_hba.enclosure_handle) {
5626 if (!(mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply,
5627 &enclosure_pg0, MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE,
5628 ioc->sas_hba.enclosure_handle)))
5629 ioc->sas_hba.enclosure_logical_id =
5630 le64_to_cpu(enclosure_pg0.EnclosureLogicalID);
5631 }
5632
5633 out:
5634 kfree(sas_iounit_pg1);
5635 kfree(sas_iounit_pg0);
5636 }
5637
5638 /**
5639 * _scsih_expander_add - creating expander object
5640 * @ioc: per adapter object
5641 * @handle: expander handle
5642 *
5643 * Creating expander object, stored in ioc->sas_expander_list.
5644 *
5645 * Return: 0 for success, else error.
5646 */
5647 static int
_scsih_expander_add(struct MPT3SAS_ADAPTER * ioc,u16 handle)5648 _scsih_expander_add(struct MPT3SAS_ADAPTER *ioc, u16 handle)
5649 {
5650 struct _sas_node *sas_expander;
5651 struct _enclosure_node *enclosure_dev;
5652 Mpi2ConfigReply_t mpi_reply;
5653 Mpi2ExpanderPage0_t expander_pg0;
5654 Mpi2ExpanderPage1_t expander_pg1;
5655 u32 ioc_status;
5656 u16 parent_handle;
5657 u64 sas_address, sas_address_parent = 0;
5658 int i;
5659 unsigned long flags;
5660 struct _sas_port *mpt3sas_port = NULL;
5661
5662 int rc = 0;
5663
5664 if (!handle)
5665 return -1;
5666
5667 if (ioc->shost_recovery || ioc->pci_error_recovery)
5668 return -1;
5669
5670 if ((mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0,
5671 MPI2_SAS_EXPAND_PGAD_FORM_HNDL, handle))) {
5672 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5673 __FILE__, __LINE__, __func__);
5674 return -1;
5675 }
5676
5677 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
5678 MPI2_IOCSTATUS_MASK;
5679 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
5680 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5681 __FILE__, __LINE__, __func__);
5682 return -1;
5683 }
5684
5685 /* handle out of order topology events */
5686 parent_handle = le16_to_cpu(expander_pg0.ParentDevHandle);
5687 if (_scsih_get_sas_address(ioc, parent_handle, &sas_address_parent)
5688 != 0) {
5689 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5690 __FILE__, __LINE__, __func__);
5691 return -1;
5692 }
5693 if (sas_address_parent != ioc->sas_hba.sas_address) {
5694 spin_lock_irqsave(&ioc->sas_node_lock, flags);
5695 sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc,
5696 sas_address_parent);
5697 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
5698 if (!sas_expander) {
5699 rc = _scsih_expander_add(ioc, parent_handle);
5700 if (rc != 0)
5701 return rc;
5702 }
5703 }
5704
5705 spin_lock_irqsave(&ioc->sas_node_lock, flags);
5706 sas_address = le64_to_cpu(expander_pg0.SASAddress);
5707 sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc,
5708 sas_address);
5709 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
5710
5711 if (sas_expander)
5712 return 0;
5713
5714 sas_expander = kzalloc(sizeof(struct _sas_node),
5715 GFP_KERNEL);
5716 if (!sas_expander) {
5717 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5718 __FILE__, __LINE__, __func__);
5719 return -1;
5720 }
5721
5722 sas_expander->handle = handle;
5723 sas_expander->num_phys = expander_pg0.NumPhys;
5724 sas_expander->sas_address_parent = sas_address_parent;
5725 sas_expander->sas_address = sas_address;
5726
5727 ioc_info(ioc, "expander_add: handle(0x%04x), parent(0x%04x), sas_addr(0x%016llx), phys(%d)\n",
5728 handle, parent_handle,
5729 (u64)sas_expander->sas_address, sas_expander->num_phys);
5730
5731 if (!sas_expander->num_phys)
5732 goto out_fail;
5733 sas_expander->phy = kcalloc(sas_expander->num_phys,
5734 sizeof(struct _sas_phy), GFP_KERNEL);
5735 if (!sas_expander->phy) {
5736 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5737 __FILE__, __LINE__, __func__);
5738 rc = -1;
5739 goto out_fail;
5740 }
5741
5742 INIT_LIST_HEAD(&sas_expander->sas_port_list);
5743 mpt3sas_port = mpt3sas_transport_port_add(ioc, handle,
5744 sas_address_parent);
5745 if (!mpt3sas_port) {
5746 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5747 __FILE__, __LINE__, __func__);
5748 rc = -1;
5749 goto out_fail;
5750 }
5751 sas_expander->parent_dev = &mpt3sas_port->rphy->dev;
5752
5753 for (i = 0 ; i < sas_expander->num_phys ; i++) {
5754 if ((mpt3sas_config_get_expander_pg1(ioc, &mpi_reply,
5755 &expander_pg1, i, handle))) {
5756 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5757 __FILE__, __LINE__, __func__);
5758 rc = -1;
5759 goto out_fail;
5760 }
5761 sas_expander->phy[i].handle = handle;
5762 sas_expander->phy[i].phy_id = i;
5763
5764 if ((mpt3sas_transport_add_expander_phy(ioc,
5765 &sas_expander->phy[i], expander_pg1,
5766 sas_expander->parent_dev))) {
5767 ioc_err(ioc, "failure at %s:%d/%s()!\n",
5768 __FILE__, __LINE__, __func__);
5769 rc = -1;
5770 goto out_fail;
5771 }
5772 }
5773
5774 if (sas_expander->enclosure_handle) {
5775 enclosure_dev =
5776 mpt3sas_scsih_enclosure_find_by_handle(ioc,
5777 sas_expander->enclosure_handle);
5778 if (enclosure_dev)
5779 sas_expander->enclosure_logical_id =
5780 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
5781 }
5782
5783 _scsih_expander_node_add(ioc, sas_expander);
5784 return 0;
5785
5786 out_fail:
5787
5788 if (mpt3sas_port)
5789 mpt3sas_transport_port_remove(ioc, sas_expander->sas_address,
5790 sas_address_parent);
5791 kfree(sas_expander);
5792 return rc;
5793 }
5794
5795 /**
5796 * mpt3sas_expander_remove - removing expander object
5797 * @ioc: per adapter object
5798 * @sas_address: expander sas_address
5799 */
5800 void
mpt3sas_expander_remove(struct MPT3SAS_ADAPTER * ioc,u64 sas_address)5801 mpt3sas_expander_remove(struct MPT3SAS_ADAPTER *ioc, u64 sas_address)
5802 {
5803 struct _sas_node *sas_expander;
5804 unsigned long flags;
5805
5806 if (ioc->shost_recovery)
5807 return;
5808
5809 spin_lock_irqsave(&ioc->sas_node_lock, flags);
5810 sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc,
5811 sas_address);
5812 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
5813 if (sas_expander)
5814 _scsih_expander_node_remove(ioc, sas_expander);
5815 }
5816
5817 /**
5818 * _scsih_done - internal SCSI_IO callback handler.
5819 * @ioc: per adapter object
5820 * @smid: system request message index
5821 * @msix_index: MSIX table index supplied by the OS
5822 * @reply: reply message frame(lower 32bit addr)
5823 *
5824 * Callback handler when sending internal generated SCSI_IO.
5825 * The callback index passed is `ioc->scsih_cb_idx`
5826 *
5827 * Return: 1 meaning mf should be freed from _base_interrupt
5828 * 0 means the mf is freed from this function.
5829 */
5830 static u8
_scsih_done(struct MPT3SAS_ADAPTER * ioc,u16 smid,u8 msix_index,u32 reply)5831 _scsih_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply)
5832 {
5833 MPI2DefaultReply_t *mpi_reply;
5834
5835 mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply);
5836 if (ioc->scsih_cmds.status == MPT3_CMD_NOT_USED)
5837 return 1;
5838 if (ioc->scsih_cmds.smid != smid)
5839 return 1;
5840 ioc->scsih_cmds.status |= MPT3_CMD_COMPLETE;
5841 if (mpi_reply) {
5842 memcpy(ioc->scsih_cmds.reply, mpi_reply,
5843 mpi_reply->MsgLength*4);
5844 ioc->scsih_cmds.status |= MPT3_CMD_REPLY_VALID;
5845 }
5846 ioc->scsih_cmds.status &= ~MPT3_CMD_PENDING;
5847 complete(&ioc->scsih_cmds.done);
5848 return 1;
5849 }
5850
5851
5852
5853
5854 #define MPT3_MAX_LUNS (255)
5855
5856
5857 /**
5858 * _scsih_check_access_status - check access flags
5859 * @ioc: per adapter object
5860 * @sas_address: sas address
5861 * @handle: sas device handle
5862 * @access_status: errors returned during discovery of the device
5863 *
5864 * Return: 0 for success, else failure
5865 */
5866 static u8
_scsih_check_access_status(struct MPT3SAS_ADAPTER * ioc,u64 sas_address,u16 handle,u8 access_status)5867 _scsih_check_access_status(struct MPT3SAS_ADAPTER *ioc, u64 sas_address,
5868 u16 handle, u8 access_status)
5869 {
5870 u8 rc = 1;
5871 char *desc = NULL;
5872
5873 switch (access_status) {
5874 case MPI2_SAS_DEVICE0_ASTATUS_NO_ERRORS:
5875 case MPI2_SAS_DEVICE0_ASTATUS_SATA_NEEDS_INITIALIZATION:
5876 rc = 0;
5877 break;
5878 case MPI2_SAS_DEVICE0_ASTATUS_SATA_CAPABILITY_FAILED:
5879 desc = "sata capability failed";
5880 break;
5881 case MPI2_SAS_DEVICE0_ASTATUS_SATA_AFFILIATION_CONFLICT:
5882 desc = "sata affiliation conflict";
5883 break;
5884 case MPI2_SAS_DEVICE0_ASTATUS_ROUTE_NOT_ADDRESSABLE:
5885 desc = "route not addressable";
5886 break;
5887 case MPI2_SAS_DEVICE0_ASTATUS_SMP_ERROR_NOT_ADDRESSABLE:
5888 desc = "smp error not addressable";
5889 break;
5890 case MPI2_SAS_DEVICE0_ASTATUS_DEVICE_BLOCKED:
5891 desc = "device blocked";
5892 break;
5893 case MPI2_SAS_DEVICE0_ASTATUS_SATA_INIT_FAILED:
5894 case MPI2_SAS_DEVICE0_ASTATUS_SIF_UNKNOWN:
5895 case MPI2_SAS_DEVICE0_ASTATUS_SIF_AFFILIATION_CONFLICT:
5896 case MPI2_SAS_DEVICE0_ASTATUS_SIF_DIAG:
5897 case MPI2_SAS_DEVICE0_ASTATUS_SIF_IDENTIFICATION:
5898 case MPI2_SAS_DEVICE0_ASTATUS_SIF_CHECK_POWER:
5899 case MPI2_SAS_DEVICE0_ASTATUS_SIF_PIO_SN:
5900 case MPI2_SAS_DEVICE0_ASTATUS_SIF_MDMA_SN:
5901 case MPI2_SAS_DEVICE0_ASTATUS_SIF_UDMA_SN:
5902 case MPI2_SAS_DEVICE0_ASTATUS_SIF_ZONING_VIOLATION:
5903 case MPI2_SAS_DEVICE0_ASTATUS_SIF_NOT_ADDRESSABLE:
5904 case MPI2_SAS_DEVICE0_ASTATUS_SIF_MAX:
5905 desc = "sata initialization failed";
5906 break;
5907 default:
5908 desc = "unknown";
5909 break;
5910 }
5911
5912 if (!rc)
5913 return 0;
5914
5915 ioc_err(ioc, "discovery errors(%s): sas_address(0x%016llx), handle(0x%04x)\n",
5916 desc, (u64)sas_address, handle);
5917 return rc;
5918 }
5919
5920 /**
5921 * _scsih_check_device - checking device responsiveness
5922 * @ioc: per adapter object
5923 * @parent_sas_address: sas address of parent expander or sas host
5924 * @handle: attached device handle
5925 * @phy_number: phy number
5926 * @link_rate: new link rate
5927 */
5928 static void
_scsih_check_device(struct MPT3SAS_ADAPTER * ioc,u64 parent_sas_address,u16 handle,u8 phy_number,u8 link_rate)5929 _scsih_check_device(struct MPT3SAS_ADAPTER *ioc,
5930 u64 parent_sas_address, u16 handle, u8 phy_number, u8 link_rate)
5931 {
5932 Mpi2ConfigReply_t mpi_reply;
5933 Mpi2SasDevicePage0_t sas_device_pg0;
5934 struct _sas_device *sas_device;
5935 struct _enclosure_node *enclosure_dev = NULL;
5936 u32 ioc_status;
5937 unsigned long flags;
5938 u64 sas_address;
5939 struct scsi_target *starget;
5940 struct MPT3SAS_TARGET *sas_target_priv_data;
5941 u32 device_info;
5942
5943 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
5944 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle)))
5945 return;
5946
5947 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
5948 if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
5949 return;
5950
5951 /* wide port handling ~ we need only handle device once for the phy that
5952 * is matched in sas device page zero
5953 */
5954 if (phy_number != sas_device_pg0.PhyNum)
5955 return;
5956
5957 /* check if this is end device */
5958 device_info = le32_to_cpu(sas_device_pg0.DeviceInfo);
5959 if (!(_scsih_is_end_device(device_info)))
5960 return;
5961
5962 spin_lock_irqsave(&ioc->sas_device_lock, flags);
5963 sas_address = le64_to_cpu(sas_device_pg0.SASAddress);
5964 sas_device = __mpt3sas_get_sdev_by_addr(ioc,
5965 sas_address);
5966
5967 if (!sas_device)
5968 goto out_unlock;
5969
5970 if (unlikely(sas_device->handle != handle)) {
5971 starget = sas_device->starget;
5972 sas_target_priv_data = starget->hostdata;
5973 starget_printk(KERN_INFO, starget,
5974 "handle changed from(0x%04x) to (0x%04x)!!!\n",
5975 sas_device->handle, handle);
5976 sas_target_priv_data->handle = handle;
5977 sas_device->handle = handle;
5978 if (le16_to_cpu(sas_device_pg0.Flags) &
5979 MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) {
5980 sas_device->enclosure_level =
5981 sas_device_pg0.EnclosureLevel;
5982 memcpy(sas_device->connector_name,
5983 sas_device_pg0.ConnectorName, 4);
5984 sas_device->connector_name[4] = '\0';
5985 } else {
5986 sas_device->enclosure_level = 0;
5987 sas_device->connector_name[0] = '\0';
5988 }
5989
5990 sas_device->enclosure_handle =
5991 le16_to_cpu(sas_device_pg0.EnclosureHandle);
5992 sas_device->is_chassis_slot_valid = 0;
5993 enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc,
5994 sas_device->enclosure_handle);
5995 if (enclosure_dev) {
5996 sas_device->enclosure_logical_id =
5997 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
5998 if (le16_to_cpu(enclosure_dev->pg0.Flags) &
5999 MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) {
6000 sas_device->is_chassis_slot_valid = 1;
6001 sas_device->chassis_slot =
6002 enclosure_dev->pg0.ChassisSlot;
6003 }
6004 }
6005 }
6006
6007 /* check if device is present */
6008 if (!(le16_to_cpu(sas_device_pg0.Flags) &
6009 MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) {
6010 ioc_err(ioc, "device is not present handle(0x%04x), flags!!!\n",
6011 handle);
6012 goto out_unlock;
6013 }
6014
6015 /* check if there were any issues with discovery */
6016 if (_scsih_check_access_status(ioc, sas_address, handle,
6017 sas_device_pg0.AccessStatus))
6018 goto out_unlock;
6019
6020 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
6021 _scsih_ublock_io_device(ioc, sas_address);
6022
6023 if (sas_device)
6024 sas_device_put(sas_device);
6025 return;
6026
6027 out_unlock:
6028 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
6029 if (sas_device)
6030 sas_device_put(sas_device);
6031 }
6032
6033 /**
6034 * _scsih_add_device - creating sas device object
6035 * @ioc: per adapter object
6036 * @handle: sas device handle
6037 * @phy_num: phy number end device attached to
6038 * @is_pd: is this hidden raid component
6039 *
6040 * Creating end device object, stored in ioc->sas_device_list.
6041 *
6042 * Return: 0 for success, non-zero for failure.
6043 */
6044 static int
_scsih_add_device(struct MPT3SAS_ADAPTER * ioc,u16 handle,u8 phy_num,u8 is_pd)6045 _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phy_num,
6046 u8 is_pd)
6047 {
6048 Mpi2ConfigReply_t mpi_reply;
6049 Mpi2SasDevicePage0_t sas_device_pg0;
6050 struct _sas_device *sas_device;
6051 struct _enclosure_node *enclosure_dev = NULL;
6052 u32 ioc_status;
6053 u64 sas_address;
6054 u32 device_info;
6055
6056 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
6057 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) {
6058 ioc_err(ioc, "failure at %s:%d/%s()!\n",
6059 __FILE__, __LINE__, __func__);
6060 return -1;
6061 }
6062
6063 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
6064 MPI2_IOCSTATUS_MASK;
6065 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
6066 ioc_err(ioc, "failure at %s:%d/%s()!\n",
6067 __FILE__, __LINE__, __func__);
6068 return -1;
6069 }
6070
6071 /* check if this is end device */
6072 device_info = le32_to_cpu(sas_device_pg0.DeviceInfo);
6073 if (!(_scsih_is_end_device(device_info)))
6074 return -1;
6075 set_bit(handle, ioc->pend_os_device_add);
6076 sas_address = le64_to_cpu(sas_device_pg0.SASAddress);
6077
6078 /* check if device is present */
6079 if (!(le16_to_cpu(sas_device_pg0.Flags) &
6080 MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) {
6081 ioc_err(ioc, "device is not present handle(0x04%x)!!!\n",
6082 handle);
6083 return -1;
6084 }
6085
6086 /* check if there were any issues with discovery */
6087 if (_scsih_check_access_status(ioc, sas_address, handle,
6088 sas_device_pg0.AccessStatus))
6089 return -1;
6090
6091 sas_device = mpt3sas_get_sdev_by_addr(ioc,
6092 sas_address);
6093 if (sas_device) {
6094 clear_bit(handle, ioc->pend_os_device_add);
6095 sas_device_put(sas_device);
6096 return -1;
6097 }
6098
6099 if (sas_device_pg0.EnclosureHandle) {
6100 enclosure_dev =
6101 mpt3sas_scsih_enclosure_find_by_handle(ioc,
6102 le16_to_cpu(sas_device_pg0.EnclosureHandle));
6103 if (enclosure_dev == NULL)
6104 ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n",
6105 sas_device_pg0.EnclosureHandle);
6106 }
6107
6108 sas_device = kzalloc(sizeof(struct _sas_device),
6109 GFP_KERNEL);
6110 if (!sas_device) {
6111 ioc_err(ioc, "failure at %s:%d/%s()!\n",
6112 __FILE__, __LINE__, __func__);
6113 return 0;
6114 }
6115
6116 kref_init(&sas_device->refcount);
6117 sas_device->handle = handle;
6118 if (_scsih_get_sas_address(ioc,
6119 le16_to_cpu(sas_device_pg0.ParentDevHandle),
6120 &sas_device->sas_address_parent) != 0)
6121 ioc_err(ioc, "failure at %s:%d/%s()!\n",
6122 __FILE__, __LINE__, __func__);
6123 sas_device->enclosure_handle =
6124 le16_to_cpu(sas_device_pg0.EnclosureHandle);
6125 if (sas_device->enclosure_handle != 0)
6126 sas_device->slot =
6127 le16_to_cpu(sas_device_pg0.Slot);
6128 sas_device->device_info = device_info;
6129 sas_device->sas_address = sas_address;
6130 sas_device->phy = sas_device_pg0.PhyNum;
6131 sas_device->fast_path = (le16_to_cpu(sas_device_pg0.Flags) &
6132 MPI25_SAS_DEVICE0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0;
6133
6134 if (le16_to_cpu(sas_device_pg0.Flags)
6135 & MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) {
6136 sas_device->enclosure_level =
6137 sas_device_pg0.EnclosureLevel;
6138 memcpy(sas_device->connector_name,
6139 sas_device_pg0.ConnectorName, 4);
6140 sas_device->connector_name[4] = '\0';
6141 } else {
6142 sas_device->enclosure_level = 0;
6143 sas_device->connector_name[0] = '\0';
6144 }
6145 /* get enclosure_logical_id & chassis_slot*/
6146 sas_device->is_chassis_slot_valid = 0;
6147 if (enclosure_dev) {
6148 sas_device->enclosure_logical_id =
6149 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
6150 if (le16_to_cpu(enclosure_dev->pg0.Flags) &
6151 MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) {
6152 sas_device->is_chassis_slot_valid = 1;
6153 sas_device->chassis_slot =
6154 enclosure_dev->pg0.ChassisSlot;
6155 }
6156 }
6157
6158 /* get device name */
6159 sas_device->device_name = le64_to_cpu(sas_device_pg0.DeviceName);
6160
6161 if (ioc->wait_for_discovery_to_complete)
6162 _scsih_sas_device_init_add(ioc, sas_device);
6163 else
6164 _scsih_sas_device_add(ioc, sas_device);
6165
6166 sas_device_put(sas_device);
6167 return 0;
6168 }
6169
6170 /**
6171 * _scsih_remove_device - removing sas device object
6172 * @ioc: per adapter object
6173 * @sas_device: the sas_device object
6174 */
6175 static void
_scsih_remove_device(struct MPT3SAS_ADAPTER * ioc,struct _sas_device * sas_device)6176 _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc,
6177 struct _sas_device *sas_device)
6178 {
6179 struct MPT3SAS_TARGET *sas_target_priv_data;
6180
6181 if ((ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM) &&
6182 (sas_device->pfa_led_on)) {
6183 _scsih_turn_off_pfa_led(ioc, sas_device);
6184 sas_device->pfa_led_on = 0;
6185 }
6186
6187 dewtprintk(ioc,
6188 ioc_info(ioc, "%s: enter: handle(0x%04x), sas_addr(0x%016llx)\n",
6189 __func__,
6190 sas_device->handle, (u64)sas_device->sas_address));
6191
6192 dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device,
6193 NULL, NULL));
6194
6195 if (sas_device->starget && sas_device->starget->hostdata) {
6196 sas_target_priv_data = sas_device->starget->hostdata;
6197 sas_target_priv_data->deleted = 1;
6198 _scsih_ublock_io_device(ioc, sas_device->sas_address);
6199 sas_target_priv_data->handle =
6200 MPT3SAS_INVALID_DEVICE_HANDLE;
6201 }
6202
6203 if (!ioc->hide_drives)
6204 mpt3sas_transport_port_remove(ioc,
6205 sas_device->sas_address,
6206 sas_device->sas_address_parent);
6207
6208 ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n",
6209 sas_device->handle, (u64)sas_device->sas_address);
6210
6211 _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL);
6212
6213 dewtprintk(ioc,
6214 ioc_info(ioc, "%s: exit: handle(0x%04x), sas_addr(0x%016llx)\n",
6215 __func__,
6216 sas_device->handle, (u64)sas_device->sas_address));
6217 dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device,
6218 NULL, NULL));
6219 }
6220
6221 /**
6222 * _scsih_sas_topology_change_event_debug - debug for topology event
6223 * @ioc: per adapter object
6224 * @event_data: event data payload
6225 * Context: user.
6226 */
6227 static void
_scsih_sas_topology_change_event_debug(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataSasTopologyChangeList_t * event_data)6228 _scsih_sas_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
6229 Mpi2EventDataSasTopologyChangeList_t *event_data)
6230 {
6231 int i;
6232 u16 handle;
6233 u16 reason_code;
6234 u8 phy_number;
6235 char *status_str = NULL;
6236 u8 link_rate, prev_link_rate;
6237
6238 switch (event_data->ExpStatus) {
6239 case MPI2_EVENT_SAS_TOPO_ES_ADDED:
6240 status_str = "add";
6241 break;
6242 case MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING:
6243 status_str = "remove";
6244 break;
6245 case MPI2_EVENT_SAS_TOPO_ES_RESPONDING:
6246 case 0:
6247 status_str = "responding";
6248 break;
6249 case MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING:
6250 status_str = "remove delay";
6251 break;
6252 default:
6253 status_str = "unknown status";
6254 break;
6255 }
6256 ioc_info(ioc, "sas topology change: (%s)\n", status_str);
6257 pr_info("\thandle(0x%04x), enclosure_handle(0x%04x) " \
6258 "start_phy(%02d), count(%d)\n",
6259 le16_to_cpu(event_data->ExpanderDevHandle),
6260 le16_to_cpu(event_data->EnclosureHandle),
6261 event_data->StartPhyNum, event_data->NumEntries);
6262 for (i = 0; i < event_data->NumEntries; i++) {
6263 handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle);
6264 if (!handle)
6265 continue;
6266 phy_number = event_data->StartPhyNum + i;
6267 reason_code = event_data->PHY[i].PhyStatus &
6268 MPI2_EVENT_SAS_TOPO_RC_MASK;
6269 switch (reason_code) {
6270 case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED:
6271 status_str = "target add";
6272 break;
6273 case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING:
6274 status_str = "target remove";
6275 break;
6276 case MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING:
6277 status_str = "delay target remove";
6278 break;
6279 case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED:
6280 status_str = "link rate change";
6281 break;
6282 case MPI2_EVENT_SAS_TOPO_RC_NO_CHANGE:
6283 status_str = "target responding";
6284 break;
6285 default:
6286 status_str = "unknown";
6287 break;
6288 }
6289 link_rate = event_data->PHY[i].LinkRate >> 4;
6290 prev_link_rate = event_data->PHY[i].LinkRate & 0xF;
6291 pr_info("\tphy(%02d), attached_handle(0x%04x): %s:" \
6292 " link rate: new(0x%02x), old(0x%02x)\n", phy_number,
6293 handle, status_str, link_rate, prev_link_rate);
6294
6295 }
6296 }
6297
6298 /**
6299 * _scsih_sas_topology_change_event - handle topology changes
6300 * @ioc: per adapter object
6301 * @fw_event: The fw_event_work object
6302 * Context: user.
6303 *
6304 */
6305 static int
_scsih_sas_topology_change_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)6306 _scsih_sas_topology_change_event(struct MPT3SAS_ADAPTER *ioc,
6307 struct fw_event_work *fw_event)
6308 {
6309 int i;
6310 u16 parent_handle, handle;
6311 u16 reason_code;
6312 u8 phy_number, max_phys;
6313 struct _sas_node *sas_expander;
6314 u64 sas_address;
6315 unsigned long flags;
6316 u8 link_rate, prev_link_rate;
6317 Mpi2EventDataSasTopologyChangeList_t *event_data =
6318 (Mpi2EventDataSasTopologyChangeList_t *)
6319 fw_event->event_data;
6320
6321 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
6322 _scsih_sas_topology_change_event_debug(ioc, event_data);
6323
6324 if (ioc->shost_recovery || ioc->remove_host || ioc->pci_error_recovery)
6325 return 0;
6326
6327 if (!ioc->sas_hba.num_phys)
6328 _scsih_sas_host_add(ioc);
6329 else
6330 _scsih_sas_host_refresh(ioc);
6331
6332 if (fw_event->ignore) {
6333 dewtprintk(ioc, ioc_info(ioc, "ignoring expander event\n"));
6334 return 0;
6335 }
6336
6337 parent_handle = le16_to_cpu(event_data->ExpanderDevHandle);
6338
6339 /* handle expander add */
6340 if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_ADDED)
6341 if (_scsih_expander_add(ioc, parent_handle) != 0)
6342 return 0;
6343
6344 spin_lock_irqsave(&ioc->sas_node_lock, flags);
6345 sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc,
6346 parent_handle);
6347 if (sas_expander) {
6348 sas_address = sas_expander->sas_address;
6349 max_phys = sas_expander->num_phys;
6350 } else if (parent_handle < ioc->sas_hba.num_phys) {
6351 sas_address = ioc->sas_hba.sas_address;
6352 max_phys = ioc->sas_hba.num_phys;
6353 } else {
6354 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
6355 return 0;
6356 }
6357 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
6358
6359 /* handle siblings events */
6360 for (i = 0; i < event_data->NumEntries; i++) {
6361 if (fw_event->ignore) {
6362 dewtprintk(ioc,
6363 ioc_info(ioc, "ignoring expander event\n"));
6364 return 0;
6365 }
6366 if (ioc->remove_host || ioc->pci_error_recovery)
6367 return 0;
6368 phy_number = event_data->StartPhyNum + i;
6369 if (phy_number >= max_phys)
6370 continue;
6371 reason_code = event_data->PHY[i].PhyStatus &
6372 MPI2_EVENT_SAS_TOPO_RC_MASK;
6373 if ((event_data->PHY[i].PhyStatus &
6374 MPI2_EVENT_SAS_TOPO_PHYSTATUS_VACANT) && (reason_code !=
6375 MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING))
6376 continue;
6377 handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle);
6378 if (!handle)
6379 continue;
6380 link_rate = event_data->PHY[i].LinkRate >> 4;
6381 prev_link_rate = event_data->PHY[i].LinkRate & 0xF;
6382 switch (reason_code) {
6383 case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED:
6384
6385 if (ioc->shost_recovery)
6386 break;
6387
6388 if (link_rate == prev_link_rate)
6389 break;
6390
6391 mpt3sas_transport_update_links(ioc, sas_address,
6392 handle, phy_number, link_rate);
6393
6394 if (link_rate < MPI2_SAS_NEG_LINK_RATE_1_5)
6395 break;
6396
6397 _scsih_check_device(ioc, sas_address, handle,
6398 phy_number, link_rate);
6399
6400 if (!test_bit(handle, ioc->pend_os_device_add))
6401 break;
6402
6403 /* fall through */
6404
6405 case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED:
6406
6407 if (ioc->shost_recovery)
6408 break;
6409
6410 mpt3sas_transport_update_links(ioc, sas_address,
6411 handle, phy_number, link_rate);
6412
6413 _scsih_add_device(ioc, handle, phy_number, 0);
6414
6415 break;
6416 case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING:
6417
6418 _scsih_device_remove_by_handle(ioc, handle);
6419 break;
6420 }
6421 }
6422
6423 /* handle expander removal */
6424 if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING &&
6425 sas_expander)
6426 mpt3sas_expander_remove(ioc, sas_address);
6427
6428 return 0;
6429 }
6430
6431 /**
6432 * _scsih_sas_device_status_change_event_debug - debug for device event
6433 * @ioc: ?
6434 * @event_data: event data payload
6435 * Context: user.
6436 */
6437 static void
_scsih_sas_device_status_change_event_debug(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataSasDeviceStatusChange_t * event_data)6438 _scsih_sas_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
6439 Mpi2EventDataSasDeviceStatusChange_t *event_data)
6440 {
6441 char *reason_str = NULL;
6442
6443 switch (event_data->ReasonCode) {
6444 case MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA:
6445 reason_str = "smart data";
6446 break;
6447 case MPI2_EVENT_SAS_DEV_STAT_RC_UNSUPPORTED:
6448 reason_str = "unsupported device discovered";
6449 break;
6450 case MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET:
6451 reason_str = "internal device reset";
6452 break;
6453 case MPI2_EVENT_SAS_DEV_STAT_RC_TASK_ABORT_INTERNAL:
6454 reason_str = "internal task abort";
6455 break;
6456 case MPI2_EVENT_SAS_DEV_STAT_RC_ABORT_TASK_SET_INTERNAL:
6457 reason_str = "internal task abort set";
6458 break;
6459 case MPI2_EVENT_SAS_DEV_STAT_RC_CLEAR_TASK_SET_INTERNAL:
6460 reason_str = "internal clear task set";
6461 break;
6462 case MPI2_EVENT_SAS_DEV_STAT_RC_QUERY_TASK_INTERNAL:
6463 reason_str = "internal query task";
6464 break;
6465 case MPI2_EVENT_SAS_DEV_STAT_RC_SATA_INIT_FAILURE:
6466 reason_str = "sata init failure";
6467 break;
6468 case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET:
6469 reason_str = "internal device reset complete";
6470 break;
6471 case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_TASK_ABORT_INTERNAL:
6472 reason_str = "internal task abort complete";
6473 break;
6474 case MPI2_EVENT_SAS_DEV_STAT_RC_ASYNC_NOTIFICATION:
6475 reason_str = "internal async notification";
6476 break;
6477 case MPI2_EVENT_SAS_DEV_STAT_RC_EXPANDER_REDUCED_FUNCTIONALITY:
6478 reason_str = "expander reduced functionality";
6479 break;
6480 case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_EXPANDER_REDUCED_FUNCTIONALITY:
6481 reason_str = "expander reduced functionality complete";
6482 break;
6483 default:
6484 reason_str = "unknown reason";
6485 break;
6486 }
6487 ioc_info(ioc, "device status change: (%s)\thandle(0x%04x), sas address(0x%016llx), tag(%d)",
6488 reason_str, le16_to_cpu(event_data->DevHandle),
6489 (u64)le64_to_cpu(event_data->SASAddress),
6490 le16_to_cpu(event_data->TaskTag));
6491 if (event_data->ReasonCode == MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA)
6492 pr_cont(", ASC(0x%x), ASCQ(0x%x)\n",
6493 event_data->ASC, event_data->ASCQ);
6494 pr_cont("\n");
6495 }
6496
6497 /**
6498 * _scsih_sas_device_status_change_event - handle device status change
6499 * @ioc: per adapter object
6500 * @event_data: The fw event
6501 * Context: user.
6502 */
6503 static void
_scsih_sas_device_status_change_event(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataSasDeviceStatusChange_t * event_data)6504 _scsih_sas_device_status_change_event(struct MPT3SAS_ADAPTER *ioc,
6505 Mpi2EventDataSasDeviceStatusChange_t *event_data)
6506 {
6507 struct MPT3SAS_TARGET *target_priv_data;
6508 struct _sas_device *sas_device;
6509 u64 sas_address;
6510 unsigned long flags;
6511
6512 /* In MPI Revision K (0xC), the internal device reset complete was
6513 * implemented, so avoid setting tm_busy flag for older firmware.
6514 */
6515 if ((ioc->facts.HeaderVersion >> 8) < 0xC)
6516 return;
6517
6518 if (event_data->ReasonCode !=
6519 MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET &&
6520 event_data->ReasonCode !=
6521 MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET)
6522 return;
6523
6524 spin_lock_irqsave(&ioc->sas_device_lock, flags);
6525 sas_address = le64_to_cpu(event_data->SASAddress);
6526 sas_device = __mpt3sas_get_sdev_by_addr(ioc,
6527 sas_address);
6528
6529 if (!sas_device || !sas_device->starget)
6530 goto out;
6531
6532 target_priv_data = sas_device->starget->hostdata;
6533 if (!target_priv_data)
6534 goto out;
6535
6536 if (event_data->ReasonCode ==
6537 MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET)
6538 target_priv_data->tm_busy = 1;
6539 else
6540 target_priv_data->tm_busy = 0;
6541
6542 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
6543 ioc_info(ioc,
6544 "%s tm_busy flag for handle(0x%04x)\n",
6545 (target_priv_data->tm_busy == 1) ? "Enable" : "Disable",
6546 target_priv_data->handle);
6547
6548 out:
6549 if (sas_device)
6550 sas_device_put(sas_device);
6551
6552 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
6553 }
6554
6555
6556 /**
6557 * _scsih_check_pcie_access_status - check access flags
6558 * @ioc: per adapter object
6559 * @wwid: wwid
6560 * @handle: sas device handle
6561 * @access_status: errors returned during discovery of the device
6562 *
6563 * Return: 0 for success, else failure
6564 */
6565 static u8
_scsih_check_pcie_access_status(struct MPT3SAS_ADAPTER * ioc,u64 wwid,u16 handle,u8 access_status)6566 _scsih_check_pcie_access_status(struct MPT3SAS_ADAPTER *ioc, u64 wwid,
6567 u16 handle, u8 access_status)
6568 {
6569 u8 rc = 1;
6570 char *desc = NULL;
6571
6572 switch (access_status) {
6573 case MPI26_PCIEDEV0_ASTATUS_NO_ERRORS:
6574 case MPI26_PCIEDEV0_ASTATUS_NEEDS_INITIALIZATION:
6575 rc = 0;
6576 break;
6577 case MPI26_PCIEDEV0_ASTATUS_CAPABILITY_FAILED:
6578 desc = "PCIe device capability failed";
6579 break;
6580 case MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED:
6581 desc = "PCIe device blocked";
6582 ioc_info(ioc,
6583 "Device with Access Status (%s): wwid(0x%016llx), "
6584 "handle(0x%04x)\n ll only be added to the internal list",
6585 desc, (u64)wwid, handle);
6586 rc = 0;
6587 break;
6588 case MPI26_PCIEDEV0_ASTATUS_MEMORY_SPACE_ACCESS_FAILED:
6589 desc = "PCIe device mem space access failed";
6590 break;
6591 case MPI26_PCIEDEV0_ASTATUS_UNSUPPORTED_DEVICE:
6592 desc = "PCIe device unsupported";
6593 break;
6594 case MPI26_PCIEDEV0_ASTATUS_MSIX_REQUIRED:
6595 desc = "PCIe device MSIx Required";
6596 break;
6597 case MPI26_PCIEDEV0_ASTATUS_INIT_FAIL_MAX:
6598 desc = "PCIe device init fail max";
6599 break;
6600 case MPI26_PCIEDEV0_ASTATUS_UNKNOWN:
6601 desc = "PCIe device status unknown";
6602 break;
6603 case MPI26_PCIEDEV0_ASTATUS_NVME_READY_TIMEOUT:
6604 desc = "nvme ready timeout";
6605 break;
6606 case MPI26_PCIEDEV0_ASTATUS_NVME_DEVCFG_UNSUPPORTED:
6607 desc = "nvme device configuration unsupported";
6608 break;
6609 case MPI26_PCIEDEV0_ASTATUS_NVME_IDENTIFY_FAILED:
6610 desc = "nvme identify failed";
6611 break;
6612 case MPI26_PCIEDEV0_ASTATUS_NVME_QCONFIG_FAILED:
6613 desc = "nvme qconfig failed";
6614 break;
6615 case MPI26_PCIEDEV0_ASTATUS_NVME_QCREATION_FAILED:
6616 desc = "nvme qcreation failed";
6617 break;
6618 case MPI26_PCIEDEV0_ASTATUS_NVME_EVENTCFG_FAILED:
6619 desc = "nvme eventcfg failed";
6620 break;
6621 case MPI26_PCIEDEV0_ASTATUS_NVME_GET_FEATURE_STAT_FAILED:
6622 desc = "nvme get feature stat failed";
6623 break;
6624 case MPI26_PCIEDEV0_ASTATUS_NVME_IDLE_TIMEOUT:
6625 desc = "nvme idle timeout";
6626 break;
6627 case MPI26_PCIEDEV0_ASTATUS_NVME_FAILURE_STATUS:
6628 desc = "nvme failure status";
6629 break;
6630 default:
6631 ioc_err(ioc, "NVMe discovery error(0x%02x): wwid(0x%016llx), handle(0x%04x)\n",
6632 access_status, (u64)wwid, handle);
6633 return rc;
6634 }
6635
6636 if (!rc)
6637 return rc;
6638
6639 ioc_info(ioc, "NVMe discovery error(%s): wwid(0x%016llx), handle(0x%04x)\n",
6640 desc, (u64)wwid, handle);
6641 return rc;
6642 }
6643
6644 /**
6645 * _scsih_pcie_device_remove_from_sml - removing pcie device
6646 * from SML and free up associated memory
6647 * @ioc: per adapter object
6648 * @pcie_device: the pcie_device object
6649 */
6650 static void
_scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER * ioc,struct _pcie_device * pcie_device)6651 _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc,
6652 struct _pcie_device *pcie_device)
6653 {
6654 struct MPT3SAS_TARGET *sas_target_priv_data;
6655
6656 dewtprintk(ioc,
6657 ioc_info(ioc, "%s: enter: handle(0x%04x), wwid(0x%016llx)\n",
6658 __func__,
6659 pcie_device->handle, (u64)pcie_device->wwid));
6660 if (pcie_device->enclosure_handle != 0)
6661 dewtprintk(ioc,
6662 ioc_info(ioc, "%s: enter: enclosure logical id(0x%016llx), slot(%d)\n",
6663 __func__,
6664 (u64)pcie_device->enclosure_logical_id,
6665 pcie_device->slot));
6666 if (pcie_device->connector_name[0] != '\0')
6667 dewtprintk(ioc,
6668 ioc_info(ioc, "%s: enter: enclosure level(0x%04x), connector name(%s)\n",
6669 __func__,
6670 pcie_device->enclosure_level,
6671 pcie_device->connector_name));
6672
6673 if (pcie_device->starget && pcie_device->starget->hostdata) {
6674 sas_target_priv_data = pcie_device->starget->hostdata;
6675 sas_target_priv_data->deleted = 1;
6676 _scsih_ublock_io_device(ioc, pcie_device->wwid);
6677 sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE;
6678 }
6679
6680 ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n",
6681 pcie_device->handle, (u64)pcie_device->wwid);
6682 if (pcie_device->enclosure_handle != 0)
6683 ioc_info(ioc, "removing : enclosure logical id(0x%016llx), slot(%d)\n",
6684 (u64)pcie_device->enclosure_logical_id,
6685 pcie_device->slot);
6686 if (pcie_device->connector_name[0] != '\0')
6687 ioc_info(ioc, "removing: enclosure level(0x%04x), connector name( %s)\n",
6688 pcie_device->enclosure_level,
6689 pcie_device->connector_name);
6690
6691 if (pcie_device->starget && (pcie_device->access_status !=
6692 MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED))
6693 scsi_remove_target(&pcie_device->starget->dev);
6694 dewtprintk(ioc,
6695 ioc_info(ioc, "%s: exit: handle(0x%04x), wwid(0x%016llx)\n",
6696 __func__,
6697 pcie_device->handle, (u64)pcie_device->wwid));
6698 if (pcie_device->enclosure_handle != 0)
6699 dewtprintk(ioc,
6700 ioc_info(ioc, "%s: exit: enclosure logical id(0x%016llx), slot(%d)\n",
6701 __func__,
6702 (u64)pcie_device->enclosure_logical_id,
6703 pcie_device->slot));
6704 if (pcie_device->connector_name[0] != '\0')
6705 dewtprintk(ioc,
6706 ioc_info(ioc, "%s: exit: enclosure level(0x%04x), connector name( %s)\n",
6707 __func__,
6708 pcie_device->enclosure_level,
6709 pcie_device->connector_name));
6710
6711 kfree(pcie_device->serial_number);
6712 }
6713
6714
6715 /**
6716 * _scsih_pcie_check_device - checking device responsiveness
6717 * @ioc: per adapter object
6718 * @handle: attached device handle
6719 */
6720 static void
_scsih_pcie_check_device(struct MPT3SAS_ADAPTER * ioc,u16 handle)6721 _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle)
6722 {
6723 Mpi2ConfigReply_t mpi_reply;
6724 Mpi26PCIeDevicePage0_t pcie_device_pg0;
6725 u32 ioc_status;
6726 struct _pcie_device *pcie_device;
6727 u64 wwid;
6728 unsigned long flags;
6729 struct scsi_target *starget;
6730 struct MPT3SAS_TARGET *sas_target_priv_data;
6731 u32 device_info;
6732
6733 if ((mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply,
6734 &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle)))
6735 return;
6736
6737 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
6738 if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
6739 return;
6740
6741 /* check if this is end device */
6742 device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo);
6743 if (!(_scsih_is_nvme_pciescsi_device(device_info)))
6744 return;
6745
6746 wwid = le64_to_cpu(pcie_device_pg0.WWID);
6747 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
6748 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid);
6749
6750 if (!pcie_device) {
6751 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
6752 return;
6753 }
6754
6755 if (unlikely(pcie_device->handle != handle)) {
6756 starget = pcie_device->starget;
6757 sas_target_priv_data = starget->hostdata;
6758 pcie_device->access_status = pcie_device_pg0.AccessStatus;
6759 starget_printk(KERN_INFO, starget,
6760 "handle changed from(0x%04x) to (0x%04x)!!!\n",
6761 pcie_device->handle, handle);
6762 sas_target_priv_data->handle = handle;
6763 pcie_device->handle = handle;
6764
6765 if (le32_to_cpu(pcie_device_pg0.Flags) &
6766 MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) {
6767 pcie_device->enclosure_level =
6768 pcie_device_pg0.EnclosureLevel;
6769 memcpy(&pcie_device->connector_name[0],
6770 &pcie_device_pg0.ConnectorName[0], 4);
6771 } else {
6772 pcie_device->enclosure_level = 0;
6773 pcie_device->connector_name[0] = '\0';
6774 }
6775 }
6776
6777 /* check if device is present */
6778 if (!(le32_to_cpu(pcie_device_pg0.Flags) &
6779 MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) {
6780 ioc_info(ioc, "device is not present handle(0x%04x), flags!!!\n",
6781 handle);
6782 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
6783 pcie_device_put(pcie_device);
6784 return;
6785 }
6786
6787 /* check if there were any issues with discovery */
6788 if (_scsih_check_pcie_access_status(ioc, wwid, handle,
6789 pcie_device_pg0.AccessStatus)) {
6790 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
6791 pcie_device_put(pcie_device);
6792 return;
6793 }
6794
6795 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
6796 pcie_device_put(pcie_device);
6797
6798 _scsih_ublock_io_device(ioc, wwid);
6799
6800 return;
6801 }
6802
6803 /**
6804 * _scsih_pcie_add_device - creating pcie device object
6805 * @ioc: per adapter object
6806 * @handle: pcie device handle
6807 *
6808 * Creating end device object, stored in ioc->pcie_device_list.
6809 *
6810 * Return: 1 means queue the event later, 0 means complete the event
6811 */
6812 static int
_scsih_pcie_add_device(struct MPT3SAS_ADAPTER * ioc,u16 handle)6813 _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle)
6814 {
6815 Mpi26PCIeDevicePage0_t pcie_device_pg0;
6816 Mpi26PCIeDevicePage2_t pcie_device_pg2;
6817 Mpi2ConfigReply_t mpi_reply;
6818 struct _pcie_device *pcie_device;
6819 struct _enclosure_node *enclosure_dev;
6820 u32 ioc_status;
6821 u64 wwid;
6822
6823 if ((mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply,
6824 &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle))) {
6825 ioc_err(ioc, "failure at %s:%d/%s()!\n",
6826 __FILE__, __LINE__, __func__);
6827 return 0;
6828 }
6829 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
6830 MPI2_IOCSTATUS_MASK;
6831 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
6832 ioc_err(ioc, "failure at %s:%d/%s()!\n",
6833 __FILE__, __LINE__, __func__);
6834 return 0;
6835 }
6836
6837 set_bit(handle, ioc->pend_os_device_add);
6838 wwid = le64_to_cpu(pcie_device_pg0.WWID);
6839
6840 /* check if device is present */
6841 if (!(le32_to_cpu(pcie_device_pg0.Flags) &
6842 MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) {
6843 ioc_err(ioc, "device is not present handle(0x04%x)!!!\n",
6844 handle);
6845 return 0;
6846 }
6847
6848 /* check if there were any issues with discovery */
6849 if (_scsih_check_pcie_access_status(ioc, wwid, handle,
6850 pcie_device_pg0.AccessStatus))
6851 return 0;
6852
6853 if (!(_scsih_is_nvme_pciescsi_device(le32_to_cpu
6854 (pcie_device_pg0.DeviceInfo))))
6855 return 0;
6856
6857 pcie_device = mpt3sas_get_pdev_by_wwid(ioc, wwid);
6858 if (pcie_device) {
6859 clear_bit(handle, ioc->pend_os_device_add);
6860 pcie_device_put(pcie_device);
6861 return 0;
6862 }
6863
6864 /* PCIe Device Page 2 contains read-only information about a
6865 * specific NVMe device; therefore, this page is only
6866 * valid for NVMe devices and skip for pcie devices of type scsi.
6867 */
6868 if (!(mpt3sas_scsih_is_pcie_scsi_device(
6869 le32_to_cpu(pcie_device_pg0.DeviceInfo)))) {
6870 if (mpt3sas_config_get_pcie_device_pg2(ioc, &mpi_reply,
6871 &pcie_device_pg2, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE,
6872 handle)) {
6873 ioc_err(ioc,
6874 "failure at %s:%d/%s()!\n", __FILE__,
6875 __LINE__, __func__);
6876 return 0;
6877 }
6878
6879 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
6880 MPI2_IOCSTATUS_MASK;
6881 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
6882 ioc_err(ioc,
6883 "failure at %s:%d/%s()!\n", __FILE__,
6884 __LINE__, __func__);
6885 return 0;
6886 }
6887 }
6888
6889 pcie_device = kzalloc(sizeof(struct _pcie_device), GFP_KERNEL);
6890 if (!pcie_device) {
6891 ioc_err(ioc, "failure at %s:%d/%s()!\n",
6892 __FILE__, __LINE__, __func__);
6893 return 0;
6894 }
6895
6896 kref_init(&pcie_device->refcount);
6897 pcie_device->id = ioc->pcie_target_id++;
6898 pcie_device->channel = PCIE_CHANNEL;
6899 pcie_device->handle = handle;
6900 pcie_device->access_status = pcie_device_pg0.AccessStatus;
6901 pcie_device->device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo);
6902 pcie_device->wwid = wwid;
6903 pcie_device->port_num = pcie_device_pg0.PortNum;
6904 pcie_device->fast_path = (le32_to_cpu(pcie_device_pg0.Flags) &
6905 MPI26_PCIEDEV0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0;
6906
6907 pcie_device->enclosure_handle =
6908 le16_to_cpu(pcie_device_pg0.EnclosureHandle);
6909 if (pcie_device->enclosure_handle != 0)
6910 pcie_device->slot = le16_to_cpu(pcie_device_pg0.Slot);
6911
6912 if (le32_to_cpu(pcie_device_pg0.Flags) &
6913 MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) {
6914 pcie_device->enclosure_level = pcie_device_pg0.EnclosureLevel;
6915 memcpy(&pcie_device->connector_name[0],
6916 &pcie_device_pg0.ConnectorName[0], 4);
6917 } else {
6918 pcie_device->enclosure_level = 0;
6919 pcie_device->connector_name[0] = '\0';
6920 }
6921
6922 /* get enclosure_logical_id */
6923 if (pcie_device->enclosure_handle) {
6924 enclosure_dev =
6925 mpt3sas_scsih_enclosure_find_by_handle(ioc,
6926 pcie_device->enclosure_handle);
6927 if (enclosure_dev)
6928 pcie_device->enclosure_logical_id =
6929 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
6930 }
6931 /* TODO -- Add device name once FW supports it */
6932 if (!(mpt3sas_scsih_is_pcie_scsi_device(
6933 le32_to_cpu(pcie_device_pg0.DeviceInfo)))) {
6934 pcie_device->nvme_mdts =
6935 le32_to_cpu(pcie_device_pg2.MaximumDataTransferSize);
6936 if (pcie_device_pg2.ControllerResetTO)
6937 pcie_device->reset_timeout =
6938 pcie_device_pg2.ControllerResetTO;
6939 else
6940 pcie_device->reset_timeout = 30;
6941 } else
6942 pcie_device->reset_timeout = 30;
6943
6944 if (ioc->wait_for_discovery_to_complete)
6945 _scsih_pcie_device_init_add(ioc, pcie_device);
6946 else
6947 _scsih_pcie_device_add(ioc, pcie_device);
6948
6949 pcie_device_put(pcie_device);
6950 return 0;
6951 }
6952
6953 /**
6954 * _scsih_pcie_topology_change_event_debug - debug for topology
6955 * event
6956 * @ioc: per adapter object
6957 * @event_data: event data payload
6958 * Context: user.
6959 */
6960 static void
_scsih_pcie_topology_change_event_debug(struct MPT3SAS_ADAPTER * ioc,Mpi26EventDataPCIeTopologyChangeList_t * event_data)6961 _scsih_pcie_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
6962 Mpi26EventDataPCIeTopologyChangeList_t *event_data)
6963 {
6964 int i;
6965 u16 handle;
6966 u16 reason_code;
6967 u8 port_number;
6968 char *status_str = NULL;
6969 u8 link_rate, prev_link_rate;
6970
6971 switch (event_data->SwitchStatus) {
6972 case MPI26_EVENT_PCIE_TOPO_SS_ADDED:
6973 status_str = "add";
6974 break;
6975 case MPI26_EVENT_PCIE_TOPO_SS_NOT_RESPONDING:
6976 status_str = "remove";
6977 break;
6978 case MPI26_EVENT_PCIE_TOPO_SS_RESPONDING:
6979 case 0:
6980 status_str = "responding";
6981 break;
6982 case MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING:
6983 status_str = "remove delay";
6984 break;
6985 default:
6986 status_str = "unknown status";
6987 break;
6988 }
6989 ioc_info(ioc, "pcie topology change: (%s)\n", status_str);
6990 pr_info("\tswitch_handle(0x%04x), enclosure_handle(0x%04x)"
6991 "start_port(%02d), count(%d)\n",
6992 le16_to_cpu(event_data->SwitchDevHandle),
6993 le16_to_cpu(event_data->EnclosureHandle),
6994 event_data->StartPortNum, event_data->NumEntries);
6995 for (i = 0; i < event_data->NumEntries; i++) {
6996 handle =
6997 le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle);
6998 if (!handle)
6999 continue;
7000 port_number = event_data->StartPortNum + i;
7001 reason_code = event_data->PortEntry[i].PortStatus;
7002 switch (reason_code) {
7003 case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED:
7004 status_str = "target add";
7005 break;
7006 case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING:
7007 status_str = "target remove";
7008 break;
7009 case MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING:
7010 status_str = "delay target remove";
7011 break;
7012 case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED:
7013 status_str = "link rate change";
7014 break;
7015 case MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE:
7016 status_str = "target responding";
7017 break;
7018 default:
7019 status_str = "unknown";
7020 break;
7021 }
7022 link_rate = event_data->PortEntry[i].CurrentPortInfo &
7023 MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK;
7024 prev_link_rate = event_data->PortEntry[i].PreviousPortInfo &
7025 MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK;
7026 pr_info("\tport(%02d), attached_handle(0x%04x): %s:"
7027 " link rate: new(0x%02x), old(0x%02x)\n", port_number,
7028 handle, status_str, link_rate, prev_link_rate);
7029 }
7030 }
7031
7032 /**
7033 * _scsih_pcie_topology_change_event - handle PCIe topology
7034 * changes
7035 * @ioc: per adapter object
7036 * @fw_event: The fw_event_work object
7037 * Context: user.
7038 *
7039 */
7040 static void
_scsih_pcie_topology_change_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)7041 _scsih_pcie_topology_change_event(struct MPT3SAS_ADAPTER *ioc,
7042 struct fw_event_work *fw_event)
7043 {
7044 int i;
7045 u16 handle;
7046 u16 reason_code;
7047 u8 link_rate, prev_link_rate;
7048 unsigned long flags;
7049 int rc;
7050 Mpi26EventDataPCIeTopologyChangeList_t *event_data =
7051 (Mpi26EventDataPCIeTopologyChangeList_t *) fw_event->event_data;
7052 struct _pcie_device *pcie_device;
7053
7054 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
7055 _scsih_pcie_topology_change_event_debug(ioc, event_data);
7056
7057 if (ioc->shost_recovery || ioc->remove_host ||
7058 ioc->pci_error_recovery)
7059 return;
7060
7061 if (fw_event->ignore) {
7062 dewtprintk(ioc, ioc_info(ioc, "ignoring switch event\n"));
7063 return;
7064 }
7065
7066 /* handle siblings events */
7067 for (i = 0; i < event_data->NumEntries; i++) {
7068 if (fw_event->ignore) {
7069 dewtprintk(ioc,
7070 ioc_info(ioc, "ignoring switch event\n"));
7071 return;
7072 }
7073 if (ioc->remove_host || ioc->pci_error_recovery)
7074 return;
7075 reason_code = event_data->PortEntry[i].PortStatus;
7076 handle =
7077 le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle);
7078 if (!handle)
7079 continue;
7080
7081 link_rate = event_data->PortEntry[i].CurrentPortInfo
7082 & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK;
7083 prev_link_rate = event_data->PortEntry[i].PreviousPortInfo
7084 & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK;
7085
7086 switch (reason_code) {
7087 case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED:
7088 if (ioc->shost_recovery)
7089 break;
7090 if (link_rate == prev_link_rate)
7091 break;
7092 if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5)
7093 break;
7094
7095 _scsih_pcie_check_device(ioc, handle);
7096
7097 /* This code after this point handles the test case
7098 * where a device has been added, however its returning
7099 * BUSY for sometime. Then before the Device Missing
7100 * Delay expires and the device becomes READY, the
7101 * device is removed and added back.
7102 */
7103 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
7104 pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle);
7105 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
7106
7107 if (pcie_device) {
7108 pcie_device_put(pcie_device);
7109 break;
7110 }
7111
7112 if (!test_bit(handle, ioc->pend_os_device_add))
7113 break;
7114
7115 dewtprintk(ioc,
7116 ioc_info(ioc, "handle(0x%04x) device not found: convert event to a device add\n",
7117 handle));
7118 event_data->PortEntry[i].PortStatus &= 0xF0;
7119 event_data->PortEntry[i].PortStatus |=
7120 MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED;
7121 /* fall through */
7122 case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED:
7123 if (ioc->shost_recovery)
7124 break;
7125 if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5)
7126 break;
7127
7128 rc = _scsih_pcie_add_device(ioc, handle);
7129 if (!rc) {
7130 /* mark entry vacant */
7131 /* TODO This needs to be reviewed and fixed,
7132 * we dont have an entry
7133 * to make an event void like vacant
7134 */
7135 event_data->PortEntry[i].PortStatus |=
7136 MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE;
7137 }
7138 break;
7139 case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING:
7140 _scsih_pcie_device_remove_by_handle(ioc, handle);
7141 break;
7142 }
7143 }
7144 }
7145
7146 /**
7147 * _scsih_pcie_device_status_change_event_debug - debug for device event
7148 * @ioc: ?
7149 * @event_data: event data payload
7150 * Context: user.
7151 */
7152 static void
_scsih_pcie_device_status_change_event_debug(struct MPT3SAS_ADAPTER * ioc,Mpi26EventDataPCIeDeviceStatusChange_t * event_data)7153 _scsih_pcie_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
7154 Mpi26EventDataPCIeDeviceStatusChange_t *event_data)
7155 {
7156 char *reason_str = NULL;
7157
7158 switch (event_data->ReasonCode) {
7159 case MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA:
7160 reason_str = "smart data";
7161 break;
7162 case MPI26_EVENT_PCIDEV_STAT_RC_UNSUPPORTED:
7163 reason_str = "unsupported device discovered";
7164 break;
7165 case MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET:
7166 reason_str = "internal device reset";
7167 break;
7168 case MPI26_EVENT_PCIDEV_STAT_RC_TASK_ABORT_INTERNAL:
7169 reason_str = "internal task abort";
7170 break;
7171 case MPI26_EVENT_PCIDEV_STAT_RC_ABORT_TASK_SET_INTERNAL:
7172 reason_str = "internal task abort set";
7173 break;
7174 case MPI26_EVENT_PCIDEV_STAT_RC_CLEAR_TASK_SET_INTERNAL:
7175 reason_str = "internal clear task set";
7176 break;
7177 case MPI26_EVENT_PCIDEV_STAT_RC_QUERY_TASK_INTERNAL:
7178 reason_str = "internal query task";
7179 break;
7180 case MPI26_EVENT_PCIDEV_STAT_RC_DEV_INIT_FAILURE:
7181 reason_str = "device init failure";
7182 break;
7183 case MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET:
7184 reason_str = "internal device reset complete";
7185 break;
7186 case MPI26_EVENT_PCIDEV_STAT_RC_CMP_TASK_ABORT_INTERNAL:
7187 reason_str = "internal task abort complete";
7188 break;
7189 case MPI26_EVENT_PCIDEV_STAT_RC_ASYNC_NOTIFICATION:
7190 reason_str = "internal async notification";
7191 break;
7192 case MPI26_EVENT_PCIDEV_STAT_RC_PCIE_HOT_RESET_FAILED:
7193 reason_str = "pcie hot reset failed";
7194 break;
7195 default:
7196 reason_str = "unknown reason";
7197 break;
7198 }
7199
7200 ioc_info(ioc, "PCIE device status change: (%s)\n"
7201 "\thandle(0x%04x), WWID(0x%016llx), tag(%d)",
7202 reason_str, le16_to_cpu(event_data->DevHandle),
7203 (u64)le64_to_cpu(event_data->WWID),
7204 le16_to_cpu(event_data->TaskTag));
7205 if (event_data->ReasonCode == MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA)
7206 pr_cont(", ASC(0x%x), ASCQ(0x%x)\n",
7207 event_data->ASC, event_data->ASCQ);
7208 pr_cont("\n");
7209 }
7210
7211 /**
7212 * _scsih_pcie_device_status_change_event - handle device status
7213 * change
7214 * @ioc: per adapter object
7215 * @fw_event: The fw_event_work object
7216 * Context: user.
7217 */
7218 static void
_scsih_pcie_device_status_change_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)7219 _scsih_pcie_device_status_change_event(struct MPT3SAS_ADAPTER *ioc,
7220 struct fw_event_work *fw_event)
7221 {
7222 struct MPT3SAS_TARGET *target_priv_data;
7223 struct _pcie_device *pcie_device;
7224 u64 wwid;
7225 unsigned long flags;
7226 Mpi26EventDataPCIeDeviceStatusChange_t *event_data =
7227 (Mpi26EventDataPCIeDeviceStatusChange_t *)fw_event->event_data;
7228 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
7229 _scsih_pcie_device_status_change_event_debug(ioc,
7230 event_data);
7231
7232 if (event_data->ReasonCode !=
7233 MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET &&
7234 event_data->ReasonCode !=
7235 MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET)
7236 return;
7237
7238 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
7239 wwid = le64_to_cpu(event_data->WWID);
7240 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid);
7241
7242 if (!pcie_device || !pcie_device->starget)
7243 goto out;
7244
7245 target_priv_data = pcie_device->starget->hostdata;
7246 if (!target_priv_data)
7247 goto out;
7248
7249 if (event_data->ReasonCode ==
7250 MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET)
7251 target_priv_data->tm_busy = 1;
7252 else
7253 target_priv_data->tm_busy = 0;
7254 out:
7255 if (pcie_device)
7256 pcie_device_put(pcie_device);
7257
7258 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
7259 }
7260
7261 /**
7262 * _scsih_sas_enclosure_dev_status_change_event_debug - debug for enclosure
7263 * event
7264 * @ioc: per adapter object
7265 * @event_data: event data payload
7266 * Context: user.
7267 */
7268 static void
_scsih_sas_enclosure_dev_status_change_event_debug(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataSasEnclDevStatusChange_t * event_data)7269 _scsih_sas_enclosure_dev_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
7270 Mpi2EventDataSasEnclDevStatusChange_t *event_data)
7271 {
7272 char *reason_str = NULL;
7273
7274 switch (event_data->ReasonCode) {
7275 case MPI2_EVENT_SAS_ENCL_RC_ADDED:
7276 reason_str = "enclosure add";
7277 break;
7278 case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING:
7279 reason_str = "enclosure remove";
7280 break;
7281 default:
7282 reason_str = "unknown reason";
7283 break;
7284 }
7285
7286 ioc_info(ioc, "enclosure status change: (%s)\n"
7287 "\thandle(0x%04x), enclosure logical id(0x%016llx) number slots(%d)\n",
7288 reason_str,
7289 le16_to_cpu(event_data->EnclosureHandle),
7290 (u64)le64_to_cpu(event_data->EnclosureLogicalID),
7291 le16_to_cpu(event_data->StartSlot));
7292 }
7293
7294 /**
7295 * _scsih_sas_enclosure_dev_status_change_event - handle enclosure events
7296 * @ioc: per adapter object
7297 * @fw_event: The fw_event_work object
7298 * Context: user.
7299 */
7300 static void
_scsih_sas_enclosure_dev_status_change_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)7301 _scsih_sas_enclosure_dev_status_change_event(struct MPT3SAS_ADAPTER *ioc,
7302 struct fw_event_work *fw_event)
7303 {
7304 Mpi2ConfigReply_t mpi_reply;
7305 struct _enclosure_node *enclosure_dev = NULL;
7306 Mpi2EventDataSasEnclDevStatusChange_t *event_data =
7307 (Mpi2EventDataSasEnclDevStatusChange_t *)fw_event->event_data;
7308 int rc;
7309 u16 enclosure_handle = le16_to_cpu(event_data->EnclosureHandle);
7310
7311 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
7312 _scsih_sas_enclosure_dev_status_change_event_debug(ioc,
7313 (Mpi2EventDataSasEnclDevStatusChange_t *)
7314 fw_event->event_data);
7315 if (ioc->shost_recovery)
7316 return;
7317
7318 if (enclosure_handle)
7319 enclosure_dev =
7320 mpt3sas_scsih_enclosure_find_by_handle(ioc,
7321 enclosure_handle);
7322 switch (event_data->ReasonCode) {
7323 case MPI2_EVENT_SAS_ENCL_RC_ADDED:
7324 if (!enclosure_dev) {
7325 enclosure_dev =
7326 kzalloc(sizeof(struct _enclosure_node),
7327 GFP_KERNEL);
7328 if (!enclosure_dev) {
7329 ioc_info(ioc, "failure at %s:%d/%s()!\n",
7330 __FILE__, __LINE__, __func__);
7331 return;
7332 }
7333 rc = mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply,
7334 &enclosure_dev->pg0,
7335 MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE,
7336 enclosure_handle);
7337
7338 if (rc || (le16_to_cpu(mpi_reply.IOCStatus) &
7339 MPI2_IOCSTATUS_MASK)) {
7340 kfree(enclosure_dev);
7341 return;
7342 }
7343
7344 list_add_tail(&enclosure_dev->list,
7345 &ioc->enclosure_list);
7346 }
7347 break;
7348 case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING:
7349 if (enclosure_dev) {
7350 list_del(&enclosure_dev->list);
7351 kfree(enclosure_dev);
7352 }
7353 break;
7354 default:
7355 break;
7356 }
7357 }
7358
7359 /**
7360 * _scsih_sas_broadcast_primitive_event - handle broadcast events
7361 * @ioc: per adapter object
7362 * @fw_event: The fw_event_work object
7363 * Context: user.
7364 */
7365 static void
_scsih_sas_broadcast_primitive_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)7366 _scsih_sas_broadcast_primitive_event(struct MPT3SAS_ADAPTER *ioc,
7367 struct fw_event_work *fw_event)
7368 {
7369 struct scsi_cmnd *scmd;
7370 struct scsi_device *sdev;
7371 struct scsiio_tracker *st;
7372 u16 smid, handle;
7373 u32 lun;
7374 struct MPT3SAS_DEVICE *sas_device_priv_data;
7375 u32 termination_count;
7376 u32 query_count;
7377 Mpi2SCSITaskManagementReply_t *mpi_reply;
7378 Mpi2EventDataSasBroadcastPrimitive_t *event_data =
7379 (Mpi2EventDataSasBroadcastPrimitive_t *)
7380 fw_event->event_data;
7381 u16 ioc_status;
7382 unsigned long flags;
7383 int r;
7384 u8 max_retries = 0;
7385 u8 task_abort_retries;
7386
7387 mutex_lock(&ioc->tm_cmds.mutex);
7388 ioc_info(ioc, "%s: enter: phy number(%d), width(%d)\n",
7389 __func__, event_data->PhyNum, event_data->PortWidth);
7390
7391 _scsih_block_io_all_device(ioc);
7392
7393 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
7394 mpi_reply = ioc->tm_cmds.reply;
7395 broadcast_aen_retry:
7396
7397 /* sanity checks for retrying this loop */
7398 if (max_retries++ == 5) {
7399 dewtprintk(ioc, ioc_info(ioc, "%s: giving up\n", __func__));
7400 goto out;
7401 } else if (max_retries > 1)
7402 dewtprintk(ioc,
7403 ioc_info(ioc, "%s: %d retry\n",
7404 __func__, max_retries - 1));
7405
7406 termination_count = 0;
7407 query_count = 0;
7408 for (smid = 1; smid <= ioc->scsiio_depth; smid++) {
7409 if (ioc->shost_recovery)
7410 goto out;
7411 scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid);
7412 if (!scmd)
7413 continue;
7414 st = scsi_cmd_priv(scmd);
7415 sdev = scmd->device;
7416 sas_device_priv_data = sdev->hostdata;
7417 if (!sas_device_priv_data || !sas_device_priv_data->sas_target)
7418 continue;
7419 /* skip hidden raid components */
7420 if (sas_device_priv_data->sas_target->flags &
7421 MPT_TARGET_FLAGS_RAID_COMPONENT)
7422 continue;
7423 /* skip volumes */
7424 if (sas_device_priv_data->sas_target->flags &
7425 MPT_TARGET_FLAGS_VOLUME)
7426 continue;
7427 /* skip PCIe devices */
7428 if (sas_device_priv_data->sas_target->flags &
7429 MPT_TARGET_FLAGS_PCIE_DEVICE)
7430 continue;
7431
7432 handle = sas_device_priv_data->sas_target->handle;
7433 lun = sas_device_priv_data->lun;
7434 query_count++;
7435
7436 if (ioc->shost_recovery)
7437 goto out;
7438
7439 spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags);
7440 r = mpt3sas_scsih_issue_tm(ioc, handle, lun,
7441 MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK, st->smid,
7442 st->msix_io, 30, 0);
7443 if (r == FAILED) {
7444 sdev_printk(KERN_WARNING, sdev,
7445 "mpt3sas_scsih_issue_tm: FAILED when sending "
7446 "QUERY_TASK: scmd(%p)\n", scmd);
7447 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
7448 goto broadcast_aen_retry;
7449 }
7450 ioc_status = le16_to_cpu(mpi_reply->IOCStatus)
7451 & MPI2_IOCSTATUS_MASK;
7452 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
7453 sdev_printk(KERN_WARNING, sdev,
7454 "query task: FAILED with IOCSTATUS(0x%04x), scmd(%p)\n",
7455 ioc_status, scmd);
7456 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
7457 goto broadcast_aen_retry;
7458 }
7459
7460 /* see if IO is still owned by IOC and target */
7461 if (mpi_reply->ResponseCode ==
7462 MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED ||
7463 mpi_reply->ResponseCode ==
7464 MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC) {
7465 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
7466 continue;
7467 }
7468 task_abort_retries = 0;
7469 tm_retry:
7470 if (task_abort_retries++ == 60) {
7471 dewtprintk(ioc,
7472 ioc_info(ioc, "%s: ABORT_TASK: giving up\n",
7473 __func__));
7474 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
7475 goto broadcast_aen_retry;
7476 }
7477
7478 if (ioc->shost_recovery)
7479 goto out_no_lock;
7480
7481 r = mpt3sas_scsih_issue_tm(ioc, handle, sdev->lun,
7482 MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK, st->smid,
7483 st->msix_io, 30, 0);
7484 if (r == FAILED || st->cb_idx != 0xFF) {
7485 sdev_printk(KERN_WARNING, sdev,
7486 "mpt3sas_scsih_issue_tm: ABORT_TASK: FAILED : "
7487 "scmd(%p)\n", scmd);
7488 goto tm_retry;
7489 }
7490
7491 if (task_abort_retries > 1)
7492 sdev_printk(KERN_WARNING, sdev,
7493 "mpt3sas_scsih_issue_tm: ABORT_TASK: RETRIES (%d):"
7494 " scmd(%p)\n",
7495 task_abort_retries - 1, scmd);
7496
7497 termination_count += le32_to_cpu(mpi_reply->TerminationCount);
7498 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
7499 }
7500
7501 if (ioc->broadcast_aen_pending) {
7502 dewtprintk(ioc,
7503 ioc_info(ioc,
7504 "%s: loop back due to pending AEN\n",
7505 __func__));
7506 ioc->broadcast_aen_pending = 0;
7507 goto broadcast_aen_retry;
7508 }
7509
7510 out:
7511 spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags);
7512 out_no_lock:
7513
7514 dewtprintk(ioc,
7515 ioc_info(ioc, "%s - exit, query_count = %d termination_count = %d\n",
7516 __func__, query_count, termination_count));
7517
7518 ioc->broadcast_aen_busy = 0;
7519 if (!ioc->shost_recovery)
7520 _scsih_ublock_io_all_device(ioc);
7521 mutex_unlock(&ioc->tm_cmds.mutex);
7522 }
7523
7524 /**
7525 * _scsih_sas_discovery_event - handle discovery events
7526 * @ioc: per adapter object
7527 * @fw_event: The fw_event_work object
7528 * Context: user.
7529 */
7530 static void
_scsih_sas_discovery_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)7531 _scsih_sas_discovery_event(struct MPT3SAS_ADAPTER *ioc,
7532 struct fw_event_work *fw_event)
7533 {
7534 Mpi2EventDataSasDiscovery_t *event_data =
7535 (Mpi2EventDataSasDiscovery_t *) fw_event->event_data;
7536
7537 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) {
7538 ioc_info(ioc, "discovery event: (%s)",
7539 event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED ?
7540 "start" : "stop");
7541 if (event_data->DiscoveryStatus)
7542 pr_cont("discovery_status(0x%08x)",
7543 le32_to_cpu(event_data->DiscoveryStatus));
7544 pr_cont("\n");
7545 }
7546
7547 if (event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED &&
7548 !ioc->sas_hba.num_phys) {
7549 if (disable_discovery > 0 && ioc->shost_recovery) {
7550 /* Wait for the reset to complete */
7551 while (ioc->shost_recovery)
7552 ssleep(1);
7553 }
7554 _scsih_sas_host_add(ioc);
7555 }
7556 }
7557
7558 /**
7559 * _scsih_sas_device_discovery_error_event - display SAS device discovery error
7560 * events
7561 * @ioc: per adapter object
7562 * @fw_event: The fw_event_work object
7563 * Context: user.
7564 */
7565 static void
_scsih_sas_device_discovery_error_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)7566 _scsih_sas_device_discovery_error_event(struct MPT3SAS_ADAPTER *ioc,
7567 struct fw_event_work *fw_event)
7568 {
7569 Mpi25EventDataSasDeviceDiscoveryError_t *event_data =
7570 (Mpi25EventDataSasDeviceDiscoveryError_t *)fw_event->event_data;
7571
7572 switch (event_data->ReasonCode) {
7573 case MPI25_EVENT_SAS_DISC_ERR_SMP_FAILED:
7574 ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has failed\n",
7575 le16_to_cpu(event_data->DevHandle),
7576 (u64)le64_to_cpu(event_data->SASAddress),
7577 event_data->PhysicalPort);
7578 break;
7579 case MPI25_EVENT_SAS_DISC_ERR_SMP_TIMEOUT:
7580 ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has timed out\n",
7581 le16_to_cpu(event_data->DevHandle),
7582 (u64)le64_to_cpu(event_data->SASAddress),
7583 event_data->PhysicalPort);
7584 break;
7585 default:
7586 break;
7587 }
7588 }
7589
7590 /**
7591 * _scsih_pcie_enumeration_event - handle enumeration events
7592 * @ioc: per adapter object
7593 * @fw_event: The fw_event_work object
7594 * Context: user.
7595 */
7596 static void
_scsih_pcie_enumeration_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)7597 _scsih_pcie_enumeration_event(struct MPT3SAS_ADAPTER *ioc,
7598 struct fw_event_work *fw_event)
7599 {
7600 Mpi26EventDataPCIeEnumeration_t *event_data =
7601 (Mpi26EventDataPCIeEnumeration_t *)fw_event->event_data;
7602
7603 if (!(ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK))
7604 return;
7605
7606 ioc_info(ioc, "pcie enumeration event: (%s) Flag 0x%02x",
7607 (event_data->ReasonCode == MPI26_EVENT_PCIE_ENUM_RC_STARTED) ?
7608 "started" : "completed",
7609 event_data->Flags);
7610 if (event_data->EnumerationStatus)
7611 pr_cont("enumeration_status(0x%08x)",
7612 le32_to_cpu(event_data->EnumerationStatus));
7613 pr_cont("\n");
7614 }
7615
7616 /**
7617 * _scsih_ir_fastpath - turn on fastpath for IR physdisk
7618 * @ioc: per adapter object
7619 * @handle: device handle for physical disk
7620 * @phys_disk_num: physical disk number
7621 *
7622 * Return: 0 for success, else failure.
7623 */
7624 static int
_scsih_ir_fastpath(struct MPT3SAS_ADAPTER * ioc,u16 handle,u8 phys_disk_num)7625 _scsih_ir_fastpath(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phys_disk_num)
7626 {
7627 Mpi2RaidActionRequest_t *mpi_request;
7628 Mpi2RaidActionReply_t *mpi_reply;
7629 u16 smid;
7630 u8 issue_reset = 0;
7631 int rc = 0;
7632 u16 ioc_status;
7633 u32 log_info;
7634
7635 if (ioc->hba_mpi_version_belonged == MPI2_VERSION)
7636 return rc;
7637
7638 mutex_lock(&ioc->scsih_cmds.mutex);
7639
7640 if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) {
7641 ioc_err(ioc, "%s: scsih_cmd in use\n", __func__);
7642 rc = -EAGAIN;
7643 goto out;
7644 }
7645 ioc->scsih_cmds.status = MPT3_CMD_PENDING;
7646
7647 smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx);
7648 if (!smid) {
7649 ioc_err(ioc, "%s: failed obtaining a smid\n", __func__);
7650 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
7651 rc = -EAGAIN;
7652 goto out;
7653 }
7654
7655 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
7656 ioc->scsih_cmds.smid = smid;
7657 memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t));
7658
7659 mpi_request->Function = MPI2_FUNCTION_RAID_ACTION;
7660 mpi_request->Action = MPI2_RAID_ACTION_PHYSDISK_HIDDEN;
7661 mpi_request->PhysDiskNum = phys_disk_num;
7662
7663 dewtprintk(ioc,
7664 ioc_info(ioc, "IR RAID_ACTION: turning fast path on for handle(0x%04x), phys_disk_num (0x%02x)\n",
7665 handle, phys_disk_num));
7666
7667 init_completion(&ioc->scsih_cmds.done);
7668 ioc->put_smid_default(ioc, smid);
7669 wait_for_completion_timeout(&ioc->scsih_cmds.done, 10*HZ);
7670
7671 if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) {
7672 issue_reset =
7673 mpt3sas_base_check_cmd_timeout(ioc,
7674 ioc->scsih_cmds.status, mpi_request,
7675 sizeof(Mpi2RaidActionRequest_t)/4);
7676 rc = -EFAULT;
7677 goto out;
7678 }
7679
7680 if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) {
7681
7682 mpi_reply = ioc->scsih_cmds.reply;
7683 ioc_status = le16_to_cpu(mpi_reply->IOCStatus);
7684 if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE)
7685 log_info = le32_to_cpu(mpi_reply->IOCLogInfo);
7686 else
7687 log_info = 0;
7688 ioc_status &= MPI2_IOCSTATUS_MASK;
7689 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
7690 dewtprintk(ioc,
7691 ioc_info(ioc, "IR RAID_ACTION: failed: ioc_status(0x%04x), loginfo(0x%08x)!!!\n",
7692 ioc_status, log_info));
7693 rc = -EFAULT;
7694 } else
7695 dewtprintk(ioc,
7696 ioc_info(ioc, "IR RAID_ACTION: completed successfully\n"));
7697 }
7698
7699 out:
7700 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
7701 mutex_unlock(&ioc->scsih_cmds.mutex);
7702
7703 if (issue_reset)
7704 mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
7705 return rc;
7706 }
7707
7708 /**
7709 * _scsih_reprobe_lun - reprobing lun
7710 * @sdev: scsi device struct
7711 * @no_uld_attach: sdev->no_uld_attach flag setting
7712 *
7713 **/
7714 static void
_scsih_reprobe_lun(struct scsi_device * sdev,void * no_uld_attach)7715 _scsih_reprobe_lun(struct scsi_device *sdev, void *no_uld_attach)
7716 {
7717 sdev->no_uld_attach = no_uld_attach ? 1 : 0;
7718 sdev_printk(KERN_INFO, sdev, "%s raid component\n",
7719 sdev->no_uld_attach ? "hiding" : "exposing");
7720 WARN_ON(scsi_device_reprobe(sdev));
7721 }
7722
7723 /**
7724 * _scsih_sas_volume_add - add new volume
7725 * @ioc: per adapter object
7726 * @element: IR config element data
7727 * Context: user.
7728 */
7729 static void
_scsih_sas_volume_add(struct MPT3SAS_ADAPTER * ioc,Mpi2EventIrConfigElement_t * element)7730 _scsih_sas_volume_add(struct MPT3SAS_ADAPTER *ioc,
7731 Mpi2EventIrConfigElement_t *element)
7732 {
7733 struct _raid_device *raid_device;
7734 unsigned long flags;
7735 u64 wwid;
7736 u16 handle = le16_to_cpu(element->VolDevHandle);
7737 int rc;
7738
7739 mpt3sas_config_get_volume_wwid(ioc, handle, &wwid);
7740 if (!wwid) {
7741 ioc_err(ioc, "failure at %s:%d/%s()!\n",
7742 __FILE__, __LINE__, __func__);
7743 return;
7744 }
7745
7746 spin_lock_irqsave(&ioc->raid_device_lock, flags);
7747 raid_device = _scsih_raid_device_find_by_wwid(ioc, wwid);
7748 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
7749
7750 if (raid_device)
7751 return;
7752
7753 raid_device = kzalloc(sizeof(struct _raid_device), GFP_KERNEL);
7754 if (!raid_device) {
7755 ioc_err(ioc, "failure at %s:%d/%s()!\n",
7756 __FILE__, __LINE__, __func__);
7757 return;
7758 }
7759
7760 raid_device->id = ioc->sas_id++;
7761 raid_device->channel = RAID_CHANNEL;
7762 raid_device->handle = handle;
7763 raid_device->wwid = wwid;
7764 _scsih_raid_device_add(ioc, raid_device);
7765 if (!ioc->wait_for_discovery_to_complete) {
7766 rc = scsi_add_device(ioc->shost, RAID_CHANNEL,
7767 raid_device->id, 0);
7768 if (rc)
7769 _scsih_raid_device_remove(ioc, raid_device);
7770 } else {
7771 spin_lock_irqsave(&ioc->raid_device_lock, flags);
7772 _scsih_determine_boot_device(ioc, raid_device, 1);
7773 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
7774 }
7775 }
7776
7777 /**
7778 * _scsih_sas_volume_delete - delete volume
7779 * @ioc: per adapter object
7780 * @handle: volume device handle
7781 * Context: user.
7782 */
7783 static void
_scsih_sas_volume_delete(struct MPT3SAS_ADAPTER * ioc,u16 handle)7784 _scsih_sas_volume_delete(struct MPT3SAS_ADAPTER *ioc, u16 handle)
7785 {
7786 struct _raid_device *raid_device;
7787 unsigned long flags;
7788 struct MPT3SAS_TARGET *sas_target_priv_data;
7789 struct scsi_target *starget = NULL;
7790
7791 spin_lock_irqsave(&ioc->raid_device_lock, flags);
7792 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
7793 if (raid_device) {
7794 if (raid_device->starget) {
7795 starget = raid_device->starget;
7796 sas_target_priv_data = starget->hostdata;
7797 sas_target_priv_data->deleted = 1;
7798 }
7799 ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n",
7800 raid_device->handle, (u64)raid_device->wwid);
7801 list_del(&raid_device->list);
7802 kfree(raid_device);
7803 }
7804 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
7805 if (starget)
7806 scsi_remove_target(&starget->dev);
7807 }
7808
7809 /**
7810 * _scsih_sas_pd_expose - expose pd component to /dev/sdX
7811 * @ioc: per adapter object
7812 * @element: IR config element data
7813 * Context: user.
7814 */
7815 static void
_scsih_sas_pd_expose(struct MPT3SAS_ADAPTER * ioc,Mpi2EventIrConfigElement_t * element)7816 _scsih_sas_pd_expose(struct MPT3SAS_ADAPTER *ioc,
7817 Mpi2EventIrConfigElement_t *element)
7818 {
7819 struct _sas_device *sas_device;
7820 struct scsi_target *starget = NULL;
7821 struct MPT3SAS_TARGET *sas_target_priv_data;
7822 unsigned long flags;
7823 u16 handle = le16_to_cpu(element->PhysDiskDevHandle);
7824
7825 spin_lock_irqsave(&ioc->sas_device_lock, flags);
7826 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
7827 if (sas_device) {
7828 sas_device->volume_handle = 0;
7829 sas_device->volume_wwid = 0;
7830 clear_bit(handle, ioc->pd_handles);
7831 if (sas_device->starget && sas_device->starget->hostdata) {
7832 starget = sas_device->starget;
7833 sas_target_priv_data = starget->hostdata;
7834 sas_target_priv_data->flags &=
7835 ~MPT_TARGET_FLAGS_RAID_COMPONENT;
7836 }
7837 }
7838 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
7839 if (!sas_device)
7840 return;
7841
7842 /* exposing raid component */
7843 if (starget)
7844 starget_for_each_device(starget, NULL, _scsih_reprobe_lun);
7845
7846 sas_device_put(sas_device);
7847 }
7848
7849 /**
7850 * _scsih_sas_pd_hide - hide pd component from /dev/sdX
7851 * @ioc: per adapter object
7852 * @element: IR config element data
7853 * Context: user.
7854 */
7855 static void
_scsih_sas_pd_hide(struct MPT3SAS_ADAPTER * ioc,Mpi2EventIrConfigElement_t * element)7856 _scsih_sas_pd_hide(struct MPT3SAS_ADAPTER *ioc,
7857 Mpi2EventIrConfigElement_t *element)
7858 {
7859 struct _sas_device *sas_device;
7860 struct scsi_target *starget = NULL;
7861 struct MPT3SAS_TARGET *sas_target_priv_data;
7862 unsigned long flags;
7863 u16 handle = le16_to_cpu(element->PhysDiskDevHandle);
7864 u16 volume_handle = 0;
7865 u64 volume_wwid = 0;
7866
7867 mpt3sas_config_get_volume_handle(ioc, handle, &volume_handle);
7868 if (volume_handle)
7869 mpt3sas_config_get_volume_wwid(ioc, volume_handle,
7870 &volume_wwid);
7871
7872 spin_lock_irqsave(&ioc->sas_device_lock, flags);
7873 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
7874 if (sas_device) {
7875 set_bit(handle, ioc->pd_handles);
7876 if (sas_device->starget && sas_device->starget->hostdata) {
7877 starget = sas_device->starget;
7878 sas_target_priv_data = starget->hostdata;
7879 sas_target_priv_data->flags |=
7880 MPT_TARGET_FLAGS_RAID_COMPONENT;
7881 sas_device->volume_handle = volume_handle;
7882 sas_device->volume_wwid = volume_wwid;
7883 }
7884 }
7885 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
7886 if (!sas_device)
7887 return;
7888
7889 /* hiding raid component */
7890 _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum);
7891
7892 if (starget)
7893 starget_for_each_device(starget, (void *)1, _scsih_reprobe_lun);
7894
7895 sas_device_put(sas_device);
7896 }
7897
7898 /**
7899 * _scsih_sas_pd_delete - delete pd component
7900 * @ioc: per adapter object
7901 * @element: IR config element data
7902 * Context: user.
7903 */
7904 static void
_scsih_sas_pd_delete(struct MPT3SAS_ADAPTER * ioc,Mpi2EventIrConfigElement_t * element)7905 _scsih_sas_pd_delete(struct MPT3SAS_ADAPTER *ioc,
7906 Mpi2EventIrConfigElement_t *element)
7907 {
7908 u16 handle = le16_to_cpu(element->PhysDiskDevHandle);
7909
7910 _scsih_device_remove_by_handle(ioc, handle);
7911 }
7912
7913 /**
7914 * _scsih_sas_pd_add - remove pd component
7915 * @ioc: per adapter object
7916 * @element: IR config element data
7917 * Context: user.
7918 */
7919 static void
_scsih_sas_pd_add(struct MPT3SAS_ADAPTER * ioc,Mpi2EventIrConfigElement_t * element)7920 _scsih_sas_pd_add(struct MPT3SAS_ADAPTER *ioc,
7921 Mpi2EventIrConfigElement_t *element)
7922 {
7923 struct _sas_device *sas_device;
7924 u16 handle = le16_to_cpu(element->PhysDiskDevHandle);
7925 Mpi2ConfigReply_t mpi_reply;
7926 Mpi2SasDevicePage0_t sas_device_pg0;
7927 u32 ioc_status;
7928 u64 sas_address;
7929 u16 parent_handle;
7930
7931 set_bit(handle, ioc->pd_handles);
7932
7933 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
7934 if (sas_device) {
7935 _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum);
7936 sas_device_put(sas_device);
7937 return;
7938 }
7939
7940 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
7941 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) {
7942 ioc_err(ioc, "failure at %s:%d/%s()!\n",
7943 __FILE__, __LINE__, __func__);
7944 return;
7945 }
7946
7947 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
7948 MPI2_IOCSTATUS_MASK;
7949 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
7950 ioc_err(ioc, "failure at %s:%d/%s()!\n",
7951 __FILE__, __LINE__, __func__);
7952 return;
7953 }
7954
7955 parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle);
7956 if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address))
7957 mpt3sas_transport_update_links(ioc, sas_address, handle,
7958 sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5);
7959
7960 _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum);
7961 _scsih_add_device(ioc, handle, 0, 1);
7962 }
7963
7964 /**
7965 * _scsih_sas_ir_config_change_event_debug - debug for IR Config Change events
7966 * @ioc: per adapter object
7967 * @event_data: event data payload
7968 * Context: user.
7969 */
7970 static void
_scsih_sas_ir_config_change_event_debug(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataIrConfigChangeList_t * event_data)7971 _scsih_sas_ir_config_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
7972 Mpi2EventDataIrConfigChangeList_t *event_data)
7973 {
7974 Mpi2EventIrConfigElement_t *element;
7975 u8 element_type;
7976 int i;
7977 char *reason_str = NULL, *element_str = NULL;
7978
7979 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
7980
7981 ioc_info(ioc, "raid config change: (%s), elements(%d)\n",
7982 le32_to_cpu(event_data->Flags) & MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG ?
7983 "foreign" : "native",
7984 event_data->NumElements);
7985 for (i = 0; i < event_data->NumElements; i++, element++) {
7986 switch (element->ReasonCode) {
7987 case MPI2_EVENT_IR_CHANGE_RC_ADDED:
7988 reason_str = "add";
7989 break;
7990 case MPI2_EVENT_IR_CHANGE_RC_REMOVED:
7991 reason_str = "remove";
7992 break;
7993 case MPI2_EVENT_IR_CHANGE_RC_NO_CHANGE:
7994 reason_str = "no change";
7995 break;
7996 case MPI2_EVENT_IR_CHANGE_RC_HIDE:
7997 reason_str = "hide";
7998 break;
7999 case MPI2_EVENT_IR_CHANGE_RC_UNHIDE:
8000 reason_str = "unhide";
8001 break;
8002 case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED:
8003 reason_str = "volume_created";
8004 break;
8005 case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED:
8006 reason_str = "volume_deleted";
8007 break;
8008 case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED:
8009 reason_str = "pd_created";
8010 break;
8011 case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED:
8012 reason_str = "pd_deleted";
8013 break;
8014 default:
8015 reason_str = "unknown reason";
8016 break;
8017 }
8018 element_type = le16_to_cpu(element->ElementFlags) &
8019 MPI2_EVENT_IR_CHANGE_EFLAGS_ELEMENT_TYPE_MASK;
8020 switch (element_type) {
8021 case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLUME_ELEMENT:
8022 element_str = "volume";
8023 break;
8024 case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLPHYSDISK_ELEMENT:
8025 element_str = "phys disk";
8026 break;
8027 case MPI2_EVENT_IR_CHANGE_EFLAGS_HOTSPARE_ELEMENT:
8028 element_str = "hot spare";
8029 break;
8030 default:
8031 element_str = "unknown element";
8032 break;
8033 }
8034 pr_info("\t(%s:%s), vol handle(0x%04x), " \
8035 "pd handle(0x%04x), pd num(0x%02x)\n", element_str,
8036 reason_str, le16_to_cpu(element->VolDevHandle),
8037 le16_to_cpu(element->PhysDiskDevHandle),
8038 element->PhysDiskNum);
8039 }
8040 }
8041
8042 /**
8043 * _scsih_sas_ir_config_change_event - handle ir configuration change events
8044 * @ioc: per adapter object
8045 * @fw_event: The fw_event_work object
8046 * Context: user.
8047 */
8048 static void
_scsih_sas_ir_config_change_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)8049 _scsih_sas_ir_config_change_event(struct MPT3SAS_ADAPTER *ioc,
8050 struct fw_event_work *fw_event)
8051 {
8052 Mpi2EventIrConfigElement_t *element;
8053 int i;
8054 u8 foreign_config;
8055 Mpi2EventDataIrConfigChangeList_t *event_data =
8056 (Mpi2EventDataIrConfigChangeList_t *)
8057 fw_event->event_data;
8058
8059 if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) &&
8060 (!ioc->hide_ir_msg))
8061 _scsih_sas_ir_config_change_event_debug(ioc, event_data);
8062
8063 foreign_config = (le32_to_cpu(event_data->Flags) &
8064 MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) ? 1 : 0;
8065
8066 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
8067 if (ioc->shost_recovery &&
8068 ioc->hba_mpi_version_belonged != MPI2_VERSION) {
8069 for (i = 0; i < event_data->NumElements; i++, element++) {
8070 if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_HIDE)
8071 _scsih_ir_fastpath(ioc,
8072 le16_to_cpu(element->PhysDiskDevHandle),
8073 element->PhysDiskNum);
8074 }
8075 return;
8076 }
8077
8078 for (i = 0; i < event_data->NumElements; i++, element++) {
8079
8080 switch (element->ReasonCode) {
8081 case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED:
8082 case MPI2_EVENT_IR_CHANGE_RC_ADDED:
8083 if (!foreign_config)
8084 _scsih_sas_volume_add(ioc, element);
8085 break;
8086 case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED:
8087 case MPI2_EVENT_IR_CHANGE_RC_REMOVED:
8088 if (!foreign_config)
8089 _scsih_sas_volume_delete(ioc,
8090 le16_to_cpu(element->VolDevHandle));
8091 break;
8092 case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED:
8093 if (!ioc->is_warpdrive)
8094 _scsih_sas_pd_hide(ioc, element);
8095 break;
8096 case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED:
8097 if (!ioc->is_warpdrive)
8098 _scsih_sas_pd_expose(ioc, element);
8099 break;
8100 case MPI2_EVENT_IR_CHANGE_RC_HIDE:
8101 if (!ioc->is_warpdrive)
8102 _scsih_sas_pd_add(ioc, element);
8103 break;
8104 case MPI2_EVENT_IR_CHANGE_RC_UNHIDE:
8105 if (!ioc->is_warpdrive)
8106 _scsih_sas_pd_delete(ioc, element);
8107 break;
8108 }
8109 }
8110 }
8111
8112 /**
8113 * _scsih_sas_ir_volume_event - IR volume event
8114 * @ioc: per adapter object
8115 * @fw_event: The fw_event_work object
8116 * Context: user.
8117 */
8118 static void
_scsih_sas_ir_volume_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)8119 _scsih_sas_ir_volume_event(struct MPT3SAS_ADAPTER *ioc,
8120 struct fw_event_work *fw_event)
8121 {
8122 u64 wwid;
8123 unsigned long flags;
8124 struct _raid_device *raid_device;
8125 u16 handle;
8126 u32 state;
8127 int rc;
8128 Mpi2EventDataIrVolume_t *event_data =
8129 (Mpi2EventDataIrVolume_t *) fw_event->event_data;
8130
8131 if (ioc->shost_recovery)
8132 return;
8133
8134 if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED)
8135 return;
8136
8137 handle = le16_to_cpu(event_data->VolDevHandle);
8138 state = le32_to_cpu(event_data->NewValue);
8139 if (!ioc->hide_ir_msg)
8140 dewtprintk(ioc,
8141 ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n",
8142 __func__, handle,
8143 le32_to_cpu(event_data->PreviousValue),
8144 state));
8145 switch (state) {
8146 case MPI2_RAID_VOL_STATE_MISSING:
8147 case MPI2_RAID_VOL_STATE_FAILED:
8148 _scsih_sas_volume_delete(ioc, handle);
8149 break;
8150
8151 case MPI2_RAID_VOL_STATE_ONLINE:
8152 case MPI2_RAID_VOL_STATE_DEGRADED:
8153 case MPI2_RAID_VOL_STATE_OPTIMAL:
8154
8155 spin_lock_irqsave(&ioc->raid_device_lock, flags);
8156 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
8157 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
8158
8159 if (raid_device)
8160 break;
8161
8162 mpt3sas_config_get_volume_wwid(ioc, handle, &wwid);
8163 if (!wwid) {
8164 ioc_err(ioc, "failure at %s:%d/%s()!\n",
8165 __FILE__, __LINE__, __func__);
8166 break;
8167 }
8168
8169 raid_device = kzalloc(sizeof(struct _raid_device), GFP_KERNEL);
8170 if (!raid_device) {
8171 ioc_err(ioc, "failure at %s:%d/%s()!\n",
8172 __FILE__, __LINE__, __func__);
8173 break;
8174 }
8175
8176 raid_device->id = ioc->sas_id++;
8177 raid_device->channel = RAID_CHANNEL;
8178 raid_device->handle = handle;
8179 raid_device->wwid = wwid;
8180 _scsih_raid_device_add(ioc, raid_device);
8181 rc = scsi_add_device(ioc->shost, RAID_CHANNEL,
8182 raid_device->id, 0);
8183 if (rc)
8184 _scsih_raid_device_remove(ioc, raid_device);
8185 break;
8186
8187 case MPI2_RAID_VOL_STATE_INITIALIZING:
8188 default:
8189 break;
8190 }
8191 }
8192
8193 /**
8194 * _scsih_sas_ir_physical_disk_event - PD event
8195 * @ioc: per adapter object
8196 * @fw_event: The fw_event_work object
8197 * Context: user.
8198 */
8199 static void
_scsih_sas_ir_physical_disk_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)8200 _scsih_sas_ir_physical_disk_event(struct MPT3SAS_ADAPTER *ioc,
8201 struct fw_event_work *fw_event)
8202 {
8203 u16 handle, parent_handle;
8204 u32 state;
8205 struct _sas_device *sas_device;
8206 Mpi2ConfigReply_t mpi_reply;
8207 Mpi2SasDevicePage0_t sas_device_pg0;
8208 u32 ioc_status;
8209 Mpi2EventDataIrPhysicalDisk_t *event_data =
8210 (Mpi2EventDataIrPhysicalDisk_t *) fw_event->event_data;
8211 u64 sas_address;
8212
8213 if (ioc->shost_recovery)
8214 return;
8215
8216 if (event_data->ReasonCode != MPI2_EVENT_IR_PHYSDISK_RC_STATE_CHANGED)
8217 return;
8218
8219 handle = le16_to_cpu(event_data->PhysDiskDevHandle);
8220 state = le32_to_cpu(event_data->NewValue);
8221
8222 if (!ioc->hide_ir_msg)
8223 dewtprintk(ioc,
8224 ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n",
8225 __func__, handle,
8226 le32_to_cpu(event_data->PreviousValue),
8227 state));
8228
8229 switch (state) {
8230 case MPI2_RAID_PD_STATE_ONLINE:
8231 case MPI2_RAID_PD_STATE_DEGRADED:
8232 case MPI2_RAID_PD_STATE_REBUILDING:
8233 case MPI2_RAID_PD_STATE_OPTIMAL:
8234 case MPI2_RAID_PD_STATE_HOT_SPARE:
8235
8236 if (!ioc->is_warpdrive)
8237 set_bit(handle, ioc->pd_handles);
8238
8239 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
8240 if (sas_device) {
8241 sas_device_put(sas_device);
8242 return;
8243 }
8244
8245 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
8246 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE,
8247 handle))) {
8248 ioc_err(ioc, "failure at %s:%d/%s()!\n",
8249 __FILE__, __LINE__, __func__);
8250 return;
8251 }
8252
8253 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
8254 MPI2_IOCSTATUS_MASK;
8255 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
8256 ioc_err(ioc, "failure at %s:%d/%s()!\n",
8257 __FILE__, __LINE__, __func__);
8258 return;
8259 }
8260
8261 parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle);
8262 if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address))
8263 mpt3sas_transport_update_links(ioc, sas_address, handle,
8264 sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5);
8265
8266 _scsih_add_device(ioc, handle, 0, 1);
8267
8268 break;
8269
8270 case MPI2_RAID_PD_STATE_OFFLINE:
8271 case MPI2_RAID_PD_STATE_NOT_CONFIGURED:
8272 case MPI2_RAID_PD_STATE_NOT_COMPATIBLE:
8273 default:
8274 break;
8275 }
8276 }
8277
8278 /**
8279 * _scsih_sas_ir_operation_status_event_debug - debug for IR op event
8280 * @ioc: per adapter object
8281 * @event_data: event data payload
8282 * Context: user.
8283 */
8284 static void
_scsih_sas_ir_operation_status_event_debug(struct MPT3SAS_ADAPTER * ioc,Mpi2EventDataIrOperationStatus_t * event_data)8285 _scsih_sas_ir_operation_status_event_debug(struct MPT3SAS_ADAPTER *ioc,
8286 Mpi2EventDataIrOperationStatus_t *event_data)
8287 {
8288 char *reason_str = NULL;
8289
8290 switch (event_data->RAIDOperation) {
8291 case MPI2_EVENT_IR_RAIDOP_RESYNC:
8292 reason_str = "resync";
8293 break;
8294 case MPI2_EVENT_IR_RAIDOP_ONLINE_CAP_EXPANSION:
8295 reason_str = "online capacity expansion";
8296 break;
8297 case MPI2_EVENT_IR_RAIDOP_CONSISTENCY_CHECK:
8298 reason_str = "consistency check";
8299 break;
8300 case MPI2_EVENT_IR_RAIDOP_BACKGROUND_INIT:
8301 reason_str = "background init";
8302 break;
8303 case MPI2_EVENT_IR_RAIDOP_MAKE_DATA_CONSISTENT:
8304 reason_str = "make data consistent";
8305 break;
8306 }
8307
8308 if (!reason_str)
8309 return;
8310
8311 ioc_info(ioc, "raid operational status: (%s)\thandle(0x%04x), percent complete(%d)\n",
8312 reason_str,
8313 le16_to_cpu(event_data->VolDevHandle),
8314 event_data->PercentComplete);
8315 }
8316
8317 /**
8318 * _scsih_sas_ir_operation_status_event - handle RAID operation events
8319 * @ioc: per adapter object
8320 * @fw_event: The fw_event_work object
8321 * Context: user.
8322 */
8323 static void
_scsih_sas_ir_operation_status_event(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)8324 _scsih_sas_ir_operation_status_event(struct MPT3SAS_ADAPTER *ioc,
8325 struct fw_event_work *fw_event)
8326 {
8327 Mpi2EventDataIrOperationStatus_t *event_data =
8328 (Mpi2EventDataIrOperationStatus_t *)
8329 fw_event->event_data;
8330 static struct _raid_device *raid_device;
8331 unsigned long flags;
8332 u16 handle;
8333
8334 if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) &&
8335 (!ioc->hide_ir_msg))
8336 _scsih_sas_ir_operation_status_event_debug(ioc,
8337 event_data);
8338
8339 /* code added for raid transport support */
8340 if (event_data->RAIDOperation == MPI2_EVENT_IR_RAIDOP_RESYNC) {
8341
8342 spin_lock_irqsave(&ioc->raid_device_lock, flags);
8343 handle = le16_to_cpu(event_data->VolDevHandle);
8344 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
8345 if (raid_device)
8346 raid_device->percent_complete =
8347 event_data->PercentComplete;
8348 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
8349 }
8350 }
8351
8352 /**
8353 * _scsih_prep_device_scan - initialize parameters prior to device scan
8354 * @ioc: per adapter object
8355 *
8356 * Set the deleted flag prior to device scan. If the device is found during
8357 * the scan, then we clear the deleted flag.
8358 */
8359 static void
_scsih_prep_device_scan(struct MPT3SAS_ADAPTER * ioc)8360 _scsih_prep_device_scan(struct MPT3SAS_ADAPTER *ioc)
8361 {
8362 struct MPT3SAS_DEVICE *sas_device_priv_data;
8363 struct scsi_device *sdev;
8364
8365 shost_for_each_device(sdev, ioc->shost) {
8366 sas_device_priv_data = sdev->hostdata;
8367 if (sas_device_priv_data && sas_device_priv_data->sas_target)
8368 sas_device_priv_data->sas_target->deleted = 1;
8369 }
8370 }
8371
8372 /**
8373 * _scsih_mark_responding_sas_device - mark a sas_devices as responding
8374 * @ioc: per adapter object
8375 * @sas_device_pg0: SAS Device page 0
8376 *
8377 * After host reset, find out whether devices are still responding.
8378 * Used in _scsih_remove_unresponsive_sas_devices.
8379 */
8380 static void
_scsih_mark_responding_sas_device(struct MPT3SAS_ADAPTER * ioc,Mpi2SasDevicePage0_t * sas_device_pg0)8381 _scsih_mark_responding_sas_device(struct MPT3SAS_ADAPTER *ioc,
8382 Mpi2SasDevicePage0_t *sas_device_pg0)
8383 {
8384 struct MPT3SAS_TARGET *sas_target_priv_data = NULL;
8385 struct scsi_target *starget;
8386 struct _sas_device *sas_device = NULL;
8387 struct _enclosure_node *enclosure_dev = NULL;
8388 unsigned long flags;
8389
8390 if (sas_device_pg0->EnclosureHandle) {
8391 enclosure_dev =
8392 mpt3sas_scsih_enclosure_find_by_handle(ioc,
8393 le16_to_cpu(sas_device_pg0->EnclosureHandle));
8394 if (enclosure_dev == NULL)
8395 ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n",
8396 sas_device_pg0->EnclosureHandle);
8397 }
8398 spin_lock_irqsave(&ioc->sas_device_lock, flags);
8399 list_for_each_entry(sas_device, &ioc->sas_device_list, list) {
8400 if ((sas_device->sas_address == le64_to_cpu(
8401 sas_device_pg0->SASAddress)) && (sas_device->slot ==
8402 le16_to_cpu(sas_device_pg0->Slot))) {
8403 sas_device->responding = 1;
8404 starget = sas_device->starget;
8405 if (starget && starget->hostdata) {
8406 sas_target_priv_data = starget->hostdata;
8407 sas_target_priv_data->tm_busy = 0;
8408 sas_target_priv_data->deleted = 0;
8409 } else
8410 sas_target_priv_data = NULL;
8411 if (starget) {
8412 starget_printk(KERN_INFO, starget,
8413 "handle(0x%04x), sas_addr(0x%016llx)\n",
8414 le16_to_cpu(sas_device_pg0->DevHandle),
8415 (unsigned long long)
8416 sas_device->sas_address);
8417
8418 if (sas_device->enclosure_handle != 0)
8419 starget_printk(KERN_INFO, starget,
8420 "enclosure logical id(0x%016llx),"
8421 " slot(%d)\n",
8422 (unsigned long long)
8423 sas_device->enclosure_logical_id,
8424 sas_device->slot);
8425 }
8426 if (le16_to_cpu(sas_device_pg0->Flags) &
8427 MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) {
8428 sas_device->enclosure_level =
8429 sas_device_pg0->EnclosureLevel;
8430 memcpy(&sas_device->connector_name[0],
8431 &sas_device_pg0->ConnectorName[0], 4);
8432 } else {
8433 sas_device->enclosure_level = 0;
8434 sas_device->connector_name[0] = '\0';
8435 }
8436
8437 sas_device->enclosure_handle =
8438 le16_to_cpu(sas_device_pg0->EnclosureHandle);
8439 sas_device->is_chassis_slot_valid = 0;
8440 if (enclosure_dev) {
8441 sas_device->enclosure_logical_id = le64_to_cpu(
8442 enclosure_dev->pg0.EnclosureLogicalID);
8443 if (le16_to_cpu(enclosure_dev->pg0.Flags) &
8444 MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) {
8445 sas_device->is_chassis_slot_valid = 1;
8446 sas_device->chassis_slot =
8447 enclosure_dev->pg0.ChassisSlot;
8448 }
8449 }
8450
8451 if (sas_device->handle == le16_to_cpu(
8452 sas_device_pg0->DevHandle))
8453 goto out;
8454 pr_info("\thandle changed from(0x%04x)!!!\n",
8455 sas_device->handle);
8456 sas_device->handle = le16_to_cpu(
8457 sas_device_pg0->DevHandle);
8458 if (sas_target_priv_data)
8459 sas_target_priv_data->handle =
8460 le16_to_cpu(sas_device_pg0->DevHandle);
8461 goto out;
8462 }
8463 }
8464 out:
8465 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
8466 }
8467
8468 /**
8469 * _scsih_create_enclosure_list_after_reset - Free Existing list,
8470 * And create enclosure list by scanning all Enclosure Page(0)s
8471 * @ioc: per adapter object
8472 */
8473 static void
_scsih_create_enclosure_list_after_reset(struct MPT3SAS_ADAPTER * ioc)8474 _scsih_create_enclosure_list_after_reset(struct MPT3SAS_ADAPTER *ioc)
8475 {
8476 struct _enclosure_node *enclosure_dev;
8477 Mpi2ConfigReply_t mpi_reply;
8478 u16 enclosure_handle;
8479 int rc;
8480
8481 /* Free existing enclosure list */
8482 mpt3sas_free_enclosure_list(ioc);
8483
8484 /* Re constructing enclosure list after reset*/
8485 enclosure_handle = 0xFFFF;
8486 do {
8487 enclosure_dev =
8488 kzalloc(sizeof(struct _enclosure_node), GFP_KERNEL);
8489 if (!enclosure_dev) {
8490 ioc_err(ioc, "failure at %s:%d/%s()!\n",
8491 __FILE__, __LINE__, __func__);
8492 return;
8493 }
8494 rc = mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply,
8495 &enclosure_dev->pg0,
8496 MPI2_SAS_ENCLOS_PGAD_FORM_GET_NEXT_HANDLE,
8497 enclosure_handle);
8498
8499 if (rc || (le16_to_cpu(mpi_reply.IOCStatus) &
8500 MPI2_IOCSTATUS_MASK)) {
8501 kfree(enclosure_dev);
8502 return;
8503 }
8504 list_add_tail(&enclosure_dev->list,
8505 &ioc->enclosure_list);
8506 enclosure_handle =
8507 le16_to_cpu(enclosure_dev->pg0.EnclosureHandle);
8508 } while (1);
8509 }
8510
8511 /**
8512 * _scsih_search_responding_sas_devices -
8513 * @ioc: per adapter object
8514 *
8515 * After host reset, find out whether devices are still responding.
8516 * If not remove.
8517 */
8518 static void
_scsih_search_responding_sas_devices(struct MPT3SAS_ADAPTER * ioc)8519 _scsih_search_responding_sas_devices(struct MPT3SAS_ADAPTER *ioc)
8520 {
8521 Mpi2SasDevicePage0_t sas_device_pg0;
8522 Mpi2ConfigReply_t mpi_reply;
8523 u16 ioc_status;
8524 u16 handle;
8525 u32 device_info;
8526
8527 ioc_info(ioc, "search for end-devices: start\n");
8528
8529 if (list_empty(&ioc->sas_device_list))
8530 goto out;
8531
8532 handle = 0xFFFF;
8533 while (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
8534 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE,
8535 handle))) {
8536 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
8537 MPI2_IOCSTATUS_MASK;
8538 if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
8539 break;
8540 handle = le16_to_cpu(sas_device_pg0.DevHandle);
8541 device_info = le32_to_cpu(sas_device_pg0.DeviceInfo);
8542 if (!(_scsih_is_end_device(device_info)))
8543 continue;
8544 _scsih_mark_responding_sas_device(ioc, &sas_device_pg0);
8545 }
8546
8547 out:
8548 ioc_info(ioc, "search for end-devices: complete\n");
8549 }
8550
8551 /**
8552 * _scsih_mark_responding_pcie_device - mark a pcie_device as responding
8553 * @ioc: per adapter object
8554 * @pcie_device_pg0: PCIe Device page 0
8555 *
8556 * After host reset, find out whether devices are still responding.
8557 * Used in _scsih_remove_unresponding_devices.
8558 */
8559 static void
_scsih_mark_responding_pcie_device(struct MPT3SAS_ADAPTER * ioc,Mpi26PCIeDevicePage0_t * pcie_device_pg0)8560 _scsih_mark_responding_pcie_device(struct MPT3SAS_ADAPTER *ioc,
8561 Mpi26PCIeDevicePage0_t *pcie_device_pg0)
8562 {
8563 struct MPT3SAS_TARGET *sas_target_priv_data = NULL;
8564 struct scsi_target *starget;
8565 struct _pcie_device *pcie_device;
8566 unsigned long flags;
8567
8568 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
8569 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) {
8570 if ((pcie_device->wwid == le64_to_cpu(pcie_device_pg0->WWID))
8571 && (pcie_device->slot == le16_to_cpu(
8572 pcie_device_pg0->Slot))) {
8573 pcie_device->access_status =
8574 pcie_device_pg0->AccessStatus;
8575 pcie_device->responding = 1;
8576 starget = pcie_device->starget;
8577 if (starget && starget->hostdata) {
8578 sas_target_priv_data = starget->hostdata;
8579 sas_target_priv_data->tm_busy = 0;
8580 sas_target_priv_data->deleted = 0;
8581 } else
8582 sas_target_priv_data = NULL;
8583 if (starget) {
8584 starget_printk(KERN_INFO, starget,
8585 "handle(0x%04x), wwid(0x%016llx) ",
8586 pcie_device->handle,
8587 (unsigned long long)pcie_device->wwid);
8588 if (pcie_device->enclosure_handle != 0)
8589 starget_printk(KERN_INFO, starget,
8590 "enclosure logical id(0x%016llx), "
8591 "slot(%d)\n",
8592 (unsigned long long)
8593 pcie_device->enclosure_logical_id,
8594 pcie_device->slot);
8595 }
8596
8597 if (((le32_to_cpu(pcie_device_pg0->Flags)) &
8598 MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) &&
8599 (ioc->hba_mpi_version_belonged != MPI2_VERSION)) {
8600 pcie_device->enclosure_level =
8601 pcie_device_pg0->EnclosureLevel;
8602 memcpy(&pcie_device->connector_name[0],
8603 &pcie_device_pg0->ConnectorName[0], 4);
8604 } else {
8605 pcie_device->enclosure_level = 0;
8606 pcie_device->connector_name[0] = '\0';
8607 }
8608
8609 if (pcie_device->handle == le16_to_cpu(
8610 pcie_device_pg0->DevHandle))
8611 goto out;
8612 pr_info("\thandle changed from(0x%04x)!!!\n",
8613 pcie_device->handle);
8614 pcie_device->handle = le16_to_cpu(
8615 pcie_device_pg0->DevHandle);
8616 if (sas_target_priv_data)
8617 sas_target_priv_data->handle =
8618 le16_to_cpu(pcie_device_pg0->DevHandle);
8619 goto out;
8620 }
8621 }
8622
8623 out:
8624 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
8625 }
8626
8627 /**
8628 * _scsih_search_responding_pcie_devices -
8629 * @ioc: per adapter object
8630 *
8631 * After host reset, find out whether devices are still responding.
8632 * If not remove.
8633 */
8634 static void
_scsih_search_responding_pcie_devices(struct MPT3SAS_ADAPTER * ioc)8635 _scsih_search_responding_pcie_devices(struct MPT3SAS_ADAPTER *ioc)
8636 {
8637 Mpi26PCIeDevicePage0_t pcie_device_pg0;
8638 Mpi2ConfigReply_t mpi_reply;
8639 u16 ioc_status;
8640 u16 handle;
8641 u32 device_info;
8642
8643 ioc_info(ioc, "search for end-devices: start\n");
8644
8645 if (list_empty(&ioc->pcie_device_list))
8646 goto out;
8647
8648 handle = 0xFFFF;
8649 while (!(mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply,
8650 &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE,
8651 handle))) {
8652 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
8653 MPI2_IOCSTATUS_MASK;
8654 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
8655 ioc_info(ioc, "\tbreak from %s: ioc_status(0x%04x), loginfo(0x%08x)\n",
8656 __func__, ioc_status,
8657 le32_to_cpu(mpi_reply.IOCLogInfo));
8658 break;
8659 }
8660 handle = le16_to_cpu(pcie_device_pg0.DevHandle);
8661 device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo);
8662 if (!(_scsih_is_nvme_pciescsi_device(device_info)))
8663 continue;
8664 _scsih_mark_responding_pcie_device(ioc, &pcie_device_pg0);
8665 }
8666 out:
8667 ioc_info(ioc, "search for PCIe end-devices: complete\n");
8668 }
8669
8670 /**
8671 * _scsih_mark_responding_raid_device - mark a raid_device as responding
8672 * @ioc: per adapter object
8673 * @wwid: world wide identifier for raid volume
8674 * @handle: device handle
8675 *
8676 * After host reset, find out whether devices are still responding.
8677 * Used in _scsih_remove_unresponsive_raid_devices.
8678 */
8679 static void
_scsih_mark_responding_raid_device(struct MPT3SAS_ADAPTER * ioc,u64 wwid,u16 handle)8680 _scsih_mark_responding_raid_device(struct MPT3SAS_ADAPTER *ioc, u64 wwid,
8681 u16 handle)
8682 {
8683 struct MPT3SAS_TARGET *sas_target_priv_data = NULL;
8684 struct scsi_target *starget;
8685 struct _raid_device *raid_device;
8686 unsigned long flags;
8687
8688 spin_lock_irqsave(&ioc->raid_device_lock, flags);
8689 list_for_each_entry(raid_device, &ioc->raid_device_list, list) {
8690 if (raid_device->wwid == wwid && raid_device->starget) {
8691 starget = raid_device->starget;
8692 if (starget && starget->hostdata) {
8693 sas_target_priv_data = starget->hostdata;
8694 sas_target_priv_data->deleted = 0;
8695 } else
8696 sas_target_priv_data = NULL;
8697 raid_device->responding = 1;
8698 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
8699 starget_printk(KERN_INFO, raid_device->starget,
8700 "handle(0x%04x), wwid(0x%016llx)\n", handle,
8701 (unsigned long long)raid_device->wwid);
8702
8703 /*
8704 * WARPDRIVE: The handles of the PDs might have changed
8705 * across the host reset so re-initialize the
8706 * required data for Direct IO
8707 */
8708 mpt3sas_init_warpdrive_properties(ioc, raid_device);
8709 spin_lock_irqsave(&ioc->raid_device_lock, flags);
8710 if (raid_device->handle == handle) {
8711 spin_unlock_irqrestore(&ioc->raid_device_lock,
8712 flags);
8713 return;
8714 }
8715 pr_info("\thandle changed from(0x%04x)!!!\n",
8716 raid_device->handle);
8717 raid_device->handle = handle;
8718 if (sas_target_priv_data)
8719 sas_target_priv_data->handle = handle;
8720 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
8721 return;
8722 }
8723 }
8724 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
8725 }
8726
8727 /**
8728 * _scsih_search_responding_raid_devices -
8729 * @ioc: per adapter object
8730 *
8731 * After host reset, find out whether devices are still responding.
8732 * If not remove.
8733 */
8734 static void
_scsih_search_responding_raid_devices(struct MPT3SAS_ADAPTER * ioc)8735 _scsih_search_responding_raid_devices(struct MPT3SAS_ADAPTER *ioc)
8736 {
8737 Mpi2RaidVolPage1_t volume_pg1;
8738 Mpi2RaidVolPage0_t volume_pg0;
8739 Mpi2RaidPhysDiskPage0_t pd_pg0;
8740 Mpi2ConfigReply_t mpi_reply;
8741 u16 ioc_status;
8742 u16 handle;
8743 u8 phys_disk_num;
8744
8745 if (!ioc->ir_firmware)
8746 return;
8747
8748 ioc_info(ioc, "search for raid volumes: start\n");
8749
8750 if (list_empty(&ioc->raid_device_list))
8751 goto out;
8752
8753 handle = 0xFFFF;
8754 while (!(mpt3sas_config_get_raid_volume_pg1(ioc, &mpi_reply,
8755 &volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) {
8756 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
8757 MPI2_IOCSTATUS_MASK;
8758 if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
8759 break;
8760 handle = le16_to_cpu(volume_pg1.DevHandle);
8761
8762 if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply,
8763 &volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle,
8764 sizeof(Mpi2RaidVolPage0_t)))
8765 continue;
8766
8767 if (volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL ||
8768 volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_ONLINE ||
8769 volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_DEGRADED)
8770 _scsih_mark_responding_raid_device(ioc,
8771 le64_to_cpu(volume_pg1.WWID), handle);
8772 }
8773
8774 /* refresh the pd_handles */
8775 if (!ioc->is_warpdrive) {
8776 phys_disk_num = 0xFF;
8777 memset(ioc->pd_handles, 0, ioc->pd_handles_sz);
8778 while (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply,
8779 &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM,
8780 phys_disk_num))) {
8781 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
8782 MPI2_IOCSTATUS_MASK;
8783 if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
8784 break;
8785 phys_disk_num = pd_pg0.PhysDiskNum;
8786 handle = le16_to_cpu(pd_pg0.DevHandle);
8787 set_bit(handle, ioc->pd_handles);
8788 }
8789 }
8790 out:
8791 ioc_info(ioc, "search for responding raid volumes: complete\n");
8792 }
8793
8794 /**
8795 * _scsih_mark_responding_expander - mark a expander as responding
8796 * @ioc: per adapter object
8797 * @expander_pg0:SAS Expander Config Page0
8798 *
8799 * After host reset, find out whether devices are still responding.
8800 * Used in _scsih_remove_unresponsive_expanders.
8801 */
8802 static void
_scsih_mark_responding_expander(struct MPT3SAS_ADAPTER * ioc,Mpi2ExpanderPage0_t * expander_pg0)8803 _scsih_mark_responding_expander(struct MPT3SAS_ADAPTER *ioc,
8804 Mpi2ExpanderPage0_t *expander_pg0)
8805 {
8806 struct _sas_node *sas_expander = NULL;
8807 unsigned long flags;
8808 int i;
8809 struct _enclosure_node *enclosure_dev = NULL;
8810 u16 handle = le16_to_cpu(expander_pg0->DevHandle);
8811 u16 enclosure_handle = le16_to_cpu(expander_pg0->EnclosureHandle);
8812 u64 sas_address = le64_to_cpu(expander_pg0->SASAddress);
8813
8814 if (enclosure_handle)
8815 enclosure_dev =
8816 mpt3sas_scsih_enclosure_find_by_handle(ioc,
8817 enclosure_handle);
8818
8819 spin_lock_irqsave(&ioc->sas_node_lock, flags);
8820 list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) {
8821 if (sas_expander->sas_address != sas_address)
8822 continue;
8823 sas_expander->responding = 1;
8824
8825 if (enclosure_dev) {
8826 sas_expander->enclosure_logical_id =
8827 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
8828 sas_expander->enclosure_handle =
8829 le16_to_cpu(expander_pg0->EnclosureHandle);
8830 }
8831
8832 if (sas_expander->handle == handle)
8833 goto out;
8834 pr_info("\texpander(0x%016llx): handle changed" \
8835 " from(0x%04x) to (0x%04x)!!!\n",
8836 (unsigned long long)sas_expander->sas_address,
8837 sas_expander->handle, handle);
8838 sas_expander->handle = handle;
8839 for (i = 0 ; i < sas_expander->num_phys ; i++)
8840 sas_expander->phy[i].handle = handle;
8841 goto out;
8842 }
8843 out:
8844 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
8845 }
8846
8847 /**
8848 * _scsih_search_responding_expanders -
8849 * @ioc: per adapter object
8850 *
8851 * After host reset, find out whether devices are still responding.
8852 * If not remove.
8853 */
8854 static void
_scsih_search_responding_expanders(struct MPT3SAS_ADAPTER * ioc)8855 _scsih_search_responding_expanders(struct MPT3SAS_ADAPTER *ioc)
8856 {
8857 Mpi2ExpanderPage0_t expander_pg0;
8858 Mpi2ConfigReply_t mpi_reply;
8859 u16 ioc_status;
8860 u64 sas_address;
8861 u16 handle;
8862
8863 ioc_info(ioc, "search for expanders: start\n");
8864
8865 if (list_empty(&ioc->sas_expander_list))
8866 goto out;
8867
8868 handle = 0xFFFF;
8869 while (!(mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0,
8870 MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) {
8871
8872 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
8873 MPI2_IOCSTATUS_MASK;
8874 if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
8875 break;
8876
8877 handle = le16_to_cpu(expander_pg0.DevHandle);
8878 sas_address = le64_to_cpu(expander_pg0.SASAddress);
8879 pr_info("\texpander present: handle(0x%04x), sas_addr(0x%016llx)\n",
8880 handle,
8881 (unsigned long long)sas_address);
8882 _scsih_mark_responding_expander(ioc, &expander_pg0);
8883 }
8884
8885 out:
8886 ioc_info(ioc, "search for expanders: complete\n");
8887 }
8888
8889 /**
8890 * _scsih_remove_unresponding_devices - removing unresponding devices
8891 * @ioc: per adapter object
8892 */
8893 static void
_scsih_remove_unresponding_devices(struct MPT3SAS_ADAPTER * ioc)8894 _scsih_remove_unresponding_devices(struct MPT3SAS_ADAPTER *ioc)
8895 {
8896 struct _sas_device *sas_device, *sas_device_next;
8897 struct _sas_node *sas_expander, *sas_expander_next;
8898 struct _raid_device *raid_device, *raid_device_next;
8899 struct _pcie_device *pcie_device, *pcie_device_next;
8900 struct list_head tmp_list;
8901 unsigned long flags;
8902 LIST_HEAD(head);
8903
8904 ioc_info(ioc, "removing unresponding devices: start\n");
8905
8906 /* removing unresponding end devices */
8907 ioc_info(ioc, "removing unresponding devices: end-devices\n");
8908 /*
8909 * Iterate, pulling off devices marked as non-responding. We become the
8910 * owner for the reference the list had on any object we prune.
8911 */
8912 spin_lock_irqsave(&ioc->sas_device_lock, flags);
8913 list_for_each_entry_safe(sas_device, sas_device_next,
8914 &ioc->sas_device_list, list) {
8915 if (!sas_device->responding)
8916 list_move_tail(&sas_device->list, &head);
8917 else
8918 sas_device->responding = 0;
8919 }
8920 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
8921
8922 /*
8923 * Now, uninitialize and remove the unresponding devices we pruned.
8924 */
8925 list_for_each_entry_safe(sas_device, sas_device_next, &head, list) {
8926 _scsih_remove_device(ioc, sas_device);
8927 list_del_init(&sas_device->list);
8928 sas_device_put(sas_device);
8929 }
8930
8931 ioc_info(ioc, "Removing unresponding devices: pcie end-devices\n");
8932 INIT_LIST_HEAD(&head);
8933 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
8934 list_for_each_entry_safe(pcie_device, pcie_device_next,
8935 &ioc->pcie_device_list, list) {
8936 if (!pcie_device->responding)
8937 list_move_tail(&pcie_device->list, &head);
8938 else
8939 pcie_device->responding = 0;
8940 }
8941 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
8942
8943 list_for_each_entry_safe(pcie_device, pcie_device_next, &head, list) {
8944 _scsih_pcie_device_remove_from_sml(ioc, pcie_device);
8945 list_del_init(&pcie_device->list);
8946 pcie_device_put(pcie_device);
8947 }
8948
8949 /* removing unresponding volumes */
8950 if (ioc->ir_firmware) {
8951 ioc_info(ioc, "removing unresponding devices: volumes\n");
8952 list_for_each_entry_safe(raid_device, raid_device_next,
8953 &ioc->raid_device_list, list) {
8954 if (!raid_device->responding)
8955 _scsih_sas_volume_delete(ioc,
8956 raid_device->handle);
8957 else
8958 raid_device->responding = 0;
8959 }
8960 }
8961
8962 /* removing unresponding expanders */
8963 ioc_info(ioc, "removing unresponding devices: expanders\n");
8964 spin_lock_irqsave(&ioc->sas_node_lock, flags);
8965 INIT_LIST_HEAD(&tmp_list);
8966 list_for_each_entry_safe(sas_expander, sas_expander_next,
8967 &ioc->sas_expander_list, list) {
8968 if (!sas_expander->responding)
8969 list_move_tail(&sas_expander->list, &tmp_list);
8970 else
8971 sas_expander->responding = 0;
8972 }
8973 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
8974 list_for_each_entry_safe(sas_expander, sas_expander_next, &tmp_list,
8975 list) {
8976 _scsih_expander_node_remove(ioc, sas_expander);
8977 }
8978
8979 ioc_info(ioc, "removing unresponding devices: complete\n");
8980
8981 /* unblock devices */
8982 _scsih_ublock_io_all_device(ioc);
8983 }
8984
8985 static void
_scsih_refresh_expander_links(struct MPT3SAS_ADAPTER * ioc,struct _sas_node * sas_expander,u16 handle)8986 _scsih_refresh_expander_links(struct MPT3SAS_ADAPTER *ioc,
8987 struct _sas_node *sas_expander, u16 handle)
8988 {
8989 Mpi2ExpanderPage1_t expander_pg1;
8990 Mpi2ConfigReply_t mpi_reply;
8991 int i;
8992
8993 for (i = 0 ; i < sas_expander->num_phys ; i++) {
8994 if ((mpt3sas_config_get_expander_pg1(ioc, &mpi_reply,
8995 &expander_pg1, i, handle))) {
8996 ioc_err(ioc, "failure at %s:%d/%s()!\n",
8997 __FILE__, __LINE__, __func__);
8998 return;
8999 }
9000
9001 mpt3sas_transport_update_links(ioc, sas_expander->sas_address,
9002 le16_to_cpu(expander_pg1.AttachedDevHandle), i,
9003 expander_pg1.NegotiatedLinkRate >> 4);
9004 }
9005 }
9006
9007 /**
9008 * _scsih_scan_for_devices_after_reset - scan for devices after host reset
9009 * @ioc: per adapter object
9010 */
9011 static void
_scsih_scan_for_devices_after_reset(struct MPT3SAS_ADAPTER * ioc)9012 _scsih_scan_for_devices_after_reset(struct MPT3SAS_ADAPTER *ioc)
9013 {
9014 Mpi2ExpanderPage0_t expander_pg0;
9015 Mpi2SasDevicePage0_t sas_device_pg0;
9016 Mpi26PCIeDevicePage0_t pcie_device_pg0;
9017 Mpi2RaidVolPage1_t volume_pg1;
9018 Mpi2RaidVolPage0_t volume_pg0;
9019 Mpi2RaidPhysDiskPage0_t pd_pg0;
9020 Mpi2EventIrConfigElement_t element;
9021 Mpi2ConfigReply_t mpi_reply;
9022 u8 phys_disk_num;
9023 u16 ioc_status;
9024 u16 handle, parent_handle;
9025 u64 sas_address;
9026 struct _sas_device *sas_device;
9027 struct _pcie_device *pcie_device;
9028 struct _sas_node *expander_device;
9029 static struct _raid_device *raid_device;
9030 u8 retry_count;
9031 unsigned long flags;
9032
9033 ioc_info(ioc, "scan devices: start\n");
9034
9035 _scsih_sas_host_refresh(ioc);
9036
9037 ioc_info(ioc, "\tscan devices: expanders start\n");
9038
9039 /* expanders */
9040 handle = 0xFFFF;
9041 while (!(mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0,
9042 MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) {
9043 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9044 MPI2_IOCSTATUS_MASK;
9045 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9046 ioc_info(ioc, "\tbreak from expander scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
9047 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
9048 break;
9049 }
9050 handle = le16_to_cpu(expander_pg0.DevHandle);
9051 spin_lock_irqsave(&ioc->sas_node_lock, flags);
9052 expander_device = mpt3sas_scsih_expander_find_by_sas_address(
9053 ioc, le64_to_cpu(expander_pg0.SASAddress));
9054 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
9055 if (expander_device)
9056 _scsih_refresh_expander_links(ioc, expander_device,
9057 handle);
9058 else {
9059 ioc_info(ioc, "\tBEFORE adding expander: handle (0x%04x), sas_addr(0x%016llx)\n",
9060 handle,
9061 (u64)le64_to_cpu(expander_pg0.SASAddress));
9062 _scsih_expander_add(ioc, handle);
9063 ioc_info(ioc, "\tAFTER adding expander: handle (0x%04x), sas_addr(0x%016llx)\n",
9064 handle,
9065 (u64)le64_to_cpu(expander_pg0.SASAddress));
9066 }
9067 }
9068
9069 ioc_info(ioc, "\tscan devices: expanders complete\n");
9070
9071 if (!ioc->ir_firmware)
9072 goto skip_to_sas;
9073
9074 ioc_info(ioc, "\tscan devices: phys disk start\n");
9075
9076 /* phys disk */
9077 phys_disk_num = 0xFF;
9078 while (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply,
9079 &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM,
9080 phys_disk_num))) {
9081 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9082 MPI2_IOCSTATUS_MASK;
9083 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9084 ioc_info(ioc, "\tbreak from phys disk scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
9085 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
9086 break;
9087 }
9088 phys_disk_num = pd_pg0.PhysDiskNum;
9089 handle = le16_to_cpu(pd_pg0.DevHandle);
9090 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
9091 if (sas_device) {
9092 sas_device_put(sas_device);
9093 continue;
9094 }
9095 if (mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
9096 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE,
9097 handle) != 0)
9098 continue;
9099 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9100 MPI2_IOCSTATUS_MASK;
9101 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9102 ioc_info(ioc, "\tbreak from phys disk scan ioc_status(0x%04x), loginfo(0x%08x)\n",
9103 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
9104 break;
9105 }
9106 parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle);
9107 if (!_scsih_get_sas_address(ioc, parent_handle,
9108 &sas_address)) {
9109 ioc_info(ioc, "\tBEFORE adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n",
9110 handle,
9111 (u64)le64_to_cpu(sas_device_pg0.SASAddress));
9112 mpt3sas_transport_update_links(ioc, sas_address,
9113 handle, sas_device_pg0.PhyNum,
9114 MPI2_SAS_NEG_LINK_RATE_1_5);
9115 set_bit(handle, ioc->pd_handles);
9116 retry_count = 0;
9117 /* This will retry adding the end device.
9118 * _scsih_add_device() will decide on retries and
9119 * return "1" when it should be retried
9120 */
9121 while (_scsih_add_device(ioc, handle, retry_count++,
9122 1)) {
9123 ssleep(1);
9124 }
9125 ioc_info(ioc, "\tAFTER adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n",
9126 handle,
9127 (u64)le64_to_cpu(sas_device_pg0.SASAddress));
9128 }
9129 }
9130
9131 ioc_info(ioc, "\tscan devices: phys disk complete\n");
9132
9133 ioc_info(ioc, "\tscan devices: volumes start\n");
9134
9135 /* volumes */
9136 handle = 0xFFFF;
9137 while (!(mpt3sas_config_get_raid_volume_pg1(ioc, &mpi_reply,
9138 &volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) {
9139 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9140 MPI2_IOCSTATUS_MASK;
9141 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9142 ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
9143 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
9144 break;
9145 }
9146 handle = le16_to_cpu(volume_pg1.DevHandle);
9147 spin_lock_irqsave(&ioc->raid_device_lock, flags);
9148 raid_device = _scsih_raid_device_find_by_wwid(ioc,
9149 le64_to_cpu(volume_pg1.WWID));
9150 spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
9151 if (raid_device)
9152 continue;
9153 if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply,
9154 &volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle,
9155 sizeof(Mpi2RaidVolPage0_t)))
9156 continue;
9157 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9158 MPI2_IOCSTATUS_MASK;
9159 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9160 ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
9161 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
9162 break;
9163 }
9164 if (volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL ||
9165 volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_ONLINE ||
9166 volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_DEGRADED) {
9167 memset(&element, 0, sizeof(Mpi2EventIrConfigElement_t));
9168 element.ReasonCode = MPI2_EVENT_IR_CHANGE_RC_ADDED;
9169 element.VolDevHandle = volume_pg1.DevHandle;
9170 ioc_info(ioc, "\tBEFORE adding volume: handle (0x%04x)\n",
9171 volume_pg1.DevHandle);
9172 _scsih_sas_volume_add(ioc, &element);
9173 ioc_info(ioc, "\tAFTER adding volume: handle (0x%04x)\n",
9174 volume_pg1.DevHandle);
9175 }
9176 }
9177
9178 ioc_info(ioc, "\tscan devices: volumes complete\n");
9179
9180 skip_to_sas:
9181
9182 ioc_info(ioc, "\tscan devices: end devices start\n");
9183
9184 /* sas devices */
9185 handle = 0xFFFF;
9186 while (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
9187 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE,
9188 handle))) {
9189 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9190 MPI2_IOCSTATUS_MASK;
9191 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9192 ioc_info(ioc, "\tbreak from end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
9193 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
9194 break;
9195 }
9196 handle = le16_to_cpu(sas_device_pg0.DevHandle);
9197 if (!(_scsih_is_end_device(
9198 le32_to_cpu(sas_device_pg0.DeviceInfo))))
9199 continue;
9200 sas_device = mpt3sas_get_sdev_by_addr(ioc,
9201 le64_to_cpu(sas_device_pg0.SASAddress));
9202 if (sas_device) {
9203 sas_device_put(sas_device);
9204 continue;
9205 }
9206 parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle);
9207 if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) {
9208 ioc_info(ioc, "\tBEFORE adding end device: handle (0x%04x), sas_addr(0x%016llx)\n",
9209 handle,
9210 (u64)le64_to_cpu(sas_device_pg0.SASAddress));
9211 mpt3sas_transport_update_links(ioc, sas_address, handle,
9212 sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5);
9213 retry_count = 0;
9214 /* This will retry adding the end device.
9215 * _scsih_add_device() will decide on retries and
9216 * return "1" when it should be retried
9217 */
9218 while (_scsih_add_device(ioc, handle, retry_count++,
9219 0)) {
9220 ssleep(1);
9221 }
9222 ioc_info(ioc, "\tAFTER adding end device: handle (0x%04x), sas_addr(0x%016llx)\n",
9223 handle,
9224 (u64)le64_to_cpu(sas_device_pg0.SASAddress));
9225 }
9226 }
9227 ioc_info(ioc, "\tscan devices: end devices complete\n");
9228 ioc_info(ioc, "\tscan devices: pcie end devices start\n");
9229
9230 /* pcie devices */
9231 handle = 0xFFFF;
9232 while (!(mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply,
9233 &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE,
9234 handle))) {
9235 ioc_status = le16_to_cpu(mpi_reply.IOCStatus)
9236 & MPI2_IOCSTATUS_MASK;
9237 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9238 ioc_info(ioc, "\tbreak from pcie end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
9239 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
9240 break;
9241 }
9242 handle = le16_to_cpu(pcie_device_pg0.DevHandle);
9243 if (!(_scsih_is_nvme_pciescsi_device(
9244 le32_to_cpu(pcie_device_pg0.DeviceInfo))))
9245 continue;
9246 pcie_device = mpt3sas_get_pdev_by_wwid(ioc,
9247 le64_to_cpu(pcie_device_pg0.WWID));
9248 if (pcie_device) {
9249 pcie_device_put(pcie_device);
9250 continue;
9251 }
9252 retry_count = 0;
9253 parent_handle = le16_to_cpu(pcie_device_pg0.ParentDevHandle);
9254 _scsih_pcie_add_device(ioc, handle);
9255
9256 ioc_info(ioc, "\tAFTER adding pcie end device: handle (0x%04x), wwid(0x%016llx)\n",
9257 handle, (u64)le64_to_cpu(pcie_device_pg0.WWID));
9258 }
9259 ioc_info(ioc, "\tpcie devices: pcie end devices complete\n");
9260 ioc_info(ioc, "scan devices: complete\n");
9261 }
9262
9263 /**
9264 * mpt3sas_scsih_reset_handler - reset callback handler (for scsih)
9265 * @ioc: per adapter object
9266 *
9267 * The handler for doing any required cleanup or initialization.
9268 */
mpt3sas_scsih_pre_reset_handler(struct MPT3SAS_ADAPTER * ioc)9269 void mpt3sas_scsih_pre_reset_handler(struct MPT3SAS_ADAPTER *ioc)
9270 {
9271 dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_PRE_RESET\n", __func__));
9272 }
9273
9274 /**
9275 * mpt3sas_scsih_after_reset_handler - reset callback handler (for scsih)
9276 * @ioc: per adapter object
9277 *
9278 * The handler for doing any required cleanup or initialization.
9279 */
9280 void
mpt3sas_scsih_after_reset_handler(struct MPT3SAS_ADAPTER * ioc)9281 mpt3sas_scsih_after_reset_handler(struct MPT3SAS_ADAPTER *ioc)
9282 {
9283 dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_AFTER_RESET\n", __func__));
9284 if (ioc->scsih_cmds.status & MPT3_CMD_PENDING) {
9285 ioc->scsih_cmds.status |= MPT3_CMD_RESET;
9286 mpt3sas_base_free_smid(ioc, ioc->scsih_cmds.smid);
9287 complete(&ioc->scsih_cmds.done);
9288 }
9289 if (ioc->tm_cmds.status & MPT3_CMD_PENDING) {
9290 ioc->tm_cmds.status |= MPT3_CMD_RESET;
9291 mpt3sas_base_free_smid(ioc, ioc->tm_cmds.smid);
9292 complete(&ioc->tm_cmds.done);
9293 }
9294
9295 memset(ioc->pend_os_device_add, 0, ioc->pend_os_device_add_sz);
9296 memset(ioc->device_remove_in_progress, 0,
9297 ioc->device_remove_in_progress_sz);
9298 _scsih_fw_event_cleanup_queue(ioc);
9299 _scsih_flush_running_cmds(ioc);
9300 }
9301
9302 /**
9303 * mpt3sas_scsih_reset_handler - reset callback handler (for scsih)
9304 * @ioc: per adapter object
9305 *
9306 * The handler for doing any required cleanup or initialization.
9307 */
9308 void
mpt3sas_scsih_reset_done_handler(struct MPT3SAS_ADAPTER * ioc)9309 mpt3sas_scsih_reset_done_handler(struct MPT3SAS_ADAPTER *ioc)
9310 {
9311 dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_DONE_RESET\n", __func__));
9312 if ((!ioc->is_driver_loading) && !(disable_discovery > 0 &&
9313 !ioc->sas_hba.num_phys)) {
9314 _scsih_prep_device_scan(ioc);
9315 _scsih_create_enclosure_list_after_reset(ioc);
9316 _scsih_search_responding_sas_devices(ioc);
9317 _scsih_search_responding_pcie_devices(ioc);
9318 _scsih_search_responding_raid_devices(ioc);
9319 _scsih_search_responding_expanders(ioc);
9320 _scsih_error_recovery_delete_devices(ioc);
9321 }
9322 }
9323
9324 /**
9325 * _mpt3sas_fw_work - delayed task for processing firmware events
9326 * @ioc: per adapter object
9327 * @fw_event: The fw_event_work object
9328 * Context: user.
9329 */
9330 static void
_mpt3sas_fw_work(struct MPT3SAS_ADAPTER * ioc,struct fw_event_work * fw_event)9331 _mpt3sas_fw_work(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event)
9332 {
9333 _scsih_fw_event_del_from_list(ioc, fw_event);
9334
9335 /* the queue is being flushed so ignore this event */
9336 if (ioc->remove_host || ioc->pci_error_recovery) {
9337 fw_event_work_put(fw_event);
9338 return;
9339 }
9340
9341 switch (fw_event->event) {
9342 case MPT3SAS_PROCESS_TRIGGER_DIAG:
9343 mpt3sas_process_trigger_data(ioc,
9344 (struct SL_WH_TRIGGERS_EVENT_DATA_T *)
9345 fw_event->event_data);
9346 break;
9347 case MPT3SAS_REMOVE_UNRESPONDING_DEVICES:
9348 while (scsi_host_in_recovery(ioc->shost) ||
9349 ioc->shost_recovery) {
9350 /*
9351 * If we're unloading, bail. Otherwise, this can become
9352 * an infinite loop.
9353 */
9354 if (ioc->remove_host)
9355 goto out;
9356 ssleep(1);
9357 }
9358 _scsih_remove_unresponding_devices(ioc);
9359 _scsih_scan_for_devices_after_reset(ioc);
9360 break;
9361 case MPT3SAS_PORT_ENABLE_COMPLETE:
9362 ioc->start_scan = 0;
9363 if (missing_delay[0] != -1 && missing_delay[1] != -1)
9364 mpt3sas_base_update_missing_delay(ioc, missing_delay[0],
9365 missing_delay[1]);
9366 dewtprintk(ioc,
9367 ioc_info(ioc, "port enable: complete from worker thread\n"));
9368 break;
9369 case MPT3SAS_TURN_ON_PFA_LED:
9370 _scsih_turn_on_pfa_led(ioc, fw_event->device_handle);
9371 break;
9372 case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST:
9373 _scsih_sas_topology_change_event(ioc, fw_event);
9374 break;
9375 case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE:
9376 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
9377 _scsih_sas_device_status_change_event_debug(ioc,
9378 (Mpi2EventDataSasDeviceStatusChange_t *)
9379 fw_event->event_data);
9380 break;
9381 case MPI2_EVENT_SAS_DISCOVERY:
9382 _scsih_sas_discovery_event(ioc, fw_event);
9383 break;
9384 case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR:
9385 _scsih_sas_device_discovery_error_event(ioc, fw_event);
9386 break;
9387 case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE:
9388 _scsih_sas_broadcast_primitive_event(ioc, fw_event);
9389 break;
9390 case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE:
9391 _scsih_sas_enclosure_dev_status_change_event(ioc,
9392 fw_event);
9393 break;
9394 case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST:
9395 _scsih_sas_ir_config_change_event(ioc, fw_event);
9396 break;
9397 case MPI2_EVENT_IR_VOLUME:
9398 _scsih_sas_ir_volume_event(ioc, fw_event);
9399 break;
9400 case MPI2_EVENT_IR_PHYSICAL_DISK:
9401 _scsih_sas_ir_physical_disk_event(ioc, fw_event);
9402 break;
9403 case MPI2_EVENT_IR_OPERATION_STATUS:
9404 _scsih_sas_ir_operation_status_event(ioc, fw_event);
9405 break;
9406 case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE:
9407 _scsih_pcie_device_status_change_event(ioc, fw_event);
9408 break;
9409 case MPI2_EVENT_PCIE_ENUMERATION:
9410 _scsih_pcie_enumeration_event(ioc, fw_event);
9411 break;
9412 case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST:
9413 _scsih_pcie_topology_change_event(ioc, fw_event);
9414 return;
9415 break;
9416 }
9417 out:
9418 fw_event_work_put(fw_event);
9419 }
9420
9421 /**
9422 * _firmware_event_work
9423 * @work: The fw_event_work object
9424 * Context: user.
9425 *
9426 * wrappers for the work thread handling firmware events
9427 */
9428
9429 static void
_firmware_event_work(struct work_struct * work)9430 _firmware_event_work(struct work_struct *work)
9431 {
9432 struct fw_event_work *fw_event = container_of(work,
9433 struct fw_event_work, work);
9434
9435 _mpt3sas_fw_work(fw_event->ioc, fw_event);
9436 }
9437
9438 /**
9439 * mpt3sas_scsih_event_callback - firmware event handler (called at ISR time)
9440 * @ioc: per adapter object
9441 * @msix_index: MSIX table index supplied by the OS
9442 * @reply: reply message frame(lower 32bit addr)
9443 * Context: interrupt.
9444 *
9445 * This function merely adds a new work task into ioc->firmware_event_thread.
9446 * The tasks are worked from _firmware_event_work in user context.
9447 *
9448 * Return: 1 meaning mf should be freed from _base_interrupt
9449 * 0 means the mf is freed from this function.
9450 */
9451 u8
mpt3sas_scsih_event_callback(struct MPT3SAS_ADAPTER * ioc,u8 msix_index,u32 reply)9452 mpt3sas_scsih_event_callback(struct MPT3SAS_ADAPTER *ioc, u8 msix_index,
9453 u32 reply)
9454 {
9455 struct fw_event_work *fw_event;
9456 Mpi2EventNotificationReply_t *mpi_reply;
9457 u16 event;
9458 u16 sz;
9459 Mpi26EventDataActiveCableExcept_t *ActiveCableEventData;
9460
9461 /* events turned off due to host reset */
9462 if (ioc->pci_error_recovery)
9463 return 1;
9464
9465 mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply);
9466
9467 if (unlikely(!mpi_reply)) {
9468 ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n",
9469 __FILE__, __LINE__, __func__);
9470 return 1;
9471 }
9472
9473 event = le16_to_cpu(mpi_reply->Event);
9474
9475 if (event != MPI2_EVENT_LOG_ENTRY_ADDED)
9476 mpt3sas_trigger_event(ioc, event, 0);
9477
9478 switch (event) {
9479 /* handle these */
9480 case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE:
9481 {
9482 Mpi2EventDataSasBroadcastPrimitive_t *baen_data =
9483 (Mpi2EventDataSasBroadcastPrimitive_t *)
9484 mpi_reply->EventData;
9485
9486 if (baen_data->Primitive !=
9487 MPI2_EVENT_PRIMITIVE_ASYNCHRONOUS_EVENT)
9488 return 1;
9489
9490 if (ioc->broadcast_aen_busy) {
9491 ioc->broadcast_aen_pending++;
9492 return 1;
9493 } else
9494 ioc->broadcast_aen_busy = 1;
9495 break;
9496 }
9497
9498 case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST:
9499 _scsih_check_topo_delete_events(ioc,
9500 (Mpi2EventDataSasTopologyChangeList_t *)
9501 mpi_reply->EventData);
9502 break;
9503 case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST:
9504 _scsih_check_pcie_topo_remove_events(ioc,
9505 (Mpi26EventDataPCIeTopologyChangeList_t *)
9506 mpi_reply->EventData);
9507 break;
9508 case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST:
9509 _scsih_check_ir_config_unhide_events(ioc,
9510 (Mpi2EventDataIrConfigChangeList_t *)
9511 mpi_reply->EventData);
9512 break;
9513 case MPI2_EVENT_IR_VOLUME:
9514 _scsih_check_volume_delete_events(ioc,
9515 (Mpi2EventDataIrVolume_t *)
9516 mpi_reply->EventData);
9517 break;
9518 case MPI2_EVENT_LOG_ENTRY_ADDED:
9519 {
9520 Mpi2EventDataLogEntryAdded_t *log_entry;
9521 u32 *log_code;
9522
9523 if (!ioc->is_warpdrive)
9524 break;
9525
9526 log_entry = (Mpi2EventDataLogEntryAdded_t *)
9527 mpi_reply->EventData;
9528 log_code = (u32 *)log_entry->LogData;
9529
9530 if (le16_to_cpu(log_entry->LogEntryQualifier)
9531 != MPT2_WARPDRIVE_LOGENTRY)
9532 break;
9533
9534 switch (le32_to_cpu(*log_code)) {
9535 case MPT2_WARPDRIVE_LC_SSDT:
9536 ioc_warn(ioc, "WarpDrive Warning: IO Throttling has occurred in the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n");
9537 break;
9538 case MPT2_WARPDRIVE_LC_SSDLW:
9539 ioc_warn(ioc, "WarpDrive Warning: Program/Erase Cycles for the WarpDrive subsystem in degraded range. Check WarpDrive documentation for additional details.\n");
9540 break;
9541 case MPT2_WARPDRIVE_LC_SSDLF:
9542 ioc_err(ioc, "WarpDrive Fatal Error: There are no Program/Erase Cycles for the WarpDrive subsystem. The storage device will be in read-only mode. Check WarpDrive documentation for additional details.\n");
9543 break;
9544 case MPT2_WARPDRIVE_LC_BRMF:
9545 ioc_err(ioc, "WarpDrive Fatal Error: The Backup Rail Monitor has failed on the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n");
9546 break;
9547 }
9548
9549 break;
9550 }
9551 case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE:
9552 _scsih_sas_device_status_change_event(ioc,
9553 (Mpi2EventDataSasDeviceStatusChange_t *)
9554 mpi_reply->EventData);
9555 break;
9556 case MPI2_EVENT_IR_OPERATION_STATUS:
9557 case MPI2_EVENT_SAS_DISCOVERY:
9558 case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR:
9559 case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE:
9560 case MPI2_EVENT_IR_PHYSICAL_DISK:
9561 case MPI2_EVENT_PCIE_ENUMERATION:
9562 case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE:
9563 break;
9564
9565 case MPI2_EVENT_TEMP_THRESHOLD:
9566 _scsih_temp_threshold_events(ioc,
9567 (Mpi2EventDataTemperature_t *)
9568 mpi_reply->EventData);
9569 break;
9570 case MPI2_EVENT_ACTIVE_CABLE_EXCEPTION:
9571 ActiveCableEventData =
9572 (Mpi26EventDataActiveCableExcept_t *) mpi_reply->EventData;
9573 switch (ActiveCableEventData->ReasonCode) {
9574 case MPI26_EVENT_ACTIVE_CABLE_INSUFFICIENT_POWER:
9575 ioc_notice(ioc, "Currently an active cable with ReceptacleID %d\n",
9576 ActiveCableEventData->ReceptacleID);
9577 pr_notice("cannot be powered and devices connected\n");
9578 pr_notice("to this active cable will not be seen\n");
9579 pr_notice("This active cable requires %d mW of power\n",
9580 ActiveCableEventData->ActiveCablePowerRequirement);
9581 break;
9582
9583 case MPI26_EVENT_ACTIVE_CABLE_DEGRADED:
9584 ioc_notice(ioc, "Currently a cable with ReceptacleID %d\n",
9585 ActiveCableEventData->ReceptacleID);
9586 pr_notice(
9587 "is not running at optimal speed(12 Gb/s rate)\n");
9588 break;
9589 }
9590
9591 break;
9592
9593 default: /* ignore the rest */
9594 return 1;
9595 }
9596
9597 sz = le16_to_cpu(mpi_reply->EventDataLength) * 4;
9598 fw_event = alloc_fw_event_work(sz);
9599 if (!fw_event) {
9600 ioc_err(ioc, "failure at %s:%d/%s()!\n",
9601 __FILE__, __LINE__, __func__);
9602 return 1;
9603 }
9604
9605 memcpy(fw_event->event_data, mpi_reply->EventData, sz);
9606 fw_event->ioc = ioc;
9607 fw_event->VF_ID = mpi_reply->VF_ID;
9608 fw_event->VP_ID = mpi_reply->VP_ID;
9609 fw_event->event = event;
9610 _scsih_fw_event_add(ioc, fw_event);
9611 fw_event_work_put(fw_event);
9612 return 1;
9613 }
9614
9615 /**
9616 * _scsih_expander_node_remove - removing expander device from list.
9617 * @ioc: per adapter object
9618 * @sas_expander: the sas_device object
9619 *
9620 * Removing object and freeing associated memory from the
9621 * ioc->sas_expander_list.
9622 */
9623 static void
_scsih_expander_node_remove(struct MPT3SAS_ADAPTER * ioc,struct _sas_node * sas_expander)9624 _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc,
9625 struct _sas_node *sas_expander)
9626 {
9627 struct _sas_port *mpt3sas_port, *next;
9628 unsigned long flags;
9629
9630 /* remove sibling ports attached to this expander */
9631 list_for_each_entry_safe(mpt3sas_port, next,
9632 &sas_expander->sas_port_list, port_list) {
9633 if (ioc->shost_recovery)
9634 return;
9635 if (mpt3sas_port->remote_identify.device_type ==
9636 SAS_END_DEVICE)
9637 mpt3sas_device_remove_by_sas_address(ioc,
9638 mpt3sas_port->remote_identify.sas_address);
9639 else if (mpt3sas_port->remote_identify.device_type ==
9640 SAS_EDGE_EXPANDER_DEVICE ||
9641 mpt3sas_port->remote_identify.device_type ==
9642 SAS_FANOUT_EXPANDER_DEVICE)
9643 mpt3sas_expander_remove(ioc,
9644 mpt3sas_port->remote_identify.sas_address);
9645 }
9646
9647 mpt3sas_transport_port_remove(ioc, sas_expander->sas_address,
9648 sas_expander->sas_address_parent);
9649
9650 ioc_info(ioc, "expander_remove: handle(0x%04x), sas_addr(0x%016llx)\n",
9651 sas_expander->handle, (unsigned long long)
9652 sas_expander->sas_address);
9653
9654 spin_lock_irqsave(&ioc->sas_node_lock, flags);
9655 list_del(&sas_expander->list);
9656 spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
9657
9658 kfree(sas_expander->phy);
9659 kfree(sas_expander);
9660 }
9661
9662 /**
9663 * _scsih_ir_shutdown - IR shutdown notification
9664 * @ioc: per adapter object
9665 *
9666 * Sending RAID Action to alert the Integrated RAID subsystem of the IOC that
9667 * the host system is shutting down.
9668 */
9669 static void
_scsih_ir_shutdown(struct MPT3SAS_ADAPTER * ioc)9670 _scsih_ir_shutdown(struct MPT3SAS_ADAPTER *ioc)
9671 {
9672 Mpi2RaidActionRequest_t *mpi_request;
9673 Mpi2RaidActionReply_t *mpi_reply;
9674 u16 smid;
9675
9676 /* is IR firmware build loaded ? */
9677 if (!ioc->ir_firmware)
9678 return;
9679
9680 /* are there any volumes ? */
9681 if (list_empty(&ioc->raid_device_list))
9682 return;
9683
9684 mutex_lock(&ioc->scsih_cmds.mutex);
9685
9686 if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) {
9687 ioc_err(ioc, "%s: scsih_cmd in use\n", __func__);
9688 goto out;
9689 }
9690 ioc->scsih_cmds.status = MPT3_CMD_PENDING;
9691
9692 smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx);
9693 if (!smid) {
9694 ioc_err(ioc, "%s: failed obtaining a smid\n", __func__);
9695 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
9696 goto out;
9697 }
9698
9699 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
9700 ioc->scsih_cmds.smid = smid;
9701 memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t));
9702
9703 mpi_request->Function = MPI2_FUNCTION_RAID_ACTION;
9704 mpi_request->Action = MPI2_RAID_ACTION_SYSTEM_SHUTDOWN_INITIATED;
9705
9706 if (!ioc->hide_ir_msg)
9707 ioc_info(ioc, "IR shutdown (sending)\n");
9708 init_completion(&ioc->scsih_cmds.done);
9709 ioc->put_smid_default(ioc, smid);
9710 wait_for_completion_timeout(&ioc->scsih_cmds.done, 10*HZ);
9711
9712 if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) {
9713 ioc_err(ioc, "%s: timeout\n", __func__);
9714 goto out;
9715 }
9716
9717 if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) {
9718 mpi_reply = ioc->scsih_cmds.reply;
9719 if (!ioc->hide_ir_msg)
9720 ioc_info(ioc, "IR shutdown (complete): ioc_status(0x%04x), loginfo(0x%08x)\n",
9721 le16_to_cpu(mpi_reply->IOCStatus),
9722 le32_to_cpu(mpi_reply->IOCLogInfo));
9723 }
9724
9725 out:
9726 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
9727 mutex_unlock(&ioc->scsih_cmds.mutex);
9728 }
9729
9730 /**
9731 * scsih_remove - detach and remove add host
9732 * @pdev: PCI device struct
9733 *
9734 * Routine called when unloading the driver.
9735 */
scsih_remove(struct pci_dev * pdev)9736 static void scsih_remove(struct pci_dev *pdev)
9737 {
9738 struct Scsi_Host *shost = pci_get_drvdata(pdev);
9739 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
9740 struct _sas_port *mpt3sas_port, *next_port;
9741 struct _raid_device *raid_device, *next;
9742 struct MPT3SAS_TARGET *sas_target_priv_data;
9743 struct _pcie_device *pcie_device, *pcienext;
9744 struct workqueue_struct *wq;
9745 unsigned long flags;
9746 Mpi2ConfigReply_t mpi_reply;
9747
9748 ioc->remove_host = 1;
9749
9750 mpt3sas_wait_for_commands_to_complete(ioc);
9751 _scsih_flush_running_cmds(ioc);
9752
9753 _scsih_fw_event_cleanup_queue(ioc);
9754
9755 spin_lock_irqsave(&ioc->fw_event_lock, flags);
9756 wq = ioc->firmware_event_thread;
9757 ioc->firmware_event_thread = NULL;
9758 spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
9759 if (wq)
9760 destroy_workqueue(wq);
9761 /*
9762 * Copy back the unmodified ioc page1. so that on next driver load,
9763 * current modified changes on ioc page1 won't take effect.
9764 */
9765 if (ioc->is_aero_ioc)
9766 mpt3sas_config_set_ioc_pg1(ioc, &mpi_reply,
9767 &ioc->ioc_pg1_copy);
9768 /* release all the volumes */
9769 _scsih_ir_shutdown(ioc);
9770 sas_remove_host(shost);
9771 list_for_each_entry_safe(raid_device, next, &ioc->raid_device_list,
9772 list) {
9773 if (raid_device->starget) {
9774 sas_target_priv_data =
9775 raid_device->starget->hostdata;
9776 sas_target_priv_data->deleted = 1;
9777 scsi_remove_target(&raid_device->starget->dev);
9778 }
9779 ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n",
9780 raid_device->handle, (u64)raid_device->wwid);
9781 _scsih_raid_device_remove(ioc, raid_device);
9782 }
9783 list_for_each_entry_safe(pcie_device, pcienext, &ioc->pcie_device_list,
9784 list) {
9785 _scsih_pcie_device_remove_from_sml(ioc, pcie_device);
9786 list_del_init(&pcie_device->list);
9787 pcie_device_put(pcie_device);
9788 }
9789
9790 /* free ports attached to the sas_host */
9791 list_for_each_entry_safe(mpt3sas_port, next_port,
9792 &ioc->sas_hba.sas_port_list, port_list) {
9793 if (mpt3sas_port->remote_identify.device_type ==
9794 SAS_END_DEVICE)
9795 mpt3sas_device_remove_by_sas_address(ioc,
9796 mpt3sas_port->remote_identify.sas_address);
9797 else if (mpt3sas_port->remote_identify.device_type ==
9798 SAS_EDGE_EXPANDER_DEVICE ||
9799 mpt3sas_port->remote_identify.device_type ==
9800 SAS_FANOUT_EXPANDER_DEVICE)
9801 mpt3sas_expander_remove(ioc,
9802 mpt3sas_port->remote_identify.sas_address);
9803 }
9804
9805 /* free phys attached to the sas_host */
9806 if (ioc->sas_hba.num_phys) {
9807 kfree(ioc->sas_hba.phy);
9808 ioc->sas_hba.phy = NULL;
9809 ioc->sas_hba.num_phys = 0;
9810 }
9811
9812 mpt3sas_base_detach(ioc);
9813 spin_lock(&gioc_lock);
9814 list_del(&ioc->list);
9815 spin_unlock(&gioc_lock);
9816 scsi_host_put(shost);
9817 }
9818
9819 /**
9820 * scsih_shutdown - routine call during system shutdown
9821 * @pdev: PCI device struct
9822 */
9823 static void
scsih_shutdown(struct pci_dev * pdev)9824 scsih_shutdown(struct pci_dev *pdev)
9825 {
9826 struct Scsi_Host *shost = pci_get_drvdata(pdev);
9827 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
9828 struct workqueue_struct *wq;
9829 unsigned long flags;
9830 Mpi2ConfigReply_t mpi_reply;
9831
9832 ioc->remove_host = 1;
9833
9834 mpt3sas_wait_for_commands_to_complete(ioc);
9835 _scsih_flush_running_cmds(ioc);
9836
9837 _scsih_fw_event_cleanup_queue(ioc);
9838
9839 spin_lock_irqsave(&ioc->fw_event_lock, flags);
9840 wq = ioc->firmware_event_thread;
9841 ioc->firmware_event_thread = NULL;
9842 spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
9843 if (wq)
9844 destroy_workqueue(wq);
9845 /*
9846 * Copy back the unmodified ioc page1 so that on next driver load,
9847 * current modified changes on ioc page1 won't take effect.
9848 */
9849 if (ioc->is_aero_ioc)
9850 mpt3sas_config_set_ioc_pg1(ioc, &mpi_reply,
9851 &ioc->ioc_pg1_copy);
9852
9853 _scsih_ir_shutdown(ioc);
9854 mpt3sas_base_detach(ioc);
9855 }
9856
9857
9858 /**
9859 * _scsih_probe_boot_devices - reports 1st device
9860 * @ioc: per adapter object
9861 *
9862 * If specified in bios page 2, this routine reports the 1st
9863 * device scsi-ml or sas transport for persistent boot device
9864 * purposes. Please refer to function _scsih_determine_boot_device()
9865 */
9866 static void
_scsih_probe_boot_devices(struct MPT3SAS_ADAPTER * ioc)9867 _scsih_probe_boot_devices(struct MPT3SAS_ADAPTER *ioc)
9868 {
9869 u32 channel;
9870 void *device;
9871 struct _sas_device *sas_device;
9872 struct _raid_device *raid_device;
9873 struct _pcie_device *pcie_device;
9874 u16 handle;
9875 u64 sas_address_parent;
9876 u64 sas_address;
9877 unsigned long flags;
9878 int rc;
9879 int tid;
9880
9881 /* no Bios, return immediately */
9882 if (!ioc->bios_pg3.BiosVersion)
9883 return;
9884
9885 device = NULL;
9886 if (ioc->req_boot_device.device) {
9887 device = ioc->req_boot_device.device;
9888 channel = ioc->req_boot_device.channel;
9889 } else if (ioc->req_alt_boot_device.device) {
9890 device = ioc->req_alt_boot_device.device;
9891 channel = ioc->req_alt_boot_device.channel;
9892 } else if (ioc->current_boot_device.device) {
9893 device = ioc->current_boot_device.device;
9894 channel = ioc->current_boot_device.channel;
9895 }
9896
9897 if (!device)
9898 return;
9899
9900 if (channel == RAID_CHANNEL) {
9901 raid_device = device;
9902 rc = scsi_add_device(ioc->shost, RAID_CHANNEL,
9903 raid_device->id, 0);
9904 if (rc)
9905 _scsih_raid_device_remove(ioc, raid_device);
9906 } else if (channel == PCIE_CHANNEL) {
9907 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
9908 pcie_device = device;
9909 tid = pcie_device->id;
9910 list_move_tail(&pcie_device->list, &ioc->pcie_device_list);
9911 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
9912 rc = scsi_add_device(ioc->shost, PCIE_CHANNEL, tid, 0);
9913 if (rc)
9914 _scsih_pcie_device_remove(ioc, pcie_device);
9915 } else {
9916 spin_lock_irqsave(&ioc->sas_device_lock, flags);
9917 sas_device = device;
9918 handle = sas_device->handle;
9919 sas_address_parent = sas_device->sas_address_parent;
9920 sas_address = sas_device->sas_address;
9921 list_move_tail(&sas_device->list, &ioc->sas_device_list);
9922 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
9923
9924 if (ioc->hide_drives)
9925 return;
9926 if (!mpt3sas_transport_port_add(ioc, handle,
9927 sas_address_parent)) {
9928 _scsih_sas_device_remove(ioc, sas_device);
9929 } else if (!sas_device->starget) {
9930 if (!ioc->is_driver_loading) {
9931 mpt3sas_transport_port_remove(ioc,
9932 sas_address,
9933 sas_address_parent);
9934 _scsih_sas_device_remove(ioc, sas_device);
9935 }
9936 }
9937 }
9938 }
9939
9940 /**
9941 * _scsih_probe_raid - reporting raid volumes to scsi-ml
9942 * @ioc: per adapter object
9943 *
9944 * Called during initial loading of the driver.
9945 */
9946 static void
_scsih_probe_raid(struct MPT3SAS_ADAPTER * ioc)9947 _scsih_probe_raid(struct MPT3SAS_ADAPTER *ioc)
9948 {
9949 struct _raid_device *raid_device, *raid_next;
9950 int rc;
9951
9952 list_for_each_entry_safe(raid_device, raid_next,
9953 &ioc->raid_device_list, list) {
9954 if (raid_device->starget)
9955 continue;
9956 rc = scsi_add_device(ioc->shost, RAID_CHANNEL,
9957 raid_device->id, 0);
9958 if (rc)
9959 _scsih_raid_device_remove(ioc, raid_device);
9960 }
9961 }
9962
get_next_sas_device(struct MPT3SAS_ADAPTER * ioc)9963 static struct _sas_device *get_next_sas_device(struct MPT3SAS_ADAPTER *ioc)
9964 {
9965 struct _sas_device *sas_device = NULL;
9966 unsigned long flags;
9967
9968 spin_lock_irqsave(&ioc->sas_device_lock, flags);
9969 if (!list_empty(&ioc->sas_device_init_list)) {
9970 sas_device = list_first_entry(&ioc->sas_device_init_list,
9971 struct _sas_device, list);
9972 sas_device_get(sas_device);
9973 }
9974 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
9975
9976 return sas_device;
9977 }
9978
sas_device_make_active(struct MPT3SAS_ADAPTER * ioc,struct _sas_device * sas_device)9979 static void sas_device_make_active(struct MPT3SAS_ADAPTER *ioc,
9980 struct _sas_device *sas_device)
9981 {
9982 unsigned long flags;
9983
9984 spin_lock_irqsave(&ioc->sas_device_lock, flags);
9985
9986 /*
9987 * Since we dropped the lock during the call to port_add(), we need to
9988 * be careful here that somebody else didn't move or delete this item
9989 * while we were busy with other things.
9990 *
9991 * If it was on the list, we need a put() for the reference the list
9992 * had. Either way, we need a get() for the destination list.
9993 */
9994 if (!list_empty(&sas_device->list)) {
9995 list_del_init(&sas_device->list);
9996 sas_device_put(sas_device);
9997 }
9998
9999 sas_device_get(sas_device);
10000 list_add_tail(&sas_device->list, &ioc->sas_device_list);
10001
10002 spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
10003 }
10004
10005 /**
10006 * _scsih_probe_sas - reporting sas devices to sas transport
10007 * @ioc: per adapter object
10008 *
10009 * Called during initial loading of the driver.
10010 */
10011 static void
_scsih_probe_sas(struct MPT3SAS_ADAPTER * ioc)10012 _scsih_probe_sas(struct MPT3SAS_ADAPTER *ioc)
10013 {
10014 struct _sas_device *sas_device;
10015
10016 if (ioc->hide_drives)
10017 return;
10018
10019 while ((sas_device = get_next_sas_device(ioc))) {
10020 if (!mpt3sas_transport_port_add(ioc, sas_device->handle,
10021 sas_device->sas_address_parent)) {
10022 _scsih_sas_device_remove(ioc, sas_device);
10023 sas_device_put(sas_device);
10024 continue;
10025 } else if (!sas_device->starget) {
10026 /*
10027 * When asyn scanning is enabled, its not possible to
10028 * remove devices while scanning is turned on due to an
10029 * oops in scsi_sysfs_add_sdev()->add_device()->
10030 * sysfs_addrm_start()
10031 */
10032 if (!ioc->is_driver_loading) {
10033 mpt3sas_transport_port_remove(ioc,
10034 sas_device->sas_address,
10035 sas_device->sas_address_parent);
10036 _scsih_sas_device_remove(ioc, sas_device);
10037 sas_device_put(sas_device);
10038 continue;
10039 }
10040 }
10041 sas_device_make_active(ioc, sas_device);
10042 sas_device_put(sas_device);
10043 }
10044 }
10045
10046 /**
10047 * get_next_pcie_device - Get the next pcie device
10048 * @ioc: per adapter object
10049 *
10050 * Get the next pcie device from pcie_device_init_list list.
10051 *
10052 * Return: pcie device structure if pcie_device_init_list list is not empty
10053 * otherwise returns NULL
10054 */
get_next_pcie_device(struct MPT3SAS_ADAPTER * ioc)10055 static struct _pcie_device *get_next_pcie_device(struct MPT3SAS_ADAPTER *ioc)
10056 {
10057 struct _pcie_device *pcie_device = NULL;
10058 unsigned long flags;
10059
10060 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
10061 if (!list_empty(&ioc->pcie_device_init_list)) {
10062 pcie_device = list_first_entry(&ioc->pcie_device_init_list,
10063 struct _pcie_device, list);
10064 pcie_device_get(pcie_device);
10065 }
10066 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
10067
10068 return pcie_device;
10069 }
10070
10071 /**
10072 * pcie_device_make_active - Add pcie device to pcie_device_list list
10073 * @ioc: per adapter object
10074 * @pcie_device: pcie device object
10075 *
10076 * Add the pcie device which has registered with SCSI Transport Later to
10077 * pcie_device_list list
10078 */
pcie_device_make_active(struct MPT3SAS_ADAPTER * ioc,struct _pcie_device * pcie_device)10079 static void pcie_device_make_active(struct MPT3SAS_ADAPTER *ioc,
10080 struct _pcie_device *pcie_device)
10081 {
10082 unsigned long flags;
10083
10084 spin_lock_irqsave(&ioc->pcie_device_lock, flags);
10085
10086 if (!list_empty(&pcie_device->list)) {
10087 list_del_init(&pcie_device->list);
10088 pcie_device_put(pcie_device);
10089 }
10090 pcie_device_get(pcie_device);
10091 list_add_tail(&pcie_device->list, &ioc->pcie_device_list);
10092
10093 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
10094 }
10095
10096 /**
10097 * _scsih_probe_pcie - reporting PCIe devices to scsi-ml
10098 * @ioc: per adapter object
10099 *
10100 * Called during initial loading of the driver.
10101 */
10102 static void
_scsih_probe_pcie(struct MPT3SAS_ADAPTER * ioc)10103 _scsih_probe_pcie(struct MPT3SAS_ADAPTER *ioc)
10104 {
10105 struct _pcie_device *pcie_device;
10106 int rc;
10107
10108 /* PCIe Device List */
10109 while ((pcie_device = get_next_pcie_device(ioc))) {
10110 if (pcie_device->starget) {
10111 pcie_device_put(pcie_device);
10112 continue;
10113 }
10114 if (pcie_device->access_status ==
10115 MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) {
10116 pcie_device_make_active(ioc, pcie_device);
10117 pcie_device_put(pcie_device);
10118 continue;
10119 }
10120 rc = scsi_add_device(ioc->shost, PCIE_CHANNEL,
10121 pcie_device->id, 0);
10122 if (rc) {
10123 _scsih_pcie_device_remove(ioc, pcie_device);
10124 pcie_device_put(pcie_device);
10125 continue;
10126 } else if (!pcie_device->starget) {
10127 /*
10128 * When async scanning is enabled, its not possible to
10129 * remove devices while scanning is turned on due to an
10130 * oops in scsi_sysfs_add_sdev()->add_device()->
10131 * sysfs_addrm_start()
10132 */
10133 if (!ioc->is_driver_loading) {
10134 /* TODO-- Need to find out whether this condition will
10135 * occur or not
10136 */
10137 _scsih_pcie_device_remove(ioc, pcie_device);
10138 pcie_device_put(pcie_device);
10139 continue;
10140 }
10141 }
10142 pcie_device_make_active(ioc, pcie_device);
10143 pcie_device_put(pcie_device);
10144 }
10145 }
10146
10147 /**
10148 * _scsih_probe_devices - probing for devices
10149 * @ioc: per adapter object
10150 *
10151 * Called during initial loading of the driver.
10152 */
10153 static void
_scsih_probe_devices(struct MPT3SAS_ADAPTER * ioc)10154 _scsih_probe_devices(struct MPT3SAS_ADAPTER *ioc)
10155 {
10156 u16 volume_mapping_flags;
10157
10158 if (!(ioc->facts.ProtocolFlags & MPI2_IOCFACTS_PROTOCOL_SCSI_INITIATOR))
10159 return; /* return when IOC doesn't support initiator mode */
10160
10161 _scsih_probe_boot_devices(ioc);
10162
10163 if (ioc->ir_firmware) {
10164 volume_mapping_flags =
10165 le16_to_cpu(ioc->ioc_pg8.IRVolumeMappingFlags) &
10166 MPI2_IOCPAGE8_IRFLAGS_MASK_VOLUME_MAPPING_MODE;
10167 if (volume_mapping_flags ==
10168 MPI2_IOCPAGE8_IRFLAGS_LOW_VOLUME_MAPPING) {
10169 _scsih_probe_raid(ioc);
10170 _scsih_probe_sas(ioc);
10171 } else {
10172 _scsih_probe_sas(ioc);
10173 _scsih_probe_raid(ioc);
10174 }
10175 } else {
10176 _scsih_probe_sas(ioc);
10177 _scsih_probe_pcie(ioc);
10178 }
10179 }
10180
10181 /**
10182 * scsih_scan_start - scsi lld callback for .scan_start
10183 * @shost: SCSI host pointer
10184 *
10185 * The shost has the ability to discover targets on its own instead
10186 * of scanning the entire bus. In our implemention, we will kick off
10187 * firmware discovery.
10188 */
10189 static void
scsih_scan_start(struct Scsi_Host * shost)10190 scsih_scan_start(struct Scsi_Host *shost)
10191 {
10192 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
10193 int rc;
10194 if (diag_buffer_enable != -1 && diag_buffer_enable != 0)
10195 mpt3sas_enable_diag_buffer(ioc, diag_buffer_enable);
10196
10197 if (disable_discovery > 0)
10198 return;
10199
10200 ioc->start_scan = 1;
10201 rc = mpt3sas_port_enable(ioc);
10202
10203 if (rc != 0)
10204 ioc_info(ioc, "port enable: FAILED\n");
10205 }
10206
10207 /**
10208 * scsih_scan_finished - scsi lld callback for .scan_finished
10209 * @shost: SCSI host pointer
10210 * @time: elapsed time of the scan in jiffies
10211 *
10212 * This function will be called periodicallyn until it returns 1 with the
10213 * scsi_host and the elapsed time of the scan in jiffies. In our implemention,
10214 * we wait for firmware discovery to complete, then return 1.
10215 */
10216 static int
scsih_scan_finished(struct Scsi_Host * shost,unsigned long time)10217 scsih_scan_finished(struct Scsi_Host *shost, unsigned long time)
10218 {
10219 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
10220
10221 if (disable_discovery > 0) {
10222 ioc->is_driver_loading = 0;
10223 ioc->wait_for_discovery_to_complete = 0;
10224 return 1;
10225 }
10226
10227 if (time >= (300 * HZ)) {
10228 ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED;
10229 ioc_info(ioc, "port enable: FAILED with timeout (timeout=300s)\n");
10230 ioc->is_driver_loading = 0;
10231 return 1;
10232 }
10233
10234 if (ioc->start_scan)
10235 return 0;
10236
10237 if (ioc->start_scan_failed) {
10238 ioc_info(ioc, "port enable: FAILED with (ioc_status=0x%08x)\n",
10239 ioc->start_scan_failed);
10240 ioc->is_driver_loading = 0;
10241 ioc->wait_for_discovery_to_complete = 0;
10242 ioc->remove_host = 1;
10243 return 1;
10244 }
10245
10246 ioc_info(ioc, "port enable: SUCCESS\n");
10247 ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED;
10248
10249 if (ioc->wait_for_discovery_to_complete) {
10250 ioc->wait_for_discovery_to_complete = 0;
10251 _scsih_probe_devices(ioc);
10252 }
10253 mpt3sas_base_start_watchdog(ioc);
10254 ioc->is_driver_loading = 0;
10255 return 1;
10256 }
10257
10258 /* shost template for SAS 2.0 HBA devices */
10259 static struct scsi_host_template mpt2sas_driver_template = {
10260 .module = THIS_MODULE,
10261 .name = "Fusion MPT SAS Host",
10262 .proc_name = MPT2SAS_DRIVER_NAME,
10263 .queuecommand = scsih_qcmd,
10264 .target_alloc = scsih_target_alloc,
10265 .slave_alloc = scsih_slave_alloc,
10266 .slave_configure = scsih_slave_configure,
10267 .target_destroy = scsih_target_destroy,
10268 .slave_destroy = scsih_slave_destroy,
10269 .scan_finished = scsih_scan_finished,
10270 .scan_start = scsih_scan_start,
10271 .change_queue_depth = scsih_change_queue_depth,
10272 .eh_abort_handler = scsih_abort,
10273 .eh_device_reset_handler = scsih_dev_reset,
10274 .eh_target_reset_handler = scsih_target_reset,
10275 .eh_host_reset_handler = scsih_host_reset,
10276 .bios_param = scsih_bios_param,
10277 .can_queue = 1,
10278 .this_id = -1,
10279 .sg_tablesize = MPT2SAS_SG_DEPTH,
10280 .max_sectors = 32767,
10281 .cmd_per_lun = 7,
10282 .shost_attrs = mpt3sas_host_attrs,
10283 .sdev_attrs = mpt3sas_dev_attrs,
10284 .track_queue_depth = 1,
10285 .cmd_size = sizeof(struct scsiio_tracker),
10286 };
10287
10288 /* raid transport support for SAS 2.0 HBA devices */
10289 static struct raid_function_template mpt2sas_raid_functions = {
10290 .cookie = &mpt2sas_driver_template,
10291 .is_raid = scsih_is_raid,
10292 .get_resync = scsih_get_resync,
10293 .get_state = scsih_get_state,
10294 };
10295
10296 /* shost template for SAS 3.0 HBA devices */
10297 static struct scsi_host_template mpt3sas_driver_template = {
10298 .module = THIS_MODULE,
10299 .name = "Fusion MPT SAS Host",
10300 .proc_name = MPT3SAS_DRIVER_NAME,
10301 .queuecommand = scsih_qcmd,
10302 .target_alloc = scsih_target_alloc,
10303 .slave_alloc = scsih_slave_alloc,
10304 .slave_configure = scsih_slave_configure,
10305 .target_destroy = scsih_target_destroy,
10306 .slave_destroy = scsih_slave_destroy,
10307 .scan_finished = scsih_scan_finished,
10308 .scan_start = scsih_scan_start,
10309 .change_queue_depth = scsih_change_queue_depth,
10310 .eh_abort_handler = scsih_abort,
10311 .eh_device_reset_handler = scsih_dev_reset,
10312 .eh_target_reset_handler = scsih_target_reset,
10313 .eh_host_reset_handler = scsih_host_reset,
10314 .bios_param = scsih_bios_param,
10315 .can_queue = 1,
10316 .this_id = -1,
10317 .sg_tablesize = MPT3SAS_SG_DEPTH,
10318 .max_sectors = 32767,
10319 .max_segment_size = 0xffffffff,
10320 .cmd_per_lun = 7,
10321 .shost_attrs = mpt3sas_host_attrs,
10322 .sdev_attrs = mpt3sas_dev_attrs,
10323 .track_queue_depth = 1,
10324 .cmd_size = sizeof(struct scsiio_tracker),
10325 };
10326
10327 /* raid transport support for SAS 3.0 HBA devices */
10328 static struct raid_function_template mpt3sas_raid_functions = {
10329 .cookie = &mpt3sas_driver_template,
10330 .is_raid = scsih_is_raid,
10331 .get_resync = scsih_get_resync,
10332 .get_state = scsih_get_state,
10333 };
10334
10335 /**
10336 * _scsih_determine_hba_mpi_version - determine in which MPI version class
10337 * this device belongs to.
10338 * @pdev: PCI device struct
10339 *
10340 * return MPI2_VERSION for SAS 2.0 HBA devices,
10341 * MPI25_VERSION for SAS 3.0 HBA devices, and
10342 * MPI26 VERSION for Cutlass & Invader SAS 3.0 HBA devices
10343 */
10344 static u16
_scsih_determine_hba_mpi_version(struct pci_dev * pdev)10345 _scsih_determine_hba_mpi_version(struct pci_dev *pdev)
10346 {
10347
10348 switch (pdev->device) {
10349 case MPI2_MFGPAGE_DEVID_SSS6200:
10350 case MPI2_MFGPAGE_DEVID_SAS2004:
10351 case MPI2_MFGPAGE_DEVID_SAS2008:
10352 case MPI2_MFGPAGE_DEVID_SAS2108_1:
10353 case MPI2_MFGPAGE_DEVID_SAS2108_2:
10354 case MPI2_MFGPAGE_DEVID_SAS2108_3:
10355 case MPI2_MFGPAGE_DEVID_SAS2116_1:
10356 case MPI2_MFGPAGE_DEVID_SAS2116_2:
10357 case MPI2_MFGPAGE_DEVID_SAS2208_1:
10358 case MPI2_MFGPAGE_DEVID_SAS2208_2:
10359 case MPI2_MFGPAGE_DEVID_SAS2208_3:
10360 case MPI2_MFGPAGE_DEVID_SAS2208_4:
10361 case MPI2_MFGPAGE_DEVID_SAS2208_5:
10362 case MPI2_MFGPAGE_DEVID_SAS2208_6:
10363 case MPI2_MFGPAGE_DEVID_SAS2308_1:
10364 case MPI2_MFGPAGE_DEVID_SAS2308_2:
10365 case MPI2_MFGPAGE_DEVID_SAS2308_3:
10366 case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP:
10367 case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1:
10368 return MPI2_VERSION;
10369 case MPI25_MFGPAGE_DEVID_SAS3004:
10370 case MPI25_MFGPAGE_DEVID_SAS3008:
10371 case MPI25_MFGPAGE_DEVID_SAS3108_1:
10372 case MPI25_MFGPAGE_DEVID_SAS3108_2:
10373 case MPI25_MFGPAGE_DEVID_SAS3108_5:
10374 case MPI25_MFGPAGE_DEVID_SAS3108_6:
10375 return MPI25_VERSION;
10376 case MPI26_MFGPAGE_DEVID_SAS3216:
10377 case MPI26_MFGPAGE_DEVID_SAS3224:
10378 case MPI26_MFGPAGE_DEVID_SAS3316_1:
10379 case MPI26_MFGPAGE_DEVID_SAS3316_2:
10380 case MPI26_MFGPAGE_DEVID_SAS3316_3:
10381 case MPI26_MFGPAGE_DEVID_SAS3316_4:
10382 case MPI26_MFGPAGE_DEVID_SAS3324_1:
10383 case MPI26_MFGPAGE_DEVID_SAS3324_2:
10384 case MPI26_MFGPAGE_DEVID_SAS3324_3:
10385 case MPI26_MFGPAGE_DEVID_SAS3324_4:
10386 case MPI26_MFGPAGE_DEVID_SAS3508:
10387 case MPI26_MFGPAGE_DEVID_SAS3508_1:
10388 case MPI26_MFGPAGE_DEVID_SAS3408:
10389 case MPI26_MFGPAGE_DEVID_SAS3516:
10390 case MPI26_MFGPAGE_DEVID_SAS3516_1:
10391 case MPI26_MFGPAGE_DEVID_SAS3416:
10392 case MPI26_MFGPAGE_DEVID_SAS3616:
10393 case MPI26_ATLAS_PCIe_SWITCH_DEVID:
10394 case MPI26_MFGPAGE_DEVID_CFG_SEC_3916:
10395 case MPI26_MFGPAGE_DEVID_HARD_SEC_3916:
10396 case MPI26_MFGPAGE_DEVID_CFG_SEC_3816:
10397 case MPI26_MFGPAGE_DEVID_HARD_SEC_3816:
10398 return MPI26_VERSION;
10399 }
10400 return 0;
10401 }
10402
10403 /**
10404 * _scsih_probe - attach and add scsi host
10405 * @pdev: PCI device struct
10406 * @id: pci device id
10407 *
10408 * Return: 0 success, anything else error.
10409 */
10410 static int
_scsih_probe(struct pci_dev * pdev,const struct pci_device_id * id)10411 _scsih_probe(struct pci_dev *pdev, const struct pci_device_id *id)
10412 {
10413 struct MPT3SAS_ADAPTER *ioc;
10414 struct Scsi_Host *shost = NULL;
10415 int rv;
10416 u16 hba_mpi_version;
10417
10418 /* Determine in which MPI version class this pci device belongs */
10419 hba_mpi_version = _scsih_determine_hba_mpi_version(pdev);
10420 if (hba_mpi_version == 0)
10421 return -ENODEV;
10422
10423 /* Enumerate only SAS 2.0 HBA's if hbas_to_enumerate is one,
10424 * for other generation HBA's return with -ENODEV
10425 */
10426 if ((hbas_to_enumerate == 1) && (hba_mpi_version != MPI2_VERSION))
10427 return -ENODEV;
10428
10429 /* Enumerate only SAS 3.0 HBA's if hbas_to_enumerate is two,
10430 * for other generation HBA's return with -ENODEV
10431 */
10432 if ((hbas_to_enumerate == 2) && (!(hba_mpi_version == MPI25_VERSION
10433 || hba_mpi_version == MPI26_VERSION)))
10434 return -ENODEV;
10435
10436 switch (hba_mpi_version) {
10437 case MPI2_VERSION:
10438 pci_disable_link_state(pdev, PCIE_LINK_STATE_L0S |
10439 PCIE_LINK_STATE_L1 | PCIE_LINK_STATE_CLKPM);
10440 /* Use mpt2sas driver host template for SAS 2.0 HBA's */
10441 shost = scsi_host_alloc(&mpt2sas_driver_template,
10442 sizeof(struct MPT3SAS_ADAPTER));
10443 if (!shost)
10444 return -ENODEV;
10445 ioc = shost_priv(shost);
10446 memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER));
10447 ioc->hba_mpi_version_belonged = hba_mpi_version;
10448 ioc->id = mpt2_ids++;
10449 sprintf(ioc->driver_name, "%s", MPT2SAS_DRIVER_NAME);
10450 switch (pdev->device) {
10451 case MPI2_MFGPAGE_DEVID_SSS6200:
10452 ioc->is_warpdrive = 1;
10453 ioc->hide_ir_msg = 1;
10454 break;
10455 case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP:
10456 case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1:
10457 ioc->is_mcpu_endpoint = 1;
10458 break;
10459 default:
10460 ioc->mfg_pg10_hide_flag = MFG_PAGE10_EXPOSE_ALL_DISKS;
10461 break;
10462 }
10463 break;
10464 case MPI25_VERSION:
10465 case MPI26_VERSION:
10466 /* Use mpt3sas driver host template for SAS 3.0 HBA's */
10467 shost = scsi_host_alloc(&mpt3sas_driver_template,
10468 sizeof(struct MPT3SAS_ADAPTER));
10469 if (!shost)
10470 return -ENODEV;
10471 ioc = shost_priv(shost);
10472 memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER));
10473 ioc->hba_mpi_version_belonged = hba_mpi_version;
10474 ioc->id = mpt3_ids++;
10475 sprintf(ioc->driver_name, "%s", MPT3SAS_DRIVER_NAME);
10476 switch (pdev->device) {
10477 case MPI26_MFGPAGE_DEVID_SAS3508:
10478 case MPI26_MFGPAGE_DEVID_SAS3508_1:
10479 case MPI26_MFGPAGE_DEVID_SAS3408:
10480 case MPI26_MFGPAGE_DEVID_SAS3516:
10481 case MPI26_MFGPAGE_DEVID_SAS3516_1:
10482 case MPI26_MFGPAGE_DEVID_SAS3416:
10483 case MPI26_MFGPAGE_DEVID_SAS3616:
10484 case MPI26_ATLAS_PCIe_SWITCH_DEVID:
10485 ioc->is_gen35_ioc = 1;
10486 break;
10487 case MPI26_MFGPAGE_DEVID_CFG_SEC_3816:
10488 case MPI26_MFGPAGE_DEVID_CFG_SEC_3916:
10489 dev_info(&pdev->dev,
10490 "HBA is in Configurable Secure mode\n");
10491 /* fall through */
10492 case MPI26_MFGPAGE_DEVID_HARD_SEC_3816:
10493 case MPI26_MFGPAGE_DEVID_HARD_SEC_3916:
10494 ioc->is_aero_ioc = ioc->is_gen35_ioc = 1;
10495 break;
10496 default:
10497 ioc->is_gen35_ioc = ioc->is_aero_ioc = 0;
10498 }
10499 if ((ioc->hba_mpi_version_belonged == MPI25_VERSION &&
10500 pdev->revision >= SAS3_PCI_DEVICE_C0_REVISION) ||
10501 (ioc->hba_mpi_version_belonged == MPI26_VERSION)) {
10502 ioc->combined_reply_queue = 1;
10503 if (ioc->is_gen35_ioc)
10504 ioc->combined_reply_index_count =
10505 MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G35;
10506 else
10507 ioc->combined_reply_index_count =
10508 MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G3;
10509 }
10510 break;
10511 default:
10512 return -ENODEV;
10513 }
10514
10515 INIT_LIST_HEAD(&ioc->list);
10516 spin_lock(&gioc_lock);
10517 list_add_tail(&ioc->list, &mpt3sas_ioc_list);
10518 spin_unlock(&gioc_lock);
10519 ioc->shost = shost;
10520 ioc->pdev = pdev;
10521 ioc->scsi_io_cb_idx = scsi_io_cb_idx;
10522 ioc->tm_cb_idx = tm_cb_idx;
10523 ioc->ctl_cb_idx = ctl_cb_idx;
10524 ioc->base_cb_idx = base_cb_idx;
10525 ioc->port_enable_cb_idx = port_enable_cb_idx;
10526 ioc->transport_cb_idx = transport_cb_idx;
10527 ioc->scsih_cb_idx = scsih_cb_idx;
10528 ioc->config_cb_idx = config_cb_idx;
10529 ioc->tm_tr_cb_idx = tm_tr_cb_idx;
10530 ioc->tm_tr_volume_cb_idx = tm_tr_volume_cb_idx;
10531 ioc->tm_sas_control_cb_idx = tm_sas_control_cb_idx;
10532 ioc->logging_level = logging_level;
10533 ioc->schedule_dead_ioc_flush_running_cmds = &_scsih_flush_running_cmds;
10534 /*
10535 * Enable MEMORY MOVE support flag.
10536 */
10537 ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_MEMMOVE;
10538
10539 ioc->enable_sdev_max_qd = enable_sdev_max_qd;
10540
10541 /* misc semaphores and spin locks */
10542 mutex_init(&ioc->reset_in_progress_mutex);
10543 /* initializing pci_access_mutex lock */
10544 mutex_init(&ioc->pci_access_mutex);
10545 spin_lock_init(&ioc->ioc_reset_in_progress_lock);
10546 spin_lock_init(&ioc->scsi_lookup_lock);
10547 spin_lock_init(&ioc->sas_device_lock);
10548 spin_lock_init(&ioc->sas_node_lock);
10549 spin_lock_init(&ioc->fw_event_lock);
10550 spin_lock_init(&ioc->raid_device_lock);
10551 spin_lock_init(&ioc->pcie_device_lock);
10552 spin_lock_init(&ioc->diag_trigger_lock);
10553
10554 INIT_LIST_HEAD(&ioc->sas_device_list);
10555 INIT_LIST_HEAD(&ioc->sas_device_init_list);
10556 INIT_LIST_HEAD(&ioc->sas_expander_list);
10557 INIT_LIST_HEAD(&ioc->enclosure_list);
10558 INIT_LIST_HEAD(&ioc->pcie_device_list);
10559 INIT_LIST_HEAD(&ioc->pcie_device_init_list);
10560 INIT_LIST_HEAD(&ioc->fw_event_list);
10561 INIT_LIST_HEAD(&ioc->raid_device_list);
10562 INIT_LIST_HEAD(&ioc->sas_hba.sas_port_list);
10563 INIT_LIST_HEAD(&ioc->delayed_tr_list);
10564 INIT_LIST_HEAD(&ioc->delayed_sc_list);
10565 INIT_LIST_HEAD(&ioc->delayed_event_ack_list);
10566 INIT_LIST_HEAD(&ioc->delayed_tr_volume_list);
10567 INIT_LIST_HEAD(&ioc->reply_queue_list);
10568
10569 sprintf(ioc->name, "%s_cm%d", ioc->driver_name, ioc->id);
10570
10571 /* init shost parameters */
10572 shost->max_cmd_len = 32;
10573 shost->max_lun = max_lun;
10574 shost->transportt = mpt3sas_transport_template;
10575 shost->unique_id = ioc->id;
10576
10577 if (ioc->is_mcpu_endpoint) {
10578 /* mCPU MPI support 64K max IO */
10579 shost->max_sectors = 128;
10580 ioc_info(ioc, "The max_sectors value is set to %d\n",
10581 shost->max_sectors);
10582 } else {
10583 if (max_sectors != 0xFFFF) {
10584 if (max_sectors < 64) {
10585 shost->max_sectors = 64;
10586 ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767. Assigning value of 64.\n",
10587 max_sectors);
10588 } else if (max_sectors > 32767) {
10589 shost->max_sectors = 32767;
10590 ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767.Assigning default value of 32767.\n",
10591 max_sectors);
10592 } else {
10593 shost->max_sectors = max_sectors & 0xFFFE;
10594 ioc_info(ioc, "The max_sectors value is set to %d\n",
10595 shost->max_sectors);
10596 }
10597 }
10598 }
10599 /* register EEDP capabilities with SCSI layer */
10600 if (prot_mask > 0)
10601 scsi_host_set_prot(shost, prot_mask);
10602 else
10603 scsi_host_set_prot(shost, SHOST_DIF_TYPE1_PROTECTION
10604 | SHOST_DIF_TYPE2_PROTECTION
10605 | SHOST_DIF_TYPE3_PROTECTION);
10606
10607 scsi_host_set_guard(shost, SHOST_DIX_GUARD_CRC);
10608
10609 /* event thread */
10610 snprintf(ioc->firmware_event_name, sizeof(ioc->firmware_event_name),
10611 "fw_event_%s%d", ioc->driver_name, ioc->id);
10612 ioc->firmware_event_thread = alloc_ordered_workqueue(
10613 ioc->firmware_event_name, 0);
10614 if (!ioc->firmware_event_thread) {
10615 ioc_err(ioc, "failure at %s:%d/%s()!\n",
10616 __FILE__, __LINE__, __func__);
10617 rv = -ENODEV;
10618 goto out_thread_fail;
10619 }
10620
10621 ioc->is_driver_loading = 1;
10622 if ((mpt3sas_base_attach(ioc))) {
10623 ioc_err(ioc, "failure at %s:%d/%s()!\n",
10624 __FILE__, __LINE__, __func__);
10625 rv = -ENODEV;
10626 goto out_attach_fail;
10627 }
10628
10629 if (ioc->is_warpdrive) {
10630 if (ioc->mfg_pg10_hide_flag == MFG_PAGE10_EXPOSE_ALL_DISKS)
10631 ioc->hide_drives = 0;
10632 else if (ioc->mfg_pg10_hide_flag == MFG_PAGE10_HIDE_ALL_DISKS)
10633 ioc->hide_drives = 1;
10634 else {
10635 if (mpt3sas_get_num_volumes(ioc))
10636 ioc->hide_drives = 1;
10637 else
10638 ioc->hide_drives = 0;
10639 }
10640 } else
10641 ioc->hide_drives = 0;
10642
10643 rv = scsi_add_host(shost, &pdev->dev);
10644 if (rv) {
10645 ioc_err(ioc, "failure at %s:%d/%s()!\n",
10646 __FILE__, __LINE__, __func__);
10647 goto out_add_shost_fail;
10648 }
10649
10650 scsi_scan_host(shost);
10651 return 0;
10652 out_add_shost_fail:
10653 mpt3sas_base_detach(ioc);
10654 out_attach_fail:
10655 destroy_workqueue(ioc->firmware_event_thread);
10656 out_thread_fail:
10657 spin_lock(&gioc_lock);
10658 list_del(&ioc->list);
10659 spin_unlock(&gioc_lock);
10660 scsi_host_put(shost);
10661 return rv;
10662 }
10663
10664 #ifdef CONFIG_PM
10665 /**
10666 * scsih_suspend - power management suspend main entry point
10667 * @pdev: PCI device struct
10668 * @state: PM state change to (usually PCI_D3)
10669 *
10670 * Return: 0 success, anything else error.
10671 */
10672 static int
scsih_suspend(struct pci_dev * pdev,pm_message_t state)10673 scsih_suspend(struct pci_dev *pdev, pm_message_t state)
10674 {
10675 struct Scsi_Host *shost = pci_get_drvdata(pdev);
10676 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
10677 pci_power_t device_state;
10678
10679 mpt3sas_base_stop_watchdog(ioc);
10680 flush_scheduled_work();
10681 scsi_block_requests(shost);
10682 device_state = pci_choose_state(pdev, state);
10683 ioc_info(ioc, "pdev=0x%p, slot=%s, entering operating state [D%d]\n",
10684 pdev, pci_name(pdev), device_state);
10685
10686 pci_save_state(pdev);
10687 mpt3sas_base_free_resources(ioc);
10688 pci_set_power_state(pdev, device_state);
10689 return 0;
10690 }
10691
10692 /**
10693 * scsih_resume - power management resume main entry point
10694 * @pdev: PCI device struct
10695 *
10696 * Return: 0 success, anything else error.
10697 */
10698 static int
scsih_resume(struct pci_dev * pdev)10699 scsih_resume(struct pci_dev *pdev)
10700 {
10701 struct Scsi_Host *shost = pci_get_drvdata(pdev);
10702 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
10703 pci_power_t device_state = pdev->current_state;
10704 int r;
10705
10706 ioc_info(ioc, "pdev=0x%p, slot=%s, previous operating state [D%d]\n",
10707 pdev, pci_name(pdev), device_state);
10708
10709 pci_set_power_state(pdev, PCI_D0);
10710 pci_enable_wake(pdev, PCI_D0, 0);
10711 pci_restore_state(pdev);
10712 ioc->pdev = pdev;
10713 r = mpt3sas_base_map_resources(ioc);
10714 if (r)
10715 return r;
10716
10717 mpt3sas_base_hard_reset_handler(ioc, SOFT_RESET);
10718 scsi_unblock_requests(shost);
10719 mpt3sas_base_start_watchdog(ioc);
10720 return 0;
10721 }
10722 #endif /* CONFIG_PM */
10723
10724 /**
10725 * scsih_pci_error_detected - Called when a PCI error is detected.
10726 * @pdev: PCI device struct
10727 * @state: PCI channel state
10728 *
10729 * Description: Called when a PCI error is detected.
10730 *
10731 * Return: PCI_ERS_RESULT_NEED_RESET or PCI_ERS_RESULT_DISCONNECT.
10732 */
10733 static pci_ers_result_t
scsih_pci_error_detected(struct pci_dev * pdev,pci_channel_state_t state)10734 scsih_pci_error_detected(struct pci_dev *pdev, pci_channel_state_t state)
10735 {
10736 struct Scsi_Host *shost = pci_get_drvdata(pdev);
10737 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
10738
10739 ioc_info(ioc, "PCI error: detected callback, state(%d)!!\n", state);
10740
10741 switch (state) {
10742 case pci_channel_io_normal:
10743 return PCI_ERS_RESULT_CAN_RECOVER;
10744 case pci_channel_io_frozen:
10745 /* Fatal error, prepare for slot reset */
10746 ioc->pci_error_recovery = 1;
10747 scsi_block_requests(ioc->shost);
10748 mpt3sas_base_stop_watchdog(ioc);
10749 mpt3sas_base_free_resources(ioc);
10750 return PCI_ERS_RESULT_NEED_RESET;
10751 case pci_channel_io_perm_failure:
10752 /* Permanent error, prepare for device removal */
10753 ioc->pci_error_recovery = 1;
10754 mpt3sas_base_stop_watchdog(ioc);
10755 _scsih_flush_running_cmds(ioc);
10756 return PCI_ERS_RESULT_DISCONNECT;
10757 }
10758 return PCI_ERS_RESULT_NEED_RESET;
10759 }
10760
10761 /**
10762 * scsih_pci_slot_reset - Called when PCI slot has been reset.
10763 * @pdev: PCI device struct
10764 *
10765 * Description: This routine is called by the pci error recovery
10766 * code after the PCI slot has been reset, just before we
10767 * should resume normal operations.
10768 */
10769 static pci_ers_result_t
scsih_pci_slot_reset(struct pci_dev * pdev)10770 scsih_pci_slot_reset(struct pci_dev *pdev)
10771 {
10772 struct Scsi_Host *shost = pci_get_drvdata(pdev);
10773 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
10774 int rc;
10775
10776 ioc_info(ioc, "PCI error: slot reset callback!!\n");
10777
10778 ioc->pci_error_recovery = 0;
10779 ioc->pdev = pdev;
10780 pci_restore_state(pdev);
10781 rc = mpt3sas_base_map_resources(ioc);
10782 if (rc)
10783 return PCI_ERS_RESULT_DISCONNECT;
10784
10785 rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
10786
10787 ioc_warn(ioc, "hard reset: %s\n",
10788 (rc == 0) ? "success" : "failed");
10789
10790 if (!rc)
10791 return PCI_ERS_RESULT_RECOVERED;
10792 else
10793 return PCI_ERS_RESULT_DISCONNECT;
10794 }
10795
10796 /**
10797 * scsih_pci_resume() - resume normal ops after PCI reset
10798 * @pdev: pointer to PCI device
10799 *
10800 * Called when the error recovery driver tells us that its
10801 * OK to resume normal operation. Use completion to allow
10802 * halted scsi ops to resume.
10803 */
10804 static void
scsih_pci_resume(struct pci_dev * pdev)10805 scsih_pci_resume(struct pci_dev *pdev)
10806 {
10807 struct Scsi_Host *shost = pci_get_drvdata(pdev);
10808 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
10809
10810 ioc_info(ioc, "PCI error: resume callback!!\n");
10811
10812 mpt3sas_base_start_watchdog(ioc);
10813 scsi_unblock_requests(ioc->shost);
10814 }
10815
10816 /**
10817 * scsih_pci_mmio_enabled - Enable MMIO and dump debug registers
10818 * @pdev: pointer to PCI device
10819 */
10820 static pci_ers_result_t
scsih_pci_mmio_enabled(struct pci_dev * pdev)10821 scsih_pci_mmio_enabled(struct pci_dev *pdev)
10822 {
10823 struct Scsi_Host *shost = pci_get_drvdata(pdev);
10824 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
10825
10826 ioc_info(ioc, "PCI error: mmio enabled callback!!\n");
10827
10828 /* TODO - dump whatever for debugging purposes */
10829
10830 /* This called only if scsih_pci_error_detected returns
10831 * PCI_ERS_RESULT_CAN_RECOVER. Read/write to the device still
10832 * works, no need to reset slot.
10833 */
10834 return PCI_ERS_RESULT_RECOVERED;
10835 }
10836
10837 /**
10838 * scsih__ncq_prio_supp - Check for NCQ command priority support
10839 * @sdev: scsi device struct
10840 *
10841 * This is called when a user indicates they would like to enable
10842 * ncq command priorities. This works only on SATA devices.
10843 */
scsih_ncq_prio_supp(struct scsi_device * sdev)10844 bool scsih_ncq_prio_supp(struct scsi_device *sdev)
10845 {
10846 unsigned char *buf;
10847 bool ncq_prio_supp = false;
10848
10849 if (!scsi_device_supports_vpd(sdev))
10850 return ncq_prio_supp;
10851
10852 buf = kmalloc(SCSI_VPD_PG_LEN, GFP_KERNEL);
10853 if (!buf)
10854 return ncq_prio_supp;
10855
10856 if (!scsi_get_vpd_page(sdev, 0x89, buf, SCSI_VPD_PG_LEN))
10857 ncq_prio_supp = (buf[213] >> 4) & 1;
10858
10859 kfree(buf);
10860 return ncq_prio_supp;
10861 }
10862 /*
10863 * The pci device ids are defined in mpi/mpi2_cnfg.h.
10864 */
10865 static const struct pci_device_id mpt3sas_pci_table[] = {
10866 /* Spitfire ~ 2004 */
10867 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2004,
10868 PCI_ANY_ID, PCI_ANY_ID },
10869 /* Falcon ~ 2008 */
10870 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2008,
10871 PCI_ANY_ID, PCI_ANY_ID },
10872 /* Liberator ~ 2108 */
10873 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_1,
10874 PCI_ANY_ID, PCI_ANY_ID },
10875 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_2,
10876 PCI_ANY_ID, PCI_ANY_ID },
10877 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_3,
10878 PCI_ANY_ID, PCI_ANY_ID },
10879 /* Meteor ~ 2116 */
10880 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_1,
10881 PCI_ANY_ID, PCI_ANY_ID },
10882 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_2,
10883 PCI_ANY_ID, PCI_ANY_ID },
10884 /* Thunderbolt ~ 2208 */
10885 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_1,
10886 PCI_ANY_ID, PCI_ANY_ID },
10887 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_2,
10888 PCI_ANY_ID, PCI_ANY_ID },
10889 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_3,
10890 PCI_ANY_ID, PCI_ANY_ID },
10891 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_4,
10892 PCI_ANY_ID, PCI_ANY_ID },
10893 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_5,
10894 PCI_ANY_ID, PCI_ANY_ID },
10895 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_6,
10896 PCI_ANY_ID, PCI_ANY_ID },
10897 /* Mustang ~ 2308 */
10898 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_1,
10899 PCI_ANY_ID, PCI_ANY_ID },
10900 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_2,
10901 PCI_ANY_ID, PCI_ANY_ID },
10902 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_3,
10903 PCI_ANY_ID, PCI_ANY_ID },
10904 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP,
10905 PCI_ANY_ID, PCI_ANY_ID },
10906 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1,
10907 PCI_ANY_ID, PCI_ANY_ID },
10908 /* SSS6200 */
10909 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SSS6200,
10910 PCI_ANY_ID, PCI_ANY_ID },
10911 /* Fury ~ 3004 and 3008 */
10912 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3004,
10913 PCI_ANY_ID, PCI_ANY_ID },
10914 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3008,
10915 PCI_ANY_ID, PCI_ANY_ID },
10916 /* Invader ~ 3108 */
10917 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_1,
10918 PCI_ANY_ID, PCI_ANY_ID },
10919 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_2,
10920 PCI_ANY_ID, PCI_ANY_ID },
10921 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_5,
10922 PCI_ANY_ID, PCI_ANY_ID },
10923 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_6,
10924 PCI_ANY_ID, PCI_ANY_ID },
10925 /* Cutlass ~ 3216 and 3224 */
10926 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3216,
10927 PCI_ANY_ID, PCI_ANY_ID },
10928 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3224,
10929 PCI_ANY_ID, PCI_ANY_ID },
10930 /* Intruder ~ 3316 and 3324 */
10931 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_1,
10932 PCI_ANY_ID, PCI_ANY_ID },
10933 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_2,
10934 PCI_ANY_ID, PCI_ANY_ID },
10935 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_3,
10936 PCI_ANY_ID, PCI_ANY_ID },
10937 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_4,
10938 PCI_ANY_ID, PCI_ANY_ID },
10939 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_1,
10940 PCI_ANY_ID, PCI_ANY_ID },
10941 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_2,
10942 PCI_ANY_ID, PCI_ANY_ID },
10943 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_3,
10944 PCI_ANY_ID, PCI_ANY_ID },
10945 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_4,
10946 PCI_ANY_ID, PCI_ANY_ID },
10947 /* Ventura, Crusader, Harpoon & Tomcat ~ 3516, 3416, 3508 & 3408*/
10948 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508,
10949 PCI_ANY_ID, PCI_ANY_ID },
10950 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508_1,
10951 PCI_ANY_ID, PCI_ANY_ID },
10952 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3408,
10953 PCI_ANY_ID, PCI_ANY_ID },
10954 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516,
10955 PCI_ANY_ID, PCI_ANY_ID },
10956 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516_1,
10957 PCI_ANY_ID, PCI_ANY_ID },
10958 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3416,
10959 PCI_ANY_ID, PCI_ANY_ID },
10960 /* Mercator ~ 3616*/
10961 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3616,
10962 PCI_ANY_ID, PCI_ANY_ID },
10963
10964 /* Aero SI 0x00E1 Configurable Secure
10965 * 0x00E2 Hard Secure
10966 */
10967 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3916,
10968 PCI_ANY_ID, PCI_ANY_ID },
10969 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3916,
10970 PCI_ANY_ID, PCI_ANY_ID },
10971
10972 /* Atlas PCIe Switch Management Port */
10973 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_ATLAS_PCIe_SWITCH_DEVID,
10974 PCI_ANY_ID, PCI_ANY_ID },
10975
10976 /* Sea SI 0x00E5 Configurable Secure
10977 * 0x00E6 Hard Secure
10978 */
10979 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3816,
10980 PCI_ANY_ID, PCI_ANY_ID },
10981 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3816,
10982 PCI_ANY_ID, PCI_ANY_ID },
10983
10984 {0} /* Terminating entry */
10985 };
10986 MODULE_DEVICE_TABLE(pci, mpt3sas_pci_table);
10987
10988 static struct pci_error_handlers _mpt3sas_err_handler = {
10989 .error_detected = scsih_pci_error_detected,
10990 .mmio_enabled = scsih_pci_mmio_enabled,
10991 .slot_reset = scsih_pci_slot_reset,
10992 .resume = scsih_pci_resume,
10993 };
10994
10995 static struct pci_driver mpt3sas_driver = {
10996 .name = MPT3SAS_DRIVER_NAME,
10997 .id_table = mpt3sas_pci_table,
10998 .probe = _scsih_probe,
10999 .remove = scsih_remove,
11000 .shutdown = scsih_shutdown,
11001 .err_handler = &_mpt3sas_err_handler,
11002 #ifdef CONFIG_PM
11003 .suspend = scsih_suspend,
11004 .resume = scsih_resume,
11005 #endif
11006 };
11007
11008 /**
11009 * scsih_init - main entry point for this driver.
11010 *
11011 * Return: 0 success, anything else error.
11012 */
11013 static int
scsih_init(void)11014 scsih_init(void)
11015 {
11016 mpt2_ids = 0;
11017 mpt3_ids = 0;
11018
11019 mpt3sas_base_initialize_callback_handler();
11020
11021 /* queuecommand callback hander */
11022 scsi_io_cb_idx = mpt3sas_base_register_callback_handler(_scsih_io_done);
11023
11024 /* task management callback handler */
11025 tm_cb_idx = mpt3sas_base_register_callback_handler(_scsih_tm_done);
11026
11027 /* base internal commands callback handler */
11028 base_cb_idx = mpt3sas_base_register_callback_handler(mpt3sas_base_done);
11029 port_enable_cb_idx = mpt3sas_base_register_callback_handler(
11030 mpt3sas_port_enable_done);
11031
11032 /* transport internal commands callback handler */
11033 transport_cb_idx = mpt3sas_base_register_callback_handler(
11034 mpt3sas_transport_done);
11035
11036 /* scsih internal commands callback handler */
11037 scsih_cb_idx = mpt3sas_base_register_callback_handler(_scsih_done);
11038
11039 /* configuration page API internal commands callback handler */
11040 config_cb_idx = mpt3sas_base_register_callback_handler(
11041 mpt3sas_config_done);
11042
11043 /* ctl module callback handler */
11044 ctl_cb_idx = mpt3sas_base_register_callback_handler(mpt3sas_ctl_done);
11045
11046 tm_tr_cb_idx = mpt3sas_base_register_callback_handler(
11047 _scsih_tm_tr_complete);
11048
11049 tm_tr_volume_cb_idx = mpt3sas_base_register_callback_handler(
11050 _scsih_tm_volume_tr_complete);
11051
11052 tm_sas_control_cb_idx = mpt3sas_base_register_callback_handler(
11053 _scsih_sas_control_complete);
11054
11055 return 0;
11056 }
11057
11058 /**
11059 * scsih_exit - exit point for this driver (when it is a module).
11060 *
11061 * Return: 0 success, anything else error.
11062 */
11063 static void
scsih_exit(void)11064 scsih_exit(void)
11065 {
11066
11067 mpt3sas_base_release_callback_handler(scsi_io_cb_idx);
11068 mpt3sas_base_release_callback_handler(tm_cb_idx);
11069 mpt3sas_base_release_callback_handler(base_cb_idx);
11070 mpt3sas_base_release_callback_handler(port_enable_cb_idx);
11071 mpt3sas_base_release_callback_handler(transport_cb_idx);
11072 mpt3sas_base_release_callback_handler(scsih_cb_idx);
11073 mpt3sas_base_release_callback_handler(config_cb_idx);
11074 mpt3sas_base_release_callback_handler(ctl_cb_idx);
11075
11076 mpt3sas_base_release_callback_handler(tm_tr_cb_idx);
11077 mpt3sas_base_release_callback_handler(tm_tr_volume_cb_idx);
11078 mpt3sas_base_release_callback_handler(tm_sas_control_cb_idx);
11079
11080 /* raid transport support */
11081 if (hbas_to_enumerate != 1)
11082 raid_class_release(mpt3sas_raid_template);
11083 if (hbas_to_enumerate != 2)
11084 raid_class_release(mpt2sas_raid_template);
11085 sas_release_transport(mpt3sas_transport_template);
11086 }
11087
11088 /**
11089 * _mpt3sas_init - main entry point for this driver.
11090 *
11091 * Return: 0 success, anything else error.
11092 */
11093 static int __init
_mpt3sas_init(void)11094 _mpt3sas_init(void)
11095 {
11096 int error;
11097
11098 pr_info("%s version %s loaded\n", MPT3SAS_DRIVER_NAME,
11099 MPT3SAS_DRIVER_VERSION);
11100
11101 mpt3sas_transport_template =
11102 sas_attach_transport(&mpt3sas_transport_functions);
11103 if (!mpt3sas_transport_template)
11104 return -ENODEV;
11105
11106 /* No need attach mpt3sas raid functions template
11107 * if hbas_to_enumarate value is one.
11108 */
11109 if (hbas_to_enumerate != 1) {
11110 mpt3sas_raid_template =
11111 raid_class_attach(&mpt3sas_raid_functions);
11112 if (!mpt3sas_raid_template) {
11113 sas_release_transport(mpt3sas_transport_template);
11114 return -ENODEV;
11115 }
11116 }
11117
11118 /* No need to attach mpt2sas raid functions template
11119 * if hbas_to_enumarate value is two
11120 */
11121 if (hbas_to_enumerate != 2) {
11122 mpt2sas_raid_template =
11123 raid_class_attach(&mpt2sas_raid_functions);
11124 if (!mpt2sas_raid_template) {
11125 sas_release_transport(mpt3sas_transport_template);
11126 return -ENODEV;
11127 }
11128 }
11129
11130 error = scsih_init();
11131 if (error) {
11132 scsih_exit();
11133 return error;
11134 }
11135
11136 mpt3sas_ctl_init(hbas_to_enumerate);
11137
11138 error = pci_register_driver(&mpt3sas_driver);
11139 if (error)
11140 scsih_exit();
11141
11142 return error;
11143 }
11144
11145 /**
11146 * _mpt3sas_exit - exit point for this driver (when it is a module).
11147 *
11148 */
11149 static void __exit
_mpt3sas_exit(void)11150 _mpt3sas_exit(void)
11151 {
11152 pr_info("mpt3sas version %s unloading\n",
11153 MPT3SAS_DRIVER_VERSION);
11154
11155 mpt3sas_ctl_exit(hbas_to_enumerate);
11156
11157 pci_unregister_driver(&mpt3sas_driver);
11158
11159 scsih_exit();
11160 }
11161
11162 module_init(_mpt3sas_init);
11163 module_exit(_mpt3sas_exit);
11164