Commit 9fa82023 authored by Kevin Barnett's avatar Kevin Barnett Committed by Martin K. Petersen

scsi: smartpqi: Update suspend/resume and shutdown

For suspend/resume and shutdown prevent: Controller events, any new I/O
requests, controller requests, REGNEWD, and reset operations.

Wait for any pending completions from the controller to complete to avoid
controller NMI events.

Link: https://lore.kernel.org/r/161549380398.25025.12266769502766103580.stgit@brunhildaReviewed-by: default avatarScott Teel <scott.teel@microchip.com>
Reviewed-by: default avatarScott Benesh <scott.benesh@microchip.com>
Signed-off-by: default avatarKevin Barnett <kevin.barnett@microchip.com>
Signed-off-by: default avatarDon Brace <don.brace@microchip.com>
Signed-off-by: default avatarMartin K. Petersen <martin.petersen@oracle.com>
parent 37f33181
...@@ -1295,6 +1295,7 @@ struct pqi_ctrl_info { ...@@ -1295,6 +1295,7 @@ struct pqi_ctrl_info {
struct mutex ofa_mutex; /* serialize ofa */ struct mutex ofa_mutex; /* serialize ofa */
bool controller_online; bool controller_online;
bool block_requests; bool block_requests;
bool scan_blocked;
bool in_ofa; bool in_ofa;
bool in_shutdown; bool in_shutdown;
u8 inbound_spanning_supported : 1; u8 inbound_spanning_supported : 1;
...@@ -1624,16 +1625,6 @@ struct bmic_diag_options { ...@@ -1624,16 +1625,6 @@ struct bmic_diag_options {
#pragma pack() #pragma pack()
static inline void pqi_ctrl_busy(struct pqi_ctrl_info *ctrl_info)
{
atomic_inc(&ctrl_info->num_busy_threads);
}
static inline void pqi_ctrl_unbusy(struct pqi_ctrl_info *ctrl_info)
{
atomic_dec(&ctrl_info->num_busy_threads);
}
static inline struct pqi_ctrl_info *shost_to_hba(struct Scsi_Host *shost) static inline struct pqi_ctrl_info *shost_to_hba(struct Scsi_Host *shost)
{ {
void *hostdata = shost_priv(shost); void *hostdata = shost_priv(shost);
......
...@@ -54,7 +54,6 @@ MODULE_LICENSE("GPL"); ...@@ -54,7 +54,6 @@ MODULE_LICENSE("GPL");
static void pqi_take_ctrl_offline(struct pqi_ctrl_info *ctrl_info); static void pqi_take_ctrl_offline(struct pqi_ctrl_info *ctrl_info);
static void pqi_ctrl_offline_worker(struct work_struct *work); static void pqi_ctrl_offline_worker(struct work_struct *work);
static void pqi_retry_raid_bypass_requests(struct pqi_ctrl_info *ctrl_info);
static int pqi_scan_scsi_devices(struct pqi_ctrl_info *ctrl_info); static int pqi_scan_scsi_devices(struct pqi_ctrl_info *ctrl_info);
static void pqi_scan_start(struct Scsi_Host *shost); static void pqi_scan_start(struct Scsi_Host *shost);
static void pqi_start_io(struct pqi_ctrl_info *ctrl_info, static void pqi_start_io(struct pqi_ctrl_info *ctrl_info,
...@@ -245,6 +244,23 @@ static inline void pqi_save_ctrl_mode(struct pqi_ctrl_info *ctrl_info, ...@@ -245,6 +244,23 @@ static inline void pqi_save_ctrl_mode(struct pqi_ctrl_info *ctrl_info,
sis_write_driver_scratch(ctrl_info, mode); sis_write_driver_scratch(ctrl_info, mode);
} }
static inline void pqi_ctrl_block_scan(struct pqi_ctrl_info *ctrl_info)
{
ctrl_info->scan_blocked = true;
mutex_lock(&ctrl_info->scan_mutex);
}
static inline void pqi_ctrl_unblock_scan(struct pqi_ctrl_info *ctrl_info)
{
ctrl_info->scan_blocked = false;
mutex_unlock(&ctrl_info->scan_mutex);
}
static inline bool pqi_ctrl_scan_blocked(struct pqi_ctrl_info *ctrl_info)
{
return ctrl_info->scan_blocked;
}
static inline void pqi_ctrl_block_device_reset(struct pqi_ctrl_info *ctrl_info) static inline void pqi_ctrl_block_device_reset(struct pqi_ctrl_info *ctrl_info)
{ {
mutex_lock(&ctrl_info->lun_reset_mutex); mutex_lock(&ctrl_info->lun_reset_mutex);
...@@ -255,6 +271,41 @@ static inline void pqi_ctrl_unblock_device_reset(struct pqi_ctrl_info *ctrl_info ...@@ -255,6 +271,41 @@ static inline void pqi_ctrl_unblock_device_reset(struct pqi_ctrl_info *ctrl_info
mutex_unlock(&ctrl_info->lun_reset_mutex); mutex_unlock(&ctrl_info->lun_reset_mutex);
} }
static inline void pqi_scsi_block_requests(struct pqi_ctrl_info *ctrl_info)
{
struct Scsi_Host *shost;
unsigned int num_loops;
int msecs_sleep;
shost = ctrl_info->scsi_host;
scsi_block_requests(shost);
num_loops = 0;
msecs_sleep = 20;
while (scsi_host_busy(shost)) {
num_loops++;
if (num_loops == 10)
msecs_sleep = 500;
msleep(msecs_sleep);
}
}
static inline void pqi_scsi_unblock_requests(struct pqi_ctrl_info *ctrl_info)
{
scsi_unblock_requests(ctrl_info->scsi_host);
}
static inline void pqi_ctrl_busy(struct pqi_ctrl_info *ctrl_info)
{
atomic_inc(&ctrl_info->num_busy_threads);
}
static inline void pqi_ctrl_unbusy(struct pqi_ctrl_info *ctrl_info)
{
atomic_dec(&ctrl_info->num_busy_threads);
}
static inline bool pqi_ctrl_blocked(struct pqi_ctrl_info *ctrl_info) static inline bool pqi_ctrl_blocked(struct pqi_ctrl_info *ctrl_info)
{ {
return ctrl_info->block_requests; return ctrl_info->block_requests;
...@@ -263,15 +314,12 @@ static inline bool pqi_ctrl_blocked(struct pqi_ctrl_info *ctrl_info) ...@@ -263,15 +314,12 @@ static inline bool pqi_ctrl_blocked(struct pqi_ctrl_info *ctrl_info)
static inline void pqi_ctrl_block_requests(struct pqi_ctrl_info *ctrl_info) static inline void pqi_ctrl_block_requests(struct pqi_ctrl_info *ctrl_info)
{ {
ctrl_info->block_requests = true; ctrl_info->block_requests = true;
scsi_block_requests(ctrl_info->scsi_host);
} }
static inline void pqi_ctrl_unblock_requests(struct pqi_ctrl_info *ctrl_info) static inline void pqi_ctrl_unblock_requests(struct pqi_ctrl_info *ctrl_info)
{ {
ctrl_info->block_requests = false; ctrl_info->block_requests = false;
wake_up_all(&ctrl_info->block_requests_wait); wake_up_all(&ctrl_info->block_requests_wait);
pqi_retry_raid_bypass_requests(ctrl_info);
scsi_unblock_requests(ctrl_info->scsi_host);
} }
static void pqi_wait_if_ctrl_blocked(struct pqi_ctrl_info *ctrl_info) static void pqi_wait_if_ctrl_blocked(struct pqi_ctrl_info *ctrl_info)
...@@ -5999,18 +6047,6 @@ static int pqi_ctrl_wait_for_pending_io(struct pqi_ctrl_info *ctrl_info, ...@@ -5999,18 +6047,6 @@ static int pqi_ctrl_wait_for_pending_io(struct pqi_ctrl_info *ctrl_info,
return 0; return 0;
} }
static int pqi_ctrl_wait_for_pending_sync_cmds(struct pqi_ctrl_info *ctrl_info)
{
while (atomic_read(&ctrl_info->sync_cmds_outstanding)) {
pqi_check_ctrl_health(ctrl_info);
if (pqi_ctrl_offline(ctrl_info))
return -ENXIO;
usleep_range(1000, 2000);
}
return 0;
}
static void pqi_lun_reset_complete(struct pqi_io_request *io_request, static void pqi_lun_reset_complete(struct pqi_io_request *io_request,
void *context) void *context)
{ {
...@@ -8208,7 +8244,6 @@ static struct pqi_ctrl_info *pqi_alloc_ctrl_info(int numa_node) ...@@ -8208,7 +8244,6 @@ static struct pqi_ctrl_info *pqi_alloc_ctrl_info(int numa_node)
INIT_WORK(&ctrl_info->event_work, pqi_event_worker); INIT_WORK(&ctrl_info->event_work, pqi_event_worker);
atomic_set(&ctrl_info->num_interrupts, 0); atomic_set(&ctrl_info->num_interrupts, 0);
atomic_set(&ctrl_info->sync_cmds_outstanding, 0);
INIT_DELAYED_WORK(&ctrl_info->rescan_work, pqi_rescan_worker); INIT_DELAYED_WORK(&ctrl_info->rescan_work, pqi_rescan_worker);
INIT_DELAYED_WORK(&ctrl_info->update_time_work, pqi_update_time_worker); INIT_DELAYED_WORK(&ctrl_info->update_time_work, pqi_update_time_worker);
...@@ -8683,24 +8718,12 @@ static void pqi_shutdown(struct pci_dev *pci_dev) ...@@ -8683,24 +8718,12 @@ static void pqi_shutdown(struct pci_dev *pci_dev)
return; return;
} }
pqi_disable_events(ctrl_info);
pqi_wait_until_ofa_finished(ctrl_info); pqi_wait_until_ofa_finished(ctrl_info);
pqi_cancel_update_time_worker(ctrl_info);
pqi_cancel_rescan_worker(ctrl_info);
pqi_cancel_event_worker(ctrl_info);
pqi_ctrl_shutdown_start(ctrl_info);
pqi_ctrl_wait_until_quiesced(ctrl_info);
rc = pqi_ctrl_wait_for_pending_io(ctrl_info, NO_TIMEOUT);
if (rc) {
dev_err(&pci_dev->dev,
"wait for pending I/O failed\n");
return;
}
pqi_scsi_block_requests(ctrl_info);
pqi_ctrl_block_device_reset(ctrl_info); pqi_ctrl_block_device_reset(ctrl_info);
pqi_wait_until_lun_reset_finished(ctrl_info); pqi_ctrl_block_requests(ctrl_info);
pqi_ctrl_wait_until_quiesced(ctrl_info);
/* /*
* Write all data in the controller's battery-backed cache to * Write all data in the controller's battery-backed cache to
...@@ -8711,15 +8734,6 @@ static void pqi_shutdown(struct pci_dev *pci_dev) ...@@ -8711,15 +8734,6 @@ static void pqi_shutdown(struct pci_dev *pci_dev)
dev_err(&pci_dev->dev, dev_err(&pci_dev->dev,
"unable to flush controller cache\n"); "unable to flush controller cache\n");
pqi_ctrl_block_requests(ctrl_info);
rc = pqi_ctrl_wait_for_pending_sync_cmds(ctrl_info);
if (rc) {
dev_err(&pci_dev->dev,
"wait for pending sync cmds failed\n");
return;
}
pqi_crash_if_pending_command(ctrl_info); pqi_crash_if_pending_command(ctrl_info);
pqi_reset(ctrl_info); pqi_reset(ctrl_info);
} }
...@@ -8754,19 +8768,18 @@ static __maybe_unused int pqi_suspend(struct pci_dev *pci_dev, pm_message_t stat ...@@ -8754,19 +8768,18 @@ static __maybe_unused int pqi_suspend(struct pci_dev *pci_dev, pm_message_t stat
ctrl_info = pci_get_drvdata(pci_dev); ctrl_info = pci_get_drvdata(pci_dev);
pqi_disable_events(ctrl_info);
pqi_cancel_update_time_worker(ctrl_info);
pqi_cancel_rescan_worker(ctrl_info);
pqi_wait_until_scan_finished(ctrl_info);
pqi_wait_until_lun_reset_finished(ctrl_info);
pqi_wait_until_ofa_finished(ctrl_info); pqi_wait_until_ofa_finished(ctrl_info);
pqi_flush_cache(ctrl_info, SUSPEND);
pqi_ctrl_block_scan(ctrl_info);
pqi_scsi_block_requests(ctrl_info);
pqi_ctrl_block_device_reset(ctrl_info);
pqi_ctrl_block_requests(ctrl_info); pqi_ctrl_block_requests(ctrl_info);
pqi_ctrl_wait_until_quiesced(ctrl_info); pqi_ctrl_wait_until_quiesced(ctrl_info);
pqi_wait_until_inbound_queues_empty(ctrl_info); pqi_flush_cache(ctrl_info, SUSPEND);
pqi_ctrl_wait_for_pending_io(ctrl_info, NO_TIMEOUT);
pqi_stop_heartbeat_timer(ctrl_info); pqi_stop_heartbeat_timer(ctrl_info);
pqi_crash_if_pending_command(ctrl_info);
if (state.event == PM_EVENT_FREEZE) if (state.event == PM_EVENT_FREEZE)
return 0; return 0;
...@@ -8799,8 +8812,8 @@ static __maybe_unused int pqi_resume(struct pci_dev *pci_dev) ...@@ -8799,8 +8812,8 @@ static __maybe_unused int pqi_resume(struct pci_dev *pci_dev)
pci_dev->irq, rc); pci_dev->irq, rc);
return rc; return rc;
} }
pqi_start_heartbeat_timer(ctrl_info);
pqi_ctrl_unblock_requests(ctrl_info); pqi_ctrl_unblock_requests(ctrl_info);
pqi_scsi_unblock_requests(ctrl_info);
return 0; return 0;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment