scsi: smartpqi: Update suspend/resume and shutdown
authorKevin Barnett <kevin.barnett@microchip.com>
Thu, 11 Mar 2021 20:16:44 +0000 (14:16 -0600)
committerMartin K. Petersen <martin.petersen@oracle.com>
Tue, 6 Apr 2021 03:02:31 +0000 (23:02 -0400)
For suspend/resume and shutdown prevent: Controller events, any new I/O
requests, controller requests, REGNEWD, and reset operations.

Wait for any pending completions from the controller to complete to avoid
controller NMI events.

Link: https://lore.kernel.org/r/161549380398.25025.12266769502766103580.stgit@brunhilda
Reviewed-by: Scott Teel <scott.teel@microchip.com>
Reviewed-by: Scott Benesh <scott.benesh@microchip.com>
Signed-off-by: Kevin Barnett <kevin.barnett@microchip.com>
Signed-off-by: Don Brace <don.brace@microchip.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
drivers/scsi/smartpqi/smartpqi.h
drivers/scsi/smartpqi/smartpqi_init.c

index ba7d26364b844a8f07e1dbd015e42b556b6a9834..fa1ebeea777d7d5560edd6e1920dee78eb41dd01 100644 (file)
@@ -1295,6 +1295,7 @@ struct pqi_ctrl_info {
        struct mutex    ofa_mutex; /* serialize ofa */
        bool            controller_online;
        bool            block_requests;
+       bool            scan_blocked;
        bool            in_ofa;
        bool            in_shutdown;
        u8              inbound_spanning_supported : 1;
@@ -1624,16 +1625,6 @@ struct bmic_diag_options {
 
 #pragma pack()
 
-static inline void pqi_ctrl_busy(struct pqi_ctrl_info *ctrl_info)
-{
-       atomic_inc(&ctrl_info->num_busy_threads);
-}
-
-static inline void pqi_ctrl_unbusy(struct pqi_ctrl_info *ctrl_info)
-{
-       atomic_dec(&ctrl_info->num_busy_threads);
-}
-
 static inline struct pqi_ctrl_info *shost_to_hba(struct Scsi_Host *shost)
 {
        void *hostdata = shost_priv(shost);
index cc2b29a67ba02217e7f6fdad73ea7695d60ed0af..4f7a38b9635618b9c8771dcb96a6341b54e96045 100644 (file)
@@ -54,7 +54,6 @@ MODULE_LICENSE("GPL");
 
 static void pqi_take_ctrl_offline(struct pqi_ctrl_info *ctrl_info);
 static void pqi_ctrl_offline_worker(struct work_struct *work);
-static void pqi_retry_raid_bypass_requests(struct pqi_ctrl_info *ctrl_info);
 static int pqi_scan_scsi_devices(struct pqi_ctrl_info *ctrl_info);
 static void pqi_scan_start(struct Scsi_Host *shost);
 static void pqi_start_io(struct pqi_ctrl_info *ctrl_info,
@@ -245,6 +244,23 @@ static inline void pqi_save_ctrl_mode(struct pqi_ctrl_info *ctrl_info,
        sis_write_driver_scratch(ctrl_info, mode);
 }
 
+static inline void pqi_ctrl_block_scan(struct pqi_ctrl_info *ctrl_info)
+{
+       ctrl_info->scan_blocked = true;
+       mutex_lock(&ctrl_info->scan_mutex);
+}
+
+static inline void pqi_ctrl_unblock_scan(struct pqi_ctrl_info *ctrl_info)
+{
+       ctrl_info->scan_blocked = false;
+       mutex_unlock(&ctrl_info->scan_mutex);
+}
+
+static inline bool pqi_ctrl_scan_blocked(struct pqi_ctrl_info *ctrl_info)
+{
+       return ctrl_info->scan_blocked;
+}
+
 static inline void pqi_ctrl_block_device_reset(struct pqi_ctrl_info *ctrl_info)
 {
        mutex_lock(&ctrl_info->lun_reset_mutex);
@@ -255,6 +271,41 @@ static inline void pqi_ctrl_unblock_device_reset(struct pqi_ctrl_info *ctrl_info
        mutex_unlock(&ctrl_info->lun_reset_mutex);
 }
 
+static inline void pqi_scsi_block_requests(struct pqi_ctrl_info *ctrl_info)
+{
+       struct Scsi_Host *shost;
+       unsigned int num_loops;
+       int msecs_sleep;
+
+       shost = ctrl_info->scsi_host;
+
+       scsi_block_requests(shost);
+
+       num_loops = 0;
+       msecs_sleep = 20;
+       while (scsi_host_busy(shost)) {
+               num_loops++;
+               if (num_loops == 10)
+                       msecs_sleep = 500;
+               msleep(msecs_sleep);
+       }
+}
+
+static inline void pqi_scsi_unblock_requests(struct pqi_ctrl_info *ctrl_info)
+{
+       scsi_unblock_requests(ctrl_info->scsi_host);
+}
+
+static inline void pqi_ctrl_busy(struct pqi_ctrl_info *ctrl_info)
+{
+       atomic_inc(&ctrl_info->num_busy_threads);
+}
+
+static inline void pqi_ctrl_unbusy(struct pqi_ctrl_info *ctrl_info)
+{
+       atomic_dec(&ctrl_info->num_busy_threads);
+}
+
 static inline bool pqi_ctrl_blocked(struct pqi_ctrl_info *ctrl_info)
 {
        return ctrl_info->block_requests;
@@ -263,15 +314,12 @@ static inline bool pqi_ctrl_blocked(struct pqi_ctrl_info *ctrl_info)
 static inline void pqi_ctrl_block_requests(struct pqi_ctrl_info *ctrl_info)
 {
        ctrl_info->block_requests = true;
-       scsi_block_requests(ctrl_info->scsi_host);
 }
 
 static inline void pqi_ctrl_unblock_requests(struct pqi_ctrl_info *ctrl_info)
 {
        ctrl_info->block_requests = false;
        wake_up_all(&ctrl_info->block_requests_wait);
-       pqi_retry_raid_bypass_requests(ctrl_info);
-       scsi_unblock_requests(ctrl_info->scsi_host);
 }
 
 static void pqi_wait_if_ctrl_blocked(struct pqi_ctrl_info *ctrl_info)
@@ -5999,18 +6047,6 @@ static int pqi_ctrl_wait_for_pending_io(struct pqi_ctrl_info *ctrl_info,
        return 0;
 }
 
-static int pqi_ctrl_wait_for_pending_sync_cmds(struct pqi_ctrl_info *ctrl_info)
-{
-       while (atomic_read(&ctrl_info->sync_cmds_outstanding)) {
-               pqi_check_ctrl_health(ctrl_info);
-               if (pqi_ctrl_offline(ctrl_info))
-                       return -ENXIO;
-               usleep_range(1000, 2000);
-       }
-
-       return 0;
-}
-
 static void pqi_lun_reset_complete(struct pqi_io_request *io_request,
        void *context)
 {
@@ -8208,7 +8244,6 @@ static struct pqi_ctrl_info *pqi_alloc_ctrl_info(int numa_node)
 
        INIT_WORK(&ctrl_info->event_work, pqi_event_worker);
        atomic_set(&ctrl_info->num_interrupts, 0);
-       atomic_set(&ctrl_info->sync_cmds_outstanding, 0);
 
        INIT_DELAYED_WORK(&ctrl_info->rescan_work, pqi_rescan_worker);
        INIT_DELAYED_WORK(&ctrl_info->update_time_work, pqi_update_time_worker);
@@ -8683,24 +8718,12 @@ static void pqi_shutdown(struct pci_dev *pci_dev)
                return;
        }
 
-       pqi_disable_events(ctrl_info);
        pqi_wait_until_ofa_finished(ctrl_info);
-       pqi_cancel_update_time_worker(ctrl_info);
-       pqi_cancel_rescan_worker(ctrl_info);
-       pqi_cancel_event_worker(ctrl_info);
-
-       pqi_ctrl_shutdown_start(ctrl_info);
-       pqi_ctrl_wait_until_quiesced(ctrl_info);
-
-       rc = pqi_ctrl_wait_for_pending_io(ctrl_info, NO_TIMEOUT);
-       if (rc) {
-               dev_err(&pci_dev->dev,
-                       "wait for pending I/O failed\n");
-               return;
-       }
 
+       pqi_scsi_block_requests(ctrl_info);
        pqi_ctrl_block_device_reset(ctrl_info);
-       pqi_wait_until_lun_reset_finished(ctrl_info);
+       pqi_ctrl_block_requests(ctrl_info);
+       pqi_ctrl_wait_until_quiesced(ctrl_info);
 
        /*
         * Write all data in the controller's battery-backed cache to
@@ -8711,15 +8734,6 @@ static void pqi_shutdown(struct pci_dev *pci_dev)
                dev_err(&pci_dev->dev,
                        "unable to flush controller cache\n");
 
-       pqi_ctrl_block_requests(ctrl_info);
-
-       rc = pqi_ctrl_wait_for_pending_sync_cmds(ctrl_info);
-       if (rc) {
-               dev_err(&pci_dev->dev,
-                       "wait for pending sync cmds failed\n");
-               return;
-       }
-
        pqi_crash_if_pending_command(ctrl_info);
        pqi_reset(ctrl_info);
 }
@@ -8754,19 +8768,18 @@ static __maybe_unused int pqi_suspend(struct pci_dev *pci_dev, pm_message_t stat
 
        ctrl_info = pci_get_drvdata(pci_dev);
 
-       pqi_disable_events(ctrl_info);
-       pqi_cancel_update_time_worker(ctrl_info);
-       pqi_cancel_rescan_worker(ctrl_info);
-       pqi_wait_until_scan_finished(ctrl_info);
-       pqi_wait_until_lun_reset_finished(ctrl_info);
        pqi_wait_until_ofa_finished(ctrl_info);
-       pqi_flush_cache(ctrl_info, SUSPEND);
+
+       pqi_ctrl_block_scan(ctrl_info);
+       pqi_scsi_block_requests(ctrl_info);
+       pqi_ctrl_block_device_reset(ctrl_info);
        pqi_ctrl_block_requests(ctrl_info);
        pqi_ctrl_wait_until_quiesced(ctrl_info);
-       pqi_wait_until_inbound_queues_empty(ctrl_info);
-       pqi_ctrl_wait_for_pending_io(ctrl_info, NO_TIMEOUT);
+       pqi_flush_cache(ctrl_info, SUSPEND);
        pqi_stop_heartbeat_timer(ctrl_info);
 
+       pqi_crash_if_pending_command(ctrl_info);
+
        if (state.event == PM_EVENT_FREEZE)
                return 0;
 
@@ -8799,8 +8812,8 @@ static __maybe_unused int pqi_resume(struct pci_dev *pci_dev)
                                pci_dev->irq, rc);
                        return rc;
                }
-               pqi_start_heartbeat_timer(ctrl_info);
                pqi_ctrl_unblock_requests(ctrl_info);
+               pqi_scsi_unblock_requests(ctrl_info);
                return 0;
        }