Commit edc0596c authored by Bart Van Assche's avatar Bart Van Assche Committed by Martin K. Petersen

scsi: ufs: core: Stop clearing UNIT ATTENTIONS

Commit aa53f580 ("scsi: ufs: Minor adjustments to error handling")
introduced a ufshcd_clear_ua_wluns() call in
ufshcd_err_handling_unprepare(). As explained in detail by Adrian Hunter,
this can trigger a deadlock. Avoid that deadlock by removing the code that
clears the unit attention. This is safe because the only software that
relies on clearing unit attentions is the Android Trusty software and
because support for handling unit attentions has been added in the Trusty
software.

See also https://lore.kernel.org/linux-scsi/20210930124224.114031-2-adrian.hunter@intel.com/

Note that "scsi: ufs: Retry START_STOP on UNIT_ATTENTION" is a prerequisite
for this commit.

Link: https://lore.kernel.org/r/20211001182015.1347587-3-jaegeuk@kernel.org
Fixes: aa53f580 ("scsi: ufs: Minor adjustments to error handling")
Cc: Adrian Hunter <adrian.hunter@intel.com>
Signed-off-by: default avatarBart Van Assche <bvanassche@google.com>
Signed-off-by: default avatarJaegeuk Kim <jaegeuk@kernel.org>
Signed-off-by: default avatarMartin K. Petersen <martin.petersen@oracle.com>
parent af21c3fd
......@@ -224,7 +224,6 @@ static int ufshcd_reset_and_restore(struct ufs_hba *hba);
static int ufshcd_eh_host_reset_handler(struct scsi_cmnd *cmd);
static int ufshcd_clear_tm_cmd(struct ufs_hba *hba, int tag);
static void ufshcd_hba_exit(struct ufs_hba *hba);
static int ufshcd_clear_ua_wluns(struct ufs_hba *hba);
static int ufshcd_probe_hba(struct ufs_hba *hba, bool init_dev_params);
static int ufshcd_setup_clocks(struct ufs_hba *hba, bool on);
static inline void ufshcd_add_delay_before_dme_cmd(struct ufs_hba *hba);
......@@ -4108,8 +4107,6 @@ int ufshcd_link_recovery(struct ufs_hba *hba)
if (ret)
dev_err(hba->dev, "%s: link recovery failed, err %d",
__func__, ret);
else
ufshcd_clear_ua_wluns(hba);
return ret;
}
......@@ -5994,7 +5991,6 @@ static void ufshcd_err_handling_unprepare(struct ufs_hba *hba)
ufshcd_release(hba);
if (ufshcd_is_clkscaling_supported(hba))
ufshcd_clk_scaling_suspend(hba, false);
ufshcd_clear_ua_wluns(hba);
ufshcd_rpm_put(hba);
}
......@@ -7952,8 +7948,6 @@ static int ufshcd_add_lus(struct ufs_hba *hba)
if (ret)
goto out;
ufshcd_clear_ua_wluns(hba);
/* Initialize devfreq after UFS device is detected */
if (ufshcd_is_clkscaling_supported(hba)) {
memcpy(&hba->clk_scaling.saved_pwr_info.info,
......@@ -7979,116 +7973,6 @@ static int ufshcd_add_lus(struct ufs_hba *hba)
return ret;
}
static void ufshcd_request_sense_done(struct request *rq, blk_status_t error)
{
if (error != BLK_STS_OK)
pr_err("%s: REQUEST SENSE failed (%d)\n", __func__, error);
kfree(rq->end_io_data);
blk_put_request(rq);
}
static int
ufshcd_request_sense_async(struct ufs_hba *hba, struct scsi_device *sdev)
{
/*
* Some UFS devices clear unit attention condition only if the sense
* size used (UFS_SENSE_SIZE in this case) is non-zero.
*/
static const u8 cmd[6] = {REQUEST_SENSE, 0, 0, 0, UFS_SENSE_SIZE, 0};
struct scsi_request *rq;
struct request *req;
char *buffer;
int ret;
buffer = kzalloc(UFS_SENSE_SIZE, GFP_KERNEL);
if (!buffer)
return -ENOMEM;
req = blk_get_request(sdev->request_queue, REQ_OP_DRV_IN,
/*flags=*/BLK_MQ_REQ_PM);
if (IS_ERR(req)) {
ret = PTR_ERR(req);
goto out_free;
}
ret = blk_rq_map_kern(sdev->request_queue, req,
buffer, UFS_SENSE_SIZE, GFP_NOIO);
if (ret)
goto out_put;
rq = scsi_req(req);
rq->cmd_len = ARRAY_SIZE(cmd);
memcpy(rq->cmd, cmd, rq->cmd_len);
rq->retries = 3;
req->timeout = 1 * HZ;
req->rq_flags |= RQF_PM | RQF_QUIET;
req->end_io_data = buffer;
blk_execute_rq_nowait(/*bd_disk=*/NULL, req, /*at_head=*/true,
ufshcd_request_sense_done);
return 0;
out_put:
blk_put_request(req);
out_free:
kfree(buffer);
return ret;
}
static int ufshcd_clear_ua_wlun(struct ufs_hba *hba, u8 wlun)
{
struct scsi_device *sdp;
unsigned long flags;
int ret = 0;
spin_lock_irqsave(hba->host->host_lock, flags);
if (wlun == UFS_UPIU_UFS_DEVICE_WLUN)
sdp = hba->sdev_ufs_device;
else if (wlun == UFS_UPIU_RPMB_WLUN)
sdp = hba->sdev_rpmb;
else
BUG();
if (sdp) {
ret = scsi_device_get(sdp);
if (!ret && !scsi_device_online(sdp)) {
ret = -ENODEV;
scsi_device_put(sdp);
}
} else {
ret = -ENODEV;
}
spin_unlock_irqrestore(hba->host->host_lock, flags);
if (ret)
goto out_err;
ret = ufshcd_request_sense_async(hba, sdp);
scsi_device_put(sdp);
out_err:
if (ret)
dev_err(hba->dev, "%s: UAC clear LU=%x ret = %d\n",
__func__, wlun, ret);
return ret;
}
static int ufshcd_clear_ua_wluns(struct ufs_hba *hba)
{
int ret = 0;
if (!hba->wlun_dev_clr_ua)
goto out;
ret = ufshcd_clear_ua_wlun(hba, UFS_UPIU_UFS_DEVICE_WLUN);
if (!ret)
ret = ufshcd_clear_ua_wlun(hba, UFS_UPIU_RPMB_WLUN);
if (!ret)
hba->wlun_dev_clr_ua = false;
out:
if (ret)
dev_err(hba->dev, "%s: Failed to clear UAC WLUNS ret = %d\n",
__func__, ret);
return ret;
}
/**
* ufshcd_probe_hba - probe hba to detect device and initialize it
* @hba: per-adapter instance
......@@ -8139,8 +8023,6 @@ static int ufshcd_probe_hba(struct ufs_hba *hba, bool init_dev_params)
/* UFS device is also active now */
ufshcd_set_ufs_dev_active(hba);
ufshcd_force_reset_auto_bkops(hba);
hba->wlun_dev_clr_ua = true;
hba->wlun_rpmb_clr_ua = true;
/* Gear up to HS gear if supported */
if (hba->max_pwr_info.is_valid) {
......@@ -8700,8 +8582,6 @@ static int ufshcd_set_dev_pwr_mode(struct ufs_hba *hba,
* handling context.
*/
hba->host->eh_noresume = 1;
if (hba->wlun_dev_clr_ua)
ufshcd_clear_ua_wlun(hba, UFS_UPIU_UFS_DEVICE_WLUN);
cmd[4] = pwr_mode << 4;
......@@ -9767,10 +9647,6 @@ void ufshcd_resume_complete(struct device *dev)
ufshcd_rpm_put(hba);
hba->complete_put = false;
}
if (hba->rpmb_complete_put) {
ufshcd_rpmb_rpm_put(hba);
hba->rpmb_complete_put = false;
}
}
EXPORT_SYMBOL_GPL(ufshcd_resume_complete);
......@@ -9793,10 +9669,6 @@ int ufshcd_suspend_prepare(struct device *dev)
}
hba->complete_put = true;
}
if (hba->sdev_rpmb) {
ufshcd_rpmb_rpm_get_sync(hba);
hba->rpmb_complete_put = true;
}
return 0;
}
EXPORT_SYMBOL_GPL(ufshcd_suspend_prepare);
......@@ -9865,49 +9737,6 @@ static struct scsi_driver ufs_dev_wlun_template = {
},
};
static int ufshcd_rpmb_probe(struct device *dev)
{
return is_rpmb_wlun(to_scsi_device(dev)) ? 0 : -ENODEV;
}
static inline int ufshcd_clear_rpmb_uac(struct ufs_hba *hba)
{
int ret = 0;
if (!hba->wlun_rpmb_clr_ua)
return 0;
ret = ufshcd_clear_ua_wlun(hba, UFS_UPIU_RPMB_WLUN);
if (!ret)
hba->wlun_rpmb_clr_ua = 0;
return ret;
}
#ifdef CONFIG_PM
static int ufshcd_rpmb_resume(struct device *dev)
{
struct ufs_hba *hba = wlun_dev_to_hba(dev);
if (hba->sdev_rpmb)
ufshcd_clear_rpmb_uac(hba);
return 0;
}
#endif
static const struct dev_pm_ops ufs_rpmb_pm_ops = {
SET_RUNTIME_PM_OPS(NULL, ufshcd_rpmb_resume, NULL)
SET_SYSTEM_SLEEP_PM_OPS(NULL, ufshcd_rpmb_resume)
};
/* ufs_rpmb_wlun_template - Describes UFS RPMB WLUN. Used only to send UAC. */
static struct scsi_driver ufs_rpmb_wlun_template = {
.gendrv = {
.name = "ufs_rpmb_wlun",
.owner = THIS_MODULE,
.probe = ufshcd_rpmb_probe,
.pm = &ufs_rpmb_pm_ops,
},
};
static int __init ufshcd_core_init(void)
{
int ret;
......@@ -9916,24 +9745,13 @@ static int __init ufshcd_core_init(void)
ret = scsi_register_driver(&ufs_dev_wlun_template.gendrv);
if (ret)
goto debugfs_exit;
ret = scsi_register_driver(&ufs_rpmb_wlun_template.gendrv);
if (ret)
goto unregister;
return ret;
unregister:
scsi_unregister_driver(&ufs_dev_wlun_template.gendrv);
debugfs_exit:
ufs_debugfs_exit();
ufs_debugfs_exit();
return ret;
}
static void __exit ufshcd_core_exit(void)
{
ufs_debugfs_exit();
scsi_unregister_driver(&ufs_rpmb_wlun_template.gendrv);
scsi_unregister_driver(&ufs_dev_wlun_template.gendrv);
}
......
......@@ -875,9 +875,6 @@ struct ufs_hba {
struct ufs_vreg_info vreg_info;
struct list_head clk_list_head;
bool wlun_dev_clr_ua;
bool wlun_rpmb_clr_ua;
/* Number of requests aborts */
int req_abort_count;
......@@ -924,7 +921,6 @@ struct ufs_hba {
#endif
u32 luns_avail;
bool complete_put;
bool rpmb_complete_put;
};
/* Returns true if clocks can be gated. Otherwise false */
......@@ -1408,14 +1404,4 @@ static inline int ufshcd_rpm_put(struct ufs_hba *hba)
return pm_runtime_put(&hba->sdev_ufs_device->sdev_gendev);
}
static inline int ufshcd_rpmb_rpm_get_sync(struct ufs_hba *hba)
{
return pm_runtime_get_sync(&hba->sdev_rpmb->sdev_gendev);
}
static inline int ufshcd_rpmb_rpm_put(struct ufs_hba *hba)
{
return pm_runtime_put(&hba->sdev_rpmb->sdev_gendev);
}
#endif /* End of Header */
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment