Commit 0a27ae51 authored by Matthew R. Ochs's avatar Matthew R. Ochs Committed by James Bottomley

cxlflash: Fix to avoid CXL services during EEH

During an EEH freeze event, certain CXL services should not be
called until after the hardware reset has taken place. Doing so
can result in unnecessary failures and possibly cause other ill
effects by triggering hardware accesses. This translates to a
requirement to quiesce all threads that may potentially use CXL
runtime service during this window. In particular, multiple ioctls
make use of the CXL services when acting on contexts on behalf of
the user. Thus, it is essential to 'drain' running ioctls _before_
proceeding with handling the EEH freeze event.

Create the ability to drain ioctls by wrapping the ioctl handler
call in a read semaphore and then implementing a small routine that
obtains the write semaphore, effectively creating a wait point for
all currently executing ioctls.
Signed-off-by: default avatarMatthew R. Ochs <mrochs@linux.vnet.ibm.com>
Signed-off-by: default avatarManoj N. Kumar <manoj@linux.vnet.ibm.com>
Reviewed-by: default avatarBrian King <brking@linux.vnet.ibm.com>
Reviewed-by: default avatarDaniel Axtens <dja@axtens.net>
Reviewed-by: default avatarTomas Henzl <thenzl@redhat.com>
Signed-off-by: default avatarJames Bottomley <JBottomley@Odin.com>
parent a76df368
...@@ -16,6 +16,7 @@ ...@@ -16,6 +16,7 @@
#define _CXLFLASH_COMMON_H #define _CXLFLASH_COMMON_H
#include <linux/list.h> #include <linux/list.h>
#include <linux/rwsem.h>
#include <linux/types.h> #include <linux/types.h>
#include <scsi/scsi.h> #include <scsi/scsi.h>
#include <scsi/scsi_device.h> #include <scsi/scsi_device.h>
...@@ -110,6 +111,7 @@ struct cxlflash_cfg { ...@@ -110,6 +111,7 @@ struct cxlflash_cfg {
atomic_t recovery_threads; atomic_t recovery_threads;
struct mutex ctx_recovery_mutex; struct mutex ctx_recovery_mutex;
struct mutex ctx_tbl_list_mutex; struct mutex ctx_tbl_list_mutex;
struct rw_semaphore ioctl_rwsem;
struct ctx_info *ctx_tbl[MAX_CONTEXT]; struct ctx_info *ctx_tbl[MAX_CONTEXT];
struct list_head ctx_err_recovery; /* contexts w/ recovery pending */ struct list_head ctx_err_recovery; /* contexts w/ recovery pending */
struct file_operations cxl_fops; struct file_operations cxl_fops;
......
...@@ -2311,6 +2311,7 @@ static int cxlflash_probe(struct pci_dev *pdev, ...@@ -2311,6 +2311,7 @@ static int cxlflash_probe(struct pci_dev *pdev,
cfg->lr_port = -1; cfg->lr_port = -1;
mutex_init(&cfg->ctx_tbl_list_mutex); mutex_init(&cfg->ctx_tbl_list_mutex);
mutex_init(&cfg->ctx_recovery_mutex); mutex_init(&cfg->ctx_recovery_mutex);
init_rwsem(&cfg->ioctl_rwsem);
INIT_LIST_HEAD(&cfg->ctx_err_recovery); INIT_LIST_HEAD(&cfg->ctx_err_recovery);
INIT_LIST_HEAD(&cfg->lluns); INIT_LIST_HEAD(&cfg->lluns);
...@@ -2364,6 +2365,19 @@ static int cxlflash_probe(struct pci_dev *pdev, ...@@ -2364,6 +2365,19 @@ static int cxlflash_probe(struct pci_dev *pdev,
goto out; goto out;
} }
/**
* drain_ioctls() - wait until all currently executing ioctls have completed
* @cfg: Internal structure associated with the host.
*
* Obtain write access to read/write semaphore that wraps ioctl
* handling to 'drain' ioctls currently executing.
*/
static void drain_ioctls(struct cxlflash_cfg *cfg)
{
down_write(&cfg->ioctl_rwsem);
up_write(&cfg->ioctl_rwsem);
}
/** /**
* cxlflash_pci_error_detected() - called when a PCI error is detected * cxlflash_pci_error_detected() - called when a PCI error is detected
* @pdev: PCI device struct. * @pdev: PCI device struct.
...@@ -2383,16 +2397,14 @@ static pci_ers_result_t cxlflash_pci_error_detected(struct pci_dev *pdev, ...@@ -2383,16 +2397,14 @@ static pci_ers_result_t cxlflash_pci_error_detected(struct pci_dev *pdev,
switch (state) { switch (state) {
case pci_channel_io_frozen: case pci_channel_io_frozen:
cfg->state = STATE_LIMBO; cfg->state = STATE_LIMBO;
/* Turn off legacy I/O */
scsi_block_requests(cfg->host); scsi_block_requests(cfg->host);
drain_ioctls(cfg);
rc = cxlflash_mark_contexts_error(cfg); rc = cxlflash_mark_contexts_error(cfg);
if (unlikely(rc)) if (unlikely(rc))
dev_err(dev, "%s: Failed to mark user contexts!(%d)\n", dev_err(dev, "%s: Failed to mark user contexts!(%d)\n",
__func__, rc); __func__, rc);
term_mc(cfg, UNDO_START); term_mc(cfg, UNDO_START);
stop_afu(cfg); stop_afu(cfg);
return PCI_ERS_RESULT_NEED_RESET; return PCI_ERS_RESULT_NEED_RESET;
case pci_channel_io_perm_failure: case pci_channel_io_perm_failure:
cfg->state = STATE_FAILTERM; cfg->state = STATE_FAILTERM;
......
...@@ -1213,6 +1213,46 @@ static const struct file_operations null_fops = { ...@@ -1213,6 +1213,46 @@ static const struct file_operations null_fops = {
.owner = THIS_MODULE, .owner = THIS_MODULE,
}; };
/**
* check_state() - checks and responds to the current adapter state
* @cfg: Internal structure associated with the host.
*
* This routine can block and should only be used on process context.
* It assumes that the caller is an ioctl thread and holding the ioctl
* read semaphore. This is temporarily let up across the wait to allow
* for draining actively running ioctls. Also note that when waking up
* from waiting in reset, the state is unknown and must be checked again
* before proceeding.
*
* Return: 0 on success, -errno on failure
*/
static int check_state(struct cxlflash_cfg *cfg)
{
struct device *dev = &cfg->dev->dev;
int rc = 0;
retry:
switch (cfg->state) {
case STATE_LIMBO:
dev_dbg(dev, "%s: Limbo state, going to wait...\n", __func__);
up_read(&cfg->ioctl_rwsem);
rc = wait_event_interruptible(cfg->limbo_waitq,
cfg->state != STATE_LIMBO);
down_read(&cfg->ioctl_rwsem);
if (unlikely(rc))
break;
goto retry;
case STATE_FAILTERM:
dev_dbg(dev, "%s: Failed/Terminating!\n", __func__);
rc = -ENODEV;
break;
default:
break;
}
return rc;
}
/** /**
* cxlflash_disk_attach() - attach a LUN to a context * cxlflash_disk_attach() - attach a LUN to a context
* @sdev: SCSI device associated with LUN. * @sdev: SCSI device associated with LUN.
...@@ -1522,41 +1562,6 @@ static int recover_context(struct cxlflash_cfg *cfg, struct ctx_info *ctxi) ...@@ -1522,41 +1562,6 @@ static int recover_context(struct cxlflash_cfg *cfg, struct ctx_info *ctxi)
goto out; goto out;
} }
/**
* check_state() - checks and responds to the current adapter state
* @cfg: Internal structure associated with the host.
*
* This routine can block and should only be used on process context.
* Note that when waking up from waiting in limbo, the state is unknown
* and must be checked again before proceeding.
*
* Return: 0 on success, -errno on failure
*/
static int check_state(struct cxlflash_cfg *cfg)
{
struct device *dev = &cfg->dev->dev;
int rc = 0;
retry:
switch (cfg->state) {
case STATE_LIMBO:
dev_dbg(dev, "%s: Limbo, going to wait...\n", __func__);
rc = wait_event_interruptible(cfg->limbo_waitq,
cfg->state != STATE_LIMBO);
if (unlikely(rc))
break;
goto retry;
case STATE_FAILTERM:
dev_dbg(dev, "%s: Failed/Terminating!\n", __func__);
rc = -ENODEV;
break;
default:
break;
}
return rc;
}
/** /**
* cxlflash_afu_recover() - initiates AFU recovery * cxlflash_afu_recover() - initiates AFU recovery
* @sdev: SCSI device associated with LUN. * @sdev: SCSI device associated with LUN.
...@@ -1646,9 +1651,14 @@ static int cxlflash_afu_recover(struct scsi_device *sdev, ...@@ -1646,9 +1651,14 @@ static int cxlflash_afu_recover(struct scsi_device *sdev,
/* Test if in error state */ /* Test if in error state */
reg = readq_be(&afu->ctrl_map->mbox_r); reg = readq_be(&afu->ctrl_map->mbox_r);
if (reg == -1) { if (reg == -1) {
dev_dbg(dev, "%s: MMIO read fail! Wait for recovery...\n", dev_dbg(dev, "%s: MMIO fail, wait for recovery.\n", __func__);
__func__);
mutex_unlock(&ctxi->mutex); /*
* Before checking the state, put back the context obtained with
* get_context() as it is no longer needed and sleep for a short
* period of time (see prolog notes).
*/
put_context(ctxi);
ctxi = NULL; ctxi = NULL;
ssleep(1); ssleep(1);
rc = check_state(cfg); rc = check_state(cfg);
...@@ -1967,6 +1977,14 @@ static int ioctl_common(struct scsi_device *sdev, int cmd) ...@@ -1967,6 +1977,14 @@ static int ioctl_common(struct scsi_device *sdev, int cmd)
* @cmd: IOCTL command. * @cmd: IOCTL command.
* @arg: Userspace ioctl data structure. * @arg: Userspace ioctl data structure.
* *
* A read/write semaphore is used to implement a 'drain' of currently
* running ioctls. The read semaphore is taken at the beginning of each
* ioctl thread and released upon concluding execution. Additionally the
* semaphore should be released and then reacquired in any ioctl execution
* path which will wait for an event to occur that is outside the scope of
* the ioctl (i.e. an adapter reset). To drain the ioctls currently running,
* a thread simply needs to acquire the write semaphore.
*
* Return: 0 on success, -errno on failure * Return: 0 on success, -errno on failure
*/ */
int cxlflash_ioctl(struct scsi_device *sdev, int cmd, void __user *arg) int cxlflash_ioctl(struct scsi_device *sdev, int cmd, void __user *arg)
...@@ -2001,6 +2019,9 @@ int cxlflash_ioctl(struct scsi_device *sdev, int cmd, void __user *arg) ...@@ -2001,6 +2019,9 @@ int cxlflash_ioctl(struct scsi_device *sdev, int cmd, void __user *arg)
{sizeof(struct dk_cxlflash_clone), (sioctl)cxlflash_disk_clone}, {sizeof(struct dk_cxlflash_clone), (sioctl)cxlflash_disk_clone},
}; };
/* Hold read semaphore so we can drain if needed */
down_read(&cfg->ioctl_rwsem);
/* Restrict command set to physical support only for internal LUN */ /* Restrict command set to physical support only for internal LUN */
if (afu->internal_lun) if (afu->internal_lun)
switch (cmd) { switch (cmd) {
...@@ -2082,6 +2103,7 @@ int cxlflash_ioctl(struct scsi_device *sdev, int cmd, void __user *arg) ...@@ -2082,6 +2103,7 @@ int cxlflash_ioctl(struct scsi_device *sdev, int cmd, void __user *arg)
/* fall through to exit */ /* fall through to exit */
cxlflash_ioctl_exit: cxlflash_ioctl_exit:
up_read(&cfg->ioctl_rwsem);
if (unlikely(rc && known_ioctl)) if (unlikely(rc && known_ioctl))
dev_err(dev, "%s: ioctl %s (%08X) on dev(%d/%d/%d/%llu) " dev_err(dev, "%s: ioctl %s (%08X) on dev(%d/%d/%d/%llu) "
"returned rc %d\n", __func__, "returned rc %d\n", __func__,
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment