nvme: Move transports to use nvme-core workqueue
authorSagi Grimberg <sagi@grimberg.me>
Wed, 7 Jun 2017 18:31:55 +0000 (20:31 +0200)
committerChristoph Hellwig <hch@lst.de>
Thu, 15 Jun 2017 12:29:43 +0000 (14:29 +0200)
Instead of each transport using it's own workqueue, export
a single nvme-core workqueue and use that instead.

In the future, this will help us moving towards some unification
if controller setup/teardown flows.

Signed-off-by: Sagi Grimberg <sagi@grimberg.me>
Signed-off-by: Christoph Hellwig <hch@lst.de>
drivers/nvme/host/core.c
drivers/nvme/host/fc.c
drivers/nvme/host/nvme.h
drivers/nvme/host/pci.c
drivers/nvme/host/rdma.c
drivers/nvme/target/loop.c

index 6d53094f4b8e52f201e64de868f408c6191b7eca..9a7fcad62d81af5413b247f1463c8a473a380613 100644 (file)
@@ -65,6 +65,9 @@ static bool force_apst;
 module_param(force_apst, bool, 0644);
 MODULE_PARM_DESC(force_apst, "allow APST for newly enumerated devices even if quirked off");
 
+struct workqueue_struct *nvme_wq;
+EXPORT_SYMBOL_GPL(nvme_wq);
+
 static LIST_HEAD(nvme_ctrl_list);
 static DEFINE_SPINLOCK(dev_list_lock);
 
@@ -2538,10 +2541,15 @@ int __init nvme_core_init(void)
 {
        int result;
 
+       nvme_wq = alloc_workqueue("nvme-wq",
+                       WQ_UNBOUND | WQ_MEM_RECLAIM | WQ_SYSFS, 0);
+       if (!nvme_wq)
+               return -ENOMEM;
+
        result = __register_chrdev(nvme_char_major, 0, NVME_MINORS, "nvme",
                                                        &nvme_dev_fops);
        if (result < 0)
-               return result;
+               goto destroy_wq;
        else if (result > 0)
                nvme_char_major = result;
 
@@ -2553,8 +2561,10 @@ int __init nvme_core_init(void)
 
        return 0;
 
- unregister_chrdev:
+unregister_chrdev:
        __unregister_chrdev(nvme_char_major, 0, NVME_MINORS, "nvme");
+destroy_wq:
+       destroy_workqueue(nvme_wq);
        return result;
 }
 
@@ -2562,6 +2572,7 @@ void nvme_core_exit(void)
 {
        class_destroy(nvme_class);
        __unregister_chrdev(nvme_char_major, 0, NVME_MINORS, "nvme");
+       destroy_workqueue(nvme_wq);
 }
 
 MODULE_LICENSE("GPL");
index 1df653ae3638a5d764c9699eb554c85ac182fd4f..e6084f3b365f67eb359e1696e1007d363cbbc4fd 100644 (file)
@@ -214,7 +214,6 @@ static LIST_HEAD(nvme_fc_lport_list);
 static DEFINE_IDA(nvme_fc_local_port_cnt);
 static DEFINE_IDA(nvme_fc_ctrl_cnt);
 
-static struct workqueue_struct *nvme_fc_wq;
 
 
 
@@ -1775,7 +1774,7 @@ nvme_fc_error_recovery(struct nvme_fc_ctrl *ctrl, char *errmsg)
                return;
        }
 
-       if (!queue_work(nvme_fc_wq, &ctrl->reset_work))
+       if (!queue_work(nvme_wq, &ctrl->reset_work))
                dev_err(ctrl->ctrl.device,
                        "NVME-FC{%d}: error_recovery: Failed to schedule "
                        "reset work\n", ctrl->cnum);
@@ -2555,7 +2554,7 @@ __nvme_fc_schedule_delete_work(struct nvme_fc_ctrl *ctrl)
        if (!nvme_change_ctrl_state(&ctrl->ctrl, NVME_CTRL_DELETING))
                return true;
 
-       if (!queue_work(nvme_fc_wq, &ctrl->delete_work))
+       if (!queue_work(nvme_wq, &ctrl->delete_work))
                return true;
 
        return false;
@@ -2582,7 +2581,7 @@ nvme_fc_del_nvme_ctrl(struct nvme_ctrl *nctrl)
        ret = __nvme_fc_del_ctrl(ctrl);
 
        if (!ret)
-               flush_workqueue(nvme_fc_wq);
+               flush_workqueue(nvme_wq);
 
        nvme_put_ctrl(&ctrl->ctrl);
 
@@ -2607,7 +2606,7 @@ nvme_fc_reconnect_or_delete(struct nvme_fc_ctrl *ctrl, int status)
                dev_info(ctrl->ctrl.device,
                        "NVME-FC{%d}: Reconnect attempt in %d seconds.\n",
                        ctrl->cnum, ctrl->ctrl.opts->reconnect_delay);
-               queue_delayed_work(nvme_fc_wq, &ctrl->connect_work,
+               queue_delayed_work(nvme_wq, &ctrl->connect_work,
                                ctrl->ctrl.opts->reconnect_delay * HZ);
        } else {
                dev_warn(ctrl->ctrl.device,
@@ -2651,7 +2650,7 @@ nvme_fc_reset_nvme_ctrl(struct nvme_ctrl *nctrl)
        if (!nvme_change_ctrl_state(&ctrl->ctrl, NVME_CTRL_RESETTING))
                return -EBUSY;
 
-       if (!queue_work(nvme_fc_wq, &ctrl->reset_work))
+       if (!queue_work(nvme_wq, &ctrl->reset_work))
                return -EBUSY;
 
        flush_work(&ctrl->reset_work);
@@ -2966,20 +2965,7 @@ static struct nvmf_transport_ops nvme_fc_transport = {
 
 static int __init nvme_fc_init_module(void)
 {
-       int ret;
-
-       nvme_fc_wq = create_workqueue("nvme_fc_wq");
-       if (!nvme_fc_wq)
-               return -ENOMEM;
-
-       ret = nvmf_register_transport(&nvme_fc_transport);
-       if (ret)
-               goto err;
-
-       return 0;
-err:
-       destroy_workqueue(nvme_fc_wq);
-       return ret;
+       return nvmf_register_transport(&nvme_fc_transport);
 }
 
 static void __exit nvme_fc_exit_module(void)
@@ -2990,8 +2976,6 @@ static void __exit nvme_fc_exit_module(void)
 
        nvmf_unregister_transport(&nvme_fc_transport);
 
-       destroy_workqueue(nvme_fc_wq);
-
        ida_destroy(&nvme_fc_local_port_cnt);
        ida_destroy(&nvme_fc_ctrl_cnt);
 }
index e2e341bba619e6898dca1e23fc2622710d37b3f4..80e9adce2691af03dcdf0e4575f454e52ee22fd3 100644 (file)
@@ -33,6 +33,8 @@ extern unsigned char shutdown_timeout;
 #define NVME_DEFAULT_KATO      5
 #define NVME_KATO_GRACE                10
 
+extern struct workqueue_struct *nvme_wq;
+
 enum {
        NVME_NS_LBA             = 0,
        NVME_NS_LIGHTNVM        = 1,
index 73d9b412f2911030e858c31af76b2d921f3d8c9c..ebd5cdfc0174c9faa9cc3b54c8f53788cdea09fd 100644 (file)
@@ -71,8 +71,6 @@ module_param(max_host_mem_size_mb, uint, 0444);
 MODULE_PARM_DESC(max_host_mem_size_mb,
        "Maximum Host Memory Buffer (HMB) size per controller (in MiB)");
 
-static struct workqueue_struct *nvme_workq;
-
 struct nvme_dev;
 struct nvme_queue;
 
@@ -2190,7 +2188,7 @@ static int nvme_reset(struct nvme_dev *dev)
                return -ENODEV;
        if (!nvme_change_ctrl_state(&dev->ctrl, NVME_CTRL_RESETTING))
                return -EBUSY;
-       if (!queue_work(nvme_workq, &dev->reset_work))
+       if (!queue_work(nvme_wq, &dev->reset_work))
                return -EBUSY;
        return 0;
 }
@@ -2318,7 +2316,7 @@ static int nvme_probe(struct pci_dev *pdev, const struct pci_device_id *id)
        nvme_change_ctrl_state(&dev->ctrl, NVME_CTRL_RESETTING);
        dev_info(dev->ctrl.device, "pci function %s\n", dev_name(&pdev->dev));
 
-       queue_work(nvme_workq, &dev->reset_work);
+       queue_work(nvme_wq, &dev->reset_work);
        return 0;
 
  release_pools:
@@ -2506,22 +2504,12 @@ static struct pci_driver nvme_driver = {
 
 static int __init nvme_init(void)
 {
-       int result;
-
-       nvme_workq = alloc_workqueue("nvme", WQ_UNBOUND | WQ_MEM_RECLAIM, 0);
-       if (!nvme_workq)
-               return -ENOMEM;
-
-       result = pci_register_driver(&nvme_driver);
-       if (result)
-               destroy_workqueue(nvme_workq);
-       return result;
+       return pci_register_driver(&nvme_driver);
 }
 
 static void __exit nvme_exit(void)
 {
        pci_unregister_driver(&nvme_driver);
-       destroy_workqueue(nvme_workq);
        _nvme_check_size();
 }
 
index 17dddaf4ca3f4fb0ac75a9331e418b1745b7135d..8805d3400846e37bd16688bc4a3dc32225c5f262 100644 (file)
@@ -140,8 +140,6 @@ static DEFINE_MUTEX(device_list_mutex);
 static LIST_HEAD(nvme_rdma_ctrl_list);
 static DEFINE_MUTEX(nvme_rdma_ctrl_mutex);
 
-static struct workqueue_struct *nvme_rdma_wq;
-
 /*
  * Disabling this option makes small I/O goes faster, but is fundamentally
  * unsafe.  With it turned off we will have to register a global rkey that
@@ -712,11 +710,11 @@ static void nvme_rdma_reconnect_or_remove(struct nvme_rdma_ctrl *ctrl)
        if (nvmf_should_reconnect(&ctrl->ctrl)) {
                dev_info(ctrl->ctrl.device, "Reconnecting in %d seconds...\n",
                        ctrl->ctrl.opts->reconnect_delay);
-               queue_delayed_work(nvme_rdma_wq, &ctrl->reconnect_work,
+               queue_delayed_work(nvme_wq, &ctrl->reconnect_work,
                                ctrl->ctrl.opts->reconnect_delay * HZ);
        } else {
                dev_info(ctrl->ctrl.device, "Removing controller...\n");
-               queue_work(nvme_rdma_wq, &ctrl->delete_work);
+               queue_work(nvme_wq, &ctrl->delete_work);
        }
 }
 
@@ -825,7 +823,7 @@ static void nvme_rdma_error_recovery(struct nvme_rdma_ctrl *ctrl)
        if (!nvme_change_ctrl_state(&ctrl->ctrl, NVME_CTRL_RECONNECTING))
                return;
 
-       queue_work(nvme_rdma_wq, &ctrl->err_work);
+       queue_work(nvme_wq, &ctrl->err_work);
 }
 
 static void nvme_rdma_wr_error(struct ib_cq *cq, struct ib_wc *wc,
@@ -1692,7 +1690,7 @@ static int __nvme_rdma_del_ctrl(struct nvme_rdma_ctrl *ctrl)
        if (!nvme_change_ctrl_state(&ctrl->ctrl, NVME_CTRL_DELETING))
                return -EBUSY;
 
-       if (!queue_work(nvme_rdma_wq, &ctrl->delete_work))
+       if (!queue_work(nvme_wq, &ctrl->delete_work))
                return -EBUSY;
 
        return 0;
@@ -1768,7 +1766,7 @@ static void nvme_rdma_reset_ctrl_work(struct work_struct *work)
 del_dead_ctrl:
        /* Deleting this dead controller... */
        dev_warn(ctrl->ctrl.device, "Removing after reset failure\n");
-       WARN_ON(!queue_work(nvme_rdma_wq, &ctrl->delete_work));
+       WARN_ON(!queue_work(nvme_wq, &ctrl->delete_work));
 }
 
 static int nvme_rdma_reset_ctrl(struct nvme_ctrl *nctrl)
@@ -1778,7 +1776,7 @@ static int nvme_rdma_reset_ctrl(struct nvme_ctrl *nctrl)
        if (!nvme_change_ctrl_state(&ctrl->ctrl, NVME_CTRL_RESETTING))
                return -EBUSY;
 
-       if (!queue_work(nvme_rdma_wq, &ctrl->reset_work))
+       if (!queue_work(nvme_wq, &ctrl->reset_work))
                return -EBUSY;
 
        flush_work(&ctrl->reset_work);
@@ -2015,7 +2013,7 @@ static void nvme_rdma_remove_one(struct ib_device *ib_device, void *client_data)
        }
        mutex_unlock(&nvme_rdma_ctrl_mutex);
 
-       flush_workqueue(nvme_rdma_wq);
+       flush_workqueue(nvme_wq);
 }
 
 static struct ib_client nvme_rdma_ib_client = {
@@ -2028,13 +2026,9 @@ static int __init nvme_rdma_init_module(void)
 {
        int ret;
 
-       nvme_rdma_wq = create_workqueue("nvme_rdma_wq");
-       if (!nvme_rdma_wq)
-               return -ENOMEM;
-
        ret = ib_register_client(&nvme_rdma_ib_client);
        if (ret)
-               goto err_destroy_wq;
+               return ret;
 
        ret = nvmf_register_transport(&nvme_rdma_transport);
        if (ret)
@@ -2044,8 +2038,6 @@ static int __init nvme_rdma_init_module(void)
 
 err_unreg_client:
        ib_unregister_client(&nvme_rdma_ib_client);
-err_destroy_wq:
-       destroy_workqueue(nvme_rdma_wq);
        return ret;
 }
 
@@ -2053,7 +2045,6 @@ static void __exit nvme_rdma_cleanup_module(void)
 {
        nvmf_unregister_transport(&nvme_rdma_transport);
        ib_unregister_client(&nvme_rdma_ib_client);
-       destroy_workqueue(nvme_rdma_wq);
 }
 
 module_init(nvme_rdma_init_module);
index 1f5bd3cd50411e45499d9e04c4d212c03d910a3c..b7715b46e0215d0954a015e4a35862d8857ca601 100644 (file)
@@ -150,7 +150,7 @@ nvme_loop_timeout(struct request *rq, bool reserved)
        struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(rq);
 
        /* queue error recovery */
-       schedule_work(&iod->queue->ctrl->reset_work);
+       queue_work(nvme_wq, &iod->queue->ctrl->reset_work);
 
        /* fail with DNR on admin cmd timeout */
        nvme_req(rq)->status = NVME_SC_ABORT_REQ | NVME_SC_DNR;
@@ -465,7 +465,7 @@ static int __nvme_loop_del_ctrl(struct nvme_loop_ctrl *ctrl)
        if (!nvme_change_ctrl_state(&ctrl->ctrl, NVME_CTRL_DELETING))
                return -EBUSY;
 
-       if (!schedule_work(&ctrl->delete_work))
+       if (!queue_work(nvme_wq, &ctrl->delete_work))
                return -EBUSY;
 
        return 0;
@@ -545,7 +545,7 @@ static int nvme_loop_reset_ctrl(struct nvme_ctrl *nctrl)
        if (!nvme_change_ctrl_state(&ctrl->ctrl, NVME_CTRL_RESETTING))
                return -EBUSY;
 
-       if (!schedule_work(&ctrl->reset_work))
+       if (!queue_work(nvme_wq, &ctrl->reset_work))
                return -EBUSY;
 
        flush_work(&ctrl->reset_work);
@@ -762,7 +762,7 @@ static void __exit nvme_loop_cleanup_module(void)
                __nvme_loop_del_ctrl(ctrl);
        mutex_unlock(&nvme_loop_ctrl_mutex);
 
-       flush_scheduled_work();
+       flush_workqueue(nvme_wq);
 }
 
 module_init(nvme_loop_init_module);