Loading drivers/nvme/host/pci.c +15 −34 Original line number Original line Diff line number Diff line Loading @@ -86,8 +86,6 @@ static void nvme_dev_shutdown(struct nvme_dev *dev); struct async_cmd_info { struct async_cmd_info { struct kthread_work work; struct kthread_work work; struct kthread_worker *worker; struct kthread_worker *worker; struct request *req; u32 result; int status; int status; void *ctx; void *ctx; }; }; Loading Loading @@ -391,16 +389,6 @@ static void abort_completion(struct nvme_queue *nvmeq, void *ctx, atomic_inc(&nvmeq->dev->ctrl.abort_limit); atomic_inc(&nvmeq->dev->ctrl.abort_limit); } } static void async_completion(struct nvme_queue *nvmeq, void *ctx, struct nvme_completion *cqe) { struct async_cmd_info *cmdinfo = ctx; cmdinfo->result = le32_to_cpup(&cqe->result); cmdinfo->status = le16_to_cpup(&cqe->status) >> 1; queue_kthread_work(cmdinfo->worker, &cmdinfo->work); blk_mq_free_request(cmdinfo->req); } static inline struct nvme_cmd_info *get_cmd_from_tag(struct nvme_queue *nvmeq, static inline struct nvme_cmd_info *get_cmd_from_tag(struct nvme_queue *nvmeq, unsigned int tag) unsigned int tag) { { Loading Loading @@ -985,28 +973,13 @@ static int nvme_submit_async_admin_req(struct nvme_dev *dev) return 0; return 0; } } static int nvme_submit_admin_async_cmd(struct nvme_dev *dev, static void async_cmd_info_endio(struct request *req, int error) struct nvme_command *cmd, struct async_cmd_info *cmdinfo, unsigned timeout) { { struct nvme_queue *nvmeq = dev->queues[0]; struct async_cmd_info *cmdinfo = req->end_io_data; struct request *req; struct nvme_cmd_info *cmd_rq; req = blk_mq_alloc_request(dev->ctrl.admin_q, WRITE, 0); if (IS_ERR(req)) return PTR_ERR(req); req->timeout = timeout; cmdinfo->status = req->errors; cmd_rq = blk_mq_rq_to_pdu(req); queue_kthread_work(cmdinfo->worker, &cmdinfo->work); cmdinfo->req = req; blk_mq_free_request(req); nvme_set_info(cmd_rq, cmdinfo, async_completion); cmdinfo->status = -EINTR; cmd->common.command_id = req->tag; nvme_submit_cmd(nvmeq, cmd); return 0; } } static int adapter_delete_queue(struct nvme_dev *dev, u8 opcode, u16 id) static int adapter_delete_queue(struct nvme_dev *dev, u8 opcode, u16 id) Loading Loading @@ -1920,6 +1893,7 @@ static void nvme_del_queue_end(struct nvme_queue *nvmeq) static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode, static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode, kthread_work_func_t fn) kthread_work_func_t fn) { { struct request *req; struct nvme_command c; struct nvme_command c; memset(&c, 0, sizeof(c)); memset(&c, 0, sizeof(c)); Loading @@ -1927,8 +1901,15 @@ static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode, c.delete_queue.qid = cpu_to_le16(nvmeq->qid); c.delete_queue.qid = cpu_to_le16(nvmeq->qid); init_kthread_work(&nvmeq->cmdinfo.work, fn); init_kthread_work(&nvmeq->cmdinfo.work, fn); return nvme_submit_admin_async_cmd(nvmeq->dev, &c, &nvmeq->cmdinfo, ADMIN_TIMEOUT); req = nvme_alloc_request(nvmeq->dev->ctrl.admin_q, &c, 0); if (IS_ERR(req)) return PTR_ERR(req); req->timeout = ADMIN_TIMEOUT; req->end_io_data = &nvmeq->cmdinfo; blk_execute_rq_nowait(req->q, NULL, req, 0, async_cmd_info_endio); return 0; } } static void nvme_del_cq_work_handler(struct kthread_work *work) static void nvme_del_cq_work_handler(struct kthread_work *work) Loading Loading
drivers/nvme/host/pci.c +15 −34 Original line number Original line Diff line number Diff line Loading @@ -86,8 +86,6 @@ static void nvme_dev_shutdown(struct nvme_dev *dev); struct async_cmd_info { struct async_cmd_info { struct kthread_work work; struct kthread_work work; struct kthread_worker *worker; struct kthread_worker *worker; struct request *req; u32 result; int status; int status; void *ctx; void *ctx; }; }; Loading Loading @@ -391,16 +389,6 @@ static void abort_completion(struct nvme_queue *nvmeq, void *ctx, atomic_inc(&nvmeq->dev->ctrl.abort_limit); atomic_inc(&nvmeq->dev->ctrl.abort_limit); } } static void async_completion(struct nvme_queue *nvmeq, void *ctx, struct nvme_completion *cqe) { struct async_cmd_info *cmdinfo = ctx; cmdinfo->result = le32_to_cpup(&cqe->result); cmdinfo->status = le16_to_cpup(&cqe->status) >> 1; queue_kthread_work(cmdinfo->worker, &cmdinfo->work); blk_mq_free_request(cmdinfo->req); } static inline struct nvme_cmd_info *get_cmd_from_tag(struct nvme_queue *nvmeq, static inline struct nvme_cmd_info *get_cmd_from_tag(struct nvme_queue *nvmeq, unsigned int tag) unsigned int tag) { { Loading Loading @@ -985,28 +973,13 @@ static int nvme_submit_async_admin_req(struct nvme_dev *dev) return 0; return 0; } } static int nvme_submit_admin_async_cmd(struct nvme_dev *dev, static void async_cmd_info_endio(struct request *req, int error) struct nvme_command *cmd, struct async_cmd_info *cmdinfo, unsigned timeout) { { struct nvme_queue *nvmeq = dev->queues[0]; struct async_cmd_info *cmdinfo = req->end_io_data; struct request *req; struct nvme_cmd_info *cmd_rq; req = blk_mq_alloc_request(dev->ctrl.admin_q, WRITE, 0); if (IS_ERR(req)) return PTR_ERR(req); req->timeout = timeout; cmdinfo->status = req->errors; cmd_rq = blk_mq_rq_to_pdu(req); queue_kthread_work(cmdinfo->worker, &cmdinfo->work); cmdinfo->req = req; blk_mq_free_request(req); nvme_set_info(cmd_rq, cmdinfo, async_completion); cmdinfo->status = -EINTR; cmd->common.command_id = req->tag; nvme_submit_cmd(nvmeq, cmd); return 0; } } static int adapter_delete_queue(struct nvme_dev *dev, u8 opcode, u16 id) static int adapter_delete_queue(struct nvme_dev *dev, u8 opcode, u16 id) Loading Loading @@ -1920,6 +1893,7 @@ static void nvme_del_queue_end(struct nvme_queue *nvmeq) static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode, static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode, kthread_work_func_t fn) kthread_work_func_t fn) { { struct request *req; struct nvme_command c; struct nvme_command c; memset(&c, 0, sizeof(c)); memset(&c, 0, sizeof(c)); Loading @@ -1927,8 +1901,15 @@ static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode, c.delete_queue.qid = cpu_to_le16(nvmeq->qid); c.delete_queue.qid = cpu_to_le16(nvmeq->qid); init_kthread_work(&nvmeq->cmdinfo.work, fn); init_kthread_work(&nvmeq->cmdinfo.work, fn); return nvme_submit_admin_async_cmd(nvmeq->dev, &c, &nvmeq->cmdinfo, ADMIN_TIMEOUT); req = nvme_alloc_request(nvmeq->dev->ctrl.admin_q, &c, 0); if (IS_ERR(req)) return PTR_ERR(req); req->timeout = ADMIN_TIMEOUT; req->end_io_data = &nvmeq->cmdinfo; blk_execute_rq_nowait(req->q, NULL, req, 0, async_cmd_info_endio); return 0; } } static void nvme_del_cq_work_handler(struct kthread_work *work) static void nvme_del_cq_work_handler(struct kthread_work *work) Loading