Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit d8f32166 authored by Christoph Hellwig's avatar Christoph Hellwig Committed by Jens Axboe
Browse files

nvme: switch delete SQ/CQ to blk_execute_rq_nowait

parent 7688faa6
Loading
Loading
Loading
Loading
+15 −34
Original line number Original line Diff line number Diff line
@@ -86,8 +86,6 @@ static void nvme_dev_shutdown(struct nvme_dev *dev);
struct async_cmd_info {
struct async_cmd_info {
	struct kthread_work work;
	struct kthread_work work;
	struct kthread_worker *worker;
	struct kthread_worker *worker;
	struct request *req;
	u32 result;
	int status;
	int status;
	void *ctx;
	void *ctx;
};
};
@@ -391,16 +389,6 @@ static void abort_completion(struct nvme_queue *nvmeq, void *ctx,
	atomic_inc(&nvmeq->dev->ctrl.abort_limit);
	atomic_inc(&nvmeq->dev->ctrl.abort_limit);
}
}


static void async_completion(struct nvme_queue *nvmeq, void *ctx,
						struct nvme_completion *cqe)
{
	struct async_cmd_info *cmdinfo = ctx;
	cmdinfo->result = le32_to_cpup(&cqe->result);
	cmdinfo->status = le16_to_cpup(&cqe->status) >> 1;
	queue_kthread_work(cmdinfo->worker, &cmdinfo->work);
	blk_mq_free_request(cmdinfo->req);
}

static inline struct nvme_cmd_info *get_cmd_from_tag(struct nvme_queue *nvmeq,
static inline struct nvme_cmd_info *get_cmd_from_tag(struct nvme_queue *nvmeq,
				  unsigned int tag)
				  unsigned int tag)
{
{
@@ -985,28 +973,13 @@ static int nvme_submit_async_admin_req(struct nvme_dev *dev)
	return 0;
	return 0;
}
}


static int nvme_submit_admin_async_cmd(struct nvme_dev *dev,
static void async_cmd_info_endio(struct request *req, int error)
			struct nvme_command *cmd,
			struct async_cmd_info *cmdinfo, unsigned timeout)
{
{
	struct nvme_queue *nvmeq = dev->queues[0];
	struct async_cmd_info *cmdinfo = req->end_io_data;
	struct request *req;
	struct nvme_cmd_info *cmd_rq;

	req = blk_mq_alloc_request(dev->ctrl.admin_q, WRITE, 0);
	if (IS_ERR(req))
		return PTR_ERR(req);


	req->timeout = timeout;
	cmdinfo->status = req->errors;
	cmd_rq = blk_mq_rq_to_pdu(req);
	queue_kthread_work(cmdinfo->worker, &cmdinfo->work);
	cmdinfo->req = req;
	blk_mq_free_request(req);
	nvme_set_info(cmd_rq, cmdinfo, async_completion);
	cmdinfo->status = -EINTR;

	cmd->common.command_id = req->tag;

	nvme_submit_cmd(nvmeq, cmd);
	return 0;
}
}


static int adapter_delete_queue(struct nvme_dev *dev, u8 opcode, u16 id)
static int adapter_delete_queue(struct nvme_dev *dev, u8 opcode, u16 id)
@@ -1920,6 +1893,7 @@ static void nvme_del_queue_end(struct nvme_queue *nvmeq)
static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode,
static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode,
						kthread_work_func_t fn)
						kthread_work_func_t fn)
{
{
	struct request *req;
	struct nvme_command c;
	struct nvme_command c;


	memset(&c, 0, sizeof(c));
	memset(&c, 0, sizeof(c));
@@ -1927,8 +1901,15 @@ static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode,
	c.delete_queue.qid = cpu_to_le16(nvmeq->qid);
	c.delete_queue.qid = cpu_to_le16(nvmeq->qid);


	init_kthread_work(&nvmeq->cmdinfo.work, fn);
	init_kthread_work(&nvmeq->cmdinfo.work, fn);
	return nvme_submit_admin_async_cmd(nvmeq->dev, &c, &nvmeq->cmdinfo,

								ADMIN_TIMEOUT);
	req = nvme_alloc_request(nvmeq->dev->ctrl.admin_q, &c, 0);
	if (IS_ERR(req))
		return PTR_ERR(req);

	req->timeout = ADMIN_TIMEOUT;
	req->end_io_data = &nvmeq->cmdinfo;
	blk_execute_rq_nowait(req->q, NULL, req, 0, async_cmd_info_endio);
	return 0;
}
}


static void nvme_del_cq_work_handler(struct kthread_work *work)
static void nvme_del_cq_work_handler(struct kthread_work *work)