[PATCH 2/3] nvme: add polling options for loop target
Keith Busch
kbusch at meta.com
Tue Mar 21 17:23:49 PDT 2023
From: Keith Busch <kbusch at kernel.org>
This is for mostly for testing purposes.
Signed-off-by: Keith Busch <kbusch at kernel.org>
---
drivers/nvme/target/loop.c | 63 +++++++++++++++++++++++++++++++++++---
1 file changed, 58 insertions(+), 5 deletions(-)
diff --git a/drivers/nvme/target/loop.c b/drivers/nvme/target/loop.c
index f2d24b2d992f8..0587ead60b09e 100644
--- a/drivers/nvme/target/loop.c
+++ b/drivers/nvme/target/loop.c
@@ -22,6 +22,7 @@ struct nvme_loop_iod {
struct nvmet_req req;
struct nvme_loop_queue *queue;
struct work_struct work;
+ struct work_struct poll;
struct sg_table sg_table;
struct scatterlist first_sgl[];
};
@@ -37,6 +38,7 @@ struct nvme_loop_ctrl {
struct nvme_ctrl ctrl;
struct nvmet_port *port;
+ u32 io_queues[HCTX_MAX_TYPES];
};
static inline struct nvme_loop_ctrl *to_loop_ctrl(struct nvme_ctrl *ctrl)
@@ -76,7 +78,11 @@ static void nvme_loop_complete_rq(struct request *req)
struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req);
sg_free_table_chained(&iod->sg_table, NVME_INLINE_SG_CNT);
- nvme_complete_rq(req);
+
+ if (req->mq_hctx->type != HCTX_TYPE_POLL || !in_interrupt())
+ nvme_complete_rq(req);
+ else
+ queue_work(nvmet_wq, &iod->poll);
}
static struct blk_mq_tags *nvme_loop_tagset(struct nvme_loop_queue *queue)
@@ -120,6 +126,15 @@ static void nvme_loop_queue_response(struct nvmet_req *req)
}
}
+static void nvme_loop_poll_work(struct work_struct *work)
+{
+ struct nvme_loop_iod *iod =
+ container_of(work, struct nvme_loop_iod, poll);
+ struct request *req = blk_mq_rq_from_pdu(iod);
+
+ nvme_complete_rq(req);
+}
+
static void nvme_loop_execute_work(struct work_struct *work)
{
struct nvme_loop_iod *iod =
@@ -170,6 +185,30 @@ static blk_status_t nvme_loop_queue_rq(struct blk_mq_hw_ctx *hctx,
return BLK_STS_OK;
}
+static bool nvme_loop_poll_iter(struct sbitmap *bitmap, unsigned int bitnr, void *data)
+{
+ struct blk_mq_hw_ctx *hctx = data;
+ struct nvme_loop_iod *iod;
+ struct request *rq;
+
+ rq = blk_mq_tag_to_rq(hctx->tags, bitnr);
+ if (!rq)
+ return true;
+
+ iod = blk_mq_rq_to_pdu(rq);
+ flush_work(&iod->poll);
+ return true;
+}
+
+static int nvme_loop_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob)
+{
+ struct blk_mq_tags *tags = hctx->tags;
+ struct sbitmap_queue *btags = &tags->bitmap_tags;
+
+ sbitmap_for_each_set(&btags->sb, nvme_loop_poll_iter, hctx);
+ return 1;
+}
+
static void nvme_loop_submit_async_event(struct nvme_ctrl *arg)
{
struct nvme_loop_ctrl *ctrl = to_loop_ctrl(arg);
@@ -197,6 +236,7 @@ static int nvme_loop_init_iod(struct nvme_loop_ctrl *ctrl,
iod->req.cqe = &iod->cqe;
iod->queue = &ctrl->queues[queue_idx];
INIT_WORK(&iod->work, nvme_loop_execute_work);
+ INIT_WORK(&iod->poll, nvme_loop_poll_work);
return 0;
}
@@ -247,11 +287,20 @@ static int nvme_loop_init_admin_hctx(struct blk_mq_hw_ctx *hctx, void *data,
return 0;
}
+static void nvme_loop_map_queues(struct blk_mq_tag_set *set)
+{
+ struct nvme_loop_ctrl *ctrl = to_loop_ctrl(set->driver_data);
+
+ nvme_map_queues(set, &ctrl->ctrl, NULL, ctrl->io_queues);
+}
+
static const struct blk_mq_ops nvme_loop_mq_ops = {
.queue_rq = nvme_loop_queue_rq,
.complete = nvme_loop_complete_rq,
.init_request = nvme_loop_init_request,
.init_hctx = nvme_loop_init_hctx,
+ .map_queues = nvme_loop_map_queues,
+ .poll = nvme_loop_poll,
};
static const struct blk_mq_ops nvme_loop_admin_mq_ops = {
@@ -305,7 +354,7 @@ static int nvme_loop_init_io_queues(struct nvme_loop_ctrl *ctrl)
unsigned int nr_io_queues;
int ret, i;
- nr_io_queues = min(opts->nr_io_queues, num_online_cpus());
+ nr_io_queues = nvme_nr_io_queues(opts);
ret = nvme_set_queue_count(&ctrl->ctrl, &nr_io_queues);
if (ret || !nr_io_queues)
return ret;
@@ -321,6 +370,7 @@ static int nvme_loop_init_io_queues(struct nvme_loop_ctrl *ctrl)
ctrl->ctrl.queue_count++;
}
+ nvme_set_io_queues(opts, nr_io_queues, ctrl->io_queues);
return 0;
out_destroy_queues:
@@ -494,7 +544,7 @@ static int nvme_loop_create_io_queues(struct nvme_loop_ctrl *ctrl)
return ret;
ret = nvme_alloc_io_tag_set(&ctrl->ctrl, &ctrl->tag_set,
- &nvme_loop_mq_ops, 1,
+ &nvme_loop_mq_ops, ctrl->ctrl.opts->nr_poll_queues ? 3 : 2,
sizeof(struct nvme_loop_iod) +
NVME_INLINE_SG_CNT * sizeof(struct scatterlist));
if (ret)
@@ -534,6 +584,7 @@ static struct nvme_ctrl *nvme_loop_create_ctrl(struct device *dev,
struct nvmf_ctrl_options *opts)
{
struct nvme_loop_ctrl *ctrl;
+ unsigned int nr_io_queues;
int ret;
ctrl = kzalloc(sizeof(*ctrl), GFP_KERNEL);
@@ -559,7 +610,8 @@ static struct nvme_ctrl *nvme_loop_create_ctrl(struct device *dev,
ctrl->ctrl.kato = opts->kato;
ctrl->port = nvme_loop_find_port(&ctrl->ctrl);
- ctrl->queues = kcalloc(opts->nr_io_queues + 1, sizeof(*ctrl->queues),
+ nr_io_queues = nvme_nr_io_queues(ctrl->ctrl.opts);;
+ ctrl->queues = kcalloc(nr_io_queues + 1, sizeof(*ctrl->queues),
GFP_KERNEL);
if (!ctrl->queues)
goto out_uninit_ctrl;
@@ -648,7 +700,8 @@ static struct nvmf_transport_ops nvme_loop_transport = {
.name = "loop",
.module = THIS_MODULE,
.create_ctrl = nvme_loop_create_ctrl,
- .allowed_opts = NVMF_OPT_TRADDR,
+ .allowed_opts = NVMF_OPT_TRADDR | NVMF_OPT_NR_WRITE_QUEUES |
+ NVMF_OPT_NR_POLL_QUEUES,
};
static int __init nvme_loop_init_module(void)
--
2.34.1
More information about the Linux-nvme
mailing list