[RFC-v2 03/11] nvmet: Add support for configfs-ng multi-tenant logic
Nicholas A. Bellinger
nab at linux-iscsi.org
Mon Jun 13 21:35:38 PDT 2016
From: Nicholas Bellinger <nab at linux-iscsi.org>
This patch introduces support for configfs-ng, that allows
for multi-tenant /sys/kernel/config/nvmet/subsystems/$SUBSYS_NQN/
operation, using existing /sys/kernel/config/target/core/
backends from target-core to be configfs symlinked as
per nvme-target subsystem NQN namespaces.
Here's how the layout looks:
/sys/kernel/config/nvmet/
└── subsystems
└── nqn.2003-01.org.linux-iscsi.NVMf.skylake-ep
├── namespaces
│ └── 1
│ └── ramdisk0 -> ../../../../../target/core/rd_mcp_1/ramdisk0
└── ports
└── loop
├── addr_adrfam
├── addr_portid
├── addr_traddr
├── addr_treq
├── addr_trsvcid
├── addr_trtype
└── enable
It convert nvmet_find_get_subsys to port_binding_list, and
do the same for nvmet_host_discovery_allowed.
Also convert nvmet_genctr to atomic_long_t, so it can be used
outside of nvmet_config_sem.
Cc: Jens Axboe <axboe at fb.com>
Cc: Christoph Hellwig <hch at lst.de>
Cc: Martin Petersen <martin.petersen at oracle.com>
Cc: Sagi Grimberg <sagi at grimberg.me>
Cc: Hannes Reinecke <hare at suse.de>
Cc: Mike Christie <michaelc at cs.wisc.edu>
Signed-off-by: Nicholas Bellinger <nab at linux-iscsi.org>
---
drivers/nvme/target/Makefile | 2 +-
drivers/nvme/target/configfs-ng.c | 662 ++++++++++++++++++++++++++++++++++++++
drivers/nvme/target/configfs.c | 12 +-
drivers/nvme/target/core.c | 91 ++++--
drivers/nvme/target/discovery.c | 31 +-
drivers/nvme/target/nvmet.h | 50 ++-
6 files changed, 812 insertions(+), 36 deletions(-)
create mode 100644 drivers/nvme/target/configfs-ng.c
diff --git a/drivers/nvme/target/Makefile b/drivers/nvme/target/Makefile
index b7a0623..2799e07 100644
--- a/drivers/nvme/target/Makefile
+++ b/drivers/nvme/target/Makefile
@@ -3,7 +3,7 @@ obj-$(CONFIG_NVME_TARGET) += nvmet.o
obj-$(CONFIG_NVME_TARGET_LOOP) += nvme-loop.o
obj-$(CONFIG_NVME_TARGET_RDMA) += nvmet-rdma.o
-nvmet-y += core.o configfs.o admin-cmd.o io-cmd.o fabrics-cmd.o \
+nvmet-y += core.o configfs-ng.o admin-cmd.o io-cmd.o fabrics-cmd.o \
discovery.o
nvme-loop-y += loop.o
nvmet-rdma-y += rdma.o
diff --git a/drivers/nvme/target/configfs-ng.c b/drivers/nvme/target/configfs-ng.c
new file mode 100644
index 0000000..28dc24b
--- /dev/null
+++ b/drivers/nvme/target/configfs-ng.c
@@ -0,0 +1,662 @@
+/*
+ * Based on target_core_fabric_configfs.c code
+ */
+#include <linux/kernel.h>
+#include <linux/module.h>
+#include <linux/slab.h>
+#include <linux/stat.h>
+#include <linux/ctype.h>
+#include <target/target_core_base.h>
+#include <target/target_core_backend.h>
+
+#include "nvmet.h"
+
+/*
+ * NVMf host CIT
+ */
+static void nvmet_host_release(struct config_item *item)
+{
+ struct nvmet_host *host = to_host(item);
+ struct nvmet_subsys *subsys = host->subsys;
+
+ mutex_lock(&subsys->hosts_mutex);
+ list_del_init(&host->node);
+ mutex_unlock(&subsys->hosts_mutex);
+
+ kfree(host);
+}
+
+static struct configfs_item_operations nvmet_host_item_opts = {
+ .release = nvmet_host_release,
+};
+
+static struct config_item_type nvmet_host_type = {
+ .ct_item_ops = &nvmet_host_item_opts,
+ .ct_attrs = NULL,
+ .ct_owner = THIS_MODULE,
+
+};
+
+static struct config_group *nvmet_make_hosts(struct config_group *group,
+ const char *name)
+{
+ struct nvmet_subsys *subsys = ports_to_subsys(&group->cg_item);
+ struct nvmet_host *host;
+
+ host = kzalloc(sizeof(*host), GFP_KERNEL);
+ if (!host)
+ return ERR_PTR(-ENOMEM);
+
+ INIT_LIST_HEAD(&host->node);
+ host->subsys = subsys;
+
+ mutex_lock(&subsys->hosts_mutex);
+ list_add_tail(&host->node, &subsys->hosts);
+ mutex_unlock(&subsys->hosts_mutex);
+
+ config_group_init_type_name(&host->group, name, &nvmet_host_type);
+
+ return &host->group;
+}
+
+static void nvmet_drop_hosts(struct config_group *group, struct config_item *item)
+{
+ config_item_put(item);
+}
+
+static struct configfs_group_operations nvmet_hosts_group_ops = {
+ .make_group = nvmet_make_hosts,
+ .drop_item = nvmet_drop_hosts,
+};
+
+static struct config_item_type nvmet_hosts_type = {
+ .ct_group_ops = &nvmet_hosts_group_ops,
+ .ct_item_ops = NULL,
+ .ct_attrs = NULL,
+ .ct_owner = THIS_MODULE,
+};
+
+/*
+ * nvmet_port Generic ConfigFS definitions.
+ */
+static ssize_t nvmet_port_addr_adrfam_show(struct config_item *item,
+ char *page)
+{
+ switch (to_nvmet_port_binding(item)->disc_addr.adrfam) {
+ case NVMF_ADDR_FAMILY_IP4:
+ return sprintf(page, "ipv4\n");
+ case NVMF_ADDR_FAMILY_IP6:
+ return sprintf(page, "ipv6\n");
+ case NVMF_ADDR_FAMILY_IB:
+ return sprintf(page, "ib\n");
+ default:
+ return sprintf(page, "\n");
+ }
+}
+
+static ssize_t nvmet_port_addr_adrfam_store(struct config_item *item,
+ const char *page, size_t count)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ if (pb->enabled) {
+ pr_err("Cannot modify address while enabled\n");
+ pr_err("Disable the address before modifying\n");
+ return -EACCES;
+ }
+
+ if (sysfs_streq(page, "ipv4")) {
+ pb->disc_addr.adrfam = NVMF_ADDR_FAMILY_IP4;
+ } else if (sysfs_streq(page, "ipv6")) {
+ pb->disc_addr.adrfam = NVMF_ADDR_FAMILY_IP6;
+ } else if (sysfs_streq(page, "ib")) {
+ pb->disc_addr.adrfam = NVMF_ADDR_FAMILY_IB;
+ } else {
+ pr_err("Invalid value '%s' for adrfam\n", page);
+ return -EINVAL;
+ }
+
+ return count;
+}
+
+CONFIGFS_ATTR(nvmet_port_, addr_adrfam);
+
+static ssize_t nvmet_port_addr_portid_show(struct config_item *item,
+ char *page)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ return snprintf(page, PAGE_SIZE, "%d\n",
+ le16_to_cpu(pb->disc_addr.portid));
+}
+
+static ssize_t nvmet_port_addr_portid_store(struct config_item *item,
+ const char *page, size_t count)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+ u16 portid = 0;
+
+ if (kstrtou16(page, 0, &portid)) {
+ pr_err("Invalid value '%s' for portid\n", page);
+ return -EINVAL;
+ }
+
+ if (pb->enabled) {
+ pr_err("Cannot modify address while enabled\n");
+ pr_err("Disable the address before modifying\n");
+ return -EACCES;
+ }
+ pb->disc_addr.portid = cpu_to_le16(portid);
+ return count;
+}
+
+CONFIGFS_ATTR(nvmet_port_, addr_portid);
+
+static ssize_t nvmet_port_addr_traddr_show(struct config_item *item,
+ char *page)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ return snprintf(page, PAGE_SIZE, "%s\n",
+ pb->disc_addr.traddr);
+}
+
+static ssize_t nvmet_port_addr_traddr_store(struct config_item *item,
+ const char *page, size_t count)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ if (count > NVMF_TRADDR_SIZE) {
+ pr_err("Invalid value '%s' for traddr\n", page);
+ return -EINVAL;
+ }
+
+ if (pb->enabled) {
+ pr_err("Cannot modify address while enabled\n");
+ pr_err("Disable the address before modifying\n");
+ return -EACCES;
+ }
+ return snprintf(pb->disc_addr.traddr,
+ sizeof(pb->disc_addr.traddr), "%s", page);
+}
+
+CONFIGFS_ATTR(nvmet_port_, addr_traddr);
+
+static ssize_t nvmet_port_addr_treq_show(struct config_item *item,
+ char *page)
+{
+ switch (to_nvmet_port_binding(item)->disc_addr.treq) {
+ case NVMF_TREQ_NOT_SPECIFIED:
+ return sprintf(page, "not specified\n");
+ case NVMF_TREQ_REQUIRED:
+ return sprintf(page, "required\n");
+ case NVMF_TREQ_NOT_REQUIRED:
+ return sprintf(page, "not required\n");
+ default:
+ return sprintf(page, "\n");
+ }
+}
+
+static ssize_t nvmet_port_addr_treq_store(struct config_item *item,
+ const char *page, size_t count)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ if (pb->enabled) {
+ pr_err("Cannot modify address while enabled\n");
+ pr_err("Disable the address before modifying\n");
+ return -EACCES;
+ }
+
+ if (sysfs_streq(page, "not specified")) {
+ pb->disc_addr.treq = NVMF_TREQ_NOT_SPECIFIED;
+ } else if (sysfs_streq(page, "required")) {
+ pb->disc_addr.treq = NVMF_TREQ_REQUIRED;
+ } else if (sysfs_streq(page, "not required")) {
+ pb->disc_addr.treq = NVMF_TREQ_NOT_REQUIRED;
+ } else {
+ pr_err("Invalid value '%s' for treq\n", page);
+ return -EINVAL;
+ }
+
+ return count;
+}
+
+CONFIGFS_ATTR(nvmet_port_, addr_treq);
+
+static ssize_t nvmet_port_addr_trsvcid_show(struct config_item *item,
+ char *page)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ return snprintf(page, PAGE_SIZE, "%s\n",
+ pb->disc_addr.trsvcid);
+}
+
+static ssize_t nvmet_port_addr_trsvcid_store(struct config_item *item,
+ const char *page, size_t count)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ if (count > NVMF_TRSVCID_SIZE) {
+ pr_err("Invalid value '%s' for trsvcid\n", page);
+ return -EINVAL;
+ }
+ if (pb->enabled) {
+ pr_err("Cannot modify address while enabled\n");
+ pr_err("Disable the address before modifying\n");
+ return -EACCES;
+ }
+ return snprintf(pb->disc_addr.trsvcid,
+ sizeof(pb->disc_addr.trsvcid), "%s", page);
+}
+
+CONFIGFS_ATTR(nvmet_port_, addr_trsvcid);
+
+static ssize_t nvmet_port_addr_trtype_show(struct config_item *item,
+ char *page)
+{
+ switch (to_nvmet_port_binding(item)->disc_addr.trtype) {
+ case NVMF_TRTYPE_RDMA:
+ return sprintf(page, "rdma\n");
+ case NVMF_TRTYPE_LOOP:
+ return sprintf(page, "loop\n");
+ default:
+ return sprintf(page, "\n");
+ }
+}
+
+static void nvmet_port_init_tsas_rdma(struct nvmet_port_binding *pb)
+{
+ pb->disc_addr.trtype = NVMF_TRTYPE_RDMA;
+ memset(&pb->disc_addr.tsas.rdma, 0, NVMF_TSAS_SIZE);
+ pb->disc_addr.tsas.rdma.qptype = NVMF_RDMA_QPTYPE_CONNECTED;
+ pb->disc_addr.tsas.rdma.prtype = NVMF_RDMA_PRTYPE_NOT_SPECIFIED;
+ pb->disc_addr.tsas.rdma.cms = NVMF_RDMA_CMS_RDMA_CM;
+}
+
+static void nvmet_port_init_tsas_loop(struct nvmet_port_binding *pb)
+{
+ pb->disc_addr.trtype = NVMF_TRTYPE_LOOP;
+ memset(&pb->disc_addr.tsas, 0, NVMF_TSAS_SIZE);
+}
+
+static ssize_t nvmet_port_addr_trtype_store(struct config_item *item,
+ const char *page, size_t count)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ if (pb->enabled) {
+ pr_err("Cannot modify address while enabled\n");
+ pr_err("Disable the address before modifying\n");
+ return -EACCES;
+ }
+
+ if (sysfs_streq(page, "rdma")) {
+ nvmet_port_init_tsas_rdma(pb);
+ } else if (sysfs_streq(page, "loop")) {
+ nvmet_port_init_tsas_loop(pb);
+ } else {
+ pr_err("Invalid value '%s' for trtype\n", page);
+ return -EINVAL;
+ }
+
+ return count;
+}
+
+CONFIGFS_ATTR(nvmet_port_, addr_trtype);
+
+static void nvmet_port_disable(struct nvmet_port_binding *pb)
+{
+ struct nvmet_fabrics_ops *ops = pb->nf_ops;
+ struct nvmet_port *port = pb->port;
+
+ if (!ops || !port)
+ return;
+
+ ops->remove_port(pb);
+ nvmet_put_transport(&pb->disc_addr);
+ pb->nf_ops = NULL;
+
+ atomic64_inc(&nvmet_genctr);
+}
+
+static ssize_t nvmet_port_enable_show(struct config_item *item, char *page)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ return sprintf(page, "%d\n", pb->enabled);
+}
+
+static ssize_t nvmet_port_enable_store(struct config_item *item,
+ const char *page, size_t count)
+{
+ struct nvmet_port *port;
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+ struct nvmet_fabrics_ops *ops;
+ bool enable;
+ int rc;
+
+ if (strtobool(page, &enable))
+ return -EINVAL;
+
+ if (enable) {
+ if (pb->enabled) {
+ pr_warn("port already enabled: %d\n",
+ pb->disc_addr.trtype);
+ goto out;
+ }
+
+ ops = nvmet_get_transport(&pb->disc_addr);
+ if (IS_ERR(ops))
+ return PTR_ERR(ops);
+
+ pb->nf_ops = ops;
+
+ rc = ops->add_port(pb);
+ if (rc) {
+ nvmet_put_transport(&pb->disc_addr);
+ return rc;
+ }
+
+ atomic64_inc(&nvmet_genctr);
+ } else {
+ if (!pb->nf_ops)
+ return -EINVAL;
+
+ port = pb->port;
+ if (!port)
+ return -EINVAL;
+
+ nvmet_port_disable(pb);
+ }
+out:
+ return count;
+}
+
+CONFIGFS_ATTR(nvmet_port_, enable);
+
+static struct configfs_attribute *nvmet_port_attrs[] = {
+ &nvmet_port_attr_addr_adrfam,
+ &nvmet_port_attr_addr_portid,
+ &nvmet_port_attr_addr_traddr,
+ &nvmet_port_attr_addr_treq,
+ &nvmet_port_attr_addr_trsvcid,
+ &nvmet_port_attr_addr_trtype,
+ &nvmet_port_attr_enable,
+ NULL,
+};
+
+/*
+ * NVMf transport port CIT
+ */
+static void nvmet_port_release(struct config_item *item)
+{
+ struct nvmet_port_binding *pb = to_nvmet_port_binding(item);
+
+ nvmet_port_disable(pb);
+ kfree(pb);
+}
+
+static struct configfs_item_operations nvmet_port_item_ops = {
+ .release = nvmet_port_release,
+};
+
+static struct config_item_type nvmet_port_type = {
+ .ct_item_ops = &nvmet_port_item_ops,
+ .ct_attrs = nvmet_port_attrs,
+ .ct_owner = THIS_MODULE,
+};
+
+static struct config_group *nvmet_make_ports(struct config_group *group,
+ const char *name)
+{
+ struct nvmet_subsys *subsys = ports_to_subsys(&group->cg_item);
+ struct nvmet_port_binding *pb;
+
+ printk("Entering nvmet_make_port %s >>>>>>>>>>>>>>>>>>\n", name);
+
+ pb = kzalloc(sizeof(*pb), GFP_KERNEL);
+ if (!pb)
+ return ERR_PTR(-ENOMEM);
+
+ INIT_LIST_HEAD(&pb->node);
+ pb->nf_subsys = subsys;
+
+ config_group_init_type_name(&pb->group, name, &nvmet_port_type);
+
+ return &pb->group;
+}
+
+static void nvmet_drop_ports(struct config_group *group, struct config_item *item)
+{
+ config_item_put(item);
+}
+
+static struct configfs_group_operations nvmet_ports_group_ops = {
+ .make_group = nvmet_make_ports,
+ .drop_item = nvmet_drop_ports,
+};
+
+static struct config_item_type nvmet_ports_type = {
+ .ct_group_ops = &nvmet_ports_group_ops,
+ .ct_item_ops = NULL,
+ .ct_attrs = NULL,
+ .ct_owner = THIS_MODULE,
+};
+
+/*
+ * NVMf namespace <-> /sys/kernel/config/target/core/ backend configfs symlink
+ */
+static int nvmet_ns_link(struct config_item *ns_ci, struct config_item *dev_ci)
+{
+ struct nvmet_ns *ns = to_nvmet_ns(ns_ci);
+ struct se_device *dev =
+ container_of(to_config_group(dev_ci), struct se_device, dev_group);
+
+ if (dev->dev_link_magic != SE_DEV_LINK_MAGIC) {
+ pr_err("Bad dev->dev_link_magic, not a valid se_dev_ci pointer:"
+ " %p to struct se_device: %p\n", dev_ci, dev);
+ return -EFAULT;
+ }
+
+ if (!(dev->dev_flags & DF_CONFIGURED)) {
+ pr_err("se_device not configured yet, cannot namespace link\n");
+ return -ENODEV;
+ }
+
+ if (!dev->transport->sbc_ops) {
+ pr_err("se_device does not have sbc_ops, cannot namespace link\n");
+ return -ENOSYS;
+ }
+
+ // XXX: Pass in struct se_device into nvmet_ns_enable
+ return nvmet_ns_enable(ns);
+}
+
+static int nvmet_ns_unlink(struct config_item *ns_ci, struct config_item *dev_ci)
+{
+ struct nvmet_ns *ns = to_nvmet_ns(ns_ci);
+
+ nvmet_ns_disable(ns);
+ return 0;
+}
+
+static void nvmet_ns_release(struct config_item *item)
+{
+ struct nvmet_ns *ns = to_nvmet_ns(item);
+
+ nvmet_ns_free(ns);
+}
+
+static struct configfs_item_operations nvmet_ns_item_ops = {
+ .release = nvmet_ns_release,
+ .allow_link = nvmet_ns_link,
+ .drop_link = nvmet_ns_unlink,
+};
+
+static struct config_item_type nvmet_ns_type = {
+ .ct_item_ops = &nvmet_ns_item_ops,
+ .ct_attrs = NULL,
+ .ct_owner = THIS_MODULE,
+};
+
+static struct config_group *nvmet_make_namespace(struct config_group *group,
+ const char *name)
+{
+ struct nvmet_subsys *subsys = namespaces_to_subsys(&group->cg_item);
+ struct nvmet_ns *ns;
+ int ret;
+ u32 nsid;
+
+ ret = kstrtou32(name, 0, &nsid);
+ if (ret)
+ goto out;
+
+ ret = -EINVAL;
+ if (nsid == 0 || nsid == 0xffffffff)
+ goto out;
+
+ ret = -ENOMEM;
+ ns = nvmet_ns_alloc(subsys, nsid);
+ if (!ns)
+ goto out;
+ config_group_init_type_name(&ns->group, name, &nvmet_ns_type);
+
+ pr_info("adding nsid %d to subsystem %s\n", nsid, subsys->subsysnqn);
+
+ return &ns->group;
+out:
+ return ERR_PTR(ret);
+}
+
+static void nvmet_drop_namespace(struct config_group *group, struct config_item *item)
+{
+ /*
+ * struct nvmet_ns is released via nvmet_ns_release()
+ */
+ config_item_put(item);
+}
+
+static struct configfs_group_operations nvmet_namespaces_group_ops = {
+ .make_group = nvmet_make_namespace,
+ .drop_item = nvmet_drop_namespace,
+};
+
+static struct config_item_type nvmet_namespaces_type = {
+ .ct_group_ops = &nvmet_namespaces_group_ops,
+ .ct_owner = THIS_MODULE,
+};
+
+/*
+ * Subsystem structures & folder operation functions below
+ */
+static void nvmet_subsys_release(struct config_item *item)
+{
+ struct nvmet_subsys *subsys = to_subsys(item);
+
+ nvmet_subsys_put(subsys);
+}
+
+static struct configfs_item_operations nvmet_subsys_item_ops = {
+ .release = nvmet_subsys_release,
+};
+
+static struct config_item_type nvmet_subsys_type = {
+ .ct_item_ops = &nvmet_subsys_item_ops,
+// .ct_attrs = nvmet_subsys_attrs,
+ .ct_owner = THIS_MODULE,
+};
+
+static struct config_group *nvmet_make_subsys(struct config_group *group,
+ const char *name)
+{
+ struct nvmet_subsys *subsys;
+
+ if (sysfs_streq(name, NVME_DISC_SUBSYS_NAME)) {
+ pr_err("can't create discovery subsystem through configfs\n");
+ return ERR_PTR(-EINVAL);
+ }
+
+ subsys = nvmet_subsys_alloc(name, NVME_NQN_NVME);
+ if (!subsys)
+ return ERR_PTR(-ENOMEM);
+
+ config_group_init_type_name(&subsys->group, name, &nvmet_subsys_type);
+
+ config_group_init_type_name(&subsys->namespaces_group,
+ "namespaces", &nvmet_namespaces_type);
+ configfs_add_default_group(&subsys->namespaces_group, &subsys->group);
+
+ config_group_init_type_name(&subsys->ports_group,
+ "ports", &nvmet_ports_type);
+ configfs_add_default_group(&subsys->ports_group, &subsys->group);
+
+ config_group_init_type_name(&subsys->hosts_group,
+ "hosts", &nvmet_hosts_type);
+ configfs_add_default_group(&subsys->hosts_group, &subsys->group);
+
+// XXX: subsys->allow_any_host hardcoded to true
+ subsys->allow_any_host = true;
+
+ return &subsys->group;
+}
+
+static void nvmet_drop_subsys(struct config_group *group, struct config_item *item)
+{
+ /*
+ * struct nvmet_port is releated via nvmet_subsys_release()
+ */
+ config_item_put(item);
+}
+
+static struct configfs_group_operations nvmet_subsystems_group_ops = {
+ .make_group = nvmet_make_subsys,
+ .drop_item = nvmet_drop_subsys,
+};
+
+static struct config_item_type nvmet_subsystems_type = {
+ .ct_group_ops = &nvmet_subsystems_group_ops,
+ .ct_owner = THIS_MODULE,
+};
+
+static struct config_group nvmet_subsystems_group;
+
+static struct config_item_type nvmet_root_type = {
+ .ct_owner = THIS_MODULE,
+};
+
+static struct configfs_subsystem nvmet_configfs_subsystem = {
+ .su_group = {
+ .cg_item = {
+ .ci_namebuf = "nvmet",
+ .ci_type = &nvmet_root_type,
+ },
+ },
+};
+
+int __init nvmet_init_configfs(void)
+{
+ int ret;
+
+ config_group_init(&nvmet_configfs_subsystem.su_group);
+ mutex_init(&nvmet_configfs_subsystem.su_mutex);
+
+ config_group_init_type_name(&nvmet_subsystems_group,
+ "subsystems", &nvmet_subsystems_type);
+ configfs_add_default_group(&nvmet_subsystems_group,
+ &nvmet_configfs_subsystem.su_group);
+
+ ret = configfs_register_subsystem(&nvmet_configfs_subsystem);
+ if (ret) {
+ pr_err("configfs_register_subsystem: %d\n", ret);
+ return ret;
+ }
+
+ return 0;
+}
+
+void __exit nvmet_exit_configfs(void)
+{
+ configfs_unregister_subsystem(&nvmet_configfs_subsystem);
+}
diff --git a/drivers/nvme/target/configfs.c b/drivers/nvme/target/configfs.c
index aebe646..b30896a 100644
--- a/drivers/nvme/target/configfs.c
+++ b/drivers/nvme/target/configfs.c
@@ -441,7 +441,9 @@ static int nvmet_port_subsys_allow_link(struct config_item *parent,
if (!link)
return -ENOMEM;
link->subsys = subsys;
-
+#if 1
+ BUG_ON(1);
+#else
down_write(&nvmet_config_sem);
ret = -EEXIST;
list_for_each_entry(p, &port->subsystems, entry) {
@@ -458,6 +460,7 @@ static int nvmet_port_subsys_allow_link(struct config_item *parent,
list_add_tail(&link->entry, &port->subsystems);
nvmet_genctr++;
up_write(&nvmet_config_sem);
+#endif
return 0;
out_free_link:
@@ -469,6 +472,7 @@ out_free_link:
static int nvmet_port_subsys_drop_link(struct config_item *parent,
struct config_item *target)
{
+#if 0
struct nvmet_port *port = to_nvmet_port(parent->ci_parent);
struct nvmet_subsys *subsys = to_subsys(target);
struct nvmet_subsys_link *p;
@@ -487,7 +491,9 @@ found:
if (list_empty(&port->subsystems))
nvmet_disable_port(port);
up_write(&nvmet_config_sem);
+
kfree(p);
+#endif
return 0;
}
@@ -504,6 +510,7 @@ static struct config_item_type nvmet_port_subsys_type = {
static int nvmet_allowed_hosts_allow_link(struct config_item *parent,
struct config_item *target)
{
+#if 0
struct nvmet_subsys *subsys = to_subsys(parent->ci_parent);
struct nvmet_host *host;
struct nvmet_host_link *link, *p;
@@ -540,11 +547,13 @@ out_free_link:
up_write(&nvmet_config_sem);
kfree(link);
return ret;
+#endif
}
static int nvmet_allowed_hosts_drop_link(struct config_item *parent,
struct config_item *target)
{
+#if 0
struct nvmet_subsys *subsys = to_subsys(parent->ci_parent);
struct nvmet_host *host = to_host(target);
struct nvmet_host_link *p;
@@ -563,6 +572,7 @@ found:
up_write(&nvmet_config_sem);
kfree(p);
return 0;
+#endif
}
static struct configfs_item_operations nvmet_allowed_hosts_item_ops = {
diff --git a/drivers/nvme/target/core.c b/drivers/nvme/target/core.c
index 689ad4c..3357696 100644
--- a/drivers/nvme/target/core.c
+++ b/drivers/nvme/target/core.c
@@ -147,6 +147,7 @@ EXPORT_SYMBOL_GPL(nvmet_unregister_transport);
int nvmet_enable_port(struct nvmet_port *port)
{
+#if 0
struct nvmet_fabrics_ops *ops;
int ret;
@@ -175,11 +176,13 @@ int nvmet_enable_port(struct nvmet_port *port)
}
port->enabled = true;
+#endif
return 0;
}
void nvmet_disable_port(struct nvmet_port *port)
{
+#if 0
struct nvmet_fabrics_ops *ops;
lockdep_assert_held(&nvmet_config_sem);
@@ -189,6 +192,7 @@ void nvmet_disable_port(struct nvmet_port *port)
ops = nvmet_transports[port->disc_addr.trtype];
ops->remove_port(port);
module_put(ops->owner);
+#endif
}
struct nvmet_fabrics_ops *
@@ -681,15 +685,19 @@ out:
static bool __nvmet_host_allowed(struct nvmet_subsys *subsys,
const char *hostnqn)
{
- struct nvmet_host_link *p;
+ struct nvmet_host *h;
if (subsys->allow_any_host)
return true;
- list_for_each_entry(p, &subsys->hosts, entry) {
- if (!strcmp(nvmet_host_name(p->host), hostnqn))
+ mutex_lock(&subsys->hosts_mutex);
+ list_for_each_entry(h, &subsys->hosts, node) {
+ if (!strcmp(nvmet_host_name(h), hostnqn)) {
+ mutex_unlock(&subsys->hosts_mutex);
return true;
+ }
}
+ mutex_unlock(&subsys->hosts_mutex);
return false;
}
@@ -697,10 +705,21 @@ static bool __nvmet_host_allowed(struct nvmet_subsys *subsys,
static bool nvmet_host_discovery_allowed(struct nvmet_req *req,
const char *hostnqn)
{
- struct nvmet_subsys_link *s;
+ struct nvmet_port_binding *pb;
+ struct nvmet_port *port = req->port;
+ struct nvmet_subsys *subsys;
+
+ if (!port)
+ return false;
+
+ lockdep_assert_held(&port->port_binding_mutex);
+
+ list_for_each_entry(pb, &port->port_binding_list, node) {
+ subsys = pb->nf_subsys;
+ if (!subsys)
+ continue;
- list_for_each_entry(s, &req->port->subsystems, entry) {
- if (__nvmet_host_allowed(s->subsys, hostnqn))
+ if (__nvmet_host_allowed(subsys, hostnqn))
return true;
}
@@ -710,8 +729,6 @@ static bool nvmet_host_discovery_allowed(struct nvmet_req *req,
bool nvmet_host_allowed(struct nvmet_req *req, struct nvmet_subsys *subsys,
const char *hostnqn)
{
- lockdep_assert_held(&nvmet_config_sem);
-
if (subsys->type == NVME_NQN_DISC)
return nvmet_host_discovery_allowed(req, hostnqn);
else
@@ -721,13 +738,14 @@ bool nvmet_host_allowed(struct nvmet_req *req, struct nvmet_subsys *subsys,
u16 nvmet_alloc_ctrl(const char *subsysnqn, const char *hostnqn,
struct nvmet_req *req, u32 kato, struct nvmet_ctrl **ctrlp)
{
+ struct nvmet_port *port = req->port;
struct nvmet_subsys *subsys;
struct nvmet_ctrl *ctrl;
int ret;
u16 status;
status = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR;
- subsys = nvmet_find_get_subsys(req->port, subsysnqn);
+ subsys = nvmet_find_get_subsys(port, subsysnqn);
if (!subsys) {
pr_warn("connect request for invalid subsystem %s!\n",
subsysnqn);
@@ -736,15 +754,16 @@ u16 nvmet_alloc_ctrl(const char *subsysnqn, const char *hostnqn,
}
status = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR;
- down_read(&nvmet_config_sem);
+
+ mutex_lock(&port->port_binding_mutex);
if (!nvmet_host_allowed(req, subsys, hostnqn)) {
pr_info("connect by host %s for subsystem %s not allowed\n",
hostnqn, subsysnqn);
req->rsp->result = IPO_IATTR_CONNECT_DATA(hostnqn);
- up_read(&nvmet_config_sem);
+ mutex_unlock(&port->port_binding_mutex);
goto out_put_subsystem;
}
- up_read(&nvmet_config_sem);
+ mutex_unlock(&port->port_binding_mutex);
status = NVME_SC_INTERNAL;
ctrl = kzalloc(sizeof(*ctrl), GFP_KERNEL);
@@ -872,10 +891,29 @@ void nvmet_ctrl_fatal_error(struct nvmet_ctrl *ctrl)
}
EXPORT_SYMBOL_GPL(nvmet_ctrl_fatal_error);
+void nvmet_port_binding_enable(struct nvmet_port_binding *pb, struct nvmet_port *port)
+{
+ mutex_lock(&port->port_binding_mutex);
+ pb->enabled = true;
+ list_add_tail(&pb->node, &port->port_binding_list);
+ mutex_unlock(&port->port_binding_mutex);
+}
+EXPORT_SYMBOL_GPL(nvmet_port_binding_enable);
+
+void nvmet_port_binding_disable(struct nvmet_port_binding *pb, struct nvmet_port *port)
+{
+ mutex_lock(&port->port_binding_mutex);
+ pb->enabled = false;
+ list_del_init(&pb->node);
+ mutex_unlock(&port->port_binding_mutex);
+}
+EXPORT_SYMBOL_GPL(nvmet_port_binding_disable);
+
static struct nvmet_subsys *nvmet_find_get_subsys(struct nvmet_port *port,
const char *subsysnqn)
{
- struct nvmet_subsys_link *p;
+ struct nvmet_port_binding *pb;
+ struct nvmet_subsys *subsys;
if (!port)
return NULL;
@@ -887,17 +925,22 @@ static struct nvmet_subsys *nvmet_find_get_subsys(struct nvmet_port *port,
return nvmet_disc_subsys;
}
- down_read(&nvmet_config_sem);
- list_for_each_entry(p, &port->subsystems, entry) {
- if (!strncmp(p->subsys->subsysnqn, subsysnqn,
- NVMF_NQN_SIZE)) {
- if (!kref_get_unless_zero(&p->subsys->ref))
- break;
- up_read(&nvmet_config_sem);
- return p->subsys;
+ mutex_lock(&port->port_binding_mutex);
+ list_for_each_entry(pb, &port->port_binding_list, node) {
+ subsys = pb->nf_subsys;
+ if (!subsys)
+ continue;
+
+ if (strcmp(subsys->subsysnqn, subsysnqn))
+ continue;
+
+ if (kref_get_unless_zero(&subsys->ref)) {
+ mutex_unlock(&port->port_binding_mutex);
+ return subsys;
}
}
- up_read(&nvmet_config_sem);
+ mutex_unlock(&port->port_binding_mutex);
+
return NULL;
}
@@ -935,13 +978,13 @@ struct nvmet_subsys *nvmet_subsys_alloc(const char *subsysnqn,
kref_init(&subsys->ref);
mutex_init(&subsys->lock);
+ mutex_init(&subsys->hosts_mutex);
INIT_LIST_HEAD(&subsys->namespaces);
INIT_LIST_HEAD(&subsys->ctrls);
+ INIT_LIST_HEAD(&subsys->hosts);
ida_init(&subsys->cntlid_ida);
- INIT_LIST_HEAD(&subsys->hosts);
-
return subsys;
}
diff --git a/drivers/nvme/target/discovery.c b/drivers/nvme/target/discovery.c
index 6f65646..32dc05c 100644
--- a/drivers/nvme/target/discovery.c
+++ b/drivers/nvme/target/discovery.c
@@ -18,7 +18,7 @@
struct nvmet_subsys *nvmet_disc_subsys;
-u64 nvmet_genctr;
+atomic_long_t nvmet_genctr;
void nvmet_referral_enable(struct nvmet_port *parent, struct nvmet_port *port)
{
@@ -26,7 +26,7 @@ void nvmet_referral_enable(struct nvmet_port *parent, struct nvmet_port *port)
if (list_empty(&port->entry)) {
list_add_tail(&port->entry, &parent->referrals);
port->enabled = true;
- nvmet_genctr++;
+ atomic64_inc(&nvmet_genctr);
}
up_write(&nvmet_config_sem);
}
@@ -37,7 +37,7 @@ void nvmet_referral_disable(struct nvmet_port *port)
if (!list_empty(&port->entry)) {
port->enabled = false;
list_del_init(&port->entry);
- nvmet_genctr++;
+ atomic64_inc(&nvmet_genctr);
}
up_write(&nvmet_config_sem);
}
@@ -69,8 +69,8 @@ static void nvmet_execute_get_disc_log_page(struct nvmet_req *req)
size_t data_len = nvmet_get_log_page_len(req->cmd);
size_t alloc_len = max(data_len, sizeof(*hdr));
int residual_len = data_len - sizeof(*hdr);
- struct nvmet_subsys_link *p;
- struct nvmet_port *r;
+ struct nvmet_port *port = req->port;
+ struct nvmet_port_binding *pb;
u32 numrec = 0;
u16 status = 0;
@@ -84,7 +84,7 @@ static void nvmet_execute_get_disc_log_page(struct nvmet_req *req)
status = NVME_SC_INTERNAL;
goto out;
}
-
+#if 0
down_read(&nvmet_config_sem);
list_for_each_entry(p, &req->port->subsystems, entry) {
if (!nvmet_host_allowed(req, p->subsys, ctrl->hostnqn))
@@ -113,7 +113,26 @@ static void nvmet_execute_get_disc_log_page(struct nvmet_req *req)
hdr->recfmt = cpu_to_le16(0);
up_read(&nvmet_config_sem);
+#else
+ mutex_lock(&port->port_binding_mutex);
+ list_for_each_entry(pb, &port->port_binding_list, node) {
+ if (!nvmet_host_allowed(req, pb->nf_subsys, ctrl->hostnqn))
+ continue;
+
+ if (residual_len >= entry_size) {
+ nvmet_format_discovery_entry(hdr, port,
+ pb->nf_subsys->subsysnqn,
+ NVME_NQN_NVME, numrec);
+ residual_len -= entry_size;
+ }
+ numrec++;
+ }
+ hdr->genctr = cpu_to_le64(atomic64_read(&nvmet_genctr));
+ hdr->numrec = cpu_to_le64(numrec);
+ hdr->recfmt = cpu_to_le16(0);
+ mutex_unlock(&port->port_binding_mutex);
+#endif
status = nvmet_copy_to_sgl(req, 0, hdr, data_len);
kfree(hdr);
out:
diff --git a/drivers/nvme/target/nvmet.h b/drivers/nvme/target/nvmet.h
index 17fd217..265f56f 100644
--- a/drivers/nvme/target/nvmet.h
+++ b/drivers/nvme/target/nvmet.h
@@ -79,6 +79,8 @@ struct nvmet_sq {
struct completion free_done;
};
+struct nvmet_port_binding;
+
/**
* struct nvmet_port - Common structure to keep port
* information for the target.
@@ -98,6 +100,25 @@ struct nvmet_port {
struct list_head referrals;
void *priv;
bool enabled;
+
+ struct nvmet_subsys *nf_subsys;
+ struct nvmet_fabrics_ops *nf_ops;
+
+ struct mutex port_binding_mutex;
+ struct list_head port_binding_list;
+};
+
+struct nvmet_port_binding {
+ bool enabled;
+ struct nvmf_disc_rsp_page_entry disc_addr;
+
+ struct nvmet_port *port;
+ struct nvmet_subsys *nf_subsys;
+ struct nvmet_fabrics_ops *nf_ops;
+
+ struct list_head node;
+ struct list_head subsys_node;
+ struct config_group group;
};
static inline struct nvmet_port *to_nvmet_port(struct config_item *item)
@@ -106,6 +127,13 @@ static inline struct nvmet_port *to_nvmet_port(struct config_item *item)
group);
}
+static inline struct nvmet_port_binding *
+to_nvmet_port_binding(struct config_item *item)
+{
+ return container_of(to_config_group(item), struct nvmet_port_binding,
+ group);
+}
+
struct nvmet_ctrl {
struct nvmet_subsys *subsys;
struct nvmet_cq **cqs;
@@ -147,6 +175,7 @@ struct nvmet_subsys {
struct list_head ctrls;
struct ida cntlid_ida;
+ struct mutex hosts_mutex;
struct list_head hosts;
bool allow_any_host;
@@ -158,7 +187,8 @@ struct nvmet_subsys {
struct config_group group;
struct config_group namespaces_group;
- struct config_group allowed_hosts_group;
+ struct config_group ports_group;
+ struct config_group hosts_group;
};
static inline struct nvmet_subsys *to_subsys(struct config_item *item)
@@ -173,7 +203,17 @@ static inline struct nvmet_subsys *namespaces_to_subsys(
namespaces_group);
}
+static inline struct nvmet_subsys *ports_to_subsys(
+ struct config_item *item)
+{
+ return container_of(to_config_group(item), struct nvmet_subsys,
+ ports_group);
+}
+
struct nvmet_host {
+ struct nvmet_subsys *subsys;
+
+ struct list_head node;
struct config_group group;
};
@@ -205,8 +245,8 @@ struct nvmet_fabrics_ops {
unsigned int msdbd;
bool has_keyed_sgls : 1;
void (*queue_response)(struct nvmet_req *req);
- int (*add_port)(struct nvmet_port *port);
- void (*remove_port)(struct nvmet_port *port);
+ int (*add_port)(struct nvmet_port_binding *pb);
+ void (*remove_port)(struct nvmet_port_binding *pb);
void (*delete_ctrl)(struct nvmet_ctrl *ctrl);
};
@@ -274,6 +314,8 @@ void nvmet_sq_destroy(struct nvmet_sq *sq);
int nvmet_sq_init(struct nvmet_sq *sq);
void nvmet_ctrl_fatal_error(struct nvmet_ctrl *ctrl);
+void nvmet_port_binding_enable(struct nvmet_port_binding *pb, struct nvmet_port *port);
+void nvmet_port_binding_disable(struct nvmet_port_binding *pb, struct nvmet_port *port);
void nvmet_update_cc(struct nvmet_ctrl *ctrl, u32 new);
u16 nvmet_alloc_ctrl(const char *subsysnqn, const char *hostnqn,
@@ -326,7 +368,7 @@ int __init nvmet_init_discovery(void);
void nvmet_exit_discovery(void);
extern struct nvmet_subsys *nvmet_disc_subsys;
-extern u64 nvmet_genctr;
+extern atomic_long_t nvmet_genctr;
extern struct rw_semaphore nvmet_config_sem;
bool nvmet_host_allowed(struct nvmet_req *req, struct nvmet_subsys *subsys,
--
1.9.1
More information about the Linux-nvme
mailing list