[RFC PATCH v1] Adding Support for namespace management
Naveen
naveen.n1 at samsung.com
Fri Aug 13 02:32:22 PDT 2021
This patch supports namespace management : create and delete operations.
This patch has been tested with the following command and size of image
file for unallocated namespaces is taken as 0GB. ns_create will look into
the list of unallocated namespaces and it will initialize the same and
return the nsid of the same. A new mandatory field has been added called
tnvmcap and we have ensured that the total capacity of namespace created
does not exceed tnvmcap
-device nvme-subsys,id=subsys0,tnvmcap=8
-device nvme,serial=foo,id=nvme0,subsys=subsys0
-device nvme,serial=bar,id=nvme1,subsys=subsys0
-drive id=ns1,file=ns1.img,if=none
-device nvme-ns,drive=ns1,bus=nvme0,nsid=1,zoned=false,shared=true
-drive id=ns2,file=ns2.img,if=none
-device nvme-ns,drive=ns2,bus=nvme0,nsid=2,zoned=false,shared=true
-drive id=ns3,file=ns3.img,if=none
-device nvme-ns,drive=ns3,bus=nvme0,nsid=3,zoned=false,shared=true
-drive id=ns4,file=ns4.img,if=none
-device nvme-ns,drive=ns4,bus=nvme0,nsid=4,zoned=false,shared=true
Please review and suggest if any changes are required.
Signed-off-by: Naveen Nagar <naveen.n1 at samsung.com>
Reviewed-by: Klaus Jensen <k.jensen at samsung.com>
Reviewed-by: Padmakar Kalghatgi <p.kalghatgi at samsung.com>
Reviewed-by: Gollu Appalanaidu <anaidu.gollu at samsung.com>
Reviewed-by: Jaegyu Choi <jg123.choi at samsung.com>
---
hw/nvme/ctrl.c | 250 +++++++++++++++++++++++++++++++++++++++++--
hw/nvme/ns.c | 14 ++-
hw/nvme/nvme.h | 6 +-
hw/nvme/subsys.c | 3 +
include/block/nvme.h | 18 +++-
5 files changed, 272 insertions(+), 19 deletions(-)
diff --git a/hw/nvme/ctrl.c b/hw/nvme/ctrl.c
index 6baf9e0420..4be23a3092 100644
--- a/hw/nvme/ctrl.c
+++ b/hw/nvme/ctrl.c
@@ -219,6 +219,7 @@ static const uint32_t nvme_cse_acs[256] = {
[NVME_ADM_CMD_SET_FEATURES] = NVME_CMD_EFF_CSUPP,
[NVME_ADM_CMD_GET_FEATURES] = NVME_CMD_EFF_CSUPP,
[NVME_ADM_CMD_ASYNC_EV_REQ] = NVME_CMD_EFF_CSUPP,
+ [NVME_ADM_CMD_NS_MANAGEMENT] = NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_NIC,
[NVME_ADM_CMD_NS_ATTACHMENT] = NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_NIC,
[NVME_ADM_CMD_FORMAT_NVM] = NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_LBCC,
};
@@ -4445,6 +4446,30 @@ static uint16_t nvme_identify_ctrl_csi(NvmeCtrl *n, NvmeRequest *req)
return nvme_c2h(n, id, sizeof(id), req);
}
+static uint16_t nvme_identify_ns_common(NvmeCtrl *n, NvmeRequest *req)
+{
+ NvmeIdNs id_ns = {};
+
+ id_ns.nsfeat |= (0x4 | 0x10);
+ id_ns.dpc = 0x1f;
+
+ NvmeLBAF lbaf[16] = {
+ [0] = {.ds = 9},
+ [1] = {.ds = 9, .ms = 8},
+ [2] = {.ds = 9, .ms = 16},
+ [3] = {.ds = 9, .ms = 64},
+ [4] = {.ds = 12},
+ [5] = {.ds = 12, .ms = 8},
+ [6] = {.ds = 12, .ms = 16},
+ [7] = {.ds = 12, .ms = 64},
+ };
+
+ memcpy(&id_ns.lbaf, &lbaf, sizeof(lbaf));
+ id_ns.nlbaf = 7;
+
+ return nvme_c2h(n, (uint8_t *)&id_ns, sizeof(NvmeIdNs), req);
+}
+
static uint16_t nvme_identify_ns(NvmeCtrl *n, NvmeRequest *req, bool active)
{
NvmeNamespace *ns;
@@ -4453,8 +4478,10 @@ static uint16_t nvme_identify_ns(NvmeCtrl *n, NvmeRequest *req, bool active)
trace_pci_nvme_identify_ns(nsid);
- if (!nvme_nsid_valid(n, nsid) || nsid == NVME_NSID_BROADCAST) {
+ if (!nvme_nsid_valid(n, nsid)) {
return NVME_INVALID_NSID | NVME_DNR;
+ } else if (nsid == NVME_NSID_BROADCAST) {
+ return nvme_identify_ns_common(n, req);
}
ns = nvme_ns(n, nsid);
@@ -5184,6 +5211,195 @@ static void nvme_select_iocs_ns(NvmeCtrl *n, NvmeNamespace *ns)
}
}
+static int nvme_blk_truncate(BlockBackend *blk, size_t len, Error **errp)
+{
+ int ret;
+ uint64_t perm, shared_perm;
+
+ blk_get_perm(blk, &perm, &shared_perm);
+
+ ret = blk_set_perm(blk, perm | BLK_PERM_RESIZE, shared_perm, errp);
+ if (ret < 0) {
+ return ret;
+ }
+
+ ret = blk_truncate(blk, len, false, PREALLOC_MODE_OFF, 0, errp);
+ if (ret < 0) {
+ return ret;
+ }
+
+ ret = blk_set_perm(blk, perm, shared_perm, errp);
+ if (ret < 0) {
+ return ret;
+ }
+
+ return 0;
+}
+
+static uint32_t nvme_allocate_nsid(NvmeCtrl *n)
+{
+ uint32_t nsid = 0;
+ for (int i = 1; i <= NVME_MAX_NAMESPACES; i++) {
+ if (nvme_ns(n, i) || nvme_subsys_ns(n->subsys, i)) {
+ continue;
+ }
+
+ nsid = i;
+ return nsid;
+ }
+ return nsid;
+}
+
+static uint16_t nvme_namespace_create(NvmeCtrl *n, NvmeRequest *req)
+{
+ uint32_t ret;
+ NvmeIdNs id_ns_host;
+ NvmeSubsystem *subsys = n->subsys;
+ Error *err = NULL;
+ uint8_t flbas_host;
+ uint64_t ns_size;
+ int lba_index;
+ NvmeNamespace *ns;
+ NvmeCtrl *ctrl;
+ NvmeIdNs *id_ns;
+
+ ret = nvme_h2c(n, (uint8_t *)&id_ns_host, sizeof(id_ns_host), req);
+ if (ret) {
+ return ret;
+ }
+
+ if (id_ns_host.ncap < id_ns_host.nsze) {
+ return NVME_THIN_PROVISION_NO_SUPP | NVME_DNR;
+ } else if (id_ns_host.ncap > id_ns_host.nsze) {
+ return NVME_INVALID_FIELD | NVME_DNR;
+ }
+
+ if (!id_ns_host.nsze) {
+ return NVME_INVALID_FIELD | NVME_DNR;
+ }
+
+ if (QSLIST_EMPTY(&subsys->unallocated_namespaces)) {
+ return NVME_NS_ID_UNAVAILABLE;
+ }
+
+ ns = QSLIST_FIRST(&subsys->unallocated_namespaces);
+ id_ns = &ns->id_ns;
+ flbas_host = (id_ns_host.flbas) & (0xF);
+
+ if (flbas_host > id_ns->nlbaf) {
+ return NVME_INVALID_FORMAT | NVME_DNR;
+ }
+
+ ret = nvme_ns_setup(ns, &err);
+ if (ret) {
+ return ret;
+ }
+
+ id_ns->flbas = id_ns_host.flbas;
+ id_ns->dps = id_ns_host.dps;
+ id_ns->nmic = id_ns_host.nmic;
+
+ lba_index = NVME_ID_NS_FLBAS_INDEX(id_ns->flbas);
+ ns_size = id_ns_host.nsze * ((1 << id_ns->lbaf[lba_index].ds) +
+ (id_ns->lbaf[lba_index].ms));
+ id_ns->nvmcap = ns_size;
+
+ if (ns_size > n->id_ctrl.unvmcap) {
+ return NVME_NS_INSUFF_CAP;
+ }
+
+ ret = nvme_blk_truncate(ns->blkconf.blk, id_ns->nvmcap, &err);
+ if (ret) {
+ return ret;
+ }
+
+ ns->size = blk_getlength(ns->blkconf.blk);
+ nvme_ns_init_format(ns);
+
+ ns->params.nsid = nvme_allocate_nsid(n);
+ if (!ns->params.nsid) {
+ return NVME_NS_ID_UNAVAILABLE;
+ }
+ subsys->namespaces[ns->params.nsid] = ns;
+
+ for (int cntlid = 0; cntlid < ARRAY_SIZE(n->subsys->ctrls); cntlid++) {
+ ctrl = nvme_subsys_ctrl(n->subsys, cntlid);
+ if (ctrl) {
+ ctrl->id_ctrl.unvmcap -= le64_to_cpu(ns->size);
+ }
+ }
+
+ stl_le_p(&req->cqe.result, ns->params.nsid);
+ QSLIST_REMOVE_HEAD(&subsys->unallocated_namespaces, entry);
+ return NVME_SUCCESS;
+}
+
+static void nvme_namespace_delete(NvmeCtrl *n, NvmeNamespace *ns, uint32_t nsid)
+{
+ NvmeCtrl *ctrl;
+ NvmeSubsystem *subsys = n->subsys;
+
+ subsys->namespaces[nsid] = NULL;
+ QSLIST_INSERT_HEAD(&subsys->unallocated_namespaces, ns, entry);
+
+ for (int cntlid = 0; cntlid < ARRAY_SIZE(n->subsys->ctrls); cntlid++) {
+ ctrl = nvme_subsys_ctrl(n->subsys, cntlid);
+ if (ctrl) {
+ ctrl->id_ctrl.unvmcap += le64_to_cpu(ns->size);
+ if (nvme_ns(ctrl, nsid)) {
+ nvme_detach_ns(ctrl, ns, nsid);
+ }
+ nvme_ns_attr_changed_aer(ctrl, nsid);
+ }
+ }
+}
+
+static uint16_t nvme_ns_management(NvmeCtrl *n, NvmeRequest *req)
+{
+ uint32_t dw10 = le32_to_cpu(req->cmd.cdw10);
+ uint8_t sel = dw10 & 0x7;
+ uint32_t nsid = le32_to_cpu(req->cmd.nsid);
+ NvmeNamespace *ns;
+
+ if (!n->subsys) {
+ return NVME_INVALID_FIELD | NVME_DNR;
+ }
+
+ switch (sel) {
+ case NVME_NS_MANAGEMENT_CREATE:
+ if (nsid) {
+ return NVME_INVALID_FIELD | NVME_DNR;
+ }
+ return nvme_namespace_create(n, req);
+ case NVME_NS_MANAGEMENT_DELETE:
+ if (nsid != NVME_NSID_BROADCAST && !nvme_nsid_valid(n, nsid)) {
+ return NVME_INVALID_NSID | NVME_DNR;
+ }
+ if (nsid == NVME_NSID_BROADCAST) {
+ uint32_t i;
+
+ for (i = 1; i <= NVME_MAX_NAMESPACES; i++) {
+ ns = nvme_subsys_ns(n->subsys, i);
+ if (!ns) {
+ continue;
+ }
+ nvme_namespace_delete(n, ns, i);
+ }
+ return NVME_SUCCESS;
+ } else {
+ ns = nvme_subsys_ns(n->subsys, nsid);
+ if (!ns) {
+ return NVME_INVALID_FIELD | NVME_DNR;
+ }
+ nvme_namespace_delete(n, ns, nsid);
+ return NVME_SUCCESS;
+ }
+
+ default:
+ return NVME_INVALID_FIELD | NVME_DNR;
+ }
+}
+
static uint16_t nvme_ns_attachment(NvmeCtrl *n, NvmeRequest *req)
{
NvmeNamespace *ns;
@@ -5239,22 +5455,14 @@ static uint16_t nvme_ns_attachment(NvmeCtrl *n, NvmeRequest *req)
if (!nvme_ns(ctrl, nsid)) {
return NVME_NS_NOT_ATTACHED | NVME_DNR;
}
-
- ctrl->namespaces[nsid] = NULL;
- ns->attached--;
-
- nvme_update_dmrsl(ctrl);
+ nvme_detach_ns(ctrl, ns, nsid);
}
/*
* Add namespace id to the changed namespace id list for event clearing
* via Get Log Page command.
*/
- if (!test_and_set_bit(nsid, ctrl->changed_nsids)) {
- nvme_enqueue_event(ctrl, NVME_AER_TYPE_NOTICE,
- NVME_AER_INFO_NOTICE_NS_ATTR_CHANGED,
- NVME_LOG_CHANGED_NSLIST);
- }
+ nvme_ns_attr_changed_aer(ctrl, nsid);
}
return NVME_SUCCESS;
@@ -5487,6 +5695,8 @@ static uint16_t nvme_admin_cmd(NvmeCtrl *n, NvmeRequest *req)
return nvme_get_feature(n, req);
case NVME_ADM_CMD_ASYNC_EV_REQ:
return nvme_aer(n, req);
+ case NVME_ADM_CMD_NS_MANAGEMENT:
+ return nvme_ns_management(n, req);
case NVME_ADM_CMD_NS_ATTACHMENT:
return nvme_ns_attachment(n, req);
case NVME_ADM_CMD_FORMAT_NVM:
@@ -6471,6 +6681,8 @@ static void nvme_init_ctrl(NvmeCtrl *n, PCIDevice *pci_dev)
if (n->subsys) {
id->cmic |= NVME_CMIC_MULTI_CTRL;
+ id->tnvmcap = n->subsys->params.tnvmcap * GiB;
+ id->unvmcap = n->subsys->params.tnvmcap * GiB;
}
NVME_CAP_SET_MQES(cap, 0x7ff);
@@ -6518,6 +6730,22 @@ void nvme_attach_ns(NvmeCtrl *n, NvmeNamespace *ns)
BDRV_REQUEST_MAX_BYTES / nvme_l2b(ns, 1));
}
+void nvme_detach_ns(NvmeCtrl *n, NvmeNamespace *ns, uint32_t nsid)
+{
+ n->namespaces[nsid] = NULL;
+ ns->attached--;
+ nvme_update_dmrsl(n);
+}
+
+void nvme_ns_attr_changed_aer(NvmeCtrl *n, uint32_t nsid)
+{
+ if (!test_and_set_bit(nsid, n->changed_nsids)) {
+ nvme_enqueue_event(n, NVME_AER_TYPE_NOTICE,
+ NVME_AER_INFO_NOTICE_NS_ATTR_CHANGED,
+ NVME_LOG_CHANGED_NSLIST);
+ }
+}
+
static void nvme_realize(PCIDevice *pci_dev, Error **errp)
{
NvmeCtrl *n = NVME(pci_dev);
diff --git a/hw/nvme/ns.c b/hw/nvme/ns.c
index b7cf1494e7..c2b8c82024 100644
--- a/hw/nvme/ns.c
+++ b/hw/nvme/ns.c
@@ -402,10 +402,6 @@ int nvme_ns_setup(NvmeNamespace *ns, Error **errp)
return -1;
}
- if (nvme_ns_init_blk(ns, errp)) {
- return -1;
- }
-
if (nvme_ns_init(ns, errp)) {
return -1;
}
@@ -481,6 +477,15 @@ static void nvme_ns_realize(DeviceState *dev, Error **errp)
}
}
+ if (nvme_ns_init_blk(ns, errp)) {
+ return;
+ }
+
+ if (!ns->size) {
+ QSLIST_INSERT_HEAD(&subsys->unallocated_namespaces, ns, entry);
+ return;
+ }
+
if (nvme_ns_setup(ns, errp)) {
return;
}
@@ -519,6 +524,7 @@ static void nvme_ns_realize(DeviceState *dev, Error **errp)
if (ctrl) {
nvme_attach_ns(ctrl, ns);
+ ctrl->id_ctrl.unvmcap -= le64_to_cpu(ns->size);
}
}
diff --git a/hw/nvme/nvme.h b/hw/nvme/nvme.h
index 83ffabade4..e1e4ab8a01 100644
--- a/hw/nvme/nvme.h
+++ b/hw/nvme/nvme.h
@@ -51,9 +51,11 @@ typedef struct NvmeSubsystem {
NvmeCtrl *ctrls[NVME_MAX_CONTROLLERS];
NvmeNamespace *namespaces[NVME_MAX_NAMESPACES + 1];
+ QSLIST_HEAD(, NvmeNamespace) unallocated_namespaces;
struct {
char *nqn;
+ uint64_t tnvmcap;
} params;
} NvmeSubsystem;
@@ -130,7 +132,7 @@ typedef struct NvmeNamespace {
uint16_t status;
int attached;
- QTAILQ_ENTRY(NvmeNamespace) entry;
+ QSLIST_ENTRY(NvmeNamespace) entry;
NvmeIdNsZoned *id_ns_zoned;
NvmeZone *zone_array;
@@ -495,6 +497,8 @@ static inline uint16_t nvme_cid(NvmeRequest *req)
}
void nvme_attach_ns(NvmeCtrl *n, NvmeNamespace *ns);
+void nvme_detach_ns(NvmeCtrl *n, NvmeNamespace *ns, uint32_t nsid);
+void nvme_ns_attr_changed_aer(NvmeCtrl *n, uint32_t nsid);
uint16_t nvme_bounce_data(NvmeCtrl *n, uint8_t *ptr, uint32_t len,
NvmeTxDirection dir, NvmeRequest *req);
uint16_t nvme_bounce_mdata(NvmeCtrl *n, uint8_t *ptr, uint32_t len,
diff --git a/hw/nvme/subsys.c b/hw/nvme/subsys.c
index 93c35950d6..e30bfbca81 100644
--- a/hw/nvme/subsys.c
+++ b/hw/nvme/subsys.c
@@ -53,11 +53,14 @@ static void nvme_subsys_realize(DeviceState *dev, Error **errp)
qbus_create_inplace(&subsys->bus, sizeof(NvmeBus), TYPE_NVME_BUS, dev,
dev->id);
+ QSLIST_INIT(&subsys->unallocated_namespaces);
+
nvme_subsys_setup(subsys);
}
static Property nvme_subsystem_props[] = {
DEFINE_PROP_STRING("nqn", NvmeSubsystem, params.nqn),
+ DEFINE_PROP_SIZE("tnvmcap", NvmeSubsystem, params.tnvmcap, 0),
DEFINE_PROP_END_OF_LIST(),
};
diff --git a/include/block/nvme.h b/include/block/nvme.h
index 77aae01174..8a8ca6753b 100644
--- a/include/block/nvme.h
+++ b/include/block/nvme.h
@@ -592,6 +592,7 @@ enum NvmeAdminCommands {
NVME_ADM_CMD_SET_FEATURES = 0x09,
NVME_ADM_CMD_GET_FEATURES = 0x0a,
NVME_ADM_CMD_ASYNC_EV_REQ = 0x0c,
+ NVME_ADM_CMD_NS_MANAGEMENT = 0x0d,
NVME_ADM_CMD_ACTIVATE_FW = 0x10,
NVME_ADM_CMD_DOWNLOAD_FW = 0x11,
NVME_ADM_CMD_NS_ATTACHMENT = 0x15,
@@ -882,6 +883,9 @@ enum NvmeStatusCodes {
NVME_FEAT_NOT_CHANGEABLE = 0x010e,
NVME_FEAT_NOT_NS_SPEC = 0x010f,
NVME_FW_REQ_SUSYSTEM_RESET = 0x0110,
+ NVME_NS_INSUFF_CAP = 0x0115,
+ NVME_THIN_PROVISION_NO_SUPP = 0x011b,
+ NVME_NS_ID_UNAVAILABLE = 0x0116,
NVME_NS_ALREADY_ATTACHED = 0x0118,
NVME_NS_PRIVATE = 0x0119,
NVME_NS_NOT_ATTACHED = 0x011a,
@@ -1056,8 +1060,10 @@ typedef struct QEMU_PACKED NvmeIdCtrl {
uint16_t mtfa;
uint32_t hmpre;
uint32_t hmmin;
- uint8_t tnvmcap[16];
- uint8_t unvmcap[16];
+ uint64_t tnvmcap;
+ uint64_t tnvmcap_u;
+ uint64_t unvmcap;
+ uint64_t unvmcap_u;
uint32_t rpmbs;
uint16_t edstt;
uint8_t dsto;
@@ -1154,6 +1160,11 @@ enum NvmeIdCtrlCmic {
NVME_CMIC_MULTI_CTRL = 1 << 1,
};
+enum NvmeNsManagementOperation {
+ NVME_NS_MANAGEMENT_CREATE = 0x0,
+ NVME_NS_MANAGEMENT_DELETE = 0x1,
+};
+
#define NVME_CTRL_SQES_MIN(sqes) ((sqes) & 0xf)
#define NVME_CTRL_SQES_MAX(sqes) (((sqes) >> 4) & 0xf)
#define NVME_CTRL_CQES_MIN(cqes) ((cqes) & 0xf)
@@ -1285,7 +1296,8 @@ typedef struct QEMU_PACKED NvmeIdNs {
uint16_t nabo;
uint16_t nabspf;
uint16_t noiob;
- uint8_t nvmcap[16];
+ uint64_t nvmcap;
+ uint64_t nvmcap_u;
uint16_t npwg;
uint16_t npwa;
uint16_t npdg;
--
2.17.1
More information about the Linux-nvme
mailing list