/* * Copyright(c) 2013-2015 Intel Corporation. All rights reserved. * * This program is free software; you can redistribute it and/or modify * it under the terms of version 2 of the GNU General Public License as * published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License for more details. */ #include #include #include #include #include "nd-core.h" #include "nd.h" static void namespace_io_release(struct device *dev) { struct nd_namespace_io *nsio = to_nd_namespace_io(dev); kfree(nsio); } static void namespace_pmem_release(struct device *dev) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); kfree(nspm->alt_name); kfree(nspm->uuid); kfree(nspm); } static void namespace_blk_release(struct device *dev) { /* TODO: blk namespace support */ } static struct device_type namespace_io_device_type = { .name = "nd_namespace_io", .release = namespace_io_release, }; static struct device_type namespace_pmem_device_type = { .name = "nd_namespace_pmem", .release = namespace_pmem_release, }; static struct device_type namespace_blk_device_type = { .name = "nd_namespace_blk", .release = namespace_blk_release, }; static bool is_namespace_pmem(struct device *dev) { return dev ? dev->type == &namespace_pmem_device_type : false; } static bool is_namespace_blk(struct device *dev) { return dev ? dev->type == &namespace_blk_device_type : false; } static bool is_namespace_io(struct device *dev) { return dev ? dev->type == &namespace_io_device_type : false; } static ssize_t nstype_show(struct device *dev, struct device_attribute *attr, char *buf) { struct nd_region *nd_region = to_nd_region(dev->parent); return sprintf(buf, "%d\n", nd_region_to_nstype(nd_region)); } static DEVICE_ATTR_RO(nstype); static ssize_t __alt_name_store(struct device *dev, const char *buf, const size_t len) { char *input, *pos, *alt_name, **ns_altname; ssize_t rc; if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); ns_altname = &nspm->alt_name; } else if (is_namespace_blk(dev)) { /* TODO: blk namespace support */ return -ENXIO; } else return -ENXIO; if (dev->driver) return -EBUSY; input = kmemdup(buf, len + 1, GFP_KERNEL); if (!input) return -ENOMEM; input[len] = '\0'; pos = strim(input); if (strlen(pos) + 1 > NSLABEL_NAME_LEN) { rc = -EINVAL; goto out; } alt_name = kzalloc(NSLABEL_NAME_LEN, GFP_KERNEL); if (!alt_name) { rc = -ENOMEM; goto out; } kfree(*ns_altname); *ns_altname = alt_name; sprintf(*ns_altname, "%s", pos); rc = len; out: kfree(input); return rc; } static ssize_t alt_name_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t len) { ssize_t rc; device_lock(dev); nvdimm_bus_lock(dev); wait_nvdimm_bus_probe_idle(dev); rc = __alt_name_store(dev, buf, len); dev_dbg(dev, "%s: %s(%zd)\n", __func__, rc < 0 ? "fail " : "", rc); nvdimm_bus_unlock(dev); device_unlock(dev); return rc; } static ssize_t alt_name_show(struct device *dev, struct device_attribute *attr, char *buf) { char *ns_altname; if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); ns_altname = nspm->alt_name; } else if (is_namespace_blk(dev)) { /* TODO: blk namespace support */ return -ENXIO; } else return -ENXIO; return sprintf(buf, "%s\n", ns_altname ? ns_altname : ""); } static DEVICE_ATTR_RW(alt_name); static int scan_free(struct nd_region *nd_region, struct nd_mapping *nd_mapping, struct nd_label_id *label_id, resource_size_t n) { bool is_blk = strncmp(label_id->id, "blk", 3) == 0; struct nvdimm_drvdata *ndd = to_ndd(nd_mapping); int rc = 0; while (n) { struct resource *res, *last; resource_size_t new_start; last = NULL; for_each_dpa_resource(ndd, res) if (strcmp(res->name, label_id->id) == 0) last = res; res = last; if (!res) return 0; if (n >= resource_size(res)) { n -= resource_size(res); nd_dbg_dpa(nd_region, ndd, res, "delete %d\n", rc); nvdimm_free_dpa(ndd, res); /* retry with last resource deleted */ continue; } /* * Keep BLK allocations relegated to high DPA as much as * possible */ if (is_blk) new_start = res->start + n; else new_start = res->start; rc = adjust_resource(res, new_start, resource_size(res) - n); nd_dbg_dpa(nd_region, ndd, res, "shrink %d\n", rc); break; } return rc; } /** * shrink_dpa_allocation - for each dimm in region free n bytes for label_id * @nd_region: the set of dimms to reclaim @n bytes from * @label_id: unique identifier for the namespace consuming this dpa range * @n: number of bytes per-dimm to release * * Assumes resources are ordered. Starting from the end try to * adjust_resource() the allocation to @n, but if @n is larger than the * allocation delete it and find the 'new' last allocation in the label * set. */ static int shrink_dpa_allocation(struct nd_region *nd_region, struct nd_label_id *label_id, resource_size_t n) { int i; for (i = 0; i < nd_region->ndr_mappings; i++) { struct nd_mapping *nd_mapping = &nd_region->mapping[i]; int rc; rc = scan_free(nd_region, nd_mapping, label_id, n); if (rc) return rc; } return 0; } static resource_size_t init_dpa_allocation(struct nd_label_id *label_id, struct nd_region *nd_region, struct nd_mapping *nd_mapping, resource_size_t n) { bool is_blk = strncmp(label_id->id, "blk", 3) == 0; struct nvdimm_drvdata *ndd = to_ndd(nd_mapping); resource_size_t first_dpa; struct resource *res; int rc = 0; /* allocate blk from highest dpa first */ if (is_blk) first_dpa = nd_mapping->start + nd_mapping->size - n; else first_dpa = nd_mapping->start; /* first resource allocation for this label-id or dimm */ res = nvdimm_allocate_dpa(ndd, label_id, first_dpa, n); if (!res) rc = -EBUSY; nd_dbg_dpa(nd_region, ndd, res, "init %d\n", rc); return rc ? n : 0; } static bool space_valid(bool is_pmem, struct nd_label_id *label_id, struct resource *res) { /* * For BLK-space any space is valid, for PMEM-space, it must be * contiguous with an existing allocation. */ if (!is_pmem) return true; if (!res || strcmp(res->name, label_id->id) == 0) return true; return false; } enum alloc_loc { ALLOC_ERR = 0, ALLOC_BEFORE, ALLOC_MID, ALLOC_AFTER, }; static resource_size_t scan_allocate(struct nd_region *nd_region, struct nd_mapping *nd_mapping, struct nd_label_id *label_id, resource_size_t n) { resource_size_t mapping_end = nd_mapping->start + nd_mapping->size - 1; bool is_pmem = strncmp(label_id->id, "pmem", 4) == 0; struct nvdimm_drvdata *ndd = to_ndd(nd_mapping); const resource_size_t to_allocate = n; struct resource *res; int first; retry: first = 0; for_each_dpa_resource(ndd, res) { resource_size_t allocate, available = 0, free_start, free_end; struct resource *next = res->sibling, *new_res = NULL; enum alloc_loc loc = ALLOC_ERR; const char *action; int rc = 0; /* ignore resources outside this nd_mapping */ if (res->start > mapping_end) continue; if (res->end < nd_mapping->start) continue; /* space at the beginning of the mapping */ if (!first++ && res->start > nd_mapping->start) { free_start = nd_mapping->start; available = res->start - free_start; if (space_valid(is_pmem, label_id, NULL)) loc = ALLOC_BEFORE; } /* space between allocations */ if (!loc && next) { free_start = res->start + resource_size(res); free_end = min(mapping_end, next->start - 1); if (space_valid(is_pmem, label_id, res) && free_start < free_end) { available = free_end + 1 - free_start; loc = ALLOC_MID; } } /* space at the end of the mapping */ if (!loc && !next) { free_start = res->start + resource_size(res); free_end = mapping_end; if (space_valid(is_pmem, label_id, res) && free_start < free_end) { available = free_end + 1 - free_start; loc = ALLOC_AFTER; } } if (!loc || !available) continue; allocate = min(available, n); switch (loc) { case ALLOC_BEFORE: if (strcmp(res->name, label_id->id) == 0) { /* adjust current resource up */ if (is_pmem) return n; rc = adjust_resource(res, res->start - allocate, resource_size(res) + allocate); action = "cur grow up"; } else action = "allocate"; break; case ALLOC_MID: if (strcmp(next->name, label_id->id) == 0) { /* adjust next resource up */ if (is_pmem) return n; rc = adjust_resource(next, next->start - allocate, resource_size(next) + allocate); new_res = next; action = "next grow up"; } else if (strcmp(res->name, label_id->id) == 0) { action = "grow down"; } else action = "allocate"; break; case ALLOC_AFTER: if (strcmp(res->name, label_id->id) == 0) action = "grow down"; else action = "allocate"; break; default: return n; } if (strcmp(action, "allocate") == 0) { /* BLK allocate bottom up */ if (!is_pmem) free_start += available - allocate; else if (free_start != nd_mapping->start) return n; new_res = nvdimm_allocate_dpa(ndd, label_id, free_start, allocate); if (!new_res) rc = -EBUSY; } else if (strcmp(action, "grow down") == 0) { /* adjust current resource down */ rc = adjust_resource(res, res->start, resource_size(res) + allocate); } if (!new_res) new_res = res; nd_dbg_dpa(nd_region, ndd, new_res, "%s(%d) %d\n", action, loc, rc); if (rc) return n; n -= allocate; if (n) { /* * Retry scan with newly inserted resources. * For example, if we did an ALLOC_BEFORE * insertion there may also have been space * available for an ALLOC_AFTER insertion, so we * need to check this same resource again */ goto retry; } else return 0; } if (is_pmem && n == to_allocate) return init_dpa_allocation(label_id, nd_region, nd_mapping, n); return n; } /** * grow_dpa_allocation - for each dimm allocate n bytes for @label_id * @nd_region: the set of dimms to allocate @n more bytes from * @label_id: unique identifier for the namespace consuming this dpa range * @n: number of bytes per-dimm to add to the existing allocation * * Assumes resources are ordered. For BLK regions, first consume * BLK-only available DPA free space, then consume PMEM-aliased DPA * space starting at the highest DPA. For PMEM regions start * allocations from the start of an interleave set and end at the first * BLK allocation or the end of the interleave set, whichever comes * first. */ static int grow_dpa_allocation(struct nd_region *nd_region, struct nd_label_id *label_id, resource_size_t n) { int i; for (i = 0; i < nd_region->ndr_mappings; i++) { struct nd_mapping *nd_mapping = &nd_region->mapping[i]; int rc; rc = scan_allocate(nd_region, nd_mapping, label_id, n); if (rc) return rc; } return 0; } static void nd_namespace_pmem_set_size(struct nd_region *nd_region, struct nd_namespace_pmem *nspm, resource_size_t size) { struct resource *res = &nspm->nsio.res; res->start = nd_region->ndr_start; res->end = nd_region->ndr_start + size - 1; } static ssize_t __size_store(struct device *dev, unsigned long long val) { resource_size_t allocated = 0, available = 0; struct nd_region *nd_region = to_nd_region(dev->parent); struct nd_mapping *nd_mapping; struct nvdimm_drvdata *ndd; struct nd_label_id label_id; u32 flags = 0, remainder; u8 *uuid = NULL; int rc, i; if (dev->driver) return -EBUSY; if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); uuid = nspm->uuid; } else if (is_namespace_blk(dev)) { /* TODO: blk namespace support */ return -ENXIO; } /* * We need a uuid for the allocation-label and dimm(s) on which * to store the label. */ if (!uuid || nd_region->ndr_mappings == 0) return -ENXIO; div_u64_rem(val, SZ_4K * nd_region->ndr_mappings, &remainder); if (remainder) { dev_dbg(dev, "%llu is not %dK aligned\n", val, (SZ_4K * nd_region->ndr_mappings) / SZ_1K); return -EINVAL; } nd_label_gen_id(&label_id, uuid, flags); for (i = 0; i < nd_region->ndr_mappings; i++) { nd_mapping = &nd_region->mapping[i]; ndd = to_ndd(nd_mapping); /* * All dimms in an interleave set, or the base dimm for a blk * region, need to be enabled for the size to be changed. */ if (!ndd) return -ENXIO; allocated += nvdimm_allocated_dpa(ndd, &label_id); } available = nd_region_available_dpa(nd_region); if (val > available + allocated) return -ENOSPC; if (val == allocated) return 0; val = div_u64(val, nd_region->ndr_mappings); allocated = div_u64(allocated, nd_region->ndr_mappings); if (val < allocated) rc = shrink_dpa_allocation(nd_region, &label_id, allocated - val); else rc = grow_dpa_allocation(nd_region, &label_id, val - allocated); if (rc) return rc; if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); nd_namespace_pmem_set_size(nd_region, nspm, val * nd_region->ndr_mappings); } return rc; } static ssize_t size_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t len) { unsigned long long val; u8 **uuid = NULL; int rc; rc = kstrtoull(buf, 0, &val); if (rc) return rc; device_lock(dev); nvdimm_bus_lock(dev); wait_nvdimm_bus_probe_idle(dev); rc = __size_store(dev, val); if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); uuid = &nspm->uuid; } else if (is_namespace_blk(dev)) { /* TODO: blk namespace support */ rc = -ENXIO; } if (rc == 0 && val == 0 && uuid) { /* setting size zero == 'delete namespace' */ kfree(*uuid); *uuid = NULL; } dev_dbg(dev, "%s: %llx %s (%d)\n", __func__, val, rc < 0 ? "fail" : "success", rc); nvdimm_bus_unlock(dev); device_unlock(dev); return rc ? rc : len; } static ssize_t size_show(struct device *dev, struct device_attribute *attr, char *buf) { if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); return sprintf(buf, "%llu\n", (unsigned long long) resource_size(&nspm->nsio.res)); } else if (is_namespace_blk(dev)) { /* TODO: blk namespace support */ return -ENXIO; } else if (is_namespace_io(dev)) { struct nd_namespace_io *nsio = to_nd_namespace_io(dev); return sprintf(buf, "%llu\n", (unsigned long long) resource_size(&nsio->res)); } else return -ENXIO; } static DEVICE_ATTR(size, S_IRUGO, size_show, size_store); static ssize_t uuid_show(struct device *dev, struct device_attribute *attr, char *buf) { u8 *uuid; if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); uuid = nspm->uuid; } else if (is_namespace_blk(dev)) { /* TODO: blk namespace support */ return -ENXIO; } else return -ENXIO; if (uuid) return sprintf(buf, "%pUb\n", uuid); return sprintf(buf, "\n"); } /** * namespace_update_uuid - check for a unique uuid and whether we're "renaming" * @nd_region: parent region so we can updates all dimms in the set * @dev: namespace type for generating label_id * @new_uuid: incoming uuid * @old_uuid: reference to the uuid storage location in the namespace object */ static int namespace_update_uuid(struct nd_region *nd_region, struct device *dev, u8 *new_uuid, u8 **old_uuid) { u32 flags = is_namespace_blk(dev) ? NSLABEL_FLAG_LOCAL : 0; struct nd_label_id old_label_id; struct nd_label_id new_label_id; int i, rc; rc = nd_is_uuid_unique(dev, new_uuid) ? 0 : -EINVAL; if (rc) { kfree(new_uuid); return rc; } if (*old_uuid == NULL) goto out; nd_label_gen_id(&old_label_id, *old_uuid, flags); nd_label_gen_id(&new_label_id, new_uuid, flags); for (i = 0; i < nd_region->ndr_mappings; i++) { struct nd_mapping *nd_mapping = &nd_region->mapping[i]; struct nvdimm_drvdata *ndd = to_ndd(nd_mapping); struct resource *res; for_each_dpa_resource(ndd, res) if (strcmp(res->name, old_label_id.id) == 0) sprintf((void *) res->name, "%s", new_label_id.id); } kfree(*old_uuid); out: *old_uuid = new_uuid; return 0; } static ssize_t uuid_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t len) { struct nd_region *nd_region = to_nd_region(dev->parent); u8 *uuid = NULL; u8 **ns_uuid; ssize_t rc; if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); ns_uuid = &nspm->uuid; } else if (is_namespace_blk(dev)) { /* TODO: blk namespace support */ return -ENXIO; } else return -ENXIO; device_lock(dev); nvdimm_bus_lock(dev); wait_nvdimm_bus_probe_idle(dev); rc = nd_uuid_store(dev, &uuid, buf, len); if (rc >= 0) rc = namespace_update_uuid(nd_region, dev, uuid, ns_uuid); dev_dbg(dev, "%s: result: %zd wrote: %s%s", __func__, rc, buf, buf[len - 1] == '\n' ? "" : "\n"); nvdimm_bus_unlock(dev); device_unlock(dev); return rc ? rc : len; } static DEVICE_ATTR_RW(uuid); static ssize_t resource_show(struct device *dev, struct device_attribute *attr, char *buf) { struct resource *res; if (is_namespace_pmem(dev)) { struct nd_namespace_pmem *nspm = to_nd_namespace_pmem(dev); res = &nspm->nsio.res; } else if (is_namespace_io(dev)) { struct nd_namespace_io *nsio = to_nd_namespace_io(dev); res = &nsio->res; } else return -ENXIO; /* no address to convey if the namespace has no allocation */ if (resource_size(res) == 0) return -ENXIO; return sprintf(buf, "%#llx\n", (unsigned long long) res->start); } static DEVICE_ATTR_RO(resource); static struct attribute *nd_namespace_attributes[] = { &dev_attr_nstype.attr, &dev_attr_size.attr, &dev_attr_uuid.attr, &dev_attr_resource.attr, &dev_attr_alt_name.attr, NULL, }; static umode_t namespace_visible(struct kobject *kobj, struct attribute *a, int n) { struct device *dev = container_of(kobj, struct device, kobj); if (a == &dev_attr_resource.attr) { if (is_namespace_blk(dev)) return 0; return a->mode; } if (is_namespace_pmem(dev) || is_namespace_blk(dev)) { if (a == &dev_attr_size.attr) return S_IWUSR | S_IRUGO; return a->mode; } if (a == &dev_attr_nstype.attr || a == &dev_attr_size.attr) return a->mode; return 0; } static struct attribute_group nd_namespace_attribute_group = { .attrs = nd_namespace_attributes, .is_visible = namespace_visible, }; static const struct attribute_group *nd_namespace_attribute_groups[] = { &nd_device_attribute_group, &nd_namespace_attribute_group, NULL, }; static struct device **create_namespace_io(struct nd_region *nd_region) { struct nd_namespace_io *nsio; struct device *dev, **devs; struct resource *res; nsio = kzalloc(sizeof(*nsio), GFP_KERNEL); if (!nsio) return NULL; devs = kcalloc(2, sizeof(struct device *), GFP_KERNEL); if (!devs) { kfree(nsio); return NULL; } dev = &nsio->dev; dev->type = &namespace_io_device_type; dev->parent = &nd_region->dev; res = &nsio->res; res->name = dev_name(&nd_region->dev); res->flags = IORESOURCE_MEM; res->start = nd_region->ndr_start; res->end = res->start + nd_region->ndr_size - 1; devs[0] = dev; return devs; } static bool has_uuid_at_pos(struct nd_region *nd_region, u8 *uuid, u64 cookie, u16 pos) { struct nd_namespace_label *found = NULL; int i; for (i = 0; i < nd_region->ndr_mappings; i++) { struct nd_mapping *nd_mapping = &nd_region->mapping[i]; struct nd_namespace_label *nd_label; bool found_uuid = false; int l; for_each_label(l, nd_label, nd_mapping->labels) { u64 isetcookie = __le64_to_cpu(nd_label->isetcookie); u16 position = __le16_to_cpu(nd_label->position); u16 nlabel = __le16_to_cpu(nd_label->nlabel); if (isetcookie != cookie) continue; if (memcmp(nd_label->uuid, uuid, NSLABEL_UUID_LEN) != 0) continue; if (found_uuid) { dev_dbg(to_ndd(nd_mapping)->dev, "%s duplicate entry for uuid\n", __func__); return false; } found_uuid = true; if (nlabel != nd_region->ndr_mappings) continue; if (position != pos) continue; found = nd_label; break; } if (found) break; } return found != NULL; } static int select_pmem_id(struct nd_region *nd_region, u8 *pmem_id) { struct nd_namespace_label *select = NULL; int i; if (!pmem_id) return -ENODEV; for (i = 0; i < nd_region->ndr_mappings; i++) { struct nd_mapping *nd_mapping = &nd_region->mapping[i]; struct nd_namespace_label *nd_label; u64 hw_start, hw_end, pmem_start, pmem_end; int l; for_each_label(l, nd_label, nd_mapping->labels) if (memcmp(nd_label->uuid, pmem_id, NSLABEL_UUID_LEN) == 0) break; if (!nd_label) { WARN_ON(1); return -EINVAL; } select = nd_label; /* * Check that this label is compliant with the dpa * range published in NFIT */ hw_start = nd_mapping->start; hw_end = hw_start + nd_mapping->size; pmem_start = __le64_to_cpu(select->dpa); pmem_end = pmem_start + __le64_to_cpu(select->rawsize); if (pmem_start == hw_start && pmem_end <= hw_end) /* pass */; else return -EINVAL; nd_mapping->labels[0] = select; nd_mapping->labels[1] = NULL; } return 0; } /** * find_pmem_label_set - validate interleave set labelling, retrieve label0 * @nd_region: region with mappings to validate */ static int find_pmem_label_set(struct nd_region *nd_region, struct nd_namespace_pmem *nspm) { u64 cookie = nd_region_interleave_set_cookie(nd_region); struct nd_namespace_label *nd_label; u8 select_id[NSLABEL_UUID_LEN]; resource_size_t size = 0; u8 *pmem_id = NULL; int rc = -ENODEV, l; u16 i; if (cookie == 0) return -ENXIO; /* * Find a complete set of labels by uuid. By definition we can start * with any mapping as the reference label */ for_each_label(l, nd_label, nd_region->mapping[0].labels) { u64 isetcookie = __le64_to_cpu(nd_label->isetcookie); if (isetcookie != cookie) continue; for (i = 0; nd_region->ndr_mappings; i++) if (!has_uuid_at_pos(nd_region, nd_label->uuid, cookie, i)) break; if (i < nd_region->ndr_mappings) { /* * Give up if we don't find an instance of a * uuid at each position (from 0 to * nd_region->ndr_mappings - 1), or if we find a * dimm with two instances of the same uuid. */ rc = -EINVAL; goto err; } else if (pmem_id) { /* * If there is more than one valid uuid set, we * need userspace to clean this up. */ rc = -EBUSY; goto err; } memcpy(select_id, nd_label->uuid, NSLABEL_UUID_LEN); pmem_id = select_id; } /* * Fix up each mapping's 'labels' to have the validated pmem label for * that position at labels[0], and NULL at labels[1]. In the process, * check that the namespace aligns with interleave-set. We know * that it does not overlap with any blk namespaces by virtue of * the dimm being enabled (i.e. nd_label_reserve_dpa() * succeeded). */ rc = select_pmem_id(nd_region, pmem_id); if (rc) goto err; /* Calculate total size and populate namespace properties from label0 */ for (i = 0; i < nd_region->ndr_mappings; i++) { struct nd_mapping *nd_mapping = &nd_region->mapping[i]; struct nd_namespace_label *label0 = nd_mapping->labels[0]; size += __le64_to_cpu(label0->rawsize); if (__le16_to_cpu(label0->position) != 0) continue; WARN_ON(nspm->alt_name || nspm->uuid); nspm->alt_name = kmemdup((void __force *) label0->name, NSLABEL_NAME_LEN, GFP_KERNEL); nspm->uuid = kmemdup((void __force *) label0->uuid, NSLABEL_UUID_LEN, GFP_KERNEL); } if (!nspm->alt_name || !nspm->uuid) { rc = -ENOMEM; goto err; } nd_namespace_pmem_set_size(nd_region, nspm, size); return 0; err: switch (rc) { case -EINVAL: dev_dbg(&nd_region->dev, "%s: invalid label(s)\n", __func__); break; case -ENODEV: dev_dbg(&nd_region->dev, "%s: label not found\n", __func__); break; default: dev_dbg(&nd_region->dev, "%s: unexpected err: %d\n", __func__, rc); break; } return rc; } static struct device **create_namespace_pmem(struct nd_region *nd_region) { struct nd_namespace_pmem *nspm; struct device *dev, **devs; struct resource *res; int rc; nspm = kzalloc(sizeof(*nspm), GFP_KERNEL); if (!nspm) return NULL; dev = &nspm->nsio.dev; dev->type = &namespace_pmem_device_type; dev->parent = &nd_region->dev; res = &nspm->nsio.res; res->name = dev_name(&nd_region->dev); res->flags = IORESOURCE_MEM; rc = find_pmem_label_set(nd_region, nspm); if (rc == -ENODEV) { int i; /* Pass, try to permit namespace creation... */ for (i = 0; i < nd_region->ndr_mappings; i++) { struct nd_mapping *nd_mapping = &nd_region->mapping[i]; kfree(nd_mapping->labels); nd_mapping->labels = NULL; } /* Publish a zero-sized namespace for userspace to configure. */ nd_namespace_pmem_set_size(nd_region, nspm, 0); rc = 0; } else if (rc) goto err; devs = kcalloc(2, sizeof(struct device *), GFP_KERNEL); if (!devs) goto err; devs[0] = dev; return devs; err: namespace_pmem_release(&nspm->nsio.dev); return NULL; } static int init_active_labels(struct nd_region *nd_region) { int i; for (i = 0; i < nd_region->ndr_mappings; i++) { struct nd_mapping *nd_mapping = &nd_region->mapping[i]; struct nvdimm_drvdata *ndd = to_ndd(nd_mapping); struct nvdimm *nvdimm = nd_mapping->nvdimm; int count, j; /* * If the dimm is disabled then prevent the region from * being activated if it aliases DPA. */ if (!ndd) { if ((nvdimm->flags & NDD_ALIASING) == 0) return 0; dev_dbg(&nd_region->dev, "%s: is disabled, failing probe\n", dev_name(&nd_mapping->nvdimm->dev)); return -ENXIO; } nd_mapping->ndd = ndd; atomic_inc(&nvdimm->busy); get_ndd(ndd); count = nd_label_active_count(ndd); dev_dbg(ndd->dev, "%s: %d\n", __func__, count); if (!count) continue; nd_mapping->labels = kcalloc(count + 1, sizeof(void *), GFP_KERNEL); if (!nd_mapping->labels) return -ENOMEM; for (j = 0; j < count; j++) { struct nd_namespace_label *label; label = nd_label_active(ndd, j); nd_mapping->labels[j] = label; } } return 0; } int nd_region_register_namespaces(struct nd_region *nd_region, int *err) { struct device **devs = NULL; int i, rc = 0, type; *err = 0; nvdimm_bus_lock(&nd_region->dev); rc = init_active_labels(nd_region); if (rc) { nvdimm_bus_unlock(&nd_region->dev); return rc; } type = nd_region_to_nstype(nd_region); switch (type) { case ND_DEVICE_NAMESPACE_IO: devs = create_namespace_io(nd_region); break; case ND_DEVICE_NAMESPACE_PMEM: devs = create_namespace_pmem(nd_region); break; default: break; } nvdimm_bus_unlock(&nd_region->dev); if (!devs) return -ENODEV; nd_region->ns_seed = devs[0]; for (i = 0; devs[i]; i++) { struct device *dev = devs[i]; dev_set_name(dev, "namespace%d.%d", nd_region->id, i); dev->groups = nd_namespace_attribute_groups; nd_device_register(dev); } kfree(devs); return i; }