2018-11-24 19:00:22

by Dan Williams

[permalink] [raw]
Subject: [PATCH 1/2] tools/testing/nvdimm: Align test resources to 128M

In preparation for libnvdimm growing new restrictions to detect section
conflicts between persistent memory regions, enable nfit_test to
allocate aligned resources. Use a gen_pool to allocate nfit_test's fake
resources in a separate address space from the virtual translation of
the same.

Signed-off-by: Dan Williams <[email protected]>
---
tools/testing/nvdimm/test/nfit.c | 36 ++++++++++++++++++++++++++++++++++--
1 file changed, 34 insertions(+), 2 deletions(-)

diff --git a/tools/testing/nvdimm/test/nfit.c b/tools/testing/nvdimm/test/nfit.c
index 01ec04bf91b5..ca4e61c864d5 100644
--- a/tools/testing/nvdimm/test/nfit.c
+++ b/tools/testing/nvdimm/test/nfit.c
@@ -15,6 +15,7 @@
#include <linux/dma-mapping.h>
#include <linux/workqueue.h>
#include <linux/libnvdimm.h>
+#include <linux/genalloc.h>
#include <linux/vmalloc.h>
#include <linux/device.h>
#include <linux/module.h>
@@ -215,6 +216,8 @@ struct nfit_test {

static struct workqueue_struct *nfit_wq;

+static struct gen_pool *nfit_pool;
+
static struct nfit_test *to_nfit_test(struct device *dev)
{
struct platform_device *pdev = to_platform_device(dev);
@@ -1132,6 +1135,9 @@ static void release_nfit_res(void *data)
list_del(&nfit_res->list);
spin_unlock(&nfit_test_lock);

+ if (resource_size(&nfit_res->res) >= DIMM_SIZE)
+ gen_pool_free(nfit_pool, nfit_res->res.start,
+ resource_size(&nfit_res->res));
vfree(nfit_res->buf);
kfree(nfit_res);
}
@@ -1144,7 +1150,7 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,
GFP_KERNEL);
int rc;

- if (!buf || !nfit_res)
+ if (!buf || !nfit_res || !*dma)
goto err;
rc = devm_add_action(dev, release_nfit_res, nfit_res);
if (rc)
@@ -1164,6 +1170,8 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,

return nfit_res->buf;
err:
+ if (*dma && size >= DIMM_SIZE)
+ gen_pool_free(nfit_pool, *dma, size);
if (buf)
vfree(buf);
kfree(nfit_res);
@@ -1172,9 +1180,16 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,

static void *test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma)
{
+ struct genpool_data_align data = {
+ .align = SZ_128M,
+ };
void *buf = vmalloc(size);

- *dma = (unsigned long) buf;
+ if (size >= DIMM_SIZE)
+ *dma = gen_pool_alloc_algo(nfit_pool, size,
+ gen_pool_first_fit_align, &data);
+ else
+ *dma = (unsigned long) buf;
return __test_alloc(t, size, dma, buf);
}

@@ -2839,6 +2854,18 @@ static __init int nfit_test_init(void)
goto err_register;
}

+ nfit_pool = gen_pool_create(ilog2(SZ_4M), NUMA_NO_NODE);
+ if (!nfit_pool) {
+ rc = -ENOMEM;
+ goto err_register;
+ }
+
+ if (gen_pool_add(nfit_pool, VMALLOC_START,
+ VMALLOC_END + 1 - VMALLOC_START, NUMA_NO_NODE)) {
+ rc = -ENOMEM;
+ goto err_register;
+ }
+
for (i = 0; i < NUM_NFITS; i++) {
struct nfit_test *nfit_test;
struct platform_device *pdev;
@@ -2894,6 +2921,9 @@ static __init int nfit_test_init(void)
return 0;

err_register:
+ if (nfit_pool)
+ gen_pool_destroy(nfit_pool);
+
destroy_workqueue(nfit_wq);
for (i = 0; i < NUM_NFITS; i++)
if (instances[i])
@@ -2917,6 +2947,8 @@ static __exit void nfit_test_exit(void)
platform_driver_unregister(&nfit_test_driver);
nfit_test_teardown();

+ gen_pool_destroy(nfit_pool);
+
for (i = 0; i < NUM_NFITS; i++)
put_device(&instances[i]->pdev.dev);
class_destroy(nfit_test_dimm);



2018-11-24 19:02:00

by Dan Williams

[permalink] [raw]
Subject: [PATCH 2/2] libnvdimm, pfn: Pad pfn namespaces relative to other regions

Commit cfe30b872058 "libnvdimm, pmem: adjust for section collisions with
'System RAM'" enabled Linux to workaround occasions where platform
firmware arranges for "System RAM" and "Persistent Memory" to collide
within a single section boundary. Unfortunately, as reported in this
issue [1], platform firmware can inflict the same collision between
persistent memory regions.

The approach of interrogating iomem_resource does not work in this
case because platform firmware may merge multiple regions into a single
iomem_resource range. Instead provide a method to interrogate regions
that share the same parent bus.

This is a stop-gap until the core-MM can grow support for hotplug on
sub-section boundaries.

[1]: https://github.com/pmem/ndctl/issues/76

Fixes: cfe30b872058 ("libnvdimm, pmem: adjust for section collisions with...")
Cc: <[email protected]>
Signed-off-by: Dan Williams <[email protected]>
---
drivers/nvdimm/nd-core.h | 2 +
drivers/nvdimm/pfn_devs.c | 64 ++++++++++++++++++++++++------------------
drivers/nvdimm/region_devs.c | 41 +++++++++++++++++++++++++++
3 files changed, 80 insertions(+), 27 deletions(-)

diff --git a/drivers/nvdimm/nd-core.h b/drivers/nvdimm/nd-core.h
index 182258f64417..d0c621b32f72 100644
--- a/drivers/nvdimm/nd-core.h
+++ b/drivers/nvdimm/nd-core.h
@@ -111,6 +111,8 @@ resource_size_t nd_pmem_available_dpa(struct nd_region *nd_region,
struct nd_mapping *nd_mapping, resource_size_t *overlap);
resource_size_t nd_blk_available_dpa(struct nd_region *nd_region);
resource_size_t nd_region_available_dpa(struct nd_region *nd_region);
+int nd_region_conflict(struct nd_region *nd_region, resource_size_t start,
+ resource_size_t size);
resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd,
struct nd_label_id *label_id);
int alias_dpa_busy(struct device *dev, void *data);
diff --git a/drivers/nvdimm/pfn_devs.c b/drivers/nvdimm/pfn_devs.c
index 24c64090169e..6f22272e8d80 100644
--- a/drivers/nvdimm/pfn_devs.c
+++ b/drivers/nvdimm/pfn_devs.c
@@ -649,14 +649,47 @@ static u64 phys_pmem_align_down(struct nd_pfn *nd_pfn, u64 phys)
ALIGN_DOWN(phys, nd_pfn->align));
}

+/*
+ * Check if pmem collides with 'System RAM', or other regions when
+ * section aligned. Trim it accordingly.
+ */
+static void trim_pfn_device(struct nd_pfn *nd_pfn, u32 *start_pad, u32 *end_trunc)
+{
+ struct nd_namespace_common *ndns = nd_pfn->ndns;
+ struct nd_namespace_io *nsio = to_nd_namespace_io(&ndns->dev);
+ struct nd_region *nd_region = to_nd_region(nd_pfn->dev.parent);
+ const resource_size_t start = nsio->res.start;
+ const resource_size_t end = start + resource_size(&nsio->res);
+ resource_size_t adjust, size;
+
+ *start_pad = 0;
+ *end_trunc = 0;
+
+ adjust = start - PHYS_SECTION_ALIGN_DOWN(start);
+ size = resource_size(&nsio->res) + adjust;
+ if (region_intersects(start - adjust, size, IORESOURCE_SYSTEM_RAM,
+ IORES_DESC_NONE) == REGION_MIXED
+ || nd_region_conflict(nd_region, start - adjust, size))
+ *start_pad = PHYS_SECTION_ALIGN_UP(start) - start;
+
+ /* Now check that end of the range does not collide. */
+ adjust = PHYS_SECTION_ALIGN_UP(end) - end;
+ size = resource_size(&nsio->res) + adjust;
+ if (region_intersects(start, size, IORESOURCE_SYSTEM_RAM,
+ IORES_DESC_NONE) == REGION_MIXED
+ || !IS_ALIGNED(end, nd_pfn->align)
+ || nd_region_conflict(nd_region, start, size + adjust))
+ *end_trunc = end - phys_pmem_align_down(nd_pfn, end);
+}
+
static int nd_pfn_init(struct nd_pfn *nd_pfn)
{
u32 dax_label_reserve = is_nd_dax(&nd_pfn->dev) ? SZ_128K : 0;
struct nd_namespace_common *ndns = nd_pfn->ndns;
- u32 start_pad = 0, end_trunc = 0;
+ struct nd_namespace_io *nsio = to_nd_namespace_io(&ndns->dev);
resource_size_t start, size;
- struct nd_namespace_io *nsio;
struct nd_region *nd_region;
+ u32 start_pad, end_trunc;
struct nd_pfn_sb *pfn_sb;
unsigned long npfns;
phys_addr_t offset;
@@ -688,30 +721,7 @@ static int nd_pfn_init(struct nd_pfn *nd_pfn)

memset(pfn_sb, 0, sizeof(*pfn_sb));

- /*
- * Check if pmem collides with 'System RAM' when section aligned and
- * trim it accordingly
- */
- nsio = to_nd_namespace_io(&ndns->dev);
- start = PHYS_SECTION_ALIGN_DOWN(nsio->res.start);
- size = resource_size(&nsio->res);
- if (region_intersects(start, size, IORESOURCE_SYSTEM_RAM,
- IORES_DESC_NONE) == REGION_MIXED) {
- start = nsio->res.start;
- start_pad = PHYS_SECTION_ALIGN_UP(start) - start;
- }
-
- start = nsio->res.start;
- size = PHYS_SECTION_ALIGN_UP(start + size) - start;
- if (region_intersects(start, size, IORESOURCE_SYSTEM_RAM,
- IORES_DESC_NONE) == REGION_MIXED
- || !IS_ALIGNED(start + resource_size(&nsio->res),
- nd_pfn->align)) {
- size = resource_size(&nsio->res);
- end_trunc = start + size - phys_pmem_align_down(nd_pfn,
- start + size);
- }
-
+ trim_pfn_device(nd_pfn, &start_pad, &end_trunc);
if (start_pad + end_trunc)
dev_info(&nd_pfn->dev, "%s alignment collision, truncate %d bytes\n",
dev_name(&ndns->dev), start_pad + end_trunc);
@@ -722,7 +732,7 @@ static int nd_pfn_init(struct nd_pfn *nd_pfn)
* implementation will limit the pfns advertised through
* ->direct_access() to those that are included in the memmap.
*/
- start += start_pad;
+ start = nsio->res.start + start_pad;
size = resource_size(&nsio->res);
npfns = PFN_SECTION_ALIGN_UP((size - start_pad - end_trunc - SZ_8K)
/ PAGE_SIZE);
diff --git a/drivers/nvdimm/region_devs.c b/drivers/nvdimm/region_devs.c
index 174a418cb171..e7377f1028ef 100644
--- a/drivers/nvdimm/region_devs.c
+++ b/drivers/nvdimm/region_devs.c
@@ -1184,6 +1184,47 @@ int nvdimm_has_cache(struct nd_region *nd_region)
}
EXPORT_SYMBOL_GPL(nvdimm_has_cache);

+struct conflict_context {
+ struct nd_region *nd_region;
+ resource_size_t start, size;
+};
+
+static int region_conflict(struct device *dev, void *data)
+{
+ struct nd_region *nd_region;
+ struct conflict_context *ctx = data;
+ resource_size_t res_end, region_end, region_start;
+
+ if (!is_memory(dev))
+ return 0;
+
+ nd_region = to_nd_region(dev);
+ if (nd_region == ctx->nd_region)
+ return 0;
+
+ res_end = ctx->start + ctx->size;
+ region_start = nd_region->ndr_start;
+ region_end = region_start + nd_region->ndr_size;
+ if (ctx->start >= region_start && ctx->start < region_end)
+ return -EBUSY;
+ if (res_end > region_start && res_end <= region_end)
+ return -EBUSY;
+ return 0;
+}
+
+int nd_region_conflict(struct nd_region *nd_region, resource_size_t start,
+ resource_size_t size)
+{
+ struct nvdimm_bus *nvdimm_bus = walk_to_nvdimm_bus(&nd_region->dev);
+ struct conflict_context ctx = {
+ .nd_region = nd_region,
+ .start = start,
+ .size = size,
+ };
+
+ return device_for_each_child(&nvdimm_bus->dev, &ctx, region_conflict);
+}
+
void __exit nd_region_devs_exit(void)
{
ida_destroy(&region_ida);


2018-11-25 16:38:55

by Dan Williams

[permalink] [raw]
Subject: Re: [PATCH 2/2] libnvdimm, pfn: Pad pfn namespaces relative to other regions

On Sat, Nov 24, 2018 at 11:54 PM Sasha Levin <[email protected]> wrote:
>
> Hi,
>
> [This is an automated email]
>
> This commit has been processed because it contains a "Fixes:" tag,
> fixing commit: cfe30b872058 libnvdimm, pmem: adjust for section collisions with 'System RAM'.
>
> The bot has tested the following trees: v4.19.4, v4.14.83, v4.9.140.
>
> v4.19.4: Build OK!
> v4.14.83: Build OK!
> v4.9.140: Failed to apply! Possible dependencies:
> Unable to calculate
>
>
> How should we proceed with this patch?

4.9-stable will need a manual backport since this code was refactored
in the intervening kernel releases.

2018-12-03 18:34:42

by Verma, Vishal L

[permalink] [raw]
Subject: Re: [PATCH 1/2] tools/testing/nvdimm: Align test resources to 128M


On Sat, 2018-11-24 at 10:46 -0800, Dan Williams wrote:
> In preparation for libnvdimm growing new restrictions to detect section
> conflicts between persistent memory regions, enable nfit_test to
> allocate aligned resources. Use a gen_pool to allocate nfit_test's fake
> resources in a separate address space from the virtual translation of
> the same.
>
> Signed-off-by: Dan Williams <[email protected]>
> ---
> tools/testing/nvdimm/test/nfit.c | 36 ++++++++++++++++++++++++++++++++++--
> 1 file changed, 34 insertions(+), 2 deletions(-)
>

This looks good to me,
Reviewed-by: Vishal Verma <[email protected]>

> diff --git a/tools/testing/nvdimm/test/nfit.c b/tools/testing/nvdimm/test/nfit.c
> index 01ec04bf91b5..ca4e61c864d5 100644
> --- a/tools/testing/nvdimm/test/nfit.c
> +++ b/tools/testing/nvdimm/test/nfit.c
> @@ -15,6 +15,7 @@
> #include <linux/dma-mapping.h>
> #include <linux/workqueue.h>
> #include <linux/libnvdimm.h>
> +#include <linux/genalloc.h>
> #include <linux/vmalloc.h>
> #include <linux/device.h>
> #include <linux/module.h>
> @@ -215,6 +216,8 @@ struct nfit_test {
>
> static struct workqueue_struct *nfit_wq;
>
> +static struct gen_pool *nfit_pool;
> +
> static struct nfit_test *to_nfit_test(struct device *dev)
> {
> struct platform_device *pdev = to_platform_device(dev);
> @@ -1132,6 +1135,9 @@ static void release_nfit_res(void *data)
> list_del(&nfit_res->list);
> spin_unlock(&nfit_test_lock);
>
> + if (resource_size(&nfit_res->res) >= DIMM_SIZE)
> + gen_pool_free(nfit_pool, nfit_res->res.start,
> + resource_size(&nfit_res->res));
> vfree(nfit_res->buf);
> kfree(nfit_res);
> }
> @@ -1144,7 +1150,7 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,
> GFP_KERNEL);
> int rc;
>
> - if (!buf || !nfit_res)
> + if (!buf || !nfit_res || !*dma)
> goto err;
> rc = devm_add_action(dev, release_nfit_res, nfit_res);
> if (rc)
> @@ -1164,6 +1170,8 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,
>
> return nfit_res->buf;
> err:
> + if (*dma && size >= DIMM_SIZE)
> + gen_pool_free(nfit_pool, *dma, size);
> if (buf)
> vfree(buf);
> kfree(nfit_res);
> @@ -1172,9 +1180,16 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,
>
> static void *test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma)
> {
> + struct genpool_data_align data = {
> + .align = SZ_128M,
> + };
> void *buf = vmalloc(size);
>
> - *dma = (unsigned long) buf;
> + if (size >= DIMM_SIZE)
> + *dma = gen_pool_alloc_algo(nfit_pool, size,
> + gen_pool_first_fit_align, &data);
> + else
> + *dma = (unsigned long) buf;
> return __test_alloc(t, size, dma, buf);
> }
>
> @@ -2839,6 +2854,18 @@ static __init int nfit_test_init(void)
> goto err_register;
> }
>
> + nfit_pool = gen_pool_create(ilog2(SZ_4M), NUMA_NO_NODE);
> + if (!nfit_pool) {
> + rc = -ENOMEM;
> + goto err_register;
> + }
> +
> + if (gen_pool_add(nfit_pool, VMALLOC_START,
> + VMALLOC_END + 1 - VMALLOC_START, NUMA_NO_NODE)) {
> + rc = -ENOMEM;
> + goto err_register;
> + }
> +
> for (i = 0; i < NUM_NFITS; i++) {
> struct nfit_test *nfit_test;
> struct platform_device *pdev;
> @@ -2894,6 +2921,9 @@ static __init int nfit_test_init(void)
> return 0;
>
> err_register:
> + if (nfit_pool)
> + gen_pool_destroy(nfit_pool);
> +
> destroy_workqueue(nfit_wq);
> for (i = 0; i < NUM_NFITS; i++)
> if (instances[i])
> @@ -2917,6 +2947,8 @@ static __exit void nfit_test_exit(void)
> platform_driver_unregister(&nfit_test_driver);
> nfit_test_teardown();
>
> + gen_pool_destroy(nfit_pool);
> +
> for (i = 0; i < NUM_NFITS; i++)
> put_device(&instances[i]->pdev.dev);
> class_destroy(nfit_test_dimm);
>

2018-12-03 18:43:04

by Verma, Vishal L

[permalink] [raw]
Subject: Re: [PATCH 2/2] libnvdimm, pfn: Pad pfn namespaces relative to other regions


On Sat, 2018-11-24 at 10:47 -0800, Dan Williams wrote:
> Commit cfe30b872058 "libnvdimm, pmem: adjust for section collisions with
> 'System RAM'" enabled Linux to workaround occasions where platform
> firmware arranges for "System RAM" and "Persistent Memory" to collide
> within a single section boundary. Unfortunately, as reported in this
> issue [1], platform firmware can inflict the same collision between
> persistent memory regions.
>
> The approach of interrogating iomem_resource does not work in this
> case because platform firmware may merge multiple regions into a single
> iomem_resource range. Instead provide a method to interrogate regions
> that share the same parent bus.
>
> This is a stop-gap until the core-MM can grow support for hotplug on
> sub-section boundaries.
>
> [1]: https://github.com/pmem/ndctl/issues/76
>
> Fixes: cfe30b872058 ("libnvdimm, pmem: adjust for section collisions with...")
> Cc: <[email protected]>
> Signed-off-by: Dan Williams <[email protected]>
> ---
> drivers/nvdimm/nd-core.h | 2 +
> drivers/nvdimm/pfn_devs.c | 64 ++++++++++++++++++++++++------------------
> drivers/nvdimm/region_devs.c | 41 +++++++++++++++++++++++++++
> 3 files changed, 80 insertions(+), 27 deletions(-)

Looks good to me,
Reviewed-by: Vishal Verma <[email protected]>

>
> diff --git a/drivers/nvdimm/nd-core.h b/drivers/nvdimm/nd-core.h
> index 182258f64417..d0c621b32f72 100644
> --- a/drivers/nvdimm/nd-core.h
> +++ b/drivers/nvdimm/nd-core.h
> @@ -111,6 +111,8 @@ resource_size_t nd_pmem_available_dpa(struct nd_region *nd_region,
> struct nd_mapping *nd_mapping, resource_size_t *overlap);
> resource_size_t nd_blk_available_dpa(struct nd_region *nd_region);
> resource_size_t nd_region_available_dpa(struct nd_region *nd_region);
> +int nd_region_conflict(struct nd_region *nd_region, resource_size_t start,
> + resource_size_t size);
> resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd,
> struct nd_label_id *label_id);
> int alias_dpa_busy(struct device *dev, void *data);
> diff --git a/drivers/nvdimm/pfn_devs.c b/drivers/nvdimm/pfn_devs.c
> index 24c64090169e..6f22272e8d80 100644
> --- a/drivers/nvdimm/pfn_devs.c
> +++ b/drivers/nvdimm/pfn_devs.c
> @@ -649,14 +649,47 @@ static u64 phys_pmem_align_down(struct nd_pfn *nd_pfn, u64 phys)
> ALIGN_DOWN(phys, nd_pfn->align));
> }
>
> +/*
> + * Check if pmem collides with 'System RAM', or other regions when
> + * section aligned. Trim it accordingly.
> + */
> +static void trim_pfn_device(struct nd_pfn *nd_pfn, u32 *start_pad, u32 *end_trunc)
> +{
> + struct nd_namespace_common *ndns = nd_pfn->ndns;
> + struct nd_namespace_io *nsio = to_nd_namespace_io(&ndns->dev);
> + struct nd_region *nd_region = to_nd_region(nd_pfn->dev.parent);
> + const resource_size_t start = nsio->res.start;
> + const resource_size_t end = start + resource_size(&nsio->res);
> + resource_size_t adjust, size;
> +
> + *start_pad = 0;
> + *end_trunc = 0;
> +
> + adjust = start - PHYS_SECTION_ALIGN_DOWN(start);
> + size = resource_size(&nsio->res) + adjust;
> + if (region_intersects(start - adjust, size, IORESOURCE_SYSTEM_RAM,
> + IORES_DESC_NONE) == REGION_MIXED
> + || nd_region_conflict(nd_region, start - adjust, size))
> + *start_pad = PHYS_SECTION_ALIGN_UP(start) - start;
> +
> + /* Now check that end of the range does not collide. */
> + adjust = PHYS_SECTION_ALIGN_UP(end) - end;
> + size = resource_size(&nsio->res) + adjust;
> + if (region_intersects(start, size, IORESOURCE_SYSTEM_RAM,
> + IORES_DESC_NONE) == REGION_MIXED
> + || !IS_ALIGNED(end, nd_pfn->align)
> + || nd_region_conflict(nd_region, start, size + adjust))
> + *end_trunc = end - phys_pmem_align_down(nd_pfn, end);
> +}
> +
> static int nd_pfn_init(struct nd_pfn *nd_pfn)
> {
> u32 dax_label_reserve = is_nd_dax(&nd_pfn->dev) ? SZ_128K : 0;
> struct nd_namespace_common *ndns = nd_pfn->ndns;
> - u32 start_pad = 0, end_trunc = 0;
> + struct nd_namespace_io *nsio = to_nd_namespace_io(&ndns->dev);
> resource_size_t start, size;
> - struct nd_namespace_io *nsio;
> struct nd_region *nd_region;
> + u32 start_pad, end_trunc;
> struct nd_pfn_sb *pfn_sb;
> unsigned long npfns;
> phys_addr_t offset;
> @@ -688,30 +721,7 @@ static int nd_pfn_init(struct nd_pfn *nd_pfn)
>
> memset(pfn_sb, 0, sizeof(*pfn_sb));
>
> - /*
> - * Check if pmem collides with 'System RAM' when section aligned and
> - * trim it accordingly
> - */
> - nsio = to_nd_namespace_io(&ndns->dev);
> - start = PHYS_SECTION_ALIGN_DOWN(nsio->res.start);
> - size = resource_size(&nsio->res);
> - if (region_intersects(start, size, IORESOURCE_SYSTEM_RAM,
> - IORES_DESC_NONE) == REGION_MIXED) {
> - start = nsio->res.start;
> - start_pad = PHYS_SECTION_ALIGN_UP(start) - start;
> - }
> -
> - start = nsio->res.start;
> - size = PHYS_SECTION_ALIGN_UP(start + size) - start;
> - if (region_intersects(start, size, IORESOURCE_SYSTEM_RAM,
> - IORES_DESC_NONE) == REGION_MIXED
> - || !IS_ALIGNED(start + resource_size(&nsio->res),
> - nd_pfn->align)) {
> - size = resource_size(&nsio->res);
> - end_trunc = start + size - phys_pmem_align_down(nd_pfn,
> - start + size);
> - }
> -
> + trim_pfn_device(nd_pfn, &start_pad, &end_trunc);
> if (start_pad + end_trunc)
> dev_info(&nd_pfn->dev, "%s alignment collision, truncate %d bytes\n",
> dev_name(&ndns->dev), start_pad + end_trunc);
> @@ -722,7 +732,7 @@ static int nd_pfn_init(struct nd_pfn *nd_pfn)
> * implementation will limit the pfns advertised through
> * ->direct_access() to those that are included in the memmap.
> */
> - start += start_pad;
> + start = nsio->res.start + start_pad;
> size = resource_size(&nsio->res);
> npfns = PFN_SECTION_ALIGN_UP((size - start_pad - end_trunc - SZ_8K)
> / PAGE_SIZE);
> diff --git a/drivers/nvdimm/region_devs.c b/drivers/nvdimm/region_devs.c
> index 174a418cb171..e7377f1028ef 100644
> --- a/drivers/nvdimm/region_devs.c
> +++ b/drivers/nvdimm/region_devs.c
> @@ -1184,6 +1184,47 @@ int nvdimm_has_cache(struct nd_region *nd_region)
> }
> EXPORT_SYMBOL_GPL(nvdimm_has_cache);
>
> +struct conflict_context {
> + struct nd_region *nd_region;
> + resource_size_t start, size;
> +};
> +
> +static int region_conflict(struct device *dev, void *data)
> +{
> + struct nd_region *nd_region;
> + struct conflict_context *ctx = data;
> + resource_size_t res_end, region_end, region_start;
> +
> + if (!is_memory(dev))
> + return 0;
> +
> + nd_region = to_nd_region(dev);
> + if (nd_region == ctx->nd_region)
> + return 0;
> +
> + res_end = ctx->start + ctx->size;
> + region_start = nd_region->ndr_start;
> + region_end = region_start + nd_region->ndr_size;
> + if (ctx->start >= region_start && ctx->start < region_end)
> + return -EBUSY;
> + if (res_end > region_start && res_end <= region_end)
> + return -EBUSY;
> + return 0;
> +}
> +
> +int nd_region_conflict(struct nd_region *nd_region, resource_size_t start,
> + resource_size_t size)
> +{
> + struct nvdimm_bus *nvdimm_bus = walk_to_nvdimm_bus(&nd_region->dev);
> + struct conflict_context ctx = {
> + .nd_region = nd_region,
> + .start = start,
> + .size = size,
> + };
> +
> + return device_for_each_child(&nvdimm_bus->dev, &ctx, region_conflict);
> +}
> +
> void __exit nd_region_devs_exit(void)
> {
> ida_destroy(&region_ida);
>

2018-12-05 22:30:39

by Dan Williams

[permalink] [raw]
Subject: [PATCH v2] tools/testing/nvdimm: Align test resources to 128M

In preparation for libnvdimm growing new restrictions to detect section
conflicts between persistent memory regions, enable nfit_test to
allocate aligned resources. Use a gen_pool to allocate nfit_test's fake
resources in a separate address space from the virtual translation of
the same.

Reviewed-by: Vishal Verma <[email protected]>
Tested-by: Vishal Verma <[email protected]>
Signed-off-by: Dan Williams <[email protected]>
---
Changes in v2:
* Fix intermittent crash from stale vmalloc() alias usage (Vishal)

tools/testing/nvdimm/test/nfit.c | 35 +++++++++++++++++++++++++++++++++--
1 file changed, 33 insertions(+), 2 deletions(-)

diff --git a/tools/testing/nvdimm/test/nfit.c b/tools/testing/nvdimm/test/nfit.c
index 01ec04bf91b5..6c16ac36d482 100644
--- a/tools/testing/nvdimm/test/nfit.c
+++ b/tools/testing/nvdimm/test/nfit.c
@@ -15,6 +15,7 @@
#include <linux/dma-mapping.h>
#include <linux/workqueue.h>
#include <linux/libnvdimm.h>
+#include <linux/genalloc.h>
#include <linux/vmalloc.h>
#include <linux/device.h>
#include <linux/module.h>
@@ -215,6 +216,8 @@ struct nfit_test {

static struct workqueue_struct *nfit_wq;

+static struct gen_pool *nfit_pool;
+
static struct nfit_test *to_nfit_test(struct device *dev)
{
struct platform_device *pdev = to_platform_device(dev);
@@ -1132,6 +1135,9 @@ static void release_nfit_res(void *data)
list_del(&nfit_res->list);
spin_unlock(&nfit_test_lock);

+ if (resource_size(&nfit_res->res) >= DIMM_SIZE)
+ gen_pool_free(nfit_pool, nfit_res->res.start,
+ resource_size(&nfit_res->res));
vfree(nfit_res->buf);
kfree(nfit_res);
}
@@ -1144,7 +1150,7 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,
GFP_KERNEL);
int rc;

- if (!buf || !nfit_res)
+ if (!buf || !nfit_res || !*dma)
goto err;
rc = devm_add_action(dev, release_nfit_res, nfit_res);
if (rc)
@@ -1164,6 +1170,8 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,

return nfit_res->buf;
err:
+ if (*dma && size >= DIMM_SIZE)
+ gen_pool_free(nfit_pool, *dma, size);
if (buf)
vfree(buf);
kfree(nfit_res);
@@ -1172,9 +1180,16 @@ static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,

static void *test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma)
{
+ struct genpool_data_align data = {
+ .align = SZ_128M,
+ };
void *buf = vmalloc(size);

- *dma = (unsigned long) buf;
+ if (size >= DIMM_SIZE)
+ *dma = gen_pool_alloc_algo(nfit_pool, size,
+ gen_pool_first_fit_align, &data);
+ else
+ *dma = (unsigned long) buf;
return __test_alloc(t, size, dma, buf);
}

@@ -2839,6 +2854,17 @@ static __init int nfit_test_init(void)
goto err_register;
}

+ nfit_pool = gen_pool_create(ilog2(SZ_4M), NUMA_NO_NODE);
+ if (!nfit_pool) {
+ rc = -ENOMEM;
+ goto err_register;
+ }
+
+ if (gen_pool_add(nfit_pool, SZ_4G, SZ_4G, NUMA_NO_NODE)) {
+ rc = -ENOMEM;
+ goto err_register;
+ }
+
for (i = 0; i < NUM_NFITS; i++) {
struct nfit_test *nfit_test;
struct platform_device *pdev;
@@ -2894,6 +2920,9 @@ static __init int nfit_test_init(void)
return 0;

err_register:
+ if (nfit_pool)
+ gen_pool_destroy(nfit_pool);
+
destroy_workqueue(nfit_wq);
for (i = 0; i < NUM_NFITS; i++)
if (instances[i])
@@ -2917,6 +2946,8 @@ static __exit void nfit_test_exit(void)
platform_driver_unregister(&nfit_test_driver);
nfit_test_teardown();

+ gen_pool_destroy(nfit_pool);
+
for (i = 0; i < NUM_NFITS; i++)
put_device(&instances[i]->pdev.dev);
class_destroy(nfit_test_dimm);