[RFC] mm: cma: move init_cma_reserved_pageblock to cma.c
Joonsoo Kim
iamjoonsoo.kim at lge.com
Thu Jun 26 23:00:47 PDT 2014
On Wed, Jun 25, 2014 at 12:48:02AM +0200, Michal Nazarewicz wrote:
> With [f495d26: “generalize CMA reserved area management
> functionality”] patch CMA has its place under mm directory now so
> there is no need to shoehorn a highly CMA specific functions inside of
> page_alloc.c.
>
> As such move init_cma_reserved_pageblock from mm/page_alloc.c to
> mm/cma.c, rename it to cma_init_reserved_pageblock and refactor
> a little.
>
> Most importantly, if a !pfn_valid(pfn) is encountered, just
> return -EINVAL instead of warning and trying to continue the
> initialisation of the area. It's not clear, to me at least, what good
> is continuing the work on a PFN that is known to be invalid.
>
> Signed-off-by: Michal Nazarewicz <mina86 at mina86.com>
Acked-by: Joonsoo Kim <iamjoonsoo.kim at lge.com>
One question below.
> ---
> include/linux/gfp.h | 3 --
> mm/cma.c | 85 +++++++++++++++++++++++++++++++++++++++++------------
> mm/page_alloc.c | 31 -------------------
> 3 files changed, 66 insertions(+), 53 deletions(-)
>
> diff --git a/include/linux/gfp.h b/include/linux/gfp.h
> index 5e7219d..107793e9 100644
> --- a/include/linux/gfp.h
> +++ b/include/linux/gfp.h
> @@ -415,9 +415,6 @@ extern int alloc_contig_range(unsigned long start, unsigned long end,
> unsigned migratetype);
> extern void free_contig_range(unsigned long pfn, unsigned nr_pages);
>
> -/* CMA stuff */
> -extern void init_cma_reserved_pageblock(struct page *page);
> -
> #endif
>
> #endif /* __LINUX_GFP_H */
> diff --git a/mm/cma.c b/mm/cma.c
> index c17751c..843b2b6 100644
> --- a/mm/cma.c
> +++ b/mm/cma.c
> @@ -28,11 +28,14 @@
> #include <linux/err.h>
> #include <linux/mm.h>
> #include <linux/mutex.h>
> +#include <linux/page-isolation.h>
> #include <linux/sizes.h>
> #include <linux/slab.h>
> #include <linux/log2.h>
> #include <linux/cma.h>
>
> +#include "internal.h"
> +
> struct cma {
> unsigned long base_pfn;
> unsigned long count;
> @@ -83,37 +86,81 @@ static void cma_clear_bitmap(struct cma *cma, unsigned long pfn, int count)
> mutex_unlock(&cma->lock);
> }
>
> +/* Free whole pageblock and set its migration type to MIGRATE_CMA. */
> +static int __init cma_init_reserved_pageblock(struct zone *zone,
> + unsigned long pageblock_pfn)
> +{
> + unsigned long pfn, nr_pages, i;
> + struct page *page, *p;
> + unsigned order;
> +
> + pfn = pageblock_pfn;
> + if (!pfn_valid(pfn))
> + goto invalid_pfn;
> + page = pfn_to_page(pfn);
> +
> + p = page;
> + i = pageblock_nr_pages;
> + do {
> + if (!pfn_valid(pfn))
> + goto invalid_pfn;
> +
> + /*
> + * alloc_contig_range requires the pfn range specified to be
> + * in the same zone. Make this simple by forcing the entire
> + * CMA resv range to be in the same zone.
> + */
> + if (page_zone(p) != zone) {
> + pr_err("pfn %lu belongs to %s, expecting %s\n",
> + pfn, page_zone(p)->name, zone->name);
> + return -EINVAL;
> + }
> +
> + __ClearPageReserved(p);
> + set_page_count(p, 0);
> + } while (++p, ++pfn, --i);
So, when we meet fail condition, __ClearPageReserved, set_page_count()
are already executed for some pages. Is that no problem?
Thanks.
More information about the linux-arm-kernel
mailing list