[Linaro-mm-sig] [PATCHv23 12/16] mm: trigger page reclaim in alloc_contig_range() to stabilise watermarks

Barry Song 21cnbao at gmail.com
Wed Feb 29 04:48:49 EST 2012


2012/2/23 Marek Szyprowski <m.szyprowski at samsung.com>:
> alloc_contig_range() performs memory allocation so it also should keep
> track on keeping the correct level of memory watermarks. This commit adds
> a call to *_slowpath style reclaim to grab enough pages to make sure that
> the final collection of contiguous pages from freelists will not starve
> the system.
>
> Signed-off-by: Marek Szyprowski <m.szyprowski at samsung.com>
> Signed-off-by: Kyungmin Park <kyungmin.park at samsung.com>
> CC: Michal Nazarewicz <mina86 at mina86.com>
> Tested-by: Rob Clark <rob.clark at linaro.org>
> Tested-by: Ohad Ben-Cohen <ohad at wizery.com>
> Tested-by: Benjamin Gaignard <benjamin.gaignard at linaro.org>
> Tested-by: Robert Nelson <robertcnelson at gmail.com>
> ---
>  include/linux/mmzone.h |    9 +++++++
>  mm/page_alloc.c        |   62 ++++++++++++++++++++++++++++++++++++++++++++++++
>  2 files changed, 71 insertions(+), 0 deletions(-)
>
> diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
> index 4781f30..77db8c0 100644
> --- a/include/linux/mmzone.h
> +++ b/include/linux/mmzone.h
> @@ -63,8 +63,10 @@ enum {
>
>  #ifdef CONFIG_CMA
>  #  define is_migrate_cma(migratetype) unlikely((migratetype) == MIGRATE_CMA)
> +#  define cma_wmark_pages(zone)        zone->min_cma_pages
>  #else
>  #  define is_migrate_cma(migratetype) false
> +#  define cma_wmark_pages(zone) 0
>  #endif
>
>  #define for_each_migratetype_order(order, type) \
> @@ -371,6 +373,13 @@ struct zone {
>        /* see spanned/present_pages for more description */
>        seqlock_t               span_seqlock;
>  #endif
> +#ifdef CONFIG_CMA
> +       /*
> +        * CMA needs to increase watermark levels during the allocation
> +        * process to make sure that the system is not starved.
> +        */
> +       unsigned long           min_cma_pages;
> +#endif
>        struct free_area        free_area[MAX_ORDER];
>
>  #ifndef CONFIG_SPARSEMEM
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index 7a0d286..39cd74f 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -5092,6 +5092,11 @@ static void __setup_per_zone_wmarks(void)
>                                        low + (min >> 2);
>                zone->watermark[WMARK_HIGH] = min_wmark_pages(zone) +
>                                        low + (min >> 1);
> +
> +               zone->watermark[WMARK_MIN] += cma_wmark_pages(zone);
> +               zone->watermark[WMARK_LOW] += cma_wmark_pages(zone);
> +               zone->watermark[WMARK_HIGH] += cma_wmark_pages(zone);
> +
>                setup_zone_migrate_reserve(zone);
>                spin_unlock_irqrestore(&zone->lock, flags);
>        }
> @@ -5695,6 +5700,56 @@ static int __alloc_contig_migrate_range(unsigned long start, unsigned long end)
>        return ret > 0 ? 0 : ret;
>  }
>
> +/*
> + * Update zone's cma pages counter used for watermark level calculation.
> + */
> +static inline void __update_cma_watermarks(struct zone *zone, int count)
> +{
> +       unsigned long flags;
> +       spin_lock_irqsave(&zone->lock, flags);
> +       zone->min_cma_pages += count;
> +       spin_unlock_irqrestore(&zone->lock, flags);
> +       setup_per_zone_wmarks();
> +}
> +
> +/*
> + * Trigger memory pressure bump to reclaim some pages in order to be able to
> + * allocate 'count' pages in single page units. Does similar work as
> + *__alloc_pages_slowpath() function.
> + */
> +static int __reclaim_pages(struct zone *zone, gfp_t gfp_mask, int count)
> +{
> +       enum zone_type high_zoneidx = gfp_zone(gfp_mask);
> +       struct zonelist *zonelist = node_zonelist(0, gfp_mask);
> +       int did_some_progress = 0;
> +       int order = 1;
> +       unsigned long watermark;
> +
> +       /*
> +        * Increase level of watermarks to force kswapd do his job
> +        * to stabilise at new watermark level.
> +        */
> +       __update_cma_watermarks(zone, count);
> +
> +       /* Obey watermarks as if the page was being allocated */
> +       watermark = low_wmark_pages(zone) + count;
> +       while (!zone_watermark_ok(zone, 0, watermark, 0, 0)) {
> +               wake_all_kswapd(order, zonelist, high_zoneidx, zone_idx(zone));
> +
> +               did_some_progress = __perform_reclaim(gfp_mask, order, zonelist,
> +                                                     NULL);
> +               if (!did_some_progress) {
> +                       /* Exhausted what can be done so it's blamo time */
> +                       out_of_memory(zonelist, gfp_mask, order, NULL);

out_of_memory() has got another param in the newest next/master tree,
out_of_memory(zonelist, gfp_mask, order, NULL, false) should be OK.

-barry



More information about the linux-arm-kernel mailing list