[PATCH] arm64:mm: free the useless initial page table

Catalin Marinas catalin.marinas at arm.com
Mon Nov 24 06:32:08 PST 2014


On Fri, Nov 21, 2014 at 08:27:40AM +0000, zhichang.yuan at linaro.org wrote:
> From: "zhichang.yuan" <zhichang.yuan at linaro.org>
> 
> For 64K page system, after mapping a PMD section, the corresponding initial
> page table is not needed any more. That page can be freed.
> 
> Signed-off-by: Zhichang Yuan <zhichang.yuan at linaro.org>
> ---
>  arch/arm64/mm/mmu.c |    5 ++++-
>  1 file changed, 4 insertions(+), 1 deletion(-)
> 
> diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c
> index f4f8b50..12a336b 100644
> --- a/arch/arm64/mm/mmu.c
> +++ b/arch/arm64/mm/mmu.c
> @@ -191,8 +191,11 @@ static void __init alloc_init_pmd(pud_t *pud, unsigned long addr,
>  			 * Check for previous table entries created during
>  			 * boot (__create_page_tables) and flush them.
>  			 */
> -			if (!pmd_none(old_pmd))
> +			if (!pmd_none(old_pmd)) {
>  				flush_tlb_all();
> +				if (pmd_table(old_pmd))
> +					memblock_free(pte_pfn(pmd_pte(old_pmd)) << PAGE_SHIFT, PAGE_SIZE);
> +			}

For consistency with alloc_init_pud(), could you do:

	phys_addr_t table = __pa(pte_offset(&old_pmd, 0));
	memblock_free(table, PAGE_SIZE);

Since you are at this, for alloc_init_pud() could you please move the
flush_tlb_all() before memblock_free()? Theoretical problem really but
it's nice for consistency.

Thanks.

-- 
Catalin



More information about the linux-arm-kernel mailing list