[PATCH V6] ARM: LPAE: Fix mapping in alloc_init_section for unaligned addresses
Christoffer Dall
chris at cloudcar.com
Sun Mar 17 15:52:43 EDT 2013
On Sat, Mar 16, 2013 at 10:05 PM, Sricharan R <r.sricharan at ti.com> wrote:
> From: R Sricharan <r.sricharan at ti.com>
>
> With LPAE enabled, alloc_init_section() does not map the entire
> address space for unaligned addresses.
>
> The issue also reproduced with CMA + LPAE. CMA tries to map 16MB
> with page granularity mappings during boot. alloc_init_pte()
> is called and out of 16MB, only 2MB gets mapped and rest remains
> unaccessible.
>
> Because of this OMAP5 boot is broken with CMA + LPAE enabled.
> Fix the issue by ensuring that the entire addresses are
> mapped.
>
> Signed-off-by: R Sricharan <r.sricharan at ti.com>
> Cc: Catalin Marinas <catalin.marinas at arm.com>
> Cc: Christoffer Dall <chris at cloudcar.com>
> Cc: Russell King <linux at arm.linux.org.uk>
> Cc: Santosh Shilimkar <santosh.shilimkar at ti.com>
> Tested-by: Laura Abbott <lauraa at codeaurora.org>
> Acked-by: Catalin Marinas <catalin.marinas at arm.com>
> ---
> [V2] Moved the loop to alloc_init_pte as per Russell's
> feedback and changed the subject accordingly.
> Using PMD_XXX instead of SECTION_XXX to avoid
> different loop increments with/without LPAE.
>
> [v3] Removed the dummy variable phys and updated
> the commit log for CMA case.
>
> [v4] Resending with updated change log and
> updating the tags.
>
> [v5] Renamed alloc_init_section to alloc_init_pmd
> and moved the loop back there. Also introduced
> map_init_section as per Catalin's comments.
>
> [v6] Corrected tags and updated the comments for code.
>
> arch/arm/mm/mmu.c | 73 ++++++++++++++++++++++++++++++++++-------------------
> 1 file changed, 47 insertions(+), 26 deletions(-)
>
> diff --git a/arch/arm/mm/mmu.c b/arch/arm/mm/mmu.c
> index e95a996..7897894 100644
> --- a/arch/arm/mm/mmu.c
> +++ b/arch/arm/mm/mmu.c
> @@ -598,39 +598,60 @@ static void __init alloc_init_pte(pmd_t *pmd, unsigned long addr,
> } while (pte++, addr += PAGE_SIZE, addr != end);
> }
>
> -static void __init alloc_init_section(pud_t *pud, unsigned long addr,
> - unsigned long end, phys_addr_t phys,
> - const struct mem_type *type)
> +static void __init map_init_section(pmd_t *pmd, unsigned long addr,
> + unsigned long end, phys_addr_t phys,
> + const struct mem_type *type)
> {
> - pmd_t *pmd = pmd_offset(pud, addr);
> -
> +#ifndef CONFIG_ARM_LPAE
> /*
> - * Try a section mapping - end, addr and phys must all be aligned
> - * to a section boundary. Note that PMDs refer to the individual
> - * L1 entries, whereas PGDs refer to a group of L1 entries making
> - * up one logical pointer to an L2 table.
> + * In classic MMU format, puds and pmds are folded in to
> + * the pgds. pmd_offset gives the PGD entry. PGDs refer to a
> + * group of L1 entries making up one logical pointer to
> + * an L2 table (2MB), where as PMDs refer to the individual
> + * L1 entries (1MB). Hence increment to get the correct
> + * offset for odd 1MB sections.
> + * (See arch/arm/include/asm/pgtable-2level.h)
> */
> - if (type->prot_sect && ((addr | end | phys) & ~SECTION_MASK) == 0) {
> - pmd_t *p = pmd;
> -
> -#ifndef CONFIG_ARM_LPAE
> - if (addr & SECTION_SIZE)
> - pmd++;
> + if (addr & SECTION_SIZE)
> + pmd++;
> #endif
> + do {
> + *pmd = __pmd(phys | type->prot_sect);
> + phys += SECTION_SIZE;
> + } while (pmd++, addr += SECTION_SIZE, addr != end);
>
> - do {
> - *pmd = __pmd(phys | type->prot_sect);
> - phys += SECTION_SIZE;
> - } while (pmd++, addr += SECTION_SIZE, addr != end);
> + flush_pmd_entry(pmd);
> +}
>
> - flush_pmd_entry(p);
> - } else {
> +static void __init alloc_init_pmd(pud_t *pud, unsigned long addr,
> + unsigned long end, phys_addr_t phys,
> + const struct mem_type *type)
> +{
> + pmd_t *pmd = pmd_offset(pud, addr);
> + unsigned long next;
> +
> + do {
> /*
> - * No need to loop; pte's aren't interested in the
> - * individual L1 entries.
> + * With LPAE, we must loop over to map
> + * all the pmds for the given range.
> */
> - alloc_init_pte(pmd, addr, end, __phys_to_pfn(phys), type);
> - }
> + next = pmd_addr_end(addr, end);
> +
> + /*
> + * Try a section mapping - addr, next and phys must all be
> + * aligned to a section boundary.
> + */
> + if (type->prot_sect &&
> + ((addr | next | phys) & ~SECTION_MASK) == 0) {
> + map_init_section(pmd, addr, next, phys, type);
> + } else {
> + alloc_init_pte(pmd, addr, next,
> + __phys_to_pfn(phys), type);
> + }
> +
> + phys += next - addr;
> +
> + } while (pmd++, addr = next, addr != end);
> }
>
> static void __init alloc_init_pud(pgd_t *pgd, unsigned long addr,
> @@ -641,7 +662,7 @@ static void __init alloc_init_pud(pgd_t *pgd, unsigned long addr,
>
> do {
> next = pud_addr_end(addr, end);
> - alloc_init_section(pud, addr, next, phys, type);
> + alloc_init_pmd(pud, addr, next, phys, type);
> phys += next - addr;
> } while (pud++, addr = next, addr != end);
> }
> --
> 1.7.9.5
>
Acked-by: Christoffer Dall <chris at cloudcar.com>
More information about the linux-arm-kernel
mailing list