[PATCH v2 1/1] ARM: LPAE: Fix mapping in alloc_init_pte for unaligned addresses.
Catalin Marinas
catalin.marinas at arm.com
Thu Mar 14 01:14:21 EDT 2013
(sorry for if you got this message twice, gmail's new reply method
decided to send html)
On 18 September 2012 12:52, R, Sricharan <r.sricharan at ti.com> wrote:
> Hi,
> On Fri, Aug 17, 2012 at 4:32 PM, R, Sricharan <r.sricharan at ti.com> wrote:
>> Hi,
>>
>>> With LPAE, When either the start address or end address
>>> or physical address to be mapped is unaligned,
>>> alloc_init_section creates page granularity mappings.
>>> alloc_init_section calls alloc_init_pte which populates
>>> one pmd entry and sets up the ptes. But if the size is
>>> greater than what can be mapped by one pmd entry,
>>> then the rest remains unmapped.
>>>
>>> The issue becomes visible when LPAE is enabled, where we have
>>> the 3 levels with seperate pgd and pmd's.
>>> When a static mapping for 3MB is requested, only 2MB is mapped
>>> and the remaining 1MB is unmapped. Fixing this here, by looping
>>> in to map the entire unaligned address range.
>>>
>>> Boot tested on OMAP5 evm with both LPAE enabled/disabled
>>> and verified that static mappings with unaligned addresses
>>> are properly mapped.
>>>
>>> Signed-off-by: R Sricharan <r.sricharan at ti.com>
>>> Reviewed-by: Santosh Shilimkar <santosh.shilimkar at ti.com>
>>> Cc: Catalin Marinas <catalin.marinas at arm.com>
>>> ---
>>> [V2] Moved the loop to alloc_init_pte as per Russell's
>>> feedback and changed the subject accordingly.
>>> Using PMD_XXX instead of SECTION_XXX to avoid
>>> different loop increments with/without LPAE.
>>>
>>> arch/arm/mm/mmu.c | 22 ++++++++++++++++++----
>>> 1 file changed, 18 insertions(+), 4 deletions(-)
>>>
>>> diff --git a/arch/arm/mm/mmu.c b/arch/arm/mm/mmu.c
>>> index cf4528d..0ed8808 100644
>>> --- a/arch/arm/mm/mmu.c
>>> +++ b/arch/arm/mm/mmu.c
>>> @@ -585,11 +585,25 @@ static void __init alloc_init_pte(pmd_t *pmd, unsigned long addr,
>>> unsigned long end, unsigned long pfn,
>>> const struct mem_type *type)
>>> {
>>> - pte_t *pte = early_pte_alloc(pmd, addr, type->prot_l1);
>>> + unsigned long next;
>>> + pte_t *pte;
>>> + phys_addr_t phys;
>>> +
>>> do {
>>> - set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)), 0);
>>> - pfn++;
>>> - } while (pte++, addr += PAGE_SIZE, addr != end);
>>> + if ((end-addr) & PMD_MASK)
>>> + next = (addr + PMD_SIZE) & PMD_MASK;
>>> + else
>>> + next = end;
>>> +
>>> + pte = early_pte_alloc(pmd, addr, type->prot_l1);
>>> + do {
>>> + set_pte_ext(pte, pfn_pte(pfn,
>>> + __pgprot(type->prot_pte)), 0);
>>> + pfn++;
>>> + } while (pte++, addr += PAGE_SIZE, addr != next);
>>> +
>>> + phys += next - addr;
>>> + } while (pmd++, addr = next, addr != end);
>>> }
>>>
>> ping..
>
> Ping again.
> The issue is reproducible in mainline with CMA + LPAE enabled.
> CMA tries to reserve/map 16 MB with 2 level table entries and
> crashes in alloc_init_pte.
>
> This patch fixes that. Just posted a V3 of the same patch.
>
> https://patchwork.kernel.org/patch/1472031/
I thought there was another patch where the looping was in an
alloc_init_pmd() function, or there are just two different threads. I
acked the other but not this one as I don't think looping over pmd
inside the alloc_init_pte() function is the right thing.
--
Catalin
More information about the linux-arm-kernel
mailing list