[PATCH 0/4] Patches for -next
Shilimkar, Santosh
santosh.shilimkar at ti.com
Fri May 21 06:57:03 EDT 2010
Catalin,
> -----Original Message-----
> From: Catalin Marinas [mailto:catalin.marinas at arm.com]
> Sent: Friday, May 21, 2010 3:22 PM
> To: Shilimkar, Santosh
> Cc: Russell King - ARM Linux; linux-arm-kernel at lists.infradead.org
> Subject: RE: [PATCH 0/4] Patches for -next
>
> Santosh,
>
> On Fri, 2010-05-21 at 06:59 +0100, Shilimkar, Santosh wrote:
> > > Catalin Marinas (4):
> > > ARM: Remove the domain switching on ARMv6k/v7 CPUs
> > > ARM: Use lazy cache flushing on ARMv7 SMP systems
> > > ARM: Assume new page cache pages have dirty D-cache
> > > ARM: Defer the L_PTE_EXEC flag setting to update_mmu_cache() on SMP
> > >
> > >
> > I just gave quick try with these patches on OMAP4 yesterday.
> > "ARM: Defer the L_PTE_EXEC flag setting to update_mmu_cache() on SMP"
> > patch is creating regression when loading the module.
>
> Thanks for testing and reporting this. I missed the fact the modules
> allocation uses set_pte_at(). Below is an updated patch which makes sure
> that L_PTE_EXEC is set for addresses greater than TASK_SIZE. We
> shouldn't have a race for modules as initialisation only happens on a
> single CPU.
>
> Could you please try the updated patch below?
>
This one works fine!!
You can add, Tested-by: Santosh Shilimkar <Santosh.shilimkar at ti.com>
>
> ARM: Defer the L_PTE_EXEC flag setting to update_mmu_cache() on SMP
>
> From: Catalin Marinas <catalin.marinas at arm.com>
>
> On SMP systems, there is a small chance of a PTE becoming visible to a
> different CPU before the cache maintenance operations in
> update_mmu_cache(). This patch clears the L_PTE_EXEC bit in set_pte_at()
> but sets it later in update_mmu_cache() if vm_flags & VM_EXEC.
>
> Signed-off-by: Catalin Marinas <catalin.marinas at arm.com>
> ---
> arch/arm/include/asm/pgtable.h | 15 +++++++++++++++
> arch/arm/mm/fault-armv.c | 17 ++++++++++++-----
> 2 files changed, 27 insertions(+), 5 deletions(-)
>
> diff --git a/arch/arm/include/asm/pgtable.h b/arch/arm/include/asm/pgtable.h
> index 1139768..ee8cc13 100644
> --- a/arch/arm/include/asm/pgtable.h
> +++ b/arch/arm/include/asm/pgtable.h
> @@ -278,9 +278,24 @@ extern struct page *empty_zero_page;
>
> #define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,pte,ext)
>
> +#ifndef CONFIG_SMP
> #define set_pte_at(mm,addr,ptep,pteval) do { \
> set_pte_ext(ptep, pteval, (addr) >= TASK_SIZE ? 0 : PTE_EXT_NG); \
> } while (0)
> +#else
> +/*
> + * The L_PTE_EXEC attribute is later set in update_mmu_cache() to avoid a
> + * race with SMP systems executing from the new mapping before the cache
> + * flushing took place.
> + */
> +#define set_pte_at(mm,addr,ptep,pteval) do { \
> + if ((addr) >= TASK_SIZE) \
> + set_pte_ext(ptep, pteval, 0); \
> + else \
> + set_pte_ext(ptep, __pte(pte_val(pteval) & ~L_PTE_EXEC), \
> + PTE_EXT_NG); \
> + } while (0)
> +#endif
>
> /*
> * The following only work if pte_present() is true.
> diff --git a/arch/arm/mm/fault-armv.c b/arch/arm/mm/fault-armv.c
> index f9e9cbb..c01356d 100644
> --- a/arch/arm/mm/fault-armv.c
> +++ b/arch/arm/mm/fault-armv.c
> @@ -172,11 +172,18 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long addr,
> mapping = page_mapping(page);
> if (!test_and_set_bit(PG_dcache_clean, &page->flags))
> __flush_dcache_page(mapping, page);
> - if (mapping) {
> - if (cache_is_vivt())
> - make_coherent(mapping, vma, addr, ptep, pfn);
> - else if (vma->vm_flags & VM_EXEC)
> - __flush_icache_all();
> + if (!mapping)
> + return;
> +
> + if (cache_is_vivt())
> + make_coherent(mapping, vma, addr, ptep, pfn);
> + else if (vma->vm_flags & VM_EXEC) {
> + __flush_icache_all();
> +#ifdef CONFIG_SMP
> + set_pte_ext(ptep, __pte(pte_val(*ptep) | L_PTE_EXEC),
> + addr >= TASK_SIZE ? 0 : PTE_EXT_NG);
> + flush_tlb_page(vma, addr);
> +#endif
> }
> }
>
>
>
> --
Regards,
Santosh
More information about the linux-arm-kernel
mailing list