[PATCH] ARC: mm: PAE40: Cast pfn to pte_t in pfn_pte() macro
Yuriy Kolerov
Yuriy.Kolerov at synopsys.com
Tue Nov 29 03:37:39 PST 2016
> -----Original Message-----
> From: Alexey Brodkin [mailto:abrodkin at synopsys.com]
> Sent: Monday, November 28, 2016 2:43 PM
> To: yuriy.kolerov at synopsys.com
> Cc: linux-kernel at vger.kernel.org; Alexey.Brodkin at synopsys.com; Vineet
> Gupta <vgupta at synopsys.com>; linux-snps-arc at lists.infradead.org
> Subject: Re: [PATCH] ARC: mm: PAE40: Cast pfn to pte_t in pfn_pte() macro
>
> Hi Yuriy,
>
> Really nice catch!
> Though a couple of nitpicks below.
>
> On Mon, 2016-11-28 at 07:07 +0300, Yuriy Kolerov wrote:
> > Originally pfn_pte(pfn, prot) macro had this definition:
> >
> > __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
> >
> > The value of pfn (Page Frame Number) is shifted to the left to get the
> > value of pte (Page Table Entry). Usually a 4-byte value is passed to
> > this macro as value of pfn. However if Linux is configured with
> > support of PAE40 then value of pte has 8-byte type because it must
> > contain additional 8 bits of the physical address. Thus if value of
> > pfn represents a physical page frame above of 4GB boundary then
> > shifting of pfn to the left by PAGE_SHIFT wipes most significant bits
> > of the 40-bit physical address.
> >
> > As a result all physical addresses above of 4GB boundary in systems
> > with PAE40 are mapped to virtual address incorrectly. An error may
> > occur when the kernel tries to unmap such bad pages:
> >
> > [ECR ]: 0x00050100 => Invalid Read @ 0x41414144 by insn @
> > 0x801644c6
> > [EFA ]: 0x41414144
> > [BLINK ]: unmap_page_range+0x134/0x700
> > [ERET ]: unmap_page_range+0x17a/0x700
> > [STAT32]: 0x8008021e : IE K
> > BTA: 0x801644c6 SP: 0x901a5e84 FP: 0x5ff35de8
> > LPS: 0x8026462c LPE: 0x80264630 LPC: 0x00000000
> > r00: 0x8fcc4fc0 r01: 0x2fe68000 r02: 0x41414140
> > r03: 0x2c05c000 r04: 0x2fe6a000 r05: 0x0009ffff
> > r06: 0x901b6898 r07: 0x2fe68000 r08: 0x00000001
> > r09: 0x804a807c r10: 0x0000067e r11: 0xffffffff
> > r12: 0x80164480
> > Stack Trace:
> > unmap_page_range+0x17a/0x700
> > unmap_vmas+0x46/0x64
> > do_munmap+0x210/0x450
> > SyS_munmap+0x2c/0x50
> > EV_Trap+0xfc/0x100
>
> This example makes not much sense in its current form.
> I'd like to see how mentioned above problem leads to this failure. I.e. pfn =
> 0xXXX gave pte = 0xYYY and at truncated to 0xYYY address there's no data we
> expected thus reading from 0x41414144 end up in exception etc.
>
> > So the value of pfn must be casted to pte_t before shifting to ensure
> > that 40-bit address will not be truncated:
> >
> > __pte(((pte_t) (pfn) << PAGE_SHIFT) | pgprot_val(prot))
> >
> > Signed-off-by: Yuriy Kolerov <yuriy.kolerov at synopsys.com>
> > ---
> > arch/arc/include/asm/pgtable.h | 3 ++-
> > 1 file changed, 2 insertions(+), 1 deletion(-)
> >
> > diff --git a/arch/arc/include/asm/pgtable.h
> > b/arch/arc/include/asm/pgtable.h index 89eeb37..77bc51c 100644
> > --- a/arch/arc/include/asm/pgtable.h
> > +++ b/arch/arc/include/asm/pgtable.h
> > @@ -280,7 +280,8 @@ static inline void pmd_set(pmd_t *pmdp, pte_t
> > *ptep)
> >
> > #define pte_page(pte) pfn_to_page(pte_pfn(pte))
> > #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
> > -#define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) |
> pgprot_val(prot))
> > +#define pfn_pte(pfn, prot) \
> > + __pte(((pte_t) (pfn) << PAGE_SHIFT) | pgprot_val(prot))
>
> I think it's better to split it in a bit different manner like:
> --------------------------------->8-----------------------------
> #define pfn_pte(pfn, prot) __pte(((pte_t) (pfn) << PAGE_SHIFT) | \
> pgprot_val(prot))
> --------------------------------->8-----------------------------
>
> Also see how this macro is implemented for example on ARM:
> http://lxr.free-electrons.com/source/arch/arm/include/asm/pgtable.h#L211
> -------------------->8------------------
> #define pfn_pte(pfn,prot) __pte(__pfn_to_phys(pfn) |
> pgprot_val(prot))
> -------------------->8------------------
>
> Where __pfn_to_phys() is:
> http://lxr.free-electrons.com/source/include/asm-
> generic/memory_model.h#L78
> -------------------->8------------------
> #define __pfn_to_phys(pfn) PFN_PHYS(pfn)
> -------------------->8------------------
>
> PFN_PHYS() is:
> http://lxr.free-electrons.com/source/include/linux/pfn.h#L20
> -------------------->8------------------
> #define PFN_PHYS(x) ((phys_addr_t)(x) << PAGE_SHIFT)
> -------------------->8------------------
>
> And finally phys_addr_t is:
> http://lxr.free-electrons.com/source/include/linux/types.h#L161
> -------------------->8------------------
> #ifdef CONFIG_PHYS_ADDR_T_64BIT
> typedef u64 phys_addr_t;
> #else
> typedef u32 phys_addr_t;
> #endif
> -------------------->8------------------
>
> Not really sure though which implementation is better.
> I like your approach because its simplicity instead of another couple of layers
> of definitions but maybe there's a reason for this kind of complication. Funny
> enough other arches take their own approaches ranging from the same you
> did to this __pfn_to_phys() to casting to "long long".
Yes, you are right. __pfn_to_phys() does what is needed.
> -Alexey
More information about the linux-snps-arc
mailing list