[PATCH] Revert "arm64: mm: set the contiguous bit for kernel mappings where appropriate"
Ard Biesheuvel
ard.biesheuvel at linaro.org
Thu Feb 23 08:25:26 PST 2017
On 23 February 2017 at 16:22, Mark Rutland <mark.rutland at arm.com> wrote:
> This reverts commit 0bfc445dec9dd8130d22c9f4476eed7598524129.
>
> When we change the permissions of regions mapped using contiguous
> entries, the architecture requires us to follow a Break-Before-Make
> strategy, breaking *all* associated entries before we can change any of
> the following properties from the entries:
>
> - presence of the contiguous bit
> - output address
> - attributes
> - permissiones
>
> Failure to do so can result in a number of problems (e.g. TLB conflict
> aborts and/or erroneous results from TLB lookups).
>
> See ARM DDI 0487A.k_iss10775, "Misprogramming of the Contiguous bit",
> page D4-1762.
>
> We do not take this into account when altering the permissions of kernel
> segments in mark_rodata_ro(), where we change the permissions of live
> contiguous entires one-by-one, leaving them transiently inconsistent.
> This has been observed to result in failures on some fast model
> configurations.
>
> Unfortunately, we cannot follow Break-Before-Make here as we'd have to
> unmap kernel text and data used to perform the sequence.
>
> For the timebeing, revert commit 0bfc445dec9dd813 so as to avoid issues
> resulting from this misuse of the contiguous bit.
>
> Signed-off-by: Mark Rutland <mark.rutland at arm.com>
> Reported-by: Jean-Philippe Brucker <jean-philippe.brucker at arm.com>
> Cc: Ard Biesheuvel <ard.biesheuvel at linaro.org>
> Cc: Catalin Marinas <catalin.marinas at arm.com>
> Cc: Will Deacon <Will.Deacon at arm.com>
> Cc: stable at vger.kernel.org # v4.10
Acked-by: Ard Biesheuvel <ard.biesheuvel at linaro.org>
> ---
> arch/arm64/mm/mmu.c | 34 ++++------------------------------
> 1 file changed, 4 insertions(+), 30 deletions(-)
>
> I'm aware that our hugtlbpage code has a similar issue. I'm looking into that
> now, and will address that with separate patches. It should be possible to use
> BBM there as it's a userspace mapping.
>
Are you looking into this issue as well? Or should I?
> diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c
> index 17243e4..c391b1f 100644
> --- a/arch/arm64/mm/mmu.c
> +++ b/arch/arm64/mm/mmu.c
> @@ -108,10 +108,8 @@ static bool pgattr_change_is_safe(u64 old, u64 new)
> static void alloc_init_pte(pmd_t *pmd, unsigned long addr,
> unsigned long end, unsigned long pfn,
> pgprot_t prot,
> - phys_addr_t (*pgtable_alloc)(void),
> - bool page_mappings_only)
> + phys_addr_t (*pgtable_alloc)(void))
> {
> - pgprot_t __prot = prot;
> pte_t *pte;
>
> BUG_ON(pmd_sect(*pmd));
> @@ -129,18 +127,7 @@ static void alloc_init_pte(pmd_t *pmd, unsigned long addr,
> do {
> pte_t old_pte = *pte;
>
> - /*
> - * Set the contiguous bit for the subsequent group of PTEs if
> - * its size and alignment are appropriate.
> - */
> - if (((addr | PFN_PHYS(pfn)) & ~CONT_PTE_MASK) == 0) {
> - if (end - addr >= CONT_PTE_SIZE && !page_mappings_only)
> - __prot = __pgprot(pgprot_val(prot) | PTE_CONT);
> - else
> - __prot = prot;
> - }
> -
> - set_pte(pte, pfn_pte(pfn, __prot));
> + set_pte(pte, pfn_pte(pfn, prot));
> pfn++;
>
> /*
> @@ -159,7 +146,6 @@ static void alloc_init_pmd(pud_t *pud, unsigned long addr, unsigned long end,
> phys_addr_t (*pgtable_alloc)(void),
> bool page_mappings_only)
> {
> - pgprot_t __prot = prot;
> pmd_t *pmd;
> unsigned long next;
>
> @@ -186,18 +172,7 @@ static void alloc_init_pmd(pud_t *pud, unsigned long addr, unsigned long end,
> /* try section mapping first */
> if (((addr | next | phys) & ~SECTION_MASK) == 0 &&
> !page_mappings_only) {
> - /*
> - * Set the contiguous bit for the subsequent group of
> - * PMDs if its size and alignment are appropriate.
> - */
> - if (((addr | phys) & ~CONT_PMD_MASK) == 0) {
> - if (end - addr >= CONT_PMD_SIZE)
> - __prot = __pgprot(pgprot_val(prot) |
> - PTE_CONT);
> - else
> - __prot = prot;
> - }
> - pmd_set_huge(pmd, phys, __prot);
> + pmd_set_huge(pmd, phys, prot);
>
> /*
> * After the PMD entry has been populated once, we
> @@ -207,8 +182,7 @@ static void alloc_init_pmd(pud_t *pud, unsigned long addr, unsigned long end,
> pmd_val(*pmd)));
> } else {
> alloc_init_pte(pmd, addr, next, __phys_to_pfn(phys),
> - prot, pgtable_alloc,
> - page_mappings_only);
> + prot, pgtable_alloc);
>
> BUG_ON(pmd_val(old_pmd) != 0 &&
> pmd_val(old_pmd) != pmd_val(*pmd));
> --
> 1.9.1
>
More information about the linux-arm-kernel
mailing list