[PATCH v4 09/13] arm64: mm: explicitly bootstrap the linear mapping

Catalin Marinas catalin.marinas at arm.com
Fri May 8 09:43:28 PDT 2015


On Fri, May 08, 2015 at 05:03:37PM +0200, Ard Biesheuvel wrote:
> On 8 May 2015 at 16:44, Catalin Marinas <catalin.marinas at arm.com> wrote:
> > On Thu, May 07, 2015 at 09:21:28PM +0200, Ard Biesheuvel wrote:
> >> On 7 May 2015 at 18:54, Catalin Marinas <catalin.marinas at arm.com> wrote:
> >> > On Wed, Apr 15, 2015 at 05:34:20PM +0200, Ard Biesheuvel wrote:
> >> >> diff --git a/arch/arm64/kernel/vmlinux.lds.S b/arch/arm64/kernel/vmlinux.lds.S
> >> >> index ceec4def354b..338eaa7bcbfd 100644
> >> >> --- a/arch/arm64/kernel/vmlinux.lds.S
> >> >> +++ b/arch/arm64/kernel/vmlinux.lds.S
> >> >> @@ -68,6 +68,17 @@ PECOFF_FILE_ALIGNMENT = 0x200;
> >> >>  #define ALIGN_DEBUG_RO_MIN(min)              . = ALIGN(min);
> >> >>  #endif
> >> >>
> >> >> +/*
> >> >> + * The pgdir region needs to be mappable using a single PMD or PUD sized region,
> >> >> + * so it should not cross a 512 MB or 1 GB alignment boundary, respectively
> >> >> + * (depending on page size). So align to an upper bound of its size.
> >> >> + */
> >> >> +#if CONFIG_ARM64_PGTABLE_LEVELS == 2
> >> >> +#define PGDIR_ALIGN  (8 * PAGE_SIZE)
> >> >> +#else
> >> >> +#define PGDIR_ALIGN  (16 * PAGE_SIZE)
> >> >> +#endif
> >> >
> >> > Isn't 8 pages sufficient in both cases? Unless some other patch changes
> >> > the idmap and swapper, I can count maximum 7 pages in total.
> >>
> >> The preceding patch moves the fixmap page tables to this region as well.
> >> But the logic is still incorrect -> we only need 16x for 4 levels (7 +
> >> 3 == 10), the remaining ones are all <= 8
> >
> > You should improve the comment here to include the maths, "upper bound
> > of its size" is not very clear ;).
> 
> Yes, you are right, it should read 'power-of-2 upper bound'

And the number of pages required for the initial page tables (I figured
out it's a power of two already ;)).

> >> >> +     static struct bootstrap_pgtables linear_bs_pgtables __pgdir;
> >> >> +     const phys_addr_t swapper_phys = __pa(swapper_pg_dir);
> >> >> +     unsigned long swapper_virt = __phys_to_virt(swapper_phys) + va_offset;
> >> >> +     struct memblock_region *reg;
> >> >> +
> >> >> +     bootstrap_early_mapping(swapper_virt, &linear_bs_pgtables,
> >> >> +                             IS_ENABLED(CONFIG_ARM64_64K_PAGES));
> >> >> +
> >> >> +     /* now find the memblock that covers swapper_pg_dir, and clip */
> >> >> +     for_each_memblock(memory, reg) {
> >> >> +             phys_addr_t start = reg->base;
> >> >> +             phys_addr_t end = start + reg->size;
> >> >> +             unsigned long vstart, vend;
> >> >> +
> >> >> +             if (start > swapper_phys || end <= swapper_phys)
> >> >> +                     continue;
> >> >> +
> >> >> +#ifdef CONFIG_ARM64_64K_PAGES
> >> >> +             /* clip the region to PMD size */
> >> >> +             vstart = max(swapper_virt & PMD_MASK,
> >> >> +                          round_up(__phys_to_virt(start + va_offset),
> >> >> +                                   PAGE_SIZE));
> >> >> +             vend = min(round_up(swapper_virt, PMD_SIZE),
> >> >> +                        round_down(__phys_to_virt(end + va_offset),
> >> >> +                                   PAGE_SIZE));
> >> >> +#else
> >> >> +             /* clip the region to PUD size */
> >> >> +             vstart = max(swapper_virt & PUD_MASK,
> >> >> +                          round_up(__phys_to_virt(start + va_offset),
> >> >> +                                   PMD_SIZE));
> >> >> +             vend = min(round_up(swapper_virt, PUD_SIZE),
> >> >> +                        round_down(__phys_to_virt(end + va_offset),
> >> >> +                                   PMD_SIZE));
> >> >> +#endif
> >> >> +
> >> >> +             create_mapping(__pa(vstart - va_offset), vstart, vend - vstart,
> >> >> +                            PAGE_KERNEL_EXEC);
> >> >> +
> >> >> +             /*
> >> >> +              * Temporarily limit the memblock range. We need to do this as
> >> >> +              * create_mapping requires puds, pmds and ptes to be allocated
> >> >> +              * from memory addressable from the early linear mapping.
> >> >> +              */
> >> >> +             memblock_set_current_limit(__pa(vend - va_offset));
> >> >> +
> >> >> +             return;
> >> >> +     }
> >> >> +     BUG();
> >> >> +}
> >> >
> >> > I'll probably revisit this function after I see the whole series. But in
> >> > the meantime, if the kernel is not loaded in the first memblock (in
> >> > address order), isn't there a risk that we allocate memory from the
> >> > first memblock which is not mapped yet?
> >>
> >> memblock allocates top down, so it should only allocate from this
> >> region, unless the remaining room is completely reserved.
> >
> > I don't like to rely on this, it's not guaranteed behaviour.
> 
> Actually, it is. Allocation is always top-down unless you call
> memblock_set_bottom_up(), which is a NOP  unless CONFIG_MOVABLE_NODE
> is selected.
> That is why the memblock limit only limits at the top afaict

It currently works like this but I'm not sure it is guaranteed to always
behave this way (e.g. someone "improves" the memblock allocator in the
future). And you never know, we may need memory hotplug on arm64 at some
point in the future (together with CONFIG_MOVABLE_NODE).

-- 
Catalin



More information about the linux-arm-kernel mailing list