[PATCH] arm64: mm: move zero page from .bss to right before swapper_pg_dir

Ard Biesheuvel ard.biesheuvel at linaro.org
Mon Sep 12 07:17:11 PDT 2016


On 12 September 2016 at 13:57, Mark Rutland <mark.rutland at arm.com> wrote:
> Hi,
>
> On Sun, Sep 11, 2016 at 03:38:34PM +0100, Ard Biesheuvel wrote:
>> Move the statically allocated zero page from the .bss section to right
>> before swapper_pg_dir. This allows us to refer to its physical address
>> by simply reading TTBR1_EL1 (which always points to swapper_pg_dir and
>> always has its ASID field cleared), and subtracting PAGE_SIZE.
>
> On a conflicting note, I was hoping to move the zero page into .rodata
> so as to catch any erroneous modification.
>
> Given that we can't rely on TTBR1 poiting at the swapper_pg_dir, that
> leaves us with Image size reduction vs RO-ification.
>
> Any thoughts/preference?
>

That's a good point. v2 coming up ...


>> Inspired-by: http://marc.info/?l=linux-arm-kernel&m=147282867511801
>> Signed-off-by: Ard Biesheuvel <ard.biesheuvel at linaro.org>
>> ---
>>  arch/arm64/include/asm/mmu_context.h | 10 ++++++----
>>  arch/arm64/kernel/head.S             |  1 -
>>  arch/arm64/kernel/vmlinux.lds.S      |  2 ++
>>  arch/arm64/mm/mmu.c                  |  1 -
>>  4 files changed, 8 insertions(+), 6 deletions(-)
>>
>> diff --git a/arch/arm64/include/asm/mmu_context.h b/arch/arm64/include/asm/mmu_context.h
>> index b1892a0dbcb0..94461ba5febd 100644
>> --- a/arch/arm64/include/asm/mmu_context.h
>> +++ b/arch/arm64/include/asm/mmu_context.h
>> @@ -49,13 +49,15 @@ static inline void contextidr_thread_switch(struct task_struct *next)
>>   */
>>  static inline void cpu_set_reserved_ttbr0(void)
>>  {
>> -     unsigned long ttbr = virt_to_phys(empty_zero_page);
>> +     unsigned long dummy;
>>
>> -     asm(
>> +     asm volatile(
>> +     "       mrs     %0, ttbr1_el1                   // get TTBR1\n"
>> +     "       sub     %0, %0, %1                      // subtract PAGE_SIZE\n"
>>       "       msr     ttbr0_el1, %0                   // set TTBR0\n"
>>       "       isb"
>> -     :
>> -     : "r" (ttbr));
>> +     : "=&r" (dummy)
>> +     : "I" (PAGE_SIZE));
>>  }
>>
>>  /*
>> diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S
>> index 8bc9458f9add..6020b884b076 100644
>> --- a/arch/arm64/kernel/head.S
>> +++ b/arch/arm64/kernel/head.S
>> @@ -449,7 +449,6 @@ __primary_switched:
>>       adr_l   x2, __bss_stop
>>       sub     x2, x2, x0
>>       bl      __pi_memset
>> -     dsb     ishst                           // Make zero page visible to PTW
>>
>>  #ifdef CONFIG_KASAN
>>       bl      kasan_early_init
>> diff --git a/arch/arm64/kernel/vmlinux.lds.S b/arch/arm64/kernel/vmlinux.lds.S
>> index 659963d40bb4..a14eb8ff5144 100644
>> --- a/arch/arm64/kernel/vmlinux.lds.S
>> +++ b/arch/arm64/kernel/vmlinux.lds.S
>> @@ -193,6 +193,8 @@ SECTIONS
>>       . = ALIGN(PAGE_SIZE);
>>       idmap_pg_dir = .;
>>       . += IDMAP_DIR_SIZE;
>> +     empty_zero_page = .;
>> +     . += PAGE_SIZE;
>>       swapper_pg_dir = .;
>>       . += SWAPPER_DIR_SIZE;
>>
>> diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c
>> index 4989948d1feb..539ce9d11325 100644
>> --- a/arch/arm64/mm/mmu.c
>> +++ b/arch/arm64/mm/mmu.c
>> @@ -53,7 +53,6 @@ EXPORT_SYMBOL(kimage_voffset);
>>   * Empty_zero_page is a special page that is used for zero-initialized data
>>   * and COW.
>>   */
>> -unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)] __page_aligned_bss;
>>  EXPORT_SYMBOL(empty_zero_page);
>>
>>  static pte_t bm_pte[PTRS_PER_PTE] __page_aligned_bss;
>> --
>> 2.7.4
>>
>>
>> _______________________________________________
>> linux-arm-kernel mailing list
>> linux-arm-kernel at lists.infradead.org
>> http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
>>



More information about the linux-arm-kernel mailing list