[PATCH] arm64: reserve [_text, _stext) virtual address range
Will Deacon
will at kernel.org
Tue Mar 11 05:54:15 PDT 2025
[+Ard]
On Mon, Mar 10, 2025 at 01:05:04PM -0700, Omar Sandoval wrote:
> From: Omar Sandoval <osandov at fb.com>
>
> Since the referenced fixes commit, the kernel's .text section is only
> mapped starting from _stext; the region [_text, _stext) is omitted. As a
> result, other vmalloc/vmap allocations may use the virtual addresses
> nominally in the range [_text, _stext). This address reuse confuses
> multiple things:
>
> 1. crash_prepare_elf64_headers() sets up a segment in /proc/vmcore
> mapping the entire range [_text, _end) to
> [__pa_symbol(_text), __pa_symbol(_end)). Reading an address in
> [_text, _stext) from /proc/vmcore therefore gives the incorrect
> result.
> 2. Tools doing symbolization (either by reading /proc/kallsyms or based
> on the vmlinux ELF file) will incorrectly identify vmalloc/vmap
> allocations in [_text, _stext) as kernel symbols.
>
> In practice, both of these issues affect the drgn debugger.
> Specifically, there were cases where the vmap IRQ stacks for some CPUs
> were allocated in [_text, _stext). As a result, drgn could not get the
> stack trace for a crash in an IRQ handler because the core dump
> contained invalid data for the IRQ stack address. The stack addresses
> were also symbolized as being in the _text symbol.
>
> Fix this by creating an unmapped vm_area to cover [_text, _stext). This
> prevents other allocations from using it while still achieving the
> original goal of not mapping unpredictable data.
>
> Fixes: e2a073dde921 ("arm64: omit [_text, _stext) from permanent kernel mapping")
> Cc: stable at vger.kernel.org
> Signed-off-by: Omar Sandoval <osandov at fb.com>
> ---
> Based on v6.14-rc6.
>
> Thanks!
>
> arch/arm64/mm/mmu.c | 26 +++++++++++++++++---------
> 1 file changed, 17 insertions(+), 9 deletions(-)
>
> diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c
> index b4df5bc5b1b8..88595ea12f39 100644
> --- a/arch/arm64/mm/mmu.c
> +++ b/arch/arm64/mm/mmu.c
> @@ -703,10 +703,14 @@ static void __init declare_vma(struct vm_struct *vma,
> void *va_start, void *va_end,
> unsigned long vm_flags)
> {
> - phys_addr_t pa_start = __pa_symbol(va_start);
> + phys_addr_t pa_start = 0;
> unsigned long size = va_end - va_start;
>
> - BUG_ON(!PAGE_ALIGNED(pa_start));
> + if (vm_flags & VM_MAP) {
> + pa_start = __pa_symbol(va_start);
> + BUG_ON(!PAGE_ALIGNED(pa_start));
> + }
> +
> BUG_ON(!PAGE_ALIGNED(size));
>
> if (!(vm_flags & VM_NO_GUARD))
> @@ -715,7 +719,7 @@ static void __init declare_vma(struct vm_struct *vma,
> vma->addr = va_start;
> vma->phys_addr = pa_start;
> vma->size = size;
> - vma->flags = VM_MAP | vm_flags;
> + vma->flags = vm_flags;
> vma->caller = __builtin_return_address(0);
>
> vm_area_add_early(vma);
> @@ -765,13 +769,17 @@ core_initcall(map_entry_trampoline);
> */
> static void __init declare_kernel_vmas(void)
> {
> - static struct vm_struct vmlinux_seg[KERNEL_SEGMENT_COUNT];
> + static struct vm_struct vmlinux_seg[KERNEL_SEGMENT_COUNT + 1];
>
> - declare_vma(&vmlinux_seg[0], _stext, _etext, VM_NO_GUARD);
> - declare_vma(&vmlinux_seg[1], __start_rodata, __inittext_begin, VM_NO_GUARD);
> - declare_vma(&vmlinux_seg[2], __inittext_begin, __inittext_end, VM_NO_GUARD);
> - declare_vma(&vmlinux_seg[3], __initdata_begin, __initdata_end, VM_NO_GUARD);
> - declare_vma(&vmlinux_seg[4], _data, _end, 0);
> + declare_vma(&vmlinux_seg[0], _text, _stext, VM_NO_GUARD);
Should we also put the memblock reservation back as it was, so that this
region can't be allocated there?
In fact, if we're not allocating from here, why don't we just map it
anyway but without execute permissions?
Will
More information about the linux-arm-kernel
mailing list