[PATCH] arm64: kernel: restrict /dev/mem read() calls to linear region

Ard Biesheuvel ard.biesheuvel at linaro.org
Wed Apr 12 04:31:38 EDT 2017


On 12 April 2017 at 09:29, Alexander Graf <agraf at suse.de> wrote:
>
>
> On 12.04.17 10:26, Ard Biesheuvel wrote:
>>
>> When running lscpu on an AArch64 system that has SMBIOS version 2.0
>> tables, it will segfault in the following way:
>>
>>   Unable to handle kernel paging request at virtual address
>> ffff8000bfff0000
>>   pgd = ffff8000f9615000
>>   [ffff8000bfff0000] *pgd=0000000000000000
>>   Internal error: Oops: 96000007 [#1] PREEMPT SMP
>>   Modules linked in:
>>   CPU: 0 PID: 1284 Comm: lscpu Not tainted 4.11.0-rc3+ #103
>>   Hardware name: QEMU QEMU Virtual Machine, BIOS 0.0.0 02/06/2015
>>   task: ffff8000fa78e800 task.stack: ffff8000f9780000
>>   PC is at __arch_copy_to_user+0x90/0x220
>>   LR is at read_mem+0xcc/0x140
>>
>> This is caused by the fact that lspci issues a read() on /dev/mem at the
>> offset where it expects to find the SMBIOS structure array. However, this
>> region is classified as EFI_RUNTIME_SERVICE_DATA (as per the UEFI spec),
>> and so it is omitted from the linear mapping.
>>
>> So let's restrict /dev/mem read/write access to those areas that are
>> covered by the linear region.
>>
>> Reported-by: Alexander Graf <agraf at suse.de>
>> Fixes: 4dffbfc48d65 ("arm64/efi: mark UEFI reserved regions as
>> MEMBLOCK_NOMAP")
>> Signed-off-by: Ard Biesheuvel <ard.biesheuvel at linaro.org>
>> ---
>>  arch/arm64/mm/mmap.c | 9 +++------
>>  1 file changed, 3 insertions(+), 6 deletions(-)
>>
>> diff --git a/arch/arm64/mm/mmap.c b/arch/arm64/mm/mmap.c
>> index 7b0d55756eb1..2956240d17d7 100644
>> --- a/arch/arm64/mm/mmap.c
>> +++ b/arch/arm64/mm/mmap.c
>> @@ -18,6 +18,7 @@
>>
>>  #include <linux/elf.h>
>>  #include <linux/fs.h>
>> +#include <linux/memblock.h>
>>  #include <linux/mm.h>
>>  #include <linux/mman.h>
>>  #include <linux/export.h>
>> @@ -103,12 +104,8 @@ void arch_pick_mmap_layout(struct mm_struct *mm)
>>   */
>>  int valid_phys_addr_range(phys_addr_t addr, size_t size)
>>  {
>> -       if (addr < PHYS_OFFSET)
>> -               return 0;
>> -       if (addr + size > __pa(high_memory - 1) + 1)
>> -               return 0;
>> -
>> -       return 1;
>> +       return memblock_is_map_memory(addr) &&
>> +              memblock_is_map_memory(addr + size - 1);
>
>
> Is that safe? Are we guaranteed that size is less than one page? Otherwise,
> someone could map a region that spans over a reserved one:
>
>   [conv mem]
>   [reserved]
>   [conv mem]
>

Well, I will leave it to the maintainers to decide how elaborate they
want this logic to become, given that read()ing from /dev/mem is
something we are not eager to support in the first place.

But indeed, if the start and end of the region are covered by the
linear region, there could potentially be an uncovered hole in the
middle.



More information about the linux-arm-kernel mailing list