[PATCH] arm64: mte: Use PAGE_KERNEL_TAGGED in arch_add_memory
David Hildenbrand
david at redhat.com
Fri Mar 5 15:55:28 GMT 2021
On 05.03.21 16:43, Catalin Marinas wrote:
> On Fri, Mar 05, 2021 at 11:19:08AM +0000, Catalin Marinas wrote:
>> On Tue, Mar 02, 2021 at 08:21:03PM -0800, Patrick Daly wrote:
>>> In a system which supports MTE, the linear kernel region must allow
>>> reading/writing allocation tags. For memory present at boot this
>>> is already being done in map_mem(). Add the same in arch_add_memory().
>>>
>>> Signed-off-by: Patrick Daly <pdaly at codeaurora.org>
>>> ---
>>> arch/arm64/mm/mmu.c | 12 +++++++++++-
>>> 1 file changed, 11 insertions(+), 1 deletion(-)
>>>
>>> diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c
>>> index 9b25d60b..0fcfe90 100644
>>> --- a/arch/arm64/mm/mmu.c
>>> +++ b/arch/arm64/mm/mmu.c
>>> @@ -1463,6 +1463,7 @@ int arch_add_memory(int nid, u64 start, u64 size,
>>> struct mhp_params *params)
>>> {
>>> int ret, flags = 0;
>>> + pgprot_t pgprot;
>>>
>>> if (!inside_linear_region(start, size)) {
>>> pr_err("[%llx %llx] is outside linear mapping region\n", start, start + size);
>>> @@ -1477,8 +1478,17 @@ int arch_add_memory(int nid, u64 start, u64 size,
>>> IS_ENABLED(CONFIG_KFENCE))
>>> flags = NO_BLOCK_MAPPINGS | NO_CONT_MAPPINGS;
>>>
>>> + /*
>>> + * The linear map must allow allocation tags reading/writing
>>> + * if MTE is present. Otherwise, it has the same attributes as
>>> + * PAGE_KERNEL.
>>> + */
>>> + pgprot = params->pgprot;
>>> + if (pgprot_val(pgprot) == pgprot_val(PAGE_KERNEL))
>>> + pgprot = PAGE_KERNEL_TAGGED;
>>> +
>>> __create_pgd_mapping(swapper_pg_dir, start, __phys_to_virt(start),
>>> - size, params->pgprot, __pgd_pgtable_alloc,
>>> + size, pgprot, __pgd_pgtable_alloc,
>>> flags);
>>
>> We'll need a similar pattern for vmalloc() once we have the khwasan
>> support in place. So we could add a pgprot_tagged() function (similar to
>> pgprot_writecombine() etc.) which does the above check and returns
>> PAGE_KERNEL_TAGGED, maybe only checking the PTE_ATTRINDX_MASK bits
>> rather than the whole prot bits.
>>
>> However, the bigger problem is that arch_add_memory() is also called for
>> ZONE_DEVICE mappings and we can't always guarantee that such range
>> supports tagging (most likely it doesn't, e.g. persistent memory),
>> leading to potential external aborts.
>>
>> One option is to expand mhp_params to pass additional information so
>> that the arch code can make the right decision. Another option is to
>> make PAGE_KERNEL_TAGGED global which is just PAGE_KERNEL for all the
>> other architectures and use it in the core code.
>>
>> Yet another option which we haven't fully explored with MTE is to have
>> PAGE_KERNEL always tagged but add a new PAGE_KERNEL_DEVICE (or
>> _UNTAGGED) for specific cases like ZONE_DEVICE. We need to make sure
>> that PAGE_KERNEL doesn't end up in places where the backing memory does
>> not support tags.
>>
>> I'll give the last option a quick try and see if it falls apart (just
>> changing PAGE_KERNEL to tagged). In terms of tag cache usage, it
>> probably won't have much of an impact since the whole of the linear map
>> is tagged already.
>
> I played with this a bit and the last option is not really feasible.
> There are several places in the kernel where PAGE_KERNEL is used just
> because the expectation is for write-back memory. For MTE, we need
> tagged memory either because the kernel allocators need it (with
> KASAN_HW_TAGS) or because it is shared with the user, potentially mapped
> as tagged in user space and such tags need to be preserved by the
> kernel. So for the latter, only the linear map needs to be tagged (and
> memory hotplug falls into this category).
>
> This leaves us with one of the first two options. I think the easiest
> is:
>
> ---------------8<-------------------------------
> diff --git a/arch/arm64/include/asm/pgtable.h b/arch/arm64/include/asm/pgtable.h
> index e17b96d0e4b5..5c78b92d9ec5 100644
> --- a/arch/arm64/include/asm/pgtable.h
> +++ b/arch/arm64/include/asm/pgtable.h
> @@ -486,6 +486,8 @@ static inline pmd_t pmd_mkdevmap(pmd_t pmd)
> __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_NORMAL_NC) | PTE_PXN | PTE_UXN)
> #define pgprot_device(prot) \
> __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRE) | PTE_PXN | PTE_UXN)
> +#define pgprot_tagged(prot) \
> + __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_NORMAL_TAGGED))
> /*
> * DMA allocations for non-coherent devices use what the Arm architecture calls
> * "Normal non-cacheable" memory, which permits speculation, unaligned accesses
> diff --git a/include/linux/pgtable.h b/include/linux/pgtable.h
> index cdfc4e9f253e..f5f5044db2ce 100644
> --- a/include/linux/pgtable.h
> +++ b/include/linux/pgtable.h
> @@ -904,6 +904,10 @@ static inline void ptep_modify_prot_commit(struct vm_area_struct *vma,
> #define pgprot_device pgprot_noncached
> #endif
>
> +#ifndef pgprot_tagged
> +#define pgprot_tagged(prot) (prot)
> +#endif
> +
> #ifdef CONFIG_MMU
> #ifndef pgprot_modify
> #define pgprot_modify pgprot_modify
> diff --git a/mm/memory_hotplug.c b/mm/memory_hotplug.c
> index 5ba51a8bdaeb..4253d80a59ba 100644
> --- a/mm/memory_hotplug.c
> +++ b/mm/memory_hotplug.c
> @@ -1072,7 +1072,7 @@ static int online_memory_block(struct memory_block *mem, void *arg)
> */
> int __ref add_memory_resource(int nid, struct resource *res, mhp_t mhp_flags)
> {
> - struct mhp_params params = { .pgprot = PAGE_KERNEL };
> + struct mhp_params params = { .pgprot = pgprot_tagged(PAGE_KERNEL) };
This looks like we're pushing arch specific stuff ("tagged") in here.
Can't we generalize this to something like
pgprot_mhp_default
(or a better name)
that defaults to PAGE_KERNEL on all architectures except arm64 which
overwrites this somehow?
--
Thanks,
David / dhildenb
More information about the linux-arm-kernel
mailing list