[PATCH v1] arm64/mm: Make set_ptes() robust when OAs cross 48-bit boundary
Catalin Marinas
catalin.marinas at arm.com
Thu Jan 25 09:07:25 PST 2024
On Tue, Jan 23, 2024 at 04:17:18PM +0000, Ryan Roberts wrote:
> diff --git a/arch/arm64/include/asm/pgtable.h b/arch/arm64/include/asm/pgtable.h
> index 79ce70fbb751..734b39401a05 100644
> --- a/arch/arm64/include/asm/pgtable.h
> +++ b/arch/arm64/include/asm/pgtable.h
> @@ -92,6 +92,14 @@ static inline pteval_t __phys_to_pte_val(phys_addr_t phys)
> #define pfn_pte(pfn,prot) \
> __pte(__phys_to_pte_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
>
> +/*
> + * Select all bits except the pfn
> + */
> +static inline pgprot_t pte_pgprot(pte_t pte)
> +{
> + return __pgprot(pte_val(pte) & ~PTE_ADDR_MASK);
> +}
> +
> #define pte_none(pte) (!pte_val(pte))
> #define pte_clear(mm,addr,ptep) set_pte(ptep, __pte(0))
> #define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
> @@ -341,6 +349,12 @@ static inline void __sync_cache_and_tags(pte_t pte, unsigned int nr_pages)
> mte_sync_tags(pte, nr_pages);
> }
>
> +#define pte_next_pfn pte_next_pfn
> +static inline pte_t pte_next_pfn(pte_t pte)
> +{
> + return pfn_pte(pte_pfn(pte) + 1, pte_pgprot(pte));
> +}
While I see why you wanted to optimise this, I'd rather keep the
pte_pgprot() change separate and at a later time. This will conflict
(fail to build) with Ard's patch removing PTE_ADDR_MASK:
https://lore.kernel.org/all/20240123145258.1462979-89-ardb+git@google.com/
This masking out is no longer straightforward with support for LPA2
(especially the 52-bit physical addresses with 4K pages): bits 8 and 9
of the PTE either contain bits 50, 51 of the PA or the shareability
attribute if FEAT_LPA2 is not present. In the latter case, we need them
preserved.
--
Catalin
More information about the linux-arm-kernel
mailing list