[PATCH 3/3] mm/pagewalk: Add pre/post_pte_table callback for lazy MMU on arm64
Dev Jain
dev.jain at arm.com
Fri May 30 02:04:07 PDT 2025
arm64 implements lazy_mmu_mode to allow deferral and batching of barriers
when updating kernel PTEs, which provides a nice performance boost. arm64
currently uses apply_to_page_range() to modify kernel PTE permissions,
which runs inside lazy_mmu_mode. So to prevent a performance regression,
let's add hooks to walk_page_range_novma() to allow continued use of
lazy_mmu_mode.
Signed-off-by: Dev Jain <dev.jain at arm.com>
---
Credits to Ryan for the patch description.
arch/arm64/mm/pageattr.c | 12 ++++++++++++
include/linux/pagewalk.h | 2 ++
mm/pagewalk.c | 6 ++++++
3 files changed, 20 insertions(+)
diff --git a/arch/arm64/mm/pageattr.c b/arch/arm64/mm/pageattr.c
index a5c829c64969..9163324b12a0 100644
--- a/arch/arm64/mm/pageattr.c
+++ b/arch/arm64/mm/pageattr.c
@@ -75,11 +75,23 @@ static int pageattr_pte_entry(pte_t *pte, unsigned long addr,
return 0;
}
+static void pte_lazy_mmu_enter(void)
+{
+ arch_enter_lazy_mmu_mode();
+}
+
+static void pte_lazy_mmu_leave(void)
+{
+ arch_leave_lazy_mmu_mode();
+}
+
static const struct mm_walk_ops pageattr_ops = {
.pud_entry = pageattr_pud_entry,
.pmd_entry = pageattr_pmd_entry,
.pte_entry = pageattr_pte_entry,
.walk_lock = PGWALK_NOLOCK,
+ .pre_pte_table = pte_lazy_mmu_enter,
+ .post_pte_table = pte_lazy_mmu_leave,
};
bool rodata_full __ro_after_init = IS_ENABLED(CONFIG_RODATA_FULL_DEFAULT_ENABLED);
diff --git a/include/linux/pagewalk.h b/include/linux/pagewalk.h
index 9bc8853ed3de..2157d345974c 100644
--- a/include/linux/pagewalk.h
+++ b/include/linux/pagewalk.h
@@ -88,6 +88,8 @@ struct mm_walk_ops {
int (*pre_vma)(unsigned long start, unsigned long end,
struct mm_walk *walk);
void (*post_vma)(struct mm_walk *walk);
+ void (*pre_pte_table)(void);
+ void (*post_pte_table)(void);
int (*install_pte)(unsigned long addr, unsigned long next,
pte_t *ptep, struct mm_walk *walk);
enum page_walk_lock walk_lock;
diff --git a/mm/pagewalk.c b/mm/pagewalk.c
index 9657cf4664b2..a441f5cbbc45 100644
--- a/mm/pagewalk.c
+++ b/mm/pagewalk.c
@@ -33,6 +33,9 @@ static int walk_pte_range_inner(pte_t *pte, unsigned long addr,
const struct mm_walk_ops *ops = walk->ops;
int err = 0;
+ if (walk->ops->pre_pte_table)
+ walk->ops->pre_pte_table();
+
for (;;) {
if (ops->install_pte && pte_none(ptep_get(pte))) {
pte_t new_pte;
@@ -56,6 +59,9 @@ static int walk_pte_range_inner(pte_t *pte, unsigned long addr,
addr += PAGE_SIZE;
pte++;
}
+
+ if (walk->ops->post_pte_table)
+ walk->ops->post_pte_table();
return err;
}
--
2.30.2
More information about the linux-arm-kernel
mailing list