[PATCH v2 16/30] KVM: arm64: Move fault context to const structure
Anshuman Khandual
anshuman.khandual at arm.com
Tue Mar 31 19:53:51 PDT 2026
On 27/03/26 5:06 PM, Marc Zyngier wrote:
> In order to make it clearer what gets updated or not during fault
> handling, move a set of information that losely represents the
> fault context.
>
> This gets populated early, from handle_mem_abort(), and gets passed
> along as a const pointer. user_mem_abort()'s signature is majorly
> improved in doing so, and kvm_s2_fault loses a bunch of fields.
>
> gmem_abort() will get a similar treatment down the line.
>
> Tested-by: Fuad Tabba <tabba at google.com>
> Reviewed-by: Fuad Tabba <tabba at google.com>
> Reviewed-by: Suzuki K Poulose <suzuki.poulose at arm.com>
> Signed-off-by: Marc Zyngier <maz at kernel.org>
Reviewed-by: Anshuman Khandual <anshuman.khandual at arm.com>
> ---
> arch/arm64/kvm/mmu.c | 133 ++++++++++++++++++++++---------------------
> 1 file changed, 69 insertions(+), 64 deletions(-)
>
> diff --git a/arch/arm64/kvm/mmu.c b/arch/arm64/kvm/mmu.c
> index 496bf5903ed3d..09e32f08028e4 100644
> --- a/arch/arm64/kvm/mmu.c
> +++ b/arch/arm64/kvm/mmu.c
> @@ -1565,6 +1565,14 @@ static void adjust_nested_exec_perms(struct kvm *kvm,
> *prot &= ~KVM_PGTABLE_PROT_PX;
> }
>
> +struct kvm_s2_fault_desc {
> + struct kvm_vcpu *vcpu;
> + phys_addr_t fault_ipa;
> + struct kvm_s2_trans *nested;
> + struct kvm_memory_slot *memslot;
> + unsigned long hva;
> +};
> +
> static int gmem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
> struct kvm_s2_trans *nested,
> struct kvm_memory_slot *memslot, bool is_perm)
> @@ -1640,23 +1648,20 @@ static int gmem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
> return ret != -EAGAIN ? ret : 0;
> }
>
> -static short kvm_s2_resolve_vma_size(struct vm_area_struct *vma,
> - unsigned long hva,
> - struct kvm_memory_slot *memslot,
> - struct kvm_s2_trans *nested,
> - bool *force_pte)
> +static short kvm_s2_resolve_vma_size(const struct kvm_s2_fault_desc *s2fd,
> + struct vm_area_struct *vma, bool *force_pte)
> {
> short vma_shift;
>
> if (*force_pte)
> vma_shift = PAGE_SHIFT;
> else
> - vma_shift = get_vma_page_shift(vma, hva);
> + vma_shift = get_vma_page_shift(vma, s2fd->hva);
>
> switch (vma_shift) {
> #ifndef __PAGETABLE_PMD_FOLDED
> case PUD_SHIFT:
> - if (fault_supports_stage2_huge_mapping(memslot, hva, PUD_SIZE))
> + if (fault_supports_stage2_huge_mapping(s2fd->memslot, s2fd->hva, PUD_SIZE))
> break;
> fallthrough;
> #endif
> @@ -1664,7 +1669,7 @@ static short kvm_s2_resolve_vma_size(struct vm_area_struct *vma,
> vma_shift = PMD_SHIFT;
> fallthrough;
> case PMD_SHIFT:
> - if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE))
> + if (fault_supports_stage2_huge_mapping(s2fd->memslot, s2fd->hva, PMD_SIZE))
> break;
> fallthrough;
> case CONT_PTE_SHIFT:
> @@ -1677,7 +1682,7 @@ static short kvm_s2_resolve_vma_size(struct vm_area_struct *vma,
> WARN_ONCE(1, "Unknown vma_shift %d", vma_shift);
> }
>
> - if (nested) {
> + if (s2fd->nested) {
> unsigned long max_map_size;
>
> max_map_size = *force_pte ? PAGE_SIZE : PUD_SIZE;
> @@ -1687,7 +1692,7 @@ static short kvm_s2_resolve_vma_size(struct vm_area_struct *vma,
> * can only create a block mapping if the guest stage 2 page
> * table uses at least as big a mapping.
> */
> - max_map_size = min(kvm_s2_trans_size(nested), max_map_size);
> + max_map_size = min(kvm_s2_trans_size(s2fd->nested), max_map_size);
>
> /*
> * Be careful that if the mapping size falls between
> @@ -1706,11 +1711,6 @@ static short kvm_s2_resolve_vma_size(struct vm_area_struct *vma,
> }
>
> struct kvm_s2_fault {
> - struct kvm_vcpu *vcpu;
> - phys_addr_t fault_ipa;
> - struct kvm_s2_trans *nested;
> - struct kvm_memory_slot *memslot;
> - unsigned long hva;
> bool fault_is_perm;
>
> bool write_fault;
> @@ -1732,28 +1732,28 @@ struct kvm_s2_fault {
> vm_flags_t vm_flags;
> };
>
> -static int kvm_s2_fault_get_vma_info(struct kvm_s2_fault *fault)
> +static int kvm_s2_fault_get_vma_info(const struct kvm_s2_fault_desc *s2fd,
> + struct kvm_s2_fault *fault)
> {
> struct vm_area_struct *vma;
> - struct kvm *kvm = fault->vcpu->kvm;
> + struct kvm *kvm = s2fd->vcpu->kvm;
>
> mmap_read_lock(current->mm);
> - vma = vma_lookup(current->mm, fault->hva);
> + vma = vma_lookup(current->mm, s2fd->hva);
> if (unlikely(!vma)) {
> - kvm_err("Failed to find VMA for fault->hva 0x%lx\n", fault->hva);
> + kvm_err("Failed to find VMA for hva 0x%lx\n", s2fd->hva);
> mmap_read_unlock(current->mm);
> return -EFAULT;
> }
>
> - fault->vma_pagesize = 1UL << kvm_s2_resolve_vma_size(vma, fault->hva, fault->memslot,
> - fault->nested, &fault->force_pte);
> + fault->vma_pagesize = BIT(kvm_s2_resolve_vma_size(s2fd, vma, &fault->force_pte));
>
> /*
> * Both the canonical IPA and fault IPA must be aligned to the
> * mapping size to ensure we find the right PFN and lay down the
> * mapping in the right place.
> */
> - fault->gfn = ALIGN_DOWN(fault->fault_ipa, fault->vma_pagesize) >> PAGE_SHIFT;
> + fault->gfn = ALIGN_DOWN(s2fd->fault_ipa, fault->vma_pagesize) >> PAGE_SHIFT;
>
> fault->mte_allowed = kvm_vma_mte_allowed(vma);
>
> @@ -1775,31 +1775,33 @@ static int kvm_s2_fault_get_vma_info(struct kvm_s2_fault *fault)
> return 0;
> }
>
> -static gfn_t get_canonical_gfn(struct kvm_s2_fault *fault)
> +static gfn_t get_canonical_gfn(const struct kvm_s2_fault_desc *s2fd,
> + const struct kvm_s2_fault *fault)
> {
> phys_addr_t ipa;
>
> - if (!fault->nested)
> + if (!s2fd->nested)
> return fault->gfn;
>
> - ipa = kvm_s2_trans_output(fault->nested);
> + ipa = kvm_s2_trans_output(s2fd->nested);
> return ALIGN_DOWN(ipa, fault->vma_pagesize) >> PAGE_SHIFT;
> }
>
> -static int kvm_s2_fault_pin_pfn(struct kvm_s2_fault *fault)
> +static int kvm_s2_fault_pin_pfn(const struct kvm_s2_fault_desc *s2fd,
> + struct kvm_s2_fault *fault)
> {
> int ret;
>
> - ret = kvm_s2_fault_get_vma_info(fault);
> + ret = kvm_s2_fault_get_vma_info(s2fd, fault);
> if (ret)
> return ret;
>
> - fault->pfn = __kvm_faultin_pfn(fault->memslot, get_canonical_gfn(fault),
> + fault->pfn = __kvm_faultin_pfn(s2fd->memslot, get_canonical_gfn(s2fd, fault),
> fault->write_fault ? FOLL_WRITE : 0,
> &fault->writable, &fault->page);
> if (unlikely(is_error_noslot_pfn(fault->pfn))) {
> if (fault->pfn == KVM_PFN_ERR_HWPOISON) {
> - kvm_send_hwpoison_signal(fault->hva, __ffs(fault->vma_pagesize));
> + kvm_send_hwpoison_signal(s2fd->hva, __ffs(fault->vma_pagesize));
> return 0;
> }
> return -EFAULT;
> @@ -1808,9 +1810,10 @@ static int kvm_s2_fault_pin_pfn(struct kvm_s2_fault *fault)
> return 1;
> }
>
> -static int kvm_s2_fault_compute_prot(struct kvm_s2_fault *fault)
> +static int kvm_s2_fault_compute_prot(const struct kvm_s2_fault_desc *s2fd,
> + struct kvm_s2_fault *fault)
> {
> - struct kvm *kvm = fault->vcpu->kvm;
> + struct kvm *kvm = s2fd->vcpu->kvm;
>
> /*
> * Check if this is non-struct page memory PFN, and cannot support
> @@ -1862,13 +1865,13 @@ static int kvm_s2_fault_compute_prot(struct kvm_s2_fault *fault)
> * and trigger the exception here. Since the memslot is valid, inject
> * the fault back to the guest.
> */
> - if (esr_fsc_is_excl_atomic_fault(kvm_vcpu_get_esr(fault->vcpu))) {
> - kvm_inject_dabt_excl_atomic(fault->vcpu, kvm_vcpu_get_hfar(fault->vcpu));
> + if (esr_fsc_is_excl_atomic_fault(kvm_vcpu_get_esr(s2fd->vcpu))) {
> + kvm_inject_dabt_excl_atomic(s2fd->vcpu, kvm_vcpu_get_hfar(s2fd->vcpu));
> return 1;
> }
>
> - if (fault->nested)
> - adjust_nested_fault_perms(fault->nested, &fault->prot, &fault->writable);
> + if (s2fd->nested)
> + adjust_nested_fault_perms(s2fd->nested, &fault->prot, &fault->writable);
>
> if (fault->writable)
> fault->prot |= KVM_PGTABLE_PROT_W;
> @@ -1882,8 +1885,8 @@ static int kvm_s2_fault_compute_prot(struct kvm_s2_fault *fault)
> else if (cpus_have_final_cap(ARM64_HAS_CACHE_DIC))
> fault->prot |= KVM_PGTABLE_PROT_X;
>
> - if (fault->nested)
> - adjust_nested_exec_perms(kvm, fault->nested, &fault->prot);
> + if (s2fd->nested)
> + adjust_nested_exec_perms(kvm, s2fd->nested, &fault->prot);
>
> if (!fault->fault_is_perm && !fault->s2_force_noncacheable && kvm_has_mte(kvm)) {
> /* Check the VMM hasn't introduced a new disallowed VMA */
> @@ -1899,15 +1902,16 @@ static phys_addr_t get_ipa(const struct kvm_s2_fault *fault)
> return gfn_to_gpa(fault->gfn);
> }
>
> -static int kvm_s2_fault_map(struct kvm_s2_fault *fault, void *memcache)
> +static int kvm_s2_fault_map(const struct kvm_s2_fault_desc *s2fd,
> + struct kvm_s2_fault *fault, void *memcache)
> {
> - struct kvm *kvm = fault->vcpu->kvm;
> + struct kvm *kvm = s2fd->vcpu->kvm;
> struct kvm_pgtable *pgt;
> int ret;
> enum kvm_pgtable_walk_flags flags = KVM_PGTABLE_WALK_SHARED;
>
> kvm_fault_lock(kvm);
> - pgt = fault->vcpu->arch.hw_mmu->pgt;
> + pgt = s2fd->vcpu->arch.hw_mmu->pgt;
> ret = -EAGAIN;
> if (mmu_invalidate_retry(kvm, fault->mmu_seq))
> goto out_unlock;
> @@ -1921,8 +1925,8 @@ static int kvm_s2_fault_map(struct kvm_s2_fault *fault, void *memcache)
> if (fault->fault_is_perm && fault->fault_granule > PAGE_SIZE) {
> fault->vma_pagesize = fault->fault_granule;
> } else {
> - fault->vma_pagesize = transparent_hugepage_adjust(kvm, fault->memslot,
> - fault->hva, &fault->pfn,
> + fault->vma_pagesize = transparent_hugepage_adjust(kvm, s2fd->memslot,
> + s2fd->hva, &fault->pfn,
> &fault->gfn);
>
> if (fault->vma_pagesize < 0) {
> @@ -1960,34 +1964,27 @@ static int kvm_s2_fault_map(struct kvm_s2_fault *fault, void *memcache)
>
> /* Mark the page dirty only if the fault is handled successfully */
> if (fault->writable && !ret)
> - mark_page_dirty_in_slot(kvm, fault->memslot, get_canonical_gfn(fault));
> + mark_page_dirty_in_slot(kvm, s2fd->memslot, get_canonical_gfn(s2fd, fault));
>
> if (ret != -EAGAIN)
> return ret;
> return 0;
> }
>
> -static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
> - struct kvm_s2_trans *nested,
> - struct kvm_memory_slot *memslot, unsigned long hva,
> - bool fault_is_perm)
> +static int user_mem_abort(const struct kvm_s2_fault_desc *s2fd)
> {
> - bool write_fault = kvm_is_write_fault(vcpu);
> - bool logging_active = memslot_is_logging(memslot);
> + bool perm_fault = kvm_vcpu_trap_is_permission_fault(s2fd->vcpu);
> + bool write_fault = kvm_is_write_fault(s2fd->vcpu);
> + bool logging_active = memslot_is_logging(s2fd->memslot);
> struct kvm_s2_fault fault = {
> - .vcpu = vcpu,
> - .fault_ipa = fault_ipa,
> - .nested = nested,
> - .memslot = memslot,
> - .hva = hva,
> - .fault_is_perm = fault_is_perm,
> + .fault_is_perm = perm_fault,
> .logging_active = logging_active,
> .force_pte = logging_active,
> .prot = KVM_PGTABLE_PROT_R,
> - .fault_granule = fault_is_perm ? kvm_vcpu_trap_get_perm_fault_granule(vcpu) : 0,
> + .fault_granule = perm_fault ? kvm_vcpu_trap_get_perm_fault_granule(s2fd->vcpu) : 0,
> .write_fault = write_fault,
> - .exec_fault = kvm_vcpu_trap_is_exec_fault(vcpu),
> - .topup_memcache = !fault_is_perm || (logging_active && write_fault),
> + .exec_fault = kvm_vcpu_trap_is_exec_fault(s2fd->vcpu),
> + .topup_memcache = !perm_fault || (logging_active && write_fault),
> };
> void *memcache;
> int ret;
> @@ -2000,7 +1997,7 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
> * only exception to this is when dirty logging is enabled at runtime
> * and a write fault needs to collapse a block entry into a table.
> */
> - ret = prepare_mmu_memcache(vcpu, fault.topup_memcache, &memcache);
> + ret = prepare_mmu_memcache(s2fd->vcpu, fault.topup_memcache, &memcache);
> if (ret)
> return ret;
>
> @@ -2008,17 +2005,17 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
> * Let's check if we will get back a huge page backed by hugetlbfs, or
> * get block mapping for device MMIO region.
> */
> - ret = kvm_s2_fault_pin_pfn(&fault);
> + ret = kvm_s2_fault_pin_pfn(s2fd, &fault);
> if (ret != 1)
> return ret;
>
> - ret = kvm_s2_fault_compute_prot(&fault);
> + ret = kvm_s2_fault_compute_prot(s2fd, &fault);
> if (ret) {
> kvm_release_page_unused(fault.page);
> return ret;
> }
>
> - return kvm_s2_fault_map(&fault, memcache);
> + return kvm_s2_fault_map(s2fd, &fault, memcache);
> }
>
> /* Resolve the access fault by making the page young again. */
> @@ -2284,12 +2281,20 @@ int kvm_handle_guest_abort(struct kvm_vcpu *vcpu)
> VM_WARN_ON_ONCE(kvm_vcpu_trap_is_permission_fault(vcpu) &&
> !write_fault && !kvm_vcpu_trap_is_exec_fault(vcpu));
>
> + const struct kvm_s2_fault_desc s2fd = {
> + .vcpu = vcpu,
> + .fault_ipa = fault_ipa,
> + .nested = nested,
> + .memslot = memslot,
> + .hva = hva,
> + };
> +
> if (kvm_slot_has_gmem(memslot))
> ret = gmem_abort(vcpu, fault_ipa, nested, memslot,
> esr_fsc_is_permission_fault(esr));
> else
> - ret = user_mem_abort(vcpu, fault_ipa, nested, memslot, hva,
> - esr_fsc_is_permission_fault(esr));
> + ret = user_mem_abort(&s2fd);
> +
> if (ret == 0)
> ret = 1;
> out:
More information about the linux-arm-kernel
mailing list