[PATCH v2 10/30] KVM: arm64: Initialize struct kvm_s2_fault completely at declaration

Anshuman Khandual anshuman.khandual at arm.com
Mon Mar 30 22:43:01 PDT 2026


On 27/03/26 5:05 PM, Marc Zyngier wrote:
> From: Fuad Tabba <tabba at google.com>
> 
> Simplify the initialization of struct kvm_s2_fault in user_mem_abort().
> 
> Instead of partially initializing the struct via designated initializers
> and then sequentially assigning the remaining fields (like write_fault
> and topup_memcache) further down the function, evaluate those
> dependencies upfront.
> 
> This allows the entire struct to be fully initialized at declaration. It
> also eliminates the need for the intermediate fault_data variable and
> its associated fault pointer, reducing boilerplate.
> 
> Signed-off-by: Fuad Tabba <tabba at google.com>
> Signed-off-by: Marc Zyngier <maz at kernel.org>

Reviewed-by: Anshuman Khandual <anshuman.khandual at arm.com>

> ---
>  arch/arm64/kvm/mmu.c | 34 ++++++++++++++++------------------
>  1 file changed, 16 insertions(+), 18 deletions(-)
> 
> diff --git a/arch/arm64/kvm/mmu.c b/arch/arm64/kvm/mmu.c
> index e77b0b60697f6..2b85daaa4426b 100644
> --- a/arch/arm64/kvm/mmu.c
> +++ b/arch/arm64/kvm/mmu.c
> @@ -1962,8 +1962,9 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
>  			  struct kvm_memory_slot *memslot, unsigned long hva,
>  			  bool fault_is_perm)
>  {
> -	int ret = 0;
> -	struct kvm_s2_fault fault_data = {
> +	bool write_fault = kvm_is_write_fault(vcpu);
> +	bool logging_active = memslot_is_logging(memslot);
> +	struct kvm_s2_fault fault = {
>  		.vcpu = vcpu,
>  		.fault_ipa = fault_ipa,
>  		.nested = nested,
> @@ -1971,19 +1972,18 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
>  		.hva = hva,
>  		.fault_is_perm = fault_is_perm,
>  		.ipa = fault_ipa,
> -		.logging_active = memslot_is_logging(memslot),
> -		.force_pte = memslot_is_logging(memslot),
> -		.s2_force_noncacheable = false,
> +		.logging_active = logging_active,
> +		.force_pte = logging_active,
>  		.prot = KVM_PGTABLE_PROT_R,
> +		.fault_granule = fault_is_perm ? kvm_vcpu_trap_get_perm_fault_granule(vcpu) : 0,
> +		.write_fault = write_fault,
> +		.exec_fault = kvm_vcpu_trap_is_exec_fault(vcpu),
> +		.topup_memcache = !fault_is_perm || (logging_active && write_fault),
>  	};
> -	struct kvm_s2_fault *fault = &fault_data;
>  	void *memcache;
> +	int ret;
>  
> -	if (fault->fault_is_perm)
> -		fault->fault_granule = kvm_vcpu_trap_get_perm_fault_granule(fault->vcpu);
> -	fault->write_fault = kvm_is_write_fault(fault->vcpu);
> -	fault->exec_fault = kvm_vcpu_trap_is_exec_fault(fault->vcpu);
> -	VM_WARN_ON_ONCE(fault->write_fault && fault->exec_fault);
> +	VM_WARN_ON_ONCE(fault.write_fault && fault.exec_fault);
>  
>  	/*
>  	 * Permission faults just need to update the existing leaf entry,
> @@ -1991,9 +1991,7 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
>  	 * only exception to this is when dirty logging is enabled at runtime
>  	 * and a write fault needs to collapse a block entry into a table.
>  	 */
> -	fault->topup_memcache = !fault->fault_is_perm ||
> -				(fault->logging_active && fault->write_fault);
> -	ret = prepare_mmu_memcache(fault->vcpu, fault->topup_memcache, &memcache);
> +	ret = prepare_mmu_memcache(vcpu, fault.topup_memcache, &memcache);
>  	if (ret)
>  		return ret;
>  
> @@ -2001,17 +1999,17 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
>  	 * Let's check if we will get back a huge page backed by hugetlbfs, or
>  	 * get block mapping for device MMIO region.
>  	 */
> -	ret = kvm_s2_fault_pin_pfn(fault);
> +	ret = kvm_s2_fault_pin_pfn(&fault);
>  	if (ret != 1)
>  		return ret;
>  
> -	ret = kvm_s2_fault_compute_prot(fault);
> +	ret = kvm_s2_fault_compute_prot(&fault);
>  	if (ret) {
> -		kvm_release_page_unused(fault->page);
> +		kvm_release_page_unused(fault.page);
>  		return ret;
>  	}
>  
> -	return kvm_s2_fault_map(fault, memcache);
> +	return kvm_s2_fault_map(&fault, memcache);
>  }
>  
>  /* Resolve the access fault by making the page young again. */



More information about the linux-arm-kernel mailing list