[PATCH v4 06/21] KVM: selftests: Add "struct kvm_mmu" to track a given MMU instance
Yosry Ahmed
yosry.ahmed at linux.dev
Fri Jan 2 08:50:19 PST 2026
On Tue, Dec 30, 2025 at 03:01:35PM -0800, Sean Christopherson wrote:
> Add a "struct kvm_mmu" to track a given MMU instance, e.g. a VM's stage-1
> MMU versus a VM's stage-2 MMU, so that x86 can share MMU functionality for
> both stage-1 and stage-2 MMUs, without creating the potential for subtle
> bugs, e.g. due to consuming on vm->pgtable_levels when operating a stage-2
> MMU.
>
> Encapsulate the existing de facto MMU in "struct kvm_vm", e.g instead of
> burying the MMU details in "struct kvm_vm_arch", to avoid more #ifdefs in
> ____vm_create(), and in the hopes that other architectures can utilize the
> formalized MMU structure if/when they too support stage-2 page tables.
>
> No functional change intended.
>
> Signed-off-by: Sean Christopherson <seanjc at google.com>
Reviewed-by: Yosry Ahmed <yosry.ahmed at linux.dev>
> ---
> .../testing/selftests/kvm/include/kvm_util.h | 11 ++++--
> .../selftests/kvm/lib/arm64/processor.c | 38 +++++++++----------
> tools/testing/selftests/kvm/lib/kvm_util.c | 28 +++++++-------
> .../selftests/kvm/lib/loongarch/processor.c | 28 +++++++-------
> .../selftests/kvm/lib/riscv/processor.c | 31 +++++++--------
> .../selftests/kvm/lib/s390/processor.c | 16 ++++----
> .../testing/selftests/kvm/lib/x86/processor.c | 28 +++++++-------
> .../kvm/x86/vmx_nested_la57_state_test.c | 2 +-
> 8 files changed, 94 insertions(+), 88 deletions(-)
>
> diff --git a/tools/testing/selftests/kvm/include/kvm_util.h b/tools/testing/selftests/kvm/include/kvm_util.h
> index 81f4355ff28a..39558c05c0bf 100644
> --- a/tools/testing/selftests/kvm/include/kvm_util.h
> +++ b/tools/testing/selftests/kvm/include/kvm_util.h
> @@ -88,12 +88,17 @@ enum kvm_mem_region_type {
> NR_MEM_REGIONS,
> };
>
> +struct kvm_mmu {
> + bool pgd_created;
> + uint64_t pgd;
> + int pgtable_levels;
> +};
> +
> struct kvm_vm {
> int mode;
> unsigned long type;
> int kvm_fd;
> int fd;
> - unsigned int pgtable_levels;
> unsigned int page_size;
> unsigned int page_shift;
> unsigned int pa_bits;
> @@ -104,13 +109,13 @@ struct kvm_vm {
> struct sparsebit *vpages_valid;
> struct sparsebit *vpages_mapped;
> bool has_irqchip;
> - bool pgd_created;
> vm_paddr_t ucall_mmio_addr;
> - vm_paddr_t pgd;
> vm_vaddr_t handlers;
> uint32_t dirty_ring_size;
> uint64_t gpa_tag_mask;
>
> + struct kvm_mmu mmu;
> +
> struct kvm_vm_arch arch;
>
> struct kvm_binary_stats stats;
> diff --git a/tools/testing/selftests/kvm/lib/arm64/processor.c b/tools/testing/selftests/kvm/lib/arm64/processor.c
> index d46e4b13b92c..c40f59d48311 100644
> --- a/tools/testing/selftests/kvm/lib/arm64/processor.c
> +++ b/tools/testing/selftests/kvm/lib/arm64/processor.c
> @@ -28,7 +28,7 @@ static uint64_t page_align(struct kvm_vm *vm, uint64_t v)
>
> static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva)
> {
> - unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift;
> + unsigned int shift = (vm->mmu.pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift;
> uint64_t mask = (1UL << (vm->va_bits - shift)) - 1;
>
> return (gva >> shift) & mask;
> @@ -39,7 +39,7 @@ static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva)
> unsigned int shift = 2 * (vm->page_shift - 3) + vm->page_shift;
> uint64_t mask = (1UL << (vm->page_shift - 3)) - 1;
>
> - TEST_ASSERT(vm->pgtable_levels == 4,
> + TEST_ASSERT(vm->mmu.pgtable_levels == 4,
> "Mode %d does not have 4 page table levels", vm->mode);
>
> return (gva >> shift) & mask;
> @@ -50,7 +50,7 @@ static uint64_t pmd_index(struct kvm_vm *vm, vm_vaddr_t gva)
> unsigned int shift = (vm->page_shift - 3) + vm->page_shift;
> uint64_t mask = (1UL << (vm->page_shift - 3)) - 1;
>
> - TEST_ASSERT(vm->pgtable_levels >= 3,
> + TEST_ASSERT(vm->mmu.pgtable_levels >= 3,
> "Mode %d does not have >= 3 page table levels", vm->mode);
>
> return (gva >> shift) & mask;
> @@ -104,7 +104,7 @@ static uint64_t pte_addr(struct kvm_vm *vm, uint64_t pte)
>
> static uint64_t ptrs_per_pgd(struct kvm_vm *vm)
> {
> - unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift;
> + unsigned int shift = (vm->mmu.pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift;
> return 1 << (vm->va_bits - shift);
> }
>
> @@ -117,13 +117,13 @@ void virt_arch_pgd_alloc(struct kvm_vm *vm)
> {
> size_t nr_pages = page_align(vm, ptrs_per_pgd(vm) * 8) / vm->page_size;
>
> - if (vm->pgd_created)
> + if (vm->mmu.pgd_created)
> return;
>
> - vm->pgd = vm_phy_pages_alloc(vm, nr_pages,
> - KVM_GUEST_PAGE_TABLE_MIN_PADDR,
> - vm->memslots[MEM_REGION_PT]);
> - vm->pgd_created = true;
> + vm->mmu.pgd = vm_phy_pages_alloc(vm, nr_pages,
> + KVM_GUEST_PAGE_TABLE_MIN_PADDR,
> + vm->memslots[MEM_REGION_PT]);
> + vm->mmu.pgd_created = true;
> }
>
> static void _virt_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr,
> @@ -147,12 +147,12 @@ static void _virt_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr,
> " paddr: 0x%lx vm->max_gfn: 0x%lx vm->page_size: 0x%x",
> paddr, vm->max_gfn, vm->page_size);
>
> - ptep = addr_gpa2hva(vm, vm->pgd) + pgd_index(vm, vaddr) * 8;
> + ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pgd_index(vm, vaddr) * 8;
> if (!*ptep)
> *ptep = addr_pte(vm, vm_alloc_page_table(vm),
> PGD_TYPE_TABLE | PTE_VALID);
>
> - switch (vm->pgtable_levels) {
> + switch (vm->mmu.pgtable_levels) {
> case 4:
> ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pud_index(vm, vaddr) * 8;
> if (!*ptep)
> @@ -190,16 +190,16 @@ uint64_t *virt_get_pte_hva_at_level(struct kvm_vm *vm, vm_vaddr_t gva, int level
> {
> uint64_t *ptep;
>
> - if (!vm->pgd_created)
> + if (!vm->mmu.pgd_created)
> goto unmapped_gva;
>
> - ptep = addr_gpa2hva(vm, vm->pgd) + pgd_index(vm, gva) * 8;
> + ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pgd_index(vm, gva) * 8;
> if (!ptep)
> goto unmapped_gva;
> if (level == 0)
> return ptep;
>
> - switch (vm->pgtable_levels) {
> + switch (vm->mmu.pgtable_levels) {
> case 4:
> ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pud_index(vm, gva) * 8;
> if (!ptep)
> @@ -263,13 +263,13 @@ static void pte_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent, uint64_t p
>
> void virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
> {
> - int level = 4 - (vm->pgtable_levels - 1);
> + int level = 4 - (vm->mmu.pgtable_levels - 1);
> uint64_t pgd, *ptep;
>
> - if (!vm->pgd_created)
> + if (!vm->mmu.pgd_created)
> return;
>
> - for (pgd = vm->pgd; pgd < vm->pgd + ptrs_per_pgd(vm) * 8; pgd += 8) {
> + for (pgd = vm->mmu.pgd; pgd < vm->mmu.pgd + ptrs_per_pgd(vm) * 8; pgd += 8) {
> ptep = addr_gpa2hva(vm, pgd);
> if (!*ptep)
> continue;
> @@ -350,7 +350,7 @@ void aarch64_vcpu_setup(struct kvm_vcpu *vcpu, struct kvm_vcpu_init *init)
> TEST_FAIL("Unknown guest mode, mode: 0x%x", vm->mode);
> }
>
> - ttbr0_el1 = vm->pgd & GENMASK(47, vm->page_shift);
> + ttbr0_el1 = vm->mmu.pgd & GENMASK(47, vm->page_shift);
>
> /* Configure output size */
> switch (vm->mode) {
> @@ -358,7 +358,7 @@ void aarch64_vcpu_setup(struct kvm_vcpu *vcpu, struct kvm_vcpu_init *init)
> case VM_MODE_P52V48_16K:
> case VM_MODE_P52V48_64K:
> tcr_el1 |= TCR_IPS_52_BITS;
> - ttbr0_el1 |= FIELD_GET(GENMASK(51, 48), vm->pgd) << 2;
> + ttbr0_el1 |= FIELD_GET(GENMASK(51, 48), vm->mmu.pgd) << 2;
> break;
> case VM_MODE_P48V48_4K:
> case VM_MODE_P48V48_16K:
> diff --git a/tools/testing/selftests/kvm/lib/kvm_util.c b/tools/testing/selftests/kvm/lib/kvm_util.c
> index 8279b6ced8d2..65752daeed90 100644
> --- a/tools/testing/selftests/kvm/lib/kvm_util.c
> +++ b/tools/testing/selftests/kvm/lib/kvm_util.c
> @@ -281,34 +281,34 @@ struct kvm_vm *____vm_create(struct vm_shape shape)
> /* Setup mode specific traits. */
> switch (vm->mode) {
> case VM_MODE_P52V48_4K:
> - vm->pgtable_levels = 4;
> + vm->mmu.pgtable_levels = 4;
> break;
> case VM_MODE_P52V48_64K:
> - vm->pgtable_levels = 3;
> + vm->mmu.pgtable_levels = 3;
> break;
> case VM_MODE_P48V48_4K:
> - vm->pgtable_levels = 4;
> + vm->mmu.pgtable_levels = 4;
> break;
> case VM_MODE_P48V48_64K:
> - vm->pgtable_levels = 3;
> + vm->mmu.pgtable_levels = 3;
> break;
> case VM_MODE_P40V48_4K:
> case VM_MODE_P36V48_4K:
> - vm->pgtable_levels = 4;
> + vm->mmu.pgtable_levels = 4;
> break;
> case VM_MODE_P40V48_64K:
> case VM_MODE_P36V48_64K:
> - vm->pgtable_levels = 3;
> + vm->mmu.pgtable_levels = 3;
> break;
> case VM_MODE_P52V48_16K:
> case VM_MODE_P48V48_16K:
> case VM_MODE_P40V48_16K:
> case VM_MODE_P36V48_16K:
> - vm->pgtable_levels = 4;
> + vm->mmu.pgtable_levels = 4;
> break;
> case VM_MODE_P47V47_16K:
> case VM_MODE_P36V47_16K:
> - vm->pgtable_levels = 3;
> + vm->mmu.pgtable_levels = 3;
> break;
> case VM_MODE_PXXVYY_4K:
> #ifdef __x86_64__
> @@ -321,22 +321,22 @@ struct kvm_vm *____vm_create(struct vm_shape shape)
> vm->va_bits);
>
> if (vm->va_bits == 57) {
> - vm->pgtable_levels = 5;
> + vm->mmu.pgtable_levels = 5;
> } else {
> TEST_ASSERT(vm->va_bits == 48,
> "Unexpected guest virtual address width: %d",
> vm->va_bits);
> - vm->pgtable_levels = 4;
> + vm->mmu.pgtable_levels = 4;
> }
> #else
> TEST_FAIL("VM_MODE_PXXVYY_4K not supported on non-x86 platforms");
> #endif
> break;
> case VM_MODE_P47V64_4K:
> - vm->pgtable_levels = 5;
> + vm->mmu.pgtable_levels = 5;
> break;
> case VM_MODE_P44V64_4K:
> - vm->pgtable_levels = 5;
> + vm->mmu.pgtable_levels = 5;
> break;
> default:
> TEST_FAIL("Unknown guest mode: 0x%x", vm->mode);
> @@ -1956,8 +1956,8 @@ void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
> fprintf(stream, "%*sMapped Virtual Pages:\n", indent, "");
> sparsebit_dump(stream, vm->vpages_mapped, indent + 2);
> fprintf(stream, "%*spgd_created: %u\n", indent, "",
> - vm->pgd_created);
> - if (vm->pgd_created) {
> + vm->mmu.pgd_created);
> + if (vm->mmu.pgd_created) {
> fprintf(stream, "%*sVirtual Translation Tables:\n",
> indent + 2, "");
> virt_dump(stream, vm, indent + 4);
> diff --git a/tools/testing/selftests/kvm/lib/loongarch/processor.c b/tools/testing/selftests/kvm/lib/loongarch/processor.c
> index 07c103369ddb..17aa55a2047a 100644
> --- a/tools/testing/selftests/kvm/lib/loongarch/processor.c
> +++ b/tools/testing/selftests/kvm/lib/loongarch/processor.c
> @@ -50,11 +50,11 @@ void virt_arch_pgd_alloc(struct kvm_vm *vm)
> int i;
> vm_paddr_t child, table;
>
> - if (vm->pgd_created)
> + if (vm->mmu.pgd_created)
> return;
>
> child = table = 0;
> - for (i = 0; i < vm->pgtable_levels; i++) {
> + for (i = 0; i < vm->mmu.pgtable_levels; i++) {
> invalid_pgtable[i] = child;
> table = vm_phy_page_alloc(vm, LOONGARCH_PAGE_TABLE_PHYS_MIN,
> vm->memslots[MEM_REGION_PT]);
> @@ -62,8 +62,8 @@ void virt_arch_pgd_alloc(struct kvm_vm *vm)
> virt_set_pgtable(vm, table, child);
> child = table;
> }
> - vm->pgd = table;
> - vm->pgd_created = true;
> + vm->mmu.pgd = table;
> + vm->mmu.pgd_created = true;
> }
>
> static int virt_pte_none(uint64_t *ptep, int level)
> @@ -77,11 +77,11 @@ static uint64_t *virt_populate_pte(struct kvm_vm *vm, vm_vaddr_t gva, int alloc)
> uint64_t *ptep;
> vm_paddr_t child;
>
> - if (!vm->pgd_created)
> + if (!vm->mmu.pgd_created)
> goto unmapped_gva;
>
> - child = vm->pgd;
> - level = vm->pgtable_levels - 1;
> + child = vm->mmu.pgd;
> + level = vm->mmu.pgtable_levels - 1;
> while (level > 0) {
> ptep = addr_gpa2hva(vm, child) + virt_pte_index(vm, gva, level) * 8;
> if (virt_pte_none(ptep, level)) {
> @@ -161,11 +161,11 @@ void virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
> {
> int level;
>
> - if (!vm->pgd_created)
> + if (!vm->mmu.pgd_created)
> return;
>
> - level = vm->pgtable_levels - 1;
> - pte_dump(stream, vm, indent, vm->pgd, level);
> + level = vm->mmu.pgtable_levels - 1;
> + pte_dump(stream, vm, indent, vm->mmu.pgd, level);
> }
>
> void vcpu_arch_dump(FILE *stream, struct kvm_vcpu *vcpu, uint8_t indent)
> @@ -297,7 +297,7 @@ static void loongarch_vcpu_setup(struct kvm_vcpu *vcpu)
>
> width = vm->page_shift - 3;
>
> - switch (vm->pgtable_levels) {
> + switch (vm->mmu.pgtable_levels) {
> case 4:
> /* pud page shift and width */
> val = (vm->page_shift + width * 2) << 20 | (width << 25);
> @@ -309,15 +309,15 @@ static void loongarch_vcpu_setup(struct kvm_vcpu *vcpu)
> val |= vm->page_shift | width << 5;
> break;
> default:
> - TEST_FAIL("Got %u page table levels, expected 3 or 4", vm->pgtable_levels);
> + TEST_FAIL("Got %u page table levels, expected 3 or 4", vm->mmu.pgtable_levels);
> }
>
> loongarch_set_csr(vcpu, LOONGARCH_CSR_PWCTL0, val);
>
> /* PGD page shift and width */
> - val = (vm->page_shift + width * (vm->pgtable_levels - 1)) | width << 6;
> + val = (vm->page_shift + width * (vm->mmu.pgtable_levels - 1)) | width << 6;
> loongarch_set_csr(vcpu, LOONGARCH_CSR_PWCTL1, val);
> - loongarch_set_csr(vcpu, LOONGARCH_CSR_PGDL, vm->pgd);
> + loongarch_set_csr(vcpu, LOONGARCH_CSR_PGDL, vm->mmu.pgd);
>
> /*
> * Refill exception runs on real mode
> diff --git a/tools/testing/selftests/kvm/lib/riscv/processor.c b/tools/testing/selftests/kvm/lib/riscv/processor.c
> index 2eac7d4b59e9..e6ec7c224fc3 100644
> --- a/tools/testing/selftests/kvm/lib/riscv/processor.c
> +++ b/tools/testing/selftests/kvm/lib/riscv/processor.c
> @@ -60,7 +60,7 @@ static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva, int level)
> {
> TEST_ASSERT(level > -1,
> "Negative page table level (%d) not possible", level);
> - TEST_ASSERT(level < vm->pgtable_levels,
> + TEST_ASSERT(level < vm->mmu.pgtable_levels,
> "Invalid page table level (%d)", level);
>
> return (gva & pte_index_mask[level]) >> pte_index_shift[level];
> @@ -70,19 +70,19 @@ void virt_arch_pgd_alloc(struct kvm_vm *vm)
> {
> size_t nr_pages = page_align(vm, ptrs_per_pte(vm) * 8) / vm->page_size;
>
> - if (vm->pgd_created)
> + if (vm->mmu.pgd_created)
> return;
>
> - vm->pgd = vm_phy_pages_alloc(vm, nr_pages,
> - KVM_GUEST_PAGE_TABLE_MIN_PADDR,
> - vm->memslots[MEM_REGION_PT]);
> - vm->pgd_created = true;
> + vm->mmu.pgd = vm_phy_pages_alloc(vm, nr_pages,
> + KVM_GUEST_PAGE_TABLE_MIN_PADDR,
> + vm->memslots[MEM_REGION_PT]);
> + vm->mmu.pgd_created = true;
> }
>
> void virt_arch_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr)
> {
> uint64_t *ptep, next_ppn;
> - int level = vm->pgtable_levels - 1;
> + int level = vm->mmu.pgtable_levels - 1;
>
> TEST_ASSERT((vaddr % vm->page_size) == 0,
> "Virtual address not on page boundary,\n"
> @@ -98,7 +98,7 @@ void virt_arch_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr)
> " paddr: 0x%lx vm->max_gfn: 0x%lx vm->page_size: 0x%x",
> paddr, vm->max_gfn, vm->page_size);
>
> - ptep = addr_gpa2hva(vm, vm->pgd) + pte_index(vm, vaddr, level) * 8;
> + ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pte_index(vm, vaddr, level) * 8;
> if (!*ptep) {
> next_ppn = vm_alloc_page_table(vm) >> PGTBL_PAGE_SIZE_SHIFT;
> *ptep = (next_ppn << PGTBL_PTE_ADDR_SHIFT) |
> @@ -126,12 +126,12 @@ void virt_arch_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr)
> vm_paddr_t addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva)
> {
> uint64_t *ptep;
> - int level = vm->pgtable_levels - 1;
> + int level = vm->mmu.pgtable_levels - 1;
>
> - if (!vm->pgd_created)
> + if (!vm->mmu.pgd_created)
> goto unmapped_gva;
>
> - ptep = addr_gpa2hva(vm, vm->pgd) + pte_index(vm, gva, level) * 8;
> + ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pte_index(vm, gva, level) * 8;
> if (!ptep)
> goto unmapped_gva;
> level--;
> @@ -176,13 +176,14 @@ static void pte_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent,
>
> void virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
> {
> - int level = vm->pgtable_levels - 1;
> + struct kvm_mmu *mmu = &vm->mmu;
> + int level = mmu->pgtable_levels - 1;
> uint64_t pgd, *ptep;
>
> - if (!vm->pgd_created)
> + if (!mmu->pgd_created)
> return;
>
> - for (pgd = vm->pgd; pgd < vm->pgd + ptrs_per_pte(vm) * 8; pgd += 8) {
> + for (pgd = mmu->pgd; pgd < mmu->pgd + ptrs_per_pte(vm) * 8; pgd += 8) {
> ptep = addr_gpa2hva(vm, pgd);
> if (!*ptep)
> continue;
> @@ -211,7 +212,7 @@ void riscv_vcpu_mmu_setup(struct kvm_vcpu *vcpu)
> TEST_FAIL("Unknown guest mode, mode: 0x%x", vm->mode);
> }
>
> - satp = (vm->pgd >> PGTBL_PAGE_SIZE_SHIFT) & SATP_PPN;
> + satp = (vm->mmu.pgd >> PGTBL_PAGE_SIZE_SHIFT) & SATP_PPN;
> satp |= SATP_MODE_48;
>
> vcpu_set_reg(vcpu, RISCV_GENERAL_CSR_REG(satp), satp);
> diff --git a/tools/testing/selftests/kvm/lib/s390/processor.c b/tools/testing/selftests/kvm/lib/s390/processor.c
> index 8ceeb17c819a..6a9a660413a7 100644
> --- a/tools/testing/selftests/kvm/lib/s390/processor.c
> +++ b/tools/testing/selftests/kvm/lib/s390/processor.c
> @@ -17,7 +17,7 @@ void virt_arch_pgd_alloc(struct kvm_vm *vm)
> TEST_ASSERT(vm->page_size == PAGE_SIZE, "Unsupported page size: 0x%x",
> vm->page_size);
>
> - if (vm->pgd_created)
> + if (vm->mmu.pgd_created)
> return;
>
> paddr = vm_phy_pages_alloc(vm, PAGES_PER_REGION,
> @@ -25,8 +25,8 @@ void virt_arch_pgd_alloc(struct kvm_vm *vm)
> vm->memslots[MEM_REGION_PT]);
> memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size);
>
> - vm->pgd = paddr;
> - vm->pgd_created = true;
> + vm->mmu.pgd = paddr;
> + vm->mmu.pgd_created = true;
> }
>
> /*
> @@ -70,7 +70,7 @@ void virt_arch_pg_map(struct kvm_vm *vm, uint64_t gva, uint64_t gpa)
> gva, vm->max_gfn, vm->page_size);
>
> /* Walk through region and segment tables */
> - entry = addr_gpa2hva(vm, vm->pgd);
> + entry = addr_gpa2hva(vm, vm->mmu.pgd);
> for (ri = 1; ri <= 4; ri++) {
> idx = (gva >> (64 - 11 * ri)) & 0x7ffu;
> if (entry[idx] & REGION_ENTRY_INVALID)
> @@ -94,7 +94,7 @@ vm_paddr_t addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva)
> TEST_ASSERT(vm->page_size == PAGE_SIZE, "Unsupported page size: 0x%x",
> vm->page_size);
>
> - entry = addr_gpa2hva(vm, vm->pgd);
> + entry = addr_gpa2hva(vm, vm->mmu.pgd);
> for (ri = 1; ri <= 4; ri++) {
> idx = (gva >> (64 - 11 * ri)) & 0x7ffu;
> TEST_ASSERT(!(entry[idx] & REGION_ENTRY_INVALID),
> @@ -149,10 +149,10 @@ static void virt_dump_region(FILE *stream, struct kvm_vm *vm, uint8_t indent,
>
> void virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
> {
> - if (!vm->pgd_created)
> + if (!vm->mmu.pgd_created)
> return;
>
> - virt_dump_region(stream, vm, indent, vm->pgd);
> + virt_dump_region(stream, vm, indent, vm->mmu.pgd);
> }
>
> void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
> @@ -184,7 +184,7 @@ struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id)
>
> vcpu_sregs_get(vcpu, &sregs);
> sregs.crs[0] |= 0x00040000; /* Enable floating point regs */
> - sregs.crs[1] = vm->pgd | 0xf; /* Primary region table */
> + sregs.crs[1] = vm->mmu.pgd | 0xf; /* Primary region table */
> vcpu_sregs_set(vcpu, &sregs);
>
> vcpu->run->psw_mask = 0x0400000180000000ULL; /* DAT enabled + 64 bit mode */
> diff --git a/tools/testing/selftests/kvm/lib/x86/processor.c b/tools/testing/selftests/kvm/lib/x86/processor.c
> index c14bf2b5f28f..f027f86d1535 100644
> --- a/tools/testing/selftests/kvm/lib/x86/processor.c
> +++ b/tools/testing/selftests/kvm/lib/x86/processor.c
> @@ -162,9 +162,9 @@ void virt_arch_pgd_alloc(struct kvm_vm *vm)
> "Unknown or unsupported guest mode: 0x%x", vm->mode);
>
> /* If needed, create the top-level page table. */
> - if (!vm->pgd_created) {
> - vm->pgd = vm_alloc_page_table(vm);
> - vm->pgd_created = true;
> + if (!vm->mmu.pgd_created) {
> + vm->mmu.pgd = vm_alloc_page_table(vm);
> + vm->mmu.pgd_created = true;
> }
> }
>
> @@ -175,7 +175,7 @@ static void *virt_get_pte(struct kvm_vm *vm, uint64_t *parent_pte,
> uint64_t *page_table = addr_gpa2hva(vm, pt_gpa);
> int index = (vaddr >> PG_LEVEL_SHIFT(level)) & 0x1ffu;
>
> - TEST_ASSERT((*parent_pte & PTE_PRESENT_MASK) || parent_pte == &vm->pgd,
> + TEST_ASSERT((*parent_pte & PTE_PRESENT_MASK) || parent_pte == &vm->mmu.pgd,
> "Parent PTE (level %d) not PRESENT for gva: 0x%08lx",
> level + 1, vaddr);
>
> @@ -218,7 +218,7 @@ static uint64_t *virt_create_upper_pte(struct kvm_vm *vm,
> void __virt_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr, int level)
> {
> const uint64_t pg_size = PG_LEVEL_SIZE(level);
> - uint64_t *pte = &vm->pgd;
> + uint64_t *pte = &vm->mmu.pgd;
> int current_level;
>
> TEST_ASSERT(vm->mode == VM_MODE_PXXVYY_4K,
> @@ -243,7 +243,7 @@ void __virt_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr, int level)
> * Allocate upper level page tables, if not already present. Return
> * early if a hugepage was created.
> */
> - for (current_level = vm->pgtable_levels;
> + for (current_level = vm->mmu.pgtable_levels;
> current_level > PG_LEVEL_4K;
> current_level--) {
> pte = virt_create_upper_pte(vm, pte, vaddr, paddr,
> @@ -309,14 +309,14 @@ static bool vm_is_target_pte(uint64_t *pte, int *level, int current_level)
> static uint64_t *__vm_get_page_table_entry(struct kvm_vm *vm, uint64_t vaddr,
> int *level)
> {
> - int va_width = 12 + (vm->pgtable_levels) * 9;
> - uint64_t *pte = &vm->pgd;
> + int va_width = 12 + (vm->mmu.pgtable_levels) * 9;
> + uint64_t *pte = &vm->mmu.pgd;
> int current_level;
>
> TEST_ASSERT(!vm->arch.is_pt_protected,
> "Walking page tables of protected guests is impossible");
>
> - TEST_ASSERT(*level >= PG_LEVEL_NONE && *level <= vm->pgtable_levels,
> + TEST_ASSERT(*level >= PG_LEVEL_NONE && *level <= vm->mmu.pgtable_levels,
> "Invalid PG_LEVEL_* '%d'", *level);
>
> TEST_ASSERT(vm->mode == VM_MODE_PXXVYY_4K,
> @@ -332,7 +332,7 @@ static uint64_t *__vm_get_page_table_entry(struct kvm_vm *vm, uint64_t vaddr,
> (((int64_t)vaddr << (64 - va_width) >> (64 - va_width))),
> "Canonical check failed. The virtual address is invalid.");
>
> - for (current_level = vm->pgtable_levels;
> + for (current_level = vm->mmu.pgtable_levels;
> current_level > PG_LEVEL_4K;
> current_level--) {
> pte = virt_get_pte(vm, pte, vaddr, current_level);
> @@ -357,7 +357,7 @@ void virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
> uint64_t *pde, *pde_start;
> uint64_t *pte, *pte_start;
>
> - if (!vm->pgd_created)
> + if (!vm->mmu.pgd_created)
> return;
>
> fprintf(stream, "%*s "
> @@ -365,7 +365,7 @@ void virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
> fprintf(stream, "%*s index hvaddr gpaddr "
> "addr w exec dirty\n",
> indent, "");
> - pml4e_start = (uint64_t *) addr_gpa2hva(vm, vm->pgd);
> + pml4e_start = (uint64_t *) addr_gpa2hva(vm, vm->mmu.pgd);
> for (uint16_t n1 = 0; n1 <= 0x1ffu; n1++) {
> pml4e = &pml4e_start[n1];
> if (!(*pml4e & PTE_PRESENT_MASK))
> @@ -538,7 +538,7 @@ static void vcpu_init_sregs(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
> sregs.cr4 |= X86_CR4_PAE | X86_CR4_OSFXSR;
> if (kvm_cpu_has(X86_FEATURE_XSAVE))
> sregs.cr4 |= X86_CR4_OSXSAVE;
> - if (vm->pgtable_levels == 5)
> + if (vm->mmu.pgtable_levels == 5)
> sregs.cr4 |= X86_CR4_LA57;
> sregs.efer |= (EFER_LME | EFER_LMA | EFER_NX);
>
> @@ -549,7 +549,7 @@ static void vcpu_init_sregs(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
> kvm_seg_set_kernel_data_64bit(&sregs.gs);
> kvm_seg_set_tss_64bit(vm->arch.tss, &sregs.tr);
>
> - sregs.cr3 = vm->pgd;
> + sregs.cr3 = vm->mmu.pgd;
> vcpu_sregs_set(vcpu, &sregs);
> }
>
> diff --git a/tools/testing/selftests/kvm/x86/vmx_nested_la57_state_test.c b/tools/testing/selftests/kvm/x86/vmx_nested_la57_state_test.c
> index cf1d2d1f2a8f..915c42001dba 100644
> --- a/tools/testing/selftests/kvm/x86/vmx_nested_la57_state_test.c
> +++ b/tools/testing/selftests/kvm/x86/vmx_nested_la57_state_test.c
> @@ -90,7 +90,7 @@ int main(int argc, char *argv[])
> * L1 needs to read its own PML5 table to set up L2. Identity map
> * the PML5 table to facilitate this.
> */
> - virt_map(vm, vm->pgd, vm->pgd, 1);
> + virt_map(vm, vm->mmu.pgd, vm->mmu.pgd, 1);
>
> vcpu_alloc_vmx(vm, &vmx_pages_gva);
> vcpu_args_set(vcpu, 1, vmx_pages_gva);
> --
> 2.52.0.351.gbe84eed79e-goog
>
More information about the linux-riscv
mailing list