[kvm-unit-tests PATCH v2 3/7] riscv: Support up to 34-bit physical addresses on rv32, sort of
Andrew Jones
andrew.jones at linux.dev
Mon Aug 12 06:44:55 PDT 2024
Change virt_to_phys() and phys_to_virt() to use phys_addr_t instead
of unsigned long. This allows 32-bit builds to use physical addresses
over 32 bits wide (the spec allows up to 34 bits). But, to keep
things simple, we don't expect physical addresses wider than 32 bits
in most the library code (and that's ensured by sprinkling around
some asserts). IOW, the support is really only for unit tests which
want to test with an additional high memory region.
Signed-off-by: Andrew Jones <andrew.jones at linux.dev>
---
lib/riscv/asm/io.h | 4 ++--
lib/riscv/mmu.c | 32 ++++++++++++++++++++------------
lib/riscv/smp.c | 7 ++++++-
3 files changed, 28 insertions(+), 15 deletions(-)
diff --git a/lib/riscv/asm/io.h b/lib/riscv/asm/io.h
index 37a130e533c9..a48a9aa654dd 100644
--- a/lib/riscv/asm/io.h
+++ b/lib/riscv/asm/io.h
@@ -77,10 +77,10 @@ static inline u64 __raw_readq(const volatile void __iomem *addr)
void __iomem *ioremap(phys_addr_t phys_addr, size_t size);
#define virt_to_phys virt_to_phys
-unsigned long virt_to_phys(volatile void *address);
+phys_addr_t virt_to_phys(volatile void *address);
#define phys_to_virt phys_to_virt
-void *phys_to_virt(unsigned long address);
+void *phys_to_virt(phys_addr_t address);
#include <asm-generic/io.h>
diff --git a/lib/riscv/mmu.c b/lib/riscv/mmu.c
index 2c9c4f376ac9..6ab1f15a99ae 100644
--- a/lib/riscv/mmu.c
+++ b/lib/riscv/mmu.c
@@ -18,9 +18,16 @@ static int pte_index(uintptr_t vaddr, int level)
return (vaddr >> (PGDIR_BITS * level + PAGE_SHIFT)) & PGDIR_MASK;
}
+static phys_addr_t pteval_to_phys_addr(pteval_t pteval)
+{
+ return (phys_addr_t)((pteval & PTE_PPN) >> PPN_SHIFT) << PAGE_SHIFT;
+}
+
static pte_t *pteval_to_ptep(pteval_t pteval)
{
- return (pte_t *)(((pteval & PTE_PPN) >> PPN_SHIFT) << PAGE_SHIFT);
+ phys_addr_t paddr = pteval_to_phys_addr(pteval);
+ assert(paddr == __pa(paddr));
+ return (pte_t *)__pa(paddr);
}
static pteval_t ptep_to_pteval(pte_t *ptep)
@@ -106,7 +113,7 @@ void __mmu_enable(unsigned long satp)
void mmu_enable(unsigned long mode, pgd_t *pgtable)
{
- unsigned long ppn = (unsigned long)pgtable >> PAGE_SHIFT;
+ unsigned long ppn = __pa(pgtable) >> PAGE_SHIFT;
unsigned long satp = mode | ppn;
assert(!(ppn & ~SATP_PPN));
@@ -118,6 +125,9 @@ void *setup_mmu(phys_addr_t top, void *opaque)
struct mem_region *r;
pgd_t *pgtable;
+ /* The initial page table uses an identity mapping. */
+ assert(top == __pa(top));
+
if (!__initial_pgtable)
__initial_pgtable = alloc_page();
pgtable = __initial_pgtable;
@@ -146,7 +156,8 @@ void __iomem *ioremap(phys_addr_t phys_addr, size_t size)
pgd_t *pgtable = current_pgtable();
bool flush = true;
- assert(sizeof(long) == 8 || !(phys_addr >> 32));
+ /* I/O is always identity mapped. */
+ assert(end == __pa(end));
if (!pgtable) {
if (!__initial_pgtable)
@@ -158,7 +169,7 @@ void __iomem *ioremap(phys_addr_t phys_addr, size_t size)
mmu_set_range_ptes(pgtable, start, start, end,
__pgprot(_PAGE_READ | _PAGE_WRITE), flush);
- return (void __iomem *)(unsigned long)phys_addr;
+ return (void __iomem *)__pa(phys_addr);
}
phys_addr_t virt_to_pte_phys(pgd_t *pgtable, void *virt)
@@ -179,27 +190,24 @@ phys_addr_t virt_to_pte_phys(pgd_t *pgtable, void *virt)
if (!pte_val(*ptep))
return 0;
- return __pa(pteval_to_ptep(pte_val(*ptep))) | offset_in_page(virt);
+ return pteval_to_phys_addr(pte_val(*ptep)) | offset_in_page(virt);
}
-unsigned long virt_to_phys(volatile void *address)
+phys_addr_t virt_to_phys(volatile void *address)
{
unsigned long satp = csr_read(CSR_SATP);
pgd_t *pgtable = (pgd_t *)((satp & SATP_PPN) << PAGE_SHIFT);
- phys_addr_t paddr;
if ((satp >> SATP_MODE_SHIFT) == 0)
return __pa(address);
- paddr = virt_to_pte_phys(pgtable, (void *)address);
- assert(sizeof(long) == 8 || !(paddr >> 32));
-
- return (unsigned long)paddr;
+ return virt_to_pte_phys(pgtable, (void *)address);
}
-void *phys_to_virt(unsigned long address)
+void *phys_to_virt(phys_addr_t address)
{
/* @address must have an identity mapping for this to work. */
+ assert(address == __pa(address));
assert(virt_to_phys(__va(address)) == address);
return __va(address);
}
diff --git a/lib/riscv/smp.c b/lib/riscv/smp.c
index 7e4bb5b76903..4d373e0a29a8 100644
--- a/lib/riscv/smp.c
+++ b/lib/riscv/smp.c
@@ -8,6 +8,7 @@
#include <alloc_page.h>
#include <cpumask.h>
#include <asm/csr.h>
+#include <asm/io.h>
#include <asm/mmu.h>
#include <asm/page.h>
#include <asm/processor.h>
@@ -36,6 +37,7 @@ secondary_func_t secondary_cinit(struct secondary_data *data)
static void __smp_boot_secondary(int cpu, secondary_func_t func)
{
struct secondary_data *sp = alloc_pages(1) + SZ_8K - 16;
+ phys_addr_t sp_phys;
struct sbiret ret;
sp -= sizeof(struct secondary_data);
@@ -43,7 +45,10 @@ static void __smp_boot_secondary(int cpu, secondary_func_t func)
sp->stvec = csr_read(CSR_STVEC);
sp->func = func;
- ret = sbi_hart_start(cpus[cpu].hartid, (unsigned long)&secondary_entry, __pa(sp));
+ sp_phys = virt_to_phys(sp);
+ assert(sp_phys == __pa(sp_phys));
+
+ ret = sbi_hart_start(cpus[cpu].hartid, (unsigned long)&secondary_entry, __pa(sp_phys));
assert(ret.error == SBI_SUCCESS);
}
--
2.45.2
More information about the kvm-riscv
mailing list