[PATCH v4 0/5] Invalidate secondary IOMMU TLB on permission upgrade
Alistair Popple
apopple at nvidia.com
Tue Jul 25 06:42:02 PDT 2023
The main change is to move secondary TLB invalidation mmu notifier
callbacks into the architecture specific TLB flushing functions. This
makes secondary TLB invalidation mostly match CPU invalidation while
still allowing efficient range based invalidations based on the
existing TLB batching code.
Changes for v4:
- Fixed a NULL pointer dereference when registering the first
notifier with mmu_interval_notifier_insert() instead of
mmu_notifier_register() - Thanks Dan and Chaitanya for the bug
reports.
- Collected Acked/Reviewed tags.
- Don't call the notifier from radix__local_flush_tlb_page() on
PowerPC.
Changes for v3:
- On x86 call the invalidation when adding pending TLB invalidates
rather than when flushing the batch. This is because the mm is
required. It also matches what happens on ARM64. Fixes a bug
reported by SeongJae Park (thanks!)
Changes for v2:
- Rebased on linux-next commit 906fa30154ef ("mm/rmap: correct stale
comment of rmap_walk_anon and rmap_walk_file") to fix a minor
integration conflict with "arm64: support batched/deferred tlb
shootdown during page reclamation/migration". This series will need
to be applied after the conflicting patch.
- Reordered the function rename until the end of the series as many
places that were getting renamed ended up being removed anyway.
- Fixed a couple of build issues which broke bisection.
- Added a minor patch to fix up a stale/incorrect comment.
==========
Background
==========
The arm64 architecture specifies TLB permission bits may be cached and
therefore the TLB must be invalidated during permission upgrades. For
the CPU this currently occurs in the architecture specific
ptep_set_access_flags() routine.
Secondary TLBs such as implemented by the SMMU IOMMU match the CPU
architecture specification and may also cache permission bits and
require the same TLB invalidations. This may be achieved in one of two
ways.
Some SMMU implementations implement broadcast TLB maintenance
(BTM). This snoops CPU TLB invalidates and will invalidate any
secondary TLB at the same time as the CPU. However implementations are
not required to implement BTM.
Implementations without BTM rely on mmu notifier callbacks to send
explicit TLB invalidation commands to invalidate SMMU TLB. Therefore
either generic kernel code or architecture specific code needs to call
the mmu notifier on permission upgrade.
Currently that doesn't happen so devices will fault indefinitely when
writing to a PTE that was previously read-only as nothing invalidates
the SMMU TLB.
========
Solution
========
To fix this the series first renames the .invalidate_range() callback
to .arch_invalidate_secondary_tlbs() as suggested by Jason and Sean to
make it clear this callback is only used for secondary TLBs. That was
made possible thanks to Sean's series [1] to remove KVM's incorrect
usage.
Based on feedback from Jason [2] the proposed solution to the bug is
to move the calls to mmu_notifier_arch_invalidate_secondary_tlbs()
closer to the architecture specific TLB invalidation code. This
ensures the secondary TLB won't miss invalidations, including the
existing invalidation in the ARM64 code to deal with permission
upgrade.
Currently only ARM64, PowerPC and x86 have IOMMU with secondary TLBs
requiring SW invalidation so the notifier is only called for those
architectures. It is also not called for invalidation of kernel
mappings as no secondary IOMMU implementations can access those and
hence it is not required.
[1] - https://lore.kernel.org/all/20230602011518.787006-1-seanjc@google.com/
[2] - https://lore.kernel.org/linux-mm/ZJMR5bw8l+BbzdJ7@ziepe.ca/
Alistair Popple (5):
arm64/smmu: Use TLBI ASID when invalidating entire range
mmu_notifiers: Fixup comment in mmu_interval_read_begin()
mmu_notifiers: Call invalidate_range() when invalidating TLBs
mmu_notifiers: Don't invalidate secondary TLBs as part of mmu_notifier_invalidate_range_end()
mmu_notifiers: Rename invalidate_range notifier
arch/arm64/include/asm/tlbflush.h | 5 +-
arch/powerpc/include/asm/book3s/64/tlbflush.h | 1 +-
arch/powerpc/mm/book3s64/radix_hugetlbpage.c | 1 +-
arch/powerpc/mm/book3s64/radix_tlb.c | 4 +-
arch/x86/include/asm/tlbflush.h | 2 +-
arch/x86/mm/tlb.c | 2 +-
drivers/iommu/amd/iommu_v2.c | 10 +-
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c | 29 +++--
drivers/iommu/intel/svm.c | 8 +-
drivers/misc/ocxl/link.c | 8 +-
include/asm-generic/tlb.h | 1 +-
include/linux/mmu_notifier.h | 104 ++++-------------
kernel/events/uprobes.c | 2 +-
mm/huge_memory.c | 29 +----
mm/hugetlb.c | 8 +-
mm/memory.c | 8 +-
mm/migrate_device.c | 9 +-
mm/mmu_notifier.c | 50 +++-----
mm/rmap.c | 40 +-------
19 files changed, 110 insertions(+), 211 deletions(-)
base-commit: 906fa30154ef42f93d28d7322860e76c6ae392ac
--
git-series 0.9.1
More information about the linux-arm-kernel
mailing list