[PATCH v3 13/29] iommu/arm-smmu-v3: Move TLB range invalidation into a macro
Mostafa Saleh
smostafa at google.com
Mon Jul 28 10:53:00 PDT 2025
Range TLB invalidation has a very specific algorithm, instead of
re-writing it for the hypervisor, put it in a macro so it can be
re-used.
Signed-off-by: Mostafa Saleh <smostafa at google.com>
---
.../arm/arm-smmu-v3/arm-smmu-v3-common.h | 65 +++++++++++++++++++
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c | 59 +----------------
2 files changed, 68 insertions(+), 56 deletions(-)
diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-common.h b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-common.h
index f115ff9a547f..ab6db8b4b9f2 100644
--- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-common.h
+++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-common.h
@@ -540,4 +540,69 @@ static inline void queue_write(__le64 *dst, u64 *src, size_t n_dwords)
for (i = 0; i < n_dwords; ++i)
*dst++ = cpu_to_le64(*src++);
}
+
+/**
+ * arm_smmu_tlb_inv_build - Create a range invalidation command
+ * @cmd: Base command initialized with OPCODE (S1, S2..), vmid and asid.
+ * @iova: Start IOVA to invalidate
+ * @size: Size of range
+ * @granule: Granule of invalidation
+ * @pgsize_bitmap: Page size bit map of the page table.
+ * @smmu: Struct for the smmu, must have ::features
+ * @add_cmd: Function to send/batch the invalidation command
+ * @cmds: Incase of batching, it includes the pointer to the batch
+ */
+#define arm_smmu_tlb_inv_build(cmd, iova, size, granule, pgsize_bitmap, smmu, add_cmd, cmds) \
+{ \
+ unsigned long _iova = (iova); \
+ size_t _size = (size); \
+ size_t _granule = (granule); \
+ unsigned long end = _iova + _size, num_pages = 0, tg = 0; \
+ size_t inv_range = _granule; \
+ if (smmu->features & ARM_SMMU_FEAT_RANGE_INV) { \
+ /* Get the leaf page size */ \
+ tg = __ffs(pgsize_bitmap); \
+ num_pages = _size >> tg; \
+ /* Convert page size of 12,14,16 (log2) to 1,2,3 */ \
+ cmd->tlbi.tg = (tg - 10) / 2; \
+ /*
+ * Determine what level the granule is at. For non-leaf, both
+ * io-pgtable and SVA pass a nominal last-level granule because
+ * they don't know what level(s) actually apply, so ignore that
+ * and leave TTL=0. However for various errata reasons we still
+ * want to use a range command, so avoid the SVA corner case
+ * where both scale and num could be 0 as well.
+ */ \
+ if (cmd->tlbi.leaf) \
+ cmd->tlbi.ttl = 4 - ((ilog2(_granule) - 3) / (tg - 3)); \
+ else if ((num_pages & CMDQ_TLBI_RANGE_NUM_MAX) == 1) \
+ num_pages++; \
+ } \
+ while (_iova < end) { \
+ if (smmu->features & ARM_SMMU_FEAT_RANGE_INV) { \
+ /*
+ * On each iteration of the loop, the range is 5 bits
+ * worth of the aligned size remaining.
+ * The range in pages is:
+ *
+ * range = (num_pages & (0x1f << __ffs(num_pages)))
+ */ \
+ unsigned long scale, num; \
+ /* Determine the power of 2 multiple number of pages */ \
+ scale = __ffs(num_pages); \
+ cmd->tlbi.scale = scale; \
+ /* Determine how many chunks of 2^scale size we have */ \
+ num = (num_pages >> scale) & CMDQ_TLBI_RANGE_NUM_MAX; \
+ cmd->tlbi.num = num - 1; \
+ /* range is num * 2^scale * pgsize */ \
+ inv_range = num << (scale + tg); \
+ /* Clear out the lower order bits for the next iteration */ \
+ num_pages -= num << scale; \
+ } \
+ cmd->tlbi.addr = iova; \
+ add_cmd(smmu, cmds, cmd); \
+ _iova += inv_range; \
+ } \
+} \
+
#endif /* _ARM_SMMU_V3_COMMON_H */
\ No newline at end of file
diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
index 066d9d9e8088..245b39ddef66 100644
--- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
+++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
@@ -2075,68 +2075,15 @@ static void __arm_smmu_tlb_inv_range(struct arm_smmu_cmdq_ent *cmd,
struct arm_smmu_domain *smmu_domain)
{
struct arm_smmu_device *smmu = smmu_domain->smmu;
- unsigned long end = iova + size, num_pages = 0, tg = 0;
- size_t inv_range = granule;
struct arm_smmu_cmdq_batch cmds;
if (!size)
return;
- if (smmu->features & ARM_SMMU_FEAT_RANGE_INV) {
- /* Get the leaf page size */
- tg = __ffs(smmu_domain->domain.pgsize_bitmap);
-
- num_pages = size >> tg;
-
- /* Convert page size of 12,14,16 (log2) to 1,2,3 */
- cmd->tlbi.tg = (tg - 10) / 2;
-
- /*
- * Determine what level the granule is at. For non-leaf, both
- * io-pgtable and SVA pass a nominal last-level granule because
- * they don't know what level(s) actually apply, so ignore that
- * and leave TTL=0. However for various errata reasons we still
- * want to use a range command, so avoid the SVA corner case
- * where both scale and num could be 0 as well.
- */
- if (cmd->tlbi.leaf)
- cmd->tlbi.ttl = 4 - ((ilog2(granule) - 3) / (tg - 3));
- else if ((num_pages & CMDQ_TLBI_RANGE_NUM_MAX) == 1)
- num_pages++;
- }
-
arm_smmu_cmdq_batch_init(smmu, &cmds, cmd);
-
- while (iova < end) {
- if (smmu->features & ARM_SMMU_FEAT_RANGE_INV) {
- /*
- * On each iteration of the loop, the range is 5 bits
- * worth of the aligned size remaining.
- * The range in pages is:
- *
- * range = (num_pages & (0x1f << __ffs(num_pages)))
- */
- unsigned long scale, num;
-
- /* Determine the power of 2 multiple number of pages */
- scale = __ffs(num_pages);
- cmd->tlbi.scale = scale;
-
- /* Determine how many chunks of 2^scale size we have */
- num = (num_pages >> scale) & CMDQ_TLBI_RANGE_NUM_MAX;
- cmd->tlbi.num = num - 1;
-
- /* range is num * 2^scale * pgsize */
- inv_range = num << (scale + tg);
-
- /* Clear out the lower order bits for the next iteration */
- num_pages -= num << scale;
- }
-
- cmd->tlbi.addr = iova;
- arm_smmu_cmdq_batch_add(smmu, &cmds, cmd);
- iova += inv_range;
- }
+ arm_smmu_tlb_inv_build(cmd, iova, size, granule,
+ smmu_domain->domain.pgsize_bitmap,
+ smmu, arm_smmu_cmdq_batch_add, &cmds);
arm_smmu_cmdq_batch_submit(smmu, &cmds);
}
--
2.50.1.552.g942d659e1b-goog
More information about the linux-arm-kernel
mailing list