[PATCH 8/9] RISC-V: lib: Use named labels in memset
Andrew Jones
ajones at ventanamicro.com
Thu Oct 27 06:02:46 PDT 2022
In a coming patch we'll be adding more branch targets. Let's
change the numeric labels to named labels to make it easier
to read and integrate with.
No functional change intended.
Signed-off-by: Andrew Jones <ajones at ventanamicro.com>
---
arch/riscv/lib/memset.S | 29 +++++++++++++++--------------
1 file changed, 15 insertions(+), 14 deletions(-)
diff --git a/arch/riscv/lib/memset.S b/arch/riscv/lib/memset.S
index e613c5c27998..74e4c7feec00 100644
--- a/arch/riscv/lib/memset.S
+++ b/arch/riscv/lib/memset.S
@@ -13,7 +13,7 @@ WEAK(memset)
/* Defer to byte-oriented fill for small sizes */
sltiu a3, a2, 16
- bnez a3, 4f
+ bnez a3, .Lfinish
/*
* Round to nearest XLEN-aligned address
@@ -21,17 +21,18 @@ WEAK(memset)
*/
addi a3, t0, SZREG-1
andi a3, a3, ~(SZREG-1)
- beq a3, t0, 2f /* Skip if already aligned */
+ beq a3, t0, .Ldo_duff /* Skip if already aligned */
/* Handle initial misalignment */
sub a4, a3, t0
-1:
+.Lmisaligned1:
sb a1, 0(t0)
addi t0, t0, 1
- bltu t0, a3, 1b
+ bltu t0, a3, .Lmisaligned1
sub a2, a2, a4 /* Update count */
-2: /* Duff's device with 32 XLEN stores per iteration */
+.Ldo_duff:
+ /* Duff's device with 32 XLEN stores per iteration */
/* Broadcast value into all bytes */
andi a1, a1, 0xff
slli a3, a1, 8
@@ -48,7 +49,7 @@ WEAK(memset)
add a3, t0, a4
andi a4, a4, 31*SZREG /* Calculate remainder */
- beqz a4, 3f /* Shortcut if no remainder */
+ beqz a4, .Lduff_loop /* Shortcut if no remainder */
neg a4, a4
addi a4, a4, 32*SZREG /* Calculate initial offset */
@@ -57,13 +58,13 @@ WEAK(memset)
/* Jump into loop body */
/* Assumes 32-bit instruction lengths */
- la a5, 3f
+ la a5, .Lduff_loop
#ifdef CONFIG_64BIT
srli a4, a4, 1
#endif
add a5, a5, a4
jr a5
-3:
+.Lduff_loop:
REG_S a1, 0(t0)
REG_S a1, SZREG(t0)
REG_S a1, 2*SZREG(t0)
@@ -98,17 +99,17 @@ WEAK(memset)
REG_S a1, 31*SZREG(t0)
addi t0, t0, 32*SZREG
- bltu t0, a3, 3b
+ bltu t0, a3, .Lduff_loop
andi a2, a2, SZREG-1 /* Update count */
-4:
+.Lfinish:
/* Handle trailing misalignment */
- beqz a2, 6f
+ beqz a2, .Ldone
add a3, t0, a2
-5:
+.Lmisaligned2:
sb a1, 0(t0)
addi t0, t0, 1
- bltu t0, a3, 5b
-6:
+ bltu t0, a3, .Lmisaligned2
+.Ldone:
ret
END(__memset)
--
2.37.3
More information about the kvm-riscv
mailing list