[PATCH 6/7] v2: Fix most of the "Checks" from "checkpatch.pl"

Shahab Vahedi list+bpf at vahedi.org
Tue Apr 30 08:09:36 PDT 2024


From: Shahab Vahedi <shahab at synopsys.com>

If they're left untouched, then it was decided like that.

The command that was used for checkpatch.pl:

$ checkpatch.pl ... --strict --no-signoff                          \
                    --ignore AVOID_BUG,SPLIT_STRING,COMMIT_MESSAGE \
                    --git <before_start>..<end>
---
 arch/arc/net/bpf_jit.h       | 121 ++++++++++++++++++-----------------
 arch/arc/net/bpf_jit_arcv2.c |  84 ++++++++++++------------
 arch/arc/net/bpf_jit_core.c  |  67 ++++++++++---------
 3 files changed, 135 insertions(+), 137 deletions(-)

diff --git a/arch/arc/net/bpf_jit.h b/arch/arc/net/bpf_jit.h
index 9fc70d97415b..ec44873c42d1 100644
--- a/arch/arc/net/bpf_jit.h
+++ b/arch/arc/net/bpf_jit.h
@@ -39,75 +39,75 @@
 
 /************** Functions that the back-end must provide **************/
 /* Extension for 32-bit operations. */
-extern inline u8 zext(u8 *buf, u8 rd);
+inline u8 zext(u8 *buf, u8 rd);
 /***** Moves *****/
-extern u8 mov_r32(u8 *buf, u8 rd, u8 rs, u8 sign_ext);
-extern u8 mov_r32_i32(u8 *buf, u8 reg, s32 imm);
-extern u8 mov_r64(u8 *buf, u8 rd, u8 rs, u8 sign_ext);
-extern u8 mov_r64_i32(u8 *buf, u8 reg, s32 imm);
-extern u8 mov_r64_i64(u8 *buf, u8 reg, u32 lo, u32 hi);
+u8 mov_r32(u8 *buf, u8 rd, u8 rs, u8 sign_ext);
+u8 mov_r32_i32(u8 *buf, u8 reg, s32 imm);
+u8 mov_r64(u8 *buf, u8 rd, u8 rs, u8 sign_ext);
+u8 mov_r64_i32(u8 *buf, u8 reg, s32 imm);
+u8 mov_r64_i64(u8 *buf, u8 reg, u32 lo, u32 hi);
 /***** Loads and stores *****/
-extern u8 load_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size, bool sign_ext);
-extern u8 store_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size);
-extern u8 store_i(u8 *buf, s32 imm, u8 rd, s16 off, u8 size);
+u8 load_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size, bool sign_ext);
+u8 store_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size);
+u8 store_i(u8 *buf, s32 imm, u8 rd, s16 off, u8 size);
 /***** Addition *****/
-extern u8 add_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 add_r32_i32(u8 *buf, u8 rd, s32 imm);
-extern u8 add_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 add_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 add_r32(u8 *buf, u8 rd, u8 rs);
+u8 add_r32_i32(u8 *buf, u8 rd, s32 imm);
+u8 add_r64(u8 *buf, u8 rd, u8 rs);
+u8 add_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Subtraction *****/
-extern u8 sub_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 sub_r32_i32(u8 *buf, u8 rd, s32 imm);
-extern u8 sub_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 sub_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 sub_r32(u8 *buf, u8 rd, u8 rs);
+u8 sub_r32_i32(u8 *buf, u8 rd, s32 imm);
+u8 sub_r64(u8 *buf, u8 rd, u8 rs);
+u8 sub_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Multiplication *****/
-extern u8 mul_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 mul_r32_i32(u8 *buf, u8 rd, s32 imm);
-extern u8 mul_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 mul_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 mul_r32(u8 *buf, u8 rd, u8 rs);
+u8 mul_r32_i32(u8 *buf, u8 rd, s32 imm);
+u8 mul_r64(u8 *buf, u8 rd, u8 rs);
+u8 mul_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Division *****/
-extern u8 div_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext);
-extern u8 div_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext);
+u8 div_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext);
+u8 div_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext);
 /***** Remainder *****/
-extern u8 mod_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext);
-extern u8 mod_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext);
+u8 mod_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext);
+u8 mod_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext);
 /***** Bitwise AND *****/
-extern u8 and_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 and_r32_i32(u8 *buf, u8 rd, s32 imm);
-extern u8 and_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 and_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 and_r32(u8 *buf, u8 rd, u8 rs);
+u8 and_r32_i32(u8 *buf, u8 rd, s32 imm);
+u8 and_r64(u8 *buf, u8 rd, u8 rs);
+u8 and_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Bitwise OR *****/
-extern u8 or_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 or_r32_i32(u8 *buf, u8 rd, s32 imm);
-extern u8 or_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 or_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 or_r32(u8 *buf, u8 rd, u8 rs);
+u8 or_r32_i32(u8 *buf, u8 rd, s32 imm);
+u8 or_r64(u8 *buf, u8 rd, u8 rs);
+u8 or_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Bitwise XOR *****/
-extern u8 xor_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 xor_r32_i32(u8 *buf, u8 rd, s32 imm);
-extern u8 xor_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 xor_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 xor_r32(u8 *buf, u8 rd, u8 rs);
+u8 xor_r32_i32(u8 *buf, u8 rd, s32 imm);
+u8 xor_r64(u8 *buf, u8 rd, u8 rs);
+u8 xor_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Bitwise Negate *****/
-extern u8 neg_r32(u8 *buf, u8 r);
-extern u8 neg_r64(u8 *buf, u8 r);
+u8 neg_r32(u8 *buf, u8 r);
+u8 neg_r64(u8 *buf, u8 r);
 /***** Bitwise left shift *****/
-extern u8 lsh_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 lsh_r32_i32(u8 *buf, u8 rd, u8 imm);
-extern u8 lsh_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 lsh_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 lsh_r32(u8 *buf, u8 rd, u8 rs);
+u8 lsh_r32_i32(u8 *buf, u8 rd, u8 imm);
+u8 lsh_r64(u8 *buf, u8 rd, u8 rs);
+u8 lsh_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Bitwise right shift (logical) *****/
-extern u8 rsh_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 rsh_r32_i32(u8 *buf, u8 rd, u8 imm);
-extern u8 rsh_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 rsh_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 rsh_r32(u8 *buf, u8 rd, u8 rs);
+u8 rsh_r32_i32(u8 *buf, u8 rd, u8 imm);
+u8 rsh_r64(u8 *buf, u8 rd, u8 rs);
+u8 rsh_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Bitwise right shift (arithmetic) *****/
-extern u8 arsh_r32(u8 *buf, u8 rd, u8 rs);
-extern u8 arsh_r32_i32(u8 *buf, u8 rd, u8 imm);
-extern u8 arsh_r64(u8 *buf, u8 rd, u8 rs);
-extern u8 arsh_r64_i32(u8 *buf, u8 rd, s32 imm);
+u8 arsh_r32(u8 *buf, u8 rd, u8 rs);
+u8 arsh_r32_i32(u8 *buf, u8 rd, u8 imm);
+u8 arsh_r64(u8 *buf, u8 rd, u8 rs);
+u8 arsh_r64_i32(u8 *buf, u8 rd, s32 imm);
 /***** Frame related *****/
-extern u32 mask_for_used_regs(u8 bpf_reg, bool is_call);
-extern u8 arc_prologue(u8 *buf, u32 usage, u16 frame_size);
-extern u8 arc_epilogue(u8 *buf, u32 usage, u16 frame_size);
+u32 mask_for_used_regs(u8 bpf_reg, bool is_call);
+u8 arc_prologue(u8 *buf, u32 usage, u16 frame_size);
+u8 arc_epilogue(u8 *buf, u32 usage, u16 frame_size);
 /***** Jumps *****/
 /*
  * Different sorts of conditions (ARC enum as opposed to BPF_*).
@@ -130,6 +130,7 @@ enum ARC_CC {
 	ARC_CC_SET,		/* test        */
 	ARC_CC_LAST
 };
+
 /*
  * A few notes:
  *
@@ -144,13 +145,13 @@ enum ARC_CC {
  *   things simpler (offsets are in the range of u32 which is more than
  *   enough).
  */
-extern bool check_jmp_32(u32 curr_off, u32 targ_off, u8 cond);
-extern bool check_jmp_64(u32 curr_off, u32 targ_off, u8 cond);
-extern u8 gen_jmp_32(u8 *buf, u8 rd, u8 rs, u8 cond, u32 c_off, u32 t_off);
-extern u8 gen_jmp_64(u8 *buf, u8 rd, u8 rs, u8 cond, u32 c_off, u32 t_off);
+bool check_jmp_32(u32 curr_off, u32 targ_off, u8 cond);
+bool check_jmp_64(u32 curr_off, u32 targ_off, u8 cond);
+u8 gen_jmp_32(u8 *buf, u8 rd, u8 rs, u8 cond, u32 c_off, u32 t_off);
+u8 gen_jmp_64(u8 *buf, u8 rd, u8 rs, u8 cond, u32 c_off, u32 t_off);
 /***** Miscellaneous *****/
-extern u8 gen_func_call(u8 *buf, ARC_ADDR func_addr, bool external_func);
-extern u8 arc_to_bpf_return(u8 *buf);
+u8 gen_func_call(u8 *buf, ARC_ADDR func_addr, bool external_func);
+u8 arc_to_bpf_return(u8 *buf);
 /*
  * - Perform byte swaps on "rd" based on the "size".
  * - If "force" is set, do it unconditionally. Otherwise, consider the
diff --git a/arch/arc/net/bpf_jit_arcv2.c b/arch/arc/net/bpf_jit_arcv2.c
index 8b7ae2f11f38..31bfb6e9ce00 100644
--- a/arch/arc/net/bpf_jit_arcv2.c
+++ b/arch/arc/net/bpf_jit_arcv2.c
@@ -5,7 +5,7 @@
  * Copyright (c) 2024 Synopsys Inc.
  * Author: Shahab Vahedi <shahab at synopsys.com>
  */
-#include <asm/bug.h>
+#include <linux/bug.h>
 #include "bpf_jit.h"
 
 /* ARC core registers. */
@@ -91,7 +91,6 @@ const u8 bpf2arc[][2] = {
 #define REG_LO(r) (bpf2arc[(r)][0])
 #define REG_HI(r) (bpf2arc[(r)][1])
 
-
 /*
  * To comply with ARCv2 ABI, BPF's arg5 must be put on stack. After which,
  * the stack needs to be restored by ARG5_SIZE.
@@ -201,7 +200,7 @@ enum {
  * c:  cccccc		source
  */
 #define OPC_MOV_CC	0x20ca0000
-#define MOV_CC_I	(1 << 5)
+#define MOV_CC_I	BIT(5)
 #define OPC_MOVU_CC	(OPC_MOV_CC | MOV_CC_I)
 
 /*
@@ -289,7 +288,7 @@ enum {
 #define OPC_ADD		0x20000000
 /* Addition with updating the pertinent flags in "status32" register. */
 #define OPC_ADDF	(OPC_ADD | FLAG(1))
-#define ADDI		(1 << 22)
+#define ADDI		BIT(22)
 #define ADDI_U6(x)	OP_C(x)
 #define OPC_ADDI	(OPC_ADD | ADDI)
 #define OPC_ADDIF	(OPC_ADDI | FLAG(1))
@@ -307,7 +306,7 @@ enum {
  * c:  cccccc		the 2nd input operand
  */
 #define OPC_ADC		0x20010000
-#define ADCI		(1 << 22)
+#define ADCI		BIT(22)
 #define ADCI_U6(x)	OP_C(x)
 #define OPC_ADCI	(OPC_ADC | ADCI)
 
@@ -326,7 +325,7 @@ enum {
 #define OPC_SUB		0x20020000
 /* Subtraction with updating the pertinent flags in "status32" register. */
 #define OPC_SUBF	(OPC_SUB | FLAG(1))
-#define SUBI		(1 << 22)
+#define SUBI		BIT(22)
 #define SUBI_U6(x)	OP_C(x)
 #define OPC_SUBI	(OPC_SUB | SUBI)
 #define OPC_SUB_I	(OPC_SUB | OP_IMM)
@@ -526,7 +525,7 @@ enum {
  * c:  cccccc		amount to be shifted
  */
 #define OPC_ASL		0x28000000
-#define ASL_I		(1 << 22)
+#define ASL_I		BIT(22)
 #define ASLI_U6(x)	OP_C((x) & 31)
 #define OPC_ASLI	(OPC_ASL | ASL_I)
 
@@ -629,13 +628,13 @@ enum {
 
 static inline void emit_2_bytes(u8 *buf, u16 bytes)
 {
-	*((u16 *) buf) = bytes;
+	*((u16 *)buf) = bytes;
 }
 
 static inline void emit_4_bytes(u8 *buf, u32 bytes)
 {
-	emit_2_bytes(buf+0, bytes >>     16);
-	emit_2_bytes(buf+2, bytes  & 0xffff);
+	emit_2_bytes(buf, bytes >> 16);
+	emit_2_bytes(buf + 2, bytes & 0xffff);
 }
 
 static inline u8 bpf_to_arc_size(u8 size)
@@ -686,7 +685,7 @@ static u8 arc_mov_i(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -698,7 +697,7 @@ static u8 arc_mov_i_fixed(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -843,7 +842,7 @@ static u8 arc_add_i(u8 *buf, u8 ra, u8 rb, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -905,7 +904,7 @@ static u8 arc_sub_i(u8 *buf, u8 ra, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -974,7 +973,7 @@ static u8 arc_mpy_i(u8 *buf, u8 ra, u8 rb, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -996,7 +995,7 @@ static u8 arc_mpydu_i(u8 *buf, u8 ra, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -1018,7 +1017,7 @@ static u8 arc_divu_i(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -1040,7 +1039,7 @@ static u8 arc_divs_i(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -1062,7 +1061,7 @@ static u8 arc_remu_i(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -1084,7 +1083,7 @@ static u8 arc_rems_i(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -1106,7 +1105,7 @@ static u8 arc_and_i(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -1151,7 +1150,7 @@ static u8 arc_or_i(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -1171,7 +1170,7 @@ static u8 arc_xor_i(u8 *buf, u8 rd, s32 imm)
 
 	if (buf) {
 		emit_4_bytes(buf, insn);
-		emit_4_bytes(buf+INSN_len_normal, imm);
+		emit_4_bytes(buf + INSN_len_normal, imm);
 	}
 	return INSN_len_normal + INSN_len_imm;
 }
@@ -1449,7 +1448,7 @@ static u8 adjust_mem_access(u8 *buf, s16 *off, u8 size,
 	if (!IN_S9_RANGE(*off) ||
 	    (size == BPF_DW && !IN_S9_RANGE(*off + 4))) {
 		len += arc_add_i(BUF(buf, len),
-				 REG_LO(JIT_REG_TMP), REG_LO(rm), (u32) (*off));
+				 REG_LO(JIT_REG_TMP), REG_LO(rm), (u32)(*off));
 		*arc_reg_mem = REG_LO(JIT_REG_TMP);
 		*off = 0;
 	}
@@ -1468,7 +1467,7 @@ u8 store_r(u8 *buf, u8 rs, u8 rd, s16 off, u8 size)
 		len += arc_st_r(BUF(buf, len), REG_LO(rs), arc_reg_mem,
 				off, ZZ_4_byte);
 		len += arc_st_r(BUF(buf, len), REG_HI(rs), arc_reg_mem,
-				off+4, ZZ_4_byte);
+				off + 4, ZZ_4_byte);
 	} else {
 		u8 zz = bpf_to_arc_size(size);
 
@@ -1504,7 +1503,7 @@ u8 store_i(u8 *buf, s32 imm, u8 rd, s16 off, u8 size)
 		imm = (imm >= 0 ? 0 : -1);
 		len += arc_mov_i(BUF(buf, len), arc_rs, imm);
 		len += arc_st_r(BUF(buf, len), arc_rs, arc_reg_mem,
-				off+4, ZZ_4_byte);
+				off + 4, ZZ_4_byte);
 	} else {
 		u8 zz = bpf_to_arc_size(size);
 
@@ -1579,14 +1578,14 @@ u8 load_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size, bool sign_ext)
 		 */
 		if (REG_LO(rd) != arc_reg_mem) {
 			len += arc_ld_r(BUF(buf, len), REG_LO(rd), arc_reg_mem,
-					off+0, ZZ_4_byte);
+					off, ZZ_4_byte);
 			len += arc_ld_r(BUF(buf, len), REG_HI(rd), arc_reg_mem,
-					off+4, ZZ_4_byte);
+					off + 4, ZZ_4_byte);
 		} else {
 			len += arc_ld_r(BUF(buf, len), REG_HI(rd), arc_reg_mem,
-					off+4, ZZ_4_byte);
+					off + 4, ZZ_4_byte);
 			len += arc_ld_r(BUF(buf, len), REG_LO(rd), arc_reg_mem,
-					off+0, ZZ_4_byte);
+					off, ZZ_4_byte);
 		}
 	}
 
@@ -1984,7 +1983,7 @@ u8 lsh_r64_i32(u8 *buf, u8 rd, s32 imm)
 	const u8 t0   = REG_LO(JIT_REG_TMP);
 	const u8 B_lo = REG_LO(rd);
 	const u8 B_hi = REG_HI(rd);
-	const u8 n    = (u8) imm;
+	const u8 n    = (u8)imm;
 	u8 len = 0;
 
 	if (n == 0) {
@@ -2079,7 +2078,7 @@ u8 rsh_r64_i32(u8 *buf, u8 rd, s32 imm)
 	const u8 t0   = REG_LO(JIT_REG_TMP);
 	const u8 B_lo = REG_LO(rd);
 	const u8 B_hi = REG_HI(rd);
-	const u8 n    = (u8) imm;
+	const u8 n    = (u8)imm;
 	u8 len = 0;
 
 	if (n == 0) {
@@ -2177,7 +2176,7 @@ u8 arsh_r64_i32(u8 *buf, u8 rd, s32 imm)
 	const u8 t0   = REG_LO(JIT_REG_TMP);
 	const u8 B_lo = REG_LO(rd);
 	const u8 B_hi = REG_HI(rd);
-	const u8 n    = (u8) imm;
+	const u8 n    = (u8)imm;
 	u8 len = 0;
 
 	if (n == 0) {
@@ -2418,14 +2417,14 @@ u8 arc_prologue(u8 *buf, u32 usage, u16 frame_size)
 	}
 
 	/* Deal with fp last. */
-	if ((usage & BIT(ARC_R_FP)) || (frame_size > 0))
+	if ((usage & BIT(ARC_R_FP)) || frame_size > 0)
 		len += arc_push_r(BUF(buf, len), ARC_R_FP);
 
 	if (frame_size > 0)
 		len += frame_create(BUF(buf, len), frame_size);
 
 #ifdef ARC_BPF_JIT_DEBUG
-	if ((usage & BIT(ARC_R_FP)) && (frame_size == 0)) {
+	if ((usage & BIT(ARC_R_FP)) && frame_size == 0) {
 		pr_err("FP is being saved while there is no frame.");
 		BUG();
 	}
@@ -2452,7 +2451,7 @@ u8 arc_epilogue(u8 *buf, u32 usage, u16 frame_size)
 	u32 gp_regs = 0;
 
 #ifdef ARC_BPF_JIT_DEBUG
-	if ((usage & BIT(ARC_R_FP)) && (frame_size == 0)) {
+	if ((usage & BIT(ARC_R_FP)) && frame_size == 0) {
 		pr_err("FP is being saved while there is no frame.");
 		BUG();
 	}
@@ -2462,7 +2461,7 @@ u8 arc_epilogue(u8 *buf, u32 usage, u16 frame_size)
 		len += frame_restore(BUF(buf, len));
 
 	/* Deal with fp first. */
-	if ((usage & BIT(ARC_R_FP)) || (frame_size > 0))
+	if ((usage & BIT(ARC_R_FP)) || frame_size > 0)
 		len += arc_pop_r(BUF(buf, len), ARC_R_FP);
 
 	gp_regs = usage & ~(BIT(ARC_R_BLINK) | BIT(ARC_R_FP));
@@ -2533,12 +2532,12 @@ const struct {
 
 	struct {
 		u8 cond[JCC64_NR_OF_JMPS];
-	} jmp[ARC_CC_SLE+1];
+	} jmp[ARC_CC_SLE + 1];
 } arcv2_64_jccs = {
 	.jit_off = {
-		INSN_len_normal*1,
-		INSN_len_normal*2,
-		INSN_len_normal*4
+		INSN_len_normal * 1,
+		INSN_len_normal * 2,
+		INSN_len_normal * 4
 	},
 	/*
 	 *   cmp  rd_hi, rs_hi
@@ -2639,7 +2638,7 @@ const struct {
  */
 static inline s32 get_displacement(u32 curr_off, u32 targ_off)
 {
-	return (s32) (targ_off - (curr_off & ~3L));
+	return (s32)(targ_off - (curr_off & ~3L));
 }
 
 /*
@@ -2704,7 +2703,6 @@ static u8 gen_jset_64(u8 *buf, u8 rd, u8 rs, u32 curr_off, u32 targ_off)
 	return len;
 }
 
-
 /*
  * Verify if all the jumps for a JITed jcc64 operation are valid,
  * by consulting the data stored at "arcv2_64_jccs".
diff --git a/arch/arc/net/bpf_jit_core.c b/arch/arc/net/bpf_jit_core.c
index 6692272fa1ac..00c99b339b4a 100644
--- a/arch/arc/net/bpf_jit_core.c
+++ b/arch/arc/net/bpf_jit_core.c
@@ -5,7 +5,7 @@
  * Copyright (c) 2024 Synopsys Inc.
  * Author: Shahab Vahedi <shahab at synopsys.com>
  */
-#include <asm/bug.h>
+#include <linux/bug.h>
 #include "bpf_jit.h"
 
 /*
@@ -30,18 +30,18 @@ static void dump_bytes(const u8 *buf, u32 len, const char *header)
 
 	for (i = 0, j = 0; i < len; i++) {
 		/* Last input byte? */
-		if (i == len-1) {
-			j += scnprintf(line+j, 64-j, "0x%02x", buf[i]);
+		if (i == len - 1) {
+			j += scnprintf(line + j, 64 - j, "0x%02x", buf[i]);
 			pr_info("%s\n", line);
 			break;
 		}
 		/* End of line? */
 		else if (i % 8 == 7) {
-			j += scnprintf(line+j, 64-j, "0x%02x", buf[i]);
+			j += scnprintf(line + j, 64 - j, "0x%02x", buf[i]);
 			pr_info("%s\n", line);
 			j = 0;
 		} else {
-			j += scnprintf(line+j, 64-j, "0x%02x, ", buf[i]);
+			j += scnprintf(line + j, 64 - j, "0x%02x, ", buf[i]);
 		}
 	}
 }
@@ -126,7 +126,7 @@ static void vm_dump(const struct bpf_prog *prog)
 {
 #ifdef ARC_BPF_JIT_DEBUG
 	if (bpf_jit_enable > 1)
-		dump_bytes((u8 *) prog->insns, 8*prog->len, " VM  ");
+		dump_bytes((u8 *)prog->insns, 8 * prog->len, " VM  ");
 #endif
 }
 
@@ -222,8 +222,8 @@ static void jit_ctx_cleanup(struct jit_context *ctx)
 			bpf_jit_prog_release_other(ctx->orig_prog, ctx->prog);
 	}
 
-	maybe_free(ctx, (void **) &ctx->bpf2insn);
-	maybe_free(ctx, (void **) &ctx->jit_data);
+	maybe_free(ctx, (void **)&ctx->bpf2insn);
+	maybe_free(ctx, (void **)&ctx->jit_data);
 
 	if (!ctx->bpf2insn)
 		ctx->bpf2insn_valid = false;
@@ -267,8 +267,8 @@ static void analyze_reg_usage(struct jit_context *ctx)
 /* Verify that no instruction will be emitted when there is no buffer. */
 static inline int jit_buffer_check(const struct jit_context *ctx)
 {
-	if (ctx->emit == true) {
-		if (ctx->jit.buf == NULL) {
+	if (ctx->emit) {
+		if (!ctx->jit.buf) {
 			pr_err("bpf-jit: inconsistence state; no "
 			       "buffer to emit instructions.\n");
 			return -EINVAL;
@@ -333,7 +333,6 @@ static inline s32 get_index_for_insn(const struct jit_context *ctx,
 	return (insn - ctx->prog->insnsi);
 }
 
-
 /*
  * In most of the cases, the "offset" is read from "insn->off". However,
  * if it is an unconditional BPF_JMP32, then it comes from "insn->imm".
@@ -608,7 +607,7 @@ static int handle_jumps(const struct jit_context *ctx,
 	 *    (curr_off) will have increased to a point where the necessary
 	 *    instructions can be inserted by "gen_jmp_{32,64}()".
 	 */
-	if (has_imm(insn) && (cond != ARC_CC_AL)) {
+	if (has_imm(insn) && cond != ARC_CC_AL) {
 		if (j32) {
 			*len += mov_r32_i32(BUF(buf, *len), JIT_REG_TMP,
 					    insn->imm);
@@ -685,7 +684,7 @@ static int handle_call(struct jit_context *ctx,
 	if (!fixed && !addr)
 		set_need_for_extra_pass(ctx);
 
-	*len = gen_func_call(buf, (ARC_ADDR) addr, in_kernel_func);
+	*len = gen_func_call(buf, (ARC_ADDR)addr, in_kernel_func);
 
 	if (insn->src_reg != BPF_PSEUDO_CALL) {
 		/* Assigning ABI's return reg to JIT's return reg. */
@@ -714,7 +713,7 @@ static int handle_ld_imm64(struct jit_context *ctx,
 		return -EINVAL;
 	}
 
-	*len = mov_r64_i64(buf, insn->dst_reg, insn->imm, (insn+1)->imm);
+	*len = mov_r64_i64(buf, insn->dst_reg, insn->imm, (insn + 1)->imm);
 
 	if (bpf_pseudo_func(insn))
 		set_need_for_extra_pass(ctx);
@@ -841,7 +840,7 @@ static int handle_insn(struct jit_context *ctx, u32 idx)
 		break;
 	/* dst = src (32-bit) */
 	case BPF_ALU | BPF_MOV | BPF_X:
-		len = mov_r32(buf, dst, src, (u8) off);
+		len = mov_r32(buf, dst, src, (u8)off);
 		break;
 	/* dst = imm32 (32-bit) */
 	case BPF_ALU | BPF_MOV | BPF_K:
@@ -934,7 +933,7 @@ static int handle_insn(struct jit_context *ctx, u32 idx)
 		break;
 	/* dst = src (64-bit) */
 	case BPF_ALU64 | BPF_MOV | BPF_X:
-		len = mov_r64(buf, dst, src, (u8) off);
+		len = mov_r64(buf, dst, src, (u8)off);
 		break;
 	/* dst = imm32 (sign extend to 64-bit) */
 	case BPF_ALU64 | BPF_MOV | BPF_K:
@@ -1074,7 +1073,7 @@ static int handle_body(struct jit_context *ctx)
 		CHECK_RET(handle_insn(ctx, i));
 		if (ret > 0) {
 			/* "ret" is 1 if two (64-bit) chunks were consumed. */
-			ctx->bpf2insn[i+1] = ctx->bpf2insn[i];
+			ctx->bpf2insn[i + 1] = ctx->bpf2insn[i];
 			i++;
 		}
 	}
@@ -1103,7 +1102,7 @@ static void fill_ill_insn(void *area, unsigned int size)
 	const u16 unimp_s = 0x79e0;
 
 	if (size & 1) {
-		*((u8 *) area + (size - 1)) = 0xff;
+		*((u8 *)area + (size - 1)) = 0xff;
 		size -= 1;
 	}
 
@@ -1141,8 +1140,7 @@ static int jit_prepare_final_mem_alloc(struct jit_context *ctx)
 	}
 
 	if (ctx->need_extra_pass) {
-		ctx->jit_data = kzalloc(sizeof(struct arc_jit_data),
-					GFP_KERNEL);
+		ctx->jit_data = kzalloc(sizeof(*ctx->jit_data), GFP_KERNEL);
 		if (!ctx->jit_data)
 			return -ENOMEM;
 	}
@@ -1224,23 +1222,23 @@ static void jit_finalize(struct jit_context *ctx)
 {
 	struct bpf_prog *prog = ctx->prog;
 
-	ctx->success    = true;
-	prog->bpf_func  = (void *) ctx->jit.buf;
+	ctx->success = true;
+	prog->bpf_func = (void *)ctx->jit.buf;
 	prog->jited_len = ctx->jit.len;
-	prog->jited     = 1;
+	prog->jited = 1;
 
 	/* We're going to need this information for the "do_extra_pass()". */
 	if (ctx->need_extra_pass) {
 		ctx->jit_data->bpf_header = ctx->bpf_header;
-		ctx->jit_data->bpf2insn   = ctx->bpf2insn;
-		prog->aux->jit_data       = (void *) ctx->jit_data;
+		ctx->jit_data->bpf2insn = ctx->bpf2insn;
+		prog->aux->jit_data = (void *)ctx->jit_data;
 	} else {
 		/*
 		 * If things seem finalised, then mark the JITed memory
 		 * as R-X and flush it.
 		 */
 		bpf_jit_binary_lock_ro(ctx->bpf_header);
-		flush_icache_range((unsigned long) ctx->bpf_header,
+		flush_icache_range((unsigned long)ctx->bpf_header,
 				   (unsigned long)
 				   BUF(ctx->jit.buf, ctx->jit.len));
 		prog->aux->jit_data = NULL;
@@ -1258,30 +1256,31 @@ static void jit_finalize(struct jit_context *ctx)
  */
 static inline int check_jit_context(const struct bpf_prog *prog)
 {
-	if (prog->aux->jit_data == NULL) {
+	if (!prog->aux->jit_data) {
 		pr_notice("bpf-jit: no jit data for the extra pass.\n");
 		return 1;
-	} else
+	} else {
 		return 0;
+	}
 }
 
 /* Reuse the previous pass's data. */
 static int jit_resume_context(struct jit_context *ctx)
 {
 	struct arc_jit_data *jdata =
-		(struct arc_jit_data *) ctx->prog->aux->jit_data;
+		(struct arc_jit_data *)ctx->prog->aux->jit_data;
 
 	if (!jdata) {
 		pr_err("bpf-jit: no jit data for the extra pass.\n");
 		return -EINVAL;
 	}
 
-	ctx->jit.buf        = (u8 *) ctx->prog->bpf_func;
-	ctx->jit.len        = ctx->prog->jited_len;
-	ctx->bpf_header     = jdata->bpf_header;
-	ctx->bpf2insn       = (u32 *) jdata->bpf2insn;
+	ctx->jit.buf = (u8 *)ctx->prog->bpf_func;
+	ctx->jit.len = ctx->prog->jited_len;
+	ctx->bpf_header = jdata->bpf_header;
+	ctx->bpf2insn = (u32 *)jdata->bpf2insn;
 	ctx->bpf2insn_valid = ctx->bpf2insn ? true : false;
-	ctx->jit_data       = jdata;
+	ctx->jit_data = jdata;
 
 	return 0;
 }
-- 
2.35.8




More information about the linux-snps-arc mailing list