diff options
author | Weiwei Li <liweiwei@iscas.ac.cn> | 2022-04-12 17:07:18 +0800 |
---|---|---|
committer | Weiwei Li <liweiwei@iscas.ac.cn> | 2022-04-14 09:40:13 +0800 |
commit | c3c04a8be2c641de2b198b90df6c1538eb204120 (patch) | |
tree | 37019ee63b7e331a1c12a164bd7294017d6b2262 | |
parent | ae7d2c3062c1a911d97444e5c6df5089a7bb52f6 (diff) | |
download | riscv-isa-sim-c3c04a8be2c641de2b198b90df6c1538eb204120.zip riscv-isa-sim-c3c04a8be2c641de2b198b90df6c1538eb204120.tar.gz riscv-isa-sim-c3c04a8be2c641de2b198b90df6c1538eb204120.tar.bz2 |
fix style problems in decode.h and processor.cc
-rw-r--r-- | riscv/decode.h | 532 | ||||
-rw-r--r-- | riscv/processor.cc | 57 |
2 files changed, 296 insertions, 293 deletions
diff --git a/riscv/decode.h b/riscv/decode.h index e0d70bf..611c910 100644 --- a/riscv/decode.h +++ b/riscv/decode.h @@ -83,7 +83,7 @@ public: int64_t i_imm() { return int64_t(b) >> 20; } int64_t shamt() { return x(20, 6); } int64_t s_imm() { return x(7, 5) + (xs(25, 7) << 5); } - int64_t sb_imm() { return (x(8, 4) << 1) + (x(25,6) << 5) + (x(7,1) << 11) + (imm_sign() << 12); } + int64_t sb_imm() { return (x(8, 4) << 1) + (x(25, 6) << 5) + (x(7, 1) << 11) + (imm_sign() << 12); } int64_t u_imm() { return int64_t(b) >> 12 << 12; } int64_t uj_imm() { return (x(21, 10) << 1) + (x(20, 1) << 11) + (x(12, 8) << 12) + (imm_sign() << 20); } uint64_t rd() { return x(7, 5); } @@ -93,8 +93,8 @@ public: uint64_t rm() { return x(12, 3); } uint64_t csr() { return x(20, 12); } uint64_t iorw() { return x(20, 8); } - uint64_t bs () {return x(30,2);} // Crypto ISE - SM4/AES32 byte select. - uint64_t rcon() {return x(20,4);} // Crypto ISE - AES64 round const. + uint64_t bs() { return x(30, 2); } // Crypto ISE - SM4/AES32 byte select. + uint64_t rcon() { return x(20, 4); } // Crypto ISE - AES64 round const. int64_t rvc_imm() { return x(2, 5) + (xs(12, 1) << 5); } int64_t rvc_zimm() { return x(2, 5) + (x(12, 1) << 5); } @@ -141,8 +141,8 @@ public: private: insn_bits_t b; - uint64_t x(int lo, int len) { return (b >> lo) & ((insn_bits_t(1) << len)-1); } - uint64_t xs(int lo, int len) { return int64_t(b) << (64-lo-len) >> (64-len); } + uint64_t x(int lo, int len) { return (b >> lo) & ((insn_bits_t(1) << len) - 1); } + uint64_t xs(int lo, int len) { return int64_t(b) << (64 - lo - len) >> (64 - len); } uint64_t imm_sign() { return xs(63, 1); } }; @@ -235,8 +235,8 @@ private: #define BRANCH_TARGET (pc + insn.sb_imm()) #define JUMP_TARGET (pc + insn.uj_imm()) #define RM ({ int rm = insn.rm(); \ - if(rm == 7) rm = STATE.frm->read(); \ - if(rm > 4) throw trap_illegal_instruction(insn.bits()); \ + if (rm == 7) rm = STATE.frm->read(); \ + if (rm > 4) throw trap_illegal_instruction(insn.bits()); \ rm; }) #define get_field(reg, mask) (((reg) & (decltype(reg))(mask)) / ((mask) & ~((mask) << 1))) @@ -258,16 +258,16 @@ private: require_extension('V'); \ require(!P.VU.vill); \ if (alu && !P.VU.vstart_alu) \ - require(P.VU.vstart->read() == 0); \ + require(P.VU.vstart->read() == 0); \ WRITE_VSTATUS; \ dirty_vs_state; \ } while (0); #define require_vector_novtype(is_log, alu) \ - do { \ + do { \ require_vector_vs; \ require_extension('V'); \ if (alu && !P.VU.vstart_alu) \ - require(P.VU.vstart->read() == 0); \ + require(P.VU.vstart->read() == 0); \ if (is_log) \ WRITE_VSTATUS; \ dirty_vs_state; \ @@ -277,7 +277,7 @@ private: require(!is_overlapped(astart, asize, bstart, bsize)) #define require_noover_widen(astart, asize, bstart, bsize) \ require(!is_overlapped_widen(astart, asize, bstart, bsize)) -#define require_vm do { if (insn.v_vm() == 0) require(insn.rd() != 0);} while(0); +#define require_vm do { if (insn.v_vm() == 0) require(insn.rd() != 0); } while (0); #define require_envcfg(field) \ do { \ if (((STATE.prv != PRV_M) && (m##field == 0)) || \ @@ -286,7 +286,7 @@ private: else if (STATE.v && ((h##field == 0) || \ ((STATE.prv == PRV_U) && (s##field == 0)))) \ throw trap_virtual_instruction(insn.bits()); \ - } while(0); + } while (0); #define set_fp_exceptions ({ if (softfloat_exceptionFlags) { \ STATE.fflags->write(STATE.fflags->read() | softfloat_exceptionFlags); \ @@ -295,20 +295,20 @@ private: #define sext32(x) ((sreg_t)(int32_t)(x)) #define zext32(x) ((reg_t)(uint32_t)(x)) -#define sext_xlen(x) (((sreg_t)(x) << (64-xlen)) >> (64-xlen)) -#define zext(x, pos) (((reg_t)(x) << (64-(pos))) >> (64-(pos))) +#define sext_xlen(x) (((sreg_t)(x) << (64 - xlen)) >> (64 - xlen)) +#define zext(x, pos) (((reg_t)(x) << (64 - (pos))) >> (64 - (pos))) #define zext_xlen(x) zext(x, xlen) #define set_pc(x) \ do { p->check_pc_alignment(x); \ npc = sext_xlen(x); \ - } while(0) + } while (0) #define set_pc_and_serialize(x) \ do { reg_t __npc = (x) & p->pc_alignment_mask(); \ npc = PC_SERIALIZE_AFTER; \ STATE.pc = __npc; \ - } while(0) + } while (0) class wait_for_interrupt_t {}; @@ -316,7 +316,7 @@ class wait_for_interrupt_t {}; do { set_pc_and_serialize(npc); \ npc = PC_SERIALIZE_WFI; \ throw wait_for_interrupt_t(); \ - } while(0) + } while (0) #define serialize() set_pc_and_serialize(npc) @@ -381,9 +381,9 @@ inline freg_t f128_negate(freg_t a) (which); }) /* For debug only. This will fail if the native machine's float types are not IEEE */ -inline float to_f(float32_t f){float r; memcpy(&r, &f, sizeof(r)); return r;} -inline double to_f(float64_t f){double r; memcpy(&r, &f, sizeof(r)); return r;} -inline long double to_f(float128_t f){long double r; memcpy(&r, &f, sizeof(r)); return r;} +inline float to_f(float32_t f) { float r; memcpy(&r, &f, sizeof(r)); return r; } +inline double to_f(float64_t f) { double r; memcpy(&r, &f, sizeof(r)); return r; } +inline long double to_f(float128_t f) { long double r; memcpy(&r, &f, sizeof(r)); return r; } // Vector macros #define e8 8 // 8b elements @@ -395,8 +395,8 @@ inline long double to_f(float128_t f){long double r; memcpy(&r, &f, sizeof(r)); #define e512 512 // 512b elements #define e1024 1024 // 1024b elements -#define vsext(x, sew) (((sreg_t)(x) << (64-sew)) >> (64-sew)) -#define vzext(x, sew) (((reg_t)(x) << (64-sew)) >> (64-sew)) +#define vsext(x, sew) (((sreg_t)(x) << (64 - sew)) >> (64 - sew)) +#define vzext(x, sew) (((reg_t)(x) << (64 - sew)) >> (64 - sew)) #define DEBUG_RVV 0 @@ -428,15 +428,15 @@ inline long double to_f(float128_t f){long double r; memcpy(&r, &f, sizeof(r)); if (insn.v_vm() == 0) { \ BODY; \ bool skip = ((P.VU.elt<uint64_t>(0, midx) >> mpos) & 0x1) == 0; \ - if (skip) {\ + if (skip) { \ continue; \ - }\ + } \ } #define VI_ELEMENT_SKIP(inx) \ if (inx >= vl) { \ continue; \ - } else if (inx < P.VU.vstart->read()) { \ + } else if (inx < P.VU.vstart->read()) { \ continue; \ } else { \ VI_LOOP_ELEMENT_SKIP(); \ @@ -446,7 +446,7 @@ inline long double to_f(float128_t f){long double r; memcpy(&r, &f, sizeof(r)); // vector: operation and register acccess check helper // static inline bool is_overlapped(const int astart, int asize, - const int bstart, int bsize) + const int bstart, int bsize) { asize = asize == 0 ? 1 : asize; bsize = bsize == 0 ? 1 : bsize; @@ -481,7 +481,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) } #define VI_NARROW_CHECK_COMMON \ - require_vector(true);\ + require_vector(true); \ require(P.VU.vflmul <= 4); \ require(P.VU.vsew * 2 <= P.VU.ELEN); \ require_align(insn.rs2(), P.VU.vflmul * 2); \ @@ -489,7 +489,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) require_vm; \ #define VI_WIDE_CHECK_COMMON \ - require_vector(true);\ + require_vector(true); \ require(P.VU.vflmul <= 4); \ require(P.VU.vsew * 2 <= P.VU.ELEN); \ require_align(insn.rd(), P.VU.vflmul * 2); \ @@ -510,14 +510,14 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) VI_CHECK_ST_INDEX(elt_width); \ for (reg_t idx = 0; idx < nf; ++idx) { \ reg_t flmul = P.VU.vflmul < 1 ? 1 : P.VU.vflmul; \ - reg_t seg_vd = insn.rd() + flmul * idx; \ + reg_t seg_vd = insn.rd() + flmul * idx; \ if (elt_width > P.VU.vsew) { \ if (seg_vd != insn.rs2()) \ require_noover(seg_vd, P.VU.vflmul, insn.rs2(), vemul); \ } else if (elt_width < P.VU.vsew) { \ - if (vemul < 1) {\ + if (vemul < 1) { \ require_noover(seg_vd, P.VU.vflmul, insn.rs2(), vemul); \ - } else {\ + } else { \ require_noover_widen(seg_vd, P.VU.vflmul, insn.rs2(), vemul); \ } \ } \ @@ -531,7 +531,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) if (insn.rd() != insn.rs2()) \ require_noover(insn.rd(), 1, insn.rs2(), P.VU.vflmul); \ require_align(insn.rs2(), P.VU.vflmul); \ - if (is_vs1) {\ + if (is_vs1) { \ if (insn.rd() != insn.rs1()) \ require_noover(insn.rd(), 1, insn.rs1(), P.VU.vflmul); \ require_align(insn.rs1(), P.VU.vflmul); \ @@ -565,16 +565,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_CHECK_DSS(is_vs1) \ VI_WIDE_CHECK_COMMON; \ require_align(insn.rs2(), P.VU.vflmul); \ - if (P.VU.vflmul < 1) {\ + if (P.VU.vflmul < 1) { \ require_noover(insn.rd(), P.VU.vflmul * 2, insn.rs2(), P.VU.vflmul); \ - } else {\ + } else { \ require_noover_widen(insn.rd(), P.VU.vflmul * 2, insn.rs2(), P.VU.vflmul); \ } \ - if (is_vs1) {\ + if (is_vs1) { \ require_align(insn.rs1(), P.VU.vflmul); \ - if (P.VU.vflmul < 1) {\ + if (P.VU.vflmul < 1) { \ require_noover(insn.rd(), P.VU.vflmul * 2, insn.rs1(), P.VU.vflmul); \ - } else {\ + } else { \ require_noover_widen(insn.rd(), P.VU.vflmul * 2, insn.rs1(), P.VU.vflmul); \ } \ } @@ -584,9 +584,9 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) require_align(insn.rs2(), P.VU.vflmul * 2); \ if (is_rs) { \ require_align(insn.rs1(), P.VU.vflmul); \ - if (P.VU.vflmul < 1) {\ + if (P.VU.vflmul < 1) { \ require_noover(insn.rd(), P.VU.vflmul * 2, insn.rs1(), P.VU.vflmul); \ - } else {\ + } else { \ require_noover_widen(insn.rd(), P.VU.vflmul * 2, insn.rs1(), P.VU.vflmul); \ } \ } @@ -599,12 +599,12 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) require_align(insn.rs1(), P.VU.vflmul); \ #define VI_CHECK_REDUCTION(is_wide) \ - require_vector(true);\ - if (is_wide) {\ + require_vector(true); \ + if (is_wide) { \ require(P.VU.vsew * 2 <= P.VU.ELEN); \ } \ require_align(insn.rs2(), P.VU.vflmul); \ - require(P.VU.vstart->read() == 0); \ + require(P.VU.vstart->read() == 0); \ #define VI_CHECK_SLIDE(is_over) \ require_align(insn.rs2(), P.VU.vflmul); \ @@ -619,13 +619,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) // #define VI_GENERAL_LOOP_BASE \ require(P.VU.vsew >= e8 && P.VU.vsew <= e64); \ - require_vector(true);\ - reg_t vl = P.VU.vl->read(); \ + require_vector(true); \ + reg_t vl = P.VU.vl->read(); \ reg_t sew = P.VU.vsew; \ reg_t rd_num = insn.rd(); \ reg_t rs1_num = insn.rs1(); \ reg_t rs2_num = insn.rs2(); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { #define VI_LOOP_BASE \ VI_GENERAL_LOOP_BASE \ @@ -665,13 +665,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_LOOP_CMP_BASE \ require(P.VU.vsew >= e8 && P.VU.vsew <= e64); \ - require_vector(true);\ - reg_t vl = P.VU.vl->read(); \ + require_vector(true); \ + reg_t vl = P.VU.vl->read(); \ reg_t sew = P.VU.vsew; \ reg_t rd_num = insn.rd(); \ reg_t rs1_num = insn.rs1(); \ reg_t rs2_num = insn.rs2(); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); \ uint64_t mmask = UINT64_C(1) << mpos; \ uint64_t &vdi = P.VU.elt<uint64_t>(insn.rd(), midx, true); \ @@ -684,8 +684,8 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_LOOP_MASK(op) \ require(P.VU.vsew <= e64); \ - require_vector(true);\ - reg_t vl = P.VU.vl->read(); \ + require_vector(true); \ + reg_t vl = P.VU.vl->read(); \ for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ int midx = i / 64; \ int mpos = i % 64; \ @@ -699,8 +699,8 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_LOOP_NSHIFT_BASE \ VI_GENERAL_LOOP_BASE; \ - VI_LOOP_ELEMENT_SKIP({\ - require(!(insn.rd() == 0 && P.VU.vflmul > 1));\ + VI_LOOP_ELEMENT_SKIP({ \ + require(!(insn.rd() == 0 && P.VU.vflmul > 1)); \ }); @@ -708,22 +708,22 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) do { \ const uint64_t lsb = 1UL << (gb); \ const uint64_t lsb_half = lsb >> 1; \ - switch (xrm) {\ - case VRM::RNU:\ + switch (xrm) { \ + case VRM::RNU: \ result += lsb_half; \ - break;\ - case VRM::RNE:\ + break; \ + case VRM::RNE: \ if ((result & lsb_half) && ((result & (lsb_half - 1)) || (result & lsb))) \ result += lsb; \ - break;\ - case VRM::RDN:\ - break;\ - case VRM::ROD:\ + break; \ + case VRM::RDN: \ + break; \ + case VRM::ROD: \ if (result & (lsb - 1)) \ result |= lsb; \ - break;\ - case VRM::INVALID_RM:\ - assert(true);\ + break; \ + case VRM::INVALID_RM: \ + assert(true); \ } \ } while (0) @@ -881,16 +881,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) // #define INSNS_BASE(PARAMS, BODY) \ - if (sew == e8){ \ + if (sew == e8) { \ PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ PARAMS(e64); \ BODY; \ } @@ -938,16 +938,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_MERGE_LOOP(BODY) \ VI_CHECK_SSS(true); \ VI_MERGE_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VV_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VV_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VV_PARAMS(e64); \ BODY; \ } \ @@ -956,16 +956,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VX_MERGE_LOOP(BODY) \ VI_CHECK_SSS(false); \ VI_MERGE_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VX_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VX_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VX_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VX_PARAMS(e64); \ BODY; \ } \ @@ -974,16 +974,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VI_MERGE_LOOP(BODY) \ VI_CHECK_SSS(false); \ VI_MERGE_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VI_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VI_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VI_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VI_PARAMS(e64); \ BODY; \ } \ @@ -992,15 +992,15 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VF_MERGE_LOOP(BODY) \ VI_CHECK_SSS(false); \ VI_VFP_COMMON \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_MERGE_VARS \ - if(P.VU.vsew == e16){ \ + if (P.VU.vsew == e16) { \ VFP_VF_PARAMS(16); \ BODY; \ - }else if(P.VU.vsew == e32){ \ + } else if (P.VU.vsew == e32) { \ VFP_VF_PARAMS(32); \ BODY; \ - }else if(P.VU.vsew == e64){ \ + } else if (P.VU.vsew == e64) { \ VFP_VF_PARAMS(64); \ BODY; \ } \ @@ -1009,13 +1009,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) // reduction loop - signed #define VI_LOOP_REDUCTION_BASE(x) \ require(x >= e8 && x <= e64); \ - reg_t vl = P.VU.vl->read(); \ + reg_t vl = P.VU.vl->read(); \ reg_t rd_num = insn.rd(); \ reg_t rs1_num = insn.rs1(); \ reg_t rs2_num = insn.rs2(); \ auto &vd_0_des = P.VU.elt<type_sew_t<x>::type>(rd_num, 0, true); \ auto vd_0_res = P.VU.elt<type_sew_t<x>::type>(rs1_num, 0); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); \ auto vs2 = P.VU.elt<type_sew_t<x>::type>(rs2_num, i); \ @@ -1029,24 +1029,24 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) reg_t sew = P.VU.vsew; \ if (sew == e8) { \ REDUCTION_LOOP(e8, BODY) \ - } else if(sew == e16) { \ + } else if (sew == e16) { \ REDUCTION_LOOP(e16, BODY) \ - } else if(sew == e32) { \ + } else if (sew == e32) { \ REDUCTION_LOOP(e32, BODY) \ - } else if(sew == e64) { \ + } else if (sew == e64) { \ REDUCTION_LOOP(e64, BODY) \ } // reduction loop - unsigned #define VI_ULOOP_REDUCTION_BASE(x) \ require(x >= e8 && x <= e64); \ - reg_t vl = P.VU.vl->read(); \ + reg_t vl = P.VU.vl->read(); \ reg_t rd_num = insn.rd(); \ reg_t rs1_num = insn.rs1(); \ reg_t rs2_num = insn.rs2(); \ auto &vd_0_des = P.VU.elt<type_usew_t<x>::type>(rd_num, 0, true); \ auto vd_0_res = P.VU.elt<type_usew_t<x>::type>(rs1_num, 0); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); \ auto vs2 = P.VU.elt<type_usew_t<x>::type>(rs2_num, i); @@ -1058,13 +1058,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_ULOOP_REDUCTION(BODY) \ VI_CHECK_REDUCTION(false); \ reg_t sew = P.VU.vsew; \ - if (sew == e8){ \ + if (sew == e8) { \ REDUCTION_ULOOP(e8, BODY) \ - } else if(sew == e16) { \ + } else if (sew == e16) { \ REDUCTION_ULOOP(e16, BODY) \ - } else if(sew == e32) { \ + } else if (sew == e32) { \ REDUCTION_ULOOP(e32, BODY) \ - } else if(sew == e64) { \ + } else if (sew == e64) { \ REDUCTION_ULOOP(e64, BODY) \ } @@ -1073,16 +1073,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_ULOOP(BODY) \ VI_CHECK_SSS(true) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_U_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VV_U_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VV_U_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VV_U_PARAMS(e64); \ BODY; \ } \ @@ -1091,16 +1091,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_LOOP(BODY) \ VI_CHECK_SSS(true) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VV_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VV_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VV_PARAMS(e64); \ BODY; \ } \ @@ -1109,16 +1109,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VX_ULOOP(BODY) \ VI_CHECK_SSS(false) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VX_U_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VX_U_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VX_U_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VX_U_PARAMS(e64); \ BODY; \ } \ @@ -1127,16 +1127,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VX_LOOP(BODY) \ VI_CHECK_SSS(false) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VX_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VX_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VX_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VX_PARAMS(e64); \ BODY; \ } \ @@ -1145,16 +1145,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VI_ULOOP(BODY) \ VI_CHECK_SSS(false) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VI_U_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VI_U_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VI_U_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VI_U_PARAMS(e64); \ BODY; \ } \ @@ -1163,16 +1163,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VI_LOOP(BODY) \ VI_CHECK_SSS(false) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VI_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VI_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VI_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VI_PARAMS(e64); \ BODY; \ } \ @@ -1182,16 +1182,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_SU_LOOP(BODY) \ VI_CHECK_SSS(true) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_SU_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VV_SU_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VV_SU_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VV_SU_PARAMS(e64); \ BODY; \ } \ @@ -1200,16 +1200,16 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VX_SU_LOOP(BODY) \ VI_CHECK_SSS(false) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VX_SU_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VX_SU_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VX_SU_PARAMS(e32); \ BODY; \ - }else if(sew == e64){ \ + } else if (sew == e64) { \ VX_SU_PARAMS(e64); \ BODY; \ } \ @@ -1219,13 +1219,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_LOOP_NARROW(BODY) \ VI_CHECK_SDS(true); \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_NARROW_PARAMS(e8, e16) \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VV_NARROW_PARAMS(e16, e32) \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VV_NARROW_PARAMS(e32, e64) \ BODY; \ } \ @@ -1234,13 +1234,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VX_LOOP_NARROW(BODY) \ VI_CHECK_SDS(false); \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VX_NARROW_PARAMS(e8, e16) \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VX_NARROW_PARAMS(e16, e32) \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VX_NARROW_PARAMS(e32, e64) \ BODY; \ } \ @@ -1249,13 +1249,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VI_LOOP_NARROW(BODY) \ VI_CHECK_SDS(false); \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VI_NARROW_PARAMS(e8, e16) \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VI_NARROW_PARAMS(e16, e32) \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VI_NARROW_PARAMS(e32, e64) \ BODY; \ } \ @@ -1264,7 +1264,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VI_LOOP_NSHIFT(BODY) \ VI_CHECK_SDS(false); \ VI_LOOP_NSHIFT_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VI_NARROW_PARAMS(e8, e16) \ BODY; \ } else if (sew == e16) { \ @@ -1279,7 +1279,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VX_LOOP_NSHIFT(BODY) \ VI_CHECK_SDS(false); \ VI_LOOP_NSHIFT_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VX_NARROW_PARAMS(e8, e16) \ BODY; \ } else if (sew == e16) { \ @@ -1294,7 +1294,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_LOOP_NSHIFT(BODY) \ VI_CHECK_SDS(true); \ VI_LOOP_NSHIFT_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_NARROW_PARAMS(e8, e16) \ BODY; \ } else if (sew == e16) { \ @@ -1309,13 +1309,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) // widen operation loop #define VI_VV_LOOP_WIDEN(BODY) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VV_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VV_PARAMS(e32); \ BODY; \ } \ @@ -1323,20 +1323,20 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VX_LOOP_WIDEN(BODY) \ VI_LOOP_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VX_PARAMS(e8); \ BODY; \ - }else if(sew == e16){ \ + } else if (sew == e16) { \ VX_PARAMS(e16); \ BODY; \ - }else if(sew == e32){ \ + } else if (sew == e32) { \ VX_PARAMS(e32); \ BODY; \ } \ VI_LOOP_END #define VI_WIDE_OP_AND_ASSIGN(var0, var1, var2, op0, op1, sign) \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e8: { \ sign##16_t vd_w = P.VU.elt<sign##16_t>(rd_num, i); \ P.VU.elt<uint16_t>(rd_num, i, true) = \ @@ -1358,7 +1358,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) } #define VI_WIDE_OP_AND_ASSIGN_MIX(var0, var1, var2, op0, op1, sign_d, sign_1, sign_2) \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e8: { \ sign_d##16_t vd_w = P.VU.elt<sign_d##16_t>(rd_num, i); \ P.VU.elt<uint16_t>(rd_num, i, true) = \ @@ -1380,7 +1380,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) } #define VI_WIDE_WVX_OP(var0, op0, sign) \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e8: { \ sign##16_t &vd_w = P.VU.elt<sign##16_t>(rd_num, i, true); \ sign##16_t vs2_w = P.VU.elt<sign##16_t>(rs2_num, i); \ @@ -1403,13 +1403,13 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) // wide reduction loop - signed #define VI_LOOP_WIDE_REDUCTION_BASE(sew1, sew2) \ - reg_t vl = P.VU.vl->read(); \ + reg_t vl = P.VU.vl->read(); \ reg_t rd_num = insn.rd(); \ reg_t rs1_num = insn.rs1(); \ reg_t rs2_num = insn.rs2(); \ auto &vd_0_des = P.VU.elt<type_sew_t<sew2>::type>(rd_num, 0, true); \ auto vd_0_res = P.VU.elt<type_sew_t<sew2>::type>(rs1_num, 0); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); \ auto vs2 = P.VU.elt<type_sew_t<sew1>::type>(rs2_num, i); @@ -1421,23 +1421,23 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_LOOP_WIDE_REDUCTION(BODY) \ VI_CHECK_REDUCTION(true); \ reg_t sew = P.VU.vsew; \ - if (sew == e8){ \ + if (sew == e8) { \ WIDE_REDUCTION_LOOP(e8, e16, BODY) \ - } else if(sew == e16){ \ + } else if (sew == e16) { \ WIDE_REDUCTION_LOOP(e16, e32, BODY) \ - } else if(sew == e32){ \ + } else if (sew == e32) { \ WIDE_REDUCTION_LOOP(e32, e64, BODY) \ } // wide reduction loop - unsigned #define VI_ULOOP_WIDE_REDUCTION_BASE(sew1, sew2) \ - reg_t vl = P.VU.vl->read(); \ + reg_t vl = P.VU.vl->read(); \ reg_t rd_num = insn.rd(); \ reg_t rs1_num = insn.rs1(); \ reg_t rs2_num = insn.rs2(); \ auto &vd_0_des = P.VU.elt<type_usew_t<sew2>::type>(rd_num, 0, true); \ auto vd_0_res = P.VU.elt<type_usew_t<sew2>::type>(rs1_num, 0); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i) { \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); \ auto vs2 = P.VU.elt<type_usew_t<sew1>::type>(rs2_num, i); @@ -1449,11 +1449,11 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_ULOOP_WIDE_REDUCTION(BODY) \ VI_CHECK_REDUCTION(true); \ reg_t sew = P.VU.vsew; \ - if (sew == e8){ \ + if (sew == e8) { \ WIDE_REDUCTION_ULOOP(e8, e16, BODY) \ - } else if(sew == e16){ \ + } else if (sew == e16) { \ WIDE_REDUCTION_ULOOP(e16, e32, BODY) \ - } else if(sew == e32){ \ + } else if (sew == e32) { \ WIDE_REDUCTION_ULOOP(e32, e64, BODY) \ } @@ -1461,7 +1461,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_VV_LOOP_CARRY(BODY) \ VI_CHECK_MSS(true); \ VI_LOOP_CARRY_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_CARRY_PARAMS(e8) \ BODY; \ } else if (sew == e16) { \ @@ -1479,7 +1479,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) #define VI_XI_LOOP_CARRY(BODY) \ VI_CHECK_MSS(false); \ VI_LOOP_CARRY_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ XI_CARRY_PARAMS(e8) \ BODY; \ } else if (sew == e16) { \ @@ -1498,7 +1498,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) require_vm; \ VI_CHECK_SSS(true); \ VI_LOOP_WITH_CARRY_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ VV_WITH_CARRY_PARAMS(e8) \ BODY; \ } else if (sew == e16) { \ @@ -1517,7 +1517,7 @@ static inline bool is_aligned(const unsigned val, const unsigned pos) require_vm; \ VI_CHECK_SSS(false); \ VI_LOOP_WITH_CARRY_BASE \ - if (sew == e8){ \ + if (sew == e8) { \ XI_WITH_CARRY_PARAMS(e8) \ BODY; \ } else if (sew == e16) { \ @@ -1573,8 +1573,8 @@ VI_VX_ULOOP({ \ #define VI_DUPLICATE_VREG(reg_num, idx_sew) \ reg_t index[P.VU.vlmax]; \ - for (reg_t i = 0; i < P.VU.vlmax && P.VU.vl->read() != 0; ++i) { \ - switch(idx_sew) { \ + for (reg_t i = 0; i < P.VU.vlmax && P.VU.vl->read() != 0; ++i) { \ + switch (idx_sew) { \ case e8: \ index[i] = P.VU.elt<uint8_t>(reg_num, i); \ break; \ @@ -1622,7 +1622,7 @@ reg_t index[P.VU.vlmax]; \ VI_STRIP(i); \ P.VU.vstart->write(i); \ for (reg_t fn = 0; fn < nf; ++fn) { \ - switch(P.VU.vsew){ \ + switch (P.VU.vsew) { \ case e8: \ P.VU.elt<uint8_t>(vd + fn * flmul, vreg_inx, true) = \ MMU.load_uint8(baseAddr + index[i] + fn * 1); \ @@ -1670,7 +1670,7 @@ reg_t index[P.VU.vlmax]; \ if (!is_seg) \ require(nf == 1); \ VI_CHECK_ST_INDEX(elt_width); \ - VI_DUPLICATE_VREG(insn.rs2(), elt_width); \ + VI_DUPLICATE_VREG(insn.rs2(), elt_width); \ for (reg_t i = 0; i < vl; ++i) { \ VI_STRIP(i) \ VI_ELEMENT_SKIP(i); \ @@ -1720,7 +1720,7 @@ reg_t index[P.VU.vlmax]; \ throw; /* Only take exception on zeroth element */ \ /* Reduce VL if an exception occurs on a later element */ \ early_stop = true; \ - P.VU.vl->write_raw(i); \ + P.VU.vl->write_raw(i); \ break; \ } \ p->VU.elt<elt_width##_t>(rd_num + fn * emul, vreg_inx, true) = val; \ @@ -1771,7 +1771,7 @@ reg_t index[P.VU.vlmax]; \ const reg_t len = insn.v_nf() + 1; \ require_align(vs3, len); \ const reg_t size = len * P.VU.vlenb; \ - \ + \ if (P.VU.vstart->read() < size) { \ reg_t i = P.VU.vstart->read() / P.VU.vlenb; \ reg_t off = P.VU.vstart->read() % P.VU.vlenb; \ @@ -1804,15 +1804,15 @@ reg_t index[P.VU.vlmax]; \ float vemul = ((float)idx_type / P.VU.vsew * P.VU.vflmul); \ require(vemul >= 0.125 && vemul <= 8); \ require_align(insn.rs2(), vemul); \ - if (insn.v_wd()) {\ + if (insn.v_wd()) { \ require_vm; \ if (idx_type > P.VU.vsew) { \ if (insn.rd() != insn.rs2()) \ require_noover(insn.rd(), P.VU.vflmul, insn.rs2(), vemul); \ } else if (idx_type < P.VU.vsew) { \ - if (vemul < 1) {\ + if (vemul < 1) { \ require_noover(insn.rd(), P.VU.vflmul, insn.rs2(), vemul); \ - } else {\ + } else { \ require_noover_widen(insn.rd(), P.VU.vflmul, insn.rs2(), vemul); \ } \ } \ @@ -1826,16 +1826,16 @@ reg_t index[P.VU.vlmax]; \ VI_STRIP(i); \ P.VU.vstart->write(i); \ switch (P.VU.vsew) { \ - case e32: {\ + case e32: { \ auto vs3 = P.VU.elt< type ## 32_t>(vd, vreg_inx); \ - auto val = MMU.amo_uint32(baseAddr + index[i], [&]( type ## 32_t lhs) { op }); \ + auto val = MMU.amo_uint32(baseAddr + index[i], [&](type ## 32_t lhs) { op }); \ if (insn.v_wd()) \ P.VU.elt< type ## 32_t>(vd, vreg_inx, true) = val; \ } \ break; \ - case e64: {\ + case e64: { \ auto vs3 = P.VU.elt< type ## 64_t>(vd, vreg_inx); \ - auto val = MMU.amo_uint64(baseAddr + index[i], [&]( type ## 64_t lhs) { op }); \ + auto val = MMU.amo_uint64(baseAddr + index[i], [&](type ## 64_t lhs) { op }); \ if (insn.v_wd()) \ P.VU.elt< type ## 64_t>(vd, vreg_inx, true) = val; \ } \ @@ -1858,7 +1858,7 @@ reg_t index[P.VU.vlmax]; \ require_align(insn.rs2(), P.VU.vflmul / div); \ if ((P.VU.vflmul / div) < 1) { \ require_noover(insn.rd(), P.VU.vflmul, insn.rs2(), P.VU.vflmul / div); \ - } else {\ + } else { \ require_noover_widen(insn.rd(), P.VU.vflmul, insn.rs2(), P.VU.vflmul / div); \ } \ reg_t pat = (((P.VU.vsew >> 3) << 4) | from >> 3); \ @@ -1899,8 +1899,8 @@ reg_t index[P.VU.vlmax]; \ require((P.VU.vsew == e16 && p->extension_enabled(EXT_ZFH)) || \ (P.VU.vsew == e32 && p->extension_enabled('F')) || \ (P.VU.vsew == e64 && p->extension_enabled('D'))); \ - require_vector(true);\ - require(STATE.frm->read() < 0x5);\ + require_vector(true); \ + require(STATE.frm->read() < 0x5); \ reg_t vl = P.VU.vl->read(); \ reg_t rd_num = insn.rd(); \ reg_t rs1_num = insn.rs1(); \ @@ -1909,7 +1909,7 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_LOOP_BASE \ VI_VFP_COMMON \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); #define VI_VFP_LOOP_CMP_BASE \ @@ -1925,7 +1925,7 @@ reg_t index[P.VU.vlmax]; \ float##width##_t vs1_0 = P.VU.elt<float##width##_t>(rs1_num, 0); \ vd_0 = vs1_0; \ bool is_active = false; \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); \ float##width##_t vs2 = P.VU.elt<float##width##_t>(rs2_num, i); \ is_active = true; \ @@ -1933,7 +1933,7 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_LOOP_WIDE_REDUCTION_BASE \ VI_VFP_COMMON \ float64_t vd_0 = f64(P.VU.elt<float64_t>(rs1_num, 0).v); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i) { \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); #define VI_VFP_LOOP_END \ @@ -1946,7 +1946,7 @@ reg_t index[P.VU.vlmax]; \ if (vl > 0) { \ if (is_propagate && !is_active) { \ switch (x) { \ - case e16: {\ + case e16: { \ auto ret = f16_classify(f16(vd_0.v)); \ if (ret & 0x300) { \ if (ret & 0x100) { \ @@ -1972,7 +1972,7 @@ reg_t index[P.VU.vlmax]; \ } \ } \ break; \ - case e64: {\ + case e64: { \ auto ret = f64_classify(f64(vd_0.v)); \ if (ret & 0x300) { \ if (ret & 0x100) { \ @@ -1992,7 +1992,7 @@ reg_t index[P.VU.vlmax]; \ } #define VI_VFP_LOOP_CMP_END \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e16: \ case e32: \ case e64: { \ @@ -2009,25 +2009,25 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_VV_LOOP(BODY16, BODY32, BODY64) \ VI_CHECK_SSS(true); \ VI_VFP_LOOP_BASE \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ VFP_VV_PARAMS(16); \ BODY16; \ set_fp_exceptions; \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ VFP_VV_PARAMS(32); \ BODY32; \ set_fp_exceptions; \ break; \ - }\ - case e64: {\ + } \ + case e64: { \ VFP_VV_PARAMS(64); \ BODY64; \ set_fp_exceptions; \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2038,22 +2038,22 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_V_LOOP(BODY16, BODY32, BODY64) \ VI_CHECK_SSS(false); \ VI_VFP_LOOP_BASE \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ VFP_V_PARAMS(16); \ BODY16; \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ VFP_V_PARAMS(32); \ BODY32; \ break; \ - }\ - case e64: {\ + } \ + case e64: { \ VFP_V_PARAMS(64); \ BODY64; \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2064,28 +2064,28 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_VV_LOOP_REDUCTION(BODY16, BODY32, BODY64) \ VI_CHECK_REDUCTION(false) \ VI_VFP_COMMON \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ VI_VFP_LOOP_REDUCTION_BASE(16) \ BODY16; \ set_fp_exceptions; \ VI_VFP_LOOP_REDUCTION_END(e16) \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ VI_VFP_LOOP_REDUCTION_BASE(32) \ BODY32; \ set_fp_exceptions; \ VI_VFP_LOOP_REDUCTION_END(e32) \ break; \ - }\ - case e64: {\ + } \ + case e64: { \ VI_VFP_LOOP_REDUCTION_BASE(64) \ BODY64; \ set_fp_exceptions; \ VI_VFP_LOOP_REDUCTION_END(e64) \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2097,10 +2097,10 @@ reg_t index[P.VU.vlmax]; \ require((P.VU.vsew == e16 && p->extension_enabled('F')) || \ (P.VU.vsew == e32 && p->extension_enabled('D'))); \ bool is_active = false; \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ float32_t vd_0 = P.VU.elt<float32_t>(rs1_num, 0); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i) { \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); \ is_active = true; \ float32_t vs2 = f16_to_f32(P.VU.elt<float16_t>(rs2_num, i)); \ @@ -2108,10 +2108,10 @@ reg_t index[P.VU.vlmax]; \ set_fp_exceptions; \ VI_VFP_LOOP_REDUCTION_END(e32) \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ float64_t vd_0 = P.VU.elt<float64_t>(rs1_num, 0); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i) { \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); \ is_active = true; \ float64_t vs2 = f32_to_f64(P.VU.elt<float32_t>(rs2_num, i)); \ @@ -2119,7 +2119,7 @@ reg_t index[P.VU.vlmax]; \ set_fp_exceptions; \ VI_VFP_LOOP_REDUCTION_END(e64) \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2128,25 +2128,25 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_VF_LOOP(BODY16, BODY32, BODY64) \ VI_CHECK_SSS(false); \ VI_VFP_LOOP_BASE \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ VFP_VF_PARAMS(16); \ BODY16; \ set_fp_exceptions; \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ VFP_VF_PARAMS(32); \ BODY32; \ set_fp_exceptions; \ break; \ - }\ - case e64: {\ + } \ + case e64: { \ VFP_VF_PARAMS(64); \ BODY64; \ set_fp_exceptions; \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2157,25 +2157,25 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_VV_LOOP_CMP(BODY16, BODY32, BODY64) \ VI_CHECK_MSS(true); \ VI_VFP_LOOP_CMP_BASE \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ VFP_VV_PARAMS(16); \ BODY16; \ set_fp_exceptions; \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ VFP_VV_PARAMS(32); \ BODY32; \ set_fp_exceptions; \ break; \ - }\ - case e64: {\ + } \ + case e64: { \ VFP_VV_PARAMS(64); \ BODY64; \ set_fp_exceptions; \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2185,25 +2185,25 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_VF_LOOP_CMP(BODY16, BODY32, BODY64) \ VI_CHECK_MSS(false); \ VI_VFP_LOOP_CMP_BASE \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ VFP_VF_PARAMS(16); \ BODY16; \ set_fp_exceptions; \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ VFP_VF_PARAMS(32); \ BODY32; \ set_fp_exceptions; \ break; \ - }\ - case e64: {\ + } \ + case e64: { \ VFP_VF_PARAMS(64); \ BODY64; \ set_fp_exceptions; \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2213,7 +2213,7 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_VF_LOOP_WIDE(BODY16, BODY32) \ VI_CHECK_DSS(false); \ VI_VFP_LOOP_BASE \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e16: { \ float32_t &vd = P.VU.elt<float32_t>(rd_num, i, true); \ float32_t vs2 = f16_to_f32(P.VU.elt<float16_t>(rs2_num, i)); \ @@ -2222,14 +2222,14 @@ reg_t index[P.VU.vlmax]; \ set_fp_exceptions; \ break; \ } \ - case e32: {\ + case e32: { \ float64_t &vd = P.VU.elt<float64_t>(rd_num, i, true); \ float64_t vs2 = f32_to_f64(P.VU.elt<float32_t>(rs2_num, i)); \ float64_t rs1 = f32_to_f64(f32(READ_FREG(rs1_num))); \ BODY32; \ set_fp_exceptions; \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2241,23 +2241,23 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_VV_LOOP_WIDE(BODY16, BODY32) \ VI_CHECK_DSS(true); \ VI_VFP_LOOP_BASE \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ float32_t &vd = P.VU.elt<float32_t>(rd_num, i, true); \ float32_t vs2 = f16_to_f32(P.VU.elt<float16_t>(rs2_num, i)); \ float32_t vs1 = f16_to_f32(P.VU.elt<float16_t>(rs1_num, i)); \ BODY16; \ set_fp_exceptions; \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ float64_t &vd = P.VU.elt<float64_t>(rd_num, i, true); \ float64_t vs2 = f32_to_f64(P.VU.elt<float32_t>(rs2_num, i)); \ float64_t vs1 = f32_to_f64(P.VU.elt<float32_t>(rs1_num, i)); \ BODY32; \ set_fp_exceptions; \ break; \ - }\ + } \ default: \ require(0); \ break; \ @@ -2268,23 +2268,23 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_WF_LOOP_WIDE(BODY16, BODY32) \ VI_CHECK_DDS(false); \ VI_VFP_LOOP_BASE \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ float32_t &vd = P.VU.elt<float32_t>(rd_num, i, true); \ float32_t vs2 = P.VU.elt<float32_t>(rs2_num, i); \ float32_t rs1 = f16_to_f32(f16(READ_FREG(rs1_num))); \ BODY16; \ set_fp_exceptions; \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ float64_t &vd = P.VU.elt<float64_t>(rd_num, i, true); \ float64_t vs2 = P.VU.elt<float64_t>(rs2_num, i); \ float64_t rs1 = f32_to_f64(f32(READ_FREG(rs1_num))); \ BODY32; \ set_fp_exceptions; \ break; \ - }\ + } \ default: \ require(0); \ }; \ @@ -2294,23 +2294,23 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_WV_LOOP_WIDE(BODY16, BODY32) \ VI_CHECK_DDS(true); \ VI_VFP_LOOP_BASE \ - switch(P.VU.vsew) { \ - case e16: {\ + switch (P.VU.vsew) { \ + case e16: { \ float32_t &vd = P.VU.elt<float32_t>(rd_num, i, true); \ float32_t vs2 = P.VU.elt<float32_t>(rs2_num, i); \ float32_t vs1 = f16_to_f32(P.VU.elt<float16_t>(rs1_num, i)); \ BODY16; \ set_fp_exceptions; \ break; \ - }\ - case e32: {\ + } \ + case e32: { \ float64_t &vd = P.VU.elt<float64_t>(rd_num, i, true); \ float64_t vs2 = P.VU.elt<float64_t>(rs2_num, i); \ float64_t vs1 = f32_to_f64(P.VU.elt<float32_t>(rs1_num, i)); \ BODY32; \ set_fp_exceptions; \ break; \ - }\ + } \ default: \ require(0); \ }; \ @@ -2319,14 +2319,14 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_LOOP_SCALE_BASE \ require_fp; \ - require_vector(true);\ - require(STATE.frm->read() < 0x5);\ + require_vector(true); \ + require(STATE.frm->read() < 0x5); \ reg_t vl = P.VU.vl->read(); \ reg_t rd_num = insn.rd(); \ reg_t rs1_num = insn.rs1(); \ reg_t rs2_num = insn.rs2(); \ softfloat_roundingMode = STATE.frm->read(); \ - for (reg_t i=P.VU.vstart->read(); i<vl; ++i){ \ + for (reg_t i = P.VU.vstart->read(); i < vl; ++i) { \ VI_LOOP_ELEMENT_SKIP(); #define VI_VFP_CVT_LOOP(CVT_PARAMS, CHECK, BODY) \ @@ -2340,7 +2340,7 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_CVT_INT_TO_FP(BODY16, BODY32, BODY64, sign) \ VI_CHECK_SSS(false); \ VI_VFP_COMMON \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e16: \ { VI_VFP_CVT_LOOP(CVT_INT_TO_FP_PARAMS(16, 16, sign), \ { p->extension_enabled(EXT_ZFH); }, \ @@ -2364,7 +2364,7 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_CVT_FP_TO_INT(BODY16, BODY32, BODY64, sign) \ VI_CHECK_SSS(false); \ VI_VFP_COMMON \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e16: \ { VI_VFP_CVT_LOOP(CVT_FP_TO_INT_PARAMS(16, 16, sign), \ { p->extension_enabled(EXT_ZFH); }, \ @@ -2388,7 +2388,7 @@ reg_t index[P.VU.vlmax]; \ #define VI_VFP_WCVT_FP_TO_FP(BODY8, BODY16, BODY32, \ CHECK8, CHECK16, CHECK32) \ VI_CHECK_DSS(false); \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e16: \ { VI_VFP_CVT_LOOP(CVT_FP_TO_FP_PARAMS(16, 32), CHECK16, BODY16); } \ break; \ @@ -2404,7 +2404,7 @@ reg_t index[P.VU.vlmax]; \ CHECK8, CHECK16, CHECK32, \ sign) \ VI_CHECK_DSS(false); \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e8: \ { VI_VFP_CVT_LOOP(CVT_INT_TO_FP_PARAMS(8, 16, sign), CHECK8, BODY8); } \ break; \ @@ -2423,7 +2423,7 @@ reg_t index[P.VU.vlmax]; \ CHECK8, CHECK16, CHECK32, \ sign) \ VI_CHECK_DSS(false); \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e16: \ { VI_VFP_CVT_LOOP(CVT_FP_TO_INT_PARAMS(16, 32, sign), CHECK16, BODY16); } \ break; \ @@ -2436,9 +2436,9 @@ reg_t index[P.VU.vlmax]; \ } #define VI_VFP_NCVT_FP_TO_FP(BODY8, BODY16, BODY32, \ - CHECK8, CHECK16, CHECK32) \ + CHECK8, CHECK16, CHECK32) \ VI_CHECK_SDS(false); \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e16: \ { VI_VFP_CVT_LOOP(CVT_FP_TO_FP_PARAMS(32, 16), CHECK16, BODY16); } \ break; \ @@ -2451,10 +2451,10 @@ reg_t index[P.VU.vlmax]; \ } #define VI_VFP_NCVT_INT_TO_FP(BODY8, BODY16, BODY32, \ - CHECK8, CHECK16, CHECK32, \ - sign) \ + CHECK8, CHECK16, CHECK32, \ + sign) \ VI_CHECK_SDS(false); \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e16: \ { VI_VFP_CVT_LOOP(CVT_INT_TO_FP_PARAMS(32, 16, sign), CHECK16, BODY16); } \ break; \ @@ -2467,10 +2467,10 @@ reg_t index[P.VU.vlmax]; \ } #define VI_VFP_NCVT_FP_TO_INT(BODY8, BODY16, BODY32, \ - CHECK8, CHECK16, CHECK32, \ - sign) \ + CHECK8, CHECK16, CHECK32, \ + sign) \ VI_CHECK_SDS(false); \ - switch(P.VU.vsew) { \ + switch (P.VU.vsew) { \ case e8: \ { VI_VFP_CVT_LOOP(CVT_FP_TO_INT_PARAMS(16, 8, sign), CHECK8, BODY8); } \ break; \ diff --git a/riscv/processor.cc b/riscv/processor.cc index cf3084b..f2b0ab3 100644 --- a/riscv/processor.cc +++ b/riscv/processor.cc @@ -149,7 +149,7 @@ void processor_t::parse_varch_string(const char* s) } // The integer should be the power of 2 - if (!check_pow2(vlen) || !check_pow2(elen)){ + if (!check_pow2(vlen) || !check_pow2(elen)) { bad_varch_string(s, "The integer value should be the power of 2"); } @@ -211,7 +211,7 @@ void state_t::reset(processor_t* const proc, reg_t max_isa) csrmap[CSR_CYCLEH] = std::make_shared<counter_proxy_csr_t>(proc, CSR_CYCLEH, mcycleh); } } - for (reg_t i=3; i<=31; ++i) { + for (reg_t i = 3; i <= 31; ++i) { const reg_t which_mevent = CSR_MHPMEVENT3 + i - 3; const reg_t which_mcounter = CSR_MHPMCOUNTER3 + i - 3; const reg_t which_mcounterh = CSR_MHPMCOUNTER3H + i - 3; @@ -358,11 +358,11 @@ void state_t::reset(processor_t* const proc, reg_t max_isa) debug_mode = false; single_step = STEP_NONE; - for (int i=0; i < max_pmp; ++i) { + for (int i = 0; i < max_pmp; ++i) { csrmap[CSR_PMPADDR0 + i] = pmpaddr[i] = std::make_shared<pmpaddr_csr_t>(proc, CSR_PMPADDR0 + i); } - for (int i=0; i < max_pmp; i += xlen/8) { - reg_t addr = CSR_PMPCFG0 + i/4; + for (int i = 0; i < max_pmp; i += xlen / 8) { + reg_t addr = CSR_PMPCFG0 + i / 4; csrmap[addr] = std::make_shared<pmpcfg_csr_t>(proc, addr); } @@ -377,14 +377,14 @@ void state_t::reset(processor_t* const proc, reg_t max_isa) csrmap[CSR_MIMPID] = std::make_shared<const_csr_t>(proc, CSR_MIMPID, 0); csrmap[CSR_MVENDORID] = std::make_shared<const_csr_t>(proc, CSR_MVENDORID, 0); csrmap[CSR_MHARTID] = std::make_shared<const_csr_t>(proc, CSR_MHARTID, proc->get_id()); - const reg_t menvcfg_mask = (proc->extension_enabled(EXT_ZICBOM) ? MENVCFG_CBCFE | MENVCFG_CBIE: 0) | - (proc->extension_enabled(EXT_ZICBOZ) ? MENVCFG_CBZE: 0); + const reg_t menvcfg_mask = (proc->extension_enabled(EXT_ZICBOM) ? MENVCFG_CBCFE | MENVCFG_CBIE : 0) | + (proc->extension_enabled(EXT_ZICBOZ) ? MENVCFG_CBZE : 0); csrmap[CSR_MENVCFG] = menvcfg = std::make_shared<masked_csr_t>(proc, CSR_MENVCFG, menvcfg_mask, 0); - const reg_t senvcfg_mask = (proc->extension_enabled(EXT_ZICBOM) ? SENVCFG_CBCFE | SENVCFG_CBIE: 0) | - (proc->extension_enabled(EXT_ZICBOZ) ? SENVCFG_CBZE: 0); + const reg_t senvcfg_mask = (proc->extension_enabled(EXT_ZICBOM) ? SENVCFG_CBCFE | SENVCFG_CBIE : 0) | + (proc->extension_enabled(EXT_ZICBOZ) ? SENVCFG_CBZE : 0); csrmap[CSR_SENVCFG] = senvcfg = std::make_shared<masked_csr_t>(proc, CSR_SENVCFG, senvcfg_mask, 0); - const reg_t henvcfg_mask = (proc->extension_enabled(EXT_ZICBOM) ? HENVCFG_CBCFE | HENVCFG_CBIE: 0) | - (proc->extension_enabled(EXT_ZICBOZ) ? HENVCFG_CBZE: 0); + const reg_t henvcfg_mask = (proc->extension_enabled(EXT_ZICBOM) ? HENVCFG_CBCFE | HENVCFG_CBIE : 0) | + (proc->extension_enabled(EXT_ZICBOZ) ? HENVCFG_CBZE : 0); csrmap[CSR_HENVCFG] = henvcfg = std::make_shared<masked_csr_t>(proc, CSR_HENVCFG, henvcfg_mask, 0); serialized = false; @@ -399,7 +399,8 @@ void state_t::reset(processor_t* const proc, reg_t max_isa) #endif } -void processor_t::vectorUnit_t::reset(){ +void processor_t::vectorUnit_t::reset() +{ free(reg_file); VLEN = get_vlen(); ELEN = get_elen(); @@ -420,9 +421,10 @@ void processor_t::vectorUnit_t::reset(){ set_vl(0, 0, 0, -1); // default to illegal configuration } -reg_t processor_t::vectorUnit_t::set_vl(int rd, int rs1, reg_t reqVL, reg_t newType){ +reg_t processor_t::vectorUnit_t::set_vl(int rd, int rs1, reg_t reqVL, reg_t newType) +{ int new_vlmul = 0; - if (vtype->read() != newType){ + if (vtype->read() != newType) { vtype->write_raw(newType); vsew = 1 << (extract64(newType, 3, 3) + 3); new_vlmul = int8_t(extract64(newType, 0, 3) << 5) >> 5; @@ -536,7 +538,8 @@ void processor_t::set_pmp_num(reg_t n) n_pmp = n; } -void processor_t::set_pmp_granularity(reg_t gran) { +void processor_t::set_pmp_granularity(reg_t gran) +{ // check the pmp granularity is set from dtb(!=0) and is power of 2 if (gran < (1 << PMP_SHIFT) || (gran & (gran - 1)) != 0) { fprintf(stderr, "error: bad pmp granularity '%ld' from the dtb\n", (unsigned long)gran); @@ -624,7 +627,7 @@ void processor_t::take_interrupt(reg_t pending_interrupts) else abort(); - throw trap_t(((reg_t)1 << (isa->get_max_xlen()-1)) | ctz(enabled_interrupts)); + throw trap_t(((reg_t)1 << (isa->get_max_xlen() - 1)) | ctz(enabled_interrupts)); } } @@ -678,7 +681,7 @@ void processor_t::enter_debug_mode(uint8_t cause) void processor_t::debug_output_log(std::stringstream *s) { - if (log_file==stderr) { + if (log_file == stderr) { std::ostream out(sout_.rdbuf()); out << s->str(); // handles command line options -d -s -l } else { @@ -697,7 +700,7 @@ void processor_t::take_trap(trap_t& t, reg_t epc) << std::hex << std::setfill('0') << std::setw(max_xlen/4) << zext(epc, max_xlen) << std::endl; if (t.has_tval()) s << "core " << std::dec << std::setfill(' ') << std::setw(3) << id - << ": tval 0x" << std::hex << std::setfill('0') << std::setw(max_xlen/4) + << ": tval 0x" << std::hex << std::setfill('0') << std::setw(max_xlen / 4) << zext(t.get_tval(), max_xlen) << std::endl; debug_output_log(&s); } @@ -723,18 +726,18 @@ void processor_t::take_trap(trap_t& t, reg_t epc) reg_t vsdeleg, hsdeleg; reg_t bit = t.cause(); bool curr_virt = state.v; - bool interrupt = (bit & ((reg_t)1 << (max_xlen-1))) != 0; + bool interrupt = (bit & ((reg_t)1 << (max_xlen - 1))) != 0; if (interrupt) { vsdeleg = (curr_virt && state.prv <= PRV_S) ? state.hideleg->read() : 0; hsdeleg = (state.prv <= PRV_S) ? state.mideleg->read() : 0; - bit &= ~((reg_t)1 << (max_xlen-1)); + bit &= ~((reg_t)1 << (max_xlen - 1)); } else { vsdeleg = (curr_virt && state.prv <= PRV_S) ? (state.medeleg->read() & state.hedeleg->read()) : 0; hsdeleg = (state.prv <= PRV_S) ? state.medeleg->read() : 0; } if (state.prv <= PRV_S && bit < max_xlen && ((vsdeleg >> bit) & 1)) { // Handle the trap in VS-mode - reg_t vector = (state.vstvec->read() & 1) && interrupt ? 4*bit : 0; + reg_t vector = (state.vstvec->read() & 1) && interrupt ? 4 * bit : 0; state.pc = (state.vstvec->read() & ~(reg_t)1) + vector; state.vscause->write((interrupt) ? (t.cause() - 1) : t.cause()); state.vsepc->write(epc); @@ -749,7 +752,7 @@ void processor_t::take_trap(trap_t& t, reg_t epc) } else if (state.prv <= PRV_S && bit < max_xlen && ((hsdeleg >> bit) & 1)) { // Handle the trap in HS-mode set_virt(false); - reg_t vector = (state.stvec->read() & 1) && interrupt ? 4*bit : 0; + reg_t vector = (state.stvec->read() & 1) && interrupt ? 4 * bit : 0; state.pc = (state.stvec->read() & ~(reg_t)1) + vector; state.scause->write(t.cause()); state.sepc->write(epc); @@ -774,7 +777,7 @@ void processor_t::take_trap(trap_t& t, reg_t epc) } else { // Handle the trap in M-mode set_virt(false); - reg_t vector = (state.mtvec->read() & 1) && interrupt ? 4*bit : 0; + reg_t vector = (state.mtvec->read() & 1) && interrupt ? 4 * bit : 0; state.pc = (state.mtvec->read() & ~(reg_t)1) + vector; state.mepc->write(epc); state.mcause->write(t.cause()); @@ -816,7 +819,7 @@ void processor_t::disasm(insn_t insn) unsigned max_xlen = isa->get_max_xlen(); s << "core " << std::dec << std::setfill(' ') << std::setw(3) << id - << std::hex << ": 0x" << std::setfill('0') << std::setw(max_xlen/4) + << std::hex << ": 0x" << std::setfill('0') << std::setw(max_xlen / 4) << zext(state.pc, max_xlen) << " (0x" << std::setw(8) << bits << ") " << disassembler->disassemble(insn) << std::endl; @@ -885,10 +888,10 @@ insn_func_t processor_t::decode_insn(insn_t insn) desc = *p; if (p->mask != 0 && p > &instructions[0]) { - if (p->match != (p-1)->match && p->match != (p+1)->match) { + if (p->match != (p - 1)->match && p->match != (p + 1)->match) { // move to front of opcode list to reduce miss penalty while (--p >= &instructions[0]) - *(p+1) = *p; + *(p + 1) = *p; instructions[0] = desc; } } @@ -948,7 +951,7 @@ void processor_t::register_base_instructions() extern reg_t rv64i_##name(processor_t*, insn_t, reg_t); \ extern reg_t rv32e_##name(processor_t*, insn_t, reg_t); \ extern reg_t rv64e_##name(processor_t*, insn_t, reg_t); \ - register_insn((insn_desc_t){ \ + register_insn((insn_desc_t) { \ name##_match, \ name##_mask, \ rv32i_##name, \ |