aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorRichard Henderson <richard.henderson@linaro.org>2024-05-27 20:39:13 +0000
committerRichard Henderson <richard.henderson@linaro.org>2024-06-19 10:55:13 -0700
commit3683354664ad153907d92b5cacdd523d4a34bf61 (patch)
treec21adf529189a46f02a2f5afc50dcc126721a312
parent840770183f05b7f04123dba2b32360b5ff04bd21 (diff)
downloadqemu-3683354664ad153907d92b5cacdd523d4a34bf61.zip
qemu-3683354664ad153907d92b5cacdd523d4a34bf61.tar.gz
qemu-3683354664ad153907d92b5cacdd523d4a34bf61.tar.bz2
tcg/loongarch64: Split out vdvjukN in tcg_out_vec_op
Fixes a bug in the immediate shifts, because the exact encoding depends on the element size. Reviewed-by: Song Gao <gaosong@loongson.cn> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
-rw-r--r--tcg/loongarch64/tcg-target.c.inc56
1 files changed, 31 insertions, 25 deletions
diff --git a/tcg/loongarch64/tcg-target.c.inc b/tcg/loongarch64/tcg-target.c.inc
index 1d9e0bf..ab1b67e 100644
--- a/tcg/loongarch64/tcg-target.c.inc
+++ b/tcg/loongarch64/tcg-target.c.inc
@@ -1901,6 +1901,9 @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
static const LoongArchInsn rotrv_vec_insn[4] = {
OPC_VROTR_B, OPC_VROTR_H, OPC_VROTR_W, OPC_VROTR_D
};
+ static const LoongArchInsn rotri_vec_insn[4] = {
+ OPC_VROTRI_B, OPC_VROTRI_H, OPC_VROTRI_W, OPC_VROTRI_D
+ };
a0 = args[0];
a1 = args[1];
@@ -2034,15 +2037,6 @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
case INDEX_op_sarv_vec:
insn = sarv_vec_insn[vece];
goto vdvjvk;
- case INDEX_op_shli_vec:
- tcg_out32(s, encode_vdvjuk3_insn(shli_vec_insn[vece], a0, a1, a2));
- break;
- case INDEX_op_shri_vec:
- tcg_out32(s, encode_vdvjuk3_insn(shri_vec_insn[vece], a0, a1, a2));
- break;
- case INDEX_op_sari_vec:
- tcg_out32(s, encode_vdvjuk3_insn(sari_vec_insn[vece], a0, a1, a2));
- break;
case INDEX_op_rotlv_vec:
/* rotlv_vec a1, a2 = rotrv_vec a1, -a2 */
tcg_out32(s, encode_vdvj_insn(neg_vec_insn[vece], TCG_VEC_TMP0, a2));
@@ -2051,38 +2045,50 @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
case INDEX_op_rotrv_vec:
insn = rotrv_vec_insn[vece];
goto vdvjvk;
+ case INDEX_op_shli_vec:
+ insn = shli_vec_insn[vece];
+ goto vdvjukN;
+ case INDEX_op_shri_vec:
+ insn = shri_vec_insn[vece];
+ goto vdvjukN;
+ case INDEX_op_sari_vec:
+ insn = sari_vec_insn[vece];
+ goto vdvjukN;
case INDEX_op_rotli_vec:
/* rotli_vec a1, a2 = rotri_vec a1, -a2 */
a2 = extract32(-a2, 0, 3 + vece);
+ insn = rotri_vec_insn[vece];
+ goto vdvjukN;
+ case INDEX_op_bitsel_vec:
+ /* vbitsel vd, vj, vk, va = bitsel_vec vd, va, vk, vj */
+ tcg_out_opc_vbitsel_v(s, a0, a3, a2, a1);
+ break;
+ case INDEX_op_dupm_vec:
+ tcg_out_dupm_vec(s, type, vece, a0, a1, a2);
+ break;
+ default:
+ g_assert_not_reached();
+ vdvjvk:
+ tcg_out32(s, encode_vdvjvk_insn(insn, a0, a1, a2));
+ break;
+ vdvjukN:
switch (vece) {
case MO_8:
- tcg_out_opc_vrotri_b(s, a0, a1, a2);
+ tcg_out32(s, encode_vdvjuk3_insn(insn, a0, a1, a2));
break;
case MO_16:
- tcg_out_opc_vrotri_h(s, a0, a1, a2);
+ tcg_out32(s, encode_vdvjuk4_insn(insn, a0, a1, a2));
break;
case MO_32:
- tcg_out_opc_vrotri_w(s, a0, a1, a2);
+ tcg_out32(s, encode_vdvjuk5_insn(insn, a0, a1, a2));
break;
case MO_64:
- tcg_out_opc_vrotri_d(s, a0, a1, a2);
+ tcg_out32(s, encode_vdvjuk6_insn(insn, a0, a1, a2));
break;
default:
g_assert_not_reached();
}
break;
- case INDEX_op_bitsel_vec:
- /* vbitsel vd, vj, vk, va = bitsel_vec vd, va, vk, vj */
- tcg_out_opc_vbitsel_v(s, a0, a3, a2, a1);
- break;
- case INDEX_op_dupm_vec:
- tcg_out_dupm_vec(s, type, vece, a0, a1, a2);
- break;
- default:
- g_assert_not_reached();
- vdvjvk:
- tcg_out32(s, encode_vdvjvk_insn(insn, a0, a1, a2));
- break;
}
}