diff options
author | Richard Henderson <richard.henderson@linaro.org> | 2024-06-19 14:00:39 -0700 |
---|---|---|
committer | Richard Henderson <richard.henderson@linaro.org> | 2024-06-19 14:00:39 -0700 |
commit | 02d9c38236cf8c9826e5c5be61780c4444cb4ae0 (patch) | |
tree | 5860c49245d7530e20ef3756c3bd1e9d6ac98be0 | |
parent | 80748eb4fbc70f0a3ae423f2c01cb5a4584d803f (diff) | |
parent | 521d7fb3ebdf88112ed13556a93e3037742b9eb8 (diff) | |
download | qemu-02d9c38236cf8c9826e5c5be61780c4444cb4ae0.zip qemu-02d9c38236cf8c9826e5c5be61780c4444cb4ae0.tar.gz qemu-02d9c38236cf8c9826e5c5be61780c4444cb4ae0.tar.bz2 |
Merge tag 'pull-tcg-20240619' of https://gitlab.com/rth7680/qemu into staging
tcg/loongarch64: Support 64- and 256-bit vectors
tcg/loongarch64: Fix tcg_out_movi vs some pcrel pointers
util/bufferiszero: Split out host include files
util/bufferiszero: Add loongarch64 vector acceleration
accel/tcg: Fix typo causing tb->page_addr[1] to not be recorded
target/sparc: use signed denominator in sdiv helper
linux-user: Make TARGET_NR_setgroups affect only the current thread
# -----BEGIN PGP SIGNATURE-----
#
# iQFRBAABCgA7FiEEekgeeIaLTbaoWgXAZN846K9+IV8FAmZzRoMdHHJpY2hhcmQu
# aGVuZGVyc29uQGxpbmFyby5vcmcACgkQZN846K9+IV9Y7gf/ZUTGjCUdAO7W7J5e
# Z3JLUNOfUHO6PxoE05963XJc+APwKiuL6Yo2bnJo6km7WM50CoaX9/7L9CXD7STg
# s3eUJ2p7FfvOADZgO373nqRrB/2mhvoywhDbVJBl+NcRvRUDW8rMqrlSKIAwDIsC
# kwwTWlCfpBSlUgm/c6yCVmt815+sGUPD2k/p+pIzAVUG6fGYAosC2fwPzPajiDGX
# Q+obV1fryKq2SRR2dMnhmPRtr3pQBBkISLuTX6xNM2+CYhYqhBrAlQaOEGhp7Dx3
# ucKjvQFpHgPOSdQxb/HaDv81A20ZUQaydiNNmuKQcTtMx3MsQFR8NyVjH7L+fbS8
# JokjaQ==
# =yVKz
# -----END PGP SIGNATURE-----
# gpg: Signature made Wed 19 Jun 2024 01:58:43 PM PDT
# gpg: using RSA key 7A481E78868B4DB6A85A05C064DF38E8AF7E215F
# gpg: issuer "richard.henderson@linaro.org"
# gpg: Good signature from "Richard Henderson <richard.henderson@linaro.org>" [ultimate]
* tag 'pull-tcg-20240619' of https://gitlab.com/rth7680/qemu: (24 commits)
tcg/loongarch64: Fix tcg_out_movi vs some pcrel pointers
target/sparc: use signed denominator in sdiv helper
linux-user: Make TARGET_NR_setgroups affect only the current thread
accel/tcg: Fix typo causing tb->page_addr[1] to not be recorded
util/bufferiszero: Add loongarch64 vector acceleration
util/bufferiszero: Split out host include files
tcg/loongarch64: Enable v256 with LASX
tcg/loongarch64: Support LASX in tcg_out_vec_op
tcg/loongarch64: Split out vdvjukN in tcg_out_vec_op
tcg/loongarch64: Remove temp_vec from tcg_out_vec_op
tcg/loongarch64: Support LASX in tcg_out_{mov,ld,st}
tcg/loongarch64: Split out vdvjvk in tcg_out_vec_op
tcg/loongarch64: Support LASX in tcg_out_addsub_vec
tcg/loongarch64: Simplify tcg_out_addsub_vec
tcg/loongarch64: Support LASX in tcg_out_dupi_vec
tcg/loongarch64: Use tcg_out_dup_vec in tcg_out_dupi_vec
tcg/loongarch64: Support LASX in tcg_out_dupm_vec
tcg/loongarch64: Support LASX in tcg_out_dup_vec
tcg/loongarch64: Simplify tcg_out_dup_vec
util/loongarch64: Detect LASX vector support
...
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
-rw-r--r-- | accel/tcg/tb-maint.c | 4 | ||||
-rw-r--r-- | host/include/aarch64/host/bufferiszero.c.inc | 76 | ||||
-rw-r--r-- | host/include/generic/host/bufferiszero.c.inc | 10 | ||||
-rw-r--r-- | host/include/i386/host/bufferiszero.c.inc | 124 | ||||
-rw-r--r-- | host/include/loongarch64/host/bufferiszero.c.inc | 143 | ||||
-rw-r--r-- | host/include/loongarch64/host/cpuinfo.h | 1 | ||||
-rw-r--r-- | host/include/x86_64/host/bufferiszero.c.inc | 1 | ||||
-rw-r--r-- | linux-user/syscall.c | 10 | ||||
-rw-r--r-- | target/sparc/helper.c | 2 | ||||
-rw-r--r-- | tcg/loongarch64/tcg-insn-defs.c.inc | 5149 | ||||
-rw-r--r-- | tcg/loongarch64/tcg-target.c.inc | 587 | ||||
-rw-r--r-- | tcg/loongarch64/tcg-target.h | 4 | ||||
-rw-r--r-- | util/bufferiszero.c | 191 | ||||
-rw-r--r-- | util/cpuinfo-loongarch.c | 1 |
14 files changed, 2315 insertions, 3988 deletions
diff --git a/accel/tcg/tb-maint.c b/accel/tcg/tb-maint.c index 19ae679..cc0f5af 100644 --- a/accel/tcg/tb-maint.c +++ b/accel/tcg/tb-maint.c @@ -713,7 +713,7 @@ static void tb_record(TranslationBlock *tb) tb_page_addr_t paddr0 = tb_page_addr0(tb); tb_page_addr_t paddr1 = tb_page_addr1(tb); tb_page_addr_t pindex0 = paddr0 >> TARGET_PAGE_BITS; - tb_page_addr_t pindex1 = paddr0 >> TARGET_PAGE_BITS; + tb_page_addr_t pindex1 = paddr1 >> TARGET_PAGE_BITS; assert(paddr0 != -1); if (unlikely(paddr1 != -1) && pindex0 != pindex1) { @@ -745,7 +745,7 @@ static void tb_remove(TranslationBlock *tb) tb_page_addr_t paddr0 = tb_page_addr0(tb); tb_page_addr_t paddr1 = tb_page_addr1(tb); tb_page_addr_t pindex0 = paddr0 >> TARGET_PAGE_BITS; - tb_page_addr_t pindex1 = paddr0 >> TARGET_PAGE_BITS; + tb_page_addr_t pindex1 = paddr1 >> TARGET_PAGE_BITS; assert(paddr0 != -1); if (unlikely(paddr1 != -1) && pindex0 != pindex1) { diff --git a/host/include/aarch64/host/bufferiszero.c.inc b/host/include/aarch64/host/bufferiszero.c.inc new file mode 100644 index 0000000..947ee7c --- /dev/null +++ b/host/include/aarch64/host/bufferiszero.c.inc @@ -0,0 +1,76 @@ +/* + * SPDX-License-Identifier: GPL-2.0-or-later + * buffer_is_zero acceleration, aarch64 version. + */ + +#ifdef __ARM_NEON +#include <arm_neon.h> + +/* + * Helper for preventing the compiler from reassociating + * chains of binary vector operations. + */ +#define REASSOC_BARRIER(vec0, vec1) asm("" : "+w"(vec0), "+w"(vec1)) + +static bool buffer_is_zero_simd(const void *buf, size_t len) +{ + uint32x4_t t0, t1, t2, t3; + + /* Align head/tail to 16-byte boundaries. */ + const uint32x4_t *p = QEMU_ALIGN_PTR_DOWN(buf + 16, 16); + const uint32x4_t *e = QEMU_ALIGN_PTR_DOWN(buf + len - 1, 16); + + /* Unaligned loads at head/tail. */ + t0 = vld1q_u32(buf) | vld1q_u32(buf + len - 16); + + /* Collect a partial block at tail end. */ + t1 = e[-7] | e[-6]; + t2 = e[-5] | e[-4]; + t3 = e[-3] | e[-2]; + t0 |= e[-1]; + REASSOC_BARRIER(t0, t1); + REASSOC_BARRIER(t2, t3); + t0 |= t1; + t2 |= t3; + REASSOC_BARRIER(t0, t2); + t0 |= t2; + + /* + * Loop over complete 128-byte blocks. + * With the head and tail removed, e - p >= 14, so the loop + * must iterate at least once. + */ + do { + /* + * Reduce via UMAXV. Whatever the actual result, + * it will only be zero if all input bytes are zero. + */ + if (unlikely(vmaxvq_u32(t0) != 0)) { + return false; + } + + t0 = p[0] | p[1]; + t1 = p[2] | p[3]; + t2 = p[4] | p[5]; + t3 = p[6] | p[7]; + REASSOC_BARRIER(t0, t1); + REASSOC_BARRIER(t2, t3); + t0 |= t1; + t2 |= t3; + REASSOC_BARRIER(t0, t2); + t0 |= t2; + p += 8; + } while (p < e - 7); + + return vmaxvq_u32(t0) == 0; +} + +static biz_accel_fn const accel_table[] = { + buffer_is_zero_int_ge256, + buffer_is_zero_simd, +}; + +#define best_accel() 1 +#else +# include "host/include/generic/host/bufferiszero.c.inc" +#endif diff --git a/host/include/generic/host/bufferiszero.c.inc b/host/include/generic/host/bufferiszero.c.inc new file mode 100644 index 0000000..ea0875c --- /dev/null +++ b/host/include/generic/host/bufferiszero.c.inc @@ -0,0 +1,10 @@ +/* + * SPDX-License-Identifier: GPL-2.0-or-later + * buffer_is_zero acceleration, generic version. + */ + +static biz_accel_fn const accel_table[1] = { + buffer_is_zero_int_ge256 +}; + +#define best_accel() 0 diff --git a/host/include/i386/host/bufferiszero.c.inc b/host/include/i386/host/bufferiszero.c.inc new file mode 100644 index 0000000..3b9605d --- /dev/null +++ b/host/include/i386/host/bufferiszero.c.inc @@ -0,0 +1,124 @@ +/* + * SPDX-License-Identifier: GPL-2.0-or-later + * buffer_is_zero acceleration, x86 version. + */ + +#if defined(CONFIG_AVX2_OPT) || defined(__SSE2__) +#include <immintrin.h> + +/* Helper for preventing the compiler from reassociating + chains of binary vector operations. */ +#define SSE_REASSOC_BARRIER(vec0, vec1) asm("" : "+x"(vec0), "+x"(vec1)) + +/* Note that these vectorized functions may assume len >= 256. */ + +static bool __attribute__((target("sse2"))) +buffer_zero_sse2(const void *buf, size_t len) +{ + /* Unaligned loads at head/tail. */ + __m128i v = *(__m128i_u *)(buf); + __m128i w = *(__m128i_u *)(buf + len - 16); + /* Align head/tail to 16-byte boundaries. */ + const __m128i *p = QEMU_ALIGN_PTR_DOWN(buf + 16, 16); + const __m128i *e = QEMU_ALIGN_PTR_DOWN(buf + len - 1, 16); + __m128i zero = { 0 }; + + /* Collect a partial block at tail end. */ + v |= e[-1]; w |= e[-2]; + SSE_REASSOC_BARRIER(v, w); + v |= e[-3]; w |= e[-4]; + SSE_REASSOC_BARRIER(v, w); + v |= e[-5]; w |= e[-6]; + SSE_REASSOC_BARRIER(v, w); + v |= e[-7]; v |= w; + + /* + * Loop over complete 128-byte blocks. + * With the head and tail removed, e - p >= 14, so the loop + * must iterate at least once. + */ + do { + v = _mm_cmpeq_epi8(v, zero); + if (unlikely(_mm_movemask_epi8(v) != 0xFFFF)) { + return false; + } + v = p[0]; w = p[1]; + SSE_REASSOC_BARRIER(v, w); + v |= p[2]; w |= p[3]; + SSE_REASSOC_BARRIER(v, w); + v |= p[4]; w |= p[5]; + SSE_REASSOC_BARRIER(v, w); + v |= p[6]; w |= p[7]; + SSE_REASSOC_BARRIER(v, w); + v |= w; + p += 8; + } while (p < e - 7); + + return _mm_movemask_epi8(_mm_cmpeq_epi8(v, zero)) == 0xFFFF; +} + +#ifdef CONFIG_AVX2_OPT +static bool __attribute__((target("avx2"))) +buffer_zero_avx2(const void *buf, size_t len) +{ + /* Unaligned loads at head/tail. */ + __m256i v = *(__m256i_u *)(buf); + __m256i w = *(__m256i_u *)(buf + len - 32); + /* Align head/tail to 32-byte boundaries. */ + const __m256i *p = QEMU_ALIGN_PTR_DOWN(buf + 32, 32); + const __m256i *e = QEMU_ALIGN_PTR_DOWN(buf + len - 1, 32); + __m256i zero = { 0 }; + + /* Collect a partial block at tail end. */ + v |= e[-1]; w |= e[-2]; + SSE_REASSOC_BARRIER(v, w); + v |= e[-3]; w |= e[-4]; + SSE_REASSOC_BARRIER(v, w); + v |= e[-5]; w |= e[-6]; + SSE_REASSOC_BARRIER(v, w); + v |= e[-7]; v |= w; + + /* Loop over complete 256-byte blocks. */ + for (; p < e - 7; p += 8) { + /* PTEST is not profitable here. */ + v = _mm256_cmpeq_epi8(v, zero); + if (unlikely(_mm256_movemask_epi8(v) != 0xFFFFFFFF)) { + return false; + } + v = p[0]; w = p[1]; + SSE_REASSOC_BARRIER(v, w); + v |= p[2]; w |= p[3]; + SSE_REASSOC_BARRIER(v, w); + v |= p[4]; w |= p[5]; + SSE_REASSOC_BARRIER(v, w); + v |= p[6]; w |= p[7]; + SSE_REASSOC_BARRIER(v, w); + v |= w; + } + + return _mm256_movemask_epi8(_mm256_cmpeq_epi8(v, zero)) == 0xFFFFFFFF; +} +#endif /* CONFIG_AVX2_OPT */ + +static biz_accel_fn const accel_table[] = { + buffer_is_zero_int_ge256, + buffer_zero_sse2, +#ifdef CONFIG_AVX2_OPT + buffer_zero_avx2, +#endif +}; + +static unsigned best_accel(void) +{ +#ifdef CONFIG_AVX2_OPT + unsigned info = cpuinfo_init(); + if (info & CPUINFO_AVX2) { + return 2; + } +#endif + return 1; +} + +#else +# include "host/include/generic/host/bufferiszero.c.inc" +#endif diff --git a/host/include/loongarch64/host/bufferiszero.c.inc b/host/include/loongarch64/host/bufferiszero.c.inc new file mode 100644 index 0000000..69891ea --- /dev/null +++ b/host/include/loongarch64/host/bufferiszero.c.inc @@ -0,0 +1,143 @@ +/* + * SPDX-License-Identifier: GPL-2.0-or-later + * buffer_is_zero acceleration, loongarch64 version. + */ + +/* + * Builtins for LSX and LASX are introduced by gcc 14 and llvm 18, + * but as yet neither has support for attribute target, so neither + * is able to enable the optimization without globally enabling + * vector support. Since we want runtime detection, use assembly. + */ + +static bool buffer_is_zero_lsx(const void *buf, size_t len) +{ + const void *p = QEMU_ALIGN_PTR_DOWN(buf + 16, 16); + const void *e = QEMU_ALIGN_PTR_DOWN(buf + len - 1, 16) - (7 * 16); + const void *l = buf + len; + bool ret; + + asm("vld $vr0,%2,0\n\t" /* first: buf + 0 */ + "vld $vr1,%4,-16\n\t" /* last: buf + len - 16 */ + "vld $vr2,%3,0\n\t" /* e[0] */ + "vld $vr3,%3,16\n\t" /* e[1] */ + "vld $vr4,%3,32\n\t" /* e[2] */ + "vld $vr5,%3,48\n\t" /* e[3] */ + "vld $vr6,%3,64\n\t" /* e[4] */ + "vld $vr7,%3,80\n\t" /* e[5] */ + "vld $vr8,%3,96\n\t" /* e[6] */ + "vor.v $vr0,$vr0,$vr1\n\t" + "vor.v $vr2,$vr2,$vr3\n\t" + "vor.v $vr4,$vr4,$vr5\n\t" + "vor.v $vr6,$vr6,$vr7\n\t" + "vor.v $vr0,$vr0,$vr2\n\t" + "vor.v $vr4,$vr4,$vr6\n\t" + "vor.v $vr0,$vr0,$vr4\n\t" + "vor.v $vr0,$vr0,$vr8\n\t" + "or %0,$r0,$r0\n" /* prepare return false */ + "1:\n\t" + "vsetnez.v $fcc0,$vr0\n\t" + "bcnez $fcc0,2f\n\t" + "vld $vr0,%1,0\n\t" /* p[0] */ + "vld $vr1,%1,16\n\t" /* p[1] */ + "vld $vr2,%1,32\n\t" /* p[2] */ + "vld $vr3,%1,48\n\t" /* p[3] */ + "vld $vr4,%1,64\n\t" /* p[4] */ + "vld $vr5,%1,80\n\t" /* p[5] */ + "vld $vr6,%1,96\n\t" /* p[6] */ + "vld $vr7,%1,112\n\t" /* p[7] */ + "addi.d %1,%1,128\n\t" + "vor.v $vr0,$vr0,$vr1\n\t" + "vor.v $vr2,$vr2,$vr3\n\t" + "vor.v $vr4,$vr4,$vr5\n\t" + "vor.v $vr6,$vr6,$vr7\n\t" + "vor.v $vr0,$vr0,$vr2\n\t" + "vor.v $vr4,$vr4,$vr6\n\t" + "vor.v $vr0,$vr0,$vr4\n\t" + "bltu %1,%3,1b\n\t" + "vsetnez.v $fcc0,$vr0\n\t" + "bcnez $fcc0,2f\n\t" + "ori %0,$r0,1\n" + "2:" + : "=&r"(ret), "+r"(p) + : "r"(buf), "r"(e), "r"(l) + : "f0", "f1", "f2", "f3", "f4", "f5", "f6", "f7", "f8", "fcc0"); + + return ret; +} + +static bool buffer_is_zero_lasx(const void *buf, size_t len) +{ + const void *p = QEMU_ALIGN_PTR_DOWN(buf + 32, 32); + const void *e = QEMU_ALIGN_PTR_DOWN(buf + len - 1, 32) - (7 * 32); + const void *l = buf + len; + bool ret; + + asm("xvld $xr0,%2,0\n\t" /* first: buf + 0 */ + "xvld $xr1,%4,-32\n\t" /* last: buf + len - 32 */ + "xvld $xr2,%3,0\n\t" /* e[0] */ + "xvld $xr3,%3,32\n\t" /* e[1] */ + "xvld $xr4,%3,64\n\t" /* e[2] */ + "xvld $xr5,%3,96\n\t" /* e[3] */ + "xvld $xr6,%3,128\n\t" /* e[4] */ + "xvld $xr7,%3,160\n\t" /* e[5] */ + "xvld $xr8,%3,192\n\t" /* e[6] */ + "xvor.v $xr0,$xr0,$xr1\n\t" + "xvor.v $xr2,$xr2,$xr3\n\t" + "xvor.v $xr4,$xr4,$xr5\n\t" + "xvor.v $xr6,$xr6,$xr7\n\t" + "xvor.v $xr0,$xr0,$xr2\n\t" + "xvor.v $xr4,$xr4,$xr6\n\t" + "xvor.v $xr0,$xr0,$xr4\n\t" + "xvor.v $xr0,$xr0,$xr8\n\t" + "or %0,$r0,$r0\n\t" /* prepare return false */ + "bgeu %1,%3,2f\n" + "1:\n\t" + "xvsetnez.v $fcc0,$xr0\n\t" + "bcnez $fcc0,3f\n\t" + "xvld $xr0,%1,0\n\t" /* p[0] */ + "xvld $xr1,%1,32\n\t" /* p[1] */ + "xvld $xr2,%1,64\n\t" /* p[2] */ + "xvld $xr3,%1,96\n\t" /* p[3] */ + "xvld $xr4,%1,128\n\t" /* p[4] */ + "xvld $xr5,%1,160\n\t" /* p[5] */ + "xvld $xr6,%1,192\n\t" /* p[6] */ + "xvld $xr7,%1,224\n\t" /* p[7] */ + "addi.d %1,%1,256\n\t" + "xvor.v $xr0,$xr0,$xr1\n\t" + "xvor.v $xr2,$xr2,$xr3\n\t" + "xvor.v $xr4,$xr4,$xr5\n\t" + "xvor.v $xr6,$xr6,$xr7\n\t" + "xvor.v $xr0,$xr0,$xr2\n\t" + "xvor.v $xr4,$xr4,$xr6\n\t" + "xvor.v $xr0,$xr0,$xr4\n\t" + "bltu %1,%3,1b\n" + "2:\n\t" + "xvsetnez.v $fcc0,$xr0\n\t" + "bcnez $fcc0,3f\n\t" + "ori %0,$r0,1\n" + "3:" + : "=&r"(ret), "+r"(p) + : "r"(buf), "r"(e), "r"(l) + : "f0", "f1", "f2", "f3", "f4", "f5", "f6", "f7", "f8", "fcc0"); + + return ret; +} + +static biz_accel_fn const accel_table[] = { + buffer_is_zero_int_ge256, + buffer_is_zero_lsx, + buffer_is_zero_lasx, +}; + +static unsigned best_accel(void) +{ + unsigned info = cpuinfo_init(); + if (info & CPUINFO_LASX) { + return 2; + } + if (info & CPUINFO_LSX) { + return 1; + } + return 0; +} diff --git a/host/include/loongarch64/host/cpuinfo.h b/host/include/loongarch64/host/cpuinfo.h index fab664a..d7bf275 100644 --- a/host/include/loongarch64/host/cpuinfo.h +++ b/host/include/loongarch64/host/cpuinfo.h @@ -8,6 +8,7 @@ #define CPUINFO_ALWAYS (1u << 0) /* so cpuinfo is nonzero */ #define CPUINFO_LSX (1u << 1) +#define CPUINFO_LASX (1u << 2) /* Initialized with a constructor. */ extern unsigned cpuinfo; diff --git a/host/include/x86_64/host/bufferiszero.c.inc b/host/include/x86_64/host/bufferiszero.c.inc new file mode 100644 index 0000000..1d3f1fd --- /dev/null +++ b/host/include/x86_64/host/bufferiszero.c.inc @@ -0,0 +1 @@ +#include "host/include/i386/host/bufferiszero.c.inc" diff --git a/linux-user/syscall.c b/linux-user/syscall.c index b9b5a38..e280431 100644 --- a/linux-user/syscall.c +++ b/linux-user/syscall.c @@ -7209,11 +7209,17 @@ static inline int tswapid(int id) #else #define __NR_sys_setresgid __NR_setresgid #endif +#ifdef __NR_setgroups32 +#define __NR_sys_setgroups __NR_setgroups32 +#else +#define __NR_sys_setgroups __NR_setgroups +#endif _syscall1(int, sys_setuid, uid_t, uid) _syscall1(int, sys_setgid, gid_t, gid) _syscall3(int, sys_setresuid, uid_t, ruid, uid_t, euid, uid_t, suid) _syscall3(int, sys_setresgid, gid_t, rgid, gid_t, egid, gid_t, sgid) +_syscall2(int, sys_setgroups, int, size, gid_t *, grouplist) void syscall_init(void) { @@ -11891,7 +11897,7 @@ static abi_long do_syscall1(CPUArchState *cpu_env, int num, abi_long arg1, unlock_user(target_grouplist, arg2, gidsetsize * sizeof(target_id)); } - return get_errno(setgroups(gidsetsize, grouplist)); + return get_errno(sys_setgroups(gidsetsize, grouplist)); } case TARGET_NR_fchown: return get_errno(fchown(arg1, low2highuid(arg2), low2highgid(arg3))); @@ -12227,7 +12233,7 @@ static abi_long do_syscall1(CPUArchState *cpu_env, int num, abi_long arg1, } unlock_user(target_grouplist, arg2, 0); } - return get_errno(setgroups(gidsetsize, grouplist)); + return get_errno(sys_setgroups(gidsetsize, grouplist)); } #endif #ifdef TARGET_NR_fchown32 diff --git a/target/sparc/helper.c b/target/sparc/helper.c index 2247e24..7846ddd 100644 --- a/target/sparc/helper.c +++ b/target/sparc/helper.c @@ -121,7 +121,7 @@ uint64_t helper_sdiv(CPUSPARCState *env, target_ulong a, target_ulong b) return (uint32_t)(b32 < 0 ? INT32_MAX : INT32_MIN) | (-1ull << 32); } - a64 /= b; + a64 /= b32; r = a64; if (unlikely(r != a64)) { return (uint32_t)(a64 < 0 ? INT32_MIN : INT32_MAX) | (-1ull << 32); diff --git a/tcg/loongarch64/tcg-insn-defs.c.inc b/tcg/loongarch64/tcg-insn-defs.c.inc index ee3b483..6bb8656 100644 --- a/tcg/loongarch64/tcg-insn-defs.c.inc +++ b/tcg/loongarch64/tcg-insn-defs.c.inc @@ -4,11 +4,13 @@ * * This file is auto-generated by genqemutcgdefs from * https://github.com/loongson-community/loongarch-opcodes, - * from commit 8027da9a8157a8b47fc48ff1def292e09c5668bd. + * from commit 7f353fb69bd99ce6edfad7ad63948c4bb526f0bf. * DO NOT EDIT. */ typedef enum { + OPC_MOVGR2SCR = 0x00000800, + OPC_MOVSCR2GR = 0x00000c00, OPC_CLZ_W = 0x00001400, OPC_CTZ_W = 0x00001c00, OPC_CLZ_D = 0x00002400, @@ -38,6 +40,8 @@ typedef enum { OPC_SLL_D = 0x00188000, OPC_SRL_D = 0x00190000, OPC_SRA_D = 0x00198000, + OPC_ROTR_B = 0x001a0000, + OPC_ROTR_H = 0x001a8000, OPC_ROTR_W = 0x001b0000, OPC_ROTR_D = 0x001b8000, OPC_MUL_W = 0x001c0000, @@ -60,12 +64,17 @@ typedef enum { OPC_SRLI_D = 0x00450000, OPC_SRAI_W = 0x00488000, OPC_SRAI_D = 0x00490000, + OPC_ROTRI_B = 0x004c2000, + OPC_ROTRI_H = 0x004c4000, OPC_ROTRI_W = 0x004c8000, OPC_ROTRI_D = 0x004d0000, OPC_BSTRINS_W = 0x00600000, OPC_BSTRPICK_W = 0x00608000, OPC_BSTRINS_D = 0x00800000, OPC_BSTRPICK_D = 0x00c00000, + OPC_FMOV_D = 0x01149800, + OPC_MOVGR2FR_D = 0x0114a800, + OPC_MOVFR2GR_D = 0x0114b800, OPC_SLTI = 0x02000000, OPC_SLTUI = 0x02400000, OPC_ADDI_W = 0x02800000, @@ -74,60 +83,10 @@ typedef enum { OPC_ANDI = 0x03400000, OPC_ORI = 0x03800000, OPC_XORI = 0x03c00000, - OPC_VFMADD_S = 0x09100000, - OPC_VFMADD_D = 0x09200000, - OPC_VFMSUB_S = 0x09500000, - OPC_VFMSUB_D = 0x09600000, - OPC_VFNMADD_S = 0x09900000, - OPC_VFNMADD_D = 0x09a00000, - OPC_VFNMSUB_S = 0x09d00000, - OPC_VFNMSUB_D = 0x09e00000, - OPC_VFCMP_CAF_S = 0x0c500000, - OPC_VFCMP_SAF_S = 0x0c508000, - OPC_VFCMP_CLT_S = 0x0c510000, - OPC_VFCMP_SLT_S = 0x0c518000, - OPC_VFCMP_CEQ_S = 0x0c520000, - OPC_VFCMP_SEQ_S = 0x0c528000, - OPC_VFCMP_CLE_S = 0x0c530000, - OPC_VFCMP_SLE_S = 0x0c538000, - OPC_VFCMP_CUN_S = 0x0c540000, - OPC_VFCMP_SUN_S = 0x0c548000, - OPC_VFCMP_CULT_S = 0x0c550000, - OPC_VFCMP_SULT_S = 0x0c558000, - OPC_VFCMP_CUEQ_S = 0x0c560000, - OPC_VFCMP_SUEQ_S = 0x0c568000, - OPC_VFCMP_CULE_S = 0x0c570000, - OPC_VFCMP_SULE_S = 0x0c578000, - OPC_VFCMP_CNE_S = 0x0c580000, - OPC_VFCMP_SNE_S = 0x0c588000, - OPC_VFCMP_COR_S = 0x0c5a0000, - OPC_VFCMP_SOR_S = 0x0c5a8000, - OPC_VFCMP_CUNE_S = 0x0c5c0000, - OPC_VFCMP_SUNE_S = 0x0c5c8000, - OPC_VFCMP_CAF_D = 0x0c600000, - OPC_VFCMP_SAF_D = 0x0c608000, - OPC_VFCMP_CLT_D = 0x0c610000, - OPC_VFCMP_SLT_D = 0x0c618000, - OPC_VFCMP_CEQ_D = 0x0c620000, - OPC_VFCMP_SEQ_D = 0x0c628000, - OPC_VFCMP_CLE_D = 0x0c630000, - OPC_VFCMP_SLE_D = 0x0c638000, - OPC_VFCMP_CUN_D = 0x0c640000, - OPC_VFCMP_SUN_D = 0x0c648000, - OPC_VFCMP_CULT_D = 0x0c650000, - OPC_VFCMP_SULT_D = 0x0c658000, - OPC_VFCMP_CUEQ_D = 0x0c660000, - OPC_VFCMP_SUEQ_D = 0x0c668000, - OPC_VFCMP_CULE_D = 0x0c670000, - OPC_VFCMP_SULE_D = 0x0c678000, - OPC_VFCMP_CNE_D = 0x0c680000, - OPC_VFCMP_SNE_D = 0x0c688000, - OPC_VFCMP_COR_D = 0x0c6a0000, - OPC_VFCMP_SOR_D = 0x0c6a8000, - OPC_VFCMP_CUNE_D = 0x0c6c0000, - OPC_VFCMP_SUNE_D = 0x0c6c8000, OPC_VBITSEL_V = 0x0d100000, + OPC_XVBITSEL_V = 0x0d200000, OPC_VSHUF_B = 0x0d500000, + OPC_XVSHUF_B = 0x0d600000, OPC_ADDU16I_D = 0x10000000, OPC_LU12I_W = 0x14000000, OPC_CU32I_D = 0x16000000, @@ -146,8 +105,14 @@ typedef enum { OPC_LD_BU = 0x2a000000, OPC_LD_HU = 0x2a400000, OPC_LD_WU = 0x2a800000, + OPC_FLD_S = 0x2b000000, + OPC_FST_S = 0x2b400000, + OPC_FLD_D = 0x2b800000, + OPC_FST_D = 0x2bc00000, OPC_VLD = 0x2c000000, OPC_VST = 0x2c400000, + OPC_XVLD = 0x2c800000, + OPC_XVST = 0x2cc00000, OPC_VLDREPL_D = 0x30100000, OPC_VLDREPL_W = 0x30200000, OPC_VLDREPL_H = 0x30400000, @@ -156,6 +121,14 @@ typedef enum { OPC_VSTELM_W = 0x31200000, OPC_VSTELM_H = 0x31400000, OPC_VSTELM_B = 0x31800000, + OPC_XVLDREPL_D = 0x32100000, + OPC_XVLDREPL_W = 0x32200000, + OPC_XVLDREPL_H = 0x32400000, + OPC_XVLDREPL_B = 0x32800000, + OPC_XVSTELM_D = 0x33100000, + OPC_XVSTELM_W = 0x33200000, + OPC_XVSTELM_H = 0x33400000, + OPC_XVSTELM_B = 0x33800000, OPC_LDX_B = 0x38000000, OPC_LDX_H = 0x38040000, OPC_LDX_W = 0x38080000, @@ -167,9 +140,17 @@ typedef enum { OPC_LDX_BU = 0x38200000, OPC_LDX_HU = 0x38240000, OPC_LDX_WU = 0x38280000, + OPC_FLDX_S = 0x38300000, + OPC_FLDX_D = 0x38340000, + OPC_FSTX_S = 0x38380000, + OPC_FSTX_D = 0x383c0000, OPC_VLDX = 0x38400000, OPC_VSTX = 0x38440000, + OPC_XVLDX = 0x38480000, + OPC_XVSTX = 0x384c0000, OPC_DBAR = 0x38720000, + OPC_JISCR0 = 0x48000200, + OPC_JISCR1 = 0x48000300, OPC_JIRL = 0x4c000000, OPC_B = 0x50000000, OPC_BL = 0x54000000, @@ -207,46 +188,6 @@ typedef enum { OPC_VSUB_H = 0x700c8000, OPC_VSUB_W = 0x700d0000, OPC_VSUB_D = 0x700d8000, - OPC_VADDWEV_H_B = 0x701e0000, - OPC_VADDWEV_W_H = 0x701e8000, - OPC_VADDWEV_D_W = 0x701f0000, - OPC_VADDWEV_Q_D = 0x701f8000, - OPC_VSUBWEV_H_B = 0x70200000, - OPC_VSUBWEV_W_H = 0x70208000, - OPC_VSUBWEV_D_W = 0x70210000, - OPC_VSUBWEV_Q_D = 0x70218000, - OPC_VADDWOD_H_B = 0x70220000, - OPC_VADDWOD_W_H = 0x70228000, - OPC_VADDWOD_D_W = 0x70230000, - OPC_VADDWOD_Q_D = 0x70238000, - OPC_VSUBWOD_H_B = 0x70240000, - OPC_VSUBWOD_W_H = 0x70248000, - OPC_VSUBWOD_D_W = 0x70250000, - OPC_VSUBWOD_Q_D = 0x70258000, - OPC_VADDWEV_H_BU = 0x702e0000, - OPC_VADDWEV_W_HU = 0x702e8000, - OPC_VADDWEV_D_WU = 0x702f0000, - OPC_VADDWEV_Q_DU = 0x702f8000, - OPC_VSUBWEV_H_BU = 0x70300000, - OPC_VSUBWEV_W_HU = 0x70308000, - OPC_VSUBWEV_D_WU = 0x70310000, - OPC_VSUBWEV_Q_DU = 0x70318000, - OPC_VADDWOD_H_BU = 0x70320000, - OPC_VADDWOD_W_HU = 0x70328000, - OPC_VADDWOD_D_WU = 0x70330000, - OPC_VADDWOD_Q_DU = 0x70338000, - OPC_VSUBWOD_H_BU = 0x70340000, - OPC_VSUBWOD_W_HU = 0x70348000, - OPC_VSUBWOD_D_WU = 0x70350000, - OPC_VSUBWOD_Q_DU = 0x70358000, - OPC_VADDWEV_H_BU_B = 0x703e0000, - OPC_VADDWEV_W_HU_H = 0x703e8000, - OPC_VADDWEV_D_WU_W = 0x703f0000, - OPC_VADDWEV_Q_DU_D = 0x703f8000, - OPC_VADDWOD_H_BU_B = 0x70400000, - OPC_VADDWOD_W_HU_H = 0x70408000, - OPC_VADDWOD_D_WU_W = 0x70410000, - OPC_VADDWOD_Q_DU_D = 0x70418000, OPC_VSADD_B = 0x70460000, OPC_VSADD_H = 0x70468000, OPC_VSADD_W = 0x70470000, @@ -263,50 +204,6 @@ typedef enum { OPC_VSSUB_HU = 0x704c8000, OPC_VSSUB_WU = 0x704d0000, OPC_VSSUB_DU = 0x704d8000, - OPC_VHADDW_H_B = 0x70540000, - OPC_VHADDW_W_H = 0x70548000, - OPC_VHADDW_D_W = 0x70550000, - OPC_VHADDW_Q_D = 0x70558000, - OPC_VHSUBW_H_B = 0x70560000, - OPC_VHSUBW_W_H = 0x70568000, - OPC_VHSUBW_D_W = 0x70570000, - OPC_VHSUBW_Q_D = 0x70578000, - OPC_VHADDW_HU_BU = 0x70580000, - OPC_VHADDW_WU_HU = 0x70588000, - OPC_VHADDW_DU_WU = 0x70590000, - OPC_VHADDW_QU_DU = 0x70598000, - OPC_VHSUBW_HU_BU = 0x705a0000, - OPC_VHSUBW_WU_HU = 0x705a8000, - OPC_VHSUBW_DU_WU = 0x705b0000, - OPC_VHSUBW_QU_DU = 0x705b8000, - OPC_VADDA_B = 0x705c0000, - OPC_VADDA_H = 0x705c8000, - OPC_VADDA_W = 0x705d0000, - OPC_VADDA_D = 0x705d8000, - OPC_VABSD_B = 0x70600000, - OPC_VABSD_H = 0x70608000, - OPC_VABSD_W = 0x70610000, - OPC_VABSD_D = 0x70618000, - OPC_VABSD_BU = 0x70620000, - OPC_VABSD_HU = 0x70628000, - OPC_VABSD_WU = 0x70630000, - OPC_VABSD_DU = 0x70638000, - OPC_VAVG_B = 0x70640000, - OPC_VAVG_H = 0x70648000, - OPC_VAVG_W = 0x70650000, - OPC_VAVG_D = 0x70658000, - OPC_VAVG_BU = 0x70660000, - OPC_VAVG_HU = 0x70668000, - OPC_VAVG_WU = 0x70670000, - OPC_VAVG_DU = 0x70678000, - OPC_VAVGR_B = 0x70680000, - OPC_VAVGR_H = 0x70688000, - OPC_VAVGR_W = 0x70690000, - OPC_VAVGR_D = 0x70698000, - OPC_VAVGR_BU = 0x706a0000, - OPC_VAVGR_HU = 0x706a8000, - OPC_VAVGR_WU = 0x706b0000, - OPC_VAVGR_DU = 0x706b8000, OPC_VMAX_B = 0x70700000, OPC_VMAX_H = 0x70708000, OPC_VMAX_W = 0x70710000, @@ -327,86 +224,6 @@ typedef enum { OPC_VMUL_H = 0x70848000, OPC_VMUL_W = 0x70850000, OPC_VMUL_D = 0x70858000, - OPC_VMUH_B = 0x70860000, - OPC_VMUH_H = 0x70868000, - OPC_VMUH_W = 0x70870000, - OPC_VMUH_D = 0x70878000, - OPC_VMUH_BU = 0x70880000, - OPC_VMUH_HU = 0x70888000, - OPC_VMUH_WU = 0x70890000, - OPC_VMUH_DU = 0x70898000, - OPC_VMULWEV_H_B = 0x70900000, - OPC_VMULWEV_W_H = 0x70908000, - OPC_VMULWEV_D_W = 0x70910000, - OPC_VMULWEV_Q_D = 0x70918000, - OPC_VMULWOD_H_B = 0x70920000, - OPC_VMULWOD_W_H = 0x70928000, - OPC_VMULWOD_D_W = 0x70930000, - OPC_VMULWOD_Q_D = 0x70938000, - OPC_VMULWEV_H_BU = 0x70980000, - OPC_VMULWEV_W_HU = 0x70988000, - OPC_VMULWEV_D_WU = 0x70990000, - OPC_VMULWEV_Q_DU = 0x70998000, - OPC_VMULWOD_H_BU = 0x709a0000, - OPC_VMULWOD_W_HU = 0x709a8000, - OPC_VMULWOD_D_WU = 0x709b0000, - OPC_VMULWOD_Q_DU = 0x709b8000, - OPC_VMULWEV_H_BU_B = 0x70a00000, - OPC_VMULWEV_W_HU_H = 0x70a08000, - OPC_VMULWEV_D_WU_W = 0x70a10000, - OPC_VMULWEV_Q_DU_D = 0x70a18000, - OPC_VMULWOD_H_BU_B = 0x70a20000, - OPC_VMULWOD_W_HU_H = 0x70a28000, - OPC_VMULWOD_D_WU_W = 0x70a30000, - OPC_VMULWOD_Q_DU_D = 0x70a38000, - OPC_VMADD_B = 0x70a80000, - OPC_VMADD_H = 0x70a88000, - OPC_VMADD_W = 0x70a90000, - OPC_VMADD_D = 0x70a98000, - OPC_VMSUB_B = 0x70aa0000, - OPC_VMSUB_H = 0x70aa8000, - OPC_VMSUB_W = 0x70ab0000, - OPC_VMSUB_D = 0x70ab8000, - OPC_VMADDWEV_H_B = 0x70ac0000, - OPC_VMADDWEV_W_H = 0x70ac8000, - OPC_VMADDWEV_D_W = 0x70ad0000, - OPC_VMADDWEV_Q_D = 0x70ad8000, - OPC_VMADDWOD_H_B = 0x70ae0000, - OPC_VMADDWOD_W_H = 0x70ae8000, - OPC_VMADDWOD_D_W = 0x70af0000, - OPC_VMADDWOD_Q_D = 0x70af8000, - OPC_VMADDWEV_H_BU = 0x70b40000, - OPC_VMADDWEV_W_HU = 0x70b48000, - OPC_VMADDWEV_D_WU = 0x70b50000, - OPC_VMADDWEV_Q_DU = 0x70b58000, - OPC_VMADDWOD_H_BU = 0x70b60000, - OPC_VMADDWOD_W_HU = 0x70b68000, - OPC_VMADDWOD_D_WU = 0x70b70000, - OPC_VMADDWOD_Q_DU = 0x70b78000, - OPC_VMADDWEV_H_BU_B = 0x70bc0000, - OPC_VMADDWEV_W_HU_H = 0x70bc8000, - OPC_VMADDWEV_D_WU_W = 0x70bd0000, - OPC_VMADDWEV_Q_DU_D = 0x70bd8000, - OPC_VMADDWOD_H_BU_B = 0x70be0000, - OPC_VMADDWOD_W_HU_H = 0x70be8000, - OPC_VMADDWOD_D_WU_W = 0x70bf0000, - OPC_VMADDWOD_Q_DU_D = 0x70bf8000, - OPC_VDIV_B = 0x70e00000, - OPC_VDIV_H = 0x70e08000, - OPC_VDIV_W = 0x70e10000, - OPC_VDIV_D = 0x70e18000, - OPC_VMOD_B = 0x70e20000, - OPC_VMOD_H = 0x70e28000, - OPC_VMOD_W = 0x70e30000, - OPC_VMOD_D = 0x70e38000, - OPC_VDIV_BU = 0x70e40000, - OPC_VDIV_HU = 0x70e48000, - OPC_VDIV_WU = 0x70e50000, - OPC_VDIV_DU = 0x70e58000, - OPC_VMOD_BU = 0x70e60000, - OPC_VMOD_HU = 0x70e68000, - OPC_VMOD_WU = 0x70e70000, - OPC_VMOD_DU = 0x70e78000, OPC_VSLL_B = 0x70e80000, OPC_VSLL_H = 0x70e88000, OPC_VSLL_W = 0x70e90000, @@ -423,86 +240,6 @@ typedef enum { OPC_VROTR_H = 0x70ee8000, OPC_VROTR_W = 0x70ef0000, OPC_VROTR_D = 0x70ef8000, - OPC_VSRLR_B = 0x70f00000, - OPC_VSRLR_H = 0x70f08000, - OPC_VSRLR_W = 0x70f10000, - OPC_VSRLR_D = 0x70f18000, - OPC_VSRAR_B = 0x70f20000, - OPC_VSRAR_H = 0x70f28000, - OPC_VSRAR_W = 0x70f30000, - OPC_VSRAR_D = 0x70f38000, - OPC_VSRLN_B_H = 0x70f48000, - OPC_VSRLN_H_W = 0x70f50000, - OPC_VSRLN_W_D = 0x70f58000, - OPC_VSRAN_B_H = 0x70f68000, - OPC_VSRAN_H_W = 0x70f70000, - OPC_VSRAN_W_D = 0x70f78000, - OPC_VSRLRN_B_H = 0x70f88000, - OPC_VSRLRN_H_W = 0x70f90000, - OPC_VSRLRN_W_D = 0x70f98000, - OPC_VSRARN_B_H = 0x70fa8000, - OPC_VSRARN_H_W = 0x70fb0000, - OPC_VSRARN_W_D = 0x70fb8000, - OPC_VSSRLN_B_H = 0x70fc8000, - OPC_VSSRLN_H_W = 0x70fd0000, - OPC_VSSRLN_W_D = 0x70fd8000, - OPC_VSSRAN_B_H = 0x70fe8000, - OPC_VSSRAN_H_W = 0x70ff0000, - OPC_VSSRAN_W_D = 0x70ff8000, - OPC_VSSRLRN_B_H = 0x71008000, - OPC_VSSRLRN_H_W = 0x71010000, - OPC_VSSRLRN_W_D = 0x71018000, - OPC_VSSRARN_B_H = 0x71028000, - OPC_VSSRARN_H_W = 0x71030000, - OPC_VSSRARN_W_D = 0x71038000, - OPC_VSSRLN_BU_H = 0x71048000, - OPC_VSSRLN_HU_W = 0x71050000, - OPC_VSSRLN_WU_D = 0x71058000, - OPC_VSSRAN_BU_H = 0x71068000, - OPC_VSSRAN_HU_W = 0x71070000, - OPC_VSSRAN_WU_D = 0x71078000, - OPC_VSSRLRN_BU_H = 0x71088000, - OPC_VSSRLRN_HU_W = 0x71090000, - OPC_VSSRLRN_WU_D = 0x71098000, - OPC_VSSRARN_BU_H = 0x710a8000, - OPC_VSSRARN_HU_W = 0x710b0000, - OPC_VSSRARN_WU_D = 0x710b8000, - OPC_VBITCLR_B = 0x710c0000, - OPC_VBITCLR_H = 0x710c8000, - OPC_VBITCLR_W = 0x710d0000, - OPC_VBITCLR_D = 0x710d8000, - OPC_VBITSET_B = 0x710e0000, - OPC_VBITSET_H = 0x710e8000, - OPC_VBITSET_W = 0x710f0000, - OPC_VBITSET_D = 0x710f8000, - OPC_VBITREV_B = 0x71100000, - OPC_VBITREV_H = 0x71108000, - OPC_VBITREV_W = 0x71110000, - OPC_VBITREV_D = 0x71118000, - OPC_VPACKEV_B = 0x71160000, - OPC_VPACKEV_H = 0x71168000, - OPC_VPACKEV_W = 0x71170000, - OPC_VPACKEV_D = 0x71178000, - OPC_VPACKOD_B = 0x71180000, - OPC_VPACKOD_H = 0x71188000, - OPC_VPACKOD_W = 0x71190000, - OPC_VPACKOD_D = 0x71198000, - OPC_VILVL_B = 0x711a0000, - OPC_VILVL_H = 0x711a8000, - OPC_VILVL_W = 0x711b0000, - OPC_VILVL_D = 0x711b8000, - OPC_VILVH_B = 0x711c0000, - OPC_VILVH_H = 0x711c8000, - OPC_VILVH_W = 0x711d0000, - OPC_VILVH_D = 0x711d8000, - OPC_VPICKEV_B = 0x711e0000, - OPC_VPICKEV_H = 0x711e8000, - OPC_VPICKEV_W = 0x711f0000, - OPC_VPICKEV_D = 0x711f8000, - OPC_VPICKOD_B = 0x71200000, - OPC_VPICKOD_H = 0x71208000, - OPC_VPICKOD_W = 0x71210000, - OPC_VPICKOD_D = 0x71218000, OPC_VREPLVE_B = 0x71220000, OPC_VREPLVE_H = 0x71228000, OPC_VREPLVE_W = 0x71230000, @@ -513,41 +250,6 @@ typedef enum { OPC_VNOR_V = 0x71278000, OPC_VANDN_V = 0x71280000, OPC_VORN_V = 0x71288000, - OPC_VFRSTP_B = 0x712b0000, - OPC_VFRSTP_H = 0x712b8000, - OPC_VADD_Q = 0x712d0000, - OPC_VSUB_Q = 0x712d8000, - OPC_VSIGNCOV_B = 0x712e0000, - OPC_VSIGNCOV_H = 0x712e8000, - OPC_VSIGNCOV_W = 0x712f0000, - OPC_VSIGNCOV_D = 0x712f8000, - OPC_VFADD_S = 0x71308000, - OPC_VFADD_D = 0x71310000, - OPC_VFSUB_S = 0x71328000, - OPC_VFSUB_D = 0x71330000, - OPC_VFMUL_S = 0x71388000, - OPC_VFMUL_D = 0x71390000, - OPC_VFDIV_S = 0x713a8000, - OPC_VFDIV_D = 0x713b0000, - OPC_VFMAX_S = 0x713c8000, - OPC_VFMAX_D = 0x713d0000, - OPC_VFMIN_S = 0x713e8000, - OPC_VFMIN_D = 0x713f0000, - OPC_VFMAXA_S = 0x71408000, - OPC_VFMAXA_D = 0x71410000, - OPC_VFMINA_S = 0x71428000, - OPC_VFMINA_D = 0x71430000, - OPC_VFCVT_H_S = 0x71460000, - OPC_VFCVT_S_D = 0x71468000, - OPC_VFFINT_S_L = 0x71480000, - OPC_VFTINT_W_D = 0x71498000, - OPC_VFTINTRM_W_D = 0x714a0000, - OPC_VFTINTRP_W_D = 0x714a8000, - OPC_VFTINTRZ_W_D = 0x714b0000, - OPC_VFTINTRNE_W_D = 0x714b8000, - OPC_VSHUF_H = 0x717a8000, - OPC_VSHUF_W = 0x717b0000, - OPC_VSHUF_D = 0x717b8000, OPC_VSEQI_B = 0x72800000, OPC_VSEQI_H = 0x72808000, OPC_VSEQI_W = 0x72810000, @@ -576,8 +278,6 @@ typedef enum { OPC_VSUBI_HU = 0x728c8000, OPC_VSUBI_WU = 0x728d0000, OPC_VSUBI_DU = 0x728d8000, - OPC_VBSLL_V = 0x728e0000, - OPC_VBSRL_V = 0x728e8000, OPC_VMAXI_B = 0x72900000, OPC_VMAXI_H = 0x72908000, OPC_VMAXI_W = 0x72910000, @@ -594,102 +294,10 @@ typedef enum { OPC_VMINI_HU = 0x72968000, OPC_VMINI_WU = 0x72970000, OPC_VMINI_DU = 0x72978000, - OPC_VFRSTPI_B = 0x729a0000, - OPC_VFRSTPI_H = 0x729a8000, - OPC_VCLO_B = 0x729c0000, - OPC_VCLO_H = 0x729c0400, - OPC_VCLO_W = 0x729c0800, - OPC_VCLO_D = 0x729c0c00, - OPC_VCLZ_B = 0x729c1000, - OPC_VCLZ_H = 0x729c1400, - OPC_VCLZ_W = 0x729c1800, - OPC_VCLZ_D = 0x729c1c00, - OPC_VPCNT_B = 0x729c2000, - OPC_VPCNT_H = 0x729c2400, - OPC_VPCNT_W = 0x729c2800, - OPC_VPCNT_D = 0x729c2c00, OPC_VNEG_B = 0x729c3000, OPC_VNEG_H = 0x729c3400, OPC_VNEG_W = 0x729c3800, OPC_VNEG_D = 0x729c3c00, - OPC_VMSKLTZ_B = 0x729c4000, - OPC_VMSKLTZ_H = 0x729c4400, - OPC_VMSKLTZ_W = 0x729c4800, - OPC_VMSKLTZ_D = 0x729c4c00, - OPC_VMSKGEZ_B = 0x729c5000, - OPC_VMSKNZ_B = 0x729c6000, - OPC_VSETEQZ_V = 0x729c9800, - OPC_VSETNEZ_V = 0x729c9c00, - OPC_VSETANYEQZ_B = 0x729ca000, - OPC_VSETANYEQZ_H = 0x729ca400, - OPC_VSETANYEQZ_W = 0x729ca800, - OPC_VSETANYEQZ_D = 0x729cac00, - OPC_VSETALLNEZ_B = 0x729cb000, - OPC_VSETALLNEZ_H = 0x729cb400, - OPC_VSETALLNEZ_W = 0x729cb800, - OPC_VSETALLNEZ_D = 0x729cbc00, - OPC_VFLOGB_S = 0x729cc400, - OPC_VFLOGB_D = 0x729cc800, - OPC_VFCLASS_S = 0x729cd400, - OPC_VFCLASS_D = 0x729cd800, - OPC_VFSQRT_S = 0x729ce400, - OPC_VFSQRT_D = 0x729ce800, - OPC_VFRECIP_S = 0x729cf400, - OPC_VFRECIP_D = 0x729cf800, - OPC_VFRSQRT_S = 0x729d0400, - OPC_VFRSQRT_D = 0x729d0800, - OPC_VFRINT_S = 0x729d3400, - OPC_VFRINT_D = 0x729d3800, - OPC_VFRINTRM_S = 0x729d4400, - OPC_VFRINTRM_D = 0x729d4800, - OPC_VFRINTRP_S = 0x729d5400, - OPC_VFRINTRP_D = 0x729d5800, - OPC_VFRINTRZ_S = 0x729d6400, - OPC_VFRINTRZ_D = 0x729d6800, - OPC_VFRINTRNE_S = 0x729d7400, - OPC_VFRINTRNE_D = 0x729d7800, - OPC_VFCVTL_S_H = 0x729de800, - OPC_VFCVTH_S_H = 0x729dec00, - OPC_VFCVTL_D_S = 0x729df000, - OPC_VFCVTH_D_S = 0x729df400, - OPC_VFFINT_S_W = 0x729e0000, - OPC_VFFINT_S_WU = 0x729e0400, - OPC_VFFINT_D_L = 0x729e0800, - OPC_VFFINT_D_LU = 0x729e0c00, - OPC_VFFINTL_D_W = 0x729e1000, - OPC_VFFINTH_D_W = 0x729e1400, - OPC_VFTINT_W_S = 0x729e3000, - OPC_VFTINT_L_D = 0x729e3400, - OPC_VFTINTRM_W_S = 0x729e3800, - OPC_VFTINTRM_L_D = 0x729e3c00, - OPC_VFTINTRP_W_S = 0x729e4000, - OPC_VFTINTRP_L_D = 0x729e4400, - OPC_VFTINTRZ_W_S = 0x729e4800, - OPC_VFTINTRZ_L_D = 0x729e4c00, - OPC_VFTINTRNE_W_S = 0x729e5000, - OPC_VFTINTRNE_L_D = 0x729e5400, - OPC_VFTINT_WU_S = 0x729e5800, - OPC_VFTINT_LU_D = 0x729e5c00, - OPC_VFTINTRZ_WU_S = 0x729e7000, - OPC_VFTINTRZ_LU_D = 0x729e7400, - OPC_VFTINTL_L_S = 0x729e8000, - OPC_VFTINTH_L_S = 0x729e8400, - OPC_VFTINTRML_L_S = 0x729e8800, - OPC_VFTINTRMH_L_S = 0x729e8c00, - OPC_VFTINTRPL_L_S = 0x729e9000, - OPC_VFTINTRPH_L_S = 0x729e9400, - OPC_VFTINTRZL_L_S = 0x729e9800, - OPC_VFTINTRZH_L_S = 0x729e9c00, - OPC_VFTINTRNEL_L_S = 0x729ea000, - OPC_VFTINTRNEH_L_S = 0x729ea400, - OPC_VEXTH_H_B = 0x729ee000, - OPC_VEXTH_W_H = 0x729ee400, - OPC_VEXTH_D_W = 0x729ee800, - OPC_VEXTH_Q_D = 0x729eec00, - OPC_VEXTH_HU_BU = 0x729ef000, - OPC_VEXTH_WU_HU = 0x729ef400, - OPC_VEXTH_DU_WU = 0x729ef800, - OPC_VEXTH_QU_DU = 0x729efc00, OPC_VREPLGR2VR_B = 0x729f0000, OPC_VREPLGR2VR_H = 0x729f0400, OPC_VREPLGR2VR_W = 0x729f0800, @@ -698,14 +306,6 @@ typedef enum { OPC_VROTRI_H = 0x72a04000, OPC_VROTRI_W = 0x72a08000, OPC_VROTRI_D = 0x72a10000, - OPC_VSRLRI_B = 0x72a42000, - OPC_VSRLRI_H = 0x72a44000, - OPC_VSRLRI_W = 0x72a48000, - OPC_VSRLRI_D = 0x72a50000, - OPC_VSRARI_B = 0x72a82000, - OPC_VSRARI_H = 0x72a84000, - OPC_VSRARI_W = 0x72a88000, - OPC_VSRARI_D = 0x72a90000, OPC_VINSGR2VR_B = 0x72eb8000, OPC_VINSGR2VR_H = 0x72ebc000, OPC_VINSGR2VR_W = 0x72ebe000, @@ -722,14 +322,6 @@ typedef enum { OPC_VREPLVEI_H = 0x72f7c000, OPC_VREPLVEI_W = 0x72f7e000, OPC_VREPLVEI_D = 0x72f7f000, - OPC_VSLLWIL_H_B = 0x73082000, - OPC_VSLLWIL_W_H = 0x73084000, - OPC_VSLLWIL_D_W = 0x73088000, - OPC_VEXTL_Q_D = 0x73090000, - OPC_VSLLWIL_HU_BU = 0x730c2000, - OPC_VSLLWIL_WU_HU = 0x730c4000, - OPC_VSLLWIL_DU_WU = 0x730c8000, - OPC_VEXTL_QU_DU = 0x730d0000, OPC_VBITCLRI_B = 0x73102000, OPC_VBITCLRI_H = 0x73104000, OPC_VBITCLRI_W = 0x73108000, @@ -742,14 +334,6 @@ typedef enum { OPC_VBITREVI_H = 0x73184000, OPC_VBITREVI_W = 0x73188000, OPC_VBITREVI_D = 0x73190000, - OPC_VSAT_B = 0x73242000, - OPC_VSAT_H = 0x73244000, - OPC_VSAT_W = 0x73248000, - OPC_VSAT_D = 0x73250000, - OPC_VSAT_BU = 0x73282000, - OPC_VSAT_HU = 0x73284000, - OPC_VSAT_WU = 0x73288000, - OPC_VSAT_DU = 0x73290000, OPC_VSLLI_B = 0x732c2000, OPC_VSLLI_H = 0x732c4000, OPC_VSLLI_W = 0x732c8000, @@ -762,69 +346,203 @@ typedef enum { OPC_VSRAI_H = 0x73344000, OPC_VSRAI_W = 0x73348000, OPC_VSRAI_D = 0x73350000, - OPC_VSRLNI_B_H = 0x73404000, - OPC_VSRLNI_H_W = 0x73408000, - OPC_VSRLNI_W_D = 0x73410000, - OPC_VSRLNI_D_Q = 0x73420000, - OPC_VSRLRNI_B_H = 0x73444000, - OPC_VSRLRNI_H_W = 0x73448000, - OPC_VSRLRNI_W_D = 0x73450000, - OPC_VSRLRNI_D_Q = 0x73460000, - OPC_VSSRLNI_B_H = 0x73484000, - OPC_VSSRLNI_H_W = 0x73488000, - OPC_VSSRLNI_W_D = 0x73490000, - OPC_VSSRLNI_D_Q = 0x734a0000, - OPC_VSSRLNI_BU_H = 0x734c4000, - OPC_VSSRLNI_HU_W = 0x734c8000, - OPC_VSSRLNI_WU_D = 0x734d0000, - OPC_VSSRLNI_DU_Q = 0x734e0000, - OPC_VSSRLRNI_B_H = 0x73504000, - OPC_VSSRLRNI_H_W = 0x73508000, - OPC_VSSRLRNI_W_D = 0x73510000, - OPC_VSSRLRNI_D_Q = 0x73520000, - OPC_VSSRLRNI_BU_H = 0x73544000, - OPC_VSSRLRNI_HU_W = 0x73548000, - OPC_VSSRLRNI_WU_D = 0x73550000, - OPC_VSSRLRNI_DU_Q = 0x73560000, - OPC_VSRANI_B_H = 0x73584000, - OPC_VSRANI_H_W = 0x73588000, - OPC_VSRANI_W_D = 0x73590000, - OPC_VSRANI_D_Q = 0x735a0000, - OPC_VSRARNI_B_H = 0x735c4000, - OPC_VSRARNI_H_W = 0x735c8000, - OPC_VSRARNI_W_D = 0x735d0000, - OPC_VSRARNI_D_Q = 0x735e0000, - OPC_VSSRANI_B_H = 0x73604000, - OPC_VSSRANI_H_W = 0x73608000, - OPC_VSSRANI_W_D = 0x73610000, - OPC_VSSRANI_D_Q = 0x73620000, - OPC_VSSRANI_BU_H = 0x73644000, - OPC_VSSRANI_HU_W = 0x73648000, - OPC_VSSRANI_WU_D = 0x73650000, - OPC_VSSRANI_DU_Q = 0x73660000, - OPC_VSSRARNI_B_H = 0x73684000, - OPC_VSSRARNI_H_W = 0x73688000, - OPC_VSSRARNI_W_D = 0x73690000, - OPC_VSSRARNI_D_Q = 0x736a0000, - OPC_VSSRARNI_BU_H = 0x736c4000, - OPC_VSSRARNI_HU_W = 0x736c8000, - OPC_VSSRARNI_WU_D = 0x736d0000, - OPC_VSSRARNI_DU_Q = 0x736e0000, - OPC_VEXTRINS_D = 0x73800000, - OPC_VEXTRINS_W = 0x73840000, - OPC_VEXTRINS_H = 0x73880000, - OPC_VEXTRINS_B = 0x738c0000, - OPC_VSHUF4I_B = 0x73900000, - OPC_VSHUF4I_H = 0x73940000, - OPC_VSHUF4I_W = 0x73980000, - OPC_VSHUF4I_D = 0x739c0000, OPC_VBITSELI_B = 0x73c40000, OPC_VANDI_B = 0x73d00000, OPC_VORI_B = 0x73d40000, OPC_VXORI_B = 0x73d80000, OPC_VNORI_B = 0x73dc0000, OPC_VLDI = 0x73e00000, - OPC_VPERMI_W = 0x73e40000, + OPC_XVSEQ_B = 0x74000000, + OPC_XVSEQ_H = 0x74008000, + OPC_XVSEQ_W = 0x74010000, + OPC_XVSEQ_D = 0x74018000, + OPC_XVSLE_B = 0x74020000, + OPC_XVSLE_H = 0x74028000, + OPC_XVSLE_W = 0x74030000, + OPC_XVSLE_D = 0x74038000, + OPC_XVSLE_BU = 0x74040000, + OPC_XVSLE_HU = 0x74048000, + OPC_XVSLE_WU = 0x74050000, + OPC_XVSLE_DU = 0x74058000, + OPC_XVSLT_B = 0x74060000, + OPC_XVSLT_H = 0x74068000, + OPC_XVSLT_W = 0x74070000, + OPC_XVSLT_D = 0x74078000, + OPC_XVSLT_BU = 0x74080000, + OPC_XVSLT_HU = 0x74088000, + OPC_XVSLT_WU = 0x74090000, + OPC_XVSLT_DU = 0x74098000, + OPC_XVADD_B = 0x740a0000, + OPC_XVADD_H = 0x740a8000, + OPC_XVADD_W = 0x740b0000, + OPC_XVADD_D = 0x740b8000, + OPC_XVSUB_B = 0x740c0000, + OPC_XVSUB_H = 0x740c8000, + OPC_XVSUB_W = 0x740d0000, + OPC_XVSUB_D = 0x740d8000, + OPC_XVSADD_B = 0x74460000, + OPC_XVSADD_H = 0x74468000, + OPC_XVSADD_W = 0x74470000, + OPC_XVSADD_D = 0x74478000, + OPC_XVSSUB_B = 0x74480000, + OPC_XVSSUB_H = 0x74488000, + OPC_XVSSUB_W = 0x74490000, + OPC_XVSSUB_D = 0x74498000, + OPC_XVSADD_BU = 0x744a0000, + OPC_XVSADD_HU = 0x744a8000, + OPC_XVSADD_WU = 0x744b0000, + OPC_XVSADD_DU = 0x744b8000, + OPC_XVSSUB_BU = 0x744c0000, + OPC_XVSSUB_HU = 0x744c8000, + OPC_XVSSUB_WU = 0x744d0000, + OPC_XVSSUB_DU = 0x744d8000, + OPC_XVMAX_B = 0x74700000, + OPC_XVMAX_H = 0x74708000, + OPC_XVMAX_W = 0x74710000, + OPC_XVMAX_D = 0x74718000, + OPC_XVMIN_B = 0x74720000, + OPC_XVMIN_H = 0x74728000, + OPC_XVMIN_W = 0x74730000, + OPC_XVMIN_D = 0x74738000, + OPC_XVMAX_BU = 0x74740000, + OPC_XVMAX_HU = 0x74748000, + OPC_XVMAX_WU = 0x74750000, + OPC_XVMAX_DU = 0x74758000, + OPC_XVMIN_BU = 0x74760000, + OPC_XVMIN_HU = 0x74768000, + OPC_XVMIN_WU = 0x74770000, + OPC_XVMIN_DU = 0x74778000, + OPC_XVMUL_B = 0x74840000, + OPC_XVMUL_H = 0x74848000, + OPC_XVMUL_W = 0x74850000, + OPC_XVMUL_D = 0x74858000, + OPC_XVSLL_B = 0x74e80000, + OPC_XVSLL_H = 0x74e88000, + OPC_XVSLL_W = 0x74e90000, + OPC_XVSLL_D = 0x74e98000, + OPC_XVSRL_B = 0x74ea0000, + OPC_XVSRL_H = 0x74ea8000, + OPC_XVSRL_W = 0x74eb0000, + OPC_XVSRL_D = 0x74eb8000, + OPC_XVSRA_B = 0x74ec0000, + OPC_XVSRA_H = 0x74ec8000, + OPC_XVSRA_W = 0x74ed0000, + OPC_XVSRA_D = 0x74ed8000, + OPC_XVROTR_B = 0x74ee0000, + OPC_XVROTR_H = 0x74ee8000, + OPC_XVROTR_W = 0x74ef0000, + OPC_XVROTR_D = 0x74ef8000, + OPC_XVREPLVE_B = 0x75220000, + OPC_XVREPLVE_H = 0x75228000, + OPC_XVREPLVE_W = 0x75230000, + OPC_XVREPLVE_D = 0x75238000, + OPC_XVAND_V = 0x75260000, + OPC_XVOR_V = 0x75268000, + OPC_XVXOR_V = 0x75270000, + OPC_XVNOR_V = 0x75278000, + OPC_XVANDN_V = 0x75280000, + OPC_XVORN_V = 0x75288000, + OPC_XVSEQI_B = 0x76800000, + OPC_XVSEQI_H = 0x76808000, + OPC_XVSEQI_W = 0x76810000, + OPC_XVSEQI_D = 0x76818000, + OPC_XVSLEI_B = 0x76820000, + OPC_XVSLEI_H = 0x76828000, + OPC_XVSLEI_W = 0x76830000, + OPC_XVSLEI_D = 0x76838000, + OPC_XVSLEI_BU = 0x76840000, + OPC_XVSLEI_HU = 0x76848000, + OPC_XVSLEI_WU = 0x76850000, + OPC_XVSLEI_DU = 0x76858000, + OPC_XVSLTI_B = 0x76860000, + OPC_XVSLTI_H = 0x76868000, + OPC_XVSLTI_W = 0x76870000, + OPC_XVSLTI_D = 0x76878000, + OPC_XVSLTI_BU = 0x76880000, + OPC_XVSLTI_HU = 0x76888000, + OPC_XVSLTI_WU = 0x76890000, + OPC_XVSLTI_DU = 0x76898000, + OPC_XVADDI_BU = 0x768a0000, + OPC_XVADDI_HU = 0x768a8000, + OPC_XVADDI_WU = 0x768b0000, + OPC_XVADDI_DU = 0x768b8000, + OPC_XVSUBI_BU = 0x768c0000, + OPC_XVSUBI_HU = 0x768c8000, + OPC_XVSUBI_WU = 0x768d0000, + OPC_XVSUBI_DU = 0x768d8000, + OPC_XVMAXI_B = 0x76900000, + OPC_XVMAXI_H = 0x76908000, + OPC_XVMAXI_W = 0x76910000, + OPC_XVMAXI_D = 0x76918000, + OPC_XVMINI_B = 0x76920000, + OPC_XVMINI_H = 0x76928000, + OPC_XVMINI_W = 0x76930000, + OPC_XVMINI_D = 0x76938000, + OPC_XVMAXI_BU = 0x76940000, + OPC_XVMAXI_HU = 0x76948000, + OPC_XVMAXI_WU = 0x76950000, + OPC_XVMAXI_DU = 0x76958000, + OPC_XVMINI_BU = 0x76960000, + OPC_XVMINI_HU = 0x76968000, + OPC_XVMINI_WU = 0x76970000, + OPC_XVMINI_DU = 0x76978000, + OPC_XVNEG_B = 0x769c3000, + OPC_XVNEG_H = 0x769c3400, + OPC_XVNEG_W = 0x769c3800, + OPC_XVNEG_D = 0x769c3c00, + OPC_XVREPLGR2VR_B = 0x769f0000, + OPC_XVREPLGR2VR_H = 0x769f0400, + OPC_XVREPLGR2VR_W = 0x769f0800, + OPC_XVREPLGR2VR_D = 0x769f0c00, + OPC_XVROTRI_B = 0x76a02000, + OPC_XVROTRI_H = 0x76a04000, + OPC_XVROTRI_W = 0x76a08000, + OPC_XVROTRI_D = 0x76a10000, + OPC_XVINSGR2VR_W = 0x76ebc000, + OPC_XVINSGR2VR_D = 0x76ebe000, + OPC_XVPICKVE2GR_W = 0x76efc000, + OPC_XVPICKVE2GR_D = 0x76efe000, + OPC_XVPICKVE2GR_WU = 0x76f3c000, + OPC_XVPICKVE2GR_DU = 0x76f3e000, + OPC_XVREPL128VEI_B = 0x76f78000, + OPC_XVREPL128VEI_H = 0x76f7c000, + OPC_XVREPL128VEI_W = 0x76f7e000, + OPC_XVREPL128VEI_D = 0x76f7f000, + OPC_XVREPLVE0_B = 0x77070000, + OPC_XVREPLVE0_H = 0x77078000, + OPC_XVREPLVE0_W = 0x7707c000, + OPC_XVREPLVE0_D = 0x7707e000, + OPC_XVREPLVE0_Q = 0x7707f000, + OPC_XVBITCLRI_B = 0x77102000, + OPC_XVBITCLRI_H = 0x77104000, + OPC_XVBITCLRI_W = 0x77108000, + OPC_XVBITCLRI_D = 0x77110000, + OPC_XVBITSETI_B = 0x77142000, + OPC_XVBITSETI_H = 0x77144000, + OPC_XVBITSETI_W = 0x77148000, + OPC_XVBITSETI_D = 0x77150000, + OPC_XVBITREVI_B = 0x77182000, + OPC_XVBITREVI_H = 0x77184000, + OPC_XVBITREVI_W = 0x77188000, + OPC_XVBITREVI_D = 0x77190000, + OPC_XVSLLI_B = 0x772c2000, + OPC_XVSLLI_H = 0x772c4000, + OPC_XVSLLI_W = 0x772c8000, + OPC_XVSLLI_D = 0x772d0000, + OPC_XVSRLI_B = 0x77302000, + OPC_XVSRLI_H = 0x77304000, + OPC_XVSRLI_W = 0x77308000, + OPC_XVSRLI_D = 0x77310000, + OPC_XVSRAI_B = 0x77342000, + OPC_XVSRAI_H = 0x77344000, + OPC_XVSRAI_W = 0x77348000, + OPC_XVSRAI_D = 0x77350000, + OPC_XVBITSELI_B = 0x77c40000, + OPC_XVANDI_B = 0x77d00000, + OPC_XVORI_B = 0x77d40000, + OPC_XVXORI_B = 0x77d80000, + OPC_XVNORI_B = 0x77dc0000, + OPC_XVLDI = 0x77e00000, } LoongArchInsn; static int32_t __attribute__((unused)) @@ -873,11 +591,11 @@ encode_dk_slots(LoongArchInsn opc, uint32_t d, uint32_t k) } static int32_t __attribute__((unused)) -encode_cdvj_insn(LoongArchInsn opc, TCGReg cd, TCGReg vj) +encode_dfj_insn(LoongArchInsn opc, TCGReg d, TCGReg fj) { - tcg_debug_assert(cd >= 0 && cd <= 0x7); - tcg_debug_assert(vj >= 0x20 && vj <= 0x3f); - return encode_dj_slots(opc, cd, vj & 0x1f); + tcg_debug_assert(d >= 0 && d <= 0x1f); + tcg_debug_assert(fj >= 0x20 && fj <= 0x3f); + return encode_dj_slots(opc, d, fj & 0x1f); } static int32_t __attribute__((unused)) @@ -925,6 +643,24 @@ encode_djuk12_insn(LoongArchInsn opc, TCGReg d, TCGReg j, uint32_t uk12) } static int32_t __attribute__((unused)) +encode_djuk3_insn(LoongArchInsn opc, TCGReg d, TCGReg j, uint32_t uk3) +{ + tcg_debug_assert(d >= 0 && d <= 0x1f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(uk3 <= 0x7); + return encode_djk_slots(opc, d, j, uk3); +} + +static int32_t __attribute__((unused)) +encode_djuk4_insn(LoongArchInsn opc, TCGReg d, TCGReg j, uint32_t uk4) +{ + tcg_debug_assert(d >= 0 && d <= 0x1f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(uk4 <= 0xf); + return encode_djk_slots(opc, d, j, uk4); +} + +static int32_t __attribute__((unused)) encode_djuk5_insn(LoongArchInsn opc, TCGReg d, TCGReg j, uint32_t uk5) { tcg_debug_assert(d >= 0 && d <= 0x1f); @@ -973,6 +709,14 @@ encode_dsj20_insn(LoongArchInsn opc, TCGReg d, int32_t sj20) } static int32_t __attribute__((unused)) +encode_dtj_insn(LoongArchInsn opc, TCGReg d, TCGReg tj) +{ + tcg_debug_assert(d >= 0 && d <= 0x1f); + tcg_debug_assert(tj >= 0 && tj <= 0x3); + return encode_dj_slots(opc, d, tj); +} + +static int32_t __attribute__((unused)) encode_dvjuk1_insn(LoongArchInsn opc, TCGReg d, TCGReg vj, uint32_t uk1) { tcg_debug_assert(d >= 0 && d <= 0x1f); @@ -1009,6 +753,58 @@ encode_dvjuk4_insn(LoongArchInsn opc, TCGReg d, TCGReg vj, uint32_t uk4) } static int32_t __attribute__((unused)) +encode_dxjuk2_insn(LoongArchInsn opc, TCGReg d, TCGReg xj, uint32_t uk2) +{ + tcg_debug_assert(d >= 0 && d <= 0x1f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk2 <= 0x3); + return encode_djk_slots(opc, d, xj & 0x1f, uk2); +} + +static int32_t __attribute__((unused)) +encode_dxjuk3_insn(LoongArchInsn opc, TCGReg d, TCGReg xj, uint32_t uk3) +{ + tcg_debug_assert(d >= 0 && d <= 0x1f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk3 <= 0x7); + return encode_djk_slots(opc, d, xj & 0x1f, uk3); +} + +static int32_t __attribute__((unused)) +encode_fdfj_insn(LoongArchInsn opc, TCGReg fd, TCGReg fj) +{ + tcg_debug_assert(fd >= 0x20 && fd <= 0x3f); + tcg_debug_assert(fj >= 0x20 && fj <= 0x3f); + return encode_dj_slots(opc, fd & 0x1f, fj & 0x1f); +} + +static int32_t __attribute__((unused)) +encode_fdj_insn(LoongArchInsn opc, TCGReg fd, TCGReg j) +{ + tcg_debug_assert(fd >= 0x20 && fd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + return encode_dj_slots(opc, fd & 0x1f, j); +} + +static int32_t __attribute__((unused)) +encode_fdjk_insn(LoongArchInsn opc, TCGReg fd, TCGReg j, TCGReg k) +{ + tcg_debug_assert(fd >= 0x20 && fd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(k >= 0 && k <= 0x1f); + return encode_djk_slots(opc, fd & 0x1f, j, k); +} + +static int32_t __attribute__((unused)) +encode_fdjsk12_insn(LoongArchInsn opc, TCGReg fd, TCGReg j, int32_t sk12) +{ + tcg_debug_assert(fd >= 0x20 && fd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk12 >= -0x800 && sk12 <= 0x7ff); + return encode_djk_slots(opc, fd & 0x1f, j, sk12 & 0xfff); +} + +static int32_t __attribute__((unused)) encode_sd10k16_insn(LoongArchInsn opc, int32_t sd10k16) { tcg_debug_assert(sd10k16 >= -0x2000000 && sd10k16 <= 0x1ffffff); @@ -1016,6 +812,21 @@ encode_sd10k16_insn(LoongArchInsn opc, int32_t sd10k16) } static int32_t __attribute__((unused)) +encode_sd5k16_insn(LoongArchInsn opc, int32_t sd5k16) +{ + tcg_debug_assert(sd5k16 >= -0x100000 && sd5k16 <= 0xfffff); + return encode_dk_slots(opc, (sd5k16 >> 16) & 0x1f, sd5k16 & 0xffff); +} + +static int32_t __attribute__((unused)) +encode_tdj_insn(LoongArchInsn opc, TCGReg td, TCGReg j) +{ + tcg_debug_assert(td >= 0 && td <= 0x3); + tcg_debug_assert(j >= 0 && j <= 0x1f); + return encode_dj_slots(opc, td, j); +} + +static int32_t __attribute__((unused)) encode_ud15_insn(LoongArchInsn opc, uint32_t ud15) { tcg_debug_assert(ud15 <= 0x7fff); @@ -1244,15 +1055,6 @@ encode_vdvjuk6_insn(LoongArchInsn opc, TCGReg vd, TCGReg vj, uint32_t uk6) } static int32_t __attribute__((unused)) -encode_vdvjuk7_insn(LoongArchInsn opc, TCGReg vd, TCGReg vj, uint32_t uk7) -{ - tcg_debug_assert(vd >= 0x20 && vd <= 0x3f); - tcg_debug_assert(vj >= 0x20 && vj <= 0x3f); - tcg_debug_assert(uk7 <= 0x7f); - return encode_djk_slots(opc, vd & 0x1f, vj & 0x1f, uk7); -} - -static int32_t __attribute__((unused)) encode_vdvjuk8_insn(LoongArchInsn opc, TCGReg vd, TCGReg vj, uint32_t uk8) { tcg_debug_assert(vd >= 0x20 && vd <= 0x3f); @@ -1281,6 +1083,252 @@ encode_vdvjvkva_insn(LoongArchInsn opc, TCGReg vd, TCGReg vj, TCGReg vk, return encode_djka_slots(opc, vd & 0x1f, vj & 0x1f, vk & 0x1f, va & 0x1f); } +static int32_t __attribute__((unused)) +encode_xdj_insn(LoongArchInsn opc, TCGReg xd, TCGReg j) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + return encode_dj_slots(opc, xd & 0x1f, j); +} + +static int32_t __attribute__((unused)) +encode_xdjk_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, TCGReg k) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(k >= 0 && k <= 0x1f); + return encode_djk_slots(opc, xd & 0x1f, j, k); +} + +static int32_t __attribute__((unused)) +encode_xdjsk10_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, int32_t sk10) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk10 >= -0x200 && sk10 <= 0x1ff); + return encode_djk_slots(opc, xd & 0x1f, j, sk10 & 0x3ff); +} + +static int32_t __attribute__((unused)) +encode_xdjsk11_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, int32_t sk11) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk11 >= -0x400 && sk11 <= 0x3ff); + return encode_djk_slots(opc, xd & 0x1f, j, sk11 & 0x7ff); +} + +static int32_t __attribute__((unused)) +encode_xdjsk12_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, int32_t sk12) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk12 >= -0x800 && sk12 <= 0x7ff); + return encode_djk_slots(opc, xd & 0x1f, j, sk12 & 0xfff); +} + +static int32_t __attribute__((unused)) +encode_xdjsk8un2_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, int32_t sk8, + uint32_t un2) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk8 >= -0x80 && sk8 <= 0x7f); + tcg_debug_assert(un2 <= 0x3); + return encode_djkn_slots(opc, xd & 0x1f, j, sk8 & 0xff, un2); +} + +static int32_t __attribute__((unused)) +encode_xdjsk8un3_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, int32_t sk8, + uint32_t un3) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk8 >= -0x80 && sk8 <= 0x7f); + tcg_debug_assert(un3 <= 0x7); + return encode_djkn_slots(opc, xd & 0x1f, j, sk8 & 0xff, un3); +} + +static int32_t __attribute__((unused)) +encode_xdjsk8un4_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, int32_t sk8, + uint32_t un4) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk8 >= -0x80 && sk8 <= 0x7f); + tcg_debug_assert(un4 <= 0xf); + return encode_djkn_slots(opc, xd & 0x1f, j, sk8 & 0xff, un4); +} + +static int32_t __attribute__((unused)) +encode_xdjsk8un5_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, int32_t sk8, + uint32_t un5) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk8 >= -0x80 && sk8 <= 0x7f); + tcg_debug_assert(un5 <= 0x1f); + return encode_djkn_slots(opc, xd & 0x1f, j, sk8 & 0xff, un5); +} + +static int32_t __attribute__((unused)) +encode_xdjsk9_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, int32_t sk9) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(sk9 >= -0x100 && sk9 <= 0xff); + return encode_djk_slots(opc, xd & 0x1f, j, sk9 & 0x1ff); +} + +static int32_t __attribute__((unused)) +encode_xdjuk2_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, uint32_t uk2) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(uk2 <= 0x3); + return encode_djk_slots(opc, xd & 0x1f, j, uk2); +} + +static int32_t __attribute__((unused)) +encode_xdjuk3_insn(LoongArchInsn opc, TCGReg xd, TCGReg j, uint32_t uk3) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(j >= 0 && j <= 0x1f); + tcg_debug_assert(uk3 <= 0x7); + return encode_djk_slots(opc, xd & 0x1f, j, uk3); +} + +static int32_t __attribute__((unused)) +encode_xdsj13_insn(LoongArchInsn opc, TCGReg xd, int32_t sj13) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(sj13 >= -0x1000 && sj13 <= 0xfff); + return encode_dj_slots(opc, xd & 0x1f, sj13 & 0x1fff); +} + +static int32_t __attribute__((unused)) +encode_xdxj_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + return encode_dj_slots(opc, xd & 0x1f, xj & 0x1f); +} + +static int32_t __attribute__((unused)) +encode_xdxjk_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, TCGReg k) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(k >= 0 && k <= 0x1f); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, k); +} + +static int32_t __attribute__((unused)) +encode_xdxjsk5_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, int32_t sk5) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(sk5 >= -0x10 && sk5 <= 0xf); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, sk5 & 0x1f); +} + +static int32_t __attribute__((unused)) +encode_xdxjuk1_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, uint32_t uk1) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk1 <= 0x1); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, uk1); +} + +static int32_t __attribute__((unused)) +encode_xdxjuk2_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, uint32_t uk2) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk2 <= 0x3); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, uk2); +} + +static int32_t __attribute__((unused)) +encode_xdxjuk3_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, uint32_t uk3) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk3 <= 0x7); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, uk3); +} + +static int32_t __attribute__((unused)) +encode_xdxjuk4_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, uint32_t uk4) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk4 <= 0xf); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, uk4); +} + +static int32_t __attribute__((unused)) +encode_xdxjuk5_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, uint32_t uk5) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk5 <= 0x1f); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, uk5); +} + +static int32_t __attribute__((unused)) +encode_xdxjuk6_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, uint32_t uk6) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk6 <= 0x3f); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, uk6); +} + +static int32_t __attribute__((unused)) +encode_xdxjuk8_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, uint32_t uk8) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(uk8 <= 0xff); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, uk8); +} + +static int32_t __attribute__((unused)) +encode_xdxjxk_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, TCGReg xk) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(xk >= 0x20 && xk <= 0x3f); + return encode_djk_slots(opc, xd & 0x1f, xj & 0x1f, xk & 0x1f); +} + +static int32_t __attribute__((unused)) +encode_xdxjxkxa_insn(LoongArchInsn opc, TCGReg xd, TCGReg xj, TCGReg xk, + TCGReg xa) +{ + tcg_debug_assert(xd >= 0x20 && xd <= 0x3f); + tcg_debug_assert(xj >= 0x20 && xj <= 0x3f); + tcg_debug_assert(xk >= 0x20 && xk <= 0x3f); + tcg_debug_assert(xa >= 0x20 && xa <= 0x3f); + return encode_djka_slots(opc, xd & 0x1f, xj & 0x1f, xk & 0x1f, xa & 0x1f); +} + +/* Emits the `movgr2scr td, j` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_movgr2scr(TCGContext *s, TCGReg td, TCGReg j) +{ + tcg_out32(s, encode_tdj_insn(OPC_MOVGR2SCR, td, j)); +} + +/* Emits the `movscr2gr d, tj` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_movscr2gr(TCGContext *s, TCGReg d, TCGReg tj) +{ + tcg_out32(s, encode_dtj_insn(OPC_MOVSCR2GR, d, tj)); +} + /* Emits the `clz.w d, j` instruction. */ static void __attribute__((unused)) tcg_out_opc_clz_w(TCGContext *s, TCGReg d, TCGReg j) @@ -1484,6 +1532,20 @@ tcg_out_opc_sra_d(TCGContext *s, TCGReg d, TCGReg j, TCGReg k) tcg_out32(s, encode_djk_insn(OPC_SRA_D, d, j, k)); } +/* Emits the `rotr.b d, j, k` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_rotr_b(TCGContext *s, TCGReg d, TCGReg j, TCGReg k) +{ + tcg_out32(s, encode_djk_insn(OPC_ROTR_B, d, j, k)); +} + +/* Emits the `rotr.h d, j, k` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_rotr_h(TCGContext *s, TCGReg d, TCGReg j, TCGReg k) +{ + tcg_out32(s, encode_djk_insn(OPC_ROTR_H, d, j, k)); +} + /* Emits the `rotr.w d, j, k` instruction. */ static void __attribute__((unused)) tcg_out_opc_rotr_w(TCGContext *s, TCGReg d, TCGReg j, TCGReg k) @@ -1638,6 +1700,20 @@ tcg_out_opc_srai_d(TCGContext *s, TCGReg d, TCGReg j, uint32_t uk6) tcg_out32(s, encode_djuk6_insn(OPC_SRAI_D, d, j, uk6)); } +/* Emits the `rotri.b d, j, uk3` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_rotri_b(TCGContext *s, TCGReg d, TCGReg j, uint32_t uk3) +{ + tcg_out32(s, encode_djuk3_insn(OPC_ROTRI_B, d, j, uk3)); +} + +/* Emits the `rotri.h d, j, uk4` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_rotri_h(TCGContext *s, TCGReg d, TCGReg j, uint32_t uk4) +{ + tcg_out32(s, encode_djuk4_insn(OPC_ROTRI_H, d, j, uk4)); +} + /* Emits the `rotri.w d, j, uk5` instruction. */ static void __attribute__((unused)) tcg_out_opc_rotri_w(TCGContext *s, TCGReg d, TCGReg j, uint32_t uk5) @@ -1684,6 +1760,27 @@ tcg_out_opc_bstrpick_d(TCGContext *s, TCGReg d, TCGReg j, uint32_t uk6, tcg_out32(s, encode_djuk6um6_insn(OPC_BSTRPICK_D, d, j, uk6, um6)); } +/* Emits the `fmov.d fd, fj` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fmov_d(TCGContext *s, TCGReg fd, TCGReg fj) +{ + tcg_out32(s, encode_fdfj_insn(OPC_FMOV_D, fd, fj)); +} + +/* Emits the `movgr2fr.d fd, j` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_movgr2fr_d(TCGContext *s, TCGReg fd, TCGReg j) +{ + tcg_out32(s, encode_fdj_insn(OPC_MOVGR2FR_D, fd, j)); +} + +/* Emits the `movfr2gr.d d, fj` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_movfr2gr_d(TCGContext *s, TCGReg d, TCGReg fj) +{ + tcg_out32(s, encode_dfj_insn(OPC_MOVFR2GR_D, d, fj)); +} + /* Emits the `slti d, j, sk12` instruction. */ static void __attribute__((unused)) tcg_out_opc_slti(TCGContext *s, TCGReg d, TCGReg j, int32_t sk12) @@ -1740,382 +1837,33 @@ tcg_out_opc_xori(TCGContext *s, TCGReg d, TCGReg j, uint32_t uk12) tcg_out32(s, encode_djuk12_insn(OPC_XORI, d, j, uk12)); } -/* Emits the `vfmadd.s vd, vj, vk, va` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmadd_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) -{ - tcg_out32(s, encode_vdvjvkva_insn(OPC_VFMADD_S, vd, vj, vk, va)); -} - -/* Emits the `vfmadd.d vd, vj, vk, va` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmadd_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) -{ - tcg_out32(s, encode_vdvjvkva_insn(OPC_VFMADD_D, vd, vj, vk, va)); -} - -/* Emits the `vfmsub.s vd, vj, vk, va` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmsub_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) -{ - tcg_out32(s, encode_vdvjvkva_insn(OPC_VFMSUB_S, vd, vj, vk, va)); -} - -/* Emits the `vfmsub.d vd, vj, vk, va` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmsub_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) -{ - tcg_out32(s, encode_vdvjvkva_insn(OPC_VFMSUB_D, vd, vj, vk, va)); -} - -/* Emits the `vfnmadd.s vd, vj, vk, va` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfnmadd_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) -{ - tcg_out32(s, encode_vdvjvkva_insn(OPC_VFNMADD_S, vd, vj, vk, va)); -} - -/* Emits the `vfnmadd.d vd, vj, vk, va` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfnmadd_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) -{ - tcg_out32(s, encode_vdvjvkva_insn(OPC_VFNMADD_D, vd, vj, vk, va)); -} - -/* Emits the `vfnmsub.s vd, vj, vk, va` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfnmsub_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) -{ - tcg_out32(s, encode_vdvjvkva_insn(OPC_VFNMSUB_S, vd, vj, vk, va)); -} - -/* Emits the `vfnmsub.d vd, vj, vk, va` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfnmsub_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) -{ - tcg_out32(s, encode_vdvjvkva_insn(OPC_VFNMSUB_D, vd, vj, vk, va)); -} - -/* Emits the `vfcmp.caf.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_caf_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CAF_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.saf.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_saf_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SAF_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.clt.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_clt_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CLT_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.slt.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_slt_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SLT_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.ceq.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_ceq_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CEQ_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.seq.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_seq_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SEQ_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.cle.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cle_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CLE_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.sle.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sle_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SLE_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.cun.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cun_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CUN_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.sun.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sun_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SUN_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.cult.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cult_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CULT_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.sult.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sult_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SULT_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.cueq.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cueq_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CUEQ_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.sueq.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sueq_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SUEQ_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.cule.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cule_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CULE_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.sule.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sule_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SULE_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.cne.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cne_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CNE_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.sne.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sne_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SNE_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.cor.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cor_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_COR_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.sor.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sor_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SOR_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.cune.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cune_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CUNE_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.sune.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sune_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SUNE_S, vd, vj, vk)); -} - -/* Emits the `vfcmp.caf.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_caf_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CAF_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.saf.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_saf_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SAF_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.clt.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_clt_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CLT_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.slt.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_slt_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SLT_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.ceq.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_ceq_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CEQ_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.seq.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_seq_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SEQ_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.cle.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cle_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CLE_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.sle.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sle_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SLE_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.cun.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cun_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CUN_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.sun.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sun_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SUN_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.cult.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cult_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CULT_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.sult.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sult_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SULT_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.cueq.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cueq_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CUEQ_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.sueq.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sueq_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SUEQ_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.cule.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cule_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CULE_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.sule.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sule_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SULE_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.cne.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cne_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CNE_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.sne.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sne_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SNE_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.cor.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_cor_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_COR_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.sor.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcmp_sor_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SOR_D, vd, vj, vk)); -} - -/* Emits the `vfcmp.cune.d vd, vj, vk` instruction. */ +/* Emits the `vbitsel.v vd, vj, vk, va` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfcmp_cune_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) +tcg_out_opc_vbitsel_v(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) { - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_CUNE_D, vd, vj, vk)); + tcg_out32(s, encode_vdvjvkva_insn(OPC_VBITSEL_V, vd, vj, vk, va)); } -/* Emits the `vfcmp.sune.d vd, vj, vk` instruction. */ +/* Emits the `xvbitsel.v xd, xj, xk, xa` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfcmp_sune_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) +tcg_out_opc_xvbitsel_v(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk, + TCGReg xa) { - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCMP_SUNE_D, vd, vj, vk)); + tcg_out32(s, encode_xdxjxkxa_insn(OPC_XVBITSEL_V, xd, xj, xk, xa)); } -/* Emits the `vbitsel.v vd, vj, vk, va` instruction. */ +/* Emits the `vshuf.b vd, vj, vk, va` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitsel_v(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) +tcg_out_opc_vshuf_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) { - tcg_out32(s, encode_vdvjvkva_insn(OPC_VBITSEL_V, vd, vj, vk, va)); + tcg_out32(s, encode_vdvjvkva_insn(OPC_VSHUF_B, vd, vj, vk, va)); } -/* Emits the `vshuf.b vd, vj, vk, va` instruction. */ +/* Emits the `xvshuf.b xd, xj, xk, xa` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vshuf_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk, TCGReg va) +tcg_out_opc_xvshuf_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk, TCGReg xa) { - tcg_out32(s, encode_vdvjvkva_insn(OPC_VSHUF_B, vd, vj, vk, va)); + tcg_out32(s, encode_xdxjxkxa_insn(OPC_XVSHUF_B, xd, xj, xk, xa)); } /* Emits the `addu16i.d d, j, sk16` instruction. */ @@ -2244,6 +1992,34 @@ tcg_out_opc_ld_wu(TCGContext *s, TCGReg d, TCGReg j, int32_t sk12) tcg_out32(s, encode_djsk12_insn(OPC_LD_WU, d, j, sk12)); } +/* Emits the `fld.s fd, j, sk12` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fld_s(TCGContext *s, TCGReg fd, TCGReg j, int32_t sk12) +{ + tcg_out32(s, encode_fdjsk12_insn(OPC_FLD_S, fd, j, sk12)); +} + +/* Emits the `fst.s fd, j, sk12` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fst_s(TCGContext *s, TCGReg fd, TCGReg j, int32_t sk12) +{ + tcg_out32(s, encode_fdjsk12_insn(OPC_FST_S, fd, j, sk12)); +} + +/* Emits the `fld.d fd, j, sk12` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fld_d(TCGContext *s, TCGReg fd, TCGReg j, int32_t sk12) +{ + tcg_out32(s, encode_fdjsk12_insn(OPC_FLD_D, fd, j, sk12)); +} + +/* Emits the `fst.d fd, j, sk12` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fst_d(TCGContext *s, TCGReg fd, TCGReg j, int32_t sk12) +{ + tcg_out32(s, encode_fdjsk12_insn(OPC_FST_D, fd, j, sk12)); +} + /* Emits the `vld vd, j, sk12` instruction. */ static void __attribute__((unused)) tcg_out_opc_vld(TCGContext *s, TCGReg vd, TCGReg j, int32_t sk12) @@ -2258,6 +2034,20 @@ tcg_out_opc_vst(TCGContext *s, TCGReg vd, TCGReg j, int32_t sk12) tcg_out32(s, encode_vdjsk12_insn(OPC_VST, vd, j, sk12)); } +/* Emits the `xvld xd, j, sk12` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvld(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk12) +{ + tcg_out32(s, encode_xdjsk12_insn(OPC_XVLD, xd, j, sk12)); +} + +/* Emits the `xvst xd, j, sk12` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvst(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk12) +{ + tcg_out32(s, encode_xdjsk12_insn(OPC_XVST, xd, j, sk12)); +} + /* Emits the `vldrepl.d vd, j, sk9` instruction. */ static void __attribute__((unused)) tcg_out_opc_vldrepl_d(TCGContext *s, TCGReg vd, TCGReg j, int32_t sk9) @@ -2318,6 +2108,66 @@ tcg_out_opc_vstelm_b(TCGContext *s, TCGReg vd, TCGReg j, int32_t sk8, tcg_out32(s, encode_vdjsk8un4_insn(OPC_VSTELM_B, vd, j, sk8, un4)); } +/* Emits the `xvldrepl.d xd, j, sk9` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvldrepl_d(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk9) +{ + tcg_out32(s, encode_xdjsk9_insn(OPC_XVLDREPL_D, xd, j, sk9)); +} + +/* Emits the `xvldrepl.w xd, j, sk10` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvldrepl_w(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk10) +{ + tcg_out32(s, encode_xdjsk10_insn(OPC_XVLDREPL_W, xd, j, sk10)); +} + +/* Emits the `xvldrepl.h xd, j, sk11` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvldrepl_h(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk11) +{ + tcg_out32(s, encode_xdjsk11_insn(OPC_XVLDREPL_H, xd, j, sk11)); +} + +/* Emits the `xvldrepl.b xd, j, sk12` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvldrepl_b(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk12) +{ + tcg_out32(s, encode_xdjsk12_insn(OPC_XVLDREPL_B, xd, j, sk12)); +} + +/* Emits the `xvstelm.d xd, j, sk8, un2` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvstelm_d(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk8, + uint32_t un2) +{ + tcg_out32(s, encode_xdjsk8un2_insn(OPC_XVSTELM_D, xd, j, sk8, un2)); +} + +/* Emits the `xvstelm.w xd, j, sk8, un3` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvstelm_w(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk8, + uint32_t un3) +{ + tcg_out32(s, encode_xdjsk8un3_insn(OPC_XVSTELM_W, xd, j, sk8, un3)); +} + +/* Emits the `xvstelm.h xd, j, sk8, un4` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvstelm_h(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk8, + uint32_t un4) +{ + tcg_out32(s, encode_xdjsk8un4_insn(OPC_XVSTELM_H, xd, j, sk8, un4)); +} + +/* Emits the `xvstelm.b xd, j, sk8, un5` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvstelm_b(TCGContext *s, TCGReg xd, TCGReg j, int32_t sk8, + uint32_t un5) +{ + tcg_out32(s, encode_xdjsk8un5_insn(OPC_XVSTELM_B, xd, j, sk8, un5)); +} + /* Emits the `ldx.b d, j, k` instruction. */ static void __attribute__((unused)) tcg_out_opc_ldx_b(TCGContext *s, TCGReg d, TCGReg j, TCGReg k) @@ -2395,6 +2245,34 @@ tcg_out_opc_ldx_wu(TCGContext *s, TCGReg d, TCGReg j, TCGReg k) tcg_out32(s, encode_djk_insn(OPC_LDX_WU, d, j, k)); } +/* Emits the `fldx.s fd, j, k` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fldx_s(TCGContext *s, TCGReg fd, TCGReg j, TCGReg k) +{ + tcg_out32(s, encode_fdjk_insn(OPC_FLDX_S, fd, j, k)); +} + +/* Emits the `fldx.d fd, j, k` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fldx_d(TCGContext *s, TCGReg fd, TCGReg j, TCGReg k) +{ + tcg_out32(s, encode_fdjk_insn(OPC_FLDX_D, fd, j, k)); +} + +/* Emits the `fstx.s fd, j, k` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fstx_s(TCGContext *s, TCGReg fd, TCGReg j, TCGReg k) +{ + tcg_out32(s, encode_fdjk_insn(OPC_FSTX_S, fd, j, k)); +} + +/* Emits the `fstx.d fd, j, k` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_fstx_d(TCGContext *s, TCGReg fd, TCGReg j, TCGReg k) +{ + tcg_out32(s, encode_fdjk_insn(OPC_FSTX_D, fd, j, k)); +} + /* Emits the `vldx vd, j, k` instruction. */ static void __attribute__((unused)) tcg_out_opc_vldx(TCGContext *s, TCGReg vd, TCGReg j, TCGReg k) @@ -2409,6 +2287,20 @@ tcg_out_opc_vstx(TCGContext *s, TCGReg vd, TCGReg j, TCGReg k) tcg_out32(s, encode_vdjk_insn(OPC_VSTX, vd, j, k)); } +/* Emits the `xvldx xd, j, k` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvldx(TCGContext *s, TCGReg xd, TCGReg j, TCGReg k) +{ + tcg_out32(s, encode_xdjk_insn(OPC_XVLDX, xd, j, k)); +} + +/* Emits the `xvstx xd, j, k` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvstx(TCGContext *s, TCGReg xd, TCGReg j, TCGReg k) +{ + tcg_out32(s, encode_xdjk_insn(OPC_XVSTX, xd, j, k)); +} + /* Emits the `dbar ud15` instruction. */ static void __attribute__((unused)) tcg_out_opc_dbar(TCGContext *s, uint32_t ud15) @@ -2416,6 +2308,20 @@ tcg_out_opc_dbar(TCGContext *s, uint32_t ud15) tcg_out32(s, encode_ud15_insn(OPC_DBAR, ud15)); } +/* Emits the `jiscr0 sd5k16` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_jiscr0(TCGContext *s, int32_t sd5k16) +{ + tcg_out32(s, encode_sd5k16_insn(OPC_JISCR0, sd5k16)); +} + +/* Emits the `jiscr1 sd5k16` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_jiscr1(TCGContext *s, int32_t sd5k16) +{ + tcg_out32(s, encode_sd5k16_insn(OPC_JISCR1, sd5k16)); +} + /* Emits the `jirl d, j, sk16` instruction. */ static void __attribute__((unused)) tcg_out_opc_jirl(TCGContext *s, TCGReg d, TCGReg j, int32_t sk16) @@ -2675,286 +2581,6 @@ tcg_out_opc_vsub_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) tcg_out32(s, encode_vdvjvk_insn(OPC_VSUB_D, vd, vj, vk)); } -/* Emits the `vaddwev.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_H_B, vd, vj, vk)); -} - -/* Emits the `vaddwev.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_W_H, vd, vj, vk)); -} - -/* Emits the `vaddwev.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_D_W, vd, vj, vk)); -} - -/* Emits the `vaddwev.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_Q_D, vd, vj, vk)); -} - -/* Emits the `vsubwev.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwev_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWEV_H_B, vd, vj, vk)); -} - -/* Emits the `vsubwev.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwev_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWEV_W_H, vd, vj, vk)); -} - -/* Emits the `vsubwev.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwev_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWEV_D_W, vd, vj, vk)); -} - -/* Emits the `vsubwev.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwev_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWEV_Q_D, vd, vj, vk)); -} - -/* Emits the `vaddwod.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_H_B, vd, vj, vk)); -} - -/* Emits the `vaddwod.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_W_H, vd, vj, vk)); -} - -/* Emits the `vaddwod.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_D_W, vd, vj, vk)); -} - -/* Emits the `vaddwod.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_Q_D, vd, vj, vk)); -} - -/* Emits the `vsubwod.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwod_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWOD_H_B, vd, vj, vk)); -} - -/* Emits the `vsubwod.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwod_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWOD_W_H, vd, vj, vk)); -} - -/* Emits the `vsubwod.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwod_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWOD_D_W, vd, vj, vk)); -} - -/* Emits the `vsubwod.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwod_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWOD_Q_D, vd, vj, vk)); -} - -/* Emits the `vaddwev.h.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_h_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_H_BU, vd, vj, vk)); -} - -/* Emits the `vaddwev.w.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_w_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_W_HU, vd, vj, vk)); -} - -/* Emits the `vaddwev.d.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_d_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_D_WU, vd, vj, vk)); -} - -/* Emits the `vaddwev.q.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_q_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_Q_DU, vd, vj, vk)); -} - -/* Emits the `vsubwev.h.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwev_h_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWEV_H_BU, vd, vj, vk)); -} - -/* Emits the `vsubwev.w.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwev_w_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWEV_W_HU, vd, vj, vk)); -} - -/* Emits the `vsubwev.d.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwev_d_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWEV_D_WU, vd, vj, vk)); -} - -/* Emits the `vsubwev.q.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwev_q_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWEV_Q_DU, vd, vj, vk)); -} - -/* Emits the `vaddwod.h.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_h_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_H_BU, vd, vj, vk)); -} - -/* Emits the `vaddwod.w.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_w_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_W_HU, vd, vj, vk)); -} - -/* Emits the `vaddwod.d.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_d_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_D_WU, vd, vj, vk)); -} - -/* Emits the `vaddwod.q.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_q_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_Q_DU, vd, vj, vk)); -} - -/* Emits the `vsubwod.h.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwod_h_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWOD_H_BU, vd, vj, vk)); -} - -/* Emits the `vsubwod.w.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwod_w_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWOD_W_HU, vd, vj, vk)); -} - -/* Emits the `vsubwod.d.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwod_d_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWOD_D_WU, vd, vj, vk)); -} - -/* Emits the `vsubwod.q.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsubwod_q_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUBWOD_Q_DU, vd, vj, vk)); -} - -/* Emits the `vaddwev.h.bu.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_h_bu_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_H_BU_B, vd, vj, vk)); -} - -/* Emits the `vaddwev.w.hu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_w_hu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_W_HU_H, vd, vj, vk)); -} - -/* Emits the `vaddwev.d.wu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_d_wu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_D_WU_W, vd, vj, vk)); -} - -/* Emits the `vaddwev.q.du.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwev_q_du_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWEV_Q_DU_D, vd, vj, vk)); -} - -/* Emits the `vaddwod.h.bu.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_h_bu_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_H_BU_B, vd, vj, vk)); -} - -/* Emits the `vaddwod.w.hu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_w_hu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_W_HU_H, vd, vj, vk)); -} - -/* Emits the `vaddwod.d.wu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_d_wu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_D_WU_W, vd, vj, vk)); -} - -/* Emits the `vaddwod.q.du.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vaddwod_q_du_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDWOD_Q_DU_D, vd, vj, vk)); -} - /* Emits the `vsadd.b vd, vj, vk` instruction. */ static void __attribute__((unused)) tcg_out_opc_vsadd_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) @@ -3067,314 +2693,6 @@ tcg_out_opc_vssub_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) tcg_out32(s, encode_vdvjvk_insn(OPC_VSSUB_DU, vd, vj, vk)); } -/* Emits the `vhaddw.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhaddw_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHADDW_H_B, vd, vj, vk)); -} - -/* Emits the `vhaddw.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhaddw_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHADDW_W_H, vd, vj, vk)); -} - -/* Emits the `vhaddw.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhaddw_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHADDW_D_W, vd, vj, vk)); -} - -/* Emits the `vhaddw.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhaddw_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHADDW_Q_D, vd, vj, vk)); -} - -/* Emits the `vhsubw.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhsubw_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHSUBW_H_B, vd, vj, vk)); -} - -/* Emits the `vhsubw.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhsubw_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHSUBW_W_H, vd, vj, vk)); -} - -/* Emits the `vhsubw.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhsubw_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHSUBW_D_W, vd, vj, vk)); -} - -/* Emits the `vhsubw.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhsubw_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHSUBW_Q_D, vd, vj, vk)); -} - -/* Emits the `vhaddw.hu.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhaddw_hu_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHADDW_HU_BU, vd, vj, vk)); -} - -/* Emits the `vhaddw.wu.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhaddw_wu_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHADDW_WU_HU, vd, vj, vk)); -} - -/* Emits the `vhaddw.du.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhaddw_du_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHADDW_DU_WU, vd, vj, vk)); -} - -/* Emits the `vhaddw.qu.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhaddw_qu_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHADDW_QU_DU, vd, vj, vk)); -} - -/* Emits the `vhsubw.hu.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhsubw_hu_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHSUBW_HU_BU, vd, vj, vk)); -} - -/* Emits the `vhsubw.wu.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhsubw_wu_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHSUBW_WU_HU, vd, vj, vk)); -} - -/* Emits the `vhsubw.du.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhsubw_du_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHSUBW_DU_WU, vd, vj, vk)); -} - -/* Emits the `vhsubw.qu.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vhsubw_qu_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VHSUBW_QU_DU, vd, vj, vk)); -} - -/* Emits the `vadda.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vadda_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDA_B, vd, vj, vk)); -} - -/* Emits the `vadda.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vadda_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDA_H, vd, vj, vk)); -} - -/* Emits the `vadda.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vadda_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDA_W, vd, vj, vk)); -} - -/* Emits the `vadda.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vadda_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADDA_D, vd, vj, vk)); -} - -/* Emits the `vabsd.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vabsd_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VABSD_B, vd, vj, vk)); -} - -/* Emits the `vabsd.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vabsd_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VABSD_H, vd, vj, vk)); -} - -/* Emits the `vabsd.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vabsd_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VABSD_W, vd, vj, vk)); -} - -/* Emits the `vabsd.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vabsd_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VABSD_D, vd, vj, vk)); -} - -/* Emits the `vabsd.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vabsd_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VABSD_BU, vd, vj, vk)); -} - -/* Emits the `vabsd.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vabsd_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VABSD_HU, vd, vj, vk)); -} - -/* Emits the `vabsd.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vabsd_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VABSD_WU, vd, vj, vk)); -} - -/* Emits the `vabsd.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vabsd_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VABSD_DU, vd, vj, vk)); -} - -/* Emits the `vavg.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavg_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVG_B, vd, vj, vk)); -} - -/* Emits the `vavg.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavg_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVG_H, vd, vj, vk)); -} - -/* Emits the `vavg.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavg_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVG_W, vd, vj, vk)); -} - -/* Emits the `vavg.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavg_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVG_D, vd, vj, vk)); -} - -/* Emits the `vavg.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavg_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVG_BU, vd, vj, vk)); -} - -/* Emits the `vavg.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavg_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVG_HU, vd, vj, vk)); -} - -/* Emits the `vavg.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavg_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVG_WU, vd, vj, vk)); -} - -/* Emits the `vavg.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavg_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVG_DU, vd, vj, vk)); -} - -/* Emits the `vavgr.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavgr_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVGR_B, vd, vj, vk)); -} - -/* Emits the `vavgr.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavgr_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVGR_H, vd, vj, vk)); -} - -/* Emits the `vavgr.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavgr_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVGR_W, vd, vj, vk)); -} - -/* Emits the `vavgr.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavgr_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVGR_D, vd, vj, vk)); -} - -/* Emits the `vavgr.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavgr_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVGR_BU, vd, vj, vk)); -} - -/* Emits the `vavgr.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavgr_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVGR_HU, vd, vj, vk)); -} - -/* Emits the `vavgr.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavgr_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVGR_WU, vd, vj, vk)); -} - -/* Emits the `vavgr.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vavgr_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VAVGR_DU, vd, vj, vk)); -} - /* Emits the `vmax.b vd, vj, vk` instruction. */ static void __attribute__((unused)) tcg_out_opc_vmax_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) @@ -3515,566 +2833,6 @@ tcg_out_opc_vmul_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) tcg_out32(s, encode_vdvjvk_insn(OPC_VMUL_D, vd, vj, vk)); } -/* Emits the `vmuh.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmuh_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMUH_B, vd, vj, vk)); -} - -/* Emits the `vmuh.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmuh_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMUH_H, vd, vj, vk)); -} - -/* Emits the `vmuh.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmuh_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMUH_W, vd, vj, vk)); -} - -/* Emits the `vmuh.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmuh_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMUH_D, vd, vj, vk)); -} - -/* Emits the `vmuh.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmuh_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMUH_BU, vd, vj, vk)); -} - -/* Emits the `vmuh.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmuh_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMUH_HU, vd, vj, vk)); -} - -/* Emits the `vmuh.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmuh_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMUH_WU, vd, vj, vk)); -} - -/* Emits the `vmuh.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmuh_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMUH_DU, vd, vj, vk)); -} - -/* Emits the `vmulwev.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_H_B, vd, vj, vk)); -} - -/* Emits the `vmulwev.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_W_H, vd, vj, vk)); -} - -/* Emits the `vmulwev.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_D_W, vd, vj, vk)); -} - -/* Emits the `vmulwev.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_Q_D, vd, vj, vk)); -} - -/* Emits the `vmulwod.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_H_B, vd, vj, vk)); -} - -/* Emits the `vmulwod.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_W_H, vd, vj, vk)); -} - -/* Emits the `vmulwod.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_D_W, vd, vj, vk)); -} - -/* Emits the `vmulwod.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_Q_D, vd, vj, vk)); -} - -/* Emits the `vmulwev.h.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_h_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_H_BU, vd, vj, vk)); -} - -/* Emits the `vmulwev.w.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_w_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_W_HU, vd, vj, vk)); -} - -/* Emits the `vmulwev.d.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_d_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_D_WU, vd, vj, vk)); -} - -/* Emits the `vmulwev.q.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_q_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_Q_DU, vd, vj, vk)); -} - -/* Emits the `vmulwod.h.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_h_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_H_BU, vd, vj, vk)); -} - -/* Emits the `vmulwod.w.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_w_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_W_HU, vd, vj, vk)); -} - -/* Emits the `vmulwod.d.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_d_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_D_WU, vd, vj, vk)); -} - -/* Emits the `vmulwod.q.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_q_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_Q_DU, vd, vj, vk)); -} - -/* Emits the `vmulwev.h.bu.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_h_bu_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_H_BU_B, vd, vj, vk)); -} - -/* Emits the `vmulwev.w.hu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_w_hu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_W_HU_H, vd, vj, vk)); -} - -/* Emits the `vmulwev.d.wu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_d_wu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_D_WU_W, vd, vj, vk)); -} - -/* Emits the `vmulwev.q.du.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwev_q_du_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWEV_Q_DU_D, vd, vj, vk)); -} - -/* Emits the `vmulwod.h.bu.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_h_bu_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_H_BU_B, vd, vj, vk)); -} - -/* Emits the `vmulwod.w.hu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_w_hu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_W_HU_H, vd, vj, vk)); -} - -/* Emits the `vmulwod.d.wu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_d_wu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_D_WU_W, vd, vj, vk)); -} - -/* Emits the `vmulwod.q.du.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmulwod_q_du_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMULWOD_Q_DU_D, vd, vj, vk)); -} - -/* Emits the `vmadd.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmadd_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADD_B, vd, vj, vk)); -} - -/* Emits the `vmadd.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmadd_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADD_H, vd, vj, vk)); -} - -/* Emits the `vmadd.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmadd_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADD_W, vd, vj, vk)); -} - -/* Emits the `vmadd.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmadd_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADD_D, vd, vj, vk)); -} - -/* Emits the `vmsub.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmsub_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMSUB_B, vd, vj, vk)); -} - -/* Emits the `vmsub.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmsub_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMSUB_H, vd, vj, vk)); -} - -/* Emits the `vmsub.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmsub_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMSUB_W, vd, vj, vk)); -} - -/* Emits the `vmsub.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmsub_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMSUB_D, vd, vj, vk)); -} - -/* Emits the `vmaddwev.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_H_B, vd, vj, vk)); -} - -/* Emits the `vmaddwev.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_W_H, vd, vj, vk)); -} - -/* Emits the `vmaddwev.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_D_W, vd, vj, vk)); -} - -/* Emits the `vmaddwev.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_Q_D, vd, vj, vk)); -} - -/* Emits the `vmaddwod.h.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_h_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_H_B, vd, vj, vk)); -} - -/* Emits the `vmaddwod.w.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_w_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_W_H, vd, vj, vk)); -} - -/* Emits the `vmaddwod.d.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_d_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_D_W, vd, vj, vk)); -} - -/* Emits the `vmaddwod.q.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_q_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_Q_D, vd, vj, vk)); -} - -/* Emits the `vmaddwev.h.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_h_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_H_BU, vd, vj, vk)); -} - -/* Emits the `vmaddwev.w.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_w_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_W_HU, vd, vj, vk)); -} - -/* Emits the `vmaddwev.d.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_d_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_D_WU, vd, vj, vk)); -} - -/* Emits the `vmaddwev.q.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_q_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_Q_DU, vd, vj, vk)); -} - -/* Emits the `vmaddwod.h.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_h_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_H_BU, vd, vj, vk)); -} - -/* Emits the `vmaddwod.w.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_w_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_W_HU, vd, vj, vk)); -} - -/* Emits the `vmaddwod.d.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_d_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_D_WU, vd, vj, vk)); -} - -/* Emits the `vmaddwod.q.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_q_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_Q_DU, vd, vj, vk)); -} - -/* Emits the `vmaddwev.h.bu.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_h_bu_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_H_BU_B, vd, vj, vk)); -} - -/* Emits the `vmaddwev.w.hu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_w_hu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_W_HU_H, vd, vj, vk)); -} - -/* Emits the `vmaddwev.d.wu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_d_wu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_D_WU_W, vd, vj, vk)); -} - -/* Emits the `vmaddwev.q.du.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwev_q_du_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWEV_Q_DU_D, vd, vj, vk)); -} - -/* Emits the `vmaddwod.h.bu.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_h_bu_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_H_BU_B, vd, vj, vk)); -} - -/* Emits the `vmaddwod.w.hu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_w_hu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_W_HU_H, vd, vj, vk)); -} - -/* Emits the `vmaddwod.d.wu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_d_wu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_D_WU_W, vd, vj, vk)); -} - -/* Emits the `vmaddwod.q.du.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmaddwod_q_du_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMADDWOD_Q_DU_D, vd, vj, vk)); -} - -/* Emits the `vdiv.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vdiv_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VDIV_B, vd, vj, vk)); -} - -/* Emits the `vdiv.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vdiv_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VDIV_H, vd, vj, vk)); -} - -/* Emits the `vdiv.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vdiv_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VDIV_W, vd, vj, vk)); -} - -/* Emits the `vdiv.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vdiv_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VDIV_D, vd, vj, vk)); -} - -/* Emits the `vmod.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmod_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMOD_B, vd, vj, vk)); -} - -/* Emits the `vmod.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmod_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMOD_H, vd, vj, vk)); -} - -/* Emits the `vmod.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmod_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMOD_W, vd, vj, vk)); -} - -/* Emits the `vmod.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmod_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMOD_D, vd, vj, vk)); -} - -/* Emits the `vdiv.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vdiv_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VDIV_BU, vd, vj, vk)); -} - -/* Emits the `vdiv.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vdiv_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VDIV_HU, vd, vj, vk)); -} - -/* Emits the `vdiv.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vdiv_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VDIV_WU, vd, vj, vk)); -} - -/* Emits the `vdiv.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vdiv_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VDIV_DU, vd, vj, vk)); -} - -/* Emits the `vmod.bu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmod_bu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMOD_BU, vd, vj, vk)); -} - -/* Emits the `vmod.hu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmod_hu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMOD_HU, vd, vj, vk)); -} - -/* Emits the `vmod.wu vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmod_wu(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMOD_WU, vd, vj, vk)); -} - -/* Emits the `vmod.du vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vmod_du(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VMOD_DU, vd, vj, vk)); -} - /* Emits the `vsll.b vd, vj, vk` instruction. */ static void __attribute__((unused)) tcg_out_opc_vsll_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) @@ -4187,566 +2945,6 @@ tcg_out_opc_vrotr_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) tcg_out32(s, encode_vdvjvk_insn(OPC_VROTR_D, vd, vj, vk)); } -/* Emits the `vsrlr.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrlr_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLR_B, vd, vj, vk)); -} - -/* Emits the `vsrlr.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrlr_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLR_H, vd, vj, vk)); -} - -/* Emits the `vsrlr.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrlr_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLR_W, vd, vj, vk)); -} - -/* Emits the `vsrlr.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrlr_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLR_D, vd, vj, vk)); -} - -/* Emits the `vsrar.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrar_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRAR_B, vd, vj, vk)); -} - -/* Emits the `vsrar.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrar_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRAR_H, vd, vj, vk)); -} - -/* Emits the `vsrar.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrar_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRAR_W, vd, vj, vk)); -} - -/* Emits the `vsrar.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrar_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRAR_D, vd, vj, vk)); -} - -/* Emits the `vsrln.b.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrln_b_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLN_B_H, vd, vj, vk)); -} - -/* Emits the `vsrln.h.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrln_h_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLN_H_W, vd, vj, vk)); -} - -/* Emits the `vsrln.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrln_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLN_W_D, vd, vj, vk)); -} - -/* Emits the `vsran.b.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsran_b_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRAN_B_H, vd, vj, vk)); -} - -/* Emits the `vsran.h.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsran_h_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRAN_H_W, vd, vj, vk)); -} - -/* Emits the `vsran.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsran_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRAN_W_D, vd, vj, vk)); -} - -/* Emits the `vsrlrn.b.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrlrn_b_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLRN_B_H, vd, vj, vk)); -} - -/* Emits the `vsrlrn.h.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrlrn_h_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLRN_H_W, vd, vj, vk)); -} - -/* Emits the `vsrlrn.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrlrn_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRLRN_W_D, vd, vj, vk)); -} - -/* Emits the `vsrarn.b.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrarn_b_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRARN_B_H, vd, vj, vk)); -} - -/* Emits the `vsrarn.h.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrarn_h_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRARN_H_W, vd, vj, vk)); -} - -/* Emits the `vsrarn.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsrarn_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSRARN_W_D, vd, vj, vk)); -} - -/* Emits the `vssrln.b.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrln_b_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLN_B_H, vd, vj, vk)); -} - -/* Emits the `vssrln.h.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrln_h_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLN_H_W, vd, vj, vk)); -} - -/* Emits the `vssrln.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrln_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLN_W_D, vd, vj, vk)); -} - -/* Emits the `vssran.b.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssran_b_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRAN_B_H, vd, vj, vk)); -} - -/* Emits the `vssran.h.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssran_h_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRAN_H_W, vd, vj, vk)); -} - -/* Emits the `vssran.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssran_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRAN_W_D, vd, vj, vk)); -} - -/* Emits the `vssrlrn.b.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrlrn_b_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLRN_B_H, vd, vj, vk)); -} - -/* Emits the `vssrlrn.h.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrlrn_h_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLRN_H_W, vd, vj, vk)); -} - -/* Emits the `vssrlrn.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrlrn_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLRN_W_D, vd, vj, vk)); -} - -/* Emits the `vssrarn.b.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrarn_b_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRARN_B_H, vd, vj, vk)); -} - -/* Emits the `vssrarn.h.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrarn_h_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRARN_H_W, vd, vj, vk)); -} - -/* Emits the `vssrarn.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrarn_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRARN_W_D, vd, vj, vk)); -} - -/* Emits the `vssrln.bu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrln_bu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLN_BU_H, vd, vj, vk)); -} - -/* Emits the `vssrln.hu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrln_hu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLN_HU_W, vd, vj, vk)); -} - -/* Emits the `vssrln.wu.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrln_wu_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLN_WU_D, vd, vj, vk)); -} - -/* Emits the `vssran.bu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssran_bu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRAN_BU_H, vd, vj, vk)); -} - -/* Emits the `vssran.hu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssran_hu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRAN_HU_W, vd, vj, vk)); -} - -/* Emits the `vssran.wu.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssran_wu_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRAN_WU_D, vd, vj, vk)); -} - -/* Emits the `vssrlrn.bu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrlrn_bu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLRN_BU_H, vd, vj, vk)); -} - -/* Emits the `vssrlrn.hu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrlrn_hu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLRN_HU_W, vd, vj, vk)); -} - -/* Emits the `vssrlrn.wu.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrlrn_wu_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRLRN_WU_D, vd, vj, vk)); -} - -/* Emits the `vssrarn.bu.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrarn_bu_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRARN_BU_H, vd, vj, vk)); -} - -/* Emits the `vssrarn.hu.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrarn_hu_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRARN_HU_W, vd, vj, vk)); -} - -/* Emits the `vssrarn.wu.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vssrarn_wu_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSSRARN_WU_D, vd, vj, vk)); -} - -/* Emits the `vbitclr.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitclr_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITCLR_B, vd, vj, vk)); -} - -/* Emits the `vbitclr.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitclr_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITCLR_H, vd, vj, vk)); -} - -/* Emits the `vbitclr.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitclr_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITCLR_W, vd, vj, vk)); -} - -/* Emits the `vbitclr.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitclr_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITCLR_D, vd, vj, vk)); -} - -/* Emits the `vbitset.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitset_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITSET_B, vd, vj, vk)); -} - -/* Emits the `vbitset.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitset_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITSET_H, vd, vj, vk)); -} - -/* Emits the `vbitset.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitset_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITSET_W, vd, vj, vk)); -} - -/* Emits the `vbitset.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitset_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITSET_D, vd, vj, vk)); -} - -/* Emits the `vbitrev.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitrev_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITREV_B, vd, vj, vk)); -} - -/* Emits the `vbitrev.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitrev_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITREV_H, vd, vj, vk)); -} - -/* Emits the `vbitrev.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitrev_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITREV_W, vd, vj, vk)); -} - -/* Emits the `vbitrev.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbitrev_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VBITREV_D, vd, vj, vk)); -} - -/* Emits the `vpackev.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpackev_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPACKEV_B, vd, vj, vk)); -} - -/* Emits the `vpackev.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpackev_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPACKEV_H, vd, vj, vk)); -} - -/* Emits the `vpackev.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpackev_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPACKEV_W, vd, vj, vk)); -} - -/* Emits the `vpackev.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpackev_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPACKEV_D, vd, vj, vk)); -} - -/* Emits the `vpackod.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpackod_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPACKOD_B, vd, vj, vk)); -} - -/* Emits the `vpackod.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpackod_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPACKOD_H, vd, vj, vk)); -} - -/* Emits the `vpackod.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpackod_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPACKOD_W, vd, vj, vk)); -} - -/* Emits the `vpackod.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpackod_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPACKOD_D, vd, vj, vk)); -} - -/* Emits the `vilvl.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vilvl_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VILVL_B, vd, vj, vk)); -} - -/* Emits the `vilvl.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vilvl_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VILVL_H, vd, vj, vk)); -} - -/* Emits the `vilvl.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vilvl_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VILVL_W, vd, vj, vk)); -} - -/* Emits the `vilvl.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vilvl_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VILVL_D, vd, vj, vk)); -} - -/* Emits the `vilvh.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vilvh_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VILVH_B, vd, vj, vk)); -} - -/* Emits the `vilvh.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vilvh_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VILVH_H, vd, vj, vk)); -} - -/* Emits the `vilvh.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vilvh_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VILVH_W, vd, vj, vk)); -} - -/* Emits the `vilvh.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vilvh_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VILVH_D, vd, vj, vk)); -} - -/* Emits the `vpickev.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpickev_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPICKEV_B, vd, vj, vk)); -} - -/* Emits the `vpickev.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpickev_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPICKEV_H, vd, vj, vk)); -} - -/* Emits the `vpickev.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpickev_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPICKEV_W, vd, vj, vk)); -} - -/* Emits the `vpickev.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpickev_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPICKEV_D, vd, vj, vk)); -} - -/* Emits the `vpickod.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpickod_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPICKOD_B, vd, vj, vk)); -} - -/* Emits the `vpickod.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpickod_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPICKOD_H, vd, vj, vk)); -} - -/* Emits the `vpickod.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpickod_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPICKOD_W, vd, vj, vk)); -} - -/* Emits the `vpickod.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vpickod_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VPICKOD_D, vd, vj, vk)); -} - /* Emits the `vreplve.b vd, vj, k` instruction. */ static void __attribute__((unused)) tcg_out_opc_vreplve_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg k) @@ -4817,251 +3015,6 @@ tcg_out_opc_vorn_v(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) tcg_out32(s, encode_vdvjvk_insn(OPC_VORN_V, vd, vj, vk)); } -/* Emits the `vfrstp.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfrstp_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFRSTP_B, vd, vj, vk)); -} - -/* Emits the `vfrstp.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfrstp_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFRSTP_H, vd, vj, vk)); -} - -/* Emits the `vadd.q vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vadd_q(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VADD_Q, vd, vj, vk)); -} - -/* Emits the `vsub.q vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsub_q(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSUB_Q, vd, vj, vk)); -} - -/* Emits the `vsigncov.b vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsigncov_b(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSIGNCOV_B, vd, vj, vk)); -} - -/* Emits the `vsigncov.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsigncov_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSIGNCOV_H, vd, vj, vk)); -} - -/* Emits the `vsigncov.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsigncov_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSIGNCOV_W, vd, vj, vk)); -} - -/* Emits the `vsigncov.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vsigncov_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSIGNCOV_D, vd, vj, vk)); -} - -/* Emits the `vfadd.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfadd_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFADD_S, vd, vj, vk)); -} - -/* Emits the `vfadd.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfadd_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFADD_D, vd, vj, vk)); -} - -/* Emits the `vfsub.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfsub_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFSUB_S, vd, vj, vk)); -} - -/* Emits the `vfsub.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfsub_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFSUB_D, vd, vj, vk)); -} - -/* Emits the `vfmul.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmul_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMUL_S, vd, vj, vk)); -} - -/* Emits the `vfmul.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmul_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMUL_D, vd, vj, vk)); -} - -/* Emits the `vfdiv.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfdiv_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFDIV_S, vd, vj, vk)); -} - -/* Emits the `vfdiv.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfdiv_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFDIV_D, vd, vj, vk)); -} - -/* Emits the `vfmax.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmax_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMAX_S, vd, vj, vk)); -} - -/* Emits the `vfmax.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmax_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMAX_D, vd, vj, vk)); -} - -/* Emits the `vfmin.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmin_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMIN_S, vd, vj, vk)); -} - -/* Emits the `vfmin.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmin_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMIN_D, vd, vj, vk)); -} - -/* Emits the `vfmaxa.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmaxa_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMAXA_S, vd, vj, vk)); -} - -/* Emits the `vfmaxa.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmaxa_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMAXA_D, vd, vj, vk)); -} - -/* Emits the `vfmina.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmina_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMINA_S, vd, vj, vk)); -} - -/* Emits the `vfmina.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfmina_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFMINA_D, vd, vj, vk)); -} - -/* Emits the `vfcvt.h.s vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcvt_h_s(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCVT_H_S, vd, vj, vk)); -} - -/* Emits the `vfcvt.s.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vfcvt_s_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFCVT_S_D, vd, vj, vk)); -} - -/* Emits the `vffint.s.l vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vffint_s_l(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFFINT_S_L, vd, vj, vk)); -} - -/* Emits the `vftint.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vftint_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFTINT_W_D, vd, vj, vk)); -} - -/* Emits the `vftintrm.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vftintrm_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFTINTRM_W_D, vd, vj, vk)); -} - -/* Emits the `vftintrp.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vftintrp_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFTINTRP_W_D, vd, vj, vk)); -} - -/* Emits the `vftintrz.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vftintrz_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFTINTRZ_W_D, vd, vj, vk)); -} - -/* Emits the `vftintrne.w.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vftintrne_w_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VFTINTRNE_W_D, vd, vj, vk)); -} - -/* Emits the `vshuf.h vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vshuf_h(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSHUF_H, vd, vj, vk)); -} - -/* Emits the `vshuf.w vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vshuf_w(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSHUF_W, vd, vj, vk)); -} - -/* Emits the `vshuf.d vd, vj, vk` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vshuf_d(TCGContext *s, TCGReg vd, TCGReg vj, TCGReg vk) -{ - tcg_out32(s, encode_vdvjvk_insn(OPC_VSHUF_D, vd, vj, vk)); -} - /* Emits the `vseqi.b vd, vj, sk5` instruction. */ static void __attribute__((unused)) tcg_out_opc_vseqi_b(TCGContext *s, TCGReg vd, TCGReg vj, int32_t sk5) @@ -5258,20 +3211,6 @@ tcg_out_opc_vsubi_du(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) tcg_out32(s, encode_vdvjuk5_insn(OPC_VSUBI_DU, vd, vj, uk5)); } -/* Emits the `vbsll.v vd, vj, uk5` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbsll_v(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) -{ - tcg_out32(s, encode_vdvjuk5_insn(OPC_VBSLL_V, vd, vj, uk5)); -} - -/* Emits the `vbsrl.v vd, vj, uk5` instruction. */ -static void __attribute__((unused)) -tcg_out_opc_vbsrl_v(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) -{ - tcg_out32(s, encode_vdvjuk5_insn(OPC_VBSRL_V, vd, vj, uk5)); -} - /* Emits the `vmaxi.b vd, vj, sk5` instruction. */ static void __attribute__((unused)) tcg_out_opc_vmaxi_b(TCGContext *s, TCGReg vd, TCGReg vj, int32_t sk5) @@ -5384,1621 +3323,1747 @@ tcg_out_opc_vmini_du(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) tcg_out32(s, encode_vdvjuk5_insn(OPC_VMINI_DU, vd, vj, uk5)); } -/* Emits the `vfrstpi.b vd, vj, uk5` instruction. */ +/* Emits the `vneg.b vd, vj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrstpi_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_vneg_b(TCGContext *s, TCGReg vd, TCGReg vj) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VFRSTPI_B, vd, vj, uk5)); + tcg_out32(s, encode_vdvj_insn(OPC_VNEG_B, vd, vj)); } -/* Emits the `vfrstpi.h vd, vj, uk5` instruction. */ +/* Emits the `vneg.h vd, vj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrstpi_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_vneg_h(TCGContext *s, TCGReg vd, TCGReg vj) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VFRSTPI_H, vd, vj, uk5)); + tcg_out32(s, encode_vdvj_insn(OPC_VNEG_H, vd, vj)); } -/* Emits the `vclo.b vd, vj` instruction. */ +/* Emits the `vneg.w vd, vj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vclo_b(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vneg_w(TCGContext *s, TCGReg vd, TCGReg vj) { - tcg_out32(s, encode_vdvj_insn(OPC_VCLO_B, vd, vj)); + tcg_out32(s, encode_vdvj_insn(OPC_VNEG_W, vd, vj)); } -/* Emits the `vclo.h vd, vj` instruction. */ +/* Emits the `vneg.d vd, vj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vclo_h(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vneg_d(TCGContext *s, TCGReg vd, TCGReg vj) { - tcg_out32(s, encode_vdvj_insn(OPC_VCLO_H, vd, vj)); + tcg_out32(s, encode_vdvj_insn(OPC_VNEG_D, vd, vj)); } -/* Emits the `vclo.w vd, vj` instruction. */ +/* Emits the `vreplgr2vr.b vd, j` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vclo_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vreplgr2vr_b(TCGContext *s, TCGReg vd, TCGReg j) { - tcg_out32(s, encode_vdvj_insn(OPC_VCLO_W, vd, vj)); + tcg_out32(s, encode_vdj_insn(OPC_VREPLGR2VR_B, vd, j)); } -/* Emits the `vclo.d vd, vj` instruction. */ +/* Emits the `vreplgr2vr.h vd, j` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vclo_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vreplgr2vr_h(TCGContext *s, TCGReg vd, TCGReg j) { - tcg_out32(s, encode_vdvj_insn(OPC_VCLO_D, vd, vj)); + tcg_out32(s, encode_vdj_insn(OPC_VREPLGR2VR_H, vd, j)); } -/* Emits the `vclz.b vd, vj` instruction. */ +/* Emits the `vreplgr2vr.w vd, j` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vclz_b(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vreplgr2vr_w(TCGContext *s, TCGReg vd, TCGReg j) { - tcg_out32(s, encode_vdvj_insn(OPC_VCLZ_B, vd, vj)); + tcg_out32(s, encode_vdj_insn(OPC_VREPLGR2VR_W, vd, j)); } -/* Emits the `vclz.h vd, vj` instruction. */ +/* Emits the `vreplgr2vr.d vd, j` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vclz_h(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vreplgr2vr_d(TCGContext *s, TCGReg vd, TCGReg j) { - tcg_out32(s, encode_vdvj_insn(OPC_VCLZ_H, vd, vj)); + tcg_out32(s, encode_vdj_insn(OPC_VREPLGR2VR_D, vd, j)); } -/* Emits the `vclz.w vd, vj` instruction. */ +/* Emits the `vrotri.b vd, vj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vclz_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vrotri_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) { - tcg_out32(s, encode_vdvj_insn(OPC_VCLZ_W, vd, vj)); + tcg_out32(s, encode_vdvjuk3_insn(OPC_VROTRI_B, vd, vj, uk3)); } -/* Emits the `vclz.d vd, vj` instruction. */ +/* Emits the `vrotri.h vd, vj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vclz_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vrotri_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) { - tcg_out32(s, encode_vdvj_insn(OPC_VCLZ_D, vd, vj)); + tcg_out32(s, encode_vdvjuk4_insn(OPC_VROTRI_H, vd, vj, uk4)); } -/* Emits the `vpcnt.b vd, vj` instruction. */ +/* Emits the `vrotri.w vd, vj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpcnt_b(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vrotri_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) { - tcg_out32(s, encode_vdvj_insn(OPC_VPCNT_B, vd, vj)); + tcg_out32(s, encode_vdvjuk5_insn(OPC_VROTRI_W, vd, vj, uk5)); } -/* Emits the `vpcnt.h vd, vj` instruction. */ +/* Emits the `vrotri.d vd, vj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpcnt_h(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vrotri_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) { - tcg_out32(s, encode_vdvj_insn(OPC_VPCNT_H, vd, vj)); + tcg_out32(s, encode_vdvjuk6_insn(OPC_VROTRI_D, vd, vj, uk6)); } -/* Emits the `vpcnt.w vd, vj` instruction. */ +/* Emits the `vinsgr2vr.b vd, j, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpcnt_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vinsgr2vr_b(TCGContext *s, TCGReg vd, TCGReg j, uint32_t uk4) { - tcg_out32(s, encode_vdvj_insn(OPC_VPCNT_W, vd, vj)); + tcg_out32(s, encode_vdjuk4_insn(OPC_VINSGR2VR_B, vd, j, uk4)); } -/* Emits the `vpcnt.d vd, vj` instruction. */ +/* Emits the `vinsgr2vr.h vd, j, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpcnt_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vinsgr2vr_h(TCGContext *s, TCGReg vd, TCGReg j, uint32_t uk3) { - tcg_out32(s, encode_vdvj_insn(OPC_VPCNT_D, vd, vj)); + tcg_out32(s, encode_vdjuk3_insn(OPC_VINSGR2VR_H, vd, j, uk3)); } -/* Emits the `vneg.b vd, vj` instruction. */ +/* Emits the `vinsgr2vr.w vd, j, uk2` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vneg_b(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vinsgr2vr_w(TCGContext *s, TCGReg vd, TCGReg j, uint32_t uk2) { - tcg_out32(s, encode_vdvj_insn(OPC_VNEG_B, vd, vj)); + tcg_out32(s, encode_vdjuk2_insn(OPC_VINSGR2VR_W, vd, j, uk2)); } -/* Emits the `vneg.h vd, vj` instruction. */ +/* Emits the `vinsgr2vr.d vd, j, uk1` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vneg_h(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vinsgr2vr_d(TCGContext *s, TCGReg vd, TCGReg j, uint32_t uk1) { - tcg_out32(s, encode_vdvj_insn(OPC_VNEG_H, vd, vj)); + tcg_out32(s, encode_vdjuk1_insn(OPC_VINSGR2VR_D, vd, j, uk1)); } -/* Emits the `vneg.w vd, vj` instruction. */ +/* Emits the `vpickve2gr.b d, vj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vneg_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vpickve2gr_b(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk4) { - tcg_out32(s, encode_vdvj_insn(OPC_VNEG_W, vd, vj)); + tcg_out32(s, encode_dvjuk4_insn(OPC_VPICKVE2GR_B, d, vj, uk4)); } -/* Emits the `vneg.d vd, vj` instruction. */ +/* Emits the `vpickve2gr.h d, vj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vneg_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vpickve2gr_h(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk3) { - tcg_out32(s, encode_vdvj_insn(OPC_VNEG_D, vd, vj)); + tcg_out32(s, encode_dvjuk3_insn(OPC_VPICKVE2GR_H, d, vj, uk3)); +} + +/* Emits the `vpickve2gr.w d, vj, uk2` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vpickve2gr_w(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk2) +{ + tcg_out32(s, encode_dvjuk2_insn(OPC_VPICKVE2GR_W, d, vj, uk2)); +} + +/* Emits the `vpickve2gr.d d, vj, uk1` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vpickve2gr_d(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk1) +{ + tcg_out32(s, encode_dvjuk1_insn(OPC_VPICKVE2GR_D, d, vj, uk1)); +} + +/* Emits the `vpickve2gr.bu d, vj, uk4` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vpickve2gr_bu(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk4) +{ + tcg_out32(s, encode_dvjuk4_insn(OPC_VPICKVE2GR_BU, d, vj, uk4)); +} + +/* Emits the `vpickve2gr.hu d, vj, uk3` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vpickve2gr_hu(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk3) +{ + tcg_out32(s, encode_dvjuk3_insn(OPC_VPICKVE2GR_HU, d, vj, uk3)); +} + +/* Emits the `vpickve2gr.wu d, vj, uk2` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vpickve2gr_wu(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk2) +{ + tcg_out32(s, encode_dvjuk2_insn(OPC_VPICKVE2GR_WU, d, vj, uk2)); +} + +/* Emits the `vpickve2gr.du d, vj, uk1` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vpickve2gr_du(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk1) +{ + tcg_out32(s, encode_dvjuk1_insn(OPC_VPICKVE2GR_DU, d, vj, uk1)); +} + +/* Emits the `vreplvei.b vd, vj, uk4` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vreplvei_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +{ + tcg_out32(s, encode_vdvjuk4_insn(OPC_VREPLVEI_B, vd, vj, uk4)); +} + +/* Emits the `vreplvei.h vd, vj, uk3` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vreplvei_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +{ + tcg_out32(s, encode_vdvjuk3_insn(OPC_VREPLVEI_H, vd, vj, uk3)); +} + +/* Emits the `vreplvei.w vd, vj, uk2` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vreplvei_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk2) +{ + tcg_out32(s, encode_vdvjuk2_insn(OPC_VREPLVEI_W, vd, vj, uk2)); +} + +/* Emits the `vreplvei.d vd, vj, uk1` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vreplvei_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk1) +{ + tcg_out32(s, encode_vdvjuk1_insn(OPC_VREPLVEI_D, vd, vj, uk1)); +} + +/* Emits the `vbitclri.b vd, vj, uk3` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vbitclri_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +{ + tcg_out32(s, encode_vdvjuk3_insn(OPC_VBITCLRI_B, vd, vj, uk3)); +} + +/* Emits the `vbitclri.h vd, vj, uk4` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vbitclri_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +{ + tcg_out32(s, encode_vdvjuk4_insn(OPC_VBITCLRI_H, vd, vj, uk4)); +} + +/* Emits the `vbitclri.w vd, vj, uk5` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vbitclri_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +{ + tcg_out32(s, encode_vdvjuk5_insn(OPC_VBITCLRI_W, vd, vj, uk5)); +} + +/* Emits the `vbitclri.d vd, vj, uk6` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_vbitclri_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +{ + tcg_out32(s, encode_vdvjuk6_insn(OPC_VBITCLRI_D, vd, vj, uk6)); } -/* Emits the `vmskltz.b vd, vj` instruction. */ +/* Emits the `vbitseti.b vd, vj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vmskltz_b(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vbitseti_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) { - tcg_out32(s, encode_vdvj_insn(OPC_VMSKLTZ_B, vd, vj)); + tcg_out32(s, encode_vdvjuk3_insn(OPC_VBITSETI_B, vd, vj, uk3)); } -/* Emits the `vmskltz.h vd, vj` instruction. */ +/* Emits the `vbitseti.h vd, vj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vmskltz_h(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vbitseti_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) { - tcg_out32(s, encode_vdvj_insn(OPC_VMSKLTZ_H, vd, vj)); + tcg_out32(s, encode_vdvjuk4_insn(OPC_VBITSETI_H, vd, vj, uk4)); } -/* Emits the `vmskltz.w vd, vj` instruction. */ +/* Emits the `vbitseti.w vd, vj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vmskltz_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vbitseti_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) { - tcg_out32(s, encode_vdvj_insn(OPC_VMSKLTZ_W, vd, vj)); + tcg_out32(s, encode_vdvjuk5_insn(OPC_VBITSETI_W, vd, vj, uk5)); } -/* Emits the `vmskltz.d vd, vj` instruction. */ +/* Emits the `vbitseti.d vd, vj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vmskltz_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vbitseti_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) { - tcg_out32(s, encode_vdvj_insn(OPC_VMSKLTZ_D, vd, vj)); + tcg_out32(s, encode_vdvjuk6_insn(OPC_VBITSETI_D, vd, vj, uk6)); } -/* Emits the `vmskgez.b vd, vj` instruction. */ +/* Emits the `vbitrevi.b vd, vj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vmskgez_b(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vbitrevi_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) { - tcg_out32(s, encode_vdvj_insn(OPC_VMSKGEZ_B, vd, vj)); + tcg_out32(s, encode_vdvjuk3_insn(OPC_VBITREVI_B, vd, vj, uk3)); } -/* Emits the `vmsknz.b vd, vj` instruction. */ +/* Emits the `vbitrevi.h vd, vj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vmsknz_b(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vbitrevi_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) { - tcg_out32(s, encode_vdvj_insn(OPC_VMSKNZ_B, vd, vj)); + tcg_out32(s, encode_vdvjuk4_insn(OPC_VBITREVI_H, vd, vj, uk4)); } -/* Emits the `vseteqz.v cd, vj` instruction. */ +/* Emits the `vbitrevi.w vd, vj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vseteqz_v(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vbitrevi_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETEQZ_V, cd, vj)); + tcg_out32(s, encode_vdvjuk5_insn(OPC_VBITREVI_W, vd, vj, uk5)); } -/* Emits the `vsetnez.v cd, vj` instruction. */ +/* Emits the `vbitrevi.d vd, vj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetnez_v(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vbitrevi_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETNEZ_V, cd, vj)); + tcg_out32(s, encode_vdvjuk6_insn(OPC_VBITREVI_D, vd, vj, uk6)); } -/* Emits the `vsetanyeqz.b cd, vj` instruction. */ +/* Emits the `vslli.b vd, vj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetanyeqz_b(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vslli_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETANYEQZ_B, cd, vj)); + tcg_out32(s, encode_vdvjuk3_insn(OPC_VSLLI_B, vd, vj, uk3)); } -/* Emits the `vsetanyeqz.h cd, vj` instruction. */ +/* Emits the `vslli.h vd, vj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetanyeqz_h(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vslli_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETANYEQZ_H, cd, vj)); + tcg_out32(s, encode_vdvjuk4_insn(OPC_VSLLI_H, vd, vj, uk4)); } -/* Emits the `vsetanyeqz.w cd, vj` instruction. */ +/* Emits the `vslli.w vd, vj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetanyeqz_w(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vslli_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETANYEQZ_W, cd, vj)); + tcg_out32(s, encode_vdvjuk5_insn(OPC_VSLLI_W, vd, vj, uk5)); } -/* Emits the `vsetanyeqz.d cd, vj` instruction. */ +/* Emits the `vslli.d vd, vj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetanyeqz_d(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vslli_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETANYEQZ_D, cd, vj)); + tcg_out32(s, encode_vdvjuk6_insn(OPC_VSLLI_D, vd, vj, uk6)); } -/* Emits the `vsetallnez.b cd, vj` instruction. */ +/* Emits the `vsrli.b vd, vj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetallnez_b(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vsrli_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETALLNEZ_B, cd, vj)); + tcg_out32(s, encode_vdvjuk3_insn(OPC_VSRLI_B, vd, vj, uk3)); } -/* Emits the `vsetallnez.h cd, vj` instruction. */ +/* Emits the `vsrli.h vd, vj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetallnez_h(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vsrli_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETALLNEZ_H, cd, vj)); + tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRLI_H, vd, vj, uk4)); } -/* Emits the `vsetallnez.w cd, vj` instruction. */ +/* Emits the `vsrli.w vd, vj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetallnez_w(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vsrli_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETALLNEZ_W, cd, vj)); + tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRLI_W, vd, vj, uk5)); } -/* Emits the `vsetallnez.d cd, vj` instruction. */ +/* Emits the `vsrli.d vd, vj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsetallnez_d(TCGContext *s, TCGReg cd, TCGReg vj) +tcg_out_opc_vsrli_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) { - tcg_out32(s, encode_cdvj_insn(OPC_VSETALLNEZ_D, cd, vj)); + tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRLI_D, vd, vj, uk6)); } -/* Emits the `vflogb.s vd, vj` instruction. */ +/* Emits the `vsrai.b vd, vj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vflogb_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vsrai_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) { - tcg_out32(s, encode_vdvj_insn(OPC_VFLOGB_S, vd, vj)); + tcg_out32(s, encode_vdvjuk3_insn(OPC_VSRAI_B, vd, vj, uk3)); } -/* Emits the `vflogb.d vd, vj` instruction. */ +/* Emits the `vsrai.h vd, vj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vflogb_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vsrai_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) { - tcg_out32(s, encode_vdvj_insn(OPC_VFLOGB_D, vd, vj)); + tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRAI_H, vd, vj, uk4)); } -/* Emits the `vfclass.s vd, vj` instruction. */ +/* Emits the `vsrai.w vd, vj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfclass_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vsrai_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) { - tcg_out32(s, encode_vdvj_insn(OPC_VFCLASS_S, vd, vj)); + tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRAI_W, vd, vj, uk5)); } -/* Emits the `vfclass.d vd, vj` instruction. */ +/* Emits the `vsrai.d vd, vj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfclass_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vsrai_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) { - tcg_out32(s, encode_vdvj_insn(OPC_VFCLASS_D, vd, vj)); + tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRAI_D, vd, vj, uk6)); } -/* Emits the `vfsqrt.s vd, vj` instruction. */ +/* Emits the `vbitseli.b vd, vj, uk8` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfsqrt_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vbitseli_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) { - tcg_out32(s, encode_vdvj_insn(OPC_VFSQRT_S, vd, vj)); + tcg_out32(s, encode_vdvjuk8_insn(OPC_VBITSELI_B, vd, vj, uk8)); } -/* Emits the `vfsqrt.d vd, vj` instruction. */ +/* Emits the `vandi.b vd, vj, uk8` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfsqrt_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vandi_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) { - tcg_out32(s, encode_vdvj_insn(OPC_VFSQRT_D, vd, vj)); + tcg_out32(s, encode_vdvjuk8_insn(OPC_VANDI_B, vd, vj, uk8)); } -/* Emits the `vfrecip.s vd, vj` instruction. */ +/* Emits the `vori.b vd, vj, uk8` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrecip_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vori_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRECIP_S, vd, vj)); + tcg_out32(s, encode_vdvjuk8_insn(OPC_VORI_B, vd, vj, uk8)); } -/* Emits the `vfrecip.d vd, vj` instruction. */ +/* Emits the `vxori.b vd, vj, uk8` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrecip_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vxori_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRECIP_D, vd, vj)); + tcg_out32(s, encode_vdvjuk8_insn(OPC_VXORI_B, vd, vj, uk8)); } -/* Emits the `vfrsqrt.s vd, vj` instruction. */ +/* Emits the `vnori.b vd, vj, uk8` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrsqrt_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vnori_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRSQRT_S, vd, vj)); + tcg_out32(s, encode_vdvjuk8_insn(OPC_VNORI_B, vd, vj, uk8)); } -/* Emits the `vfrsqrt.d vd, vj` instruction. */ +/* Emits the `vldi vd, sj13` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrsqrt_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_vldi(TCGContext *s, TCGReg vd, int32_t sj13) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRSQRT_D, vd, vj)); + tcg_out32(s, encode_vdsj13_insn(OPC_VLDI, vd, sj13)); } -/* Emits the `vfrint.s vd, vj` instruction. */ +/* Emits the `xvseq.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrint_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvseq_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINT_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSEQ_B, xd, xj, xk)); } -/* Emits the `vfrint.d vd, vj` instruction. */ +/* Emits the `xvseq.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrint_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvseq_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINT_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSEQ_H, xd, xj, xk)); } -/* Emits the `vfrintrm.s vd, vj` instruction. */ +/* Emits the `xvseq.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrintrm_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvseq_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINTRM_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSEQ_W, xd, xj, xk)); } -/* Emits the `vfrintrm.d vd, vj` instruction. */ +/* Emits the `xvseq.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrintrm_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvseq_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINTRM_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSEQ_D, xd, xj, xk)); } -/* Emits the `vfrintrp.s vd, vj` instruction. */ +/* Emits the `xvsle.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrintrp_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsle_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINTRP_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLE_B, xd, xj, xk)); } -/* Emits the `vfrintrp.d vd, vj` instruction. */ +/* Emits the `xvsle.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrintrp_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsle_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINTRP_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLE_H, xd, xj, xk)); } -/* Emits the `vfrintrz.s vd, vj` instruction. */ +/* Emits the `xvsle.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrintrz_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsle_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINTRZ_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLE_W, xd, xj, xk)); } -/* Emits the `vfrintrz.d vd, vj` instruction. */ +/* Emits the `xvsle.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrintrz_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsle_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINTRZ_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLE_D, xd, xj, xk)); } -/* Emits the `vfrintrne.s vd, vj` instruction. */ +/* Emits the `xvsle.bu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrintrne_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsle_bu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINTRNE_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLE_BU, xd, xj, xk)); } -/* Emits the `vfrintrne.d vd, vj` instruction. */ +/* Emits the `xvsle.hu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfrintrne_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsle_hu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFRINTRNE_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLE_HU, xd, xj, xk)); } -/* Emits the `vfcvtl.s.h vd, vj` instruction. */ +/* Emits the `xvsle.wu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfcvtl_s_h(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsle_wu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFCVTL_S_H, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLE_WU, xd, xj, xk)); } -/* Emits the `vfcvth.s.h vd, vj` instruction. */ +/* Emits the `xvsle.du xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfcvth_s_h(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsle_du(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFCVTH_S_H, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLE_DU, xd, xj, xk)); } -/* Emits the `vfcvtl.d.s vd, vj` instruction. */ +/* Emits the `xvslt.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfcvtl_d_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvslt_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFCVTL_D_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLT_B, xd, xj, xk)); } -/* Emits the `vfcvth.d.s vd, vj` instruction. */ +/* Emits the `xvslt.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vfcvth_d_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvslt_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFCVTH_D_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLT_H, xd, xj, xk)); } -/* Emits the `vffint.s.w vd, vj` instruction. */ +/* Emits the `xvslt.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vffint_s_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvslt_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFFINT_S_W, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLT_W, xd, xj, xk)); } -/* Emits the `vffint.s.wu vd, vj` instruction. */ +/* Emits the `xvslt.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vffint_s_wu(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvslt_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFFINT_S_WU, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLT_D, xd, xj, xk)); } -/* Emits the `vffint.d.l vd, vj` instruction. */ +/* Emits the `xvslt.bu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vffint_d_l(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvslt_bu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFFINT_D_L, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLT_BU, xd, xj, xk)); } -/* Emits the `vffint.d.lu vd, vj` instruction. */ +/* Emits the `xvslt.hu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vffint_d_lu(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvslt_hu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFFINT_D_LU, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLT_HU, xd, xj, xk)); } -/* Emits the `vffintl.d.w vd, vj` instruction. */ +/* Emits the `xvslt.wu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vffintl_d_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvslt_wu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFFINTL_D_W, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLT_WU, xd, xj, xk)); } -/* Emits the `vffinth.d.w vd, vj` instruction. */ +/* Emits the `xvslt.du xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vffinth_d_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvslt_du(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFFINTH_D_W, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLT_DU, xd, xj, xk)); } -/* Emits the `vftint.w.s vd, vj` instruction. */ +/* Emits the `xvadd.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftint_w_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvadd_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINT_W_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVADD_B, xd, xj, xk)); } -/* Emits the `vftint.l.d vd, vj` instruction. */ +/* Emits the `xvadd.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftint_l_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvadd_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINT_L_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVADD_H, xd, xj, xk)); } -/* Emits the `vftintrm.w.s vd, vj` instruction. */ +/* Emits the `xvadd.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrm_w_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvadd_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRM_W_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVADD_W, xd, xj, xk)); } -/* Emits the `vftintrm.l.d vd, vj` instruction. */ +/* Emits the `xvadd.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrm_l_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvadd_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRM_L_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVADD_D, xd, xj, xk)); } -/* Emits the `vftintrp.w.s vd, vj` instruction. */ +/* Emits the `xvsub.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrp_w_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsub_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRP_W_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSUB_B, xd, xj, xk)); } -/* Emits the `vftintrp.l.d vd, vj` instruction. */ +/* Emits the `xvsub.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrp_l_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsub_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRP_L_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSUB_H, xd, xj, xk)); } -/* Emits the `vftintrz.w.s vd, vj` instruction. */ +/* Emits the `xvsub.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrz_w_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsub_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRZ_W_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSUB_W, xd, xj, xk)); } -/* Emits the `vftintrz.l.d vd, vj` instruction. */ +/* Emits the `xvsub.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrz_l_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsub_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRZ_L_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSUB_D, xd, xj, xk)); } -/* Emits the `vftintrne.w.s vd, vj` instruction. */ +/* Emits the `xvsadd.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrne_w_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsadd_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRNE_W_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSADD_B, xd, xj, xk)); } -/* Emits the `vftintrne.l.d vd, vj` instruction. */ +/* Emits the `xvsadd.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrne_l_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsadd_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRNE_L_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSADD_H, xd, xj, xk)); } -/* Emits the `vftint.wu.s vd, vj` instruction. */ +/* Emits the `xvsadd.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftint_wu_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsadd_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINT_WU_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSADD_W, xd, xj, xk)); } -/* Emits the `vftint.lu.d vd, vj` instruction. */ +/* Emits the `xvsadd.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftint_lu_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsadd_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINT_LU_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSADD_D, xd, xj, xk)); } -/* Emits the `vftintrz.wu.s vd, vj` instruction. */ +/* Emits the `xvssub.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrz_wu_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvssub_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRZ_WU_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSSUB_B, xd, xj, xk)); } -/* Emits the `vftintrz.lu.d vd, vj` instruction. */ +/* Emits the `xvssub.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrz_lu_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvssub_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRZ_LU_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSSUB_H, xd, xj, xk)); } -/* Emits the `vftintl.l.s vd, vj` instruction. */ +/* Emits the `xvssub.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintl_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvssub_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTL_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSSUB_W, xd, xj, xk)); } -/* Emits the `vftinth.l.s vd, vj` instruction. */ +/* Emits the `xvssub.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftinth_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvssub_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTH_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSSUB_D, xd, xj, xk)); } -/* Emits the `vftintrml.l.s vd, vj` instruction. */ +/* Emits the `xvsadd.bu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrml_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsadd_bu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRML_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSADD_BU, xd, xj, xk)); } -/* Emits the `vftintrmh.l.s vd, vj` instruction. */ +/* Emits the `xvsadd.hu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrmh_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsadd_hu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRMH_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSADD_HU, xd, xj, xk)); } -/* Emits the `vftintrpl.l.s vd, vj` instruction. */ +/* Emits the `xvsadd.wu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrpl_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsadd_wu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRPL_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSADD_WU, xd, xj, xk)); } -/* Emits the `vftintrph.l.s vd, vj` instruction. */ +/* Emits the `xvsadd.du xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrph_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvsadd_du(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRPH_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSADD_DU, xd, xj, xk)); } -/* Emits the `vftintrzl.l.s vd, vj` instruction. */ +/* Emits the `xvssub.bu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrzl_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvssub_bu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRZL_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSSUB_BU, xd, xj, xk)); } -/* Emits the `vftintrzh.l.s vd, vj` instruction. */ +/* Emits the `xvssub.hu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrzh_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvssub_hu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRZH_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSSUB_HU, xd, xj, xk)); } -/* Emits the `vftintrnel.l.s vd, vj` instruction. */ +/* Emits the `xvssub.wu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrnel_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvssub_wu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRNEL_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSSUB_WU, xd, xj, xk)); } -/* Emits the `vftintrneh.l.s vd, vj` instruction. */ +/* Emits the `xvssub.du xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vftintrneh_l_s(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvssub_du(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VFTINTRNEH_L_S, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSSUB_DU, xd, xj, xk)); } -/* Emits the `vexth.h.b vd, vj` instruction. */ +/* Emits the `xvmax.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vexth_h_b(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvmax_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTH_H_B, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMAX_B, xd, xj, xk)); } -/* Emits the `vexth.w.h vd, vj` instruction. */ +/* Emits the `xvmax.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vexth_w_h(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvmax_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTH_W_H, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMAX_H, xd, xj, xk)); } -/* Emits the `vexth.d.w vd, vj` instruction. */ +/* Emits the `xvmax.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vexth_d_w(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvmax_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTH_D_W, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMAX_W, xd, xj, xk)); } -/* Emits the `vexth.q.d vd, vj` instruction. */ +/* Emits the `xvmax.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vexth_q_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvmax_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTH_Q_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMAX_D, xd, xj, xk)); } -/* Emits the `vexth.hu.bu vd, vj` instruction. */ +/* Emits the `xvmin.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vexth_hu_bu(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvmin_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTH_HU_BU, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMIN_B, xd, xj, xk)); } -/* Emits the `vexth.wu.hu vd, vj` instruction. */ +/* Emits the `xvmin.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vexth_wu_hu(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvmin_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTH_WU_HU, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMIN_H, xd, xj, xk)); } -/* Emits the `vexth.du.wu vd, vj` instruction. */ +/* Emits the `xvmin.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vexth_du_wu(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvmin_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTH_DU_WU, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMIN_W, xd, xj, xk)); } -/* Emits the `vexth.qu.du vd, vj` instruction. */ +/* Emits the `xvmin.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vexth_qu_du(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvmin_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTH_QU_DU, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMIN_D, xd, xj, xk)); } -/* Emits the `vreplgr2vr.b vd, j` instruction. */ +/* Emits the `xvmax.bu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vreplgr2vr_b(TCGContext *s, TCGReg vd, TCGReg j) +tcg_out_opc_xvmax_bu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdj_insn(OPC_VREPLGR2VR_B, vd, j)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMAX_BU, xd, xj, xk)); } -/* Emits the `vreplgr2vr.h vd, j` instruction. */ +/* Emits the `xvmax.hu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vreplgr2vr_h(TCGContext *s, TCGReg vd, TCGReg j) +tcg_out_opc_xvmax_hu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdj_insn(OPC_VREPLGR2VR_H, vd, j)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMAX_HU, xd, xj, xk)); } -/* Emits the `vreplgr2vr.w vd, j` instruction. */ +/* Emits the `xvmax.wu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vreplgr2vr_w(TCGContext *s, TCGReg vd, TCGReg j) +tcg_out_opc_xvmax_wu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdj_insn(OPC_VREPLGR2VR_W, vd, j)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMAX_WU, xd, xj, xk)); } -/* Emits the `vreplgr2vr.d vd, j` instruction. */ +/* Emits the `xvmax.du xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vreplgr2vr_d(TCGContext *s, TCGReg vd, TCGReg j) +tcg_out_opc_xvmax_du(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdj_insn(OPC_VREPLGR2VR_D, vd, j)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMAX_DU, xd, xj, xk)); } -/* Emits the `vrotri.b vd, vj, uk3` instruction. */ +/* Emits the `xvmin.bu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vrotri_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvmin_bu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VROTRI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMIN_BU, xd, xj, xk)); } -/* Emits the `vrotri.h vd, vj, uk4` instruction. */ +/* Emits the `xvmin.hu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vrotri_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvmin_hu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VROTRI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMIN_HU, xd, xj, xk)); } -/* Emits the `vrotri.w vd, vj, uk5` instruction. */ +/* Emits the `xvmin.wu xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vrotri_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvmin_wu(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VROTRI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMIN_WU, xd, xj, xk)); } -/* Emits the `vrotri.d vd, vj, uk6` instruction. */ +/* Emits the `xvmin.du xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vrotri_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvmin_du(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VROTRI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMIN_DU, xd, xj, xk)); } -/* Emits the `vsrlri.b vd, vj, uk3` instruction. */ +/* Emits the `xvmul.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlri_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvmul_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSRLRI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMUL_B, xd, xj, xk)); } -/* Emits the `vsrlri.h vd, vj, uk4` instruction. */ +/* Emits the `xvmul.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlri_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvmul_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRLRI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMUL_H, xd, xj, xk)); } -/* Emits the `vsrlri.w vd, vj, uk5` instruction. */ +/* Emits the `xvmul.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlri_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvmul_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRLRI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMUL_W, xd, xj, xk)); } -/* Emits the `vsrlri.d vd, vj, uk6` instruction. */ +/* Emits the `xvmul.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlri_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvmul_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRLRI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVMUL_D, xd, xj, xk)); } -/* Emits the `vsrari.b vd, vj, uk3` instruction. */ +/* Emits the `xvsll.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrari_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvsll_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSRARI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLL_B, xd, xj, xk)); } -/* Emits the `vsrari.h vd, vj, uk4` instruction. */ +/* Emits the `xvsll.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrari_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvsll_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRARI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLL_H, xd, xj, xk)); } -/* Emits the `vsrari.w vd, vj, uk5` instruction. */ +/* Emits the `xvsll.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrari_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvsll_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRARI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLL_W, xd, xj, xk)); } -/* Emits the `vsrari.d vd, vj, uk6` instruction. */ +/* Emits the `xvsll.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrari_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvsll_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRARI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSLL_D, xd, xj, xk)); } -/* Emits the `vinsgr2vr.b vd, j, uk4` instruction. */ +/* Emits the `xvsrl.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vinsgr2vr_b(TCGContext *s, TCGReg vd, TCGReg j, uint32_t uk4) +tcg_out_opc_xvsrl_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdjuk4_insn(OPC_VINSGR2VR_B, vd, j, uk4)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSRL_B, xd, xj, xk)); } -/* Emits the `vinsgr2vr.h vd, j, uk3` instruction. */ +/* Emits the `xvsrl.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vinsgr2vr_h(TCGContext *s, TCGReg vd, TCGReg j, uint32_t uk3) +tcg_out_opc_xvsrl_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdjuk3_insn(OPC_VINSGR2VR_H, vd, j, uk3)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSRL_H, xd, xj, xk)); } -/* Emits the `vinsgr2vr.w vd, j, uk2` instruction. */ +/* Emits the `xvsrl.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vinsgr2vr_w(TCGContext *s, TCGReg vd, TCGReg j, uint32_t uk2) +tcg_out_opc_xvsrl_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdjuk2_insn(OPC_VINSGR2VR_W, vd, j, uk2)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSRL_W, xd, xj, xk)); } -/* Emits the `vinsgr2vr.d vd, j, uk1` instruction. */ +/* Emits the `xvsrl.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vinsgr2vr_d(TCGContext *s, TCGReg vd, TCGReg j, uint32_t uk1) +tcg_out_opc_xvsrl_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdjuk1_insn(OPC_VINSGR2VR_D, vd, j, uk1)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSRL_D, xd, xj, xk)); } -/* Emits the `vpickve2gr.b d, vj, uk4` instruction. */ +/* Emits the `xvsra.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpickve2gr_b(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvsra_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_dvjuk4_insn(OPC_VPICKVE2GR_B, d, vj, uk4)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSRA_B, xd, xj, xk)); } -/* Emits the `vpickve2gr.h d, vj, uk3` instruction. */ +/* Emits the `xvsra.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpickve2gr_h(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvsra_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_dvjuk3_insn(OPC_VPICKVE2GR_H, d, vj, uk3)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSRA_H, xd, xj, xk)); } -/* Emits the `vpickve2gr.w d, vj, uk2` instruction. */ +/* Emits the `xvsra.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpickve2gr_w(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk2) +tcg_out_opc_xvsra_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_dvjuk2_insn(OPC_VPICKVE2GR_W, d, vj, uk2)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSRA_W, xd, xj, xk)); } -/* Emits the `vpickve2gr.d d, vj, uk1` instruction. */ +/* Emits the `xvsra.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpickve2gr_d(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk1) +tcg_out_opc_xvsra_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_dvjuk1_insn(OPC_VPICKVE2GR_D, d, vj, uk1)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVSRA_D, xd, xj, xk)); } -/* Emits the `vpickve2gr.bu d, vj, uk4` instruction. */ +/* Emits the `xvrotr.b xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpickve2gr_bu(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvrotr_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_dvjuk4_insn(OPC_VPICKVE2GR_BU, d, vj, uk4)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVROTR_B, xd, xj, xk)); } -/* Emits the `vpickve2gr.hu d, vj, uk3` instruction. */ +/* Emits the `xvrotr.h xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpickve2gr_hu(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvrotr_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_dvjuk3_insn(OPC_VPICKVE2GR_HU, d, vj, uk3)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVROTR_H, xd, xj, xk)); } -/* Emits the `vpickve2gr.wu d, vj, uk2` instruction. */ +/* Emits the `xvrotr.w xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpickve2gr_wu(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk2) +tcg_out_opc_xvrotr_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_dvjuk2_insn(OPC_VPICKVE2GR_WU, d, vj, uk2)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVROTR_W, xd, xj, xk)); } -/* Emits the `vpickve2gr.du d, vj, uk1` instruction. */ +/* Emits the `xvrotr.d xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpickve2gr_du(TCGContext *s, TCGReg d, TCGReg vj, uint32_t uk1) +tcg_out_opc_xvrotr_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_dvjuk1_insn(OPC_VPICKVE2GR_DU, d, vj, uk1)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVROTR_D, xd, xj, xk)); } -/* Emits the `vreplvei.b vd, vj, uk4` instruction. */ +/* Emits the `xvreplve.b xd, xj, k` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vreplvei_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvreplve_b(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg k) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VREPLVEI_B, vd, vj, uk4)); + tcg_out32(s, encode_xdxjk_insn(OPC_XVREPLVE_B, xd, xj, k)); } -/* Emits the `vreplvei.h vd, vj, uk3` instruction. */ +/* Emits the `xvreplve.h xd, xj, k` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vreplvei_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvreplve_h(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg k) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VREPLVEI_H, vd, vj, uk3)); + tcg_out32(s, encode_xdxjk_insn(OPC_XVREPLVE_H, xd, xj, k)); } -/* Emits the `vreplvei.w vd, vj, uk2` instruction. */ +/* Emits the `xvreplve.w xd, xj, k` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vreplvei_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk2) +tcg_out_opc_xvreplve_w(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg k) { - tcg_out32(s, encode_vdvjuk2_insn(OPC_VREPLVEI_W, vd, vj, uk2)); + tcg_out32(s, encode_xdxjk_insn(OPC_XVREPLVE_W, xd, xj, k)); } -/* Emits the `vreplvei.d vd, vj, uk1` instruction. */ +/* Emits the `xvreplve.d xd, xj, k` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vreplvei_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk1) +tcg_out_opc_xvreplve_d(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg k) { - tcg_out32(s, encode_vdvjuk1_insn(OPC_VREPLVEI_D, vd, vj, uk1)); + tcg_out32(s, encode_xdxjk_insn(OPC_XVREPLVE_D, xd, xj, k)); } -/* Emits the `vsllwil.h.b vd, vj, uk3` instruction. */ +/* Emits the `xvand.v xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsllwil_h_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvand_v(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSLLWIL_H_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVAND_V, xd, xj, xk)); } -/* Emits the `vsllwil.w.h vd, vj, uk4` instruction. */ +/* Emits the `xvor.v xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsllwil_w_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvor_v(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSLLWIL_W_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVOR_V, xd, xj, xk)); } -/* Emits the `vsllwil.d.w vd, vj, uk5` instruction. */ +/* Emits the `xvxor.v xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsllwil_d_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvxor_v(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSLLWIL_D_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVXOR_V, xd, xj, xk)); } -/* Emits the `vextl.q.d vd, vj` instruction. */ +/* Emits the `xvnor.v xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vextl_q_d(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvnor_v(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTL_Q_D, vd, vj)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVNOR_V, xd, xj, xk)); } -/* Emits the `vsllwil.hu.bu vd, vj, uk3` instruction. */ +/* Emits the `xvandn.v xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsllwil_hu_bu(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvandn_v(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSLLWIL_HU_BU, vd, vj, uk3)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVANDN_V, xd, xj, xk)); } -/* Emits the `vsllwil.wu.hu vd, vj, uk4` instruction. */ +/* Emits the `xvorn.v xd, xj, xk` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsllwil_wu_hu(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvorn_v(TCGContext *s, TCGReg xd, TCGReg xj, TCGReg xk) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSLLWIL_WU_HU, vd, vj, uk4)); + tcg_out32(s, encode_xdxjxk_insn(OPC_XVORN_V, xd, xj, xk)); } -/* Emits the `vsllwil.du.wu vd, vj, uk5` instruction. */ +/* Emits the `xvseqi.b xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsllwil_du_wu(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvseqi_b(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSLLWIL_DU_WU, vd, vj, uk5)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSEQI_B, xd, xj, sk5)); } -/* Emits the `vextl.qu.du vd, vj` instruction. */ +/* Emits the `xvseqi.h xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vextl_qu_du(TCGContext *s, TCGReg vd, TCGReg vj) +tcg_out_opc_xvseqi_h(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvj_insn(OPC_VEXTL_QU_DU, vd, vj)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSEQI_H, xd, xj, sk5)); } -/* Emits the `vbitclri.b vd, vj, uk3` instruction. */ +/* Emits the `xvseqi.w xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitclri_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvseqi_w(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VBITCLRI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSEQI_W, xd, xj, sk5)); } -/* Emits the `vbitclri.h vd, vj, uk4` instruction. */ +/* Emits the `xvseqi.d xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitclri_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvseqi_d(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VBITCLRI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSEQI_D, xd, xj, sk5)); } -/* Emits the `vbitclri.w vd, vj, uk5` instruction. */ +/* Emits the `xvslei.b xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitclri_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvslei_b(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VBITCLRI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSLEI_B, xd, xj, sk5)); } -/* Emits the `vbitclri.d vd, vj, uk6` instruction. */ +/* Emits the `xvslei.h xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitclri_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvslei_h(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VBITCLRI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSLEI_H, xd, xj, sk5)); } -/* Emits the `vbitseti.b vd, vj, uk3` instruction. */ +/* Emits the `xvslei.w xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitseti_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvslei_w(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VBITSETI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSLEI_W, xd, xj, sk5)); } -/* Emits the `vbitseti.h vd, vj, uk4` instruction. */ +/* Emits the `xvslei.d xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitseti_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvslei_d(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VBITSETI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSLEI_D, xd, xj, sk5)); } -/* Emits the `vbitseti.w vd, vj, uk5` instruction. */ +/* Emits the `xvslei.bu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitseti_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvslei_bu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VBITSETI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLEI_BU, xd, xj, uk5)); } -/* Emits the `vbitseti.d vd, vj, uk6` instruction. */ +/* Emits the `xvslei.hu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitseti_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvslei_hu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VBITSETI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLEI_HU, xd, xj, uk5)); } -/* Emits the `vbitrevi.b vd, vj, uk3` instruction. */ +/* Emits the `xvslei.wu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitrevi_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvslei_wu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VBITREVI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLEI_WU, xd, xj, uk5)); } -/* Emits the `vbitrevi.h vd, vj, uk4` instruction. */ +/* Emits the `xvslei.du xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitrevi_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvslei_du(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VBITREVI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLEI_DU, xd, xj, uk5)); } -/* Emits the `vbitrevi.w vd, vj, uk5` instruction. */ +/* Emits the `xvslti.b xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitrevi_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvslti_b(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VBITREVI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSLTI_B, xd, xj, sk5)); } -/* Emits the `vbitrevi.d vd, vj, uk6` instruction. */ +/* Emits the `xvslti.h xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitrevi_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvslti_h(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VBITREVI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSLTI_H, xd, xj, sk5)); } -/* Emits the `vsat.b vd, vj, uk3` instruction. */ +/* Emits the `xvslti.w xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsat_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvslti_w(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSAT_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSLTI_W, xd, xj, sk5)); } -/* Emits the `vsat.h vd, vj, uk4` instruction. */ +/* Emits the `xvslti.d xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsat_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvslti_d(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSAT_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVSLTI_D, xd, xj, sk5)); } -/* Emits the `vsat.w vd, vj, uk5` instruction. */ +/* Emits the `xvslti.bu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsat_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvslti_bu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSAT_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLTI_BU, xd, xj, uk5)); } -/* Emits the `vsat.d vd, vj, uk6` instruction. */ +/* Emits the `xvslti.hu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsat_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvslti_hu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSAT_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLTI_HU, xd, xj, uk5)); } -/* Emits the `vsat.bu vd, vj, uk3` instruction. */ +/* Emits the `xvslti.wu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsat_bu(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvslti_wu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSAT_BU, vd, vj, uk3)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLTI_WU, xd, xj, uk5)); } -/* Emits the `vsat.hu vd, vj, uk4` instruction. */ +/* Emits the `xvslti.du xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsat_hu(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvslti_du(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSAT_HU, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLTI_DU, xd, xj, uk5)); } -/* Emits the `vsat.wu vd, vj, uk5` instruction. */ +/* Emits the `xvaddi.bu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsat_wu(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvaddi_bu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSAT_WU, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVADDI_BU, xd, xj, uk5)); } -/* Emits the `vsat.du vd, vj, uk6` instruction. */ +/* Emits the `xvaddi.hu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsat_du(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvaddi_hu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSAT_DU, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVADDI_HU, xd, xj, uk5)); } -/* Emits the `vslli.b vd, vj, uk3` instruction. */ +/* Emits the `xvaddi.wu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vslli_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvaddi_wu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSLLI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVADDI_WU, xd, xj, uk5)); } -/* Emits the `vslli.h vd, vj, uk4` instruction. */ +/* Emits the `xvaddi.du xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vslli_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvaddi_du(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSLLI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVADDI_DU, xd, xj, uk5)); } -/* Emits the `vslli.w vd, vj, uk5` instruction. */ +/* Emits the `xvsubi.bu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vslli_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvsubi_bu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSLLI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSUBI_BU, xd, xj, uk5)); } -/* Emits the `vslli.d vd, vj, uk6` instruction. */ +/* Emits the `xvsubi.hu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vslli_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvsubi_hu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSLLI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSUBI_HU, xd, xj, uk5)); } -/* Emits the `vsrli.b vd, vj, uk3` instruction. */ +/* Emits the `xvsubi.wu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrli_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvsubi_wu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSRLI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSUBI_WU, xd, xj, uk5)); } -/* Emits the `vsrli.h vd, vj, uk4` instruction. */ +/* Emits the `xvsubi.du xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrli_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvsubi_du(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRLI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSUBI_DU, xd, xj, uk5)); } -/* Emits the `vsrli.w vd, vj, uk5` instruction. */ +/* Emits the `xvmaxi.b xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrli_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvmaxi_b(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRLI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVMAXI_B, xd, xj, sk5)); } -/* Emits the `vsrli.d vd, vj, uk6` instruction. */ +/* Emits the `xvmaxi.h xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrli_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvmaxi_h(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRLI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVMAXI_H, xd, xj, sk5)); } -/* Emits the `vsrai.b vd, vj, uk3` instruction. */ +/* Emits the `xvmaxi.w xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrai_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk3) +tcg_out_opc_xvmaxi_w(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk3_insn(OPC_VSRAI_B, vd, vj, uk3)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVMAXI_W, xd, xj, sk5)); } -/* Emits the `vsrai.h vd, vj, uk4` instruction. */ +/* Emits the `xvmaxi.d xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrai_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvmaxi_d(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRAI_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVMAXI_D, xd, xj, sk5)); } -/* Emits the `vsrai.w vd, vj, uk5` instruction. */ +/* Emits the `xvmini.b xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrai_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvmini_b(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRAI_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVMINI_B, xd, xj, sk5)); } -/* Emits the `vsrai.d vd, vj, uk6` instruction. */ +/* Emits the `xvmini.h xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrai_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvmini_h(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRAI_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVMINI_H, xd, xj, sk5)); } -/* Emits the `vsrlni.b.h vd, vj, uk4` instruction. */ +/* Emits the `xvmini.w xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlni_b_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvmini_w(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRLNI_B_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVMINI_W, xd, xj, sk5)); } -/* Emits the `vsrlni.h.w vd, vj, uk5` instruction. */ +/* Emits the `xvmini.d xd, xj, sk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlni_h_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvmini_d(TCGContext *s, TCGReg xd, TCGReg xj, int32_t sk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRLNI_H_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjsk5_insn(OPC_XVMINI_D, xd, xj, sk5)); } -/* Emits the `vsrlni.w.d vd, vj, uk6` instruction. */ +/* Emits the `xvmaxi.bu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlni_w_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvmaxi_bu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRLNI_W_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVMAXI_BU, xd, xj, uk5)); } -/* Emits the `vsrlni.d.q vd, vj, uk7` instruction. */ +/* Emits the `xvmaxi.hu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlni_d_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvmaxi_hu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSRLNI_D_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVMAXI_HU, xd, xj, uk5)); } -/* Emits the `vsrlrni.b.h vd, vj, uk4` instruction. */ +/* Emits the `xvmaxi.wu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlrni_b_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvmaxi_wu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRLRNI_B_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVMAXI_WU, xd, xj, uk5)); } -/* Emits the `vsrlrni.h.w vd, vj, uk5` instruction. */ +/* Emits the `xvmaxi.du xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlrni_h_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvmaxi_du(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRLRNI_H_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVMAXI_DU, xd, xj, uk5)); } -/* Emits the `vsrlrni.w.d vd, vj, uk6` instruction. */ +/* Emits the `xvmini.bu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlrni_w_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvmini_bu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRLRNI_W_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVMINI_BU, xd, xj, uk5)); } -/* Emits the `vsrlrni.d.q vd, vj, uk7` instruction. */ +/* Emits the `xvmini.hu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrlrni_d_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvmini_hu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSRLRNI_D_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVMINI_HU, xd, xj, uk5)); } -/* Emits the `vssrlni.b.h vd, vj, uk4` instruction. */ +/* Emits the `xvmini.wu xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlni_b_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvmini_wu(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSSRLNI_B_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVMINI_WU, xd, xj, uk5)); } -/* Emits the `vssrlni.h.w vd, vj, uk5` instruction. */ +/* Emits the `xvmini.du xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlni_h_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvmini_du(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSSRLNI_H_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVMINI_DU, xd, xj, uk5)); } -/* Emits the `vssrlni.w.d vd, vj, uk6` instruction. */ +/* Emits the `xvneg.b xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlni_w_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvneg_b(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSSRLNI_W_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxj_insn(OPC_XVNEG_B, xd, xj)); } -/* Emits the `vssrlni.d.q vd, vj, uk7` instruction. */ +/* Emits the `xvneg.h xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlni_d_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvneg_h(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSSRLNI_D_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxj_insn(OPC_XVNEG_H, xd, xj)); } -/* Emits the `vssrlni.bu.h vd, vj, uk4` instruction. */ +/* Emits the `xvneg.w xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlni_bu_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvneg_w(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSSRLNI_BU_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxj_insn(OPC_XVNEG_W, xd, xj)); } -/* Emits the `vssrlni.hu.w vd, vj, uk5` instruction. */ +/* Emits the `xvneg.d xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlni_hu_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvneg_d(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSSRLNI_HU_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxj_insn(OPC_XVNEG_D, xd, xj)); } -/* Emits the `vssrlni.wu.d vd, vj, uk6` instruction. */ +/* Emits the `xvreplgr2vr.b xd, j` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlni_wu_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvreplgr2vr_b(TCGContext *s, TCGReg xd, TCGReg j) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSSRLNI_WU_D, vd, vj, uk6)); + tcg_out32(s, encode_xdj_insn(OPC_XVREPLGR2VR_B, xd, j)); } -/* Emits the `vssrlni.du.q vd, vj, uk7` instruction. */ +/* Emits the `xvreplgr2vr.h xd, j` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlni_du_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvreplgr2vr_h(TCGContext *s, TCGReg xd, TCGReg j) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSSRLNI_DU_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdj_insn(OPC_XVREPLGR2VR_H, xd, j)); } -/* Emits the `vssrlrni.b.h vd, vj, uk4` instruction. */ +/* Emits the `xvreplgr2vr.w xd, j` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlrni_b_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvreplgr2vr_w(TCGContext *s, TCGReg xd, TCGReg j) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSSRLRNI_B_H, vd, vj, uk4)); + tcg_out32(s, encode_xdj_insn(OPC_XVREPLGR2VR_W, xd, j)); } -/* Emits the `vssrlrni.h.w vd, vj, uk5` instruction. */ +/* Emits the `xvreplgr2vr.d xd, j` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlrni_h_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvreplgr2vr_d(TCGContext *s, TCGReg xd, TCGReg j) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSSRLRNI_H_W, vd, vj, uk5)); + tcg_out32(s, encode_xdj_insn(OPC_XVREPLGR2VR_D, xd, j)); } -/* Emits the `vssrlrni.w.d vd, vj, uk6` instruction. */ +/* Emits the `xvrotri.b xd, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlrni_w_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvrotri_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSSRLRNI_W_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk3_insn(OPC_XVROTRI_B, xd, xj, uk3)); } -/* Emits the `vssrlrni.d.q vd, vj, uk7` instruction. */ +/* Emits the `xvrotri.h xd, xj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlrni_d_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvrotri_h(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk4) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSSRLRNI_D_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxjuk4_insn(OPC_XVROTRI_H, xd, xj, uk4)); } -/* Emits the `vssrlrni.bu.h vd, vj, uk4` instruction. */ +/* Emits the `xvrotri.w xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlrni_bu_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvrotri_w(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSSRLRNI_BU_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVROTRI_W, xd, xj, uk5)); } -/* Emits the `vssrlrni.hu.w vd, vj, uk5` instruction. */ +/* Emits the `xvrotri.d xd, xj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlrni_hu_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvrotri_d(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk6) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSSRLRNI_HU_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk6_insn(OPC_XVROTRI_D, xd, xj, uk6)); } -/* Emits the `vssrlrni.wu.d vd, vj, uk6` instruction. */ +/* Emits the `xvinsgr2vr.w xd, j, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlrni_wu_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvinsgr2vr_w(TCGContext *s, TCGReg xd, TCGReg j, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSSRLRNI_WU_D, vd, vj, uk6)); + tcg_out32(s, encode_xdjuk3_insn(OPC_XVINSGR2VR_W, xd, j, uk3)); } -/* Emits the `vssrlrni.du.q vd, vj, uk7` instruction. */ +/* Emits the `xvinsgr2vr.d xd, j, uk2` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrlrni_du_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvinsgr2vr_d(TCGContext *s, TCGReg xd, TCGReg j, uint32_t uk2) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSSRLRNI_DU_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdjuk2_insn(OPC_XVINSGR2VR_D, xd, j, uk2)); } -/* Emits the `vsrani.b.h vd, vj, uk4` instruction. */ +/* Emits the `xvpickve2gr.w d, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrani_b_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvpickve2gr_w(TCGContext *s, TCGReg d, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRANI_B_H, vd, vj, uk4)); + tcg_out32(s, encode_dxjuk3_insn(OPC_XVPICKVE2GR_W, d, xj, uk3)); } -/* Emits the `vsrani.h.w vd, vj, uk5` instruction. */ +/* Emits the `xvpickve2gr.d d, xj, uk2` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrani_h_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvpickve2gr_d(TCGContext *s, TCGReg d, TCGReg xj, uint32_t uk2) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRANI_H_W, vd, vj, uk5)); + tcg_out32(s, encode_dxjuk2_insn(OPC_XVPICKVE2GR_D, d, xj, uk2)); } -/* Emits the `vsrani.w.d vd, vj, uk6` instruction. */ +/* Emits the `xvpickve2gr.wu d, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrani_w_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvpickve2gr_wu(TCGContext *s, TCGReg d, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRANI_W_D, vd, vj, uk6)); + tcg_out32(s, encode_dxjuk3_insn(OPC_XVPICKVE2GR_WU, d, xj, uk3)); } -/* Emits the `vsrani.d.q vd, vj, uk7` instruction. */ +/* Emits the `xvpickve2gr.du d, xj, uk2` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrani_d_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvpickve2gr_du(TCGContext *s, TCGReg d, TCGReg xj, uint32_t uk2) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSRANI_D_Q, vd, vj, uk7)); + tcg_out32(s, encode_dxjuk2_insn(OPC_XVPICKVE2GR_DU, d, xj, uk2)); } -/* Emits the `vsrarni.b.h vd, vj, uk4` instruction. */ +/* Emits the `xvrepl128vei.b xd, xj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrarni_b_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvrepl128vei_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk4) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSRARNI_B_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk4_insn(OPC_XVREPL128VEI_B, xd, xj, uk4)); } -/* Emits the `vsrarni.h.w vd, vj, uk5` instruction. */ +/* Emits the `xvrepl128vei.h xd, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrarni_h_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvrepl128vei_h(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSRARNI_H_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk3_insn(OPC_XVREPL128VEI_H, xd, xj, uk3)); } -/* Emits the `vsrarni.w.d vd, vj, uk6` instruction. */ +/* Emits the `xvrepl128vei.w xd, xj, uk2` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrarni_w_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvrepl128vei_w(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk2) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSRARNI_W_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk2_insn(OPC_XVREPL128VEI_W, xd, xj, uk2)); } -/* Emits the `vsrarni.d.q vd, vj, uk7` instruction. */ +/* Emits the `xvrepl128vei.d xd, xj, uk1` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vsrarni_d_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvrepl128vei_d(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk1) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSRARNI_D_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxjuk1_insn(OPC_XVREPL128VEI_D, xd, xj, uk1)); } -/* Emits the `vssrani.b.h vd, vj, uk4` instruction. */ +/* Emits the `xvreplve0.b xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrani_b_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvreplve0_b(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSSRANI_B_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxj_insn(OPC_XVREPLVE0_B, xd, xj)); } -/* Emits the `vssrani.h.w vd, vj, uk5` instruction. */ +/* Emits the `xvreplve0.h xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrani_h_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvreplve0_h(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSSRANI_H_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxj_insn(OPC_XVREPLVE0_H, xd, xj)); } -/* Emits the `vssrani.w.d vd, vj, uk6` instruction. */ +/* Emits the `xvreplve0.w xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrani_w_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvreplve0_w(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSSRANI_W_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxj_insn(OPC_XVREPLVE0_W, xd, xj)); } -/* Emits the `vssrani.d.q vd, vj, uk7` instruction. */ +/* Emits the `xvreplve0.d xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrani_d_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvreplve0_d(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSSRANI_D_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxj_insn(OPC_XVREPLVE0_D, xd, xj)); } -/* Emits the `vssrani.bu.h vd, vj, uk4` instruction. */ +/* Emits the `xvreplve0.q xd, xj` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrani_bu_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvreplve0_q(TCGContext *s, TCGReg xd, TCGReg xj) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSSRANI_BU_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxj_insn(OPC_XVREPLVE0_Q, xd, xj)); } -/* Emits the `vssrani.hu.w vd, vj, uk5` instruction. */ +/* Emits the `xvbitclri.b xd, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrani_hu_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvbitclri_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSSRANI_HU_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk3_insn(OPC_XVBITCLRI_B, xd, xj, uk3)); } -/* Emits the `vssrani.wu.d vd, vj, uk6` instruction. */ +/* Emits the `xvbitclri.h xd, xj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrani_wu_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvbitclri_h(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk4) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSSRANI_WU_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk4_insn(OPC_XVBITCLRI_H, xd, xj, uk4)); } -/* Emits the `vssrani.du.q vd, vj, uk7` instruction. */ +/* Emits the `xvbitclri.w xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrani_du_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvbitclri_w(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSSRANI_DU_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVBITCLRI_W, xd, xj, uk5)); } -/* Emits the `vssrarni.b.h vd, vj, uk4` instruction. */ +/* Emits the `xvbitclri.d xd, xj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrarni_b_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvbitclri_d(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk6) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSSRARNI_B_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk6_insn(OPC_XVBITCLRI_D, xd, xj, uk6)); } -/* Emits the `vssrarni.h.w vd, vj, uk5` instruction. */ +/* Emits the `xvbitseti.b xd, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrarni_h_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvbitseti_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSSRARNI_H_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk3_insn(OPC_XVBITSETI_B, xd, xj, uk3)); } -/* Emits the `vssrarni.w.d vd, vj, uk6` instruction. */ +/* Emits the `xvbitseti.h xd, xj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrarni_w_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvbitseti_h(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk4) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSSRARNI_W_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk4_insn(OPC_XVBITSETI_H, xd, xj, uk4)); } -/* Emits the `vssrarni.d.q vd, vj, uk7` instruction. */ +/* Emits the `xvbitseti.w xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrarni_d_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvbitseti_w(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSSRARNI_D_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVBITSETI_W, xd, xj, uk5)); } -/* Emits the `vssrarni.bu.h vd, vj, uk4` instruction. */ +/* Emits the `xvbitseti.d xd, xj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrarni_bu_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk4) +tcg_out_opc_xvbitseti_d(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk6) { - tcg_out32(s, encode_vdvjuk4_insn(OPC_VSSRARNI_BU_H, vd, vj, uk4)); + tcg_out32(s, encode_xdxjuk6_insn(OPC_XVBITSETI_D, xd, xj, uk6)); } -/* Emits the `vssrarni.hu.w vd, vj, uk5` instruction. */ +/* Emits the `xvbitrevi.b xd, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrarni_hu_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk5) +tcg_out_opc_xvbitrevi_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk5_insn(OPC_VSSRARNI_HU_W, vd, vj, uk5)); + tcg_out32(s, encode_xdxjuk3_insn(OPC_XVBITREVI_B, xd, xj, uk3)); } -/* Emits the `vssrarni.wu.d vd, vj, uk6` instruction. */ +/* Emits the `xvbitrevi.h xd, xj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrarni_wu_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk6) +tcg_out_opc_xvbitrevi_h(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk4) { - tcg_out32(s, encode_vdvjuk6_insn(OPC_VSSRARNI_WU_D, vd, vj, uk6)); + tcg_out32(s, encode_xdxjuk4_insn(OPC_XVBITREVI_H, xd, xj, uk4)); } -/* Emits the `vssrarni.du.q vd, vj, uk7` instruction. */ +/* Emits the `xvbitrevi.w xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vssrarni_du_q(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk7) +tcg_out_opc_xvbitrevi_w(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk7_insn(OPC_VSSRARNI_DU_Q, vd, vj, uk7)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVBITREVI_W, xd, xj, uk5)); } -/* Emits the `vextrins.d vd, vj, uk8` instruction. */ +/* Emits the `xvbitrevi.d xd, xj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vextrins_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvbitrevi_d(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk6) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VEXTRINS_D, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk6_insn(OPC_XVBITREVI_D, xd, xj, uk6)); } -/* Emits the `vextrins.w vd, vj, uk8` instruction. */ +/* Emits the `xvslli.b xd, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vextrins_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvslli_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VEXTRINS_W, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk3_insn(OPC_XVSLLI_B, xd, xj, uk3)); } -/* Emits the `vextrins.h vd, vj, uk8` instruction. */ +/* Emits the `xvslli.h xd, xj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vextrins_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvslli_h(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk4) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VEXTRINS_H, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk4_insn(OPC_XVSLLI_H, xd, xj, uk4)); } -/* Emits the `vextrins.b vd, vj, uk8` instruction. */ +/* Emits the `xvslli.w xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vextrins_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvslli_w(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VEXTRINS_B, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSLLI_W, xd, xj, uk5)); } -/* Emits the `vshuf4i.b vd, vj, uk8` instruction. */ +/* Emits the `xvslli.d xd, xj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vshuf4i_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvslli_d(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk6) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VSHUF4I_B, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk6_insn(OPC_XVSLLI_D, xd, xj, uk6)); } -/* Emits the `vshuf4i.h vd, vj, uk8` instruction. */ +/* Emits the `xvsrli.b xd, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vshuf4i_h(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvsrli_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VSHUF4I_H, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk3_insn(OPC_XVSRLI_B, xd, xj, uk3)); } -/* Emits the `vshuf4i.w vd, vj, uk8` instruction. */ +/* Emits the `xvsrli.h xd, xj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vshuf4i_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvsrli_h(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk4) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VSHUF4I_W, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk4_insn(OPC_XVSRLI_H, xd, xj, uk4)); } -/* Emits the `vshuf4i.d vd, vj, uk8` instruction. */ +/* Emits the `xvsrli.w xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vshuf4i_d(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvsrli_w(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VSHUF4I_D, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSRLI_W, xd, xj, uk5)); } -/* Emits the `vbitseli.b vd, vj, uk8` instruction. */ +/* Emits the `xvsrli.d xd, xj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vbitseli_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvsrli_d(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk6) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VBITSELI_B, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk6_insn(OPC_XVSRLI_D, xd, xj, uk6)); } -/* Emits the `vandi.b vd, vj, uk8` instruction. */ +/* Emits the `xvsrai.b xd, xj, uk3` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vandi_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvsrai_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk3) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VANDI_B, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk3_insn(OPC_XVSRAI_B, xd, xj, uk3)); } -/* Emits the `vori.b vd, vj, uk8` instruction. */ +/* Emits the `xvsrai.h xd, xj, uk4` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vori_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvsrai_h(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk4) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VORI_B, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk4_insn(OPC_XVSRAI_H, xd, xj, uk4)); } -/* Emits the `vxori.b vd, vj, uk8` instruction. */ +/* Emits the `xvsrai.w xd, xj, uk5` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vxori_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvsrai_w(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk5) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VXORI_B, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk5_insn(OPC_XVSRAI_W, xd, xj, uk5)); } -/* Emits the `vnori.b vd, vj, uk8` instruction. */ +/* Emits the `xvsrai.d xd, xj, uk6` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vnori_b(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvsrai_d(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk6) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VNORI_B, vd, vj, uk8)); + tcg_out32(s, encode_xdxjuk6_insn(OPC_XVSRAI_D, xd, xj, uk6)); } -/* Emits the `vldi vd, sj13` instruction. */ +/* Emits the `xvbitseli.b xd, xj, uk8` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vldi(TCGContext *s, TCGReg vd, int32_t sj13) +tcg_out_opc_xvbitseli_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk8) { - tcg_out32(s, encode_vdsj13_insn(OPC_VLDI, vd, sj13)); + tcg_out32(s, encode_xdxjuk8_insn(OPC_XVBITSELI_B, xd, xj, uk8)); +} + +/* Emits the `xvandi.b xd, xj, uk8` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvandi_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk8) +{ + tcg_out32(s, encode_xdxjuk8_insn(OPC_XVANDI_B, xd, xj, uk8)); +} + +/* Emits the `xvori.b xd, xj, uk8` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvori_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk8) +{ + tcg_out32(s, encode_xdxjuk8_insn(OPC_XVORI_B, xd, xj, uk8)); +} + +/* Emits the `xvxori.b xd, xj, uk8` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvxori_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk8) +{ + tcg_out32(s, encode_xdxjuk8_insn(OPC_XVXORI_B, xd, xj, uk8)); +} + +/* Emits the `xvnori.b xd, xj, uk8` instruction. */ +static void __attribute__((unused)) +tcg_out_opc_xvnori_b(TCGContext *s, TCGReg xd, TCGReg xj, uint32_t uk8) +{ + tcg_out32(s, encode_xdxjuk8_insn(OPC_XVNORI_B, xd, xj, uk8)); } -/* Emits the `vpermi.w vd, vj, uk8` instruction. */ +/* Emits the `xvldi xd, sj13` instruction. */ static void __attribute__((unused)) -tcg_out_opc_vpermi_w(TCGContext *s, TCGReg vd, TCGReg vj, uint32_t uk8) +tcg_out_opc_xvldi(TCGContext *s, TCGReg xd, int32_t sj13) { - tcg_out32(s, encode_vdvjuk8_insn(OPC_VPERMI_W, vd, vj, uk8)); + tcg_out32(s, encode_xdsj13_insn(OPC_XVLDI, xd, sj13)); } /* End of generated code. */ diff --git a/tcg/loongarch64/tcg-target.c.inc b/tcg/loongarch64/tcg-target.c.inc index 06ca1ab..5b7ed5c 100644 --- a/tcg/loongarch64/tcg-target.c.inc +++ b/tcg/loongarch64/tcg-target.c.inc @@ -303,15 +303,31 @@ static bool tcg_out_mov(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg) switch (type) { case TCG_TYPE_I32: case TCG_TYPE_I64: - /* - * Conventional register-register move used in LoongArch is - * `or dst, src, zero`. - */ - tcg_out_opc_or(s, ret, arg, TCG_REG_ZERO); + if (ret < TCG_REG_V0) { + if (arg < TCG_REG_V0) { + /* + * Conventional register-register move used in LoongArch is + * `or dst, src, zero`. + */ + tcg_out_opc_or(s, ret, arg, TCG_REG_ZERO); + } else { + tcg_out_opc_movfr2gr_d(s, ret, arg); + } + } else { + if (arg < TCG_REG_V0) { + tcg_out_opc_movgr2fr_d(s, ret, arg); + } else { + tcg_out_opc_fmov_d(s, ret, arg); + } + } break; + case TCG_TYPE_V64: case TCG_TYPE_V128: tcg_out_opc_vori_b(s, ret, arg, 0); break; + case TCG_TYPE_V256: + tcg_out_opc_xvori_b(s, ret, arg, 0); + break; default: g_assert_not_reached(); } @@ -366,8 +382,7 @@ static void tcg_out_movi(TCGContext *s, TCGType type, TCGReg rd, * back to the slow path. */ - intptr_t pc_offset; - tcg_target_long val_lo, val_hi, pc_hi, offset_hi; + intptr_t src_rx, pc_offset; tcg_target_long hi12, hi32, hi52; /* Value fits in signed i32. */ @@ -377,24 +392,23 @@ static void tcg_out_movi(TCGContext *s, TCGType type, TCGReg rd, } /* PC-relative cases. */ - pc_offset = tcg_pcrel_diff(s, (void *)val); - if (pc_offset == sextreg(pc_offset, 0, 22) && (pc_offset & 3) == 0) { - /* Single pcaddu2i. */ - tcg_out_opc_pcaddu2i(s, rd, pc_offset >> 2); - return; + src_rx = (intptr_t)tcg_splitwx_to_rx(s->code_ptr); + if ((val & 3) == 0) { + pc_offset = val - src_rx; + if (pc_offset == sextreg(pc_offset, 0, 22)) { + /* Single pcaddu2i. */ + tcg_out_opc_pcaddu2i(s, rd, pc_offset >> 2); + return; + } } - if (pc_offset == (int32_t)pc_offset) { - /* Offset within 32 bits; load with pcalau12i + ori. */ - val_lo = sextreg(val, 0, 12); - val_hi = val >> 12; - pc_hi = (val - pc_offset) >> 12; - offset_hi = val_hi - pc_hi; - - tcg_debug_assert(offset_hi == sextreg(offset_hi, 0, 20)); - tcg_out_opc_pcalau12i(s, rd, offset_hi); + pc_offset = (val >> 12) - (src_rx >> 12); + if (pc_offset == sextreg(pc_offset, 0, 20)) { + /* Load with pcalau12i + ori. */ + tcg_target_long val_lo = val & 0xfff; + tcg_out_opc_pcalau12i(s, rd, pc_offset); if (val_lo != 0) { - tcg_out_opc_ori(s, rd, rd, val_lo & 0xfff); + tcg_out_opc_ori(s, rd, rd, val_lo); } return; } @@ -803,6 +817,12 @@ static void tcg_out_ldst(TCGContext *s, LoongArchInsn opc, TCGReg data, case OPC_ST_D: tcg_out32(s, encode_djsk12_insn(opc, data, addr, imm12)); break; + case OPC_FLD_S: + case OPC_FLD_D: + case OPC_FST_S: + case OPC_FST_D: + tcg_out32(s, encode_fdjsk12_insn(opc, data, addr, imm12)); + break; default: g_assert_not_reached(); } @@ -816,14 +836,15 @@ static void tcg_out_ld(TCGContext *s, TCGType type, TCGReg dest, if (dest < TCG_REG_V0) { tcg_out_ldst(s, OPC_LD_W, dest, base, offset); } else { - tcg_out_dupm_vec(s, TCG_TYPE_I128, MO_32, dest, base, offset); + tcg_out_ldst(s, OPC_FLD_S, dest, base, offset); } break; case TCG_TYPE_I64: + case TCG_TYPE_V64: if (dest < TCG_REG_V0) { tcg_out_ldst(s, OPC_LD_D, dest, base, offset); } else { - tcg_out_dupm_vec(s, TCG_TYPE_I128, MO_64, dest, base, offset); + tcg_out_ldst(s, OPC_FLD_D, dest, base, offset); } break; case TCG_TYPE_V128: @@ -834,6 +855,14 @@ static void tcg_out_ld(TCGContext *s, TCGType type, TCGReg dest, tcg_out_opc_vldx(s, dest, base, TCG_REG_TMP0); } break; + case TCG_TYPE_V256: + if (-0x800 <= offset && offset <= 0x7ff) { + tcg_out_opc_xvld(s, dest, base, offset); + } else { + tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_TMP0, offset); + tcg_out_opc_xvldx(s, dest, base, TCG_REG_TMP0); + } + break; default: g_assert_not_reached(); } @@ -847,36 +876,15 @@ static void tcg_out_st(TCGContext *s, TCGType type, TCGReg src, if (src < TCG_REG_V0) { tcg_out_ldst(s, OPC_ST_W, src, base, offset); } else { - /* TODO: Could use fst_s, fstx_s */ - if (offset < -0x100 || offset > 0xff || (offset & 3)) { - if (-0x800 <= offset && offset <= 0x7ff) { - tcg_out_opc_addi_d(s, TCG_REG_TMP0, base, offset); - } else { - tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_TMP0, offset); - tcg_out_opc_add_d(s, TCG_REG_TMP0, TCG_REG_TMP0, base); - } - base = TCG_REG_TMP0; - offset = 0; - } - tcg_out_opc_vstelm_w(s, src, base, offset, 0); + tcg_out_ldst(s, OPC_FST_S, src, base, offset); } break; case TCG_TYPE_I64: + case TCG_TYPE_V64: if (src < TCG_REG_V0) { tcg_out_ldst(s, OPC_ST_D, src, base, offset); } else { - /* TODO: Could use fst_d, fstx_d */ - if (offset < -0x100 || offset > 0xff || (offset & 7)) { - if (-0x800 <= offset && offset <= 0x7ff) { - tcg_out_opc_addi_d(s, TCG_REG_TMP0, base, offset); - } else { - tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_TMP0, offset); - tcg_out_opc_add_d(s, TCG_REG_TMP0, TCG_REG_TMP0, base); - } - base = TCG_REG_TMP0; - offset = 0; - } - tcg_out_opc_vstelm_d(s, src, base, offset, 0); + tcg_out_ldst(s, OPC_FST_D, src, base, offset); } break; case TCG_TYPE_V128: @@ -887,6 +895,14 @@ static void tcg_out_st(TCGContext *s, TCGType type, TCGReg src, tcg_out_opc_vstx(s, src, base, TCG_REG_TMP0); } break; + case TCG_TYPE_V256: + if (-0x800 <= offset && offset <= 0x7ff) { + tcg_out_opc_xvst(s, src, base, offset); + } else { + tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_TMP0, offset); + tcg_out_opc_xvstx(s, src, base, TCG_REG_TMP0); + } + break; default: g_assert_not_reached(); } @@ -1675,30 +1691,26 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, static bool tcg_out_dup_vec(TCGContext *s, TCGType type, unsigned vece, TCGReg rd, TCGReg rs) { - switch (vece) { - case MO_8: - tcg_out_opc_vreplgr2vr_b(s, rd, rs); - break; - case MO_16: - tcg_out_opc_vreplgr2vr_h(s, rd, rs); - break; - case MO_32: - tcg_out_opc_vreplgr2vr_w(s, rd, rs); - break; - case MO_64: - tcg_out_opc_vreplgr2vr_d(s, rd, rs); - break; - default: - g_assert_not_reached(); - } + static const LoongArchInsn repl_insn[2][4] = { + { OPC_VREPLGR2VR_B, OPC_VREPLGR2VR_H, + OPC_VREPLGR2VR_W, OPC_VREPLGR2VR_D }, + { OPC_XVREPLGR2VR_B, OPC_XVREPLGR2VR_H, + OPC_XVREPLGR2VR_W, OPC_XVREPLGR2VR_D }, + }; + bool lasx = type == TCG_TYPE_V256; + + tcg_debug_assert(vece <= MO_64); + tcg_out32(s, encode_vdj_insn(repl_insn[lasx][vece], rd, rs)); return true; } static bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece, TCGReg r, TCGReg base, intptr_t offset) { - /* Handle imm overflow and division (vldrepl.d imm is divided by 8) */ - if (offset < -0x800 || offset > 0x7ff || \ + bool lasx = type == TCG_TYPE_V256; + + /* Handle imm overflow and division (vldrepl.d imm is divided by 8). */ + if (offset < -0x800 || offset > 0x7ff || (offset & ((1 << vece) - 1)) != 0) { tcg_out_addi(s, TCG_TYPE_I64, TCG_REG_TMP0, base, offset); base = TCG_REG_TMP0; @@ -1708,16 +1720,32 @@ static bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece, switch (vece) { case MO_8: - tcg_out_opc_vldrepl_b(s, r, base, offset); + if (lasx) { + tcg_out_opc_xvldrepl_b(s, r, base, offset); + } else { + tcg_out_opc_vldrepl_b(s, r, base, offset); + } break; case MO_16: - tcg_out_opc_vldrepl_h(s, r, base, offset); + if (lasx) { + tcg_out_opc_xvldrepl_h(s, r, base, offset); + } else { + tcg_out_opc_vldrepl_h(s, r, base, offset); + } break; case MO_32: - tcg_out_opc_vldrepl_w(s, r, base, offset); + if (lasx) { + tcg_out_opc_xvldrepl_w(s, r, base, offset); + } else { + tcg_out_opc_vldrepl_w(s, r, base, offset); + } break; case MO_64: - tcg_out_opc_vldrepl_d(s, r, base, offset); + if (lasx) { + tcg_out_opc_xvldrepl_d(s, r, base, offset); + } else { + tcg_out_opc_vldrepl_d(s, r, base, offset); + } break; default: g_assert_not_reached(); @@ -1732,75 +1760,69 @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece, int64_t value = sextract64(v64, 0, 8 << vece); if (-0x200 <= value && value <= 0x1FF) { uint32_t imm = (vece << 10) | ((uint32_t)v64 & 0x3FF); - tcg_out_opc_vldi(s, rd, imm); + + if (type == TCG_TYPE_V256) { + tcg_out_opc_xvldi(s, rd, imm); + } else { + tcg_out_opc_vldi(s, rd, imm); + } return; } /* TODO: vldi patterns when imm 12 is set */ - /* Fallback to vreplgr2vr */ tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_TMP0, value); - switch (vece) { - case MO_8: - tcg_out_opc_vreplgr2vr_b(s, rd, TCG_REG_TMP0); - break; - case MO_16: - tcg_out_opc_vreplgr2vr_h(s, rd, TCG_REG_TMP0); - break; - case MO_32: - tcg_out_opc_vreplgr2vr_w(s, rd, TCG_REG_TMP0); - break; - case MO_64: - tcg_out_opc_vreplgr2vr_d(s, rd, TCG_REG_TMP0); - break; - default: - g_assert_not_reached(); - } + tcg_out_dup_vec(s, type, vece, rd, TCG_REG_TMP0); } -static void tcg_out_addsub_vec(TCGContext *s, unsigned vece, const TCGArg a0, - const TCGArg a1, const TCGArg a2, +static void tcg_out_addsub_vec(TCGContext *s, bool lasx, unsigned vece, + TCGArg a0, TCGArg a1, TCGArg a2, bool a2_is_const, bool is_add) { - static const LoongArchInsn add_vec_insn[4] = { - OPC_VADD_B, OPC_VADD_H, OPC_VADD_W, OPC_VADD_D + static const LoongArchInsn add_vec_insn[2][4] = { + { OPC_VADD_B, OPC_VADD_H, OPC_VADD_W, OPC_VADD_D }, + { OPC_XVADD_B, OPC_XVADD_H, OPC_XVADD_W, OPC_XVADD_D }, }; - static const LoongArchInsn add_vec_imm_insn[4] = { - OPC_VADDI_BU, OPC_VADDI_HU, OPC_VADDI_WU, OPC_VADDI_DU + static const LoongArchInsn add_vec_imm_insn[2][4] = { + { OPC_VADDI_BU, OPC_VADDI_HU, OPC_VADDI_WU, OPC_VADDI_DU }, + { OPC_XVADDI_BU, OPC_XVADDI_HU, OPC_XVADDI_WU, OPC_XVADDI_DU }, }; - static const LoongArchInsn sub_vec_insn[4] = { - OPC_VSUB_B, OPC_VSUB_H, OPC_VSUB_W, OPC_VSUB_D + static const LoongArchInsn sub_vec_insn[2][4] = { + { OPC_VSUB_B, OPC_VSUB_H, OPC_VSUB_W, OPC_VSUB_D }, + { OPC_XVSUB_B, OPC_XVSUB_H, OPC_XVSUB_W, OPC_XVSUB_D }, }; - static const LoongArchInsn sub_vec_imm_insn[4] = { - OPC_VSUBI_BU, OPC_VSUBI_HU, OPC_VSUBI_WU, OPC_VSUBI_DU + static const LoongArchInsn sub_vec_imm_insn[2][4] = { + { OPC_VSUBI_BU, OPC_VSUBI_HU, OPC_VSUBI_WU, OPC_VSUBI_DU }, + { OPC_XVSUBI_BU, OPC_XVSUBI_HU, OPC_XVSUBI_WU, OPC_XVSUBI_DU }, }; + LoongArchInsn insn; if (a2_is_const) { int64_t value = sextract64(a2, 0, 8 << vece); + if (!is_add) { value = -value; } - - /* Try vaddi/vsubi */ - if (0 <= value && value <= 0x1f) { - tcg_out32(s, encode_vdvjuk5_insn(add_vec_imm_insn[vece], a0, \ - a1, value)); - return; - } else if (-0x1f <= value && value < 0) { - tcg_out32(s, encode_vdvjuk5_insn(sub_vec_imm_insn[vece], a0, \ - a1, -value)); - return; + if (value < 0) { + insn = sub_vec_imm_insn[lasx][vece]; + value = -value; + } else { + insn = add_vec_imm_insn[lasx][vece]; } - /* constraint TCG_CT_CONST_VADD ensures unreachable */ - g_assert_not_reached(); + /* Constraint TCG_CT_CONST_VADD ensures validity. */ + tcg_debug_assert(0 <= value && value <= 0x1f); + + tcg_out32(s, encode_vdvjuk5_insn(insn, a0, a1, value)); + return; } if (is_add) { - tcg_out32(s, encode_vdvjvk_insn(add_vec_insn[vece], a0, a1, a2)); + insn = add_vec_insn[lasx][vece]; } else { - tcg_out32(s, encode_vdvjvk_insn(sub_vec_insn[vece], a0, a1, a2)); + insn = sub_vec_insn[lasx][vece]; } + tcg_out32(s, encode_vdvjvk_insn(insn, a0, a1, a2)); } static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc, @@ -1809,74 +1831,125 @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc, const int const_args[TCG_MAX_OP_ARGS]) { TCGType type = vecl + TCG_TYPE_V64; + bool lasx = type == TCG_TYPE_V256; TCGArg a0, a1, a2, a3; - TCGReg temp_vec = TCG_VEC_TMP0; - - static const LoongArchInsn cmp_vec_insn[16][4] = { - [TCG_COND_EQ] = {OPC_VSEQ_B, OPC_VSEQ_H, OPC_VSEQ_W, OPC_VSEQ_D}, - [TCG_COND_LE] = {OPC_VSLE_B, OPC_VSLE_H, OPC_VSLE_W, OPC_VSLE_D}, - [TCG_COND_LEU] = {OPC_VSLE_BU, OPC_VSLE_HU, OPC_VSLE_WU, OPC_VSLE_DU}, - [TCG_COND_LT] = {OPC_VSLT_B, OPC_VSLT_H, OPC_VSLT_W, OPC_VSLT_D}, - [TCG_COND_LTU] = {OPC_VSLT_BU, OPC_VSLT_HU, OPC_VSLT_WU, OPC_VSLT_DU}, + LoongArchInsn insn; + + static const LoongArchInsn cmp_vec_insn[16][2][4] = { + [TCG_COND_EQ] = { + { OPC_VSEQ_B, OPC_VSEQ_H, OPC_VSEQ_W, OPC_VSEQ_D }, + { OPC_XVSEQ_B, OPC_XVSEQ_H, OPC_XVSEQ_W, OPC_XVSEQ_D }, + }, + [TCG_COND_LE] = { + { OPC_VSLE_B, OPC_VSLE_H, OPC_VSLE_W, OPC_VSLE_D }, + { OPC_XVSLE_B, OPC_XVSLE_H, OPC_XVSLE_W, OPC_XVSLE_D }, + }, + [TCG_COND_LEU] = { + { OPC_VSLE_BU, OPC_VSLE_HU, OPC_VSLE_WU, OPC_VSLE_DU }, + { OPC_XVSLE_BU, OPC_XVSLE_HU, OPC_XVSLE_WU, OPC_XVSLE_DU }, + }, + [TCG_COND_LT] = { + { OPC_VSLT_B, OPC_VSLT_H, OPC_VSLT_W, OPC_VSLT_D }, + { OPC_XVSLT_B, OPC_XVSLT_H, OPC_XVSLT_W, OPC_XVSLT_D }, + }, + [TCG_COND_LTU] = { + { OPC_VSLT_BU, OPC_VSLT_HU, OPC_VSLT_WU, OPC_VSLT_DU }, + { OPC_XVSLT_BU, OPC_XVSLT_HU, OPC_XVSLT_WU, OPC_XVSLT_DU }, + } }; - static const LoongArchInsn cmp_vec_imm_insn[16][4] = { - [TCG_COND_EQ] = {OPC_VSEQI_B, OPC_VSEQI_H, OPC_VSEQI_W, OPC_VSEQI_D}, - [TCG_COND_LE] = {OPC_VSLEI_B, OPC_VSLEI_H, OPC_VSLEI_W, OPC_VSLEI_D}, - [TCG_COND_LEU] = {OPC_VSLEI_BU, OPC_VSLEI_HU, OPC_VSLEI_WU, OPC_VSLEI_DU}, - [TCG_COND_LT] = {OPC_VSLTI_B, OPC_VSLTI_H, OPC_VSLTI_W, OPC_VSLTI_D}, - [TCG_COND_LTU] = {OPC_VSLTI_BU, OPC_VSLTI_HU, OPC_VSLTI_WU, OPC_VSLTI_DU}, + static const LoongArchInsn cmp_vec_imm_insn[16][2][4] = { + [TCG_COND_EQ] = { + { OPC_VSEQI_B, OPC_VSEQI_H, OPC_VSEQI_W, OPC_VSEQI_D }, + { OPC_XVSEQI_B, OPC_XVSEQI_H, OPC_XVSEQI_W, OPC_XVSEQI_D }, + }, + [TCG_COND_LE] = { + { OPC_VSLEI_B, OPC_VSLEI_H, OPC_VSLEI_W, OPC_VSLEI_D }, + { OPC_XVSLEI_B, OPC_XVSLEI_H, OPC_XVSLEI_W, OPC_XVSLEI_D }, + }, + [TCG_COND_LEU] = { + { OPC_VSLEI_BU, OPC_VSLEI_HU, OPC_VSLEI_WU, OPC_VSLEI_DU }, + { OPC_XVSLEI_BU, OPC_XVSLEI_HU, OPC_XVSLEI_WU, OPC_XVSLEI_DU }, + }, + [TCG_COND_LT] = { + { OPC_VSLTI_B, OPC_VSLTI_H, OPC_VSLTI_W, OPC_VSLTI_D }, + { OPC_XVSLTI_B, OPC_XVSLTI_H, OPC_XVSLTI_W, OPC_XVSLTI_D }, + }, + [TCG_COND_LTU] = { + { OPC_VSLTI_BU, OPC_VSLTI_HU, OPC_VSLTI_WU, OPC_VSLTI_DU }, + { OPC_XVSLTI_BU, OPC_XVSLTI_HU, OPC_XVSLTI_WU, OPC_XVSLTI_DU }, + } }; - LoongArchInsn insn; - static const LoongArchInsn neg_vec_insn[4] = { - OPC_VNEG_B, OPC_VNEG_H, OPC_VNEG_W, OPC_VNEG_D + static const LoongArchInsn neg_vec_insn[2][4] = { + { OPC_VNEG_B, OPC_VNEG_H, OPC_VNEG_W, OPC_VNEG_D }, + { OPC_XVNEG_B, OPC_XVNEG_H, OPC_XVNEG_W, OPC_XVNEG_D }, }; - static const LoongArchInsn mul_vec_insn[4] = { - OPC_VMUL_B, OPC_VMUL_H, OPC_VMUL_W, OPC_VMUL_D + static const LoongArchInsn mul_vec_insn[2][4] = { + { OPC_VMUL_B, OPC_VMUL_H, OPC_VMUL_W, OPC_VMUL_D }, + { OPC_XVMUL_B, OPC_XVMUL_H, OPC_XVMUL_W, OPC_XVMUL_D }, }; - static const LoongArchInsn smin_vec_insn[4] = { - OPC_VMIN_B, OPC_VMIN_H, OPC_VMIN_W, OPC_VMIN_D + static const LoongArchInsn smin_vec_insn[2][4] = { + { OPC_VMIN_B, OPC_VMIN_H, OPC_VMIN_W, OPC_VMIN_D }, + { OPC_XVMIN_B, OPC_XVMIN_H, OPC_XVMIN_W, OPC_XVMIN_D }, }; - static const LoongArchInsn umin_vec_insn[4] = { - OPC_VMIN_BU, OPC_VMIN_HU, OPC_VMIN_WU, OPC_VMIN_DU + static const LoongArchInsn umin_vec_insn[2][4] = { + { OPC_VMIN_BU, OPC_VMIN_HU, OPC_VMIN_WU, OPC_VMIN_DU }, + { OPC_XVMIN_BU, OPC_XVMIN_HU, OPC_XVMIN_WU, OPC_XVMIN_DU }, }; - static const LoongArchInsn smax_vec_insn[4] = { - OPC_VMAX_B, OPC_VMAX_H, OPC_VMAX_W, OPC_VMAX_D + static const LoongArchInsn smax_vec_insn[2][4] = { + { OPC_VMAX_B, OPC_VMAX_H, OPC_VMAX_W, OPC_VMAX_D }, + { OPC_XVMAX_B, OPC_XVMAX_H, OPC_XVMAX_W, OPC_XVMAX_D }, }; - static const LoongArchInsn umax_vec_insn[4] = { - OPC_VMAX_BU, OPC_VMAX_HU, OPC_VMAX_WU, OPC_VMAX_DU + static const LoongArchInsn umax_vec_insn[2][4] = { + { OPC_VMAX_BU, OPC_VMAX_HU, OPC_VMAX_WU, OPC_VMAX_DU }, + { OPC_XVMAX_BU, OPC_XVMAX_HU, OPC_XVMAX_WU, OPC_XVMAX_DU }, }; - static const LoongArchInsn ssadd_vec_insn[4] = { - OPC_VSADD_B, OPC_VSADD_H, OPC_VSADD_W, OPC_VSADD_D + static const LoongArchInsn ssadd_vec_insn[2][4] = { + { OPC_VSADD_B, OPC_VSADD_H, OPC_VSADD_W, OPC_VSADD_D }, + { OPC_XVSADD_B, OPC_XVSADD_H, OPC_XVSADD_W, OPC_XVSADD_D }, }; - static const LoongArchInsn usadd_vec_insn[4] = { - OPC_VSADD_BU, OPC_VSADD_HU, OPC_VSADD_WU, OPC_VSADD_DU + static const LoongArchInsn usadd_vec_insn[2][4] = { + { OPC_VSADD_BU, OPC_VSADD_HU, OPC_VSADD_WU, OPC_VSADD_DU }, + { OPC_XVSADD_BU, OPC_XVSADD_HU, OPC_XVSADD_WU, OPC_XVSADD_DU }, }; - static const LoongArchInsn sssub_vec_insn[4] = { - OPC_VSSUB_B, OPC_VSSUB_H, OPC_VSSUB_W, OPC_VSSUB_D + static const LoongArchInsn sssub_vec_insn[2][4] = { + { OPC_VSSUB_B, OPC_VSSUB_H, OPC_VSSUB_W, OPC_VSSUB_D }, + { OPC_XVSSUB_B, OPC_XVSSUB_H, OPC_XVSSUB_W, OPC_XVSSUB_D }, }; - static const LoongArchInsn ussub_vec_insn[4] = { - OPC_VSSUB_BU, OPC_VSSUB_HU, OPC_VSSUB_WU, OPC_VSSUB_DU + static const LoongArchInsn ussub_vec_insn[2][4] = { + { OPC_VSSUB_BU, OPC_VSSUB_HU, OPC_VSSUB_WU, OPC_VSSUB_DU }, + { OPC_XVSSUB_BU, OPC_XVSSUB_HU, OPC_XVSSUB_WU, OPC_XVSSUB_DU }, }; - static const LoongArchInsn shlv_vec_insn[4] = { - OPC_VSLL_B, OPC_VSLL_H, OPC_VSLL_W, OPC_VSLL_D + static const LoongArchInsn shlv_vec_insn[2][4] = { + { OPC_VSLL_B, OPC_VSLL_H, OPC_VSLL_W, OPC_VSLL_D }, + { OPC_XVSLL_B, OPC_XVSLL_H, OPC_XVSLL_W, OPC_XVSLL_D }, }; - static const LoongArchInsn shrv_vec_insn[4] = { - OPC_VSRL_B, OPC_VSRL_H, OPC_VSRL_W, OPC_VSRL_D + static const LoongArchInsn shrv_vec_insn[2][4] = { + { OPC_VSRL_B, OPC_VSRL_H, OPC_VSRL_W, OPC_VSRL_D }, + { OPC_XVSRL_B, OPC_XVSRL_H, OPC_XVSRL_W, OPC_XVSRL_D }, }; - static const LoongArchInsn sarv_vec_insn[4] = { - OPC_VSRA_B, OPC_VSRA_H, OPC_VSRA_W, OPC_VSRA_D + static const LoongArchInsn sarv_vec_insn[2][4] = { + { OPC_VSRA_B, OPC_VSRA_H, OPC_VSRA_W, OPC_VSRA_D }, + { OPC_XVSRA_B, OPC_XVSRA_H, OPC_XVSRA_W, OPC_XVSRA_D }, }; - static const LoongArchInsn shli_vec_insn[4] = { - OPC_VSLLI_B, OPC_VSLLI_H, OPC_VSLLI_W, OPC_VSLLI_D + static const LoongArchInsn shli_vec_insn[2][4] = { + { OPC_VSLLI_B, OPC_VSLLI_H, OPC_VSLLI_W, OPC_VSLLI_D }, + { OPC_XVSLLI_B, OPC_XVSLLI_H, OPC_XVSLLI_W, OPC_XVSLLI_D }, }; - static const LoongArchInsn shri_vec_insn[4] = { - OPC_VSRLI_B, OPC_VSRLI_H, OPC_VSRLI_W, OPC_VSRLI_D + static const LoongArchInsn shri_vec_insn[2][4] = { + { OPC_VSRLI_B, OPC_VSRLI_H, OPC_VSRLI_W, OPC_VSRLI_D }, + { OPC_XVSRLI_B, OPC_XVSRLI_H, OPC_XVSRLI_W, OPC_XVSRLI_D }, }; - static const LoongArchInsn sari_vec_insn[4] = { - OPC_VSRAI_B, OPC_VSRAI_H, OPC_VSRAI_W, OPC_VSRAI_D + static const LoongArchInsn sari_vec_insn[2][4] = { + { OPC_VSRAI_B, OPC_VSRAI_H, OPC_VSRAI_W, OPC_VSRAI_D }, + { OPC_XVSRAI_B, OPC_XVSRAI_H, OPC_XVSRAI_W, OPC_XVSRAI_D }, }; - static const LoongArchInsn rotrv_vec_insn[4] = { - OPC_VROTR_B, OPC_VROTR_H, OPC_VROTR_W, OPC_VROTR_D + static const LoongArchInsn rotrv_vec_insn[2][4] = { + { OPC_VROTR_B, OPC_VROTR_H, OPC_VROTR_W, OPC_VROTR_D }, + { OPC_XVROTR_B, OPC_XVROTR_H, OPC_XVROTR_W, OPC_XVROTR_D }, + }; + static const LoongArchInsn rotri_vec_insn[2][4] = { + { OPC_VROTRI_B, OPC_VROTRI_H, OPC_VROTRI_W, OPC_VROTRI_D }, + { OPC_XVROTRI_B, OPC_XVROTRI_H, OPC_XVROTRI_W, OPC_XVROTRI_D }, }; a0 = args[0]; @@ -1884,9 +1957,6 @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc, a2 = args[2]; a3 = args[3]; - /* Currently only supports V128 */ - tcg_debug_assert(type == TCG_TYPE_V128); - switch (opc) { case INDEX_op_st_vec: tcg_out_st(s, type, a0, a1, a2); @@ -1895,49 +1965,55 @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc, tcg_out_ld(s, type, a0, a1, a2); break; case INDEX_op_and_vec: - tcg_out_opc_vand_v(s, a0, a1, a2); - break; + insn = lasx ? OPC_XVAND_V : OPC_VAND_V; + goto vdvjvk; case INDEX_op_andc_vec: /* * vandn vd, vj, vk: vd = vk & ~vj * andc_vec vd, vj, vk: vd = vj & ~vk - * vk and vk are swapped + * vj and vk are swapped */ - tcg_out_opc_vandn_v(s, a0, a2, a1); - break; + a1 = a2; + a2 = args[1]; + insn = lasx ? OPC_XVANDN_V : OPC_VANDN_V; + goto vdvjvk; case INDEX_op_or_vec: - tcg_out_opc_vor_v(s, a0, a1, a2); - break; + insn = lasx ? OPC_XVOR_V : OPC_VOR_V; + goto vdvjvk; case INDEX_op_orc_vec: - tcg_out_opc_vorn_v(s, a0, a1, a2); - break; + insn = lasx ? OPC_XVORN_V : OPC_VORN_V; + goto vdvjvk; case INDEX_op_xor_vec: - tcg_out_opc_vxor_v(s, a0, a1, a2); - break; - case INDEX_op_nor_vec: - tcg_out_opc_vnor_v(s, a0, a1, a2); - break; + insn = lasx ? OPC_XVXOR_V : OPC_VXOR_V; + goto vdvjvk; case INDEX_op_not_vec: - tcg_out_opc_vnor_v(s, a0, a1, a1); - break; + a2 = a1; + /* fall through */ + case INDEX_op_nor_vec: + insn = lasx ? OPC_XVNOR_V : OPC_VNOR_V; + goto vdvjvk; case INDEX_op_cmp_vec: { TCGCond cond = args[3]; + if (const_args[2]) { /* * cmp_vec dest, src, value * Try vseqi/vslei/vslti */ int64_t value = sextract64(a2, 0, 8 << vece); - if ((cond == TCG_COND_EQ || cond == TCG_COND_LE || \ - cond == TCG_COND_LT) && (-0x10 <= value && value <= 0x0f)) { - tcg_out32(s, encode_vdvjsk5_insn(cmp_vec_imm_insn[cond][vece], \ - a0, a1, value)); + if ((cond == TCG_COND_EQ || + cond == TCG_COND_LE || + cond == TCG_COND_LT) && + (-0x10 <= value && value <= 0x0f)) { + insn = cmp_vec_imm_insn[cond][lasx][vece]; + tcg_out32(s, encode_vdvjsk5_insn(insn, a0, a1, value)); break; - } else if ((cond == TCG_COND_LEU || cond == TCG_COND_LTU) && - (0x00 <= value && value <= 0x1f)) { - tcg_out32(s, encode_vdvjuk5_insn(cmp_vec_imm_insn[cond][vece], \ - a0, a1, value)); + } else if ((cond == TCG_COND_LEU || + cond == TCG_COND_LTU) && + (0x00 <= value && value <= 0x1f)) { + insn = cmp_vec_imm_insn[cond][lasx][vece]; + tcg_out32(s, encode_vdvjuk5_insn(insn, a0, a1, value)); break; } @@ -1946,113 +2022,122 @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc, * dupi_vec temp, a2 * cmp_vec a0, a1, temp, cond */ - tcg_out_dupi_vec(s, type, vece, temp_vec, a2); - a2 = temp_vec; + tcg_out_dupi_vec(s, type, vece, TCG_VEC_TMP0, a2); + a2 = TCG_VEC_TMP0; } - insn = cmp_vec_insn[cond][vece]; + insn = cmp_vec_insn[cond][lasx][vece]; if (insn == 0) { TCGArg t; t = a1, a1 = a2, a2 = t; cond = tcg_swap_cond(cond); - insn = cmp_vec_insn[cond][vece]; + insn = cmp_vec_insn[cond][lasx][vece]; tcg_debug_assert(insn != 0); } - tcg_out32(s, encode_vdvjvk_insn(insn, a0, a1, a2)); } - break; + goto vdvjvk; case INDEX_op_add_vec: - tcg_out_addsub_vec(s, vece, a0, a1, a2, const_args[2], true); + tcg_out_addsub_vec(s, lasx, vece, a0, a1, a2, const_args[2], true); break; case INDEX_op_sub_vec: - tcg_out_addsub_vec(s, vece, a0, a1, a2, const_args[2], false); + tcg_out_addsub_vec(s, lasx, vece, a0, a1, a2, const_args[2], false); break; case INDEX_op_neg_vec: - tcg_out32(s, encode_vdvj_insn(neg_vec_insn[vece], a0, a1)); + tcg_out32(s, encode_vdvj_insn(neg_vec_insn[lasx][vece], a0, a1)); break; case INDEX_op_mul_vec: - tcg_out32(s, encode_vdvjvk_insn(mul_vec_insn[vece], a0, a1, a2)); - break; + insn = mul_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_smin_vec: - tcg_out32(s, encode_vdvjvk_insn(smin_vec_insn[vece], a0, a1, a2)); - break; + insn = smin_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_smax_vec: - tcg_out32(s, encode_vdvjvk_insn(smax_vec_insn[vece], a0, a1, a2)); - break; + insn = smax_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_umin_vec: - tcg_out32(s, encode_vdvjvk_insn(umin_vec_insn[vece], a0, a1, a2)); - break; + insn = umin_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_umax_vec: - tcg_out32(s, encode_vdvjvk_insn(umax_vec_insn[vece], a0, a1, a2)); - break; + insn = umax_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_ssadd_vec: - tcg_out32(s, encode_vdvjvk_insn(ssadd_vec_insn[vece], a0, a1, a2)); - break; + insn = ssadd_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_usadd_vec: - tcg_out32(s, encode_vdvjvk_insn(usadd_vec_insn[vece], a0, a1, a2)); - break; + insn = usadd_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_sssub_vec: - tcg_out32(s, encode_vdvjvk_insn(sssub_vec_insn[vece], a0, a1, a2)); - break; + insn = sssub_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_ussub_vec: - tcg_out32(s, encode_vdvjvk_insn(ussub_vec_insn[vece], a0, a1, a2)); - break; + insn = ussub_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_shlv_vec: - tcg_out32(s, encode_vdvjvk_insn(shlv_vec_insn[vece], a0, a1, a2)); - break; + insn = shlv_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_shrv_vec: - tcg_out32(s, encode_vdvjvk_insn(shrv_vec_insn[vece], a0, a1, a2)); - break; + insn = shrv_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_sarv_vec: - tcg_out32(s, encode_vdvjvk_insn(sarv_vec_insn[vece], a0, a1, a2)); - break; + insn = sarv_vec_insn[lasx][vece]; + goto vdvjvk; + case INDEX_op_rotlv_vec: + /* rotlv_vec a1, a2 = rotrv_vec a1, -a2 */ + tcg_out32(s, encode_vdvj_insn(neg_vec_insn[lasx][vece], + TCG_VEC_TMP0, a2)); + a2 = TCG_VEC_TMP0; + /* fall through */ + case INDEX_op_rotrv_vec: + insn = rotrv_vec_insn[lasx][vece]; + goto vdvjvk; case INDEX_op_shli_vec: - tcg_out32(s, encode_vdvjuk3_insn(shli_vec_insn[vece], a0, a1, a2)); - break; + insn = shli_vec_insn[lasx][vece]; + goto vdvjukN; case INDEX_op_shri_vec: - tcg_out32(s, encode_vdvjuk3_insn(shri_vec_insn[vece], a0, a1, a2)); - break; + insn = shri_vec_insn[lasx][vece]; + goto vdvjukN; case INDEX_op_sari_vec: - tcg_out32(s, encode_vdvjuk3_insn(sari_vec_insn[vece], a0, a1, a2)); - break; - case INDEX_op_rotrv_vec: - tcg_out32(s, encode_vdvjvk_insn(rotrv_vec_insn[vece], a0, a1, a2)); - break; - case INDEX_op_rotlv_vec: - /* rotlv_vec a1, a2 = rotrv_vec a1, -a2 */ - tcg_out32(s, encode_vdvj_insn(neg_vec_insn[vece], temp_vec, a2)); - tcg_out32(s, encode_vdvjvk_insn(rotrv_vec_insn[vece], a0, a1, - temp_vec)); - break; + insn = sari_vec_insn[lasx][vece]; + goto vdvjukN; case INDEX_op_rotli_vec: /* rotli_vec a1, a2 = rotri_vec a1, -a2 */ a2 = extract32(-a2, 0, 3 + vece); + insn = rotri_vec_insn[lasx][vece]; + goto vdvjukN; + case INDEX_op_bitsel_vec: + /* vbitsel vd, vj, vk, va = bitsel_vec vd, va, vk, vj */ + if (lasx) { + tcg_out_opc_xvbitsel_v(s, a0, a3, a2, a1); + } else { + tcg_out_opc_vbitsel_v(s, a0, a3, a2, a1); + } + break; + case INDEX_op_dupm_vec: + tcg_out_dupm_vec(s, type, vece, a0, a1, a2); + break; + default: + g_assert_not_reached(); + vdvjvk: + tcg_out32(s, encode_vdvjvk_insn(insn, a0, a1, a2)); + break; + vdvjukN: switch (vece) { case MO_8: - tcg_out_opc_vrotri_b(s, a0, a1, a2); + tcg_out32(s, encode_vdvjuk3_insn(insn, a0, a1, a2)); break; case MO_16: - tcg_out_opc_vrotri_h(s, a0, a1, a2); + tcg_out32(s, encode_vdvjuk4_insn(insn, a0, a1, a2)); break; case MO_32: - tcg_out_opc_vrotri_w(s, a0, a1, a2); + tcg_out32(s, encode_vdvjuk5_insn(insn, a0, a1, a2)); break; case MO_64: - tcg_out_opc_vrotri_d(s, a0, a1, a2); + tcg_out32(s, encode_vdvjuk6_insn(insn, a0, a1, a2)); break; default: g_assert_not_reached(); } break; - case INDEX_op_bitsel_vec: - /* vbitsel vd, vj, vk, va = bitsel_vec vd, va, vk, vj */ - tcg_out_opc_vbitsel_v(s, a0, a3, a2, a1); - break; - case INDEX_op_dupm_vec: - tcg_out_dupm_vec(s, type, vece, a0, a1, a2); - break; - default: - g_assert_not_reached(); } } @@ -2398,7 +2483,11 @@ static void tcg_target_init(TCGContext *s) tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_S9); if (cpuinfo & CPUINFO_LSX) { + tcg_target_available_regs[TCG_TYPE_V64] = ALL_VECTOR_REGS; tcg_target_available_regs[TCG_TYPE_V128] = ALL_VECTOR_REGS; + if (cpuinfo & CPUINFO_LASX) { + tcg_target_available_regs[TCG_TYPE_V256] = ALL_VECTOR_REGS; + } tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V24); tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V25); tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V26); diff --git a/tcg/loongarch64/tcg-target.h b/tcg/loongarch64/tcg-target.h index 29e4860..58bd7d2 100644 --- a/tcg/loongarch64/tcg-target.h +++ b/tcg/loongarch64/tcg-target.h @@ -171,9 +171,9 @@ typedef enum { #define TCG_TARGET_HAS_tst 0 -#define TCG_TARGET_HAS_v64 0 +#define TCG_TARGET_HAS_v64 (cpuinfo & CPUINFO_LSX) #define TCG_TARGET_HAS_v128 (cpuinfo & CPUINFO_LSX) -#define TCG_TARGET_HAS_v256 0 +#define TCG_TARGET_HAS_v256 (cpuinfo & CPUINFO_LASX) #define TCG_TARGET_HAS_not_vec 1 #define TCG_TARGET_HAS_neg_vec 1 diff --git a/util/bufferiszero.c b/util/bufferiszero.c index 11c080e..522146d 100644 --- a/util/bufferiszero.c +++ b/util/bufferiszero.c @@ -81,196 +81,7 @@ static bool buffer_is_zero_int_ge256(const void *buf, size_t len) return t == 0; } -#if defined(CONFIG_AVX2_OPT) || defined(__SSE2__) -#include <immintrin.h> - -/* Helper for preventing the compiler from reassociating - chains of binary vector operations. */ -#define SSE_REASSOC_BARRIER(vec0, vec1) asm("" : "+x"(vec0), "+x"(vec1)) - -/* Note that these vectorized functions may assume len >= 256. */ - -static bool __attribute__((target("sse2"))) -buffer_zero_sse2(const void *buf, size_t len) -{ - /* Unaligned loads at head/tail. */ - __m128i v = *(__m128i_u *)(buf); - __m128i w = *(__m128i_u *)(buf + len - 16); - /* Align head/tail to 16-byte boundaries. */ - const __m128i *p = QEMU_ALIGN_PTR_DOWN(buf + 16, 16); - const __m128i *e = QEMU_ALIGN_PTR_DOWN(buf + len - 1, 16); - __m128i zero = { 0 }; - - /* Collect a partial block at tail end. */ - v |= e[-1]; w |= e[-2]; - SSE_REASSOC_BARRIER(v, w); - v |= e[-3]; w |= e[-4]; - SSE_REASSOC_BARRIER(v, w); - v |= e[-5]; w |= e[-6]; - SSE_REASSOC_BARRIER(v, w); - v |= e[-7]; v |= w; - - /* - * Loop over complete 128-byte blocks. - * With the head and tail removed, e - p >= 14, so the loop - * must iterate at least once. - */ - do { - v = _mm_cmpeq_epi8(v, zero); - if (unlikely(_mm_movemask_epi8(v) != 0xFFFF)) { - return false; - } - v = p[0]; w = p[1]; - SSE_REASSOC_BARRIER(v, w); - v |= p[2]; w |= p[3]; - SSE_REASSOC_BARRIER(v, w); - v |= p[4]; w |= p[5]; - SSE_REASSOC_BARRIER(v, w); - v |= p[6]; w |= p[7]; - SSE_REASSOC_BARRIER(v, w); - v |= w; - p += 8; - } while (p < e - 7); - - return _mm_movemask_epi8(_mm_cmpeq_epi8(v, zero)) == 0xFFFF; -} - -#ifdef CONFIG_AVX2_OPT -static bool __attribute__((target("avx2"))) -buffer_zero_avx2(const void *buf, size_t len) -{ - /* Unaligned loads at head/tail. */ - __m256i v = *(__m256i_u *)(buf); - __m256i w = *(__m256i_u *)(buf + len - 32); - /* Align head/tail to 32-byte boundaries. */ - const __m256i *p = QEMU_ALIGN_PTR_DOWN(buf + 32, 32); - const __m256i *e = QEMU_ALIGN_PTR_DOWN(buf + len - 1, 32); - __m256i zero = { 0 }; - - /* Collect a partial block at tail end. */ - v |= e[-1]; w |= e[-2]; - SSE_REASSOC_BARRIER(v, w); - v |= e[-3]; w |= e[-4]; - SSE_REASSOC_BARRIER(v, w); - v |= e[-5]; w |= e[-6]; - SSE_REASSOC_BARRIER(v, w); - v |= e[-7]; v |= w; - - /* Loop over complete 256-byte blocks. */ - for (; p < e - 7; p += 8) { - /* PTEST is not profitable here. */ - v = _mm256_cmpeq_epi8(v, zero); - if (unlikely(_mm256_movemask_epi8(v) != 0xFFFFFFFF)) { - return false; - } - v = p[0]; w = p[1]; - SSE_REASSOC_BARRIER(v, w); - v |= p[2]; w |= p[3]; - SSE_REASSOC_BARRIER(v, w); - v |= p[4]; w |= p[5]; - SSE_REASSOC_BARRIER(v, w); - v |= p[6]; w |= p[7]; - SSE_REASSOC_BARRIER(v, w); - v |= w; - } - - return _mm256_movemask_epi8(_mm256_cmpeq_epi8(v, zero)) == 0xFFFFFFFF; -} -#endif /* CONFIG_AVX2_OPT */ - -static biz_accel_fn const accel_table[] = { - buffer_is_zero_int_ge256, - buffer_zero_sse2, -#ifdef CONFIG_AVX2_OPT - buffer_zero_avx2, -#endif -}; - -static unsigned best_accel(void) -{ -#ifdef CONFIG_AVX2_OPT - unsigned info = cpuinfo_init(); - - if (info & CPUINFO_AVX2) { - return 2; - } -#endif - return 1; -} - -#elif defined(__aarch64__) && defined(__ARM_NEON) -#include <arm_neon.h> - -/* - * Helper for preventing the compiler from reassociating - * chains of binary vector operations. - */ -#define REASSOC_BARRIER(vec0, vec1) asm("" : "+w"(vec0), "+w"(vec1)) - -static bool buffer_is_zero_simd(const void *buf, size_t len) -{ - uint32x4_t t0, t1, t2, t3; - - /* Align head/tail to 16-byte boundaries. */ - const uint32x4_t *p = QEMU_ALIGN_PTR_DOWN(buf + 16, 16); - const uint32x4_t *e = QEMU_ALIGN_PTR_DOWN(buf + len - 1, 16); - - /* Unaligned loads at head/tail. */ - t0 = vld1q_u32(buf) | vld1q_u32(buf + len - 16); - - /* Collect a partial block at tail end. */ - t1 = e[-7] | e[-6]; - t2 = e[-5] | e[-4]; - t3 = e[-3] | e[-2]; - t0 |= e[-1]; - REASSOC_BARRIER(t0, t1); - REASSOC_BARRIER(t2, t3); - t0 |= t1; - t2 |= t3; - REASSOC_BARRIER(t0, t2); - t0 |= t2; - - /* - * Loop over complete 128-byte blocks. - * With the head and tail removed, e - p >= 14, so the loop - * must iterate at least once. - */ - do { - /* - * Reduce via UMAXV. Whatever the actual result, - * it will only be zero if all input bytes are zero. - */ - if (unlikely(vmaxvq_u32(t0) != 0)) { - return false; - } - - t0 = p[0] | p[1]; - t1 = p[2] | p[3]; - t2 = p[4] | p[5]; - t3 = p[6] | p[7]; - REASSOC_BARRIER(t0, t1); - REASSOC_BARRIER(t2, t3); - t0 |= t1; - t2 |= t3; - REASSOC_BARRIER(t0, t2); - t0 |= t2; - p += 8; - } while (p < e - 7); - - return vmaxvq_u32(t0) == 0; -} - -#define best_accel() 1 -static biz_accel_fn const accel_table[] = { - buffer_is_zero_int_ge256, - buffer_is_zero_simd, -}; -#else -#define best_accel() 0 -static biz_accel_fn const accel_table[1] = { - buffer_is_zero_int_ge256 -}; -#endif +#include "host/bufferiszero.c.inc" static biz_accel_fn buffer_is_zero_accel; static unsigned accel_index; diff --git a/util/cpuinfo-loongarch.c b/util/cpuinfo-loongarch.c index 08b6d74..bb1f7f6 100644 --- a/util/cpuinfo-loongarch.c +++ b/util/cpuinfo-loongarch.c @@ -29,6 +29,7 @@ unsigned __attribute__((constructor)) cpuinfo_init(void) info = CPUINFO_ALWAYS; info |= (hwcap & HWCAP_LOONGARCH_LSX ? CPUINFO_LSX : 0); + info |= (hwcap & HWCAP_LOONGARCH_LASX ? CPUINFO_LASX : 0); cpuinfo = info; return info; |