diff options
author | Will Schmidt <will_schmidt@vnet.ibm.com> | 2021-04-12 13:17:56 -0500 |
---|---|---|
committer | Will Schmidt <will_schmidt@vnet.ibm.com> | 2021-04-12 13:17:56 -0500 |
commit | 82d9b2804751252a31342ab5a501ef6c96f4548a (patch) | |
tree | 295b3e87504b45d7614f5a8b8fc890f8d3248cbc /gdb | |
parent | b722acca420eae2ec927591aaad2d866c282b535 (diff) | |
download | gdb-82d9b2804751252a31342ab5a501ef6c96f4548a.zip gdb-82d9b2804751252a31342ab5a501ef6c96f4548a.tar.gz gdb-82d9b2804751252a31342ab5a501ef6c96f4548a.tar.bz2 |
[rs6000] Create a powerpc-power10.exp test
Inspired by the existing powerpc-power9.exp test, this is a
new test to cover the power10 instruction disassembly.
gdb/testsuite/ChangeLog:
2021-04-12 Will Schmidt <will_schmidt@vnet.ibm.com>
* gdb.arch/powerpc-power10.s: New test for instructions.
* gdb.arch/powerpc-power10.exp: Harness to run the test.
Diffstat (limited to 'gdb')
-rw-r--r-- | gdb/testsuite/ChangeLog | 5 | ||||
-rw-r--r-- | gdb/testsuite/gdb.arch/powerpc-power10.exp | 686 | ||||
-rw-r--r-- | gdb/testsuite/gdb.arch/powerpc-power10.s | 639 |
3 files changed, 1330 insertions, 0 deletions
diff --git a/gdb/testsuite/ChangeLog b/gdb/testsuite/ChangeLog index f833782..f097a02 100644 --- a/gdb/testsuite/ChangeLog +++ b/gdb/testsuite/ChangeLog @@ -1,3 +1,8 @@ +2021-04-12 Will Schmidt <will_schmidt@vnet.ibm.com> + + * gdb.arch/powerpc-power10.s: New test for instructions. + * gdb.arch/powerpc-power10.exp: Harness to run the test. + 2021-4-12 Carl Love <cel@us.ibm.com> * gdb.arch/vsx-vsr-float128.c: New test file. diff --git a/gdb/testsuite/gdb.arch/powerpc-power10.exp b/gdb/testsuite/gdb.arch/powerpc-power10.exp new file mode 100644 index 0000000..16297cc --- /dev/null +++ b/gdb/testsuite/gdb.arch/powerpc-power10.exp @@ -0,0 +1,686 @@ +# Copyright 2021 Free Software Foundation,Inc. + +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 3 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see <http://www.gnu.org/licenses/>. + +# This file is part of the gdb testsuite. + +# Test PowerPC instructions disassembly of power10 instructions. + +standard_testfile .s +set objfile [standard_output_file ${testfile}.o] + +if {![istarget "powerpc*-*-*"]} then { + verbose "Skipping PowerPC instructions disassembly." + return +} + +if { [gdb_compile "${srcdir}/${subdir}/${srcfile}" "${objfile}" object {debug}] != "" } { + untested "PowerPC instructions disassembly" + return -1 +} + +clean_restart ${objfile} + +# Disassemble the function. +set func "" + +gdb_test "set disassembler-options power10" +set test "disass func" +gdb_test_multiple $test $test { + -re "\r\nDump of assembler code for function func:(\r\n.*\r\n)End of assembler dump.\r\n$gdb_prompt $" { + set func $expect_out(1,string) + pass $test + } +} + +proc instr_to_patt {instr} { + global decimal hex + # 0x0000000000000018 <+24>: stxvd2x vs43,r4,r5 + return "\[ \t\]*$hex <\\+$decimal>:\[ \t\]*[string_to_regexp $instr]" +} + +proc func_check {instr} { + global func + + set test "found: $instr" + if [regexp -nocase -line [instr_to_patt $instr] $func] { + pass $test + } else { + fail $test + } +} + +func_check "brd r0,r1" +func_check "brh r0,r1" +func_check "brw r0,r1" +func_check "cfuged r0,r1,r2" +func_check "cntlzdm r0,r1,r2" +func_check "cnttzdm r0,r1,r2" +func_check "dcffixqq f28,v0" +func_check "dctfixqq v0,f26" +func_check "lxvkq vs0,1" +func_check "lxvkq vs0,16" +func_check "lxvkq vs0,17" +func_check "lxvkq vs0,18" +func_check "lxvkq vs0,19" +func_check "lxvkq vs0,2" +func_check "lxvkq vs0,20" +func_check "lxvkq vs0,21" +func_check "lxvkq vs0,22" +func_check "lxvkq vs0,23" +func_check "lxvkq vs0,24" +func_check "lxvkq vs0,3" +func_check "lxvkq vs0,4" +func_check "lxvkq vs0,5" +func_check "lxvkq vs0,6" +func_check "lxvkq vs0,7" +func_check "lxvkq vs0,8" +func_check "lxvkq vs0,9" +func_check "lxvp vs20,0(0)" +func_check "lxvp vs20,16(0)" +func_check "lxvp vs20,32(0)" +func_check "lxvpx vs20,0,r1" +func_check "lxvrbx vs0,r1,r2" +func_check "lxvrdx vs0,r1,r2" +func_check "lxvrhx vs0,r1,r2" +func_check "lxvrwx vs0,r1,r2" +func_check "mfvsrld r0,vs26" +func_check "mfvsrld r0,vs27" +func_check "mfvsrld r0,vs28" +func_check "mfvsrld r0,vs29" +func_check "mffprd r0,f26" +func_check "mffprd r0,f27" +func_check "mffprd r0,f28" +func_check "mffprd r0,f29" +func_check "mtvsrbm v0,r1" +func_check "mtvsrbmi v0,0" +func_check "mtvsrbmi v0,3" +func_check "mtvsrbmi v0,7" +func_check "mtvsrdm v0,r1" +func_check "mtvsrhm v0,r1" +func_check "mtvsrqm v0,r1" +func_check "mtvsrwm v0,r1" +func_check "paddi r0,r1,0" +func_check "paddi r0,r1,12" +func_check "paddi r0,r1,48" +func_check "paddi r0,r1,98" +func_check "pdepd r0,r1,r2" +func_check "pextd r0,r1,r2" +func_check "plbz r0,0(r1)" +func_check "plbz r0,16(r1)" +func_check "plbz r0,32(r1)" +func_check "plbz r0,64(r1)" +func_check "plbz r0,8(r1)" +func_check "pld r0,0(r1)" +func_check "pld r0,16(r1)" +func_check "pld r0,32(r1)" +func_check "pld r0,64(r1)" +func_check "pld r0,8(r1)" +func_check "plfd f28,0(0)" +func_check "plfd f28,16(0)" +func_check "plfd f28,32(0)" +func_check "plfd f28,4(0)" +func_check "plfd f28,64(0)" +func_check "plfd f28,8(0)" +func_check "plfs f28,0(0)" +func_check "plfs f28,16(0)" +func_check "plfs f28,32(0)" +func_check "plfs f28,4(0)" +func_check "plfs f28,64(0)" +func_check "plfs f28,8(0)" +func_check "plha r0,0(r1)" +func_check "plha r0,16(r1)" +func_check "plha r0,32(r1)" +func_check "plha r0,64(r1)" +func_check "plha r0,8(r1)" +func_check "plhz r0,0(r1)" +func_check "plhz r0,16(r1)" +func_check "plhz r0,32(r1)" +func_check "plhz r0,64(r1)" +func_check "plhz r0,8(r1)" +func_check "plq r26,0(0)" +func_check "plq r26,16(0)" +func_check "plq r26,32(0)" +func_check "plq r26,48(0)" +func_check "plq r26,64(0)" +func_check "plq r26,8(0)" +func_check "plwa r0,0(r1)" +func_check "plwa r0,16(r1)" +func_check "plwa r0,32(r1)" +func_check "plwa r0,64(r1)" +func_check "plwa r0,8(r1)" +func_check "plwz r0,0(r1)" +func_check "plwz r0,16(r1)" +func_check "plwz r0,32(r1)" +func_check "plwz r0,64(r1)" +func_check "plwz r0,8(r1)" +func_check "plxsd v0,0(r1)" +func_check "plxsd v0,16(r1)" +func_check "plxsd v0,32(r1)" +func_check "plxsd v0,4(r1)" +func_check "plxsd v0,64(r1)" +func_check "plxsd v0,8(r1)" +func_check "plxssp v0,0(r1)" +func_check "plxssp v0,16(r1)" +func_check "plxssp v0,32(r1)" +func_check "plxssp v0,4(r1)" +func_check "plxssp v0,64(r1)" +func_check "plxssp v0,8(r1)" +func_check "plxv vs0,0(r1)" +func_check "plxv vs0,16(r1)" +func_check "plxv vs0,4(r1)" +func_check "plxv vs0,8(r1)" +func_check "plxvp vs20,0(0)" +func_check "plxvp vs20,16(0)" +func_check "plxvp vs20,24(0)" +func_check "plxvp vs20,32(0)" +func_check "plxvp vs20,8(0)" +func_check "pmxvbf16ger2 a4,vs0,vs1,0,0,0" +func_check "pmxvbf16ger2 a4,vs0,vs1,0,0,1" +func_check "pmxvbf16ger2 a4,vs0,vs1,0,13,0" +func_check "pmxvbf16ger2 a4,vs0,vs1,0,13,1" +func_check "pmxvbf16ger2 a4,vs0,vs1,11,0,0" +func_check "pmxvbf16ger2 a4,vs0,vs1,11,0,1" +func_check "pmxvbf16ger2 a4,vs0,vs1,11,13,0" +func_check "pmxvbf16ger2 a4,vs0,vs1,11,13,1" +func_check "pmxvbf16ger2nn a4,vs0,vs1,0,0,0" +func_check "pmxvbf16ger2nn a4,vs0,vs1,0,0,1" +func_check "pmxvbf16ger2nn a4,vs0,vs1,0,13,0" +func_check "pmxvbf16ger2nn a4,vs0,vs1,0,13,1" +func_check "pmxvbf16ger2nn a4,vs0,vs1,11,0,0" +func_check "pmxvbf16ger2nn a4,vs0,vs1,11,0,1" +func_check "pmxvbf16ger2nn a4,vs0,vs1,11,13,0" +func_check "pmxvbf16ger2nn a4,vs0,vs1,11,13,1" +func_check "pmxvbf16ger2np a4,vs0,vs1,0,0,0" +func_check "pmxvbf16ger2np a4,vs0,vs1,0,0,1" +func_check "pmxvbf16ger2np a4,vs0,vs1,0,13,0" +func_check "pmxvbf16ger2np a4,vs0,vs1,0,13,1" +func_check "pmxvbf16ger2np a4,vs0,vs1,11,0,0" +func_check "pmxvbf16ger2np a4,vs0,vs1,11,0,1" +func_check "pmxvbf16ger2np a4,vs0,vs1,11,13,0" +func_check "pmxvbf16ger2np a4,vs0,vs1,11,13,1" +func_check "pmxvbf16ger2pn a4,vs0,vs1,0,0,0" +func_check "pmxvbf16ger2pn a4,vs0,vs1,0,0,1" +func_check "pmxvbf16ger2pn a4,vs0,vs1,0,13,0" +func_check "pmxvbf16ger2pn a4,vs0,vs1,0,13,1" +func_check "pmxvbf16ger2pn a4,vs0,vs1,11,0,0" +func_check "pmxvbf16ger2pn a4,vs0,vs1,11,0,1" +func_check "pmxvbf16ger2pn a4,vs0,vs1,11,13,0" +func_check "pmxvbf16ger2pn a4,vs0,vs1,11,13,1" +func_check "pmxvbf16ger2pp a4,vs0,vs1,0,0,0" +func_check "pmxvbf16ger2pp a4,vs0,vs1,0,0,1" +func_check "pmxvbf16ger2pp a4,vs0,vs1,0,13,0" +func_check "pmxvbf16ger2pp a4,vs0,vs1,0,13,1" +func_check "pmxvbf16ger2pp a4,vs0,vs1,11,0,0" +func_check "pmxvbf16ger2pp a4,vs0,vs1,11,0,1" +func_check "pmxvbf16ger2pp a4,vs0,vs1,11,13,0" +func_check "pmxvbf16ger2pp a4,vs0,vs1,11,13,1" +func_check "pmxvf16ger2 a4,vs0,vs1,0,0,0" +func_check "pmxvf16ger2 a4,vs0,vs1,0,0,1" +func_check "pmxvf16ger2 a4,vs0,vs1,0,13,0" +func_check "pmxvf16ger2 a4,vs0,vs1,0,13,1" +func_check "pmxvf16ger2 a4,vs0,vs1,11,0,0" +func_check "pmxvf16ger2 a4,vs0,vs1,11,0,1" +func_check "pmxvf16ger2 a4,vs0,vs1,11,13,0" +func_check "pmxvf16ger2 a4,vs0,vs1,11,13,1" +func_check "pmxvf16ger2nn a4,vs0,vs1,0,0,0" +func_check "pmxvf16ger2nn a4,vs0,vs1,0,0,1" +func_check "pmxvf16ger2nn a4,vs0,vs1,0,13,0" +func_check "pmxvf16ger2nn a4,vs0,vs1,0,13,1" +func_check "pmxvf16ger2nn a4,vs0,vs1,11,0,0" +func_check "pmxvf16ger2nn a4,vs0,vs1,11,0,1" +func_check "pmxvf16ger2nn a4,vs0,vs1,11,13,0" +func_check "pmxvf16ger2nn a4,vs0,vs1,11,13,1" +func_check "pmxvf16ger2np a4,vs0,vs1,0,0,0" +func_check "pmxvf16ger2np a4,vs0,vs1,0,0,1" +func_check "pmxvf16ger2np a4,vs0,vs1,0,13,0" +func_check "pmxvf16ger2np a4,vs0,vs1,0,13,1" +func_check "pmxvf16ger2np a4,vs0,vs1,11,0,0" +func_check "pmxvf16ger2np a4,vs0,vs1,11,0,1" +func_check "pmxvf16ger2np a4,vs0,vs1,11,13,0" +func_check "pmxvf16ger2np a4,vs0,vs1,11,13,1" +func_check "pmxvf16ger2pn a4,vs0,vs1,0,0,0" +func_check "pmxvf16ger2pn a4,vs0,vs1,0,0,1" +func_check "pmxvf16ger2pn a4,vs0,vs1,0,13,0" +func_check "pmxvf16ger2pn a4,vs0,vs1,0,13,1" +func_check "pmxvf16ger2pn a4,vs0,vs1,11,0,0" +func_check "pmxvf16ger2pn a4,vs0,vs1,11,0,1" +func_check "pmxvf16ger2pn a4,vs0,vs1,11,13,0" +func_check "pmxvf16ger2pn a4,vs0,vs1,11,13,1" +func_check "pmxvf16ger2pp a4,vs0,vs1,0,0,0" +func_check "pmxvf16ger2pp a4,vs0,vs1,0,0,1" +func_check "pmxvf16ger2pp a4,vs0,vs1,0,13,0" +func_check "pmxvf16ger2pp a4,vs0,vs1,0,13,1" +func_check "pmxvf16ger2pp a4,vs0,vs1,11,0,0" +func_check "pmxvf16ger2pp a4,vs0,vs1,11,0,1" +func_check "pmxvf16ger2pp a4,vs0,vs1,11,13,0" +func_check "pmxvf16ger2pp a4,vs0,vs1,11,13,1" +func_check "pmxvf32ger a4,vs0,vs1,0,0" +func_check "pmxvf32ger a4,vs0,vs1,0,13" +func_check "pmxvf32ger a4,vs0,vs1,11,0" +func_check "pmxvf32ger a4,vs0,vs1,11,13" +func_check "pmxvf32gernn a4,vs0,vs1,0,0" +func_check "pmxvf32gernn a4,vs0,vs1,0,13" +func_check "pmxvf32gernn a4,vs0,vs1,11,0" +func_check "pmxvf32gernn a4,vs0,vs1,11,13" +func_check "pmxvf32gernp a4,vs0,vs1,0,0" +func_check "pmxvf32gernp a4,vs0,vs1,0,13" +func_check "pmxvf32gernp a4,vs0,vs1,11,0" +func_check "pmxvf32gernp a4,vs0,vs1,11,13" +func_check "pmxvf32gerpn a4,vs0,vs1,0,0" +func_check "pmxvf32gerpn a4,vs0,vs1,0,13" +func_check "pmxvf32gerpn a4,vs0,vs1,11,0" +func_check "pmxvf32gerpn a4,vs0,vs1,11,13" +func_check "pmxvf32gerpp a4,vs0,vs1,0,0" +func_check "pmxvf32gerpp a4,vs0,vs1,0,13" +func_check "pmxvf32gerpp a4,vs0,vs1,11,0" +func_check "pmxvf32gerpp a4,vs0,vs1,11,13" +func_check "pmxvf64ger a4,vs22,vs0,0,0" +func_check "pmxvf64ger a4,vs22,vs0,0,1" +func_check "pmxvf64ger a4,vs22,vs0,11,0" +func_check "pmxvf64ger a4,vs22,vs0,11,1" +func_check "pmxvf64gernn a4,vs22,vs0,0,0" +func_check "pmxvf64gernn a4,vs22,vs0,0,1" +func_check "pmxvf64gernn a4,vs22,vs0,11,0" +func_check "pmxvf64gernn a4,vs22,vs0,11,1" +func_check "pmxvf64gernp a4,vs22,vs0,0,0" +func_check "pmxvf64gernp a4,vs22,vs0,0,1" +func_check "pmxvf64gernp a4,vs22,vs0,11,0" +func_check "pmxvf64gernp a4,vs22,vs0,11,1" +func_check "pmxvf64gerpn a4,vs22,vs0,0,0" +func_check "pmxvf64gerpn a4,vs22,vs0,0,1" +func_check "pmxvf64gerpn a4,vs22,vs0,11,0" +func_check "pmxvf64gerpn a4,vs22,vs0,11,1" +func_check "pmxvf64gerpp a4,vs22,vs0,0,0" +func_check "pmxvf64gerpp a4,vs22,vs0,0,1" +func_check "pmxvf64gerpp a4,vs22,vs0,11,0" +func_check "pmxvf64gerpp a4,vs22,vs0,11,1" +func_check "pmxvi16ger2 a4,vs0,vs1,0,0,0" +func_check "pmxvi16ger2 a4,vs0,vs1,0,0,1" +func_check "pmxvi16ger2 a4,vs0,vs1,0,13,0" +func_check "pmxvi16ger2 a4,vs0,vs1,0,13,1" +func_check "pmxvi16ger2 a4,vs0,vs1,11,0,0" +func_check "pmxvi16ger2 a4,vs0,vs1,11,0,1" +func_check "pmxvi16ger2 a4,vs0,vs1,11,13,0" +func_check "pmxvi16ger2 a4,vs0,vs1,11,13,1" +func_check "pmxvi16ger2pp a4,vs0,vs1,0,0,0" +func_check "pmxvi16ger2pp a4,vs0,vs1,0,0,1" +func_check "pmxvi16ger2pp a4,vs0,vs1,0,13,0" +func_check "pmxvi16ger2pp a4,vs0,vs1,0,13,1" +func_check "pmxvi16ger2pp a4,vs0,vs1,11,0,0" +func_check "pmxvi16ger2pp a4,vs0,vs1,11,0,1" +func_check "pmxvi16ger2pp a4,vs0,vs1,11,13,0" +func_check "pmxvi16ger2pp a4,vs0,vs1,11,13,1" +func_check "pmxvi16ger2s a4,vs0,vs1,0,0,0" +func_check "pmxvi16ger2s a4,vs0,vs1,0,0,1" +func_check "pmxvi16ger2s a4,vs0,vs1,0,13,0" +func_check "pmxvi16ger2s a4,vs0,vs1,0,13,1" +func_check "pmxvi16ger2s a4,vs0,vs1,11,0,0" +func_check "pmxvi16ger2s a4,vs0,vs1,11,0,1" +func_check "pmxvi16ger2s a4,vs0,vs1,11,13,0" +func_check "pmxvi16ger2s a4,vs0,vs1,11,13,1" +func_check "pmxvi16ger2spp a4,vs0,vs1,0,0,0" +func_check "pmxvi16ger2spp a4,vs0,vs1,0,0,1" +func_check "pmxvi16ger2spp a4,vs0,vs1,0,13,0" +func_check "pmxvi16ger2spp a4,vs0,vs1,0,13,1" +func_check "pmxvi16ger2spp a4,vs0,vs1,11,0,0" +func_check "pmxvi16ger2spp a4,vs0,vs1,11,0,1" +func_check "pmxvi16ger2spp a4,vs0,vs1,11,13,0" +func_check "pmxvi16ger2spp a4,vs0,vs1,11,13,1" +func_check "pmxvi4ger8 a4,vs0,vs1,0,0,0" +func_check "pmxvi4ger8 a4,vs0,vs1,0,0,45" +func_check "pmxvi4ger8 a4,vs0,vs1,0,1,0" +func_check "pmxvi4ger8 a4,vs0,vs1,0,1,45" +func_check "pmxvi4ger8 a4,vs0,vs1,11,0,0" +func_check "pmxvi4ger8 a4,vs0,vs1,11,0,45" +func_check "pmxvi4ger8 a4,vs0,vs1,11,1,0" +func_check "pmxvi4ger8 a4,vs0,vs1,11,1,45" +func_check "pmxvi4ger8pp a4,vs0,vs1,0,0,0" +func_check "pmxvi4ger8pp a4,vs0,vs1,0,0,45" +func_check "pmxvi4ger8pp a4,vs0,vs1,0,1,0" +func_check "pmxvi4ger8pp a4,vs0,vs1,0,1,45" +func_check "pmxvi4ger8pp a4,vs0,vs1,11,0,0" +func_check "pmxvi4ger8pp a4,vs0,vs1,11,0,45" +func_check "pmxvi4ger8pp a4,vs0,vs1,11,1,0" +func_check "pmxvi4ger8pp a4,vs0,vs1,11,1,45" +func_check "pmxvi8ger4 a4,vs0,vs1,0,0,0" +func_check "pmxvi8ger4 a4,vs0,vs1,0,0,5" +func_check "pmxvi8ger4 a4,vs0,vs1,0,13,0" +func_check "pmxvi8ger4 a4,vs0,vs1,0,13,5" +func_check "pmxvi8ger4 a4,vs0,vs1,11,0,0" +func_check "pmxvi8ger4 a4,vs0,vs1,11,0,5" +func_check "pmxvi8ger4 a4,vs0,vs1,11,13,0" +func_check "pmxvi8ger4 a4,vs0,vs1,11,13,5" +func_check "pmxvi8ger4pp a4,vs0,vs1,0,0,0" +func_check "pmxvi8ger4pp a4,vs0,vs1,0,0,5" +func_check "pmxvi8ger4pp a4,vs0,vs1,0,13,0" +func_check "pmxvi8ger4pp a4,vs0,vs1,0,13,5" +func_check "pmxvi8ger4pp a4,vs0,vs1,11,0,0" +func_check "pmxvi8ger4pp a4,vs0,vs1,11,0,5" +func_check "pmxvi8ger4pp a4,vs0,vs1,11,13,0" +func_check "pmxvi8ger4pp a4,vs0,vs1,11,13,5" +func_check "pmxvi8ger4spp a4,vs0,vs1,0,0,0" +func_check "pmxvi8ger4spp a4,vs0,vs1,0,0,5" +func_check "pmxvi8ger4spp a4,vs0,vs1,0,13,0" +func_check "pmxvi8ger4spp a4,vs0,vs1,0,13,5" +func_check "pmxvi8ger4spp a4,vs0,vs1,11,0,0" +func_check "pmxvi8ger4spp a4,vs0,vs1,11,0,5" +func_check "pmxvi8ger4spp a4,vs0,vs1,11,13,0" +func_check "pmxvi8ger4spp a4,vs0,vs1,11,13,5" +#/* pstb extended mnemonics can suppress (r1) or the trailing ,0 or ,1, see ISA. +func_check "pstb r0,0(r1)" +func_check "pstb r0,16(r1)" +func_check "pstb r0,32(r1)" +func_check "pstb r0,8(r1)" +func_check "pstd r0,0(r1)" +func_check "pstd r0,16(r1)" +func_check "pstd r0,32(r1)" +func_check "pstd r0,8(r1)" +func_check "pstfd f26,0(0)" +func_check "pstfd f26,16(0)" +func_check "pstfd f26,32(0)" +func_check "pstfd f26,4(0)" +func_check "pstfd f26,8(0)" +func_check "pstfs f26,0(0)" +func_check "pstfs f26,16(0)" +func_check "pstfs f26,32(0)" +func_check "pstfs f26,4(0)" +func_check "pstfs f26,8(0)" +func_check "psth r0,0(r1)" +func_check "psth r0,16(r1)" +func_check "psth r0,32(r1)" +func_check "psth r0,8(r1)" +func_check "pstq r24,0(0)" +func_check "pstq r24,16(0)" +func_check "pstq r24,32(0)" +func_check "pstq r24,64(0)" +func_check "pstq r24,8(0)" +func_check "pstw r0,0(r1)" +func_check "pstw r0,16(r1)" +func_check "pstw r0,32(r1)" +func_check "pstw r0,8(r1)" +func_check "pstxsd v22,0(0)" +func_check "pstxsd v22,16(0)" +func_check "pstxsd v22,32(0)" +func_check "pstxsd v22,4(0)" +func_check "pstxsd v22,64(0)" +func_check "pstxsd v22,8(0)" +func_check "pstxssp v22,0(0)" +func_check "pstxssp v22,16(0)" +func_check "pstxssp v22,32(0)" +func_check "pstxssp v22,4(0)" +func_check "pstxssp v22,64(0)" +func_check "pstxssp v22,8(0)" +func_check "pstxv vs0,0(r1)" +func_check "pstxv vs0,16(r1)" +func_check "pstxv vs0,4(r1)" +func_check "pstxv vs0,8(r1)" +func_check "pstxvp vs20,0(0)" +func_check "pstxvp vs20,16(0)" +func_check "pstxvp vs20,32(0)" +func_check "pstxvp vs20,48(0)" +# lt,gt,eq,so map to cr0 0,1,2,3 ; cr1,cr2 extend from there. lt=0 ; gt=1 ; eq=2 ; so=3 +func_check "setbc r26,lt" +func_check "setbc r26,gt" +func_check "setbc r26,eq" +func_check "setbc r26,so" +func_check "setbc r26,4*cr1+so" +func_check "setbc r26,4*cr2+lt" +func_check "setbc r26,4*cr7+so" + +func_check "setbcr r26,4*cr7+so" +func_check "setbcr r26,4*cr1+so" +func_check "setbcr r26,4*cr2+lt" + +func_check "setnbc r26,lt" +func_check "setnbc r26,so" +func_check "setnbc r26,4*cr7+so" +func_check "setnbc r26,4*cr1+so" +func_check "setnbc r26,4*cr2+lt" + +func_check "setnbcr r26,lt" +func_check "setnbcr r26,so" +func_check "setnbcr r26,4*cr7+so" +func_check "setnbcr r26,4*cr1+so" +func_check "setnbcr r26,4*cr2+lt" + +func_check "stxvp vs20,0(0)" +func_check "stxvp vs20,16(0)" +func_check "stxvp vs20,32(0)" +func_check "stxvp vs20,48(0)" +func_check "stxvpx vs20,0,r1" +func_check "stxvrbx vs0,r1,r2" +func_check "stxvrdx vs0,r1,r2" +func_check "stxvrhx vs0,r1,r2" +func_check "stxvrwx vs0,r1,r2" +func_check "vcfuged v0,v1,v2" +func_check "vclrlb v0,v1,r2" +func_check "vclrrb v0,v1,r2" +func_check "vclzdm v0,v1,v2" +func_check "vcmpequq v0,v1,v2" +func_check "vcmpgtsq v0,v1,v2" +func_check "vcmpgtuq v0,v1,v2" +func_check "vcmpsq cr3,v0,v1" +func_check "vcmpuq cr3,v0,v1" +func_check "vcntmbb r0,v1,0" +func_check "vcntmbb r0,v1,1" +func_check "vcntmbd r0,v1,0" +func_check "vcntmbd r0,v1,1" +func_check "vcntmbh r0,v1,0" +func_check "vcntmbh r0,v1,1" +func_check "vcntmbw r0,v1,0" +func_check "vcntmbw r0,v1,1" +func_check "vctzdm v0,v1,v2" +func_check "vdivesd v0,v1,v2" +func_check "vdivesq v0,v1,v2" +func_check "vdivesw v0,v1,v2" +func_check "vdiveud v0,v1,v2" +func_check "vdiveuq v0,v1,v2" +func_check "vdiveuw v0,v1,v2" +func_check "vdivsd v0,v1,v2" +func_check "vdivsq v0,v1,v2" +func_check "vdivsw v0,v1,v2" +func_check "vdivud v0,v1,v2" +func_check "vdivuq v0,v1,v2" +func_check "vdivuw v0,v1,v2" +func_check "vexpandbm v0,v1" +func_check "vexpanddm v0,v1" +func_check "vexpandhm v0,v1" +func_check "vexpandqm v0,v1" +func_check "vexpandwm v0,v1" +func_check "vextddvlx v0,v1,v2,r3" +func_check "vextddvrx v0,v1,v2,r3" +func_check "vextdubvlx v0,v1,v2,r3" +func_check "vextdubvrx v0,v1,v2,r3" +func_check "vextduhvlx v0,v1,v2,r3" +func_check "vextduhvrx v0,v1,v2,r3" +func_check "vextduwvlx v0,v1,v2,r3" +func_check "vextduwvrx v0,v1,v2,r3" +func_check "vextractbm r0,v1" +func_check "vextractdm r0,v1" +func_check "vextracthm r0,v1" +func_check "vextractqm r0,v1" +func_check "vextractwm r0,v1" +func_check "vextsd2q v0,v1" +func_check "vgnb r0,v1,2" +func_check "vgnb r0,v1,3" +func_check "vgnb r0,v1,4" +func_check "vgnb r0,v1,5" +func_check "vgnb r0,v1,6" +func_check "vgnb r0,v1,7" +func_check "vinsblx v0,r1,r2" +func_check "vinsbrx v0,r1,r2" +func_check "vinsbvlx v0,r1,v2" +func_check "vinsbvrx v0,r1,v2" +func_check "vinsd v0,r1,3" +func_check "vinsd v0,r1,7" +func_check "vinsdlx v0,r1,r2" +func_check "vinsdrx v0,r1,r2" +func_check "vinshlx v0,r1,r2" +func_check "vinshrx v0,r1,r2" +func_check "vinshvlx v0,r1,v2" +func_check "vinshvrx v0,r1,v2" +func_check "vinsw v0,r1,3" +func_check "vinsw v0,r1,7" +func_check "vinswlx v0,r1,r2" +func_check "vinswrx v0,r1,r2" +func_check "vinswvlx v0,r1,v2" +func_check "vinswvrx v0,r1,v2" +func_check "vmodsd v0,v1,v2" +func_check "vmodsq v0,v1,v2" +func_check "vmodsw v0,v1,v2" +func_check "vmodud v0,v1,v2" +func_check "vmoduq v0,v1,v2" +func_check "vmoduw v0,v1,v2" +func_check "vmsumcud v0,v1,v2,v3" +func_check "vmulesd v0,v1,v2" +func_check "vmuleud v0,v1,v2" +func_check "vmulhsd v0,v1,v2" +func_check "vmulhsw v0,v1,v2" +func_check "vmulhud v0,v1,v2" +func_check "vmulhuw v0,v1,v2" +func_check "vmulld v0,v1,v2" +func_check "vmulosd v0,v1,v2" +func_check "vmuloud v0,v1,v2" +func_check "vpdepd v0,v1,v2" +func_check "vpextd v0,v1,v2" +func_check "vrlq v0,v1,v2" +func_check "vrlqmi v0,v1,v2" +func_check "vrlqnm v0,v1,v2" +func_check "vsldbi v0,v1,v2,0" +func_check "vsldbi v0,v1,v2,4" +func_check "vslq v0,v1,v2" +func_check "vsraq v0,v1,v2" +func_check "vsrdbi v0,v1,v2,0" +func_check "vsrdbi v0,v1,v2,4" +func_check "vsrq v0,v1,v2" +func_check "vstribl v0,v1" +func_check "vstribl. v0,v1" +func_check "vstribr v0,v1" +func_check "vstribr. v0,v1" +func_check "vstrihl v0,v1" +func_check "vstrihl. v0,v1" +func_check "vstrihr v0,v1" +func_check "vstrihr. v0,v1" +func_check "xscmpeqqp v0,v1,v2" +func_check "xscmpgeqp v0,v1,v2" +func_check "xscmpgtqp v0,v1,v2" +func_check "xscvqpsqz v0,v1" +func_check "xscvqpuqz v0,v1" +func_check "xscvsqqp v0,v1" +func_check "xscvuqqp v0,v1" +func_check "xsmaxcqp v0,v1,v2" +func_check "xsmincqp v0,v1,v2" +func_check "xvbf16ger2 a4,vs0,vs1" +func_check "xvbf16ger2nn a4,vs0,vs1" +func_check "xvbf16ger2np a4,vs0,vs1" +func_check "xvbf16ger2pn a4,vs0,vs1" +func_check "xvbf16ger2pp a4,vs0,vs1" +func_check "xvcvbf16spn vs0,vs1" +func_check "xvcvspbf16 vs0,vs1" +func_check "xvf16ger2 a4,vs0,vs1" +func_check "xvf16ger2nn a4,vs0,vs1" +func_check "xvf16ger2np a4,vs0,vs1" +func_check "xvf16ger2pn a4,vs0,vs1" +func_check "xvf16ger2pp a4,vs0,vs1" +func_check "xvf32ger a4,vs0,vs1" +func_check "xvf32gernn a4,vs0,vs1" +func_check "xvf32gernp a4,vs0,vs1" +func_check "xvf32gerpn a4,vs0,vs1" +func_check "xvf32gerpp a4,vs0,vs1" +func_check "xvf64ger a4,vs22,vs0" +func_check "xvf64gernn a4,vs22,vs0" +func_check "xvf64gernp a4,vs22,vs0" +func_check "xvf64gerpn a4,vs22,vs0" +func_check "xvf64gerpp a4,vs22,vs0" +func_check "xvi16ger2 a4,vs0,vs1" +func_check "xvi16ger2pp a4,vs0,vs1" +func_check "xvi16ger2s a4,vs0,vs1" +func_check "xvi16ger2spp a4,vs0,vs1" +func_check "xvi4ger8 a4,vs0,vs1" +func_check "xvi4ger8pp a4,vs0,vs1" +func_check "xvi8ger4 a4,vs0,vs1" +func_check "xvi8ger4pp a4,vs0,vs1" +func_check "xvi8ger4spp a4,vs0,vs1" +func_check "xvtlsbb cr3,vs0" +func_check "xxblendvb vs0,vs1,vs2,vs3" +func_check "xxblendvd vs0,vs1,vs2,vs3" +func_check "xxblendvh vs0,vs1,vs2,vs3" +func_check "xxblendvw vs0,vs1,vs2,vs3" +func_check "xxeval vs0,vs1,vs2,vs3,0" +func_check "xxeval vs0,vs1,vs2,vs3,3" +func_check "xxgenpcvbm vs0,v1,0" +func_check "xxgenpcvbm vs0,v1,1" +func_check "xxgenpcvbm vs0,v1,2" +func_check "xxgenpcvbm vs0,v1,3" +func_check "xxgenpcvdm vs0,v1,0" +func_check "xxgenpcvdm vs0,v1,1" +func_check "xxgenpcvdm vs0,v1,2" +func_check "xxgenpcvdm vs0,v1,3" +func_check "xxgenpcvhm vs0,v1,0" +func_check "xxgenpcvhm vs0,v1,1" +func_check "xxgenpcvhm vs0,v1,2" +func_check "xxgenpcvhm vs0,v1,3" +func_check "xxgenpcvwm vs0,v1,0" +func_check "xxgenpcvwm vs0,v1,1" +func_check "xxgenpcvwm vs0,v1,2" +func_check "xxgenpcvwm vs0,v1,3" +func_check "xxmfacc a4" +func_check "xxmtacc a4" +func_check "xxpermx vs0,vs1,vs2,vs3,0" +func_check "xxpermx vs0,vs1,vs2,vs3,3" +func_check "xxsetaccz a4" +func_check "xxsplti32dx vs0,0,2779096485" +func_check "xxsplti32dx vs0,0,4294967295" +func_check "xxsplti32dx vs0,0,127" +func_check "xxsplti32dx vs0,0,15" +func_check "xxsplti32dx vs0,0,3" +func_check "xxsplti32dx vs0,0,31" +func_check "xxsplti32dx vs0,0,32768" +func_check "xxsplti32dx vs0,0,63" +func_check "xxsplti32dx vs0,0,66535" +func_check "xxsplti32dx vs0,0,7" +func_check "xxsplti32dx vs0,1,2779096485" +func_check "xxsplti32dx vs0,1,4294967295" +func_check "xxsplti32dx vs0,1,127" +func_check "xxsplti32dx vs0,1,15" +func_check "xxsplti32dx vs0,1,3" +func_check "xxsplti32dx vs0,1,31" +func_check "xxsplti32dx vs0,1,32768" +func_check "xxsplti32dx vs0,1,63" +func_check "xxsplti32dx vs0,1,66535" +func_check "xxsplti32dx vs0,1,7" +func_check "xxspltidp vs0,0" +func_check "xxspltidp vs0,8388608" +func_check "xxspltidp vs0,8388609" +func_check "xxspltidp vs0,8594245" +func_check "xxspltidp vs0,16777215" +func_check "xxspltidp vs0,838860800" +func_check "xxspltidp vs0,2139095040" +func_check "xxspltidp vs0,2139095041" +func_check "xxspltidp vs0,2139300677" +func_check "xxspltidp vs0,2147483647" +func_check "xxspltidp vs0,2147483648" +func_check "xxspltidp vs0,2155872256" +func_check "xxspltidp vs0,2155872257" +func_check "xxspltidp vs0,2156077893" +func_check "xxspltidp vs0,2164260863" +func_check "xxspltidp vs0,4286578688" +func_check "xxspltidp vs0,4286578689" +func_check "xxspltidp vs0,4286784325" +func_check "xxspltidp vs0,4294967295" +func_check "xxspltiw vs0,0" +func_check "xxspltiw vs0,1" +func_check "xxspltiw vs0,3" +func_check "xxspltiw vs0,8" diff --git a/gdb/testsuite/gdb.arch/powerpc-power10.s b/gdb/testsuite/gdb.arch/powerpc-power10.s new file mode 100644 index 0000000..77111af --- /dev/null +++ b/gdb/testsuite/gdb.arch/powerpc-power10.s @@ -0,0 +1,639 @@ +/* This testcase is part of GDB, the GNU debugger. + + Copyright 2021 Free Software Foundation, Inc. + + This program is free software; you can redistribute it and/or modify + it under the terms of the GNU General Public License as published by + the Free Software Foundation; either version 3 of the License, or + (at your option) any later version. + + This program is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + GNU General Public License for more details. + + You should have received a copy of the GNU General Public License + along with this program. If not, see <http://www.gnu.org/licenses/>. */ + + .text + .globl func +func: + .long 0x7c200176 /* brd r0,r1 */ + .long 0x7c2001b6 /* brh r0,r1 */ + .long 0x7c200136 /* brw r0,r1 */ + .long 0x7c2011b8 /* cfuged r0,r1,r2 */ + .long 0x7c201076 /* cntlzdm r0,r1,r2 */ + .long 0x7c201476 /* cnttzdm r0,r1,r2 */ + .long 0xff8007c4 /* dcffixqq f28,v0 */ + .long 0xfc01d7c4 /* dctfixqq v0,f26 */ + .long 0xf01f0ad0 /* lxvkq vs0,1 */ + .long 0xf01f82d0 /* lxvkq vs0,16 */ + .long 0xf01f8ad0 /* lxvkq vs0,17 */ + .long 0xf01f92d0 /* lxvkq vs0,18 */ + .long 0xf01f9ad0 /* lxvkq vs0,19 */ + .long 0xf01f12d0 /* lxvkq vs0,2 */ + .long 0xf01fa2d0 /* lxvkq vs0,20 */ + .long 0xf01faad0 /* lxvkq vs0,21 */ + .long 0xf01fb2d0 /* lxvkq vs0,22 */ + .long 0xf01fbad0 /* lxvkq vs0,23 */ + .long 0xf01fc2d0 /* lxvkq vs0,24 */ + .long 0xf01f1ad0 /* lxvkq vs0,3 */ + .long 0xf01f22d0 /* lxvkq vs0,4 */ + .long 0xf01f2ad0 /* lxvkq vs0,5 */ + .long 0xf01f32d0 /* lxvkq vs0,6 */ + .long 0xf01f3ad0 /* lxvkq vs0,7 */ + .long 0xf01f42d0 /* lxvkq vs0,8 */ + .long 0xf01f4ad0 /* lxvkq vs0,9 */ + .long 0x1a800000 /* lxvp vs20,0(0) */ + .long 0x1a800010 /* lxvp vs20,16(0) */ + .long 0x1a800020 /* lxvp vs20,32(0) */ + .long 0x7e800a9a /* lxvpx vs20,0,r1 */ + .long 0x7c01101a /* lxvrbx vs0,r1,r2 */ + .long 0x7c0110da /* lxvrdx vs0,r1,r2 */ + .long 0x7c01105a /* lxvrhx vs0,r1,r2 */ + .long 0x7c01109a /* lxvrwx vs0,r1,r2 */ + .long 0x7f400066 /* mffprd r0,f26 */ + .long 0x7f600066 /* mffprd r0,f27 */ + .long 0x7f800066 /* mffprd r0,f28 */ + .long 0x7fa00066 /* mffprd r0,f29 */ + .long 0x7f400266 /* mfvsrld r0,vs26*/ + .long 0x7f600266 /* mfvsrld r0,vs27*/ + .long 0x7f800266 /* mfvsrld r0,vs28*/ + .long 0x7fa00266 /* mfvsrld r0,vs29 */ + .long 0x10000014 /* mtvsrbmi v0,0 */ + .long 0x10010015 /* mtvsrbmi v0,3 */ + .long 0x10030015 /* mtvsrbmi v0,7 */ + .long 0x10100e42 /* mtvsrbm v0,r1 */ + .long 0x10130e42 /* mtvsrdm v0,r1 */ + .long 0x10110e42 /* mtvsrhm v0,r1 */ + .long 0x10140e42 /* mtvsrqm v0,r1 */ + .long 0x10120e42 /* mtvsrwm v0,r1 */ + .long 0x60000000 /* nop */ + .quad 0x3801000006000000 /* paddi r0,r1,0 */ + .quad 0x3801000c06000000 /* paddi r0,r1,12 */ + .quad 0x3801003006000000 /* paddi r0,r1,48 */ + .quad 0x3801006206000000 /* paddi r0,r1,98 */ + .quad 0x7c2011787c201138 /* pdepd r0,r1,r2 */ + .quad 0x8801000006000000 /* plbz r0,0(r1) */ + .quad 0x8801001006000000 /* plbz r0,16(r1) */ + .quad 0x8801002006000000 /* plbz r0,32(r1) */ + .quad 0x8801004006000000 /* plbz r0,64(r1) */ + .quad 0x8801000806000000 /* plbz r0,8(r1) */ + .quad 0xe401000004000000 /* pld r0,0(r1) */ + .quad 0xe401001004000000 /* pld r0,16(r1) */ + .quad 0xe401002004000000 /* pld r0,32(r1) */ + .quad 0xe401004004000000 /* pld r0,64(r1) */ + .quad 0xe401000804000000 /* pld r0,8(r1) */ + .quad 0xcb80000006000000 /* plfd f28,0(0) */ + .quad 0xcb80001006000000 /* plfd f28,16(0) */ + .quad 0xcb80002006000000 /* plfd f28,32(0) */ + .quad 0xcb80000406000000 /* plfd f28,4(0) */ + .quad 0xcb80004006000000 /* plfd f28,64(0) */ + .quad 0xcb80000806000000 /* plfd f28,8(0) */ + .quad 0xc380000006000000 /* plfs f28,0(0) */ + .quad 0xc380001006000000 /* plfs f28,16(0) */ + .quad 0xc380002006000000 /* plfs f28,32(0) */ + .quad 0xc380000406000000 /* plfs f28,4(0) */ + .quad 0xc380004006000000 /* plfs f28,64(0) */ + .quad 0xc380000806000000 /* plfs f28,8(0) */ + .quad 0xa801000006000000 /* plha r0,0(r1) */ + .quad 0xa801001006000000 /* plha r0,16(r1) */ + .quad 0xa801002006000000 /* plha r0,32(r1) */ + .quad 0xa801004006000000 /* plha r0,64(r1) */ + .quad 0xa801000806000000 /* plha r0,8(r1) */ + .quad 0xa001000006000000 /* plhz r0,0(r1) */ + .quad 0xa001001006000000 /* plhz r0,16(r1) */ + .quad 0xa001002006000000 /* plhz r0,32(r1) */ + .quad 0xa001004006000000 /* plhz r0,64(r1) */ + .quad 0xa001000806000000 /* plhz r0,8(r1) */ + .quad 0xe340000004000000 /* plq r26,0(0) */ + .quad 0xe340001004000000 /* plq r26,16(0) */ + .quad 0xe340002004000000 /* plq r26,32(0) */ + .quad 0xe340003004000000 /* plq r26,48(0) */ + .quad 0xe340004004000000 /* plq r26,64(0) */ + .quad 0xe340000804000000 /* plq r26,8(0) */ + .quad 0xa401000004000000 /* plwa r0,0(r1) */ + .quad 0xa401001004000000 /* plwa r0,16(r1) */ + .quad 0xa401002004000000 /* plwa r0,32(r1) */ + .quad 0xa401004004000000 /* plwa r0,64(r1) */ + .quad 0xa401000804000000 /* plwa r0,8(r1) */ + .quad 0x8001000006000000 /* plwz r0,0(r1) */ + .quad 0x8001001006000000 /* plwz r0,16(r1) */ + .quad 0x8001002006000000 /* plwz r0,32(r1) */ + .quad 0x8001004006000000 /* plwz r0,64(r1) */ + .quad 0x8001000806000000 /* plwz r0,8(r1) */ + .quad 0xa801000004000000 /* plxsd v0,0(r1) */ + .quad 0xa801001004000000 /* plxsd v0,16(r1) */ + .quad 0xa801002004000000 /* plxsd v0,32(r1) */ + .quad 0xa801000404000000 /* plxsd v0,4(r1) */ + .quad 0xa801004004000000 /* plxsd v0,64(r1) */ + .quad 0xa801000804000000 /* plxsd v0,8(r1) */ + .quad 0xac01000004000000 /* plxssp v0,0(r1) */ + .quad 0xac01001004000000 /* plxssp v0,16(r1) */ + .quad 0xac01002004000000 /* plxssp v0,32(r1) */ + .quad 0xac01000404000000 /* plxssp v0,4(r1) */ + .quad 0xac01004004000000 /* plxssp v0,64(r1) */ + .quad 0xac01000804000000 /* plxssp v0,8(r1) */ + .quad 0xea80000004000000 /* plxvp vs20,0(0) */ + .quad 0xea80001004000000 /* plxvp vs20,16(0) */ + .quad 0xea80001804000000 /* plxvp vs20,24(0) */ + .quad 0xea80002004000000 /* plxvp vs20,32(0) */ + .quad 0xea80000804000000 /* plxvp vs20,8(0) */ + .quad 0xc801000004000000 /* plxv vs0,0(r1) */ + .quad 0xc801001004000000 /* plxv vs0,16(r1) */ + .quad 0xc801000404000000 /* plxv vs0,4(r1) */ + .quad 0xc801000804000000 /* plxv vs0,8(r1) */ + .quad 0xee00099807900000 /* pmxvbf16ger2 a4,vs0,vs1,0,0,0 */ + .quad 0xee00099807904000 /* pmxvbf16ger2 a4,vs0,vs1,0,0,1 */ + .quad 0xee0009980790000d /* pmxvbf16ger2 a4,vs0,vs1,0,13,0 */ + .quad 0xee0009980790400d /* pmxvbf16ger2 a4,vs0,vs1,0,13,1 */ + .quad 0xee000998079000b0 /* pmxvbf16ger2 a4,vs0,vs1,11,0,0 */ + .quad 0xee000998079040b0 /* pmxvbf16ger2 a4,vs0,vs1,11,0,1 */ + .quad 0xee000998079000bd /* pmxvbf16ger2 a4,vs0,vs1,11,13,0 */ + .quad 0xee000998079040bd /* pmxvbf16ger2 a4,vs0,vs1,11,13,1 */ + .quad 0xee000f9007900000 /* pmxvbf16ger2nn a4,vs0,vs1,0,0,0 */ + .quad 0xee000f9007904000 /* pmxvbf16ger2nn a4,vs0,vs1,0,0,1 */ + .quad 0xee000f900790000d /* pmxvbf16ger2nn a4,vs0,vs1,0,13,0 */ + .quad 0xee000f900790400d /* pmxvbf16ger2nn a4,vs0,vs1,0,13,1 */ + .quad 0xee000f90079000b0 /* pmxvbf16ger2nn a4,vs0,vs1,11,0,0 */ + .quad 0xee000f90079040b0 /* pmxvbf16ger2nn a4,vs0,vs1,11,0,1 */ + .quad 0xee000f90079000bd /* pmxvbf16ger2nn a4,vs0,vs1,11,13,0 */ + .quad 0xee000f90079040bd /* pmxvbf16ger2nn a4,vs0,vs1,11,13,1 */ + .quad 0xee000b9007900000 /* pmxvbf16ger2np a4,vs0,vs1,0,0,0 */ + .quad 0xee000b9007904000 /* pmxvbf16ger2np a4,vs0,vs1,0,0,1 */ + .quad 0xee000b900790000d /* pmxvbf16ger2np a4,vs0,vs1,0,13,0 */ + .quad 0xee000b900790400d /* pmxvbf16ger2np a4,vs0,vs1,0,13,1 */ + .quad 0xee000b90079000b0 /* pmxvbf16ger2np a4,vs0,vs1,11,0,0 */ + .quad 0xee000b90079040b0 /* pmxvbf16ger2np a4,vs0,vs1,11,0,1 */ + .quad 0xee000b90079000bd /* pmxvbf16ger2np a4,vs0,vs1,11,13,0 */ + .quad 0xee000b90079040bd /* pmxvbf16ger2np a4,vs0,vs1,11,13,1 */ + .quad 0xee000d9007900000 /* pmxvbf16ger2pn a4,vs0,vs1,0,0,0 */ + .quad 0xee000d9007904000 /* pmxvbf16ger2pn a4,vs0,vs1,0,0,1 */ + .quad 0xee000d900790000d /* pmxvbf16ger2pn a4,vs0,vs1,0,13,0 */ + .quad 0xee000d900790400d /* pmxvbf16ger2pn a4,vs0,vs1,0,13,1 */ + .quad 0xee000d90079000b0 /* pmxvbf16ger2pn a4,vs0,vs1,11,0,0 */ + .quad 0xee000d90079040b0 /* pmxvbf16ger2pn a4,vs0,vs1,11,0,1 */ + .quad 0xee000d90079000bd /* pmxvbf16ger2pn a4,vs0,vs1,11,13,0 */ + .quad 0xee000d90079040bd /* pmxvbf16ger2pn a4,vs0,vs1,11,13,1 */ + .quad 0xee00099007900000 /* pmxvbf16ger2pp a4,vs0,vs1,0,0,0 */ + .quad 0xee00099007904000 /* pmxvbf16ger2pp a4,vs0,vs1,0,0,1 */ + .quad 0xee0009900790000d /* pmxvbf16ger2pp a4,vs0,vs1,0,13,0 */ + .quad 0xee0009900790400d /* pmxvbf16ger2pp a4,vs0,vs1,0,13,1 */ + .quad 0xee000990079000b0 /* pmxvbf16ger2pp a4,vs0,vs1,11,0,0 */ + .quad 0xee000990079040b0 /* pmxvbf16ger2pp a4,vs0,vs1,11,0,1 */ + .quad 0xee000990079000bd /* pmxvbf16ger2pp a4,vs0,vs1,11,13,0 */ + .quad 0xee000990079040bd /* pmxvbf16ger2pp a4,vs0,vs1,11,13,1 */ + .quad 0xee00089807900000 /* pmxvf16ger2 a4,vs0,vs1,0,0,0 */ + .quad 0xee00089807904000 /* pmxvf16ger2 a4,vs0,vs1,0,0,1 */ + .quad 0xee0008980790000d /* pmxvf16ger2 a4,vs0,vs1,0,13,0 */ + .quad 0xee0008980790400d /* pmxvf16ger2 a4,vs0,vs1,0,13,1 */ + .quad 0xee000898079000b0 /* pmxvf16ger2 a4,vs0,vs1,11,0,0 */ + .quad 0xee000898079040b0 /* pmxvf16ger2 a4,vs0,vs1,11,0,1 */ + .quad 0xee000898079000bd /* pmxvf16ger2 a4,vs0,vs1,11,13,0 */ + .quad 0xee000898079040bd /* pmxvf16ger2 a4,vs0,vs1,11,13,1 */ + .quad 0xee000e9007900000 /* pmxvf16ger2nn a4,vs0,vs1,0,0,0 */ + .quad 0xee000e9007904000 /* pmxvf16ger2nn a4,vs0,vs1,0,0,1 */ + .quad 0xee000e900790000d /* pmxvf16ger2nn a4,vs0,vs1,0,13,0 */ + .quad 0xee000e900790400d /* pmxvf16ger2nn a4,vs0,vs1,0,13,1 */ + .quad 0xee000e90079000b0 /* pmxvf16ger2nn a4,vs0,vs1,11,0,0 */ + .quad 0xee000e90079040b0 /* pmxvf16ger2nn a4,vs0,vs1,11,0,1 */ + .quad 0xee000e90079000bd /* pmxvf16ger2nn a4,vs0,vs1,11,13,0 */ + .quad 0xee000e90079040bd /* pmxvf16ger2nn a4,vs0,vs1,11,13,1 */ + .quad 0xee000a9007900000 /* pmxvf16ger2np a4,vs0,vs1,0,0,0 */ + .quad 0xee000a9007904000 /* pmxvf16ger2np a4,vs0,vs1,0,0,1 */ + .quad 0xee000a900790000d /* pmxvf16ger2np a4,vs0,vs1,0,13,0 */ + .quad 0xee000a900790400d /* pmxvf16ger2np a4,vs0,vs1,0,13,1 */ + .quad 0xee000a90079000b0 /* pmxvf16ger2np a4,vs0,vs1,11,0,0 */ + .quad 0xee000a90079040b0 /* pmxvf16ger2np a4,vs0,vs1,11,0,1 */ + .quad 0xee000a90079000bd /* pmxvf16ger2np a4,vs0,vs1,11,13,0 */ + .quad 0xee000a90079040bd /* pmxvf16ger2np a4,vs0,vs1,11,13,1 */ + .quad 0xee000c9007900000 /* pmxvf16ger2pn a4,vs0,vs1,0,0,0 */ + .quad 0xee000c9007904000 /* pmxvf16ger2pn a4,vs0,vs1,0,0,1 */ + .quad 0xee000c900790000d /* pmxvf16ger2pn a4,vs0,vs1,0,13,0 */ + .quad 0xee000c900790400d /* pmxvf16ger2pn a4,vs0,vs1,0,13,1 */ + .quad 0xee000c90079000b0 /* pmxvf16ger2pn a4,vs0,vs1,11,0,0 */ + .quad 0xee000c90079040b0 /* pmxvf16ger2pn a4,vs0,vs1,11,0,1 */ + .quad 0xee000c90079000bd /* pmxvf16ger2pn a4,vs0,vs1,11,13,0 */ + .quad 0xee000c90079040bd /* pmxvf16ger2pn a4,vs0,vs1,11,13,1 */ + .quad 0xee00089007900000 /* pmxvf16ger2pp a4,vs0,vs1,0,0,0 */ + .quad 0xee00089007904000 /* pmxvf16ger2pp a4,vs0,vs1,0,0,1 */ + .quad 0xee0008900790000d /* pmxvf16ger2pp a4,vs0,vs1,0,13,0 */ + .quad 0xee0008900790400d /* pmxvf16ger2pp a4,vs0,vs1,0,13,1 */ + .quad 0xee000890079000b0 /* pmxvf16ger2pp a4,vs0,vs1,11,0,0 */ + .quad 0xee000890079040b0 /* pmxvf16ger2pp a4,vs0,vs1,11,0,1 */ + .quad 0xee000890079000bd /* pmxvf16ger2pp a4,vs0,vs1,11,13,0 */ + .quad 0xee000890079040bd /* pmxvf16ger2pp a4,vs0,vs1,11,13,1 */ + .quad 0xee0008d807900000 /* pmxvf32ger a4,vs0,vs1,0,0 */ + .quad 0xee0008d80790000d /* pmxvf32ger a4,vs0,vs1,0,13 */ + .quad 0xee0008d8079000b0 /* pmxvf32ger a4,vs0,vs1,11,0 */ + .quad 0xee0008d8079000bd /* pmxvf32ger a4,vs0,vs1,11,13 */ + .quad 0xee000ed007900000 /* pmxvf32gernn a4,vs0,vs1,0,0 */ + .quad 0xee000ed00790000d /* pmxvf32gernn a4,vs0,vs1,0,13 */ + .quad 0xee000ed0079000b0 /* pmxvf32gernn a4,vs0,vs1,11,0 */ + .quad 0xee000ed0079000bd /* pmxvf32gernn a4,vs0,vs1,11,13 */ + .quad 0xee000ad007900000 /* pmxvf32gernp a4,vs0,vs1,0,0 */ + .quad 0xee000ad00790000d /* pmxvf32gernp a4,vs0,vs1,0,13 */ + .quad 0xee000ad0079000b0 /* pmxvf32gernp a4,vs0,vs1,11,0 */ + .quad 0xee000ad0079000bd /* pmxvf32gernp a4,vs0,vs1,11,13 */ + .quad 0xee000cd007900000 /* pmxvf32gerpn a4,vs0,vs1,0,0 */ + .quad 0xee000cd00790000d /* pmxvf32gerpn a4,vs0,vs1,0,13 */ + .quad 0xee000cd0079000b0 /* pmxvf32gerpn a4,vs0,vs1,11,0 */ + .quad 0xee000cd0079000bd /* pmxvf32gerpn a4,vs0,vs1,11,13 */ + .quad 0xee0008d007900000 /* pmxvf32gerpp a4,vs0,vs1,0,0 */ + .quad 0xee0008d00790000d /* pmxvf32gerpp a4,vs0,vs1,0,13 */ + .quad 0xee0008d0079000b0 /* pmxvf32gerpp a4,vs0,vs1,11,0 */ + .quad 0xee0008d0079000bd /* pmxvf32gerpp a4,vs0,vs1,11,13 */ + .quad 0xee1601d807900000 /* pmxvf64ger a4,vs22,vs0,0,0 */ + .quad 0xee1601d807900004 /* pmxvf64ger a4,vs22,vs0,0,1 */ + .quad 0xee1601d8079000b0 /* pmxvf64ger a4,vs22,vs0,11,0 */ + .quad 0xee1601d8079000b4 /* pmxvf64ger a4,vs22,vs0,11,1 */ + .quad 0xee1607d007900000 /* pmxvf64gernn a4,vs22,vs0,0,0 */ + .quad 0xee1607d007900004 /* pmxvf64gernn a4,vs22,vs0,0,1 */ + .quad 0xee1607d0079000b0 /* pmxvf64gernn a4,vs22,vs0,11,0 */ + .quad 0xee1607d0079000b4 /* pmxvf64gernn a4,vs22,vs0,11,1 */ + .quad 0xee1603d007900000 /* pmxvf64gernp a4,vs22,vs0,0,0 */ + .quad 0xee1603d007900004 /* pmxvf64gernp a4,vs22,vs0,0,1 */ + .quad 0xee1603d0079000b0 /* pmxvf64gernp a4,vs22,vs0,11,0 */ + .quad 0xee1603d0079000b4 /* pmxvf64gernp a4,vs22,vs0,11,1 */ + .quad 0xee1605d007900000 /* pmxvf64gerpn a4,vs22,vs0,0,0 */ + .quad 0xee1605d007900004 /* pmxvf64gerpn a4,vs22,vs0,0,1 */ + .quad 0xee1605d0079000b0 /* pmxvf64gerpn a4,vs22,vs0,11,0 */ + .quad 0xee1605d0079000b4 /* pmxvf64gerpn a4,vs22,vs0,11,1 */ + .quad 0xee1601d007900000 /* pmxvf64gerpp a4,vs22,vs0,0,0 */ + .quad 0xee1601d007900004 /* pmxvf64gerpp a4,vs22,vs0,0,1 */ + .quad 0xee1601d0079000b0 /* pmxvf64gerpp a4,vs22,vs0,11,0 */ + .quad 0xee1601d0079000b4 /* pmxvf64gerpp a4,vs22,vs0,11,1 */ + .quad 0xee000a5807900000 /* pmxvi16ger2 a4,vs0,vs1,0,0,0 */ + .quad 0xee000a5807904000 /* pmxvi16ger2 a4,vs0,vs1,0,0,1 */ + .quad 0xee000a580790000d /* pmxvi16ger2 a4,vs0,vs1,0,13,0 */ + .quad 0xee000a580790400d /* pmxvi16ger2 a4,vs0,vs1,0,13,1 */ + .quad 0xee000a58079000b0 /* pmxvi16ger2 a4,vs0,vs1,11,0,0 */ + .quad 0xee000a58079040b0 /* pmxvi16ger2 a4,vs0,vs1,11,0,1 */ + .quad 0xee000a58079000bd /* pmxvi16ger2 a4,vs0,vs1,11,13,0 */ + .quad 0xee000a58079040bd /* pmxvi16ger2 a4,vs0,vs1,11,13,1 */ + .quad 0xee000b5807900000 /* pmxvi16ger2pp a4,vs0,vs1,0,0,0 */ + .quad 0xee000b5807904000 /* pmxvi16ger2pp a4,vs0,vs1,0,0,1 */ + .quad 0xee000b580790000d /* pmxvi16ger2pp a4,vs0,vs1,0,13,0 */ + .quad 0xee000b580790400d /* pmxvi16ger2pp a4,vs0,vs1,0,13,1 */ + .quad 0xee000b58079000b0 /* pmxvi16ger2pp a4,vs0,vs1,11,0,0 */ + .quad 0xee000b58079040b0 /* pmxvi16ger2pp a4,vs0,vs1,11,0,1 */ + .quad 0xee000b58079000bd /* pmxvi16ger2pp a4,vs0,vs1,11,13,0 */ + .quad 0xee000b58079040bd /* pmxvi16ger2pp a4,vs0,vs1,11,13,1 */ + .quad 0xee00095807900000 /* pmxvi16ger2s a4,vs0,vs1,0,0,0 */ + .quad 0xee00095807904000 /* pmxvi16ger2s a4,vs0,vs1,0,0,1 */ + .quad 0xee0009580790000d /* pmxvi16ger2s a4,vs0,vs1,0,13,0 */ + .quad 0xee0009580790400d /* pmxvi16ger2s a4,vs0,vs1,0,13,1 */ + .quad 0xee000958079000b0 /* pmxvi16ger2s a4,vs0,vs1,11,0,0 */ + .quad 0xee000958079040b0 /* pmxvi16ger2s a4,vs0,vs1,11,0,1 */ + .quad 0xee000958079000bd /* pmxvi16ger2s a4,vs0,vs1,11,13,0 */ + .quad 0xee000958079040bd /* pmxvi16ger2s a4,vs0,vs1,11,13,1 */ + .quad 0xee00095007900000 /* pmxvi16ger2spp a4,vs0,vs1,0,0,0 */ + .quad 0xee00095007904000 /* pmxvi16ger2spp a4,vs0,vs1,0,0,1 */ + .quad 0xee0009500790000d /* pmxvi16ger2spp a4,vs0,vs1,0,13,0 */ + .quad 0xee0009500790400d /* pmxvi16ger2spp a4,vs0,vs1,0,13,1 */ + .quad 0xee000950079000b0 /* pmxvi16ger2spp a4,vs0,vs1,11,0,0 */ + .quad 0xee000950079040b0 /* pmxvi16ger2spp a4,vs0,vs1,11,0,1 */ + .quad 0xee000950079000bd /* pmxvi16ger2spp a4,vs0,vs1,11,13,0 */ + .quad 0xee000950079040bd /* pmxvi16ger2spp a4,vs0,vs1,11,13,1 */ + .quad 0xee00091807900000 /* pmxvi4ger8 a4,vs0,vs1,0,0,0 */ + .quad 0xee00091807902d00 /* pmxvi4ger8 a4,vs0,vs1,0,0,45 */ + .quad 0xee00091807900001 /* pmxvi4ger8 a4,vs0,vs1,0,1,0 */ + .quad 0xee00091807902d01 /* pmxvi4ger8 a4,vs0,vs1,0,1,45 */ + .quad 0xee000918079000b0 /* pmxvi4ger8 a4,vs0,vs1,11,0,0 */ + .quad 0xee00091807902db0 /* pmxvi4ger8 a4,vs0,vs1,11,0,45 */ + .quad 0xee000918079000b1 /* pmxvi4ger8 a4,vs0,vs1,11,1,0 */ + .quad 0xee00091807902db1 /* pmxvi4ger8 a4,vs0,vs1,11,1,45 */ + .quad 0xee00091007900000 /* pmxvi4ger8pp a4,vs0,vs1,0,0,0 */ + .quad 0xee00091007902d00 /* pmxvi4ger8pp a4,vs0,vs1,0,0,45 */ + .quad 0xee00091007900001 /* pmxvi4ger8pp a4,vs0,vs1,0,1,0 */ + .quad 0xee00091007902d01 /* pmxvi4ger8pp a4,vs0,vs1,0,1,45 */ + .quad 0xee000910079000b0 /* pmxvi4ger8pp a4,vs0,vs1,11,0,0 */ + .quad 0xee00091007902db0 /* pmxvi4ger8pp a4,vs0,vs1,11,0,45 */ + .quad 0xee000910079000b1 /* pmxvi4ger8pp a4,vs0,vs1,11,1,0 */ + .quad 0xee00091007902db1 /* pmxvi4ger8pp a4,vs0,vs1,11,1,45 */ + .quad 0xee00081807900000 /* pmxvi8ger4 a4,vs0,vs1,0,0,0 */ + .quad 0xee00081807905000 /* pmxvi8ger4 a4,vs0,vs1,0,0,5 */ + .quad 0xee0008180790000d /* pmxvi8ger4 a4,vs0,vs1,0,13,0 */ + .quad 0xee0008180790500d /* pmxvi8ger4 a4,vs0,vs1,0,13,5 */ + .quad 0xee000818079000b0 /* pmxvi8ger4 a4,vs0,vs1,11,0,0 */ + .quad 0xee000818079050b0 /* pmxvi8ger4 a4,vs0,vs1,11,0,5 */ + .quad 0xee000818079000bd /* pmxvi8ger4 a4,vs0,vs1,11,13,0 */ + .quad 0xee000818079050bd /* pmxvi8ger4 a4,vs0,vs1,11,13,5 */ + .quad 0xee00081007900000 /* pmxvi8ger4pp a4,vs0,vs1,0,0,0 */ + .quad 0xee00081007905000 /* pmxvi8ger4pp a4,vs0,vs1,0,0,5 */ + .quad 0xee0008100790000d /* pmxvi8ger4pp a4,vs0,vs1,0,13,0 */ + .quad 0xee0008100790500d /* pmxvi8ger4pp a4,vs0,vs1,0,13,5 */ + .quad 0xee000810079000b0 /* pmxvi8ger4pp a4,vs0,vs1,11,0,0 */ + .quad 0xee000810079050b0 /* pmxvi8ger4pp a4,vs0,vs1,11,0,5 */ + .quad 0xee000810079000bd /* pmxvi8ger4pp a4,vs0,vs1,11,13,0 */ + .quad 0xee000810079050bd /* pmxvi8ger4pp a4,vs0,vs1,11,13,5 */ + .quad 0xee000b1807900000 /* pmxvi8ger4spp a4,vs0,vs1,0,0,0 */ + .quad 0xee000b1807905000 /* pmxvi8ger4spp a4,vs0,vs1,0,0,5 */ + .quad 0xee000b180790000d /* pmxvi8ger4spp a4,vs0,vs1,0,13,0 */ + .quad 0xee000b180790500d /* pmxvi8ger4spp a4,vs0,vs1,0,13,5 */ + .quad 0xee000b18079000b0 /* pmxvi8ger4spp a4,vs0,vs1,11,0,0 */ + .quad 0xee000b18079050b0 /* pmxvi8ger4spp a4,vs0,vs1,11,0,5 */ + .quad 0xee000b18079000bd /* pmxvi8ger4spp a4,vs0,vs1,11,13,0 */ + .quad 0xee000b18079050bd /* pmxvi8ger4spp a4,vs0,vs1,11,13,5 */ + .quad 0x9801000006000000 /* pstb r0,0(r1) */ + .quad 0x9801001006000000 /* pstb r0,16(r1) */ + .quad 0x9801002006000000 /* pstb r0,32(r1) */ + .quad 0x9801000806000000 /* pstb r0,8(r1) */ + .quad 0xf401000004000000 /* pstd r0,0(r1) */ + .quad 0xf401001004000000 /* pstd r0,16(r1) */ + .quad 0xf401002004000000 /* pstd r0,32(r1) */ + .quad 0xf401000804000000 /* pstd r0,8(r1) */ + .quad 0xdb40000006000000 /* pstfd f26,0(0) */ + .quad 0xdb40001006000000 /* pstfd f26,16(0) */ + .quad 0xdb40002006000000 /* pstfd f26,32(0) */ + .quad 0xdb40000406000000 /* pstfd f26,4(0) */ + .quad 0xdb40000806000000 /* pstfd f26,8(0) */ + .quad 0xd340000006000000 /* pstfs f26,0(0) */ + .quad 0xd340001006000000 /* pstfs f26,16(0) */ + .quad 0xd340002006000000 /* pstfs f26,32(0) */ + .quad 0xd340000406000000 /* pstfs f26,4(0) */ + .quad 0xd340000806000000 /* pstfs f26,8(0) */ + .quad 0xb001000006000000 /* psth r0,0(r1) */ + .quad 0xb001001006000000 /* psth r0,16(r1) */ + .quad 0xb001002006000000 /* psth r0,32(r1) */ + .quad 0xb001000806000000 /* psth r0,8(r1) */ + .quad 0xf300000004000000 /* pstq r24,0(0) */ + .quad 0xf300001004000000 /* pstq r24,16(0) */ + .quad 0xf300002004000000 /* pstq r24,32(0) */ + .quad 0xf300004004000000 /* pstq r24,64(0) */ + .quad 0xf300000804000000 /* pstq r24,8(0) */ + .quad 0x9001000006000000 /* pstw r0,0(r1) */ + .quad 0x9001001006000000 /* pstw r0,16(r1) */ + .quad 0x9001002006000000 /* pstw r0,32(r1) */ + .quad 0x9001000806000000 /* pstw r0,8(r1) */ + .quad 0xbac0000004000000 /* pstxsd v22,0(0) */ + .quad 0xbac0001004000000 /* pstxsd v22,16(0) */ + .quad 0xbac0002004000000 /* pstxsd v22,32(0) */ + .quad 0xbac0000404000000 /* pstxsd v22,4(0) */ + .quad 0xbac0004004000000 /* pstxsd v22,64(0) */ + .quad 0xbac0000804000000 /* pstxsd v22,8(0) */ + .quad 0xbec0000004000000 /* pstxssp v22,0(0) */ + .quad 0xbec0001004000000 /* pstxssp v22,16(0) */ + .quad 0xbec0002004000000 /* pstxssp v22,32(0) */ + .quad 0xbec0000404000000 /* pstxssp v22,4(0) */ + .quad 0xbec0004004000000 /* pstxssp v22,64(0) */ + .quad 0xbec0000804000000 /* pstxssp v22,8(0) */ + .quad 0xfa80000004000000 /* pstxvp vs20,0(0) */ + .quad 0xfa80001004000000 /* pstxvp vs20,16(0) */ + .quad 0xfa80002004000000 /* pstxvp vs20,32(0) */ + .quad 0xfa80003004000000 /* pstxvp vs20,48(0) */ + .quad 0xd801000004000000 /* pstxv vs0,0(r1) */ + .quad 0xd801001004000000 /* pstxv vs0,16(r1) */ + .quad 0xd801000404000000 /* pstxv vs0,4(r1) */ + .quad 0xd801000804000000 /* pstxv vs0,8(r1) */ + .long 0x7f470300 /* setbc r26,4*cr1+so */ + .long 0x7f480300 /* setbc r26,4*cr2+lt */ + .long 0x7f5f0300 /* setbc r26,4*cr7+so */ + .long 0x7f420300 /* setbc r26,eq */ + .long 0x7f410300 /* setbc r26,gt */ + .long 0x7f400300 /* setbc r26,lt */ + .long 0x7f430300 /* setbc r26,so */ + .long 0x7f470340 /* setbcr r26,4*cr1+so */ + .long 0x7f480340 /* setbcr r26,4*cr2+lt */ + .long 0x7f5f0340 /* setbcr r26,4*cr7+so */ + .long 0x7f470380 /* setnbc r26,4*cr1+so */ + .long 0x7f480380 /* setnbc r26,4*cr2+lt */ + .long 0x7f5f0380 /* setnbc r26,4*cr7+so */ + .long 0x7f420380 /* setnbc r26,eq */ + .long 0x7f410380 /* setnbc r26,gt */ + .long 0x7f400380 /* setnbc r26,lt */ + .long 0x7f430380 /* setnbc r26,so */ + .long 0x7f4703c0 /* setnbcr r26,4*cr1+so */ + .long 0x7f4803c0 /* setnbcr r26,4*cr2+lt */ + .long 0x7f5f03c0 /* setnbcr r26,4*cr7+so */ + .long 0x7f4003c0 /* setnbcr r26,lt */ + .long 0x7f4303c0 /* setnbcr r26,so */ + .long 0x1a800001 /* stxvp vs20,0(0) */ + .long 0x1a800011 /* stxvp vs20,16(0) */ + .long 0x1a800021 /* stxvp vs20,32(0) */ + .long 0x1a800031 /* stxvp vs20,48(0) */ + .long 0x7e800b9a /* stxvpx vs20,0,r1 */ + .long 0x7c01111a /* stxvrbx vs0,r1,r2 */ + .long 0x7c0111da /* stxvrdx vs0,r1,r2 */ + .long 0x7c01115a /* stxvrhx vs0,r1,r2 */ + .long 0x7c01119a /* stxvrwx vs0,r1,r2 */ + .long 0x1001154d /* vcfuged v0,v1,v2 */ + .long 0x1001118d /* vclrlb v0,v1,r2 */ + .long 0x100111cd /* vclrrb v0,v1,r2 */ + .long 0x10011784 /* vclzdm v0,v1,v2 */ + .long 0x100111c7 /* vcmpequq v0,v1,v2 */ + .long 0x10011387 /* vcmpgtsq v0,v1,v2 */ + .long 0x10011287 /* vcmpgtuq v0,v1,v2 */ + .long 0x11800941 /* vcmpsq cr3,v0,v1 */ + .long 0x11800901 /* vcmpuq cr3,v0,v1 */ + .long 0x10180e42 /* vcntmbb r0,v1,0 */ + .long 0x10190e42 /* vcntmbb r0,v1,1 */ + .long 0x101e0e42 /* vcntmbd r0,v1,0 */ + .long 0x101f0e42 /* vcntmbd r0,v1,1 */ + .long 0x101a0e42 /* vcntmbh r0,v1,0 */ + .long 0x101b0e42 /* vcntmbh r0,v1,1 */ + .long 0x101c0e42 /* vcntmbw r0,v1,0 */ + .long 0x101d0e42 /* vcntmbw r0,v1,1 */ + .long 0x100117c4 /* vctzdm v0,v1,v2 */ + .long 0x100113cb /* vdivesd v0,v1,v2 */ + .long 0x1001130b /* vdivesq v0,v1,v2 */ + .long 0x1001138b /* vdivesw v0,v1,v2 */ + .long 0x100112cb /* vdiveud v0,v1,v2 */ + .long 0x1001120b /* vdiveuq v0,v1,v2 */ + .long 0x1001128b /* vdiveuw v0,v1,v2 */ + .long 0x100111cb /* vdivsd v0,v1,v2 */ + .long 0x1001110b /* vdivsq v0,v1,v2 */ + .long 0x1001118b /* vdivsw v0,v1,v2 */ + .long 0x100110cb /* vdivud v0,v1,v2 */ + .long 0x1001100b /* vdivuq v0,v1,v2 */ + .long 0x1001108b /* vdivuw v0,v1,v2 */ + .long 0x10000e42 /* vexpandbm v0,v1 */ + .long 0x10030e42 /* vexpanddm v0,v1 */ + .long 0x10010e42 /* vexpandhm v0,v1 */ + .long 0x10040e42 /* vexpandqm v0,v1 */ + .long 0x10020e42 /* vexpandwm v0,v1 */ + .long 0x100110de /* vextddvlx v0,v1,v2,r3 */ + .long 0x100110df /* vextddvrx v0,v1,v2,r3 */ + .long 0x100110d8 /* vextdubvlx v0,v1,v2,r3 */ + .long 0x100110d9 /* vextdubvrx v0,v1,v2,r3 */ + .long 0x100110da /* vextduhvlx v0,v1,v2,r3 */ + .long 0x100110db /* vextduhvrx v0,v1,v2,r3 */ + .long 0x100110dc /* vextduwvlx v0,v1,v2,r3 */ + .long 0x100110dd /* vextduwvrx v0,v1,v2,r3 */ + .long 0x10080e42 /* vextractbm r0,v1 */ + .long 0x100b0e42 /* vextractdm r0,v1 */ + .long 0x10090e42 /* vextracthm r0,v1 */ + .long 0x100c0e42 /* vextractqm r0,v1 */ + .long 0x100a0e42 /* vextractwm r0,v1 */ + .long 0x101b0e02 /* vextsd2q v0,v1 */ + .long 0x10020ccc /* vgnb r0,v1,2 */ + .long 0x10030ccc /* vgnb r0,v1,3 */ + .long 0x10040ccc /* vgnb r0,v1,4 */ + .long 0x10050ccc /* vgnb r0,v1,5 */ + .long 0x10060ccc /* vgnb r0,v1,6 */ + .long 0x10070ccc /* vgnb r0,v1,7 */ + .long 0x1001120f /* vinsblx v0,r1,r2 */ + .long 0x1001130f /* vinsbrx v0,r1,r2 */ + .long 0x1001100f /* vinsbvlx v0,r1,v2 */ + .long 0x1001110f /* vinsbvrx v0,r1,v2 */ + .long 0x100112cf /* vinsdlx v0,r1,r2 */ + .long 0x100113cf /* vinsdrx v0,r1,r2 */ + .long 0x100309cf /* vinsd v0,r1,3 */ + .long 0x100709cf /* vinsd v0,r1,7 */ + .long 0x1001124f /* vinshlx v0,r1,r2 */ + .long 0x1001134f /* vinshrx v0,r1,r2 */ + .long 0x1001104f /* vinshvlx v0,r1,v2 */ + .long 0x1001114f /* vinshvrx v0,r1,v2 */ + .long 0x1001128f /* vinswlx v0,r1,r2 */ + .long 0x1001138f /* vinswrx v0,r1,r2 */ + .long 0x100308cf /* vinsw v0,r1,3 */ + .long 0x100708cf /* vinsw v0,r1,7 */ + .long 0x1001108f /* vinswvlx v0,r1,v2 */ + .long 0x1001118f /* vinswvrx v0,r1,v2 */ + .long 0x100117cb /* vmodsd v0,v1,v2 */ + .long 0x1001170b /* vmodsq v0,v1,v2 */ + .long 0x1001178b /* vmodsw v0,v1,v2 */ + .long 0x100116cb /* vmodud v0,v1,v2 */ + .long 0x1001160b /* vmoduq v0,v1,v2 */ + .long 0x1001168b /* vmoduw v0,v1,v2 */ + .long 0x100110d7 /* vmsumcud v0,v1,v2,v3 */ + .long 0x100113c8 /* vmulesd v0,v1,v2 */ + .long 0x100112c8 /* vmuleud v0,v1,v2 */ + .long 0x100113c9 /* vmulhsd v0,v1,v2 */ + .long 0x10011389 /* vmulhsw v0,v1,v2 */ + .long 0x100112c9 /* vmulhud v0,v1,v2 */ + .long 0x10011289 /* vmulhuw v0,v1,v2 */ + .long 0x100111c9 /* vmulld v0,v1,v2 */ + .long 0x100111c8 /* vmulosd v0,v1,v2 */ + .long 0x100110c8 /* vmuloud v0,v1,v2 */ + .long 0x100115cd /* vpdepd v0,v1,v2 */ + .long 0x1001158d /* vpextd v0,v1,v2 */ + .long 0x10011045 /* vrlqmi v0,v1,v2 */ + .long 0x10011145 /* vrlqnm v0,v1,v2 */ + .long 0x10011005 /* vrlq v0,v1,v2 */ + .long 0x10011016 /* vsldbi v0,v1,v2,0 */ + .long 0x10011116 /* vsldbi v0,v1,v2,4 */ + .long 0x10011105 /* vslq v0,v1,v2 */ + .long 0x10011305 /* vsraq v0,v1,v2 */ + .long 0x10011216 /* vsrdbi v0,v1,v2,0 */ + .long 0x10011316 /* vsrdbi v0,v1,v2,4 */ + .long 0x10011205 /* vsrq v0,v1,v2 */ + .long 0x1000080d /* vstribl v0,v1 */ + .long 0x10000c0d /* vstribl. v0,v1 */ + .long 0x1001080d /* vstribr v0,v1 */ + .long 0x10010c0d /* vstribr. v0,v1 */ + .long 0x1002080d /* vstrihl v0,v1 */ + .long 0x10020c0d /* vstrihl. v0,v1 */ + .long 0x1003080d /* vstrihr v0,v1 */ + .long 0x10030c0d /* vstrihr. v0,v1 */ + .long 0xfc011088 /* xscmpeqqp v0,v1,v2 */ + .long 0xfc011188 /* xscmpgeqp v0,v1,v2 */ + .long 0xfc0111c8 /* xscmpgtqp v0,v1,v2 */ + .long 0xfc080e88 /* xscvqpsqz v0,v1 */ + .long 0xfc000e88 /* xscvqpuqz v0,v1 */ + .long 0xfc0b0e88 /* xscvsqqp v0,v1 */ + .long 0xfc030e88 /* xscvuqqp v0,v1 */ + .long 0xfc011548 /* xsmaxcqp v0,v1,v2 */ + .long 0xfc0115c8 /* xsmincqp v0,v1,v2 */ + .long 0xee000998 /* xvbf16ger2 a4,vs0,vs1 */ + .long 0xee000f90 /* xvbf16ger2nn a4,vs0,vs1 */ + .long 0xee000b90 /* xvbf16ger2np a4,vs0,vs1 */ + .long 0xee000d90 /* xvbf16ger2pn a4,vs0,vs1 */ + .long 0xee000990 /* xvbf16ger2pp a4,vs0,vs1 */ + .long 0xf0100f6c /* xvcvbf16spn vs0,vs1 */ + .long 0xf0110f6c /* xvcvspbf16 vs0,vs1 */ + .long 0xee000898 /* xvf16ger2 a4,vs0,vs1 */ + .long 0xee000e90 /* xvf16ger2nn a4,vs0,vs1 */ + .long 0xee000a90 /* xvf16ger2np a4,vs0,vs1 */ + .long 0xee000c90 /* xvf16ger2pn a4,vs0,vs1 */ + .long 0xee000890 /* xvf16ger2pp a4,vs0,vs1 */ + .long 0xee0008d8 /* xvf32ger a4,vs0,vs1 */ + .long 0xee000ed0 /* xvf32gernn a4,vs0,vs1 */ + .long 0xee000ad0 /* xvf32gernp a4,vs0,vs1 */ + .long 0xee000cd0 /* xvf32gerpn a4,vs0,vs1 */ + .long 0xee0008d0 /* xvf32gerpp a4,vs0,vs1 */ + .long 0xee1601d8 /* xvf64ger a4,vs22,vs0 */ + .long 0xee1607d0 /* xvf64gernn a4,vs22,vs0 */ + .long 0xee1603d0 /* xvf64gernp a4,vs22,vs0 */ + .long 0xee1605d0 /* xvf64gerpn a4,vs22,vs0 */ + .long 0xee1601d0 /* xvf64gerpp a4,vs22,vs0 */ + .long 0xee000a58 /* xvi16ger2 a4,vs0,vs1 */ + .long 0xee000b58 /* xvi16ger2pp a4,vs0,vs1 */ + .long 0xee000958 /* xvi16ger2s a4,vs0,vs1 */ + .long 0xee000950 /* xvi16ger2spp a4,vs0,vs1 */ + .long 0xee000918 /* xvi4ger8 a4,vs0,vs1 */ + .long 0xee000910 /* xvi4ger8pp a4,vs0,vs1 */ + .long 0xee000818 /* xvi8ger4 a4,vs0,vs1 */ + .long 0xee000810 /* xvi8ger4pp a4,vs0,vs1 */ + .long 0xee000b18 /* xvi8ger4spp a4,vs0,vs1 */ + .long 0xf182076c /* xvtlsbb cr3,vs0 */ + .quad 0x840110c005000000 /* xxblendvb vs0,vs1,vs2,vs3 */ + .quad 0x840110f005000000 /* xxblendvd vs0,vs1,vs2,vs3 */ + .quad 0x840110d005000000 /* xxblendvh vs0,vs1,vs2,vs3 */ + .quad 0x840110e005000000 /* xxblendvw vs0,vs1,vs2,vs3 */ + .quad 0x880110d005000000 /* xxeval vs0,vs1,vs2,vs3,0 */ + .quad 0x880110d005000003 /* xxeval vs0,vs1,vs2,vs3,3 */ + .long 0xf0000f28 /* xxgenpcvbm vs0,v1,0 */ + .long 0xf0010f28 /* xxgenpcvbm vs0,v1,1 */ + .long 0xf0020f28 /* xxgenpcvbm vs0,v1,2 */ + .long 0xf0030f28 /* xxgenpcvbm vs0,v1,3 */ + .long 0xf0000f6a /* xxgenpcvdm vs0,v1,0 */ + .long 0xf0010f6a /* xxgenpcvdm vs0,v1,1 */ + .long 0xf0020f6a /* xxgenpcvdm vs0,v1,2 */ + .long 0xf0030f6a /* xxgenpcvdm vs0,v1,3 */ + .long 0xf0000f2a /* xxgenpcvhm vs0,v1,0 */ + .long 0xf0010f2a /* xxgenpcvhm vs0,v1,1 */ + .long 0xf0020f2a /* xxgenpcvhm vs0,v1,2 */ + .long 0xf0030f2a /* xxgenpcvhm vs0,v1,3 */ + .long 0xf0000f68 /* xxgenpcvwm vs0,v1,0 */ + .long 0xf0010f68 /* xxgenpcvwm vs0,v1,1 */ + .long 0xf0020f68 /* xxgenpcvwm vs0,v1,2 */ + .long 0xf0030f68 /* xxgenpcvwm vs0,v1,3 */ + .long 0x7e000162 /* xxmfacc a4 */ + .long 0x7e010162 /* xxmtacc a4 */ + .long 0x7e030162 /* xxsetaccz a4 */ + .quad 0x880110c005000000 /* xxpermx vs0,vs1,vs2,vs3,0 */ + .quad 0x880110c005000003 /* xxpermx vs0,vs1,vs2,vs3,3 */ + .quad 0x8000007f05000000 /* xxsplti32dx vs0,0,127 */ + .quad 0x8000000f05000000 /* xxsplti32dx vs0,0,15 */ + .quad 0x8000a5a50500a5a5 /* xxsplti32dx vs0,0,2779096485 */ + .quad 0x8000000305000000 /* xxsplti32dx vs0,0,3 */ + .quad 0x8000001f05000000 /* xxsplti32dx vs0,0,31 */ + .quad 0x8000800005000000 /* xxsplti32dx vs0,0,32768 */ + .quad 0x8000ffff0500ffff /* xxsplti32dx vs0,0,4294967295 */ + .quad 0x8000003f05000000 /* xxsplti32dx vs0,0,63 */ + .quad 0x800003e705000001 /* xxsplti32dx vs0,0,66535 */ + .quad 0x8000000705000000 /* xxsplti32dx vs0,0,7 */ + .quad 0x8002007f05000000 /* xxsplti32dx vs0,1,127 */ + .quad 0x8002000f05000000 /* xxsplti32dx vs0,1,15 */ + .quad 0x8002a5a50500a5a5 /* xxsplti32dx vs0,1,2779096485 */ + .quad 0x8002000305000000 /* xxsplti32dx vs0,1,3 */ + .quad 0x8002001f05000000 /* xxsplti32dx vs0,1,31 */ + .quad 0x8002800005000000 /* xxsplti32dx vs0,1,32768 */ + .quad 0x8002ffff0500ffff /* xxsplti32dx vs0,1,4294967295 */ + .quad 0x8002003f05000000 /* xxsplti32dx vs0,1,63 */ + .quad 0x800203e705000001 /* xxsplti32dx vs0,1,66535 */ + .quad 0x8002000705000000 /* xxsplti32dx vs0,1,7 */ + .quad 0x8004000005000000 /* xxspltidp vs0,0 */ + .quad 0x8004000005000080 /* xxspltidp vs0,8388608 */ + .quad 0x8004000105000080 /* xxspltidp vs0,8388609 */ + .quad 0x8004234505000083 /* xxspltidp vs0,8594245 */ + .quad 0x8004ffff050000ff /* xxspltidp vs0,16777215 */ + .quad 0x8004000005003200 /* xxspltidp vs0,838860800 */ + .quad 0x8004000005007f80 /* xxspltidp vs0,2139095040 */ + .quad 0x8004000105007f80 /* xxspltidp vs0,2139095041 */ + .quad 0x8004234505007f83 /* xxspltidp vs0,2139300677 */ + .quad 0x8004ffff05007fff /* xxspltidp vs0,2147483647 */ + .quad 0x8004000005008000 /* xxspltidp vs0,2147483648 */ + .quad 0x8004000005008080 /* xxspltidp vs0,2155872256 */ + .quad 0x8004000105008080 /* xxspltidp vs0,2155872257 */ + .quad 0x8004234505008083 /* xxspltidp vs0,2156077893 */ + .quad 0x8004ffff050080ff /* xxspltidp vs0,2164260863 */ + .quad 0x800400000500ff80 /* xxspltidp vs0,4286578688 */ + .quad 0x800400010500ff80 /* xxspltidp vs0,4286578689 */ + .quad 0x800423450500ff83 /* xxspltidp vs0,4286784325 */ + .quad 0x8004ffff0500ffff /* xxspltidp vs0,4294967295 */ + .quad 0x8006000005000000 /* xxspltiw vs0,0 */ + .quad 0x8006000105000000 /* xxspltiw vs0,0 */ + .quad 0x8006000305000000 /* xxspltiw vs0,3 */ + .quad 0x8006000805000000 /* xxspltiw vs0,8 */ |