aboutsummaryrefslogtreecommitdiff
path: root/gcc
diff options
context:
space:
mode:
authorKyrylo Tkachov <kyrylo.tkachov@arm.com>2023-06-07 16:18:01 +0100
committerKyrylo Tkachov <kyrylo.tkachov@arm.com>2023-06-07 16:18:01 +0100
commita053c659f6bbc2d2eaf61fb3aad8c52899d0deb7 (patch)
treeee56405f5f777f5fa15e7c05c2c0116658ca7a11 /gcc
parent000f8b9a6a0ec7eaae9535e73c8f4fc3ea10a364 (diff)
downloadgcc-a053c659f6bbc2d2eaf61fb3aad8c52899d0deb7.zip
gcc-a053c659f6bbc2d2eaf61fb3aad8c52899d0deb7.tar.gz
gcc-a053c659f6bbc2d2eaf61fb3aad8c52899d0deb7.tar.bz2
aarch64: Improve RTL representation of ADDP instructions
Similar to the ADDLP instructions the non-widening ADDP ones can be represented by adding the odd lanes with the even lanes of a vector. These instructions take two vector inputs and the architecture spec describes the operation as concatenating them together before going through it with pairwise additions. This patch chooses to represent ADDP on 64-bit and 128-bit input vectors slightly differently, reasons explained in the comments in aarhc64-simd.md. Bootstrapped and tested on aarch64-none-linux-gnu and aarch64_be-none-elf. gcc/ChangeLog: * config/aarch64/aarch64-simd.md (aarch64_addp<mode><vczle><vczbe>): Reimplement as... (aarch64_addp<mode>_insn): ... This... (aarch64_addp<mode><vczle><vczbe>_insn): ... And this. (aarch64_addp<mode>): New define_expand.
Diffstat (limited to 'gcc')
-rw-r--r--gcc/config/aarch64/aarch64-simd.md70
1 files changed, 63 insertions, 7 deletions
diff --git a/gcc/config/aarch64/aarch64-simd.md b/gcc/config/aarch64/aarch64-simd.md
index dbd6fc6..b23067c 100644
--- a/gcc/config/aarch64/aarch64-simd.md
+++ b/gcc/config/aarch64/aarch64-simd.md
@@ -7014,17 +7014,73 @@
;; addp
-(define_insn "aarch64_addp<mode><vczle><vczbe>"
- [(set (match_operand:VDQ_I 0 "register_operand" "=w")
- (unspec:VDQ_I
- [(match_operand:VDQ_I 1 "register_operand" "w")
- (match_operand:VDQ_I 2 "register_operand" "w")]
- UNSPEC_ADDP))]
- "TARGET_SIMD"
+;; ADDP with two registers semantically concatenates them and performs
+;; a pairwise addition on the result. For 128-bit input modes represent this
+;; as a concatentation of the pairwise addition results of the two input
+;; registers. This allow us to avoid using intermediate 256-bit modes.
+(define_insn "aarch64_addp<mode>_insn"
+ [(set (match_operand:VQ_I 0 "register_operand" "=w")
+ (vec_concat:VQ_I
+ (plus:<VHALF>
+ (vec_select:<VHALF>
+ (match_operand:VQ_I 1 "register_operand" "w")
+ (match_operand:VQ_I 3 "vect_par_cnst_even_or_odd_half"))
+ (vec_select:<VHALF>
+ (match_dup 1)
+ (match_operand:VQ_I 4 "vect_par_cnst_even_or_odd_half")))
+ (plus:<VHALF>
+ (vec_select:<VHALF>
+ (match_operand:VQ_I 2 "register_operand" "w")
+ (match_dup 3))
+ (vec_select:<VHALF>
+ (match_dup 2)
+ (match_dup 4)))))]
+ "TARGET_SIMD && !rtx_equal_p (operands[3], operands[4])"
+ "addp\t%<v>0<Vmtype>, %<v>1<Vmtype>, %<v>2<Vmtype>"
+ [(set_attr "type" "neon_reduc_add<q>")]
+)
+
+;; For 64-bit input modes an ADDP is represented as a concatentation
+;; of the input registers into an 128-bit register which is then fed
+;; into a pairwise add. That way we avoid having to create intermediate
+;; 32-bit vector modes.
+(define_insn "aarch64_addp<mode><vczle><vczbe>_insn"
+ [(set (match_operand:VD_BHSI 0 "register_operand" "=w")
+ (plus:VD_BHSI
+ (vec_select:VD_BHSI
+ (vec_concat:<VDBL>
+ (match_operand:VD_BHSI 1 "register_operand" "w")
+ (match_operand:VD_BHSI 2 "register_operand" "w"))
+ (match_operand:<VDBL> 3 "vect_par_cnst_even_or_odd_half"))
+ (vec_select:VD_BHSI
+ (vec_concat:<VDBL>
+ (match_dup 1)
+ (match_dup 2))
+ (match_operand:<VDBL> 4 "vect_par_cnst_even_or_odd_half"))))]
+ "TARGET_SIMD && !rtx_equal_p (operands[3], operands[4])"
"addp\t%<v>0<Vmtype>, %<v>1<Vmtype>, %<v>2<Vmtype>"
[(set_attr "type" "neon_reduc_add<q>")]
)
+(define_expand "aarch64_addp<mode>"
+ [(match_operand:VDQ_I 0 "register_operand")
+ (match_operand:VDQ_I 1 "register_operand")
+ (match_operand:VDQ_I 2 "register_operand")]
+ "TARGET_SIMD"
+ {
+ int nunits = GET_MODE_NUNITS (<MODE>mode).to_constant ();
+ if (known_eq (GET_MODE_BITSIZE (<MODE>mode), 128))
+ nunits /= 2;
+ rtx par_even = aarch64_gen_stepped_int_parallel (nunits, 0, 2);
+ rtx par_odd = aarch64_gen_stepped_int_parallel (nunits, 1, 2);
+ if (BYTES_BIG_ENDIAN)
+ std::swap (operands[1], operands[2]);
+ emit_insn (gen_aarch64_addp<mode>_insn (operands[0], operands[1],
+ operands[2], par_even, par_odd));
+ DONE;
+ }
+)
+
;; sqrt
(define_expand "sqrt<mode>2"