; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 4 ; Test vector add reduction intrinsic ; ; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z16 | FileCheck %s ; 1 vector length declare i8 @llvm.vector.reduce.add.v16i8(<16 x i8> %a) declare i16 @llvm.vector.reduce.add.v8i16(<8 x i16> %a) declare i32 @llvm.vector.reduce.add.v4i32(<4 x i32> %a) declare i64 @llvm.vector.reduce.add.v2i64(<2 x i64> %a) declare i128 @llvm.vector.reduce.add.v1i128(<1 x i128> %a) ; 2 vector lengths declare i8 @llvm.vector.reduce.add.v32i8(<32 x i8> %a) declare i16 @llvm.vector.reduce.add.v16i16(<16 x i16> %a) declare i32 @llvm.vector.reduce.add.v8i32(<8 x i32> %a) declare i64 @llvm.vector.reduce.add.v4i64(<4 x i64> %a) declare i128 @llvm.vector.reduce.add.v2i128(<2 x i128> %a) ; ; TODO ; ; 4 vector lengths declare i8 @llvm.vector.reduce.add.v64i8(<64 x i8> %a) declare i16 @llvm.vector.reduce.add.v32i16(<32 x i16> %a) declare i32 @llvm.vector.reduce.add.v16i32(<16 x i32> %a) declare i64 @llvm.vector.reduce.add.v8i64(<8 x i64> %a) declare i128 @llvm.vector.reduce.add.v4i128(<4 x i128> %a) ; ; Subvector lengths declare i8 @llvm.vector.reduce.add.v8i8(<8 x i8> %a) declare i16 @llvm.vector.reduce.add.v4i16(<4 x i16> %a) declare i32 @llvm.vector.reduce.add.v2i32(<2 x i32> %a) declare i64 @llvm.vector.reduce.add.v1i64(<1 x i64> %a) define i8 @f1_1(<16 x i8> %a) { ; CHECK-LABEL: f1_1: ; CHECK: # %bb.0: ; CHECK-NEXT: vgbm %v0, 0 ; CHECK-NEXT: vsumb %v1, %v24, %v0 ; CHECK-NEXT: vsumqf %v0, %v1, %v0 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i8 @llvm.vector.reduce.add.v16i8(<16 x i8> %a) ret i8 %redadd } define i16 @f1_2(<8 x i16> %a) { ; CHECK-LABEL: f1_2: ; CHECK: # %bb.0: ; CHECK-NEXT: vgbm %v0, 0 ; CHECK-NEXT: vsumh %v1, %v24, %v0 ; CHECK-NEXT: vsumqf %v0, %v1, %v0 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i16 @llvm.vector.reduce.add.v8i16(<8 x i16> %a) ret i16 %redadd } define i32 @f1_3(<4 x i32> %a) { ; CHECK-LABEL: f1_3: ; CHECK: # %bb.0: ; CHECK-NEXT: vgbm %v0, 0 ; CHECK-NEXT: vsumqf %v0, %v24, %v0 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i32 @llvm.vector.reduce.add.v4i32(<4 x i32> %a) ret i32 %redadd } define i64 @f1_4(<2 x i64> %a) { ; CHECK-LABEL: f1_4: ; CHECK: # %bb.0: ; CHECK-NEXT: vrepg %v0, %v24, 1 ; CHECK-NEXT: vag %v0, %v24, %v0 ; CHECK-NEXT: vlgvg %r2, %v0, 0 ; CHECK-NEXT: br %r14 %redadd = call i64 @llvm.vector.reduce.add.v2i64(<2 x i64> %a) ret i64 %redadd } define i128 @f1_5(<1 x i128> %a) { ; CHECK-LABEL: f1_5: ; CHECK: # %bb.0: ; CHECK-NEXT: vst %v24, 0(%r2), 3 ; CHECK-NEXT: br %r14 %redadd = call i128 @llvm.vector.reduce.add.v1i128(<1 x i128> %a) ret i128 %redadd } define i8 @f2_1(<32 x i8> %a) { ; CHECK-LABEL: f2_1: ; CHECK: # %bb.0: ; CHECK-NEXT: vab %v0, %v24, %v26 ; CHECK-NEXT: vgbm %v1, 0 ; CHECK-NEXT: vsumb %v0, %v0, %v1 ; CHECK-NEXT: vsumqf %v0, %v0, %v1 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i8 @llvm.vector.reduce.add.v32i8(<32 x i8> %a) ret i8 %redadd } define i16 @f2_2(<16 x i16> %a) { ; CHECK-LABEL: f2_2: ; CHECK: # %bb.0: ; CHECK-NEXT: vah %v0, %v24, %v26 ; CHECK-NEXT: vgbm %v1, 0 ; CHECK-NEXT: vsumh %v0, %v0, %v1 ; CHECK-NEXT: vsumqf %v0, %v0, %v1 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i16 @llvm.vector.reduce.add.v16i16(<16 x i16> %a) ret i16 %redadd } define i32 @f2_3(<8 x i32> %a) { ; CHECK-LABEL: f2_3: ; CHECK: # %bb.0: ; CHECK-NEXT: vaf %v0, %v24, %v26 ; CHECK-NEXT: vgbm %v1, 0 ; CHECK-NEXT: vsumqf %v0, %v0, %v1 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> %a) ret i32 %redadd } define i64 @f2_4(<4 x i64> %a) { ; CHECK-LABEL: f2_4: ; CHECK: # %bb.0: ; CHECK-NEXT: vag %v0, %v24, %v26 ; CHECK-NEXT: vrepg %v1, %v0, 1 ; CHECK-NEXT: vag %v0, %v0, %v1 ; CHECK-NEXT: vlgvg %r2, %v0, 0 ; CHECK-NEXT: br %r14 %redadd = call i64 @llvm.vector.reduce.add.v4i64(<4 x i64> %a) ret i64 %redadd } define i128 @f2_5(<2 x i128> %a) { ; CHECK-LABEL: f2_5: ; CHECK: # %bb.0: ; CHECK-NEXT: vl %v0, 16(%r3), 3 ; CHECK-NEXT: vl %v1, 0(%r3), 3 ; CHECK-NEXT: vaq %v0, %v1, %v0 ; CHECK-NEXT: vst %v0, 0(%r2), 3 ; CHECK-NEXT: br %r14 %redadd = call i128 @llvm.vector.reduce.add.v2i128(<2 x i128> %a) ret i128 %redadd } define i8 @f3_1(<64 x i8> %a) { ; CHECK-LABEL: f3_1: ; CHECK: # %bb.0: ; CHECK-NEXT: vab %v0, %v26, %v30 ; CHECK-NEXT: vab %v1, %v24, %v28 ; CHECK-NEXT: vab %v0, %v1, %v0 ; CHECK-NEXT: vgbm %v1, 0 ; CHECK-NEXT: vsumb %v0, %v0, %v1 ; CHECK-NEXT: vsumqf %v0, %v0, %v1 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i8 @llvm.vector.reduce.add.v64i8(<64 x i8> %a) ret i8 %redadd } define i16 @f3_2(<32 x i16> %a) { ; CHECK-LABEL: f3_2: ; CHECK: # %bb.0: ; CHECK-NEXT: vah %v0, %v26, %v30 ; CHECK-NEXT: vah %v1, %v24, %v28 ; CHECK-NEXT: vah %v0, %v1, %v0 ; CHECK-NEXT: vgbm %v1, 0 ; CHECK-NEXT: vsumh %v0, %v0, %v1 ; CHECK-NEXT: vsumqf %v0, %v0, %v1 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i16 @llvm.vector.reduce.add.v32i16(<32 x i16> %a) ret i16 %redadd } define i32 @f3_3(<16 x i32> %a) { ; CHECK-LABEL: f3_3: ; CHECK: # %bb.0: ; CHECK-NEXT: vaf %v0, %v26, %v30 ; CHECK-NEXT: vaf %v1, %v24, %v28 ; CHECK-NEXT: vaf %v0, %v1, %v0 ; CHECK-NEXT: vgbm %v1, 0 ; CHECK-NEXT: vsumqf %v0, %v0, %v1 ; CHECK-NEXT: vlgvf %r2, %v0, 3 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i32 @llvm.vector.reduce.add.v16i32(<16 x i32> %a) ret i32 %redadd } define i64 @f3_4(<8 x i64> %a) { ; CHECK-LABEL: f3_4: ; CHECK: # %bb.0: ; CHECK-NEXT: vag %v0, %v26, %v30 ; CHECK-NEXT: vag %v1, %v24, %v28 ; CHECK-NEXT: vag %v0, %v1, %v0 ; CHECK-NEXT: vrepg %v1, %v0, 1 ; CHECK-NEXT: vag %v0, %v0, %v1 ; CHECK-NEXT: vlgvg %r2, %v0, 0 ; CHECK-NEXT: br %r14 %redadd = call i64 @llvm.vector.reduce.add.v8i64(<8 x i64> %a) ret i64 %redadd } define i128 @f3_5(<4 x i128> %a) { ; CHECK-LABEL: f3_5: ; CHECK: # %bb.0: ; CHECK-NEXT: vl %v0, 32(%r3), 3 ; CHECK-NEXT: vl %v1, 0(%r3), 3 ; CHECK-NEXT: vl %v2, 48(%r3), 3 ; CHECK-NEXT: vl %v3, 16(%r3), 3 ; CHECK-NEXT: vaq %v2, %v3, %v2 ; CHECK-NEXT: vaq %v0, %v1, %v0 ; CHECK-NEXT: vaq %v0, %v0, %v2 ; CHECK-NEXT: vst %v0, 0(%r2), 3 ; CHECK-NEXT: br %r14 %redadd = call i128 @llvm.vector.reduce.add.v4i128(<4 x i128> %a) ret i128 %redadd } define i8 @f4_1(<8 x i8> %a) { ; CHECK-LABEL: f4_1: ; CHECK: # %bb.0: ; CHECK-NEXT: vpkg %v0, %v24, %v24 ; CHECK-NEXT: vab %v0, %v24, %v0 ; CHECK-NEXT: vpkf %v1, %v0, %v0 ; CHECK-NEXT: vab %v0, %v0, %v1 ; CHECK-NEXT: vrepb %v1, %v0, 1 ; CHECK-NEXT: vab %v0, %v0, %v1 ; CHECK-NEXT: vlgvb %r2, %v0, 0 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i8 @llvm.vector.reduce.add.v8i8(<8 x i8> %a) ret i8 %redadd } define i16 @f4_2(<4 x i16> %a) { ; CHECK-LABEL: f4_2: ; CHECK: # %bb.0: ; CHECK-NEXT: vpkg %v0, %v24, %v24 ; CHECK-NEXT: vah %v0, %v24, %v0 ; CHECK-NEXT: vreph %v1, %v0, 1 ; CHECK-NEXT: vah %v0, %v0, %v1 ; CHECK-NEXT: vlgvh %r2, %v0, 0 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i16 @llvm.vector.reduce.add.v4i16(<4 x i16> %a) ret i16 %redadd } define i32 @f4_3(<2 x i32> %a) { ; CHECK-LABEL: f4_3: ; CHECK: # %bb.0: ; CHECK-NEXT: vrepf %v0, %v24, 1 ; CHECK-NEXT: vaf %v0, %v24, %v0 ; CHECK-NEXT: vlgvf %r2, %v0, 0 ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d ; CHECK-NEXT: br %r14 %redadd = call i32 @llvm.vector.reduce.add.v2i32(<2 x i32> %a) ret i32 %redadd } define i64 @f4_4(<1 x i64> %a) { ; CHECK-LABEL: f4_4: ; CHECK: # %bb.0: ; CHECK-NEXT: vlgvg %r2, %v24, 0 ; CHECK-NEXT: br %r14 %redadd = call i64 @llvm.vector.reduce.add.v1i64(<1 x i64> %a) ret i64 %redadd }