diff options
author | Simon Pilgrim <llvm-dev@redking.me.uk> | 2015-02-07 16:14:55 +0000 |
---|---|---|
committer | Simon Pilgrim <llvm-dev@redking.me.uk> | 2015-02-07 16:14:55 +0000 |
commit | 0238b96c065540a7ec4ed4fdf243135e34d4ad9a (patch) | |
tree | 8f4520d5d4750e47810b36c6812b23423e452678 | |
parent | 947ce78d4993903583a1ce6eac0b0391f493fb98 (diff) | |
download | llvm-0238b96c065540a7ec4ed4fdf243135e34d4ad9a.zip llvm-0238b96c065540a7ec4ed4fdf243135e34d4ad9a.tar.gz llvm-0238b96c065540a7ec4ed4fdf243135e34d4ad9a.tar.bz2 |
[X86] Force fp stack folding tests to keep to specific domain.
General boolean instructions (AND, ANDN, OR, XOR) need to use a specific domain instruction (and not just the default).
llvm-svn: 228495
-rw-r--r-- | llvm/test/CodeGen/X86/stack-folding-fp-avx1.ll | 33 | ||||
-rw-r--r-- | llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll | 16 |
2 files changed, 36 insertions, 13 deletions
diff --git a/llvm/test/CodeGen/X86/stack-folding-fp-avx1.ll b/llvm/test/CodeGen/X86/stack-folding-fp-avx1.ll index 26bd391..18cd417 100644 --- a/llvm/test/CodeGen/X86/stack-folding-fp-avx1.ll +++ b/llvm/test/CodeGen/X86/stack-folding-fp-avx1.ll @@ -147,7 +147,9 @@ define <4 x float> @stack_fold_andnps(<4 x float> %a0, <4 x float> %a1) { %4 = xor <2 x i64> %2, <i64 -1, i64 -1> %5 = and <2 x i64> %4, %3 %6 = bitcast <2 x i64> %5 to <4 x float> - ret <4 x float> %6 + ; fadd forces execution domain + %7 = fadd <4 x float> %6, <float 0x0, float 0x0, float 0x0, float 0x0> + ret <4 x float> %7 } define <8 x float> @stack_fold_andnps_ymm(<8 x float> %a0, <8 x float> %a1) { @@ -159,7 +161,9 @@ define <8 x float> @stack_fold_andnps_ymm(<8 x float> %a0, <8 x float> %a1) { %4 = xor <4 x i64> %2, <i64 -1, i64 -1, i64 -1, i64 -1> %5 = and <4 x i64> %4, %3 %6 = bitcast <4 x i64> %5 to <8 x float> - ret <8 x float> %6 + ; fadd forces execution domain + %7 = fadd <8 x float> %6, <float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0> + ret <8 x float> %7 } define <2 x double> @stack_fold_andpd(<2 x double> %a0, <2 x double> %a1) { @@ -196,7 +200,9 @@ define <4 x float> @stack_fold_andps(<4 x float> %a0, <4 x float> %a1) { %3 = bitcast <4 x float> %a1 to <2 x i64> %4 = and <2 x i64> %2, %3 %5 = bitcast <2 x i64> %4 to <4 x float> - ret <4 x float> %5 + ; fadd forces execution domain + %6 = fadd <4 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0> + ret <4 x float> %6 } define <8 x float> @stack_fold_andps_ymm(<8 x float> %a0, <8 x float> %a1) { @@ -207,7 +213,9 @@ define <8 x float> @stack_fold_andps_ymm(<8 x float> %a0, <8 x float> %a1) { %3 = bitcast <8 x float> %a1 to <4 x i64> %4 = and <4 x i64> %2, %3 %5 = bitcast <4 x i64> %4 to <8 x float> - ret <8 x float> %5 + ; fadd forces execution domain + %6 = fadd <8 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0> + ret <8 x float> %6 } define <2 x double> @stack_fold_blendpd(<2 x double> %a0, <2 x double> %a1) { @@ -843,7 +851,6 @@ declare <8 x float> @llvm.x86.avx.dp.ps.256(<8 x float>, <8 x float>, i8) nounwi define <4 x float> @stack_fold_extractf128(<8 x float> %a0, <8 x float> %a1) { ;CHECK-LABEL: stack_fold_extractf128 ;CHECK: vextractf128 $1, {{%ymm[0-9][0-9]*}}, {{-?[0-9]*}}(%rsp) {{.*#+}} 16-byte Folded Spill - ;CHECK: vmovaps {{-?[0-9]*}}(%rsp), %xmm0 {{.*#+}} 16-byte Reload %1 = shufflevector <8 x float> %a0, <8 x float> %a1, <4 x i32> <i32 4, i32 5, i32 6, i32 7> %2 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() ret <4 x float> %1 @@ -1239,7 +1246,9 @@ define <4 x float> @stack_fold_orps(<4 x float> %a0, <4 x float> %a1) { %3 = bitcast <4 x float> %a1 to <2 x i64> %4 = or <2 x i64> %2, %3 %5 = bitcast <2 x i64> %4 to <4 x float> - ret <4 x float> %5 + ; fadd forces execution domain + %6 = fadd <4 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0> + ret <4 x float> %6 } define <8 x float> @stack_fold_orps_ymm(<8 x float> %a0, <8 x float> %a1) { @@ -1250,7 +1259,9 @@ define <8 x float> @stack_fold_orps_ymm(<8 x float> %a0, <8 x float> %a1) { %3 = bitcast <8 x float> %a1 to <4 x i64> %4 = or <4 x i64> %2, %3 %5 = bitcast <4 x i64> %4 to <8 x float> - ret <8 x float> %5 + ; fadd forces execution domain + %6 = fadd <8 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0> + ret <8 x float> %6 } define <8 x float> @stack_fold_perm2f128(<8 x float> %a0, <8 x float> %a1) { @@ -1781,7 +1792,9 @@ define <4 x float> @stack_fold_xorps(<4 x float> %a0, <4 x float> %a1) { %3 = bitcast <4 x float> %a1 to <2 x i64> %4 = xor <2 x i64> %2, %3 %5 = bitcast <2 x i64> %4 to <4 x float> - ret <4 x float> %5 + ; fadd forces execution domain + %6 = fadd <4 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0> + ret <4 x float> %6 } define <8 x float> @stack_fold_xorps_ymm(<8 x float> %a0, <8 x float> %a1) { @@ -1792,5 +1805,7 @@ define <8 x float> @stack_fold_xorps_ymm(<8 x float> %a0, <8 x float> %a1) { %3 = bitcast <8 x float> %a1 to <4 x i64> %4 = xor <4 x i64> %2, %3 %5 = bitcast <4 x i64> %4 to <8 x float> - ret <8 x float> %5 + ; fadd forces execution domain + %6 = fadd <8 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0, float 0x0> + ret <8 x float> %6 } diff --git a/llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll b/llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll index 2beccb18..c26cc9d 100644 --- a/llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll +++ b/llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll @@ -99,7 +99,9 @@ define <4 x float> @stack_fold_andnps(<4 x float> %a0, <4 x float> %a1) { %4 = xor <2 x i64> %2, <i64 -1, i64 -1> %5 = and <2 x i64> %4, %3 %6 = bitcast <2 x i64> %5 to <4 x float> - ret <4 x float> %6 + ; fadd forces execution domain + %7 = fadd <4 x float> %6, <float 0x0, float 0x0, float 0x0, float 0x0> + ret <4 x float> %7 } define <2 x double> @stack_fold_andpd(<2 x double> %a0, <2 x double> %a1) { @@ -123,7 +125,9 @@ define <4 x float> @stack_fold_andps(<4 x float> %a0, <4 x float> %a1) { %3 = bitcast <4 x float> %a1 to <2 x i64> %4 = and <2 x i64> %2, %3 %5 = bitcast <2 x i64> %4 to <4 x float> - ret <4 x float> %5 + ; fadd forces execution domain + %6 = fadd <4 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0> + ret <4 x float> %6 } define <2 x double> @stack_fold_blendpd(<2 x double> %a0, <2 x double> %a1) { @@ -837,7 +841,9 @@ define <4 x float> @stack_fold_orps(<4 x float> %a0, <4 x float> %a1) { %3 = bitcast <4 x float> %a1 to <2 x i64> %4 = or <2 x i64> %2, %3 %5 = bitcast <2 x i64> %4 to <4 x float> - ret <4 x float> %5 + ; fadd forces execution domain + %6 = fadd <4 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0> + ret <4 x float> %6 } ; TODO stack_fold_rcpps @@ -1077,5 +1083,7 @@ define <4 x float> @stack_fold_xorps(<4 x float> %a0, <4 x float> %a1) { %3 = bitcast <4 x float> %a1 to <2 x i64> %4 = xor <2 x i64> %2, %3 %5 = bitcast <2 x i64> %4 to <4 x float> - ret <4 x float> %5 + ; fadd forces execution domain + %6 = fadd <4 x float> %5, <float 0x0, float 0x0, float 0x0, float 0x0> + ret <4 x float> %6 } |