diff options
Diffstat (limited to 'gcc/testsuite/gcc.target/i386/avx512fp16-vaddsh-1b.c')
-rw-r--r-- | gcc/testsuite/gcc.target/i386/avx512fp16-vaddsh-1b.c | 104 |
1 files changed, 104 insertions, 0 deletions
diff --git a/gcc/testsuite/gcc.target/i386/avx512fp16-vaddsh-1b.c b/gcc/testsuite/gcc.target/i386/avx512fp16-vaddsh-1b.c new file mode 100644 index 0000000..724112c --- /dev/null +++ b/gcc/testsuite/gcc.target/i386/avx512fp16-vaddsh-1b.c @@ -0,0 +1,104 @@ +/* { dg-do run { target avx512fp16 } } */ +/* { dg-options "-O2 -mavx512fp16 -mavx512dq" } */ + + +#define AVX512FP16 +#include "avx512fp16-helper.h" + +#define N_ELEMS 8 + +void NOINLINE +emulate_add_sh(V512 * dest, V512 op1, V512 op2, + __mmask8 k, int zero_mask) +{ + V512 v1, v2, v3, v4, v5, v6, v7, v8; + int i; + + unpack_ph_2twops(op1, &v1, &v2); + unpack_ph_2twops(op2, &v3, &v4); + unpack_ph_2twops(*dest, &v7, &v8); + + if ((k&1) || !k) + v5.f32[0] = v1.f32[0] + v3.f32[0]; + else if (zero_mask) + v5.f32[0] = 0; + else + v5.f32[0] = v7.f32[0]; + + for (i = 1; i < 8; i++) + v5.f32[i] = v1.f32[i]; + + *dest = pack_twops_2ph(v5, v6); +} + +void +test_512 (void) +{ + V512 res; + V512 exp; + + init_src(); + + emulate_add_sh(&exp, src1, src2, 0x1, 0); + res.xmmh[0] = _mm_add_sh(src1.xmmh[0], src2.xmmh[0]); + check_results(&res, &exp, N_ELEMS, "_mm_add_sh"); + + //DEST.fp16[0] := SRC1.fp16[0] + SRC2.fp16[0] + emulate_add_sh(&exp, src1, src2, 0x1, 0); + res.xmmh[0] = _mm_mask_add_sh(res.xmmh[0], 0x1, + src1.xmmh[0], src2.xmmh[0]); + check_results(&res, &exp, N_ELEMS, "_mm_mask_add_sh"); + + //dest.fp16[0] remains unchanged + init_dest(&res, &exp); + emulate_add_sh(&exp, src1, src2, 0x2, 0); + res.xmmh[0] = _mm_mask_add_sh(res.xmmh[0], 0x2, + src1.xmmh[0], src2.xmmh[0]); + check_results(&res, &exp, N_ELEMS, "_mm_mask_add_sh"); + + //dest.fp16[0] = 0 + emulate_add_sh(&exp, src1, src2, 0x2, 1); + res.xmmh[0] = _mm_maskz_add_sh(0x2, src1.xmmh[0], src2.xmmh[0]); + check_results(&res, &exp, N_ELEMS, "_mm_maskz_add_sh"); + + //DEST.fp16[0] := SRC1.fp16[0] + SRC2.fp16[0] + emulate_add_sh(&exp, src1, src2, 0x3, 1); + res.xmmh[0] = _mm_maskz_add_sh(0x3, src1.xmmh[0], src2.xmmh[0]); + check_results(&res, &exp, N_ELEMS, "_mm_maskz_add_sh"); + + //DEST.fp16[0] := SRC1.fp16[0] + SRC2.fp16[0] + emulate_add_sh(&exp, src1, src2, 0x1, 0); + res.xmmh[0] = _mm_add_round_sh(src1.xmmh[0], + src2.xmmh[0], _ROUND_NINT); + check_results(&res, &exp, N_ELEMS, "_mm_add_round_sh"); + + //DEST.fp16[0] := SRC1.fp16[0] + SRC2.fp16[0] + emulate_add_sh(&exp, src1, src2, 0x1, 0); + res.xmmh[0] = _mm_mask_add_round_sh(res.xmmh[0], 0x1, src1.xmmh[0], + src2.xmmh[0], _ROUND_NINT); + check_results(&res, &exp, N_ELEMS, "_mm_mask_add_round_sh"); + + //dest.fp16[0] remains unchanged + init_dest(&res, &exp); + emulate_add_sh(&exp, src1, src2, 0x2, 0); + res.xmmh[0] = _mm_mask_add_round_sh(res.xmmh[0], 0x2, src1.xmmh[0], + src2.xmmh[0], _ROUND_NINT); + check_results(&res, &exp, N_ELEMS, "_mm_mask_add_round_sh"); + + //dest.fp16[0] = 0 + emulate_add_sh(&exp, src1, src2, 0x2, 1); + res.xmmh[0] = _mm_maskz_add_round_sh(0x2, src1.xmmh[0], + src2.xmmh[0], _ROUND_NINT); + check_results(&res, &exp, N_ELEMS, "_mm_maskz_add_round_sh"); + + //DEST.fp16[0] := SRC1.fp16[0] + SRC2.fp16[0] + emulate_add_sh(&exp, src1, src2, 0x3, 1); + res.xmmh[0] = _mm_maskz_add_round_sh(0x3, src1.xmmh[0], + src2.xmmh[0], _ROUND_NINT); + check_results(&res, &exp, N_ELEMS, "_mm_maskz_add_round_sh"); + + if (n_errs != 0) { + abort (); + } +} + |