aboutsummaryrefslogtreecommitdiff
path: root/gcc/config
diff options
context:
space:
mode:
authorHaochen Jiang <haochen.jiang@intel.com>2025-01-23 09:52:20 +0800
committerHaochen Jiang <haochen.jiang@intel.com>2025-01-23 09:53:33 +0800
commitf1056463cb4c7950fc1bada6485c14df71ea3dd7 (patch)
tree78583d29cf21ea2c4f4eb240371e677e33e6851d /gcc/config
parentcfef82be8973c9dc481c96306ba3e2c342398e48 (diff)
downloadgcc-f1056463cb4c7950fc1bada6485c14df71ea3dd7.zip
gcc-f1056463cb4c7950fc1bada6485c14df71ea3dd7.tar.gz
gcc-f1056463cb4c7950fc1bada6485c14df71ea3dd7.tar.bz2
i386: Change mnemonics from VCVTNEPH2[B,H]F8 to VCVTPH2[B,H]F8
gcc/ChangeLog: PR target/118270 * config/i386/avx10_2-512convertintrin.h: Change intrin and builtin name according to new mnemonics. * config/i386/avx10_2convertintrin.h: Ditto. * config/i386/i386-builtin.def (BDESC): Ditto. * config/i386/sse.md (UNSPEC_VCVTPH2BF8): Rename from UNSPEC_VCVTNEPH2BF8. (UNSPEC_VCVTPH2BF8S): Rename from UNSPEC_VCVTNEPH2BF8S. (UNSPEC_VCVTPH2HF8): Rename from UNSPEC_VCVTNEPH2HF8. (UNSPEC_VCVTPH2HF8S): Rename from UNSPEC_VCVTNEPH2HF8S. (UNSPEC_CONVERTPH2FP8): Rename from UNSPEC_NECONVERTPH2FP8. Adjust UNSPEC name. (convertph2fp8): Rename from neconvertph2fp8. Adjust iterator map. (vcvt<neconvertph2fp8>v8hf): Rename to... (vcvt<neconvertph2fp8>v8hf): ...this. (*vcvt<neconvertph2fp8>v8hf): Rename to... (*vcvt<neconvertph2fp8>v8hf): ...this. (vcvt<neconvertph2fp8>v8hf_mask): Rename to... (vcvt<neconvertph2fp8>v8hf_mask): ...this. (*vcvt<neconvertph2fp8>v8hf_mask): Rename to... (*vcvt<neconvertph2fp8>v8hf_mask): ...this. (vcvt<neconvertph2fp8><mode><mask_name>): Rename to... (vcvt<convertph2fp8><mode><mask_name>): ...this. gcc/testsuite/ChangeLog: PR target/118270 * gcc.target/i386/avx10_2-512-convert-1.c: Adjust output and intrin call. * gcc.target/i386/avx10_2-512-vcvtneph2bf8-2.c: Move to... * gcc.target/i386/avx10_2-512-vcvtph2bf8-2.c: ...here. Adjust intrin call. * gcc.target/i386/avx10_2-512-vcvtneph2bf8s-2.c: Move to... * gcc.target/i386/avx10_2-512-vcvtph2bf8s-2.c: ...here. Adjust intrin call. * gcc.target/i386/avx10_2-512-vcvtneph2hf8-2.c: Move to... * gcc.target/i386/avx10_2-512-vcvtph2hf8-2.c: ...here. Adjust intrin call. * gcc.target/i386/avx10_2-512-vcvtneph2hf8s-2.c: Move to... * gcc.target/i386/avx10_2-512-vcvtph2hf8s-2.c: ...here. Adjust intrin call. * gcc.target/i386/avx10_2-convert-1.c: Adjust output and intrin call. * gcc.target/i386/avx10_2-vcvtneph2bf8-2.c: Move to... * gcc.target/i386/avx10_2-vcvtph2bf8-2.c: ...here. Adjust intrin call. * gcc.target/i386/avx10_2-vcvtneph2hf8-2.c: Move to... * gcc.target/i386/avx10_2-vcvtph2bf8s-2.c: ...here. Adjust intrin call. * gcc.target/i386/avx10_2-vcvtneph2bf8s-2.c: Move to... * gcc.target/i386/avx10_2-vcvtph2hf8-2.c: ...here. Adjust intrin call. * gcc.target/i386/avx10_2-vcvtneph2hf8s-2.c: Move to... * gcc.target/i386/avx10_2-vcvtph2hf8s-2.c: ...here. Adjust intrin call.
Diffstat (limited to 'gcc/config')
-rw-r--r--gcc/config/i386/avx10_2-512convertintrin.h112
-rw-r--r--gcc/config/i386/avx10_2convertintrin.h224
-rw-r--r--gcc/config/i386/i386-builtin.def24
-rw-r--r--gcc/config/i386/sse.md50
4 files changed, 205 insertions, 205 deletions
diff --git a/gcc/config/i386/avx10_2-512convertintrin.h b/gcc/config/i386/avx10_2-512convertintrin.h
index c753dd7..5c64b9f 100644
--- a/gcc/config/i386/avx10_2-512convertintrin.h
+++ b/gcc/config/i386/avx10_2-512convertintrin.h
@@ -426,118 +426,118 @@ _mm512_maskz_cvthf8_ph (__mmask32 __U, __m256i __A)
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtneph_pbf8 (__m512h __A)
+_mm512_cvtph_bf8 (__m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2bf8512_mask ((__v32hf) __A,
- (__v32qi) (__m256i)
- _mm256_undefined_si256 (),
- (__mmask32) -1);
+ return (__m256i) __builtin_ia32_vcvtph2bf8512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i)
+ _mm256_undefined_si256 (),
+ (__mmask32) -1);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtneph_pbf8 (__m256i __W, __mmask32 __U, __m512h __A)
+_mm512_mask_cvtph_bf8 (__m256i __W, __mmask32 __U, __m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2bf8512_mask ((__v32hf) __A,
- (__v32qi) (__m256i) __W,
- (__mmask32) __U);
+ return (__m256i) __builtin_ia32_vcvtph2bf8512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i) __W,
+ (__mmask32) __U);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtneph_pbf8 (__mmask32 __U, __m512h __A)
+_mm512_maskz_cvtph_bf8 (__mmask32 __U, __m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2bf8512_mask ((__v32hf) __A,
- (__v32qi) (__m256i)
- _mm256_setzero_si256 (),
- (__mmask32) __U);
+ return (__m256i) __builtin_ia32_vcvtph2bf8512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i)
+ _mm256_setzero_si256 (),
+ (__mmask32) __U);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtnesph_pbf8 (__m512h __A)
+_mm512_cvtsph_bf8 (__m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2bf8s512_mask ((__v32hf) __A,
- (__v32qi) (__m256i)
- _mm256_undefined_si256 (),
- (__mmask32) -1);
+ return (__m256i) __builtin_ia32_vcvtph2bf8s512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i)
+ _mm256_undefined_si256 (),
+ (__mmask32) -1);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtnesph_pbf8 (__m256i __W, __mmask32 __U, __m512h __A)
+_mm512_mask_cvtsph_bf8 (__m256i __W, __mmask32 __U, __m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2bf8s512_mask ((__v32hf) __A,
- (__v32qi) (__m256i) __W,
- (__mmask32) __U);
+ return (__m256i) __builtin_ia32_vcvtph2bf8s512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i) __W,
+ (__mmask32) __U);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtnesph_pbf8 (__mmask32 __U, __m512h __A)
+_mm512_maskz_cvtsph_bf8 (__mmask32 __U, __m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2bf8s512_mask ((__v32hf) __A,
- (__v32qi) (__m256i)
- _mm256_setzero_si256 (),
- (__mmask32) __U);
+ return (__m256i) __builtin_ia32_vcvtph2bf8s512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i)
+ _mm256_setzero_si256 (),
+ (__mmask32) __U);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtneph_phf8 (__m512h __A)
+_mm512_cvtph_hf8 (__m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2hf8512_mask ((__v32hf) __A,
- (__v32qi) (__m256i)
- _mm256_undefined_si256 (),
- (__mmask32) -1);
+ return (__m256i) __builtin_ia32_vcvtph2hf8512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i)
+ _mm256_undefined_si256 (),
+ (__mmask32) -1);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtneph_phf8 (__m256i __W, __mmask32 __U, __m512h __A)
+_mm512_mask_cvtph_hf8 (__m256i __W, __mmask32 __U, __m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2hf8512_mask ((__v32hf) __A,
- (__v32qi)(__m256i) __W,
- (__mmask32) __U);
+ return (__m256i) __builtin_ia32_vcvtph2hf8512_mask ((__v32hf) __A,
+ (__v32qi)(__m256i) __W,
+ (__mmask32) __U);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtneph_phf8 (__mmask32 __U, __m512h __A)
+_mm512_maskz_cvtph_hf8 (__mmask32 __U, __m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2hf8512_mask ((__v32hf) __A,
- (__v32qi) (__m256i)
- _mm256_setzero_si256 (),
- (__mmask32) __U);
+ return (__m256i) __builtin_ia32_vcvtph2hf8512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i)
+ _mm256_setzero_si256 (),
+ (__mmask32) __U);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtnesph_phf8 (__m512h __A)
+_mm512_cvtsph_hf8 (__m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2hf8s512_mask ((__v32hf) __A,
- (__v32qi) (__m256i)
- _mm256_undefined_si256 (),
- (__mmask32) -1);
+ return (__m256i) __builtin_ia32_vcvtph2hf8s512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i)
+ _mm256_undefined_si256 (),
+ (__mmask32) -1);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtnesph_phf8 (__m256i __W, __mmask32 __U, __m512h __A)
+_mm512_mask_cvtsph_hf8 (__m256i __W, __mmask32 __U, __m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2hf8s512_mask ((__v32hf) __A,
- (__v32qi) (__m256i) __W,
- (__mmask32) __U);
+ return (__m256i) __builtin_ia32_vcvtph2hf8s512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i) __W,
+ (__mmask32) __U);
}
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtnesph_phf8 (__mmask32 __U, __m512h __A)
+_mm512_maskz_cvtsph_hf8 (__mmask32 __U, __m512h __A)
{
- return (__m256i) __builtin_ia32_vcvtneph2hf8s512_mask ((__v32hf) __A,
- (__v32qi) (__m256i)
- _mm256_setzero_si256 (),
- (__mmask32) __U);
+ return (__m256i) __builtin_ia32_vcvtph2hf8s512_mask ((__v32hf) __A,
+ (__v32qi) (__m256i)
+ _mm256_setzero_si256 (),
+ (__mmask32) __U);
}
extern __inline __m512h
diff --git a/gcc/config/i386/avx10_2convertintrin.h b/gcc/config/i386/avx10_2convertintrin.h
index 45d2bff..8635566 100644
--- a/gcc/config/i386/avx10_2convertintrin.h
+++ b/gcc/config/i386/avx10_2convertintrin.h
@@ -740,234 +740,234 @@ _mm256_maskz_cvthf8_ph (__mmask16 __U, __m128i __A)
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtneph_pbf8 (__m128h __A)
+_mm_cvtph_bf8 (__m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8128_mask ((__v8hf) __A,
- (__v16qi)(__m128i)
- _mm_undefined_si128 (),
- (__mmask8) -1);
+ return (__m128i) __builtin_ia32_vcvtph2bf8128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i)
+ _mm_undefined_si128 (),
+ (__mmask8) -1);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtneph_pbf8 (__m128i __W, __mmask8 __U, __m128h __A)
+_mm_mask_cvtph_bf8 (__m128i __W, __mmask8 __U, __m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8128_mask ((__v8hf) __A,
- (__v16qi)(__m128i) __W,
- (__mmask8) __U);
+ return (__m128i) __builtin_ia32_vcvtph2bf8128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i) __W,
+ (__mmask8) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtneph_pbf8 (__mmask8 __U, __m128h __A)
+_mm_maskz_cvtph_bf8 (__mmask8 __U, __m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8128_mask ((__v8hf) __A,
- (__v16qi)(__m128i)
- _mm_setzero_si128 (),
- (__mmask8) __U);
+ return (__m128i) __builtin_ia32_vcvtph2bf8128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i)
+ _mm_setzero_si128 (),
+ (__mmask8) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtneph_pbf8 (__m256h __A)
+_mm256_cvtph_bf8 (__m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8256_mask ((__v16hf) __A,
- (__v16qi)(__m128i)
- _mm_undefined_si128 (),
- (__mmask16) -1);
+ return (__m128i) __builtin_ia32_vcvtph2bf8256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i)
+ _mm_undefined_si128 (),
+ (__mmask16) -1);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtneph_pbf8 (__m128i __W, __mmask16 __U, __m256h __A)
+_mm256_mask_cvtph_bf8 (__m128i __W, __mmask16 __U, __m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8256_mask ((__v16hf) __A,
- (__v16qi)(__m128i) __W,
- (__mmask16) __U);
+ return (__m128i) __builtin_ia32_vcvtph2bf8256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i) __W,
+ (__mmask16) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtneph_pbf8 (__mmask16 __U, __m256h __A)
+_mm256_maskz_cvtph_bf8 (__mmask16 __U, __m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8256_mask ((__v16hf) __A,
- (__v16qi)(__m128i)
- _mm_setzero_si128 (),
- (__mmask16) __U);
+ return (__m128i) __builtin_ia32_vcvtph2bf8256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i)
+ _mm_setzero_si128 (),
+ (__mmask16) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtnesph_pbf8 (__m128h __A)
+_mm_cvtsph_bf8 (__m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8s128_mask ((__v8hf) __A,
- (__v16qi)(__m128i)
- _mm_undefined_si128 (),
- (__mmask8) -1);
+ return (__m128i) __builtin_ia32_vcvtph2bf8s128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i)
+ _mm_undefined_si128 (),
+ (__mmask8) -1);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtnesph_pbf8 (__m128i __W, __mmask8 __U, __m128h __A)
+_mm_mask_cvtsph_bf8 (__m128i __W, __mmask8 __U, __m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8s128_mask ((__v8hf) __A,
- (__v16qi)(__m128i) __W,
- (__mmask8) __U);
+ return (__m128i) __builtin_ia32_vcvtph2bf8s128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i) __W,
+ (__mmask8) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtnesph_pbf8 (__mmask8 __U, __m128h __A)
+_mm_maskz_cvtsph_bf8 (__mmask8 __U, __m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8s128_mask ((__v8hf) __A,
- (__v16qi)(__m128i)
- _mm_setzero_si128 (),
- (__mmask8) __U);
+ return (__m128i) __builtin_ia32_vcvtph2bf8s128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i)
+ _mm_setzero_si128 (),
+ (__mmask8) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtnesph_pbf8 (__m256h __A)
+_mm256_cvtsph_bf8 (__m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8s256_mask ((__v16hf) __A,
- (__v16qi)(__m128i)
- _mm_undefined_si128 (),
- (__mmask16) -1);
+ return (__m128i) __builtin_ia32_vcvtph2bf8s256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i)
+ _mm_undefined_si128 (),
+ (__mmask16) -1);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtnesph_pbf8 (__m128i __W, __mmask16 __U, __m256h __A)
+_mm256_mask_cvtsph_bf8 (__m128i __W, __mmask16 __U, __m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8s256_mask ((__v16hf) __A,
- (__v16qi)(__m128i) __W,
- (__mmask16) __U);
+ return (__m128i) __builtin_ia32_vcvtph2bf8s256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i) __W,
+ (__mmask16) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtnesph_pbf8 (__mmask16 __U, __m256h __A)
+_mm256_maskz_cvtsph_bf8 (__mmask16 __U, __m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2bf8s256_mask ((__v16hf) __A,
- (__v16qi)(__m128i)
- _mm_setzero_si128 (),
- (__mmask16) __U);
+ return (__m128i) __builtin_ia32_vcvtph2bf8s256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i)
+ _mm_setzero_si128 (),
+ (__mmask16) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtneph_phf8 (__m128h __A)
+_mm_cvtph_hf8 (__m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8128_mask ((__v8hf) __A,
- (__v16qi)(__m128i)
- _mm_undefined_si128 (),
- (__mmask8) -1);
+ return (__m128i) __builtin_ia32_vcvtph2hf8128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i)
+ _mm_undefined_si128 (),
+ (__mmask8) -1);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtneph_phf8 (__m128i __W, __mmask8 __U, __m128h __A)
+_mm_mask_cvtph_hf8 (__m128i __W, __mmask8 __U, __m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8128_mask ((__v8hf) __A,
- (__v16qi)(__m128i) __W,
- (__mmask8) __U);
+ return (__m128i) __builtin_ia32_vcvtph2hf8128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i) __W,
+ (__mmask8) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtneph_phf8 (__mmask8 __U, __m128h __A)
+_mm_maskz_cvtph_hf8 (__mmask8 __U, __m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8128_mask ((__v8hf) __A,
- (__v16qi)(__m128i)
- _mm_setzero_si128 (),
- (__mmask8) __U);
+ return (__m128i) __builtin_ia32_vcvtph2hf8128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i)
+ _mm_setzero_si128 (),
+ (__mmask8) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtneph_phf8 (__m256h __A)
+_mm256_cvtph_hf8 (__m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8256_mask ((__v16hf) __A,
- (__v16qi)(__m128i)
- _mm_undefined_si128 (),
- (__mmask16) -1);
+ return (__m128i) __builtin_ia32_vcvtph2hf8256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i)
+ _mm_undefined_si128 (),
+ (__mmask16) -1);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtneph_phf8 (__m128i __W, __mmask16 __U, __m256h __A)
+_mm256_mask_cvtph_hf8 (__m128i __W, __mmask16 __U, __m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8256_mask ((__v16hf) __A,
- (__v16qi)(__m128i) __W,
- (__mmask16) __U);
+ return (__m128i) __builtin_ia32_vcvtph2hf8256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i) __W,
+ (__mmask16) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtneph_phf8 (__mmask16 __U, __m256h __A)
+_mm256_maskz_cvtph_hf8 (__mmask16 __U, __m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8256_mask ((__v16hf) __A,
- (__v16qi)(__m128i)
- _mm_setzero_si128 (),
- (__mmask16) __U);
+ return (__m128i) __builtin_ia32_vcvtph2hf8256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i)
+ _mm_setzero_si128 (),
+ (__mmask16) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtnesph_phf8 (__m128h __A)
+_mm_cvtsph_hf8 (__m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8s128_mask ((__v8hf) __A,
- (__v16qi)(__m128i)
- _mm_undefined_si128 (),
- (__mmask8) -1);
+ return (__m128i) __builtin_ia32_vcvtph2hf8s128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i)
+ _mm_undefined_si128 (),
+ (__mmask8) -1);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtnesph_phf8 (__m128i __W, __mmask8 __U, __m128h __A)
+_mm_mask_cvtsph_hf8 (__m128i __W, __mmask8 __U, __m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8s128_mask ((__v8hf) __A,
- (__v16qi)(__m128i) __W,
- (__mmask8) __U);
+ return (__m128i) __builtin_ia32_vcvtph2hf8s128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i) __W,
+ (__mmask8) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtnesph_phf8 (__mmask8 __U, __m128h __A)
+_mm_maskz_cvtsph_hf8 (__mmask8 __U, __m128h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8s128_mask ((__v8hf) __A,
- (__v16qi)(__m128i)
- _mm_setzero_si128 (),
- (__mmask8) __U);
+ return (__m128i) __builtin_ia32_vcvtph2hf8s128_mask ((__v8hf) __A,
+ (__v16qi)(__m128i)
+ _mm_setzero_si128 (),
+ (__mmask8) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtnesph_phf8 (__m256h __A)
+_mm256_cvtsph_hf8 (__m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8s256_mask ((__v16hf) __A,
- (__v16qi)(__m128i)
- _mm_undefined_si128 (),
- (__mmask16) -1);
+ return (__m128i) __builtin_ia32_vcvtph2hf8s256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i)
+ _mm_undefined_si128 (),
+ (__mmask16) -1);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtnesph_phf8 (__m128i __W, __mmask16 __U, __m256h __A)
+_mm256_mask_cvtsph_hf8 (__m128i __W, __mmask16 __U, __m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8s256_mask ((__v16hf) __A,
- (__v16qi)(__m128i) __W,
- (__mmask16) __U);
+ return (__m128i) __builtin_ia32_vcvtph2hf8s256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i) __W,
+ (__mmask16) __U);
}
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtnesph_phf8 (__mmask16 __U, __m256h __A)
+_mm256_maskz_cvtsph_hf8 (__mmask16 __U, __m256h __A)
{
- return (__m128i) __builtin_ia32_vcvtneph2hf8s256_mask ((__v16hf) __A,
- (__v16qi)(__m128i)
- _mm_setzero_si128 (),
- (__mmask16) __U);
+ return (__m128i) __builtin_ia32_vcvtph2hf8s256_mask ((__v16hf) __A,
+ (__v16qi)(__m128i)
+ _mm_setzero_si128 (),
+ (__mmask16) __U);
}
extern __inline __m128h
diff --git a/gcc/config/i386/i386-builtin.def b/gcc/config/i386/i386-builtin.def
index edd5c2b..5880f42 100644
--- a/gcc/config/i386/i386-builtin.def
+++ b/gcc/config/i386/i386-builtin.def
@@ -3164,18 +3164,18 @@ BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvt2ph2hf8v32hf_mask, "__built
BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvt2ph2hf8sv8hf_mask, "__builtin_ia32_vcvt2ph2hf8s128_mask", IX86_BUILTIN_VCVT2PH2HF8S128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V8HF_V16QI_UHI)
BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvt2ph2hf8sv16hf_mask, "__builtin_ia32_vcvt2ph2hf8s256_mask", IX86_BUILTIN_VCVT2PH2HF8S256_MASK, UNKNOWN, (int) V32QI_FTYPE_V16HF_V16HF_V32QI_USI)
BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvt2ph2hf8sv32hf_mask, "__builtin_ia32_vcvt2ph2hf8s512_mask", IX86_BUILTIN_VCVT2PH2HF8S512_MASK, UNKNOWN, (int) V64QI_FTYPE_V32HF_V32HF_V64QI_UDI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtneph2bf8v8hf_mask, "__builtin_ia32_vcvtneph2bf8128_mask", IX86_BUILTIN_VCVTNEPH2BF8128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V16QI_UQI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtneph2bf8v16hf_mask, "__builtin_ia32_vcvtneph2bf8256_mask", IX86_BUILTIN_VCVTNEPH2BF8256_MASK, UNKNOWN, (int) V16QI_FTYPE_V16HF_V16QI_UHI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvtneph2bf8v32hf_mask, "__builtin_ia32_vcvtneph2bf8512_mask", IX86_BUILTIN_VCVTNEPH2BF8512_MASK, UNKNOWN, (int) V32QI_FTYPE_V32HF_V32QI_USI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtneph2bf8sv8hf_mask, "__builtin_ia32_vcvtneph2bf8s128_mask", IX86_BUILTIN_VCVTNEPH2BF8S128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V16QI_UQI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtneph2bf8sv16hf_mask, "__builtin_ia32_vcvtneph2bf8s256_mask", IX86_BUILTIN_VCVTNEPH2BF8S256_MASK, UNKNOWN, (int) V16QI_FTYPE_V16HF_V16QI_UHI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvtneph2bf8sv32hf_mask, "__builtin_ia32_vcvtneph2bf8s512_mask", IX86_BUILTIN_VCVTNEPH2BF8S512_MASK, UNKNOWN, (int) V32QI_FTYPE_V32HF_V32QI_USI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtneph2hf8v8hf_mask, "__builtin_ia32_vcvtneph2hf8128_mask", IX86_BUILTIN_VCVTNEPH2HF8128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V16QI_UQI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtneph2hf8v16hf_mask, "__builtin_ia32_vcvtneph2hf8256_mask", IX86_BUILTIN_VCVTNEPH2HF8256_MASK, UNKNOWN, (int) V16QI_FTYPE_V16HF_V16QI_UHI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvtneph2hf8v32hf_mask, "__builtin_ia32_vcvtneph2hf8512_mask", IX86_BUILTIN_VCVTNEPH2HF8512_MASK, UNKNOWN, (int) V32QI_FTYPE_V32HF_V32QI_USI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtneph2hf8sv8hf_mask, "__builtin_ia32_vcvtneph2hf8s128_mask", IX86_BUILTIN_VCVTNEPH2HF8S128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V16QI_UQI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtneph2hf8sv16hf_mask, "__builtin_ia32_vcvtneph2hf8s256_mask", IX86_BUILTIN_VCVTNEPH2HF8S256_MASK, UNKNOWN, (int) V16QI_FTYPE_V16HF_V16QI_UHI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvtneph2hf8sv32hf_mask, "__builtin_ia32_vcvtneph2hf8s512_mask", IX86_BUILTIN_VCVTNEPH2HF8S512_MASK, UNKNOWN, (int) V32QI_FTYPE_V32HF_V32QI_USI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtph2bf8v8hf_mask, "__builtin_ia32_vcvtph2bf8128_mask", IX86_BUILTIN_VCVTPH2BF8128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V16QI_UQI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtph2bf8v16hf_mask, "__builtin_ia32_vcvtph2bf8256_mask", IX86_BUILTIN_VCVTPH2BF8256_MASK, UNKNOWN, (int) V16QI_FTYPE_V16HF_V16QI_UHI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvtph2bf8v32hf_mask, "__builtin_ia32_vcvtph2bf8512_mask", IX86_BUILTIN_VCVTPH2BF8512_MASK, UNKNOWN, (int) V32QI_FTYPE_V32HF_V32QI_USI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtph2bf8sv8hf_mask, "__builtin_ia32_vcvtph2bf8s128_mask", IX86_BUILTIN_VCVTPH2BF8S128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V16QI_UQI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtph2bf8sv16hf_mask, "__builtin_ia32_vcvtph2bf8s256_mask", IX86_BUILTIN_VCVTPH2BF8S256_MASK, UNKNOWN, (int) V16QI_FTYPE_V16HF_V16QI_UHI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvtph2bf8sv32hf_mask, "__builtin_ia32_vcvtph2bf8s512_mask", IX86_BUILTIN_VCVTPH2BF8S512_MASK, UNKNOWN, (int) V32QI_FTYPE_V32HF_V32QI_USI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtph2hf8v8hf_mask, "__builtin_ia32_vcvtph2hf8128_mask", IX86_BUILTIN_VCVTPH2HF8128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V16QI_UQI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtph2hf8v16hf_mask, "__builtin_ia32_vcvtph2hf8256_mask", IX86_BUILTIN_VCVTPH2HF8256_MASK, UNKNOWN, (int) V16QI_FTYPE_V16HF_V16QI_UHI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvtph2hf8v32hf_mask, "__builtin_ia32_vcvtph2hf8512_mask", IX86_BUILTIN_VCVTPH2HF8512_MASK, UNKNOWN, (int) V32QI_FTYPE_V32HF_V32QI_USI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtph2hf8sv8hf_mask, "__builtin_ia32_vcvtph2hf8s128_mask", IX86_BUILTIN_VCVTPH2HF8S128_MASK, UNKNOWN, (int) V16QI_FTYPE_V8HF_V16QI_UQI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvtph2hf8sv16hf_mask, "__builtin_ia32_vcvtph2hf8s256_mask", IX86_BUILTIN_VCVTPH2HF8S256_MASK, UNKNOWN, (int) V16QI_FTYPE_V16HF_V16QI_UHI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvtph2hf8sv32hf_mask, "__builtin_ia32_vcvtph2hf8s512_mask", IX86_BUILTIN_VCVTPH2HF8S512_MASK, UNKNOWN, (int) V32QI_FTYPE_V32HF_V32QI_USI)
BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvthf82phv8hf_mask, "__builtin_ia32_vcvthf82ph128_mask", IX86_BUILTIN_VCVTHF82PH128_MASK, UNKNOWN, (int) V8HF_FTYPE_V16QI_V8HF_UQI)
BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_vcvthf82phv16hf_mask, "__builtin_ia32_vcvthf82ph256_mask", IX86_BUILTIN_VCVTHF82PH256_MASK, UNKNOWN, (int) V16HF_FTYPE_V16QI_V16HF_UHI)
BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_vcvthf82phv32hf_mask, "__builtin_ia32_vcvthf82ph512_mask", IX86_BUILTIN_VCVTHF82PH512_MASK, UNKNOWN, (int) V32HF_FTYPE_V32QI_V32HF_USI)
diff --git a/gcc/config/i386/sse.md b/gcc/config/i386/sse.md
index 582942c..74fc141 100644
--- a/gcc/config/i386/sse.md
+++ b/gcc/config/i386/sse.md
@@ -224,10 +224,10 @@
UNSPEC_VCVT2PH2BF8S
UNSPEC_VCVT2PH2HF8
UNSPEC_VCVT2PH2HF8S
- UNSPEC_VCVTNEPH2BF8
- UNSPEC_VCVTNEPH2BF8S
- UNSPEC_VCVTNEPH2HF8
- UNSPEC_VCVTNEPH2HF8S
+ UNSPEC_VCVTPH2BF8
+ UNSPEC_VCVTPH2BF8S
+ UNSPEC_VCVTPH2HF8
+ UNSPEC_VCVTPH2HF8S
UNSPEC_VCVTHF82PH
UNSPEC_VSCALEFBF16
UNSPEC_VRNDSCALEBF16
@@ -31837,45 +31837,45 @@
(define_mode_attr ph2fp8suff
[(V32HF "") (V16HF "{y}") (V8HF "{x}")])
-(define_int_iterator UNSPEC_NECONVERTPH2FP8
- [UNSPEC_VCVTNEPH2BF8 UNSPEC_VCVTNEPH2BF8S
- UNSPEC_VCVTNEPH2HF8 UNSPEC_VCVTNEPH2HF8S])
+(define_int_iterator UNSPEC_CONVERTPH2FP8
+ [UNSPEC_VCVTPH2BF8 UNSPEC_VCVTPH2BF8S
+ UNSPEC_VCVTPH2HF8 UNSPEC_VCVTPH2HF8S])
-(define_int_attr neconvertph2fp8
- [(UNSPEC_VCVTNEPH2BF8 "neph2bf8")
- (UNSPEC_VCVTNEPH2BF8S "neph2bf8s")
- (UNSPEC_VCVTNEPH2HF8 "neph2hf8")
- (UNSPEC_VCVTNEPH2HF8S "neph2hf8s")])
+(define_int_attr convertph2fp8
+ [(UNSPEC_VCVTPH2BF8 "ph2bf8")
+ (UNSPEC_VCVTPH2BF8S "ph2bf8s")
+ (UNSPEC_VCVTPH2HF8 "ph2hf8")
+ (UNSPEC_VCVTPH2HF8S "ph2hf8s")])
-(define_expand "vcvt<neconvertph2fp8>v8hf"
+(define_expand "vcvt<convertph2fp8>v8hf"
[(set (match_operand:V16QI 0 "register_operand")
(vec_concat:V16QI
(unspec:V8QI
[(match_operand:V8HF 1 "nonimmediate_operand")]
- UNSPEC_NECONVERTPH2FP8)
+ UNSPEC_CONVERTPH2FP8)
(match_dup 2)))]
"TARGET_AVX10_2_256"
"operands[2] = CONST0_RTX (V8QImode);")
-(define_insn "*vcvt<neconvertph2fp8>v8hf"
+(define_insn "*vcvt<convertph2fp8>v8hf"
[(set (match_operand:V16QI 0 "register_operand" "=v")
(vec_concat:V16QI
(unspec:V8QI
[(match_operand:V8HF 1 "nonimmediate_operand" "vm")]
- UNSPEC_NECONVERTPH2FP8)
+ UNSPEC_CONVERTPH2FP8)
(match_operand:V8QI 2 "const0_operand")))]
"TARGET_AVX10_2_256"
- "vcvt<neconvertph2fp8>{x}\t{%1, %0|%0, %1}"
+ "vcvt<convertph2fp8>{x}\t{%1, %0|%0, %1}"
[(set_attr "prefix" "evex")
(set_attr "mode" "HF")])
-(define_expand "vcvt<neconvertph2fp8>v8hf_mask"
+(define_expand "vcvt<convertph2fp8>v8hf_mask"
[(set (match_operand:V16QI 0 "register_operand")
(vec_concat:V16QI
(vec_merge:V8QI
(unspec:V8QI
[(match_operand:V8HF 1 "nonimmediate_operand")]
- UNSPEC_NECONVERTPH2FP8)
+ UNSPEC_CONVERTPH2FP8)
(vec_select:V8QI
(match_operand:V16QI 2 "nonimm_or_0_operand")
(parallel [(const_int 0) (const_int 1)
@@ -31887,13 +31887,13 @@
"TARGET_AVX10_2_256"
"operands[4] = CONST0_RTX (V8QImode);")
-(define_insn "*vcvt<neconvertph2fp8>v8hf_mask"
+(define_insn "*vcvt<convertph2fp8>v8hf_mask"
[(set (match_operand:V16QI 0 "register_operand" "=v")
(vec_concat:V16QI
(vec_merge:V8QI
(unspec:V8QI
[(match_operand:V8HF 1 "nonimmediate_operand" "vm")]
- UNSPEC_NECONVERTPH2FP8)
+ UNSPEC_CONVERTPH2FP8)
(vec_select:V8QI
(match_operand:V16QI 2 "nonimm_or_0_operand" "0C")
(parallel [(const_int 0) (const_int 1)
@@ -31903,16 +31903,16 @@
(match_operand:QI 3 "register_operand" "Yk"))
(match_operand:V8QI 4 "const0_operand")))]
"TARGET_AVX10_2_256"
- "vcvt<neconvertph2fp8>{x}\t{%1, %0%{%3%}%N2|%0%{%3%}%N2, %1}"
+ "vcvt<convertph2fp8>{x}\t{%1, %0%{%3%}%N2|%0%{%3%}%N2, %1}"
[(set_attr "prefix" "evex")])
-(define_insn "vcvt<neconvertph2fp8><mode><mask_name>"
+(define_insn "vcvt<convertph2fp8><mode><mask_name>"
[(set (match_operand:<ssebvecmode_2> 0 "register_operand" "=v")
(unspec:<ssebvecmode_2>
[(match_operand:VHF_256_512 1 "nonimmediate_operand" "vm")]
- UNSPEC_NECONVERTPH2FP8))]
+ UNSPEC_CONVERTPH2FP8))]
"TARGET_AVX10_2_256"
- "vcvt<neconvertph2fp8><ph2fp8suff>\t{%1, %0<mask_operand2>|%0<mask_operand2>, %1}"
+ "vcvt<convertph2fp8><ph2fp8suff>\t{%1, %0<mask_operand2>|%0<mask_operand2>, %1}"
[(set_attr "prefix" "evex")])
(define_insn "vcvthf82ph<mode><mask_name>"