aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--bits/libm-simd-decl-stubs.h11
-rw-r--r--math/bits/mathcalls.h2
-rw-r--r--sysdeps/unix/sysv/linux/x86_64/libmvec.abilist8
-rw-r--r--sysdeps/x86/fpu/bits/math-vector.h4
-rw-r--r--sysdeps/x86/fpu/finclude/math-vector-fortran.h4
-rw-r--r--sysdeps/x86_64/fpu/Makeconfig1
-rw-r--r--sysdeps/x86_64/fpu/Versions2
-rw-r--r--sysdeps/x86_64/fpu/libm-test-ulps17
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S20
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c27
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S1662
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S20
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c27
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S1601
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S20
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c27
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S510
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S20
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c28
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S476
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S20
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c28
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S509
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S20
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c28
-rw-r--r--sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S457
-rw-r--r--sysdeps/x86_64/fpu/svml_d_asinh2_core.S29
-rw-r--r--sysdeps/x86_64/fpu/svml_d_asinh4_core.S29
-rw-r--r--sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S25
-rw-r--r--sysdeps/x86_64/fpu/svml_d_asinh8_core.S25
-rw-r--r--sysdeps/x86_64/fpu/svml_s_asinhf16_core.S25
-rw-r--r--sysdeps/x86_64/fpu/svml_s_asinhf4_core.S29
-rw-r--r--sysdeps/x86_64/fpu/svml_s_asinhf8_core.S29
-rw-r--r--sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S25
-rw-r--r--sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c1
-rw-r--r--sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c1
-rw-r--r--sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c1
-rw-r--r--sysdeps/x86_64/fpu/test-double-libmvec-asinh.c3
-rw-r--r--sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c1
-rw-r--r--sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c1
-rw-r--r--sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c1
-rw-r--r--sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c1
-rw-r--r--sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c1
-rw-r--r--sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c1
-rw-r--r--sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c1
-rw-r--r--sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c3
-rw-r--r--sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c1
-rw-r--r--sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c1
-rw-r--r--sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c1
-rw-r--r--sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c1
50 files changed, 5784 insertions, 1 deletions
diff --git a/bits/libm-simd-decl-stubs.h b/bits/libm-simd-decl-stubs.h
index 21f1a43..bcaddb7 100644
--- a/bits/libm-simd-decl-stubs.h
+++ b/bits/libm-simd-decl-stubs.h
@@ -296,4 +296,15 @@
#define __DECL_SIMD_tanhf32x
#define __DECL_SIMD_tanhf64x
#define __DECL_SIMD_tanhf128x
+
+#define __DECL_SIMD_asinh
+#define __DECL_SIMD_asinhf
+#define __DECL_SIMD_asinhl
+#define __DECL_SIMD_asinhf16
+#define __DECL_SIMD_asinhf32
+#define __DECL_SIMD_asinhf64
+#define __DECL_SIMD_asinhf128
+#define __DECL_SIMD_asinhf32x
+#define __DECL_SIMD_asinhf64x
+#define __DECL_SIMD_asinhf128x
#endif
diff --git a/math/bits/mathcalls.h b/math/bits/mathcalls.h
index 3d1c205..40e055e 100644
--- a/math/bits/mathcalls.h
+++ b/math/bits/mathcalls.h
@@ -84,7 +84,7 @@ __MATHDECL_VEC (void,sincos,,
/* Hyperbolic arc cosine of X. */
__MATHCALL_VEC (acosh,, (_Mdouble_ __x));
/* Hyperbolic arc sine of X. */
-__MATHCALL (asinh,, (_Mdouble_ __x));
+__MATHCALL_VEC (asinh,, (_Mdouble_ __x));
/* Hyperbolic arc tangent of X. */
__MATHCALL_VEC (atanh,, (_Mdouble_ __x));
#endif
diff --git a/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist b/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
index e178cef..df265d6 100644
--- a/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
+++ b/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
@@ -49,6 +49,7 @@ GLIBC_2.22 _ZGVeN8vvv_sincos F
GLIBC_2.35 _ZGVbN2v_acos F
GLIBC_2.35 _ZGVbN2v_acosh F
GLIBC_2.35 _ZGVbN2v_asin F
+GLIBC_2.35 _ZGVbN2v_asinh F
GLIBC_2.35 _ZGVbN2v_atan F
GLIBC_2.35 _ZGVbN2v_atanh F
GLIBC_2.35 _ZGVbN2v_cbrt F
@@ -67,6 +68,7 @@ GLIBC_2.35 _ZGVbN2vv_hypot F
GLIBC_2.35 _ZGVbN4v_acosf F
GLIBC_2.35 _ZGVbN4v_acoshf F
GLIBC_2.35 _ZGVbN4v_asinf F
+GLIBC_2.35 _ZGVbN4v_asinhf F
GLIBC_2.35 _ZGVbN4v_atanf F
GLIBC_2.35 _ZGVbN4v_atanhf F
GLIBC_2.35 _ZGVbN4v_cbrtf F
@@ -85,6 +87,7 @@ GLIBC_2.35 _ZGVbN4vv_hypotf F
GLIBC_2.35 _ZGVcN4v_acos F
GLIBC_2.35 _ZGVcN4v_acosh F
GLIBC_2.35 _ZGVcN4v_asin F
+GLIBC_2.35 _ZGVcN4v_asinh F
GLIBC_2.35 _ZGVcN4v_atan F
GLIBC_2.35 _ZGVcN4v_atanh F
GLIBC_2.35 _ZGVcN4v_cbrt F
@@ -103,6 +106,7 @@ GLIBC_2.35 _ZGVcN4vv_hypot F
GLIBC_2.35 _ZGVcN8v_acosf F
GLIBC_2.35 _ZGVcN8v_acoshf F
GLIBC_2.35 _ZGVcN8v_asinf F
+GLIBC_2.35 _ZGVcN8v_asinhf F
GLIBC_2.35 _ZGVcN8v_atanf F
GLIBC_2.35 _ZGVcN8v_atanhf F
GLIBC_2.35 _ZGVcN8v_cbrtf F
@@ -121,6 +125,7 @@ GLIBC_2.35 _ZGVcN8vv_hypotf F
GLIBC_2.35 _ZGVdN4v_acos F
GLIBC_2.35 _ZGVdN4v_acosh F
GLIBC_2.35 _ZGVdN4v_asin F
+GLIBC_2.35 _ZGVdN4v_asinh F
GLIBC_2.35 _ZGVdN4v_atan F
GLIBC_2.35 _ZGVdN4v_atanh F
GLIBC_2.35 _ZGVdN4v_cbrt F
@@ -139,6 +144,7 @@ GLIBC_2.35 _ZGVdN4vv_hypot F
GLIBC_2.35 _ZGVdN8v_acosf F
GLIBC_2.35 _ZGVdN8v_acoshf F
GLIBC_2.35 _ZGVdN8v_asinf F
+GLIBC_2.35 _ZGVdN8v_asinhf F
GLIBC_2.35 _ZGVdN8v_atanf F
GLIBC_2.35 _ZGVdN8v_atanhf F
GLIBC_2.35 _ZGVdN8v_cbrtf F
@@ -157,6 +163,7 @@ GLIBC_2.35 _ZGVdN8vv_hypotf F
GLIBC_2.35 _ZGVeN16v_acosf F
GLIBC_2.35 _ZGVeN16v_acoshf F
GLIBC_2.35 _ZGVeN16v_asinf F
+GLIBC_2.35 _ZGVeN16v_asinhf F
GLIBC_2.35 _ZGVeN16v_atanf F
GLIBC_2.35 _ZGVeN16v_atanhf F
GLIBC_2.35 _ZGVeN16v_cbrtf F
@@ -175,6 +182,7 @@ GLIBC_2.35 _ZGVeN16vv_hypotf F
GLIBC_2.35 _ZGVeN8v_acos F
GLIBC_2.35 _ZGVeN8v_acosh F
GLIBC_2.35 _ZGVeN8v_asin F
+GLIBC_2.35 _ZGVeN8v_asinh F
GLIBC_2.35 _ZGVeN8v_atan F
GLIBC_2.35 _ZGVeN8v_atanh F
GLIBC_2.35 _ZGVeN8v_cbrt F
diff --git a/sysdeps/x86/fpu/bits/math-vector.h b/sysdeps/x86/fpu/bits/math-vector.h
index 3c657f6..71b7d66 100644
--- a/sysdeps/x86/fpu/bits/math-vector.h
+++ b/sysdeps/x86/fpu/bits/math-vector.h
@@ -130,6 +130,10 @@
# define __DECL_SIMD_tanh __DECL_SIMD_x86_64
# undef __DECL_SIMD_tanhf
# define __DECL_SIMD_tanhf __DECL_SIMD_x86_64
+# undef __DECL_SIMD_asinh
+# define __DECL_SIMD_asinh __DECL_SIMD_x86_64
+# undef __DECL_SIMD_asinhf
+# define __DECL_SIMD_asinhf __DECL_SIMD_x86_64
# endif
#endif
diff --git a/sysdeps/x86/fpu/finclude/math-vector-fortran.h b/sysdeps/x86/fpu/finclude/math-vector-fortran.h
index c7f8194..4d3afdf 100644
--- a/sysdeps/x86/fpu/finclude/math-vector-fortran.h
+++ b/sysdeps/x86/fpu/finclude/math-vector-fortran.h
@@ -64,6 +64,8 @@
!GCC$ builtin (erff) attributes simd (notinbranch) if('x86_64')
!GCC$ builtin (tanh) attributes simd (notinbranch) if('x86_64')
!GCC$ builtin (tanhf) attributes simd (notinbranch) if('x86_64')
+!GCC$ builtin (asinh) attributes simd (notinbranch) if('x86_64')
+!GCC$ builtin (asinhf) attributes simd (notinbranch) if('x86_64')
!GCC$ builtin (cos) attributes simd (notinbranch) if('x32')
!GCC$ builtin (cosf) attributes simd (notinbranch) if('x32')
@@ -113,3 +115,5 @@
!GCC$ builtin (erff) attributes simd (notinbranch) if('x32')
!GCC$ builtin (tanh) attributes simd (notinbranch) if('x32')
!GCC$ builtin (tanhf) attributes simd (notinbranch) if('x32')
+!GCC$ builtin (asinh) attributes simd (notinbranch) if('x32')
+!GCC$ builtin (asinhf) attributes simd (notinbranch) if('x32')
diff --git a/sysdeps/x86_64/fpu/Makeconfig b/sysdeps/x86_64/fpu/Makeconfig
index 26df8d4..2ff33c7 100644
--- a/sysdeps/x86_64/fpu/Makeconfig
+++ b/sysdeps/x86_64/fpu/Makeconfig
@@ -25,6 +25,7 @@ libmvec-funcs = \
acos \
acosh \
asin \
+ asinh \
atan \
atan2 \
atanh \
diff --git a/sysdeps/x86_64/fpu/Versions b/sysdeps/x86_64/fpu/Versions
index adcbe0f..e6ead13 100644
--- a/sysdeps/x86_64/fpu/Versions
+++ b/sysdeps/x86_64/fpu/Versions
@@ -17,6 +17,7 @@ libmvec {
_ZGVbN2v_acos; _ZGVcN4v_acos; _ZGVdN4v_acos; _ZGVeN8v_acos;
_ZGVbN2v_acosh; _ZGVcN4v_acosh; _ZGVdN4v_acosh; _ZGVeN8v_acosh;
_ZGVbN2v_asin; _ZGVcN4v_asin; _ZGVdN4v_asin; _ZGVeN8v_asin;
+ _ZGVbN2v_asinh; _ZGVcN4v_asinh; _ZGVdN4v_asinh; _ZGVeN8v_asinh;
_ZGVbN2v_atan; _ZGVcN4v_atan; _ZGVdN4v_atan; _ZGVeN8v_atan;
_ZGVbN2v_atanh; _ZGVcN4v_atanh; _ZGVdN4v_atanh; _ZGVeN8v_atanh;
_ZGVbN2v_cbrt; _ZGVcN4v_cbrt; _ZGVdN4v_cbrt; _ZGVeN8v_cbrt;
@@ -35,6 +36,7 @@ libmvec {
_ZGVbN4v_acosf; _ZGVcN8v_acosf; _ZGVdN8v_acosf; _ZGVeN16v_acosf;
_ZGVbN4v_acoshf; _ZGVcN8v_acoshf; _ZGVdN8v_acoshf; _ZGVeN16v_acoshf;
_ZGVbN4v_asinf; _ZGVcN8v_asinf; _ZGVdN8v_asinf; _ZGVeN16v_asinf;
+ _ZGVbN4v_asinhf; _ZGVcN8v_asinhf; _ZGVdN8v_asinhf; _ZGVeN16v_asinhf;
_ZGVbN4v_atanf; _ZGVcN8v_atanf; _ZGVdN8v_atanf; _ZGVeN16v_atanf;
_ZGVbN4v_atanhf; _ZGVcN8v_atanhf; _ZGVdN8v_atanhf; _ZGVeN16v_atanhf;
_ZGVbN4v_cbrtf; _ZGVcN8v_cbrtf; _ZGVdN8v_cbrtf; _ZGVeN16v_cbrtf;
diff --git a/sysdeps/x86_64/fpu/libm-test-ulps b/sysdeps/x86_64/fpu/libm-test-ulps
index bfaad7a..71e9fce 100644
--- a/sysdeps/x86_64/fpu/libm-test-ulps
+++ b/sysdeps/x86_64/fpu/libm-test-ulps
@@ -157,6 +157,23 @@ float: 3
float128: 4
ldouble: 5
+Function: "asinh_vlen2":
+double: 1
+
+Function: "asinh_vlen4":
+double: 1
+float: 1
+
+Function: "asinh_vlen4_avx2":
+double: 1
+
+Function: "asinh_vlen8":
+double: 1
+float: 1
+
+Function: "asinh_vlen8_avx2":
+float: 1
+
Function: "atan":
double: 1
float: 1
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S
new file mode 100644
index 0000000..ddd1c3c
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S
@@ -0,0 +1,20 @@
+/* SSE2 version of vectorized asinh, vector length is 2.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define _ZGVbN2v_asinh _ZGVbN2v_asinh_sse2
+#include "../svml_d_asinh2_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c
new file mode 100644
index 0000000..37452d0
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c
@@ -0,0 +1,27 @@
+/* Multiple versions of vectorized asinh, vector length is 2.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define SYMBOL_NAME _ZGVbN2v_asinh
+#include "ifunc-mathvec-sse4_1.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVbN2v_asinh, __GI__ZGVbN2v_asinh, __redirect__ZGVbN2v_asinh)
+ __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S
new file mode 100644
index 0000000..0fe130f
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S
@@ -0,0 +1,1662 @@
+/* Function asinh vectorized with SSE4.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ https://www.gnu.org/licenses/. */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ * Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *
+ * Special cases:
+ *
+ * asinh(NaN) = quiet NaN, and raise invalid exception
+ * asinh(INF) = that INF
+ * asinh(0) = that 0
+ *
+ */
+
+/* Offsets for data table __svml_dasinh_data_internal
+ */
+#define Log_HA_table 0
+#define Log_LA_table 8208
+#define poly_coeff 12320
+#define ExpMask 12384
+#define Two10 12400
+#define MinLog1p 12416
+#define MaxLog1p 12432
+#define One 12448
+#define SgnMask 12464
+#define XThreshold 12480
+#define XhMask 12496
+#define Threshold 12512
+#define Bias 12528
+#define Bias1 12544
+#define ExpMask0 12560
+#define ExpMask2 12576
+#define L2 12592
+#define dBigThreshold 12608
+#define dC2 12624
+#define dC3 12640
+#define dC4 12656
+#define dC5 12672
+#define dHalf 12688
+#define dLargestFinite 12704
+#define dLittleThreshold 12720
+#define dSign 12736
+#define dThirtyOne 12752
+#define dTopMask12 12768
+#define dTopMask26 12784
+#define dTopMask29 12800
+#define XScale 12816
+
+/* Lookup bias for data table __svml_dasinh_data_internal. */
+#define Table_Lookup_Bias -0x405ff0
+
+#include <sysdep.h>
+
+ .text
+ .section .text.sse4,"ax",@progbits
+ENTRY(_ZGVbN2v_asinh_sse4)
+ pushq %rbp
+ cfi_def_cfa_offset(16)
+ movq %rsp, %rbp
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+ andq $-32, %rsp
+ subq $64, %rsp
+ movaps %xmm0, %xmm13
+
+/*
+ * Split X into high and low parts, XHi (<= 26 bits) and XLo (<= 27 bits)
+ * We could use either X or |X| here, but it doesn't seem to matter
+ */
+ movups dTopMask26+__svml_dasinh_data_internal(%rip), %xmm15
+ movaps %xmm13, %xmm7
+ andps %xmm13, %xmm15
+ lea Table_Lookup_Bias+__svml_dasinh_data_internal(%rip), %rsi
+
+/*
+ * Compute X^2 = (XHi + XLo)^2 = XHi^2 + XLo * (X + XHi)
+ * The two parts are shifted off by around 26 bits. So even though
+ * the low bit will not in general be exact, it's near enough
+ */
+ movaps %xmm15, %xmm8
+ mulpd %xmm15, %xmm8
+ subpd %xmm15, %xmm7
+ addpd %xmm13, %xmm15
+
+/* Load the constant 1 and a sign mask */
+ movups One+__svml_dasinh_data_internal(%rip), %xmm12
+
+/*
+ * Finally, express Y + W = X^2 + 1 accurately where Y has <= 29 bits.
+ * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
+ * as the dominant component in the compensated summation. Otherwise,
+ * if |X| >= 1, then since X2Hi only has 52 significant bits, the basic
+ * addition will be exact anyway until we get to |X| >= 2^53. But by
+ * that time the log function is well-conditioned enough that the
+ * rounding error doesn't matter. Hence we can treat 1 as dominant even
+ * if it literally isn't.
+ */
+ movaps %xmm12, %xmm3
+ movaps %xmm12, %xmm5
+ addpd %xmm8, %xmm3
+ mulpd %xmm15, %xmm7
+ subpd %xmm3, %xmm5
+ movups dTopMask29+__svml_dasinh_data_internal(%rip), %xmm6
+ andps %xmm3, %xmm6
+
+/*
+ * Compute R = 1/sqrt(Y + W) * (1 + d)
+ * Force R to <= 12 significant bits in case it isn't already
+ * This means that R * Y and R^2 * Y are exactly representable.
+ */
+ cvtpd2ps %xmm6, %xmm1
+ addpd %xmm8, %xmm5
+ subpd %xmm6, %xmm3
+
+/*
+ * Unfortunately, we can still be in trouble if |X| <= 2^-10, since
+ * the absolute error 2^-(12+53)-ish in sqrt(1 + X^2) gets scaled up
+ * by 1/X and comes close to our threshold. Hence if |X| <= 2^-9,
+ * perform an alternative computation
+ * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
+ * X2 = X^2
+ */
+ addpd %xmm7, %xmm8
+ addpd %xmm7, %xmm5
+ movlhps %xmm1, %xmm1
+ rsqrtps %xmm1, %xmm4
+ addpd %xmm3, %xmm5
+ cvtps2pd %xmm4, %xmm2
+ andps dTopMask12+__svml_dasinh_data_internal(%rip), %xmm2
+
+/*
+ * Compute e = -(2 * d + d^2)
+ * The first FMR is exact, and the rounding error in the other is acceptable
+ * since d and e are ~ 2^-12
+ */
+ movaps %xmm12, %xmm1
+
+/*
+ * Compute S = (Y/sqrt(Y + W)) * (1 + d)
+ * and T = (W/sqrt(Y + W)) * (1 + d)
+ * so that S + T = sqrt(Y + W) * (1 + d)
+ * S is exact, and the rounding error in T is OK.
+ */
+ mulpd %xmm2, %xmm6
+ mulpd %xmm2, %xmm5
+ movaps %xmm2, %xmm0
+
+/*
+ * Obtain sqrt(1 + X^2) - 1 in two pieces
+ * sqrt(1 + X^2) - 1
+ * = sqrt(Y + W) - 1
+ * = (S + T) * (1 + Corr) - 1
+ * = [S - 1] + [T + (S + T) * Corr]
+ * We need a compensated summation for the last part. We treat S - 1
+ * as the larger part; it certainly is until about X < 2^-4, and in that
+ * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
+ * Final sum is dTmp5 (hi) + dTmp7 (lo)
+ */
+ movaps %xmm6, %xmm3
+ mulpd %xmm6, %xmm0
+ mulpd %xmm5, %xmm2
+ subpd %xmm0, %xmm1
+ addpd %xmm5, %xmm3
+ subpd %xmm12, %xmm6
+ subpd %xmm2, %xmm1
+ movups SgnMask+__svml_dasinh_data_internal(%rip), %xmm9
+ movaps %xmm12, %xmm4
+
+/*
+ * Get the absolute value of the input, since we will exploit antisymmetry
+ * and mostly assume X >= 0 in the core computation
+ */
+ movaps %xmm9, %xmm10
+ andps %xmm13, %xmm10
+
+/*
+ * Check whether the input is finite, by checking |X| <= MaxFloat
+ * Otherwise set the rangemask so that the callout will get used.
+ * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
+ */
+ movaps %xmm10, %xmm14
+
+/*
+ * The following computation can go wrong for very large X, basically
+ * because X^2 overflows. But for large X we have
+ * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
+ * we can just later stick X back into the log and tweak up the exponent.
+ * Actually we scale X by 2^-30 and tweak the exponent up by 31,
+ * to stay in the safe range for the later log computation.
+ * Compute a flag now telling us when do do this.
+ */
+ movaps %xmm10, %xmm11
+ cmpnlepd dLargestFinite+__svml_dasinh_data_internal(%rip), %xmm14
+ cmpltpd dBigThreshold+__svml_dasinh_data_internal(%rip), %xmm11
+ movmskpd %xmm14, %edx
+
+/*
+ * Now 1 / (1 + d)
+ * = 1 / (1 + (sqrt(1 - e) - 1))
+ * = 1 / sqrt(1 - e)
+ * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 +
+ * 63/256 * e^5 + 231/1024 * e^6 + ....
+ * So compute the first five nonconstant terms of that, so that
+ * we have a relative correction (1 + Corr) to apply to S etc.
+ * C1 = 1/2
+ * C2 = 3/8
+ * C3 = 5/16
+ * C4 = 35/128
+ * C5 = 63/256
+ */
+ movups dC5+__svml_dasinh_data_internal(%rip), %xmm14
+ movups dHalf+__svml_dasinh_data_internal(%rip), %xmm15
+ mulpd %xmm1, %xmm14
+
+/* dX2over2 = X^2/2 */
+ mulpd %xmm15, %xmm8
+ addpd dC4+__svml_dasinh_data_internal(%rip), %xmm14
+ mulpd %xmm1, %xmm14
+ addpd dC3+__svml_dasinh_data_internal(%rip), %xmm14
+ mulpd %xmm1, %xmm14
+ addpd dC2+__svml_dasinh_data_internal(%rip), %xmm14
+ mulpd %xmm1, %xmm14
+ addpd %xmm15, %xmm14
+ mulpd %xmm14, %xmm1
+ mulpd %xmm3, %xmm1
+ addpd %xmm1, %xmm5
+ addpd %xmm6, %xmm5
+
+/* dX4over4 = X^4/4 */
+ movaps %xmm8, %xmm6
+
+/* dX46 = -X^4/4 + X^6/8 */
+ movaps %xmm8, %xmm7
+ mulpd %xmm8, %xmm6
+ mulpd %xmm6, %xmm7
+ subpd %xmm6, %xmm7
+
+/* dX46over2 = -X^4/8 + x^6/16 */
+ mulpd %xmm7, %xmm15
+
+/* Now multiplex the two possible computations */
+ movaps %xmm10, %xmm3
+ cmplepd dLittleThreshold+__svml_dasinh_data_internal(%rip), %xmm3
+ addpd %xmm15, %xmm8
+ movaps %xmm3, %xmm1
+ andps %xmm3, %xmm8
+ andnps %xmm5, %xmm1
+
+/*
+ * Now we feed into the log1p code, using H in place of _VARG1 and
+ * also adding L into Xl.
+ * compute 1+x as high, low parts
+ */
+ movaps %xmm12, %xmm5
+ orps %xmm8, %xmm1
+ movaps %xmm11, %xmm3
+
+/*
+ * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
+ * It's always safe to assume |X| is larger.
+ * This is the final 2-part argument to the log1p function
+ */
+ addpd %xmm10, %xmm1
+ maxpd %xmm1, %xmm5
+ minpd %xmm1, %xmm4
+
+/* Now multiplex to the case X = 2^-30 * |input|, Xl = dL = 0 in the "big" case. */
+ movups XScale+__svml_dasinh_data_internal(%rip), %xmm8
+ andps %xmm9, %xmm1
+ mulpd %xmm8, %xmm10
+ cmpltpd XThreshold+__svml_dasinh_data_internal(%rip), %xmm1
+ movaps %xmm5, %xmm9
+ andnps %xmm10, %xmm3
+ addpd %xmm4, %xmm9
+ orps XhMask+__svml_dasinh_data_internal(%rip), %xmm1
+ andps %xmm1, %xmm9
+ subpd %xmm9, %xmm5
+ andps %xmm11, %xmm9
+
+/* Now resume the main code. */
+ movups ExpMask+__svml_dasinh_data_internal(%rip), %xmm10
+ orps %xmm9, %xmm3
+
+/* preserve mantissa, set input exponent to 2^(-10) */
+ andps %xmm3, %xmm10
+
+/* exponent bits */
+ movaps %xmm3, %xmm7
+ orps Two10+__svml_dasinh_data_internal(%rip), %xmm10
+ psrlq $20, %xmm7
+
+/* reciprocal approximation good to at least 11 bits */
+ cvtpd2ps %xmm10, %xmm1
+ addpd %xmm5, %xmm4
+ movlhps %xmm1, %xmm1
+ andps %xmm11, %xmm4
+ rcpps %xmm1, %xmm0
+ cvtps2pd %xmm0, %xmm0
+
+/* round reciprocal to nearest integer, will have 1+9 mantissa bits */
+ movups .FLT_30(%rip), %xmm6
+ movaps %xmm11, %xmm1
+ addpd %xmm6, %xmm0
+ subpd %xmm6, %xmm0
+
+/* exponent of X needed to scale Xl */
+ movdqu ExpMask0+__svml_dasinh_data_internal(%rip), %xmm5
+
+/* 2^ (-10-exp(X) ) */
+ movdqu ExpMask2+__svml_dasinh_data_internal(%rip), %xmm2
+ pand %xmm3, %xmm5
+ psubq %xmm5, %xmm2
+
+/* scale DblRcp */
+ mulpd %xmm0, %xmm2
+
+/* argument reduction */
+ mulpd %xmm2, %xmm3
+ mulpd %xmm2, %xmm4
+ subpd %xmm12, %xmm3
+ addpd %xmm4, %xmm3
+
+/* polynomial */
+ movups poly_coeff+__svml_dasinh_data_internal(%rip), %xmm12
+ movaps %xmm3, %xmm2
+ pshufd $221, %xmm7, %xmm8
+ mulpd %xmm3, %xmm12
+
+/* biased exponent in DP format */
+ cvtdq2pd %xmm8, %xmm14
+ addpd poly_coeff+16+__svml_dasinh_data_internal(%rip), %xmm12
+ mulpd %xmm3, %xmm2
+
+/* Add 31 to the exponent in the "large" case to get log(2 * input) */
+ movups dThirtyOne+__svml_dasinh_data_internal(%rip), %xmm9
+
+/* exponent*log(2.0) */
+ movups Threshold+__svml_dasinh_data_internal(%rip), %xmm5
+ addpd %xmm14, %xmm9
+ cmpltpd %xmm0, %xmm5
+ mulpd %xmm2, %xmm12
+ andps %xmm11, %xmm14
+
+/*
+ * prepare table index
+ * table lookup
+ */
+ movaps %xmm0, %xmm11
+ movups poly_coeff+32+__svml_dasinh_data_internal(%rip), %xmm0
+ andnps %xmm9, %xmm1
+ mulpd %xmm3, %xmm0
+ addpd poly_coeff+48+__svml_dasinh_data_internal(%rip), %xmm0
+ addpd %xmm12, %xmm0
+
+/* reconstruction */
+ mulpd %xmm0, %xmm2
+ andps Bias+__svml_dasinh_data_internal(%rip), %xmm5
+ psrlq $40, %xmm11
+ orps Bias1+__svml_dasinh_data_internal(%rip), %xmm5
+ orps %xmm14, %xmm1
+ movd %xmm11, %eax
+ pshufd $2, %xmm11, %xmm11
+
+/* Finally, reincorporate the original sign. */
+ movups dSign+__svml_dasinh_data_internal(%rip), %xmm0
+ subpd %xmm5, %xmm1
+ addpd %xmm2, %xmm3
+ movd %xmm11, %ecx
+ mulpd L2+__svml_dasinh_data_internal(%rip), %xmm1
+ movslq %eax, %rax
+ andps %xmm13, %xmm0
+ movslq %ecx, %rcx
+ movsd (%rsi,%rax), %xmm6
+ movhpd (%rsi,%rcx), %xmm6
+ addpd %xmm3, %xmm6
+ addpd %xmm6, %xmm1
+ pxor %xmm1, %xmm0
+ testl %edx, %edx
+
+/* Go to special inputs processing branch */
+ jne L(SPECIAL_VALUES_BRANCH)
+ # LOE rbx r12 r13 r14 r15 edx xmm0 xmm13
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+ movq %rbp, %rsp
+ popq %rbp
+ cfi_def_cfa(7, 8)
+ cfi_restore(6)
+ ret
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+ movups %xmm13, 32(%rsp)
+ movups %xmm0, 48(%rsp)
+ # LOE rbx r12 r13 r14 r15 edx
+
+ xorl %eax, %eax
+ movq %r12, 16(%rsp)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -48; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xd0, 0xff, 0xff, 0xff, 0x22
+ movl %eax, %r12d
+ movq %r13, 8(%rsp)
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -56; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc8, 0xff, 0xff, 0xff, 0x22
+ movl %edx, %r13d
+ movq %r14, (%rsp)
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -64; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+ btl %r12d, %r13d
+
+/* Call scalar math function */
+ jc L(SCALAR_MATH_CALL)
+ # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+ incl %r12d
+ cmpl $2, %r12d
+
+/* Check bits in range mask */
+ jl L(RANGEMASK_CHECK)
+ # LOE rbx r15 r12d r13d
+
+ movq 16(%rsp), %r12
+ cfi_restore(12)
+ movq 8(%rsp), %r13
+ cfi_restore(13)
+ movq (%rsp), %r14
+ cfi_restore(14)
+ movups 48(%rsp), %xmm0
+
+/* Go to exit */
+ jmp L(EXIT)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -48; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xd0, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -56; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc8, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -64; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r12 r13 r14 r15 xmm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+ movl %r12d, %r14d
+ movsd 32(%rsp,%r14,8), %xmm0
+ call asinh@PLT
+ # LOE rbx r14 r15 r12d r13d xmm0
+
+ movsd %xmm0, 48(%rsp,%r14,8)
+
+/* Process special inputs in loop */
+ jmp L(SPECIAL_VALUES_LOOP)
+ # LOE rbx r15 r12d r13d
+END(_ZGVbN2v_asinh_sse4)
+
+ .section .rodata, "a"
+ .align 16
+
+#ifdef __svml_dasinh_data_internal_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+ __declspec(align(16)) VUINT32 Log_HA_table[(1<<10)+2][2];
+ __declspec(align(16)) VUINT32 Log_LA_table[(1<<9)+1][2];
+ __declspec(align(16)) VUINT32 poly_coeff[4][2][2];
+ __declspec(align(16)) VUINT32 ExpMask[2][2];
+ __declspec(align(16)) VUINT32 Two10[2][2];
+ __declspec(align(16)) VUINT32 MinLog1p[2][2];
+ __declspec(align(16)) VUINT32 MaxLog1p[2][2];
+ __declspec(align(16)) VUINT32 One[2][2];
+ __declspec(align(16)) VUINT32 SgnMask[2][2];
+ __declspec(align(16)) VUINT32 XThreshold[2][2];
+ __declspec(align(16)) VUINT32 XhMask[2][2];
+ __declspec(align(16)) VUINT32 Threshold[2][2];
+ __declspec(align(16)) VUINT32 Bias[2][2];
+ __declspec(align(16)) VUINT32 Bias1[2][2];
+ __declspec(align(16)) VUINT32 ExpMask0[2][2];
+ __declspec(align(16)) VUINT32 ExpMask2[2][2];
+ __declspec(align(16)) VUINT32 L2[2][2];
+ __declspec(align(16)) VUINT32 dBigThreshold[2][2];
+ __declspec(align(16)) VUINT32 dC2[2][2];
+ __declspec(align(16)) VUINT32 dC3[2][2];
+ __declspec(align(16)) VUINT32 dC4[2][2];
+ __declspec(align(16)) VUINT32 dC5[2][2];
+ __declspec(align(16)) VUINT32 dHalf[2][2];
+ __declspec(align(16)) VUINT32 dLargestFinite[2][2];
+ __declspec(align(16)) VUINT32 dLittleThreshold[2][2];
+ __declspec(align(16)) VUINT32 dSign[2][2];
+ __declspec(align(16)) VUINT32 dThirtyOne[2][2];
+ __declspec(align(16)) VUINT32 dTopMask12[2][2];
+ __declspec(align(16)) VUINT32 dTopMask26[2][2];
+ __declspec(align(16)) VUINT32 dTopMask29[2][2];
+ __declspec(align(16)) VUINT32 XScale[2][2];
+} __svml_dasinh_data_internal;
+#endif
+__svml_dasinh_data_internal:
+ /* Log_HA_table */
+ .quad 0xc086232bdd7a8300, 0xbe1ce91eef3fb100
+ .quad 0xc086232fdc7ad828, 0xbe1cefcffda73b6a
+ .quad 0xc0862333d97d2ba0, 0xbe1cef406748f1ff
+ .quad 0xc0862337d48378e0, 0xbe1cef2a9429925a
+ .quad 0xc086233bcd8fb878, 0xbe1cf138d17ebecb
+ .quad 0xc086233fc4a3e018, 0xbe1ceff2dbbbb29e
+ .quad 0xc0862343b9c1e270, 0xbe1cf1a42aae437b
+ .quad 0xc0862347acebaf68, 0xbe1cef3b152048af
+ .quad 0xc086234b9e2333f0, 0xbe1cef20e127805e
+ .quad 0xc086234f8d6a5a30, 0xbe1cf00ad6052cf4
+ .quad 0xc08623537ac30980, 0xbe1cefc4642ee597
+ .quad 0xc0862357662f2660, 0xbe1cf1f277d36e16
+ .quad 0xc086235b4fb092a0, 0xbe1ceed009e8d8e6
+ .quad 0xc086235f37492d28, 0xbe1cf1e4038cb362
+ .quad 0xc08623631cfad250, 0xbe1cf0b0873b8557
+ .quad 0xc086236700c75b98, 0xbe1cf15bb3227c0b
+ .quad 0xc086236ae2b09fe0, 0xbe1cf151ef8ca9ed
+ .quad 0xc086236ec2b87358, 0xbe1cefe1dc2cd2ed
+ .quad 0xc0862372a0e0a780, 0xbe1cf0d1eec5454f
+ .quad 0xc08623767d2b0b48, 0xbe1ceeefd570bbce
+ .quad 0xc086237a57996af0, 0xbe1cee99ae91b3a7
+ .quad 0xc086237e302d9028, 0xbe1cf0412830fbd1
+ .quad 0xc086238206e94218, 0xbe1ceee898588610
+ .quad 0xc0862385dbce4548, 0xbe1cee9a1fbcaaea
+ .quad 0xc0862389aede5bc0, 0xbe1ceed8e7cc1ad6
+ .quad 0xc086238d801b4500, 0xbe1cf10c8d059da6
+ .quad 0xc08623914f86be18, 0xbe1ceee6c63a8165
+ .quad 0xc08623951d228180, 0xbe1cf0c3592d2ff1
+ .quad 0xc0862398e8f04758, 0xbe1cf0026cc4cb1b
+ .quad 0xc086239cb2f1c538, 0xbe1cf15d48d8e670
+ .quad 0xc08623a07b28ae60, 0xbe1cef359363787c
+ .quad 0xc08623a44196b390, 0xbe1cefdf1ab2e82c
+ .quad 0xc08623a8063d8338, 0xbe1cefe43c02aa84
+ .quad 0xc08623abc91ec960, 0xbe1cf044f5ae35b7
+ .quad 0xc08623af8a3c2fb8, 0xbe1cf0b0b4001e1b
+ .quad 0xc08623b349975d98, 0xbe1cf1bae76dfbcf
+ .quad 0xc08623b70731f810, 0xbe1cef0a72e13a62
+ .quad 0xc08623bac30da1c8, 0xbe1cf184007d2b6b
+ .quad 0xc08623be7d2bfb40, 0xbe1cf16f4b239e98
+ .quad 0xc08623c2358ea2a0, 0xbe1cf0976acada87
+ .quad 0xc08623c5ec3733d0, 0xbe1cf066318a16ff
+ .quad 0xc08623c9a1274880, 0xbe1ceffaa7148798
+ .quad 0xc08623cd54607820, 0xbe1cf23ab02e9b6e
+ .quad 0xc08623d105e45800, 0xbe1cefdfef7d4fde
+ .quad 0xc08623d4b5b47b20, 0xbe1cf17fece44f2b
+ .quad 0xc08623d863d27270, 0xbe1cf18f907d0d7c
+ .quad 0xc08623dc103fccb0, 0xbe1cee61fe072c98
+ .quad 0xc08623dfbafe1668, 0xbe1cf022dd891e2f
+ .quad 0xc08623e3640eda20, 0xbe1ceecc1daf4358
+ .quad 0xc08623e70b73a028, 0xbe1cf0173c4fa380
+ .quad 0xc08623eab12deec8, 0xbe1cf16a2150c2f4
+ .quad 0xc08623ee553f4a30, 0xbe1cf1bf980b1f4b
+ .quad 0xc08623f1f7a93480, 0xbe1cef8b731663c2
+ .quad 0xc08623f5986d2dc0, 0xbe1cee9a664d7ef4
+ .quad 0xc08623f9378cb3f0, 0xbe1cf1eda2af6400
+ .quad 0xc08623fcd5094320, 0xbe1cf1923f9d68d7
+ .quad 0xc086240070e45548, 0xbe1cf0747cd3e03a
+ .quad 0xc08624040b1f6260, 0xbe1cf22ee855bd6d
+ .quad 0xc0862407a3bbe078, 0xbe1cf0d57360c00b
+ .quad 0xc086240b3abb4398, 0xbe1ceebc815cd575
+ .quad 0xc086240ed01efdd0, 0xbe1cf03bfb970951
+ .quad 0xc086241263e87f50, 0xbe1cf16e74768529
+ .quad 0xc0862415f6193658, 0xbe1cefec64b8becb
+ .quad 0xc086241986b28f30, 0xbe1cf0838d210baa
+ .quad 0xc086241d15b5f448, 0xbe1cf0ea86e75b11
+ .quad 0xc0862420a324ce28, 0xbe1cf1708d11d805
+ .quad 0xc08624242f008380, 0xbe1ceea988c5a417
+ .quad 0xc0862427b94a7910, 0xbe1cef166a7bbca5
+ .quad 0xc086242b420411d0, 0xbe1cf0c9d9e86a38
+ .quad 0xc086242ec92eaee8, 0xbe1cef0946455411
+ .quad 0xc08624324ecbaf98, 0xbe1cefea60907739
+ .quad 0xc0862435d2dc7160, 0xbe1cf1ed0934ce42
+ .quad 0xc086243955624ff8, 0xbe1cf191ba746c7d
+ .quad 0xc086243cd65ea548, 0xbe1ceeec78cf2a7e
+ .quad 0xc086244055d2c968, 0xbe1cef345284c119
+ .quad 0xc0862443d3c012b8, 0xbe1cf24f77355219
+ .quad 0xc08624475027d5e8, 0xbe1cf05bf087e114
+ .quad 0xc086244acb0b65d0, 0xbe1cef3504a32189
+ .quad 0xc086244e446c1398, 0xbe1ceff54b2a406f
+ .quad 0xc0862451bc4b2eb8, 0xbe1cf0757d54ed4f
+ .quad 0xc086245532aa04f0, 0xbe1cf0c8099fdfd5
+ .quad 0xc0862458a789e250, 0xbe1cf0b173796a31
+ .quad 0xc086245c1aec1138, 0xbe1cf11d8734540d
+ .quad 0xc086245f8cd1da60, 0xbe1cf1916a723ceb
+ .quad 0xc0862462fd3c84d8, 0xbe1cf19a911e1da7
+ .quad 0xc08624666c2d5608, 0xbe1cf23a9ef72e4f
+ .quad 0xc0862469d9a591c0, 0xbe1cef503d947663
+ .quad 0xc086246d45a67a18, 0xbe1cf0fceeb1a0b2
+ .quad 0xc0862470b0314fa8, 0xbe1cf107e27e4fbc
+ .quad 0xc086247419475160, 0xbe1cf03dd9922331
+ .quad 0xc086247780e9bc98, 0xbe1cefce1a10e129
+ .quad 0xc086247ae719cd18, 0xbe1ceea47f73c4f6
+ .quad 0xc086247e4bd8bd10, 0xbe1ceec0ac56d100
+ .quad 0xc0862481af27c528, 0xbe1cee8a6593278a
+ .quad 0xc086248511081c70, 0xbe1cf2231dd9dec7
+ .quad 0xc0862488717af888, 0xbe1cf0b4b8ed7da8
+ .quad 0xc086248bd0818d68, 0xbe1cf1bd8d835002
+ .quad 0xc086248f2e1d0d98, 0xbe1cf259acc107f4
+ .quad 0xc08624928a4eaa20, 0xbe1cee897636b00c
+ .quad 0xc0862495e5179270, 0xbe1cee757f20c326
+ .quad 0xc08624993e78f490, 0xbe1cefafd3aa54a4
+ .quad 0xc086249c9673fd10, 0xbe1cee7298d38b97
+ .quad 0xc086249fed09d6f8, 0xbe1ceedc158d4ceb
+ .quad 0xc08624a3423babe0, 0xbe1cf2282987cb2e
+ .quad 0xc08624a6960aa400, 0xbe1cefe7381ecc4b
+ .quad 0xc08624a9e877e600, 0xbe1cef328dbbce80
+ .quad 0xc08624ad39849728, 0xbe1cefde45f3cc71
+ .quad 0xc08624b08931db58, 0xbe1cefa8b89433b9
+ .quad 0xc08624b3d780d500, 0xbe1cef6773c0b139
+ .quad 0xc08624b72472a528, 0xbe1cf031c931c11f
+ .quad 0xc08624ba70086b78, 0xbe1cf088f49275e7
+ .quad 0xc08624bdba434630, 0xbe1cf17de0eaa86d
+ .quad 0xc08624c103245238, 0xbe1cefd492f1ba75
+ .quad 0xc08624c44aacab08, 0xbe1cf1253e154466
+ .quad 0xc08624c790dd6ad0, 0xbe1cf0fb09ee6d55
+ .quad 0xc08624cad5b7aa58, 0xbe1cf1f08dd048fe
+ .quad 0xc08624ce193c8120, 0xbe1ceeca0809697f
+ .quad 0xc08624d15b6d0538, 0xbe1cef8d5662d968
+ .quad 0xc08624d49c4a4b78, 0xbe1cee97b556ed78
+ .quad 0xc08624d7dbd56750, 0xbe1cf1b14b6acb75
+ .quad 0xc08624db1a0f6b00, 0xbe1cef1e860623f2
+ .quad 0xc08624de56f96758, 0xbe1ceeaf4d156f3d
+ .quad 0xc08624e192946bf0, 0xbe1ceecc12b400ed
+ .quad 0xc08624e4cce18710, 0xbe1cf180c40c794f
+ .quad 0xc08624e805e1c5c8, 0xbe1cf185a08f7f65
+ .quad 0xc08624eb3d9633d8, 0xbe1cef45fc924078
+ .quad 0xc08624ee73ffdbb0, 0xbe1cf1e4f457f32a
+ .quad 0xc08624f1a91fc6a0, 0xbe1cf040147b8a5a
+ .quad 0xc08624f4dcf6fc98, 0xbe1cf1effca0dfb2
+ .quad 0xc08624f80f868468, 0xbe1cf0470146e5bc
+ .quad 0xc08624fb40cf6390, 0xbe1cef4dd186e501
+ .quad 0xc08624fe70d29e60, 0xbe1ceebe257f66c7
+ .quad 0xc08625019f9137f0, 0xbe1ceefb7a1c395c
+ .quad 0xc0862504cd0c3220, 0xbe1cf209dedfed8c
+ .quad 0xc0862507f9448db0, 0xbe1cf082da464994
+ .quad 0xc086250b243b4a18, 0xbe1cee88694a73cf
+ .quad 0xc086250e4df165a0, 0xbe1cf0b61e8f0531
+ .quad 0xc08625117667dd78, 0xbe1cf1106599c962
+ .quad 0xc08625149d9fad98, 0xbe1ceff1ee88af1f
+ .quad 0xc0862517c399d0c8, 0xbe1cf0f746994ef6
+ .quad 0xc086251ae85740b8, 0xbe1cefe8a1d077e4
+ .quad 0xc086251e0bd8f5e0, 0xbe1cf1a1da036092
+ .quad 0xc08625212e1fe7a8, 0xbe1cf0f8a7786fcd
+ .quad 0xc08625244f2d0c48, 0xbe1cefa1174a07a7
+ .quad 0xc08625276f0158d8, 0xbe1cef1043aa5b25
+ .quad 0xc086252a8d9dc150, 0xbe1cf15d521c169d
+ .quad 0xc086252dab033898, 0xbe1cf220bba8861f
+ .quad 0xc0862530c732b078, 0xbe1cef51e310eae2
+ .quad 0xc0862533e22d1988, 0xbe1cf222fcedd8ae
+ .quad 0xc0862536fbf36370, 0xbe1cefdb4da4bda8
+ .quad 0xc086253a14867ca0, 0xbe1ceeafc1112171
+ .quad 0xc086253d2be75280, 0xbe1cee99dfb4b408
+ .quad 0xc08625404216d160, 0xbe1cf22d2536f06b
+ .quad 0xc08625435715e498, 0xbe1cef6abbf2e268
+ .quad 0xc08625466ae57648, 0xbe1cf093a14789f5
+ .quad 0xc08625497d866fa0, 0xbe1cf0f93655603c
+ .quad 0xc086254c8ef9b8b8, 0xbe1cf1cc40c9aafc
+ .quad 0xc086254f9f4038a8, 0xbe1ceeea5f4e9157
+ .quad 0xc0862552ae5ad568, 0xbe1cefa9f52d4997
+ .quad 0xc0862555bc4a7400, 0xbe1cefa490a638ff
+ .quad 0xc0862558c90ff868, 0xbe1cef7fcf797d6f
+ .quad 0xc086255bd4ac4590, 0xbe1cf1b4c51113c9
+ .quad 0xc086255edf203d78, 0xbe1cef55e5b4a55d
+ .quad 0xc0862561e86cc100, 0xbe1cf0d37a25f9dc
+ .quad 0xc0862564f092b028, 0xbe1ceebe9efc19d9
+ .quad 0xc0862567f792e9d8, 0xbe1cee8ad30a57b5
+ .quad 0xc086256afd6e4c08, 0xbe1cef4e1817b90b
+ .quad 0xc086256e0225b3b8, 0xbe1cee7fa9229996
+ .quad 0xc086257105b9fce0, 0xbe1cf0b54963d945
+ .quad 0xc0862574082c0298, 0xbe1cee5f2f3c7995
+ .quad 0xc0862577097c9ee0, 0xbe1cf0828e303a2c
+ .quad 0xc086257a09acaae0, 0xbe1cf172c3078947
+ .quad 0xc086257d08bcfec0, 0xbe1cf189252afa22
+ .quad 0xc086258006ae71b8, 0xbe1cefdb80426923
+ .quad 0xc08625830381da08, 0xbe1ceef1391a0372
+ .quad 0xc0862585ff380d00, 0xbe1cf17720c78d13
+ .quad 0xc0862588f9d1df18, 0xbe1ceef1f9027d83
+ .quad 0xc086258bf35023b8, 0xbe1cf06fac99dec9
+ .quad 0xc086258eebb3ad78, 0xbe1cf1373eeb45c0
+ .quad 0xc0862591e2fd4e00, 0xbe1cef777536bb81
+ .quad 0xc0862594d92dd600, 0xbe1cf0f43ca40766
+ .quad 0xc0862597ce461558, 0xbe1cefb2cfc6766b
+ .quad 0xc086259ac246daf0, 0xbe1ceea49e64ffa2
+ .quad 0xc086259db530f4c8, 0xbe1cf250fa457dec
+ .quad 0xc08625a0a7053018, 0xbe1cf17d8bb2a44e
+ .quad 0xc08625a397c45918, 0xbe1cf1d5906d54b7
+ .quad 0xc08625a6876f3b30, 0xbe1cf08fe7b31780
+ .quad 0xc08625a97606a0e0, 0xbe1cef13edfc9d11
+ .quad 0xc08625ac638b53c8, 0xbe1cef9d2b107219
+ .quad 0xc08625af4ffe1cb0, 0xbe1cf1ddd4ff6160
+ .quad 0xc08625b23b5fc390, 0xbe1cefa02a996495
+ .quad 0xc08625b525b10f68, 0xbe1cf166a7e37ee5
+ .quad 0xc08625b80ef2c680, 0xbe1cef0b171068a5
+ .quad 0xc08625baf725ae28, 0xbe1cf05c80779283
+ .quad 0xc08625bdde4a8af0, 0xbe1cf1bbfbffb889
+ .quad 0xc08625c0c4622090, 0xbe1cf0b8666c0124
+ .quad 0xc08625c3a96d31e0, 0xbe1cf0a8fcf47a86
+ .quad 0xc08625c68d6c80f0, 0xbe1cef46e18cb092
+ .quad 0xc08625c97060cef0, 0xbe1cf1458a350efb
+ .quad 0xc08625cc524adc58, 0xbe1ceeea1dadce12
+ .quad 0xc08625cf332b68b0, 0xbe1cf0a1bfdc44c7
+ .quad 0xc08625d2130332d0, 0xbe1cef96d02da73e
+ .quad 0xc08625d4f1d2f8a8, 0xbe1cf2451c3c7701
+ .quad 0xc08625d7cf9b7778, 0xbe1cf10d08f83812
+ .quad 0xc08625daac5d6ba0, 0xbe1ceec5b4895c5e
+ .quad 0xc08625dd881990b0, 0xbe1cf14e1325c5e4
+ .quad 0xc08625e062d0a188, 0xbe1cf21d0904be12
+ .quad 0xc08625e33c835838, 0xbe1ceed0839bcf21
+ .quad 0xc08625e615326df0, 0xbe1cf1bb944889d2
+ .quad 0xc08625e8ecde9b48, 0xbe1cee738e85eece
+ .quad 0xc08625ebc38897e0, 0xbe1cf25c2bc6ef12
+ .quad 0xc08625ee99311ac8, 0xbe1cf132b70a41ad
+ .quad 0xc08625f16dd8da28, 0xbe1cf1984236a6e3
+ .quad 0xc08625f441808b78, 0xbe1cf19ae74998f9
+ .quad 0xc08625f71428e370, 0xbe1cef3e175d61a1
+ .quad 0xc08625f9e5d295f8, 0xbe1cf101f9868fd9
+ .quad 0xc08625fcb67e5658, 0xbe1cee69db83dcd2
+ .quad 0xc08625ff862cd6f8, 0xbe1cf081b636af51
+ .quad 0xc086260254dec9a8, 0xbe1cee62c7d59b3e
+ .quad 0xc08626052294df58, 0xbe1cf1b745c57716
+ .quad 0xc0862607ef4fc868, 0xbe1cef3d2800ea23
+ .quad 0xc086260abb103458, 0xbe1cef480ff1acd2
+ .quad 0xc086260d85d6d200, 0xbe1cf2424c9a17ef
+ .quad 0xc08626104fa44f90, 0xbe1cf12cfde90fd5
+ .quad 0xc086261318795a68, 0xbe1cf21f590dd5b6
+ .quad 0xc0862615e0569f48, 0xbe1cf0c50f9cd28a
+ .quad 0xc0862618a73cca30, 0xbe1ceedbdb520545
+ .quad 0xc086261b6d2c8668, 0xbe1cf0b030396011
+ .quad 0xc086261e32267e98, 0xbe1cf19917010e96
+ .quad 0xc0862620f62b5cb0, 0xbe1cf07331355985
+ .quad 0xc0862623b93bc9e8, 0xbe1cf01ae921a1c3
+ .quad 0xc08626267b586ed0, 0xbe1cefe5cf0dbf0c
+ .quad 0xc08626293c81f348, 0xbe1cf01b258aeb50
+ .quad 0xc086262bfcb8fe88, 0xbe1cee6b9e7f4c68
+ .quad 0xc086262ebbfe3710, 0xbe1cee684a9b21c9
+ .quad 0xc08626317a5242b8, 0xbe1cf1f8bcde9a8b
+ .quad 0xc086263437b5c6c0, 0xbe1cf1d063d36238
+ .quad 0xc0862636f42967a8, 0xbe1cf1e31a19075e
+ .quad 0xc0862639afadc950, 0xbe1cf1d8efdf7e7d
+ .quad 0xc086263c6a438ef0, 0xbe1cf1812ee72dba
+ .quad 0xc086263f23eb5b18, 0xbe1cf1449a9a2279
+ .quad 0xc0862641dca5cfb8, 0xbe1cee96edce5085
+ .quad 0xc086264494738e08, 0xbe1cf06797bd03b2
+ .quad 0xc08626474b5536b8, 0xbe1cef91b9b7ffc1
+ .quad 0xc086264a014b69c0, 0xbe1cef4b6721278f
+ .quad 0xc086264cb656c678, 0xbe1cf1942925eb4a
+ .quad 0xc086264f6a77eba8, 0xbe1cefa2c7bc2e39
+ .quad 0xc08626521daf7758, 0xbe1cf252595aceb3
+ .quad 0xc0862654cffe0718, 0xbe1cee8e9ae47ec2
+ .quad 0xc0862657816437a8, 0xbe1cf1bf913828fa
+ .quad 0xc086265a31e2a558, 0xbe1cf23475d6b366
+ .quad 0xc086265ce179ebc8, 0xbe1cef8df00a922b
+ .quad 0xc086265f902aa5f0, 0xbe1cef279bfa43e0
+ .quad 0xc08626623df56e38, 0xbe1cf080e10b8365
+ .quad 0xc0862664eadade70, 0xbe1cf1a518f9b544
+ .quad 0xc086266796db8fd0, 0xbe1cef9308fed9e9
+ .quad 0xc086266a41f81ae8, 0xbe1ceea3ae6b19c9
+ .quad 0xc086266cec3117b8, 0xbe1ceef06003d4c2
+ .quad 0xc086266f95871da8, 0xbe1cf0b8457ffb0c
+ .quad 0xc08626723dfac390, 0xbe1cf0c526745ad6
+ .quad 0xc0862674e58c9fa8, 0xbe1cf0cf91ff7b5d
+ .quad 0xc08626778c3d4798, 0xbe1cefe260819380
+ .quad 0xc086267a320d5070, 0xbe1ceebd90aa27a3
+ .quad 0xc086267cd6fd4ea8, 0xbe1cf0388121dffa
+ .quad 0xc086267f7b0dd630, 0xbe1cf1a3881435f1
+ .quad 0xc08626821e3f7a68, 0xbe1cef28e9d9ac52
+ .quad 0xc0862684c092ce08, 0xbe1cf02d300062dd
+ .quad 0xc086268762086350, 0xbe1cefaee1edfa35
+ .quad 0xc086268a02a0cbe0, 0xbe1cf0a5a052e936
+ .quad 0xc086268ca25c98d8, 0xbe1cee60a4a497ed
+ .quad 0xc086268f413c5ab0, 0xbe1cf0e4a5d0cf49
+ .quad 0xc0862691df40a170, 0xbe1cf149235a4e6e
+ .quad 0xc08626947c69fc80, 0xbe1cf215180b9fcc
+ .quad 0xc086269718b8fac8, 0xbe1cef9b156a9840
+ .quad 0xc0862699b42e2a90, 0xbe1cf054c91441be
+ .quad 0xc086269c4eca19a8, 0xbe1cf13ded26512c
+ .quad 0xc086269ee88d5550, 0xbe1cf22ea4d8ac06
+ .quad 0xc08626a181786a40, 0xbe1cf2354666ee2e
+ .quad 0xc08626a4198be4a8, 0xbe1cefef936752b3
+ .quad 0xc08626a6b0c85020, 0xbe1cf1e360a9db68
+ .quad 0xc08626a9472e37d8, 0xbe1ceed6aeb812c5
+ .quad 0xc08626abdcbe2650, 0xbe1cf227340b4986
+ .quad 0xc08626ae7178a5b0, 0xbe1cf0215a0cbe0d
+ .quad 0xc08626b1055e3f70, 0xbe1cf256adf0ae26
+ .quad 0xc08626b3986f7ca8, 0xbe1ceff3c67aed06
+ .quad 0xc08626b62aace5c8, 0xbe1cf2159fb93652
+ .quad 0xc08626b8bc1702e0, 0xbe1cf01e6dbd1c7f
+ .quad 0xc08626bb4cae5b60, 0xbe1cf009e75d1c0c
+ .quad 0xc08626bddc737648, 0xbe1ceec10a020e73
+ .quad 0xc08626c06b66da08, 0xbe1cf06d5783eee7
+ .quad 0xc08626c2f9890ca0, 0xbe1cf0cb8f169ffe
+ .quad 0xc08626c586da9388, 0xbe1cef7de2452430
+ .quad 0xc08626c8135bf3b0, 0xbe1cf05da6f783ae
+ .quad 0xc08626ca9f0db198, 0xbe1cefcc877d681d
+ .quad 0xc08626cd29f05138, 0xbe1cef0531954ab3
+ .quad 0xc08626cfb4045608, 0xbe1cf06b8565ea3d
+ .quad 0xc08626d23d4a4310, 0xbe1cefdc455d9d7e
+ .quad 0xc08626d4c5c29ad0, 0xbe1ceefc47e8fa64
+ .quad 0xc08626d74d6ddf48, 0xbe1cf1872bf033f2
+ .quad 0xc08626d9d44c9210, 0xbe1cf19d91087f9d
+ .quad 0xc08626dc5a5f3438, 0xbe1cf012d444c6ab
+ .quad 0xc08626dedfa64650, 0xbe1cf0ba528ee153
+ .quad 0xc08626e164224880, 0xbe1ceeb431709788
+ .quad 0xc08626e3e7d3ba60, 0xbe1cf0b9af31a6a5
+ .quad 0xc08626e66abb1b28, 0xbe1cf168fb2e135b
+ .quad 0xc08626e8ecd8e990, 0xbe1cef9097461c93
+ .quad 0xc08626eb6e2da3d0, 0xbe1cee7a434735d8
+ .quad 0xc08626edeeb9c7a8, 0xbe1cf235732b86f2
+ .quad 0xc08626f06e7dd280, 0xbe1cefe1510b89e6
+ .quad 0xc08626f2ed7a4120, 0xbe1cf1f64b9b80ef
+ .quad 0xc08626f56baf9000, 0xbe1cf08f320ca339
+ .quad 0xc08626f7e91e3b08, 0xbe1cf1b1de2808a1
+ .quad 0xc08626fa65c6bdc0, 0xbe1cf1976d778b28
+ .quad 0xc08626fce1a99338, 0xbe1ceef40a4f076f
+ .quad 0xc08626ff5cc73600, 0xbe1cef3e45869ce3
+ .quad 0xc0862701d7202048, 0xbe1ceef601b4c9d6
+ .quad 0xc086270450b4cbc0, 0xbe1cf1eaf0b57fd6
+ .quad 0xc0862706c985b1c0, 0xbe1cef82a44990f3
+ .quad 0xc086270941934b10, 0xbe1ceefe32981f2c
+ .quad 0xc086270bb8de1018, 0xbe1cefbf6f5a0445
+ .quad 0xc086270e2f6678d0, 0xbe1cf18dba75792c
+ .quad 0xc0862710a52cfcc8, 0xbe1cf0da64ce995f
+ .quad 0xc08627131a321318, 0xbe1cef04ac0fb802
+ .quad 0xc08627158e763268, 0xbe1cee9d4e2ad9bd
+ .quad 0xc086271801f9d0f8, 0xbe1cefa9b55407b5
+ .quad 0xc086271a74bd64a0, 0xbe1cefe6bd329570
+ .quad 0xc086271ce6c162c8, 0xbe1cef0b1205dc85
+ .quad 0xc086271f58064068, 0xbe1cef092a785e3f
+ .quad 0xc0862721c88c7210, 0xbe1cf050dcdaac30
+ .quad 0xc086272438546be8, 0xbe1cf210907ded8b
+ .quad 0xc0862726a75ea1b8, 0xbe1cee760be44f99
+ .quad 0xc086272915ab86c0, 0xbe1ceeeee07c2bcc
+ .quad 0xc086272b833b8df0, 0xbe1cf06874992df5
+ .quad 0xc086272df00f29d0, 0xbe1cef8fac5d4899
+ .quad 0xc08627305c26cc70, 0xbe1cf1103241cc99
+ .quad 0xc0862732c782e788, 0xbe1cf1d35fef83fe
+ .quad 0xc08627353223ec68, 0xbe1cef3ec8133e1d
+ .quad 0xc08627379c0a4be8, 0xbe1cef7261daccd8
+ .quad 0xc086273a05367688, 0xbe1cf18656c50806
+ .quad 0xc086273c6da8dc68, 0xbe1cf1c8736e049a
+ .quad 0xc086273ed561ed38, 0xbe1cf1f93bff4911
+ .quad 0xc08627413c621848, 0xbe1cf188a4ea680c
+ .quad 0xc0862743a2a9cc80, 0xbe1cf1d270930c80
+ .quad 0xc086274608397868, 0xbe1cf25a328c28e2
+ .quad 0xc08627486d118a28, 0xbe1cf106f90aa3b8
+ .quad 0xc086274ad1326f80, 0xbe1cee5e9d2e885a
+ .quad 0xc086274d349c95c0, 0xbe1cf1c0bac27228
+ .quad 0xc086274f975069f8, 0xbe1cf1a1500f9b1c
+ .quad 0xc0862751f94e58c0, 0xbe1cefc30663ac44
+ .quad 0xc08627545a96ce48, 0xbe1cf17123e427a2
+ .quad 0xc0862756bb2a3678, 0xbe1cefb92749fea4
+ .quad 0xc08627591b08fcc0, 0xbe1cefa40e1ea74a
+ .quad 0xc086275b7a338c40, 0xbe1cee6f4612c3e9
+ .quad 0xc086275dd8aa4fa8, 0xbe1cf1c54a053627
+ .quad 0xc0862760366db168, 0xbe1ceff5eb503d9e
+ .quad 0xc0862762937e1b70, 0xbe1cf02e47f10cee
+ .quad 0xc0862764efdbf768, 0xbe1ceeb06e1d0dad
+ .quad 0xc08627674b87ae88, 0xbe1cf10aadd6dba5
+ .quad 0xc0862769a681a9c0, 0xbe1cf24e9913d30f
+ .quad 0xc086276c00ca51a0, 0xbe1cef47b301e312
+ .quad 0xc086276e5a620e48, 0xbe1ceeb1cefc2e85
+ .quad 0xc0862770b3494788, 0xbe1cf16f1fbbe011
+ .quad 0xc08627730b8064e8, 0xbe1ceebdf75174c7
+ .quad 0xc08627756307cd70, 0xbe1cf06e3871a0da
+ .quad 0xc0862777b9dfe7f0, 0xbe1cef16799fd554
+ .quad 0xc086277a10091ac0, 0xbe1cf248dabf5377
+ .quad 0xc086277c6583cc00, 0xbe1cf0c78d92a2cd
+ .quad 0xc086277eba506158, 0xbe1cf0b911b029f0
+ .quad 0xc08627810e6f4028, 0xbe1cefdc24719766
+ .quad 0xc086278361e0cd70, 0xbe1cefbb6562b7e7
+ .quad 0xc0862785b4a56dd8, 0xbe1cf1e0afb349ec
+ .quad 0xc086278806bd85c0, 0xbe1cf008292e52fc
+ .quad 0xc086278a58297918, 0xbe1cf053073872bf
+ .quad 0xc086278ca8e9ab88, 0xbe1cf17a0a55a947
+ .quad 0xc086278ef8fe8068, 0xbe1ceeffb0b60234
+ .quad 0xc086279148685aa0, 0xbe1cf162204794a8
+ .quad 0xc086279397279ce0, 0xbe1cf24cc8cb48ac
+ .quad 0xc0862795e53ca978, 0xbe1cf0c9be68d5c3
+ .quad 0xc086279832a7e258, 0xbe1cf172cd3d7388
+ .quad 0xc086279a7f69a930, 0xbe1ceea2465fbce5
+ .quad 0xc086279ccb825f40, 0xbe1cf0a386d2500f
+ .quad 0xc086279f16f26590, 0xbe1cf1e338ddc18a
+ .quad 0xc08627a161ba1cd0, 0xbe1cef1f5049867f
+ .quad 0xc08627a3abd9e548, 0xbe1cef96c1ea8b1f
+ .quad 0xc08627a5f5521f00, 0xbe1cf138f6fd3c26
+ .quad 0xc08627a83e2329b0, 0xbe1cf0d4fcbfdf3a
+ .quad 0xc08627aa864d64b0, 0xbe1cf24870c12c81
+ .quad 0xc08627accdd12f18, 0xbe1cf0ae2a56348d
+ .quad 0xc08627af14aee7a0, 0xbe1cee8ca1a9b893
+ .quad 0xc08627b15ae6eca8, 0xbe1cf20414d637b0
+ .quad 0xc08627b3a0799c60, 0xbe1cf0fc6b7b12d8
+ .quad 0xc08627b5e5675488, 0xbe1cf152d93c4a00
+ .quad 0xc08627b829b072a0, 0xbe1cf1073f9b77c2
+ .quad 0xc08627ba6d5553d8, 0xbe1cee694f97d5a4
+ .quad 0xc08627bcb0565500, 0xbe1cf0456b8239d7
+ .quad 0xc08627bef2b3d2b0, 0xbe1cf211497127e3
+ .quad 0xc08627c1346e2930, 0xbe1cf01856c0384d
+ .quad 0xc08627c37585b468, 0xbe1cefa7dd05479e
+ .quad 0xc08627c5b5fad000, 0xbe1cef3ae8e50b93
+ .quad 0xc08627c7f5cdd750, 0xbe1ceea5f32fdd3a
+ .quad 0xc08627ca34ff2560, 0xbe1cef424caeb8d9
+ .quad 0xc08627cc738f14f0, 0xbe1cf0194d07a81f
+ .quad 0xc08627ceb17e0070, 0xbe1cf20f452000c1
+ .quad 0xc08627d0eecc4210, 0xbe1cf00e356218e4
+ .quad 0xc08627d32b7a33a0, 0xbe1cef30484b4bcb
+ .quad 0xc08627d567882eb0, 0xbe1ceeea11a6641b
+ .quad 0xc08627d7a2f68c80, 0xbe1cf13492d5bd7b
+ .quad 0xc08627d9ddc5a618, 0xbe1ceeb7048fad96
+ .quad 0xc08627dc17f5d418, 0xbe1ceef0666f0477
+ .quad 0xc08627de51876ee8, 0xbe1cf060d4b8b5c2
+ .quad 0xc08627e08a7acea8, 0xbe1cf0b2a4b6ff8c
+ .quad 0xc08627e2c2d04b28, 0xbe1cf0e34809a875
+ .quad 0xc08627e4fa883bf0, 0xbe1cf16bf74a3522
+ .quad 0xc08627e731a2f848, 0xbe1cee6a24623d57
+ .quad 0xc08627e96820d718, 0xbe1cefc7b4f1528e
+ .quad 0xc08627eb9e022f18, 0xbe1cf163051f3548
+ .quad 0xc08627edd34756b8, 0xbe1cef36b3366305
+ .quad 0xc08627f007f0a408, 0xbe1cf18134625550
+ .quad 0xc08627f23bfe6cf0, 0xbe1cf0ec32ec1a11
+ .quad 0xc08627f46f710700, 0xbe1ceeb3b64f3edc
+ .quad 0xc08627f6a248c778, 0xbe1cf0cd15805bc8
+ .quad 0xc08627f8d4860368, 0xbe1cf20db3bddebe
+ .quad 0xc08627fb06290f90, 0xbe1cf25188430e25
+ .quad 0xc08627fd37324070, 0xbe1ceea1713490f9
+ .quad 0xc08627ff67a1ea28, 0xbe1cf159521d234c
+ .quad 0xc0862801977860b8, 0xbe1cf24dfe50783b
+ .quad 0xc0862803c6b5f7d0, 0xbe1ceef2ef89a60b
+ .quad 0xc0862805f55b02c8, 0xbe1cee7fc919d62c
+ .quad 0xc08628082367d4c0, 0xbe1cf215a7fb513a
+ .quad 0xc086280a50dcc0a8, 0xbe1cf0e4401c5ed4
+ .quad 0xc086280c7dba1910, 0xbe1cf04ec734d256
+ .quad 0xc086280eaa003050, 0xbe1cf010ad787fea
+ .quad 0xc0862810d5af5880, 0xbe1cee622478393d
+ .quad 0xc086281300c7e368, 0xbe1cf01c7482564f
+ .quad 0xc08628152b4a22a0, 0xbe1cf0de20d33536
+ .quad 0xc086281755366778, 0xbe1cef2edae5837d
+ .quad 0xc08628197e8d02f0, 0xbe1cf0a345318cc9
+ .quad 0xc086281ba74e45d8, 0xbe1cf20085aa34b8
+ .quad 0xc086281dcf7a80c0, 0xbe1cef5fa845ad83
+ .quad 0xc086281ff71203e0, 0xbe1cf050d1df69c4
+ .quad 0xc08628221e151f48, 0xbe1ceffe43c035b9
+ .quad 0xc0862824448422b8, 0xbe1cf14f3018d3c2
+ .quad 0xc08628266a5f5dc0, 0xbe1cef0a5fbae83d
+ .quad 0xc08628288fa71f98, 0xbe1ceff8a95b72a1
+ .quad 0xc086282ab45bb750, 0xbe1cef073aa9849b
+ .quad 0xc086282cd87d73a8, 0xbe1cef69b3835c02
+ .quad 0xc086282efc0ca328, 0xbe1cf0bc139379a9
+ .quad 0xc08628311f099420, 0xbe1cef247a9ec596
+ .quad 0xc086283341749490, 0xbe1cef74bbcc488a
+ .quad 0xc0862835634df248, 0xbe1cef4bc42e7b8e
+ .quad 0xc08628378495fad0, 0xbe1cf136d4d5a810
+ .quad 0xc0862839a54cfb80, 0xbe1cf0d290b24dd8
+ .quad 0xc086283bc5734168, 0xbe1ceeebde8e0065
+ .quad 0xc086283de5091950, 0xbe1cf1a09f60aa1e
+ .quad 0xc0862840040ecfe0, 0xbe1cf0803947a234
+ .quad 0xc08628422284b168, 0xbe1cf0abf7638127
+ .quad 0xc0862844406b0a08, 0xbe1cf0f73ee12058
+ .quad 0xc08628465dc225a0, 0xbe1cf2079971b26c
+ .quad 0xc08628487a8a4fe0, 0xbe1cee74957564b1
+ .quad 0xc086284a96c3d420, 0xbe1ceee77c1b7d43
+ .quad 0xc086284cb26efd90, 0xbe1cf23addba6e09
+ .quad 0xc086284ecd8c1730, 0xbe1cf199f4a1da60
+ .quad 0xc0862850e81b6bb0, 0xbe1cf09fdea81393
+ .quad 0xc0862853021d4588, 0xbe1cf176adb417f7
+ .quad 0xc08628551b91ef00, 0xbe1cf0f64f84a8da
+ .quad 0xc08628573479b220, 0xbe1ceec34cf49523
+ .quad 0xc08628594cd4d8a8, 0xbe1cf16d60fbe0bb
+ .quad 0xc086285b64a3ac40, 0xbe1cee8de7acfc7b
+ .quad 0xc086285d7be67630, 0xbe1ceee6256cce8d
+ .quad 0xc086285f929d7fa0, 0xbe1cee7d66a3d8a5
+ .quad 0xc0862861a8c91170, 0xbe1cf0bef8265792
+ .quad 0xc0862863be697458, 0xbe1cf097f890c6f8
+ .quad 0xc0862865d37ef0c8, 0xbe1cf09502d5c3fc
+ .quad 0xc0862867e809cf00, 0xbe1ceeffb239dac7
+ .quad 0xc0862869fc0a56f8, 0xbe1cf1fbfff95c98
+ .quad 0xc086286c0f80d090, 0xbe1cefa57ad3eef7
+ .quad 0xc086286e226d8348, 0xbe1cf22c58b9183d
+ .quad 0xc086287034d0b690, 0xbe1ceff262d0a248
+ .quad 0xc086287246aab180, 0xbe1cefa7bc194186
+ .quad 0xc086287457fbbb08, 0xbe1cf06782d784d9
+ .quad 0xc086287668c419e0, 0xbe1cf1d44d0eaa07
+ .quad 0xc086287879041490, 0xbe1cf034803c8a48
+ .quad 0xc086287a88bbf158, 0xbe1cf08e84916b6f
+ .quad 0xc086287c97ebf650, 0xbe1cf0c4d3dc1bc7
+ .quad 0xc086287ea6946958, 0xbe1cefb1e4625943
+ .quad 0xc0862880b4b59010, 0xbe1cf143efdd1fd0
+ .quad 0xc0862882c24faff8, 0xbe1cee9896d016da
+ .quad 0xc0862884cf630e38, 0xbe1cf2186072f2cc
+ .quad 0xc0862886dbefeff0, 0xbe1cef9217633d34
+ .quad 0xc0862888e7f699e0, 0xbe1cf05603549486
+ .quad 0xc086288af37750b0, 0xbe1cef50fff513d3
+ .quad 0xc086288cfe7258c0, 0xbe1cf127713b32d0
+ .quad 0xc086288f08e7f650, 0xbe1cf05015520f3d
+ .quad 0xc086289112d86d58, 0xbe1cf12eb458b26f
+ .quad 0xc08628931c4401a8, 0xbe1cf22eae2887ed
+ .quad 0xc0862895252af6e0, 0xbe1cefdd6656dd2d
+ .quad 0xc08628972d8d9058, 0xbe1cf1048ea4e646
+ .quad 0xc0862899356c1150, 0xbe1ceec4501167e9
+ .quad 0xc086289b3cc6bcb8, 0xbe1cf0ad52becc3f
+ .quad 0xc086289d439dd568, 0xbe1cf0daa4e00e35
+ .quad 0xc086289f49f19df8, 0xbe1cf00b80de8d6a
+ .quad 0xc08628a14fc258c8, 0xbe1cf1bcf2ea8464
+ .quad 0xc08628a355104818, 0xbe1cf0435e2782b0
+ .quad 0xc08628a559dbade0, 0xbe1cf0e3e1a5f56c
+ .quad 0xc08628a75e24cbf8, 0xbe1cefed9d5a721d
+ .quad 0xc08628a961ebe3f8, 0xbe1cf0d2d74321e2
+ .quad 0xc08628ab65313750, 0xbe1cf24200eb55e9
+ .quad 0xc08628ad67f50740, 0xbe1cf23e9d7cf979
+ .quad 0xc08628af6a3794d0, 0xbe1cf23a088f421c
+ .quad 0xc08628b16bf920e0, 0xbe1cef2c1de1ab32
+ .quad 0xc08628b36d39ec08, 0xbe1cf1abc231f7b2
+ .quad 0xc08628b56dfa36d0, 0xbe1cf2074d5ba303
+ .quad 0xc08628b76e3a4180, 0xbe1cf05cd5eed880
+ /*== Log_LA_table ==*/
+ .align 16
+ .quad 0x8000000000000000
+ .quad 0xbf5ff802a9ab10e6
+ .quad 0xbf6ff00aa2b10bc0
+ .quad 0xbf77ee11ebd82e94
+ .quad 0xbf7fe02a6b106789
+ .quad 0xbf83e7295d25a7d9
+ .quad 0xbf87dc475f810a77
+ .quad 0xbf8bcf712c74384c
+ .quad 0xbf8fc0a8b0fc03e4
+ .quad 0xbf91d7f7eb9eebe7
+ .quad 0xbf93cea44346a575
+ .quad 0xbf95c45a51b8d389
+ .quad 0xbf97b91b07d5b11b
+ .quad 0xbf99ace7551cc514
+ .quad 0xbf9b9fc027af9198
+ .quad 0xbf9d91a66c543cc4
+ .quad 0xbf9f829b0e783300
+ .quad 0xbfa0b94f7c196176
+ .quad 0xbfa1b0d98923d980
+ .quad 0xbfa2a7ec2214e873
+ .quad 0xbfa39e87b9febd60
+ .quad 0xbfa494acc34d911c
+ .quad 0xbfa58a5bafc8e4d5
+ .quad 0xbfa67f94f094bd98
+ .quad 0xbfa77458f632dcfc
+ .quad 0xbfa868a83083f6cf
+ .quad 0xbfa95c830ec8e3eb
+ .quad 0xbfaa4fe9ffa3d235
+ .quad 0xbfab42dd711971bf
+ .quad 0xbfac355dd0921f2d
+ .quad 0xbfad276b8adb0b52
+ .quad 0xbfae19070c276016
+ .quad 0xbfaf0a30c01162a6
+ .quad 0xbfaffae9119b9303
+ .quad 0xbfb075983598e471
+ .quad 0xbfb0ed839b5526fe
+ .quad 0xbfb16536eea37ae1
+ .quad 0xbfb1dcb263db1944
+ .quad 0xbfb253f62f0a1417
+ .quad 0xbfb2cb0283f5de1f
+ .quad 0xbfb341d7961bd1d1
+ .quad 0xbfb3b87598b1b6ee
+ .quad 0xbfb42edcbea646f0
+ .quad 0xbfb4a50d3aa1b040
+ .quad 0xbfb51b073f06183f
+ .quad 0xbfb590cafdf01c28
+ .quad 0xbfb60658a93750c4
+ .quad 0xbfb67bb0726ec0fc
+ .quad 0xbfb6f0d28ae56b4c
+ .quad 0xbfb765bf23a6be13
+ .quad 0xbfb7da766d7b12cd
+ .quad 0xbfb84ef898e8282a
+ .quad 0xbfb8c345d6319b21
+ .quad 0xbfb9375e55595ede
+ .quad 0xbfb9ab42462033ad
+ .quad 0xbfba1ef1d8061cd4
+ .quad 0xbfba926d3a4ad563
+ .quad 0xbfbb05b49bee43fe
+ .quad 0xbfbb78c82bb0eda1
+ .quad 0xbfbbeba818146765
+ .quad 0xbfbc5e548f5bc743
+ .quad 0xbfbcd0cdbf8c13e1
+ .quad 0xbfbd4313d66cb35d
+ .quad 0xbfbdb5270187d927
+ .quad 0xbfbe27076e2af2e6
+ .quad 0xbfbe98b549671467
+ .quad 0xbfbf0a30c01162a6
+ .quad 0xbfbf7b79fec37ddf
+ .quad 0xbfbfec9131dbeabb
+ .quad 0xbfc02ebb42bf3d4b
+ .quad 0xbfc0671512ca596e
+ .quad 0xbfc09f561ee719c3
+ .quad 0xbfc0d77e7cd08e59
+ .quad 0xbfc10f8e422539b1
+ .quad 0xbfc14785846742ac
+ .quad 0xbfc17f6458fca611
+ .quad 0xbfc1b72ad52f67a0
+ .quad 0xbfc1eed90e2dc2c3
+ .quad 0xbfc2266f190a5acb
+ .quad 0xbfc25ded0abc6ad2
+ .quad 0xbfc29552f81ff523
+ .quad 0xbfc2cca0f5f5f251
+ .quad 0xbfc303d718e47fd3
+ .quad 0xbfc33af575770e4f
+ .quad 0xbfc371fc201e8f74
+ .quad 0xbfc3a8eb2d31a376
+ .quad 0xbfc3dfc2b0ecc62a
+ .quad 0xbfc41682bf727bc0
+ .quad 0xbfc44d2b6ccb7d1e
+ .quad 0xbfc483bccce6e3dd
+ .quad 0xbfc4ba36f39a55e5
+ .quad 0xbfc4f099f4a230b2
+ .quad 0xbfc526e5e3a1b438
+ .quad 0xbfc55d1ad4232d6f
+ .quad 0xbfc59338d9982086
+ .quad 0xbfc5c940075972b9
+ .quad 0xbfc5ff3070a793d4
+ .quad 0xbfc6350a28aaa758
+ .quad 0xbfc66acd4272ad51
+ .quad 0xbfc6a079d0f7aad2
+ .quad 0xbfc6d60fe719d21d
+ .quad 0xbfc70b8f97a1aa75
+ .quad 0xbfc740f8f54037a5
+ .quad 0xbfc7764c128f2127
+ .quad 0xbfc7ab890210d909
+ .quad 0xbfc7e0afd630c274
+ .quad 0xbfc815c0a14357eb
+ .quad 0xbfc84abb75865139
+ .quad 0xbfc87fa06520c911
+ .quad 0xbfc8b46f8223625b
+ .quad 0xbfc8e928de886d41
+ .quad 0xbfc91dcc8c340bde
+ .quad 0xbfc9525a9cf456b4
+ .quad 0xbfc986d3228180ca
+ .quad 0xbfc9bb362e7dfb83
+ .quad 0xbfc9ef83d2769a34
+ .quad 0xbfca23bc1fe2b563
+ .quad 0xbfca57df28244dcd
+ .quad 0xbfca8becfc882f19
+ .quad 0xbfcabfe5ae46124c
+ .quad 0xbfcaf3c94e80bff3
+ .quad 0xbfcb2797ee46320c
+ .quad 0xbfcb5b519e8fb5a4
+ .quad 0xbfcb8ef670420c3b
+ .quad 0xbfcbc286742d8cd6
+ .quad 0xbfcbf601bb0e44e2
+ .quad 0xbfcc2968558c18c1
+ .quad 0xbfcc5cba543ae425
+ .quad 0xbfcc8ff7c79a9a22
+ .quad 0xbfccc320c0176502
+ .quad 0xbfccf6354e09c5dc
+ .quad 0xbfcd293581b6b3e7
+ .quad 0xbfcd5c216b4fbb91
+ .quad 0xbfcd8ef91af31d5e
+ .quad 0xbfcdc1bca0abec7d
+ .quad 0xbfcdf46c0c722d2f
+ .quad 0xbfce27076e2af2e6
+ .quad 0xbfce598ed5a87e2f
+ .quad 0xbfce8c0252aa5a60
+ .quad 0xbfcebe61f4dd7b0b
+ .quad 0xbfcef0adcbdc5936
+ .quad 0xbfcf22e5e72f105d
+ .quad 0xbfcf550a564b7b37
+ .quad 0xbfcf871b28955045
+ .quad 0xbfcfb9186d5e3e2b
+ .quad 0xbfcfeb0233e607cc
+ .quad 0xbfd00e6c45ad501d
+ .quad 0xbfd0274dc16c232f
+ .quad 0xbfd0402594b4d041
+ .quad 0xbfd058f3c703ebc6
+ .quad 0xbfd071b85fcd590d
+ .quad 0xbfd08a73667c57af
+ .quad 0xbfd0a324e27390e3
+ .quad 0xbfd0bbccdb0d24bd
+ .quad 0xbfd0d46b579ab74b
+ .quad 0xbfd0ed005f657da4
+ .quad 0xbfd1058bf9ae4ad5
+ .quad 0xbfd11e0e2dad9cb7
+ .quad 0xbfd136870293a8b0
+ .quad 0xbfd14ef67f88685a
+ .quad 0xbfd1675cababa60e
+ .quad 0xbfd17fb98e15095d
+ .quad 0xbfd1980d2dd4236f
+ .quad 0xbfd1b05791f07b49
+ .quad 0xbfd1c898c16999fb
+ .quad 0xbfd1e0d0c33716be
+ .quad 0xbfd1f8ff9e48a2f3
+ .quad 0xbfd211255986160c
+ .quad 0xbfd22941fbcf7966
+ .quad 0xbfd241558bfd1404
+ .quad 0xbfd2596010df763a
+ .quad 0xbfd27161913f853d
+ .quad 0xbfd2895a13de86a3
+ .quad 0xbfd2a1499f762bc9
+ .quad 0xbfd2b9303ab89d25
+ .quad 0xbfd2d10dec508583
+ .quad 0xbfd2e8e2bae11d31
+ .quad 0xbfd300aead06350c
+ .quad 0xbfd31871c9544185
+ .quad 0xbfd3302c16586588
+ .quad 0xbfd347dd9a987d55
+ .quad 0xbfd35f865c93293e
+ .quad 0xbfd3772662bfd85b
+ .quad 0xbfd38ebdb38ed321
+ .quad 0xbfd3a64c556945ea
+ .quad 0xbfd3bdd24eb14b6a
+ .quad 0xbfd3d54fa5c1f710
+ .quad 0xbfd3ecc460ef5f50
+ .quad 0xbfd404308686a7e4
+ .quad 0xbfd41b941cce0bee
+ .quad 0xbfd432ef2a04e814
+ .quad 0xbfd44a41b463c47c
+ .quad 0xbfd4618bc21c5ec2
+ .quad 0xbfd478cd5959b3d9
+ .quad 0xbfd49006804009d1
+ .quad 0xbfd4a7373cecf997
+ .quad 0xbfd4be5f957778a1
+ .quad 0xbfd4d57f8fefe27f
+ .quad 0xbfd4ec973260026a
+ .quad 0xbfd503a682cb1cb3
+ .quad 0xbfd51aad872df82d
+ .quad 0xbfd531ac457ee77e
+ .quad 0xbfd548a2c3add263
+ .quad 0xbfd55f9107a43ee2
+ .quad 0xbfd5767717455a6c
+ .quad 0xbfd58d54f86e02f2
+ .quad 0xbfd5a42ab0f4cfe2
+ .quad 0xbfd5baf846aa1b19
+ .quad 0xbfd5d1bdbf5809ca
+ .quad 0xbfd5e87b20c2954a
+ .quad 0xbfd5ff3070a793d4
+ .quad 0xbfd615ddb4bec13c
+ .quad 0xbfd62c82f2b9c795
+ .quad 0x3fd61965cdb02c1f
+ .quad 0x3fd602d08af091ec
+ .quad 0x3fd5ec433d5c35ae
+ .quad 0x3fd5d5bddf595f30
+ .quad 0x3fd5bf406b543db2
+ .quad 0x3fd5a8cadbbedfa1
+ .quad 0x3fd5925d2b112a59
+ .quad 0x3fd57bf753c8d1fb
+ .quad 0x3fd565995069514c
+ .quad 0x3fd54f431b7be1a9
+ .quad 0x3fd538f4af8f72fe
+ .quad 0x3fd522ae0738a3d8
+ .quad 0x3fd50c6f1d11b97c
+ .quad 0x3fd4f637ebba9810
+ .quad 0x3fd4e0086dd8baca
+ .quad 0x3fd4c9e09e172c3c
+ .quad 0x3fd4b3c077267e9a
+ .quad 0x3fd49da7f3bcc41f
+ .quad 0x3fd487970e958770
+ .quad 0x3fd4718dc271c41b
+ .quad 0x3fd45b8c0a17df13
+ .quad 0x3fd44591e0539f49
+ .quad 0x3fd42f9f3ff62642
+ .quad 0x3fd419b423d5e8c7
+ .quad 0x3fd403d086cea79c
+ .quad 0x3fd3edf463c1683e
+ .quad 0x3fd3d81fb5946dba
+ .quad 0x3fd3c25277333184
+ .quad 0x3fd3ac8ca38e5c5f
+ .quad 0x3fd396ce359bbf54
+ .quad 0x3fd3811728564cb2
+ .quad 0x3fd36b6776be1117
+ .quad 0x3fd355bf1bd82c8b
+ .quad 0x3fd3401e12aecba1
+ .quad 0x3fd32a84565120a8
+ .quad 0x3fd314f1e1d35ce4
+ .quad 0x3fd2ff66b04ea9d4
+ .quad 0x3fd2e9e2bce12286
+ .quad 0x3fd2d46602adccee
+ .quad 0x3fd2bef07cdc9354
+ .quad 0x3fd2a982269a3dbf
+ .quad 0x3fd2941afb186b7c
+ .quad 0x3fd27ebaf58d8c9d
+ .quad 0x3fd269621134db92
+ .quad 0x3fd25410494e56c7
+ .quad 0x3fd23ec5991eba49
+ .quad 0x3fd22981fbef797b
+ .quad 0x3fd214456d0eb8d4
+ .quad 0x3fd1ff0fe7cf47a7
+ .quad 0x3fd1e9e1678899f4
+ .quad 0x3fd1d4b9e796c245
+ .quad 0x3fd1bf99635a6b95
+ .quad 0x3fd1aa7fd638d33f
+ .quad 0x3fd1956d3b9bc2fa
+ .quad 0x3fd180618ef18adf
+ .quad 0x3fd16b5ccbacfb73
+ .quad 0x3fd1565eed455fc3
+ .quad 0x3fd14167ef367783
+ .quad 0x3fd12c77cd00713b
+ .quad 0x3fd1178e8227e47c
+ .quad 0x3fd102ac0a35cc1c
+ .quad 0x3fd0edd060b78081
+ .quad 0x3fd0d8fb813eb1ef
+ .quad 0x3fd0c42d676162e3
+ .quad 0x3fd0af660eb9e279
+ .quad 0x3fd09aa572e6c6d4
+ .quad 0x3fd085eb8f8ae797
+ .quad 0x3fd07138604d5862
+ .quad 0x3fd05c8be0d9635a
+ .quad 0x3fd047e60cde83b8
+ .quad 0x3fd03346e0106062
+ .quad 0x3fd01eae5626c691
+ .quad 0x3fd00a1c6adda473
+ .quad 0x3fcfeb2233ea07cd
+ .quad 0x3fcfc218be620a5e
+ .quad 0x3fcf991c6cb3b379
+ .quad 0x3fcf702d36777df0
+ .quad 0x3fcf474b134df229
+ .quad 0x3fcf1e75fadf9bde
+ .quad 0x3fcef5ade4dcffe6
+ .quad 0x3fceccf2c8fe920a
+ .quad 0x3fcea4449f04aaf5
+ .quad 0x3fce7ba35eb77e2a
+ .quad 0x3fce530effe71012
+ .quad 0x3fce2a877a6b2c12
+ .quad 0x3fce020cc6235ab5
+ .quad 0x3fcdd99edaf6d7e9
+ .quad 0x3fcdb13db0d48940
+ .quad 0x3fcd88e93fb2f450
+ .quad 0x3fcd60a17f903515
+ .quad 0x3fcd38666871f465
+ .quad 0x3fcd1037f2655e7b
+ .quad 0x3fcce816157f1988
+ .quad 0x3fccc000c9db3c52
+ .quad 0x3fcc97f8079d44ec
+ .quad 0x3fcc6ffbc6f00f71
+ .quad 0x3fcc480c0005ccd1
+ .quad 0x3fcc2028ab17f9b4
+ .quad 0x3fcbf851c067555f
+ .quad 0x3fcbd087383bd8ad
+ .quad 0x3fcba8c90ae4ad19
+ .quad 0x3fcb811730b823d2
+ .quad 0x3fcb5971a213acdb
+ .quad 0x3fcb31d8575bce3d
+ .quad 0x3fcb0a4b48fc1b46
+ .quad 0x3fcae2ca6f672bd4
+ .quad 0x3fcabb55c31693ad
+ .quad 0x3fca93ed3c8ad9e3
+ .quad 0x3fca6c90d44b704e
+ .quad 0x3fca454082e6ab05
+ .quad 0x3fca1dfc40f1b7f1
+ .quad 0x3fc9f6c407089664
+ .quad 0x3fc9cf97cdce0ec3
+ .quad 0x3fc9a8778debaa38
+ .quad 0x3fc981634011aa75
+ .quad 0x3fc95a5adcf7017f
+ .quad 0x3fc9335e5d594989
+ .quad 0x3fc90c6db9fcbcd9
+ .quad 0x3fc8e588ebac2dbf
+ .quad 0x3fc8beafeb38fe8c
+ .quad 0x3fc897e2b17b19a5
+ .quad 0x3fc871213750e994
+ .quad 0x3fc84a6b759f512f
+ .quad 0x3fc823c16551a3c2
+ .quad 0x3fc7fd22ff599d4f
+ .quad 0x3fc7d6903caf5ad0
+ .quad 0x3fc7b0091651528c
+ .quad 0x3fc7898d85444c73
+ .quad 0x3fc7631d82935a86
+ .quad 0x3fc73cb9074fd14d
+ .quad 0x3fc716600c914054
+ .quad 0x3fc6f0128b756abc
+ .quad 0x3fc6c9d07d203fc7
+ .quad 0x3fc6a399dabbd383
+ .quad 0x3fc67d6e9d785771
+ .quad 0x3fc6574ebe8c133a
+ .quad 0x3fc6313a37335d76
+ .quad 0x3fc60b3100b09476
+ .quad 0x3fc5e533144c1719
+ .quad 0x3fc5bf406b543db2
+ .quad 0x3fc59958ff1d52f1
+ .quad 0x3fc5737cc9018cdd
+ .quad 0x3fc54dabc26105d2
+ .quad 0x3fc527e5e4a1b58d
+ .quad 0x3fc5022b292f6a45
+ .quad 0x3fc4dc7b897bc1c8
+ .quad 0x3fc4b6d6fefe22a4
+ .quad 0x3fc4913d8333b561
+ .quad 0x3fc46baf0f9f5db7
+ .quad 0x3fc4462b9dc9b3dc
+ .quad 0x3fc420b32740fdd4
+ .quad 0x3fc3fb45a59928cc
+ .quad 0x3fc3d5e3126bc27f
+ .quad 0x3fc3b08b6757f2a9
+ .quad 0x3fc38b3e9e027479
+ .quad 0x3fc365fcb0159016
+ .quad 0x3fc340c59741142e
+ .quad 0x3fc31b994d3a4f85
+ .quad 0x3fc2f677cbbc0a96
+ .quad 0x3fc2d1610c86813a
+ .quad 0x3fc2ac55095f5c59
+ .quad 0x3fc28753bc11aba5
+ .quad 0x3fc2625d1e6ddf57
+ .quad 0x3fc23d712a49c202
+ .quad 0x3fc2188fd9807263
+ .quad 0x3fc1f3b925f25d41
+ .quad 0x3fc1ceed09853752
+ .quad 0x3fc1aa2b7e23f72a
+ .quad 0x3fc185747dbecf34
+ .quad 0x3fc160c8024b27b1
+ .quad 0x3fc13c2605c398c3
+ .quad 0x3fc1178e8227e47c
+ .quad 0x3fc0f301717cf0fb
+ .quad 0x3fc0ce7ecdccc28d
+ .quad 0x3fc0aa06912675d5
+ .quad 0x3fc08598b59e3a07
+ .quad 0x3fc06135354d4b18
+ .quad 0x3fc03cdc0a51ec0d
+ .quad 0x3fc0188d2ecf6140
+ .quad 0x3fbfe89139dbd566
+ .quad 0x3fbfa01c9db57ce2
+ .quad 0x3fbf57bc7d9005db
+ .quad 0x3fbf0f70cdd992e3
+ .quad 0x3fbec739830a1120
+ .quad 0x3fbe7f1691a32d3e
+ .quad 0x3fbe3707ee30487b
+ .quad 0x3fbdef0d8d466db9
+ .quad 0x3fbda727638446a2
+ .quad 0x3fbd5f55659210e2
+ .quad 0x3fbd179788219364
+ .quad 0x3fbccfedbfee13a8
+ .quad 0x3fbc885801bc4b23
+ .quad 0x3fbc40d6425a5cb1
+ .quad 0x3fbbf968769fca11
+ .quad 0x3fbbb20e936d6974
+ .quad 0x3fbb6ac88dad5b1c
+ .quad 0x3fbb23965a52ff00
+ .quad 0x3fbadc77ee5aea8c
+ .quad 0x3fba956d3ecade63
+ .quad 0x3fba4e7640b1bc38
+ .quad 0x3fba0792e9277cac
+ .quad 0x3fb9c0c32d4d2548
+ .quad 0x3fb97a07024cbe74
+ .quad 0x3fb9335e5d594989
+ .quad 0x3fb8ecc933aeb6e8
+ .quad 0x3fb8a6477a91dc29
+ .quad 0x3fb85fd927506a48
+ .quad 0x3fb8197e2f40e3f0
+ .quad 0x3fb7d33687c293c9
+ .quad 0x3fb78d02263d82d3
+ .quad 0x3fb746e100226ed9
+ .quad 0x3fb700d30aeac0e1
+ .quad 0x3fb6bad83c1883b6
+ .quad 0x3fb674f089365a7a
+ .quad 0x3fb62f1be7d77743
+ .quad 0x3fb5e95a4d9791cb
+ .quad 0x3fb5a3abb01ade25
+ .quad 0x3fb55e10050e0384
+ .quad 0x3fb518874226130a
+ .quad 0x3fb4d3115d207eac
+ .quad 0x3fb48dae4bc31018
+ .quad 0x3fb4485e03dbdfad
+ .quad 0x3fb403207b414b7f
+ .quad 0x3fb3bdf5a7d1ee64
+ .quad 0x3fb378dd7f749714
+ .quad 0x3fb333d7f8183f4b
+ .quad 0x3fb2eee507b40301
+ .quad 0x3fb2aa04a44717a5
+ .quad 0x3fb26536c3d8c369
+ .quad 0x3fb2207b5c78549e
+ .quad 0x3fb1dbd2643d190b
+ .quad 0x3fb1973bd1465567
+ .quad 0x3fb152b799bb3cc9
+ .quad 0x3fb10e45b3cae831
+ .quad 0x3fb0c9e615ac4e17
+ .quad 0x3fb08598b59e3a07
+ .quad 0x3fb0415d89e74444
+ .quad 0x3faffa6911ab9301
+ .quad 0x3faf723b517fc523
+ .quad 0x3faeea31c006b87c
+ .quad 0x3fae624c4a0b5e1b
+ .quad 0x3fadda8adc67ee4e
+ .quad 0x3fad52ed6405d86f
+ .quad 0x3faccb73cdddb2cc
+ .quad 0x3fac441e06f72a9e
+ .quad 0x3fabbcebfc68f420
+ .quad 0x3fab35dd9b58baad
+ .quad 0x3faaaef2d0fb10fc
+ .quad 0x3faa282b8a936171
+ .quad 0x3fa9a187b573de7c
+ .quad 0x3fa91b073efd7314
+ .quad 0x3fa894aa149fb343
+ .quad 0x3fa80e7023d8ccc4
+ .quad 0x3fa788595a3577ba
+ .quad 0x3fa70265a550e777
+ .quad 0x3fa67c94f2d4bb58
+ .quad 0x3fa5f6e73078efb8
+ .quad 0x3fa5715c4c03ceef
+ .quad 0x3fa4ebf43349e26f
+ .quad 0x3fa466aed42de3ea
+ .quad 0x3fa3e18c1ca0ae92
+ .quad 0x3fa35c8bfaa1306b
+ .quad 0x3fa2d7ae5c3c5bae
+ .quad 0x3fa252f32f8d183f
+ .quad 0x3fa1ce5a62bc353a
+ .quad 0x3fa149e3e4005a8d
+ .quad 0x3fa0c58fa19dfaaa
+ .quad 0x3fa0415d89e74444
+ .quad 0x3f9f7a9b16782856
+ .quad 0x3f9e72bf2813ce51
+ .quad 0x3f9d6b2725979802
+ .quad 0x3f9c63d2ec14aaf2
+ .quad 0x3f9b5cc258b718e6
+ .quad 0x3f9a55f548c5c43f
+ .quad 0x3f994f6b99a24475
+ .quad 0x3f98492528c8cabf
+ .quad 0x3f974321d3d006d3
+ .quad 0x3f963d6178690bd6
+ .quad 0x3f9537e3f45f3565
+ .quad 0x3f9432a925980cc1
+ .quad 0x3f932db0ea132e22
+ .quad 0x3f9228fb1fea2e28
+ .quad 0x3f912487a5507f70
+ .quad 0x3f90205658935847
+ .quad 0x3f8e38ce3033310c
+ .quad 0x3f8c317384c75f06
+ .quad 0x3f8a2a9c6c170462
+ .quad 0x3f882448a388a2aa
+ .quad 0x3f861e77e8b53fc6
+ .quad 0x3f841929f96832f0
+ .quad 0x3f82145e939ef1e9
+ .quad 0x3f8010157588de71
+ .quad 0x3f7c189cbb0e27fb
+ .quad 0x3f78121214586b54
+ .quad 0x3f740c8a747878e2
+ .quad 0x3f70080559588b35
+ .quad 0x3f680904828985c0
+ .quad 0x3f60040155d5889e
+ .quad 0x3f50020055655889
+ .quad 0x0000000000000000
+ /*== poly_coeff[4] ==*/
+ .align 16
+ .quad 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A /* coeff4 */
+ .quad 0xbfd0000148058EE1, 0xbfd0000148058EE1 /* coeff3 */
+ .quad 0x3fd55555555543C5, 0x3fd55555555543C5 /* coeff2 */
+ .quad 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F /* coeff1 */
+ /*== ExpMask ==*/
+ .align 16
+ .quad 0x000fffffffffffff, 0x000fffffffffffff
+ /*== Two10 ==*/
+ .align 16
+ .quad 0x3f50000000000000, 0x3f50000000000000
+ /*== MinLog1p = -1+2^(-53) ==*/
+ .align 16
+ .quad 0xbfefffffffffffff, 0xbfefffffffffffff
+ /*== MaxLog1p ==*/
+ .align 16
+ .quad 0x7f3ffffffffff000, 0x7f3ffffffffff000
+ /*== One ==*/
+ .align 16
+ .quad 0x3ff0000000000000, 0x3ff0000000000000
+ /*== SgnMask ==*/
+ .align 16
+ .quad 0x7fffffffffffffff, 0x7fffffffffffffff
+ /*== XThreshold ==*/
+ .align 16
+ .quad 0x3e00000000000000, 0x3e00000000000000
+ /*== XhMask ==*/
+ .align 16
+ .quad 0xfffffffffffffc00, 0xfffffffffffffc00
+ /*== Threshold ==*/
+ .align 16
+ .quad 0x4086a00000000000, 0x4086a00000000000
+ /*== Bias ==*/
+ .align 16
+ .quad 0x408ff80000000000, 0x408ff80000000000
+ /*== Bias1 ==*/
+ .align 16
+ .quad 0x408ff00000000000, 0x408ff00000000000
+ /*== ExpMask ==*/
+ .align 16
+ .quad 0x7ff0000000000000, 0x7ff0000000000000
+ /*== ExpMask2 ==*/
+ .align 16
+ .quad 0x7f40000000000000, 0x7f40000000000000
+ /*== L2L ==*/
+ .align 16
+ .quad 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF
+ /*== dBigThreshold ==*/
+ .align 16
+ .quad 0x41D0000000000000, 0x41D0000000000000
+ /*== dC2 ==*/
+ .align 16
+ .quad 0x3FD8000000000000, 0x3FD8000000000000
+ /*== dC3 ==*/
+ .align 16
+ .quad 0x3FD4000000000000, 0x3FD4000000000000
+ /*== dC4 ==*/
+ .align 16
+ .quad 0x3FD1800000000000, 0x3FD1800000000000
+ /*== dC5 ==*/
+ .align 16
+ .quad 0x3FCF800000000000, 0x3FCF800000000000
+ /*== dHalf ==*/
+ .align 16
+ .quad 0x3FE0000000000000, 0x3FE0000000000000
+ /*== dLargestFinite ==*/
+ .align 16
+ .quad 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF
+ /*== dLittleThreshold ==*/
+ .align 16
+ .quad 0x3F60000000000000, 0x3F60000000000000
+ /*== dSign ==*/
+ .align 16
+ .quad 0x8000000000000000, 0x8000000000000000
+ /*== dThirtyOne ==*/
+ .align 16
+ .quad 0x403F000000000000, 0x403F000000000000
+ /*== dTopMask12 ==*/
+ .align 16
+ .quad 0xFFFFFE0000000000, 0xFFFFFE0000000000
+ /*== dTopMask26 ==*/
+ .align 16
+ .quad 0xFFFFFFFFF8000000, 0xFFFFFFFFF8000000
+ /*== dTopMask29 ==*/
+ .align 16
+ .quad 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000
+ /*== XScale ==*/
+ .align 16
+ .quad 0x3E10000000000000, 0x3E10000000000000
+ .align 16
+ .type __svml_dasinh_data_internal,@object
+ .size __svml_dasinh_data_internal,.-__svml_dasinh_data_internal
+ .align 16
+
+.FLT_30:
+ .long 0x00000000,0x43380000,0x00000000,0x43380000
+ .type .FLT_30,@object
+ .size .FLT_30,16
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S
new file mode 100644
index 0000000..903b5f0
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S
@@ -0,0 +1,20 @@
+/* SSE version of vectorized asinh, vector length is 4.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define _ZGVdN4v_asinh _ZGVdN4v_asinh_sse_wrapper
+#include "../svml_d_asinh4_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c
new file mode 100644
index 0000000..e7acd03
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c
@@ -0,0 +1,27 @@
+/* Multiple versions of vectorized asinh, vector length is 4.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define SYMBOL_NAME _ZGVdN4v_asinh
+#include "ifunc-mathvec-avx2.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVdN4v_asinh, __GI__ZGVdN4v_asinh, __redirect__ZGVdN4v_asinh)
+ __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S
new file mode 100644
index 0000000..d691d1e
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S
@@ -0,0 +1,1601 @@
+/* Function asinh vectorized with AVX2.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ https://www.gnu.org/licenses/. */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ * Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *
+ * Special cases:
+ *
+ * asinh(NaN) = quiet NaN, and raise invalid exception
+ * asinh(INF) = that INF
+ * asinh(0) = that 0
+ *
+ */
+
+/* Offsets for data table __svml_dasinh_data_internal
+ */
+#define Log_HA_table 0
+#define Log_LA_table 8224
+#define poly_coeff 12352
+#define ExpMask 12480
+#define Two10 12512
+#define MinLog1p 12544
+#define MaxLog1p 12576
+#define One 12608
+#define SgnMask 12640
+#define XThreshold 12672
+#define XhMask 12704
+#define Threshold 12736
+#define Bias 12768
+#define Bias1 12800
+#define ExpMask0 12832
+#define ExpMask2 12864
+#define L2 12896
+#define dBigThreshold 12928
+#define dC2 12960
+#define dC3 12992
+#define dC4 13024
+#define dC5 13056
+#define dHalf 13088
+#define dLargestFinite 13120
+#define dLittleThreshold 13152
+#define dSign 13184
+#define dThirtyOne 13216
+#define dTopMask12 13248
+#define dTopMask29 13280
+#define XScale 13312
+
+/* Lookup bias for data table __svml_dasinh_data_internal. */
+#define Table_Lookup_Bias -0x405fe0
+
+#include <sysdep.h>
+
+ .text
+ .section .text.avx2,"ax",@progbits
+ENTRY(_ZGVdN4v_asinh_avx2)
+ pushq %rbp
+ cfi_def_cfa_offset(16)
+ movq %rsp, %rbp
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+ andq $-32, %rsp
+ subq $96, %rsp
+ lea Table_Lookup_Bias+__svml_dasinh_data_internal(%rip), %r8
+ vmovapd %ymm0, %ymm13
+ vmovupd SgnMask+__svml_dasinh_data_internal(%rip), %ymm9
+
+/* Load the constant 1 and a sign mask */
+ vmovupd One+__svml_dasinh_data_internal(%rip), %ymm12
+
+/* No need to split X when FMA is available in hardware. */
+ vmulpd %ymm13, %ymm13, %ymm8
+
+/*
+ * Get the absolute value of the input, since we will exploit antisymmetry
+ * and mostly assume X >= 0 in the core computation
+ */
+ vandpd %ymm9, %ymm13, %ymm10
+
+/*
+ * Check whether the input is finite, by checking |X| <= MaxFloat
+ * Otherwise set the rangemask so that the callout will get used.
+ * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
+ */
+ vcmpnle_uqpd dLargestFinite+__svml_dasinh_data_internal(%rip), %ymm10, %ymm14
+
+/*
+ * Finally, express Y + W = X^2 + 1 accurately where Y has <= 29 bits.
+ * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
+ * as the dominant component in the compensated summation. Otherwise,
+ * if |X| >= 1, then since X2Hi only has 52 significant bits, the basic
+ * addition will be exact anyway until we get to |X| >= 2^53. But by
+ * that time the log function is well-conditioned enough that the
+ * rounding error doesn't matter. Hence we can treat 1 as dominant even
+ * if it literally isn't.
+ */
+ vaddpd %ymm8, %ymm12, %ymm5
+
+/*
+ * The following computation can go wrong for very large X, basically
+ * because X^2 overflows. But for large X we have
+ * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
+ * we can just later stick X back into the log and tweak up the exponent.
+ * Actually we scale X by 2^-30 and tweak the exponent up by 31,
+ * to stay in the safe range for the later log computation.
+ * Compute a flag now telling us when do do this.
+ */
+ vcmplt_oqpd dBigThreshold+__svml_dasinh_data_internal(%rip), %ymm10, %ymm11
+ vsubpd %ymm5, %ymm12, %ymm15
+ vmovmskpd %ymm14, %eax
+ vandpd dTopMask29+__svml_dasinh_data_internal(%rip), %ymm5, %ymm14
+
+/*
+ * Compute R = 1/sqrt(Y + W) * (1 + d)
+ * Force R to <= 12 significant bits in case it isn't already
+ * This means that R * Y and R^2 * Y are exactly representable.
+ */
+ vcvtpd2ps %ymm14, %xmm1
+ vaddpd %ymm15, %ymm8, %ymm0
+ vsubpd %ymm14, %ymm5, %ymm2
+ vrsqrtps %xmm1, %xmm3
+ vmovapd %ymm13, %ymm7
+ vfmsub213pd %ymm8, %ymm13, %ymm7
+ vcvtps2pd %xmm3, %ymm6
+ vaddpd %ymm0, %ymm7, %ymm4
+
+/*
+ * Unfortunately, we can still be in trouble if |X| <= 2^-10, since
+ * the absolute error 2^-(12+53)-ish in sqrt(1 + X^2) gets scaled up
+ * by 1/X and comes close to our threshold. Hence if |X| <= 2^-9,
+ * perform an alternative computation
+ * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
+ * X2 = X^2
+ */
+ vaddpd %ymm7, %ymm8, %ymm7
+ vaddpd %ymm2, %ymm4, %ymm15
+
+/*
+ * Now 1 / (1 + d)
+ * = 1 / (1 + (sqrt(1 - e) - 1))
+ * = 1 / sqrt(1 - e)
+ * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 +
+ * 63/256 * e^5 + 231/1024 * e^6 + ....
+ * So compute the first five nonconstant terms of that, so that
+ * we have a relative correction (1 + Corr) to apply to S etc.
+ * C1 = 1/2
+ * C2 = 3/8
+ * C3 = 5/16
+ * C4 = 35/128
+ * C5 = 63/256
+ */
+ vmovupd dC5+__svml_dasinh_data_internal(%rip), %ymm4
+ vandpd dTopMask12+__svml_dasinh_data_internal(%rip), %ymm6, %ymm0
+
+/*
+ * Compute S = (Y/sqrt(Y + W)) * (1 + d)
+ * and T = (W/sqrt(Y + W)) * (1 + d)
+ * so that S + T = sqrt(Y + W) * (1 + d)
+ * S is exact, and the rounding error in T is OK.
+ */
+ vmulpd %ymm0, %ymm14, %ymm3
+ vmulpd %ymm15, %ymm0, %ymm1
+ vmovupd dHalf+__svml_dasinh_data_internal(%rip), %ymm6
+ vsubpd %ymm12, %ymm3, %ymm14
+
+/*
+ * Obtain sqrt(1 + X^2) - 1 in two pieces
+ * sqrt(1 + X^2) - 1
+ * = sqrt(Y + W) - 1
+ * = (S + T) * (1 + Corr) - 1
+ * = [S - 1] + [T + (S + T) * Corr]
+ * We need a compensated summation for the last part. We treat S - 1
+ * as the larger part; it certainly is until about X < 2^-4, and in that
+ * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
+ * Final sum is dTmp5 (hi) + dTmp7 (lo)
+ */
+ vaddpd %ymm1, %ymm3, %ymm2
+
+/*
+ * Compute e = -(2 * d + d^2)
+ * The first FMR is exact, and the rounding error in the other is acceptable
+ * since d and e are ~ 2^-12
+ */
+ vmovapd %ymm12, %ymm5
+ vfnmadd231pd %ymm3, %ymm0, %ymm5
+ vfnmadd231pd %ymm1, %ymm0, %ymm5
+ vfmadd213pd dC4+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
+ vfmadd213pd dC3+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
+ vfmadd213pd dC2+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
+ vfmadd213pd %ymm6, %ymm5, %ymm4
+ vmulpd %ymm4, %ymm5, %ymm0
+ vfmadd213pd %ymm1, %ymm2, %ymm0
+
+/* Now multiplex the two possible computations */
+ vcmple_oqpd dLittleThreshold+__svml_dasinh_data_internal(%rip), %ymm10, %ymm2
+ vaddpd %ymm14, %ymm0, %ymm15
+
+/* dX2over2 = X^2/2 */
+ vmulpd %ymm7, %ymm6, %ymm0
+
+/* dX4over4 = X^4/4 */
+ vmulpd %ymm0, %ymm0, %ymm8
+
+/* dX46 = -X^4/4 + X^6/8 */
+ vfmsub231pd %ymm0, %ymm8, %ymm8
+
+/* dX46over2 = -X^4/8 + x^6/16 */
+ vmulpd %ymm8, %ymm6, %ymm5
+
+/* 2^ (-10-exp(X) ) */
+ vmovupd ExpMask2+__svml_dasinh_data_internal(%rip), %ymm8
+ vaddpd %ymm5, %ymm0, %ymm4
+ vblendvpd %ymm2, %ymm4, %ymm15, %ymm1
+
+/*
+ * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
+ * It's always safe to assume |X| is larger.
+ * This is the final 2-part argument to the log1p function
+ */
+ vaddpd %ymm1, %ymm10, %ymm3
+
+/* Now multiplex to the case X = 2^-30 * |input|, Xl = dL = 0 in the "big" case. */
+ vmulpd XScale+__svml_dasinh_data_internal(%rip), %ymm10, %ymm10
+
+/*
+ * Now we feed into the log1p code, using H in place of _VARG1 and
+ * also adding L into Xl.
+ * compute 1+x as high, low parts
+ */
+ vmaxpd %ymm3, %ymm12, %ymm6
+ vminpd %ymm3, %ymm12, %ymm7
+ vandpd %ymm9, %ymm3, %ymm9
+ vcmplt_oqpd XThreshold+__svml_dasinh_data_internal(%rip), %ymm9, %ymm0
+ vaddpd %ymm7, %ymm6, %ymm5
+ vorpd XhMask+__svml_dasinh_data_internal(%rip), %ymm0, %ymm4
+ vandpd %ymm4, %ymm5, %ymm1
+ vblendvpd %ymm11, %ymm1, %ymm10, %ymm5
+ vsubpd %ymm1, %ymm6, %ymm2
+
+/* exponent bits */
+ vpsrlq $20, %ymm5, %ymm10
+ vaddpd %ymm2, %ymm7, %ymm3
+
+/*
+ * Now resume the main code.
+ * preserve mantissa, set input exponent to 2^(-10)
+ */
+ vandpd ExpMask+__svml_dasinh_data_internal(%rip), %ymm5, %ymm0
+ vorpd Two10+__svml_dasinh_data_internal(%rip), %ymm0, %ymm2
+
+/* reciprocal approximation good to at least 11 bits */
+ vcvtpd2ps %ymm2, %xmm6
+ vrcpps %xmm6, %xmm7
+ vcvtps2pd %xmm7, %ymm15
+
+/* exponent of X needed to scale Xl */
+ vandps ExpMask0+__svml_dasinh_data_internal(%rip), %ymm5, %ymm9
+ vpsubq %ymm9, %ymm8, %ymm0
+ vandpd %ymm11, %ymm3, %ymm4
+
+/* round reciprocal to nearest integer, will have 1+9 mantissa bits */
+ vroundpd $0, %ymm15, %ymm3
+
+/* scale DblRcp */
+ vmulpd %ymm0, %ymm3, %ymm2
+
+/* argument reduction */
+ vfmsub213pd %ymm12, %ymm2, %ymm5
+ vmulpd %ymm2, %ymm4, %ymm12
+ vmovupd poly_coeff+64+__svml_dasinh_data_internal(%rip), %ymm2
+ vaddpd %ymm12, %ymm5, %ymm5
+ vfmadd213pd poly_coeff+96+__svml_dasinh_data_internal(%rip), %ymm5, %ymm2
+ vmulpd %ymm5, %ymm5, %ymm4
+ vextractf128 $1, %ymm10, %xmm14
+ vshufps $221, %xmm14, %xmm10, %xmm1
+
+/* biased exponent in DP format */
+ vcvtdq2pd %xmm1, %ymm7
+
+/* exponent*log(2.0) */
+ vmovupd Threshold+__svml_dasinh_data_internal(%rip), %ymm10
+
+/* Add 31 to the exponent in the "large" case to get log(2 * input) */
+ vaddpd dThirtyOne+__svml_dasinh_data_internal(%rip), %ymm7, %ymm6
+ vblendvpd %ymm11, %ymm7, %ymm6, %ymm1
+
+/*
+ * prepare table index
+ * table lookup
+ */
+ vpsrlq $40, %ymm3, %ymm11
+ vcmplt_oqpd %ymm3, %ymm10, %ymm3
+ vandpd Bias+__svml_dasinh_data_internal(%rip), %ymm3, %ymm14
+ vorpd Bias1+__svml_dasinh_data_internal(%rip), %ymm14, %ymm15
+ vsubpd %ymm15, %ymm1, %ymm1
+ vmulpd L2+__svml_dasinh_data_internal(%rip), %ymm1, %ymm3
+
+/* polynomial */
+ vmovupd poly_coeff+__svml_dasinh_data_internal(%rip), %ymm1
+ vfmadd213pd poly_coeff+32+__svml_dasinh_data_internal(%rip), %ymm5, %ymm1
+ vfmadd213pd %ymm2, %ymm4, %ymm1
+
+/* reconstruction */
+ vfmadd213pd %ymm5, %ymm4, %ymm1
+ vextractf128 $1, %ymm11, %xmm7
+ vmovd %xmm11, %edx
+ vmovd %xmm7, %esi
+ movslq %edx, %rdx
+ vpextrd $2, %xmm11, %ecx
+ movslq %esi, %rsi
+ vpextrd $2, %xmm7, %edi
+ movslq %ecx, %rcx
+ movslq %edi, %rdi
+ vmovsd (%r8,%rdx), %xmm0
+ vmovsd (%r8,%rsi), %xmm8
+ vmovhpd (%r8,%rcx), %xmm0, %xmm6
+ vmovhpd (%r8,%rdi), %xmm8, %xmm9
+ vinsertf128 $1, %xmm9, %ymm6, %ymm0
+ vaddpd %ymm1, %ymm0, %ymm0
+ vaddpd %ymm0, %ymm3, %ymm7
+
+/* Finally, reincorporate the original sign. */
+ vandpd dSign+__svml_dasinh_data_internal(%rip), %ymm13, %ymm6
+ vxorpd %ymm7, %ymm6, %ymm0
+ testl %eax, %eax
+
+/* Go to special inputs processing branch */
+ jne L(SPECIAL_VALUES_BRANCH)
+ # LOE rbx r12 r13 r14 r15 eax ymm0 ymm13
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+ movq %rbp, %rsp
+ popq %rbp
+ cfi_def_cfa(7, 8)
+ cfi_restore(6)
+ ret
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+ vmovupd %ymm13, 32(%rsp)
+ vmovupd %ymm0, 64(%rsp)
+ # LOE rbx r12 r13 r14 r15 eax ymm0
+
+ xorl %edx, %edx
+ # LOE rbx r12 r13 r14 r15 eax edx
+
+ vzeroupper
+ movq %r12, 16(%rsp)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
+ movl %edx, %r12d
+ movq %r13, 8(%rsp)
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
+ movl %eax, %r13d
+ movq %r14, (%rsp)
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+ btl %r12d, %r13d
+
+/* Call scalar math function */
+ jc L(SCALAR_MATH_CALL)
+ # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+ incl %r12d
+ cmpl $4, %r12d
+
+/* Check bits in range mask */
+ jl L(RANGEMASK_CHECK)
+ # LOE rbx r15 r12d r13d
+
+ movq 16(%rsp), %r12
+ cfi_restore(12)
+ movq 8(%rsp), %r13
+ cfi_restore(13)
+ movq (%rsp), %r14
+ cfi_restore(14)
+ vmovupd 64(%rsp), %ymm0
+
+/* Go to exit */
+ jmp L(EXIT)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r12 r13 r14 r15 ymm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+ movl %r12d, %r14d
+ movsd 32(%rsp,%r14,8), %xmm0
+ call asinh@PLT
+ # LOE rbx r14 r15 r12d r13d xmm0
+
+ movsd %xmm0, 64(%rsp,%r14,8)
+
+/* Process special inputs in loop */
+ jmp L(SPECIAL_VALUES_LOOP)
+ # LOE rbx r15 r12d r13d
+END(_ZGVdN4v_asinh_avx2)
+
+ .section .rodata, "a"
+ .align 32
+
+#ifdef __svml_dasinh_data_internal_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+ __declspec(align(32)) VUINT32 Log_HA_table[(1<<10)+2][2];
+ __declspec(align(32)) VUINT32 Log_LA_table[(1<<9)+1][2];
+ __declspec(align(32)) VUINT32 poly_coeff[4][4][2];
+ __declspec(align(32)) VUINT32 ExpMask[4][2];
+ __declspec(align(32)) VUINT32 Two10[4][2];
+ __declspec(align(32)) VUINT32 MinLog1p[4][2];
+ __declspec(align(32)) VUINT32 MaxLog1p[4][2];
+ __declspec(align(32)) VUINT32 One[4][2];
+ __declspec(align(32)) VUINT32 SgnMask[4][2];
+ __declspec(align(32)) VUINT32 XThreshold[4][2];
+ __declspec(align(32)) VUINT32 XhMask[4][2];
+ __declspec(align(32)) VUINT32 Threshold[4][2];
+ __declspec(align(32)) VUINT32 Bias[4][2];
+ __declspec(align(32)) VUINT32 Bias1[4][2];
+ __declspec(align(32)) VUINT32 ExpMask0[4][2];
+ __declspec(align(32)) VUINT32 ExpMask2[4][2];
+ __declspec(align(32)) VUINT32 L2[4][2];
+ __declspec(align(32)) VUINT32 dBigThreshold[4][2];
+ __declspec(align(32)) VUINT32 dC2[4][2];
+ __declspec(align(32)) VUINT32 dC3[4][2];
+ __declspec(align(32)) VUINT32 dC4[4][2];
+ __declspec(align(32)) VUINT32 dC5[4][2];
+ __declspec(align(32)) VUINT32 dHalf[4][2];
+ __declspec(align(32)) VUINT32 dLargestFinite[4][2];
+ __declspec(align(32)) VUINT32 dLittleThreshold[4][2];
+ __declspec(align(32)) VUINT32 dSign[4][2];
+ __declspec(align(32)) VUINT32 dThirtyOne[4][2];
+ __declspec(align(32)) VUINT32 dTopMask12[4][2];
+ __declspec(align(32)) VUINT32 dTopMask29[4][2];
+ __declspec(align(32)) VUINT32 XScale[4][2];
+} __svml_dasinh_data_internal;
+#endif
+__svml_dasinh_data_internal:
+ /* Log_HA_table */
+ .quad 0xc086232bdd7a8300, 0xbe1ce91eef3fb100
+ .quad 0xc086232fdc7ad828, 0xbe1cefcffda73b6a
+ .quad 0xc0862333d97d2ba0, 0xbe1cef406748f1ff
+ .quad 0xc0862337d48378e0, 0xbe1cef2a9429925a
+ .quad 0xc086233bcd8fb878, 0xbe1cf138d17ebecb
+ .quad 0xc086233fc4a3e018, 0xbe1ceff2dbbbb29e
+ .quad 0xc0862343b9c1e270, 0xbe1cf1a42aae437b
+ .quad 0xc0862347acebaf68, 0xbe1cef3b152048af
+ .quad 0xc086234b9e2333f0, 0xbe1cef20e127805e
+ .quad 0xc086234f8d6a5a30, 0xbe1cf00ad6052cf4
+ .quad 0xc08623537ac30980, 0xbe1cefc4642ee597
+ .quad 0xc0862357662f2660, 0xbe1cf1f277d36e16
+ .quad 0xc086235b4fb092a0, 0xbe1ceed009e8d8e6
+ .quad 0xc086235f37492d28, 0xbe1cf1e4038cb362
+ .quad 0xc08623631cfad250, 0xbe1cf0b0873b8557
+ .quad 0xc086236700c75b98, 0xbe1cf15bb3227c0b
+ .quad 0xc086236ae2b09fe0, 0xbe1cf151ef8ca9ed
+ .quad 0xc086236ec2b87358, 0xbe1cefe1dc2cd2ed
+ .quad 0xc0862372a0e0a780, 0xbe1cf0d1eec5454f
+ .quad 0xc08623767d2b0b48, 0xbe1ceeefd570bbce
+ .quad 0xc086237a57996af0, 0xbe1cee99ae91b3a7
+ .quad 0xc086237e302d9028, 0xbe1cf0412830fbd1
+ .quad 0xc086238206e94218, 0xbe1ceee898588610
+ .quad 0xc0862385dbce4548, 0xbe1cee9a1fbcaaea
+ .quad 0xc0862389aede5bc0, 0xbe1ceed8e7cc1ad6
+ .quad 0xc086238d801b4500, 0xbe1cf10c8d059da6
+ .quad 0xc08623914f86be18, 0xbe1ceee6c63a8165
+ .quad 0xc08623951d228180, 0xbe1cf0c3592d2ff1
+ .quad 0xc0862398e8f04758, 0xbe1cf0026cc4cb1b
+ .quad 0xc086239cb2f1c538, 0xbe1cf15d48d8e670
+ .quad 0xc08623a07b28ae60, 0xbe1cef359363787c
+ .quad 0xc08623a44196b390, 0xbe1cefdf1ab2e82c
+ .quad 0xc08623a8063d8338, 0xbe1cefe43c02aa84
+ .quad 0xc08623abc91ec960, 0xbe1cf044f5ae35b7
+ .quad 0xc08623af8a3c2fb8, 0xbe1cf0b0b4001e1b
+ .quad 0xc08623b349975d98, 0xbe1cf1bae76dfbcf
+ .quad 0xc08623b70731f810, 0xbe1cef0a72e13a62
+ .quad 0xc08623bac30da1c8, 0xbe1cf184007d2b6b
+ .quad 0xc08623be7d2bfb40, 0xbe1cf16f4b239e98
+ .quad 0xc08623c2358ea2a0, 0xbe1cf0976acada87
+ .quad 0xc08623c5ec3733d0, 0xbe1cf066318a16ff
+ .quad 0xc08623c9a1274880, 0xbe1ceffaa7148798
+ .quad 0xc08623cd54607820, 0xbe1cf23ab02e9b6e
+ .quad 0xc08623d105e45800, 0xbe1cefdfef7d4fde
+ .quad 0xc08623d4b5b47b20, 0xbe1cf17fece44f2b
+ .quad 0xc08623d863d27270, 0xbe1cf18f907d0d7c
+ .quad 0xc08623dc103fccb0, 0xbe1cee61fe072c98
+ .quad 0xc08623dfbafe1668, 0xbe1cf022dd891e2f
+ .quad 0xc08623e3640eda20, 0xbe1ceecc1daf4358
+ .quad 0xc08623e70b73a028, 0xbe1cf0173c4fa380
+ .quad 0xc08623eab12deec8, 0xbe1cf16a2150c2f4
+ .quad 0xc08623ee553f4a30, 0xbe1cf1bf980b1f4b
+ .quad 0xc08623f1f7a93480, 0xbe1cef8b731663c2
+ .quad 0xc08623f5986d2dc0, 0xbe1cee9a664d7ef4
+ .quad 0xc08623f9378cb3f0, 0xbe1cf1eda2af6400
+ .quad 0xc08623fcd5094320, 0xbe1cf1923f9d68d7
+ .quad 0xc086240070e45548, 0xbe1cf0747cd3e03a
+ .quad 0xc08624040b1f6260, 0xbe1cf22ee855bd6d
+ .quad 0xc0862407a3bbe078, 0xbe1cf0d57360c00b
+ .quad 0xc086240b3abb4398, 0xbe1ceebc815cd575
+ .quad 0xc086240ed01efdd0, 0xbe1cf03bfb970951
+ .quad 0xc086241263e87f50, 0xbe1cf16e74768529
+ .quad 0xc0862415f6193658, 0xbe1cefec64b8becb
+ .quad 0xc086241986b28f30, 0xbe1cf0838d210baa
+ .quad 0xc086241d15b5f448, 0xbe1cf0ea86e75b11
+ .quad 0xc0862420a324ce28, 0xbe1cf1708d11d805
+ .quad 0xc08624242f008380, 0xbe1ceea988c5a417
+ .quad 0xc0862427b94a7910, 0xbe1cef166a7bbca5
+ .quad 0xc086242b420411d0, 0xbe1cf0c9d9e86a38
+ .quad 0xc086242ec92eaee8, 0xbe1cef0946455411
+ .quad 0xc08624324ecbaf98, 0xbe1cefea60907739
+ .quad 0xc0862435d2dc7160, 0xbe1cf1ed0934ce42
+ .quad 0xc086243955624ff8, 0xbe1cf191ba746c7d
+ .quad 0xc086243cd65ea548, 0xbe1ceeec78cf2a7e
+ .quad 0xc086244055d2c968, 0xbe1cef345284c119
+ .quad 0xc0862443d3c012b8, 0xbe1cf24f77355219
+ .quad 0xc08624475027d5e8, 0xbe1cf05bf087e114
+ .quad 0xc086244acb0b65d0, 0xbe1cef3504a32189
+ .quad 0xc086244e446c1398, 0xbe1ceff54b2a406f
+ .quad 0xc0862451bc4b2eb8, 0xbe1cf0757d54ed4f
+ .quad 0xc086245532aa04f0, 0xbe1cf0c8099fdfd5
+ .quad 0xc0862458a789e250, 0xbe1cf0b173796a31
+ .quad 0xc086245c1aec1138, 0xbe1cf11d8734540d
+ .quad 0xc086245f8cd1da60, 0xbe1cf1916a723ceb
+ .quad 0xc0862462fd3c84d8, 0xbe1cf19a911e1da7
+ .quad 0xc08624666c2d5608, 0xbe1cf23a9ef72e4f
+ .quad 0xc0862469d9a591c0, 0xbe1cef503d947663
+ .quad 0xc086246d45a67a18, 0xbe1cf0fceeb1a0b2
+ .quad 0xc0862470b0314fa8, 0xbe1cf107e27e4fbc
+ .quad 0xc086247419475160, 0xbe1cf03dd9922331
+ .quad 0xc086247780e9bc98, 0xbe1cefce1a10e129
+ .quad 0xc086247ae719cd18, 0xbe1ceea47f73c4f6
+ .quad 0xc086247e4bd8bd10, 0xbe1ceec0ac56d100
+ .quad 0xc0862481af27c528, 0xbe1cee8a6593278a
+ .quad 0xc086248511081c70, 0xbe1cf2231dd9dec7
+ .quad 0xc0862488717af888, 0xbe1cf0b4b8ed7da8
+ .quad 0xc086248bd0818d68, 0xbe1cf1bd8d835002
+ .quad 0xc086248f2e1d0d98, 0xbe1cf259acc107f4
+ .quad 0xc08624928a4eaa20, 0xbe1cee897636b00c
+ .quad 0xc0862495e5179270, 0xbe1cee757f20c326
+ .quad 0xc08624993e78f490, 0xbe1cefafd3aa54a4
+ .quad 0xc086249c9673fd10, 0xbe1cee7298d38b97
+ .quad 0xc086249fed09d6f8, 0xbe1ceedc158d4ceb
+ .quad 0xc08624a3423babe0, 0xbe1cf2282987cb2e
+ .quad 0xc08624a6960aa400, 0xbe1cefe7381ecc4b
+ .quad 0xc08624a9e877e600, 0xbe1cef328dbbce80
+ .quad 0xc08624ad39849728, 0xbe1cefde45f3cc71
+ .quad 0xc08624b08931db58, 0xbe1cefa8b89433b9
+ .quad 0xc08624b3d780d500, 0xbe1cef6773c0b139
+ .quad 0xc08624b72472a528, 0xbe1cf031c931c11f
+ .quad 0xc08624ba70086b78, 0xbe1cf088f49275e7
+ .quad 0xc08624bdba434630, 0xbe1cf17de0eaa86d
+ .quad 0xc08624c103245238, 0xbe1cefd492f1ba75
+ .quad 0xc08624c44aacab08, 0xbe1cf1253e154466
+ .quad 0xc08624c790dd6ad0, 0xbe1cf0fb09ee6d55
+ .quad 0xc08624cad5b7aa58, 0xbe1cf1f08dd048fe
+ .quad 0xc08624ce193c8120, 0xbe1ceeca0809697f
+ .quad 0xc08624d15b6d0538, 0xbe1cef8d5662d968
+ .quad 0xc08624d49c4a4b78, 0xbe1cee97b556ed78
+ .quad 0xc08624d7dbd56750, 0xbe1cf1b14b6acb75
+ .quad 0xc08624db1a0f6b00, 0xbe1cef1e860623f2
+ .quad 0xc08624de56f96758, 0xbe1ceeaf4d156f3d
+ .quad 0xc08624e192946bf0, 0xbe1ceecc12b400ed
+ .quad 0xc08624e4cce18710, 0xbe1cf180c40c794f
+ .quad 0xc08624e805e1c5c8, 0xbe1cf185a08f7f65
+ .quad 0xc08624eb3d9633d8, 0xbe1cef45fc924078
+ .quad 0xc08624ee73ffdbb0, 0xbe1cf1e4f457f32a
+ .quad 0xc08624f1a91fc6a0, 0xbe1cf040147b8a5a
+ .quad 0xc08624f4dcf6fc98, 0xbe1cf1effca0dfb2
+ .quad 0xc08624f80f868468, 0xbe1cf0470146e5bc
+ .quad 0xc08624fb40cf6390, 0xbe1cef4dd186e501
+ .quad 0xc08624fe70d29e60, 0xbe1ceebe257f66c7
+ .quad 0xc08625019f9137f0, 0xbe1ceefb7a1c395c
+ .quad 0xc0862504cd0c3220, 0xbe1cf209dedfed8c
+ .quad 0xc0862507f9448db0, 0xbe1cf082da464994
+ .quad 0xc086250b243b4a18, 0xbe1cee88694a73cf
+ .quad 0xc086250e4df165a0, 0xbe1cf0b61e8f0531
+ .quad 0xc08625117667dd78, 0xbe1cf1106599c962
+ .quad 0xc08625149d9fad98, 0xbe1ceff1ee88af1f
+ .quad 0xc0862517c399d0c8, 0xbe1cf0f746994ef6
+ .quad 0xc086251ae85740b8, 0xbe1cefe8a1d077e4
+ .quad 0xc086251e0bd8f5e0, 0xbe1cf1a1da036092
+ .quad 0xc08625212e1fe7a8, 0xbe1cf0f8a7786fcd
+ .quad 0xc08625244f2d0c48, 0xbe1cefa1174a07a7
+ .quad 0xc08625276f0158d8, 0xbe1cef1043aa5b25
+ .quad 0xc086252a8d9dc150, 0xbe1cf15d521c169d
+ .quad 0xc086252dab033898, 0xbe1cf220bba8861f
+ .quad 0xc0862530c732b078, 0xbe1cef51e310eae2
+ .quad 0xc0862533e22d1988, 0xbe1cf222fcedd8ae
+ .quad 0xc0862536fbf36370, 0xbe1cefdb4da4bda8
+ .quad 0xc086253a14867ca0, 0xbe1ceeafc1112171
+ .quad 0xc086253d2be75280, 0xbe1cee99dfb4b408
+ .quad 0xc08625404216d160, 0xbe1cf22d2536f06b
+ .quad 0xc08625435715e498, 0xbe1cef6abbf2e268
+ .quad 0xc08625466ae57648, 0xbe1cf093a14789f5
+ .quad 0xc08625497d866fa0, 0xbe1cf0f93655603c
+ .quad 0xc086254c8ef9b8b8, 0xbe1cf1cc40c9aafc
+ .quad 0xc086254f9f4038a8, 0xbe1ceeea5f4e9157
+ .quad 0xc0862552ae5ad568, 0xbe1cefa9f52d4997
+ .quad 0xc0862555bc4a7400, 0xbe1cefa490a638ff
+ .quad 0xc0862558c90ff868, 0xbe1cef7fcf797d6f
+ .quad 0xc086255bd4ac4590, 0xbe1cf1b4c51113c9
+ .quad 0xc086255edf203d78, 0xbe1cef55e5b4a55d
+ .quad 0xc0862561e86cc100, 0xbe1cf0d37a25f9dc
+ .quad 0xc0862564f092b028, 0xbe1ceebe9efc19d9
+ .quad 0xc0862567f792e9d8, 0xbe1cee8ad30a57b5
+ .quad 0xc086256afd6e4c08, 0xbe1cef4e1817b90b
+ .quad 0xc086256e0225b3b8, 0xbe1cee7fa9229996
+ .quad 0xc086257105b9fce0, 0xbe1cf0b54963d945
+ .quad 0xc0862574082c0298, 0xbe1cee5f2f3c7995
+ .quad 0xc0862577097c9ee0, 0xbe1cf0828e303a2c
+ .quad 0xc086257a09acaae0, 0xbe1cf172c3078947
+ .quad 0xc086257d08bcfec0, 0xbe1cf189252afa22
+ .quad 0xc086258006ae71b8, 0xbe1cefdb80426923
+ .quad 0xc08625830381da08, 0xbe1ceef1391a0372
+ .quad 0xc0862585ff380d00, 0xbe1cf17720c78d13
+ .quad 0xc0862588f9d1df18, 0xbe1ceef1f9027d83
+ .quad 0xc086258bf35023b8, 0xbe1cf06fac99dec9
+ .quad 0xc086258eebb3ad78, 0xbe1cf1373eeb45c0
+ .quad 0xc0862591e2fd4e00, 0xbe1cef777536bb81
+ .quad 0xc0862594d92dd600, 0xbe1cf0f43ca40766
+ .quad 0xc0862597ce461558, 0xbe1cefb2cfc6766b
+ .quad 0xc086259ac246daf0, 0xbe1ceea49e64ffa2
+ .quad 0xc086259db530f4c8, 0xbe1cf250fa457dec
+ .quad 0xc08625a0a7053018, 0xbe1cf17d8bb2a44e
+ .quad 0xc08625a397c45918, 0xbe1cf1d5906d54b7
+ .quad 0xc08625a6876f3b30, 0xbe1cf08fe7b31780
+ .quad 0xc08625a97606a0e0, 0xbe1cef13edfc9d11
+ .quad 0xc08625ac638b53c8, 0xbe1cef9d2b107219
+ .quad 0xc08625af4ffe1cb0, 0xbe1cf1ddd4ff6160
+ .quad 0xc08625b23b5fc390, 0xbe1cefa02a996495
+ .quad 0xc08625b525b10f68, 0xbe1cf166a7e37ee5
+ .quad 0xc08625b80ef2c680, 0xbe1cef0b171068a5
+ .quad 0xc08625baf725ae28, 0xbe1cf05c80779283
+ .quad 0xc08625bdde4a8af0, 0xbe1cf1bbfbffb889
+ .quad 0xc08625c0c4622090, 0xbe1cf0b8666c0124
+ .quad 0xc08625c3a96d31e0, 0xbe1cf0a8fcf47a86
+ .quad 0xc08625c68d6c80f0, 0xbe1cef46e18cb092
+ .quad 0xc08625c97060cef0, 0xbe1cf1458a350efb
+ .quad 0xc08625cc524adc58, 0xbe1ceeea1dadce12
+ .quad 0xc08625cf332b68b0, 0xbe1cf0a1bfdc44c7
+ .quad 0xc08625d2130332d0, 0xbe1cef96d02da73e
+ .quad 0xc08625d4f1d2f8a8, 0xbe1cf2451c3c7701
+ .quad 0xc08625d7cf9b7778, 0xbe1cf10d08f83812
+ .quad 0xc08625daac5d6ba0, 0xbe1ceec5b4895c5e
+ .quad 0xc08625dd881990b0, 0xbe1cf14e1325c5e4
+ .quad 0xc08625e062d0a188, 0xbe1cf21d0904be12
+ .quad 0xc08625e33c835838, 0xbe1ceed0839bcf21
+ .quad 0xc08625e615326df0, 0xbe1cf1bb944889d2
+ .quad 0xc08625e8ecde9b48, 0xbe1cee738e85eece
+ .quad 0xc08625ebc38897e0, 0xbe1cf25c2bc6ef12
+ .quad 0xc08625ee99311ac8, 0xbe1cf132b70a41ad
+ .quad 0xc08625f16dd8da28, 0xbe1cf1984236a6e3
+ .quad 0xc08625f441808b78, 0xbe1cf19ae74998f9
+ .quad 0xc08625f71428e370, 0xbe1cef3e175d61a1
+ .quad 0xc08625f9e5d295f8, 0xbe1cf101f9868fd9
+ .quad 0xc08625fcb67e5658, 0xbe1cee69db83dcd2
+ .quad 0xc08625ff862cd6f8, 0xbe1cf081b636af51
+ .quad 0xc086260254dec9a8, 0xbe1cee62c7d59b3e
+ .quad 0xc08626052294df58, 0xbe1cf1b745c57716
+ .quad 0xc0862607ef4fc868, 0xbe1cef3d2800ea23
+ .quad 0xc086260abb103458, 0xbe1cef480ff1acd2
+ .quad 0xc086260d85d6d200, 0xbe1cf2424c9a17ef
+ .quad 0xc08626104fa44f90, 0xbe1cf12cfde90fd5
+ .quad 0xc086261318795a68, 0xbe1cf21f590dd5b6
+ .quad 0xc0862615e0569f48, 0xbe1cf0c50f9cd28a
+ .quad 0xc0862618a73cca30, 0xbe1ceedbdb520545
+ .quad 0xc086261b6d2c8668, 0xbe1cf0b030396011
+ .quad 0xc086261e32267e98, 0xbe1cf19917010e96
+ .quad 0xc0862620f62b5cb0, 0xbe1cf07331355985
+ .quad 0xc0862623b93bc9e8, 0xbe1cf01ae921a1c3
+ .quad 0xc08626267b586ed0, 0xbe1cefe5cf0dbf0c
+ .quad 0xc08626293c81f348, 0xbe1cf01b258aeb50
+ .quad 0xc086262bfcb8fe88, 0xbe1cee6b9e7f4c68
+ .quad 0xc086262ebbfe3710, 0xbe1cee684a9b21c9
+ .quad 0xc08626317a5242b8, 0xbe1cf1f8bcde9a8b
+ .quad 0xc086263437b5c6c0, 0xbe1cf1d063d36238
+ .quad 0xc0862636f42967a8, 0xbe1cf1e31a19075e
+ .quad 0xc0862639afadc950, 0xbe1cf1d8efdf7e7d
+ .quad 0xc086263c6a438ef0, 0xbe1cf1812ee72dba
+ .quad 0xc086263f23eb5b18, 0xbe1cf1449a9a2279
+ .quad 0xc0862641dca5cfb8, 0xbe1cee96edce5085
+ .quad 0xc086264494738e08, 0xbe1cf06797bd03b2
+ .quad 0xc08626474b5536b8, 0xbe1cef91b9b7ffc1
+ .quad 0xc086264a014b69c0, 0xbe1cef4b6721278f
+ .quad 0xc086264cb656c678, 0xbe1cf1942925eb4a
+ .quad 0xc086264f6a77eba8, 0xbe1cefa2c7bc2e39
+ .quad 0xc08626521daf7758, 0xbe1cf252595aceb3
+ .quad 0xc0862654cffe0718, 0xbe1cee8e9ae47ec2
+ .quad 0xc0862657816437a8, 0xbe1cf1bf913828fa
+ .quad 0xc086265a31e2a558, 0xbe1cf23475d6b366
+ .quad 0xc086265ce179ebc8, 0xbe1cef8df00a922b
+ .quad 0xc086265f902aa5f0, 0xbe1cef279bfa43e0
+ .quad 0xc08626623df56e38, 0xbe1cf080e10b8365
+ .quad 0xc0862664eadade70, 0xbe1cf1a518f9b544
+ .quad 0xc086266796db8fd0, 0xbe1cef9308fed9e9
+ .quad 0xc086266a41f81ae8, 0xbe1ceea3ae6b19c9
+ .quad 0xc086266cec3117b8, 0xbe1ceef06003d4c2
+ .quad 0xc086266f95871da8, 0xbe1cf0b8457ffb0c
+ .quad 0xc08626723dfac390, 0xbe1cf0c526745ad6
+ .quad 0xc0862674e58c9fa8, 0xbe1cf0cf91ff7b5d
+ .quad 0xc08626778c3d4798, 0xbe1cefe260819380
+ .quad 0xc086267a320d5070, 0xbe1ceebd90aa27a3
+ .quad 0xc086267cd6fd4ea8, 0xbe1cf0388121dffa
+ .quad 0xc086267f7b0dd630, 0xbe1cf1a3881435f1
+ .quad 0xc08626821e3f7a68, 0xbe1cef28e9d9ac52
+ .quad 0xc0862684c092ce08, 0xbe1cf02d300062dd
+ .quad 0xc086268762086350, 0xbe1cefaee1edfa35
+ .quad 0xc086268a02a0cbe0, 0xbe1cf0a5a052e936
+ .quad 0xc086268ca25c98d8, 0xbe1cee60a4a497ed
+ .quad 0xc086268f413c5ab0, 0xbe1cf0e4a5d0cf49
+ .quad 0xc0862691df40a170, 0xbe1cf149235a4e6e
+ .quad 0xc08626947c69fc80, 0xbe1cf215180b9fcc
+ .quad 0xc086269718b8fac8, 0xbe1cef9b156a9840
+ .quad 0xc0862699b42e2a90, 0xbe1cf054c91441be
+ .quad 0xc086269c4eca19a8, 0xbe1cf13ded26512c
+ .quad 0xc086269ee88d5550, 0xbe1cf22ea4d8ac06
+ .quad 0xc08626a181786a40, 0xbe1cf2354666ee2e
+ .quad 0xc08626a4198be4a8, 0xbe1cefef936752b3
+ .quad 0xc08626a6b0c85020, 0xbe1cf1e360a9db68
+ .quad 0xc08626a9472e37d8, 0xbe1ceed6aeb812c5
+ .quad 0xc08626abdcbe2650, 0xbe1cf227340b4986
+ .quad 0xc08626ae7178a5b0, 0xbe1cf0215a0cbe0d
+ .quad 0xc08626b1055e3f70, 0xbe1cf256adf0ae26
+ .quad 0xc08626b3986f7ca8, 0xbe1ceff3c67aed06
+ .quad 0xc08626b62aace5c8, 0xbe1cf2159fb93652
+ .quad 0xc08626b8bc1702e0, 0xbe1cf01e6dbd1c7f
+ .quad 0xc08626bb4cae5b60, 0xbe1cf009e75d1c0c
+ .quad 0xc08626bddc737648, 0xbe1ceec10a020e73
+ .quad 0xc08626c06b66da08, 0xbe1cf06d5783eee7
+ .quad 0xc08626c2f9890ca0, 0xbe1cf0cb8f169ffe
+ .quad 0xc08626c586da9388, 0xbe1cef7de2452430
+ .quad 0xc08626c8135bf3b0, 0xbe1cf05da6f783ae
+ .quad 0xc08626ca9f0db198, 0xbe1cefcc877d681d
+ .quad 0xc08626cd29f05138, 0xbe1cef0531954ab3
+ .quad 0xc08626cfb4045608, 0xbe1cf06b8565ea3d
+ .quad 0xc08626d23d4a4310, 0xbe1cefdc455d9d7e
+ .quad 0xc08626d4c5c29ad0, 0xbe1ceefc47e8fa64
+ .quad 0xc08626d74d6ddf48, 0xbe1cf1872bf033f2
+ .quad 0xc08626d9d44c9210, 0xbe1cf19d91087f9d
+ .quad 0xc08626dc5a5f3438, 0xbe1cf012d444c6ab
+ .quad 0xc08626dedfa64650, 0xbe1cf0ba528ee153
+ .quad 0xc08626e164224880, 0xbe1ceeb431709788
+ .quad 0xc08626e3e7d3ba60, 0xbe1cf0b9af31a6a5
+ .quad 0xc08626e66abb1b28, 0xbe1cf168fb2e135b
+ .quad 0xc08626e8ecd8e990, 0xbe1cef9097461c93
+ .quad 0xc08626eb6e2da3d0, 0xbe1cee7a434735d8
+ .quad 0xc08626edeeb9c7a8, 0xbe1cf235732b86f2
+ .quad 0xc08626f06e7dd280, 0xbe1cefe1510b89e6
+ .quad 0xc08626f2ed7a4120, 0xbe1cf1f64b9b80ef
+ .quad 0xc08626f56baf9000, 0xbe1cf08f320ca339
+ .quad 0xc08626f7e91e3b08, 0xbe1cf1b1de2808a1
+ .quad 0xc08626fa65c6bdc0, 0xbe1cf1976d778b28
+ .quad 0xc08626fce1a99338, 0xbe1ceef40a4f076f
+ .quad 0xc08626ff5cc73600, 0xbe1cef3e45869ce3
+ .quad 0xc0862701d7202048, 0xbe1ceef601b4c9d6
+ .quad 0xc086270450b4cbc0, 0xbe1cf1eaf0b57fd6
+ .quad 0xc0862706c985b1c0, 0xbe1cef82a44990f3
+ .quad 0xc086270941934b10, 0xbe1ceefe32981f2c
+ .quad 0xc086270bb8de1018, 0xbe1cefbf6f5a0445
+ .quad 0xc086270e2f6678d0, 0xbe1cf18dba75792c
+ .quad 0xc0862710a52cfcc8, 0xbe1cf0da64ce995f
+ .quad 0xc08627131a321318, 0xbe1cef04ac0fb802
+ .quad 0xc08627158e763268, 0xbe1cee9d4e2ad9bd
+ .quad 0xc086271801f9d0f8, 0xbe1cefa9b55407b5
+ .quad 0xc086271a74bd64a0, 0xbe1cefe6bd329570
+ .quad 0xc086271ce6c162c8, 0xbe1cef0b1205dc85
+ .quad 0xc086271f58064068, 0xbe1cef092a785e3f
+ .quad 0xc0862721c88c7210, 0xbe1cf050dcdaac30
+ .quad 0xc086272438546be8, 0xbe1cf210907ded8b
+ .quad 0xc0862726a75ea1b8, 0xbe1cee760be44f99
+ .quad 0xc086272915ab86c0, 0xbe1ceeeee07c2bcc
+ .quad 0xc086272b833b8df0, 0xbe1cf06874992df5
+ .quad 0xc086272df00f29d0, 0xbe1cef8fac5d4899
+ .quad 0xc08627305c26cc70, 0xbe1cf1103241cc99
+ .quad 0xc0862732c782e788, 0xbe1cf1d35fef83fe
+ .quad 0xc08627353223ec68, 0xbe1cef3ec8133e1d
+ .quad 0xc08627379c0a4be8, 0xbe1cef7261daccd8
+ .quad 0xc086273a05367688, 0xbe1cf18656c50806
+ .quad 0xc086273c6da8dc68, 0xbe1cf1c8736e049a
+ .quad 0xc086273ed561ed38, 0xbe1cf1f93bff4911
+ .quad 0xc08627413c621848, 0xbe1cf188a4ea680c
+ .quad 0xc0862743a2a9cc80, 0xbe1cf1d270930c80
+ .quad 0xc086274608397868, 0xbe1cf25a328c28e2
+ .quad 0xc08627486d118a28, 0xbe1cf106f90aa3b8
+ .quad 0xc086274ad1326f80, 0xbe1cee5e9d2e885a
+ .quad 0xc086274d349c95c0, 0xbe1cf1c0bac27228
+ .quad 0xc086274f975069f8, 0xbe1cf1a1500f9b1c
+ .quad 0xc0862751f94e58c0, 0xbe1cefc30663ac44
+ .quad 0xc08627545a96ce48, 0xbe1cf17123e427a2
+ .quad 0xc0862756bb2a3678, 0xbe1cefb92749fea4
+ .quad 0xc08627591b08fcc0, 0xbe1cefa40e1ea74a
+ .quad 0xc086275b7a338c40, 0xbe1cee6f4612c3e9
+ .quad 0xc086275dd8aa4fa8, 0xbe1cf1c54a053627
+ .quad 0xc0862760366db168, 0xbe1ceff5eb503d9e
+ .quad 0xc0862762937e1b70, 0xbe1cf02e47f10cee
+ .quad 0xc0862764efdbf768, 0xbe1ceeb06e1d0dad
+ .quad 0xc08627674b87ae88, 0xbe1cf10aadd6dba5
+ .quad 0xc0862769a681a9c0, 0xbe1cf24e9913d30f
+ .quad 0xc086276c00ca51a0, 0xbe1cef47b301e312
+ .quad 0xc086276e5a620e48, 0xbe1ceeb1cefc2e85
+ .quad 0xc0862770b3494788, 0xbe1cf16f1fbbe011
+ .quad 0xc08627730b8064e8, 0xbe1ceebdf75174c7
+ .quad 0xc08627756307cd70, 0xbe1cf06e3871a0da
+ .quad 0xc0862777b9dfe7f0, 0xbe1cef16799fd554
+ .quad 0xc086277a10091ac0, 0xbe1cf248dabf5377
+ .quad 0xc086277c6583cc00, 0xbe1cf0c78d92a2cd
+ .quad 0xc086277eba506158, 0xbe1cf0b911b029f0
+ .quad 0xc08627810e6f4028, 0xbe1cefdc24719766
+ .quad 0xc086278361e0cd70, 0xbe1cefbb6562b7e7
+ .quad 0xc0862785b4a56dd8, 0xbe1cf1e0afb349ec
+ .quad 0xc086278806bd85c0, 0xbe1cf008292e52fc
+ .quad 0xc086278a58297918, 0xbe1cf053073872bf
+ .quad 0xc086278ca8e9ab88, 0xbe1cf17a0a55a947
+ .quad 0xc086278ef8fe8068, 0xbe1ceeffb0b60234
+ .quad 0xc086279148685aa0, 0xbe1cf162204794a8
+ .quad 0xc086279397279ce0, 0xbe1cf24cc8cb48ac
+ .quad 0xc0862795e53ca978, 0xbe1cf0c9be68d5c3
+ .quad 0xc086279832a7e258, 0xbe1cf172cd3d7388
+ .quad 0xc086279a7f69a930, 0xbe1ceea2465fbce5
+ .quad 0xc086279ccb825f40, 0xbe1cf0a386d2500f
+ .quad 0xc086279f16f26590, 0xbe1cf1e338ddc18a
+ .quad 0xc08627a161ba1cd0, 0xbe1cef1f5049867f
+ .quad 0xc08627a3abd9e548, 0xbe1cef96c1ea8b1f
+ .quad 0xc08627a5f5521f00, 0xbe1cf138f6fd3c26
+ .quad 0xc08627a83e2329b0, 0xbe1cf0d4fcbfdf3a
+ .quad 0xc08627aa864d64b0, 0xbe1cf24870c12c81
+ .quad 0xc08627accdd12f18, 0xbe1cf0ae2a56348d
+ .quad 0xc08627af14aee7a0, 0xbe1cee8ca1a9b893
+ .quad 0xc08627b15ae6eca8, 0xbe1cf20414d637b0
+ .quad 0xc08627b3a0799c60, 0xbe1cf0fc6b7b12d8
+ .quad 0xc08627b5e5675488, 0xbe1cf152d93c4a00
+ .quad 0xc08627b829b072a0, 0xbe1cf1073f9b77c2
+ .quad 0xc08627ba6d5553d8, 0xbe1cee694f97d5a4
+ .quad 0xc08627bcb0565500, 0xbe1cf0456b8239d7
+ .quad 0xc08627bef2b3d2b0, 0xbe1cf211497127e3
+ .quad 0xc08627c1346e2930, 0xbe1cf01856c0384d
+ .quad 0xc08627c37585b468, 0xbe1cefa7dd05479e
+ .quad 0xc08627c5b5fad000, 0xbe1cef3ae8e50b93
+ .quad 0xc08627c7f5cdd750, 0xbe1ceea5f32fdd3a
+ .quad 0xc08627ca34ff2560, 0xbe1cef424caeb8d9
+ .quad 0xc08627cc738f14f0, 0xbe1cf0194d07a81f
+ .quad 0xc08627ceb17e0070, 0xbe1cf20f452000c1
+ .quad 0xc08627d0eecc4210, 0xbe1cf00e356218e4
+ .quad 0xc08627d32b7a33a0, 0xbe1cef30484b4bcb
+ .quad 0xc08627d567882eb0, 0xbe1ceeea11a6641b
+ .quad 0xc08627d7a2f68c80, 0xbe1cf13492d5bd7b
+ .quad 0xc08627d9ddc5a618, 0xbe1ceeb7048fad96
+ .quad 0xc08627dc17f5d418, 0xbe1ceef0666f0477
+ .quad 0xc08627de51876ee8, 0xbe1cf060d4b8b5c2
+ .quad 0xc08627e08a7acea8, 0xbe1cf0b2a4b6ff8c
+ .quad 0xc08627e2c2d04b28, 0xbe1cf0e34809a875
+ .quad 0xc08627e4fa883bf0, 0xbe1cf16bf74a3522
+ .quad 0xc08627e731a2f848, 0xbe1cee6a24623d57
+ .quad 0xc08627e96820d718, 0xbe1cefc7b4f1528e
+ .quad 0xc08627eb9e022f18, 0xbe1cf163051f3548
+ .quad 0xc08627edd34756b8, 0xbe1cef36b3366305
+ .quad 0xc08627f007f0a408, 0xbe1cf18134625550
+ .quad 0xc08627f23bfe6cf0, 0xbe1cf0ec32ec1a11
+ .quad 0xc08627f46f710700, 0xbe1ceeb3b64f3edc
+ .quad 0xc08627f6a248c778, 0xbe1cf0cd15805bc8
+ .quad 0xc08627f8d4860368, 0xbe1cf20db3bddebe
+ .quad 0xc08627fb06290f90, 0xbe1cf25188430e25
+ .quad 0xc08627fd37324070, 0xbe1ceea1713490f9
+ .quad 0xc08627ff67a1ea28, 0xbe1cf159521d234c
+ .quad 0xc0862801977860b8, 0xbe1cf24dfe50783b
+ .quad 0xc0862803c6b5f7d0, 0xbe1ceef2ef89a60b
+ .quad 0xc0862805f55b02c8, 0xbe1cee7fc919d62c
+ .quad 0xc08628082367d4c0, 0xbe1cf215a7fb513a
+ .quad 0xc086280a50dcc0a8, 0xbe1cf0e4401c5ed4
+ .quad 0xc086280c7dba1910, 0xbe1cf04ec734d256
+ .quad 0xc086280eaa003050, 0xbe1cf010ad787fea
+ .quad 0xc0862810d5af5880, 0xbe1cee622478393d
+ .quad 0xc086281300c7e368, 0xbe1cf01c7482564f
+ .quad 0xc08628152b4a22a0, 0xbe1cf0de20d33536
+ .quad 0xc086281755366778, 0xbe1cef2edae5837d
+ .quad 0xc08628197e8d02f0, 0xbe1cf0a345318cc9
+ .quad 0xc086281ba74e45d8, 0xbe1cf20085aa34b8
+ .quad 0xc086281dcf7a80c0, 0xbe1cef5fa845ad83
+ .quad 0xc086281ff71203e0, 0xbe1cf050d1df69c4
+ .quad 0xc08628221e151f48, 0xbe1ceffe43c035b9
+ .quad 0xc0862824448422b8, 0xbe1cf14f3018d3c2
+ .quad 0xc08628266a5f5dc0, 0xbe1cef0a5fbae83d
+ .quad 0xc08628288fa71f98, 0xbe1ceff8a95b72a1
+ .quad 0xc086282ab45bb750, 0xbe1cef073aa9849b
+ .quad 0xc086282cd87d73a8, 0xbe1cef69b3835c02
+ .quad 0xc086282efc0ca328, 0xbe1cf0bc139379a9
+ .quad 0xc08628311f099420, 0xbe1cef247a9ec596
+ .quad 0xc086283341749490, 0xbe1cef74bbcc488a
+ .quad 0xc0862835634df248, 0xbe1cef4bc42e7b8e
+ .quad 0xc08628378495fad0, 0xbe1cf136d4d5a810
+ .quad 0xc0862839a54cfb80, 0xbe1cf0d290b24dd8
+ .quad 0xc086283bc5734168, 0xbe1ceeebde8e0065
+ .quad 0xc086283de5091950, 0xbe1cf1a09f60aa1e
+ .quad 0xc0862840040ecfe0, 0xbe1cf0803947a234
+ .quad 0xc08628422284b168, 0xbe1cf0abf7638127
+ .quad 0xc0862844406b0a08, 0xbe1cf0f73ee12058
+ .quad 0xc08628465dc225a0, 0xbe1cf2079971b26c
+ .quad 0xc08628487a8a4fe0, 0xbe1cee74957564b1
+ .quad 0xc086284a96c3d420, 0xbe1ceee77c1b7d43
+ .quad 0xc086284cb26efd90, 0xbe1cf23addba6e09
+ .quad 0xc086284ecd8c1730, 0xbe1cf199f4a1da60
+ .quad 0xc0862850e81b6bb0, 0xbe1cf09fdea81393
+ .quad 0xc0862853021d4588, 0xbe1cf176adb417f7
+ .quad 0xc08628551b91ef00, 0xbe1cf0f64f84a8da
+ .quad 0xc08628573479b220, 0xbe1ceec34cf49523
+ .quad 0xc08628594cd4d8a8, 0xbe1cf16d60fbe0bb
+ .quad 0xc086285b64a3ac40, 0xbe1cee8de7acfc7b
+ .quad 0xc086285d7be67630, 0xbe1ceee6256cce8d
+ .quad 0xc086285f929d7fa0, 0xbe1cee7d66a3d8a5
+ .quad 0xc0862861a8c91170, 0xbe1cf0bef8265792
+ .quad 0xc0862863be697458, 0xbe1cf097f890c6f8
+ .quad 0xc0862865d37ef0c8, 0xbe1cf09502d5c3fc
+ .quad 0xc0862867e809cf00, 0xbe1ceeffb239dac7
+ .quad 0xc0862869fc0a56f8, 0xbe1cf1fbfff95c98
+ .quad 0xc086286c0f80d090, 0xbe1cefa57ad3eef7
+ .quad 0xc086286e226d8348, 0xbe1cf22c58b9183d
+ .quad 0xc086287034d0b690, 0xbe1ceff262d0a248
+ .quad 0xc086287246aab180, 0xbe1cefa7bc194186
+ .quad 0xc086287457fbbb08, 0xbe1cf06782d784d9
+ .quad 0xc086287668c419e0, 0xbe1cf1d44d0eaa07
+ .quad 0xc086287879041490, 0xbe1cf034803c8a48
+ .quad 0xc086287a88bbf158, 0xbe1cf08e84916b6f
+ .quad 0xc086287c97ebf650, 0xbe1cf0c4d3dc1bc7
+ .quad 0xc086287ea6946958, 0xbe1cefb1e4625943
+ .quad 0xc0862880b4b59010, 0xbe1cf143efdd1fd0
+ .quad 0xc0862882c24faff8, 0xbe1cee9896d016da
+ .quad 0xc0862884cf630e38, 0xbe1cf2186072f2cc
+ .quad 0xc0862886dbefeff0, 0xbe1cef9217633d34
+ .quad 0xc0862888e7f699e0, 0xbe1cf05603549486
+ .quad 0xc086288af37750b0, 0xbe1cef50fff513d3
+ .quad 0xc086288cfe7258c0, 0xbe1cf127713b32d0
+ .quad 0xc086288f08e7f650, 0xbe1cf05015520f3d
+ .quad 0xc086289112d86d58, 0xbe1cf12eb458b26f
+ .quad 0xc08628931c4401a8, 0xbe1cf22eae2887ed
+ .quad 0xc0862895252af6e0, 0xbe1cefdd6656dd2d
+ .quad 0xc08628972d8d9058, 0xbe1cf1048ea4e646
+ .quad 0xc0862899356c1150, 0xbe1ceec4501167e9
+ .quad 0xc086289b3cc6bcb8, 0xbe1cf0ad52becc3f
+ .quad 0xc086289d439dd568, 0xbe1cf0daa4e00e35
+ .quad 0xc086289f49f19df8, 0xbe1cf00b80de8d6a
+ .quad 0xc08628a14fc258c8, 0xbe1cf1bcf2ea8464
+ .quad 0xc08628a355104818, 0xbe1cf0435e2782b0
+ .quad 0xc08628a559dbade0, 0xbe1cf0e3e1a5f56c
+ .quad 0xc08628a75e24cbf8, 0xbe1cefed9d5a721d
+ .quad 0xc08628a961ebe3f8, 0xbe1cf0d2d74321e2
+ .quad 0xc08628ab65313750, 0xbe1cf24200eb55e9
+ .quad 0xc08628ad67f50740, 0xbe1cf23e9d7cf979
+ .quad 0xc08628af6a3794d0, 0xbe1cf23a088f421c
+ .quad 0xc08628b16bf920e0, 0xbe1cef2c1de1ab32
+ .quad 0xc08628b36d39ec08, 0xbe1cf1abc231f7b2
+ .quad 0xc08628b56dfa36d0, 0xbe1cf2074d5ba303
+ .quad 0xc08628b76e3a4180, 0xbe1cf05cd5eed880
+ /*== Log_LA_table ==*/
+ .align 32
+ .quad 0x8000000000000000
+ .quad 0xbf5ff802a9ab10e6
+ .quad 0xbf6ff00aa2b10bc0
+ .quad 0xbf77ee11ebd82e94
+ .quad 0xbf7fe02a6b106789
+ .quad 0xbf83e7295d25a7d9
+ .quad 0xbf87dc475f810a77
+ .quad 0xbf8bcf712c74384c
+ .quad 0xbf8fc0a8b0fc03e4
+ .quad 0xbf91d7f7eb9eebe7
+ .quad 0xbf93cea44346a575
+ .quad 0xbf95c45a51b8d389
+ .quad 0xbf97b91b07d5b11b
+ .quad 0xbf99ace7551cc514
+ .quad 0xbf9b9fc027af9198
+ .quad 0xbf9d91a66c543cc4
+ .quad 0xbf9f829b0e783300
+ .quad 0xbfa0b94f7c196176
+ .quad 0xbfa1b0d98923d980
+ .quad 0xbfa2a7ec2214e873
+ .quad 0xbfa39e87b9febd60
+ .quad 0xbfa494acc34d911c
+ .quad 0xbfa58a5bafc8e4d5
+ .quad 0xbfa67f94f094bd98
+ .quad 0xbfa77458f632dcfc
+ .quad 0xbfa868a83083f6cf
+ .quad 0xbfa95c830ec8e3eb
+ .quad 0xbfaa4fe9ffa3d235
+ .quad 0xbfab42dd711971bf
+ .quad 0xbfac355dd0921f2d
+ .quad 0xbfad276b8adb0b52
+ .quad 0xbfae19070c276016
+ .quad 0xbfaf0a30c01162a6
+ .quad 0xbfaffae9119b9303
+ .quad 0xbfb075983598e471
+ .quad 0xbfb0ed839b5526fe
+ .quad 0xbfb16536eea37ae1
+ .quad 0xbfb1dcb263db1944
+ .quad 0xbfb253f62f0a1417
+ .quad 0xbfb2cb0283f5de1f
+ .quad 0xbfb341d7961bd1d1
+ .quad 0xbfb3b87598b1b6ee
+ .quad 0xbfb42edcbea646f0
+ .quad 0xbfb4a50d3aa1b040
+ .quad 0xbfb51b073f06183f
+ .quad 0xbfb590cafdf01c28
+ .quad 0xbfb60658a93750c4
+ .quad 0xbfb67bb0726ec0fc
+ .quad 0xbfb6f0d28ae56b4c
+ .quad 0xbfb765bf23a6be13
+ .quad 0xbfb7da766d7b12cd
+ .quad 0xbfb84ef898e8282a
+ .quad 0xbfb8c345d6319b21
+ .quad 0xbfb9375e55595ede
+ .quad 0xbfb9ab42462033ad
+ .quad 0xbfba1ef1d8061cd4
+ .quad 0xbfba926d3a4ad563
+ .quad 0xbfbb05b49bee43fe
+ .quad 0xbfbb78c82bb0eda1
+ .quad 0xbfbbeba818146765
+ .quad 0xbfbc5e548f5bc743
+ .quad 0xbfbcd0cdbf8c13e1
+ .quad 0xbfbd4313d66cb35d
+ .quad 0xbfbdb5270187d927
+ .quad 0xbfbe27076e2af2e6
+ .quad 0xbfbe98b549671467
+ .quad 0xbfbf0a30c01162a6
+ .quad 0xbfbf7b79fec37ddf
+ .quad 0xbfbfec9131dbeabb
+ .quad 0xbfc02ebb42bf3d4b
+ .quad 0xbfc0671512ca596e
+ .quad 0xbfc09f561ee719c3
+ .quad 0xbfc0d77e7cd08e59
+ .quad 0xbfc10f8e422539b1
+ .quad 0xbfc14785846742ac
+ .quad 0xbfc17f6458fca611
+ .quad 0xbfc1b72ad52f67a0
+ .quad 0xbfc1eed90e2dc2c3
+ .quad 0xbfc2266f190a5acb
+ .quad 0xbfc25ded0abc6ad2
+ .quad 0xbfc29552f81ff523
+ .quad 0xbfc2cca0f5f5f251
+ .quad 0xbfc303d718e47fd3
+ .quad 0xbfc33af575770e4f
+ .quad 0xbfc371fc201e8f74
+ .quad 0xbfc3a8eb2d31a376
+ .quad 0xbfc3dfc2b0ecc62a
+ .quad 0xbfc41682bf727bc0
+ .quad 0xbfc44d2b6ccb7d1e
+ .quad 0xbfc483bccce6e3dd
+ .quad 0xbfc4ba36f39a55e5
+ .quad 0xbfc4f099f4a230b2
+ .quad 0xbfc526e5e3a1b438
+ .quad 0xbfc55d1ad4232d6f
+ .quad 0xbfc59338d9982086
+ .quad 0xbfc5c940075972b9
+ .quad 0xbfc5ff3070a793d4
+ .quad 0xbfc6350a28aaa758
+ .quad 0xbfc66acd4272ad51
+ .quad 0xbfc6a079d0f7aad2
+ .quad 0xbfc6d60fe719d21d
+ .quad 0xbfc70b8f97a1aa75
+ .quad 0xbfc740f8f54037a5
+ .quad 0xbfc7764c128f2127
+ .quad 0xbfc7ab890210d909
+ .quad 0xbfc7e0afd630c274
+ .quad 0xbfc815c0a14357eb
+ .quad 0xbfc84abb75865139
+ .quad 0xbfc87fa06520c911
+ .quad 0xbfc8b46f8223625b
+ .quad 0xbfc8e928de886d41
+ .quad 0xbfc91dcc8c340bde
+ .quad 0xbfc9525a9cf456b4
+ .quad 0xbfc986d3228180ca
+ .quad 0xbfc9bb362e7dfb83
+ .quad 0xbfc9ef83d2769a34
+ .quad 0xbfca23bc1fe2b563
+ .quad 0xbfca57df28244dcd
+ .quad 0xbfca8becfc882f19
+ .quad 0xbfcabfe5ae46124c
+ .quad 0xbfcaf3c94e80bff3
+ .quad 0xbfcb2797ee46320c
+ .quad 0xbfcb5b519e8fb5a4
+ .quad 0xbfcb8ef670420c3b
+ .quad 0xbfcbc286742d8cd6
+ .quad 0xbfcbf601bb0e44e2
+ .quad 0xbfcc2968558c18c1
+ .quad 0xbfcc5cba543ae425
+ .quad 0xbfcc8ff7c79a9a22
+ .quad 0xbfccc320c0176502
+ .quad 0xbfccf6354e09c5dc
+ .quad 0xbfcd293581b6b3e7
+ .quad 0xbfcd5c216b4fbb91
+ .quad 0xbfcd8ef91af31d5e
+ .quad 0xbfcdc1bca0abec7d
+ .quad 0xbfcdf46c0c722d2f
+ .quad 0xbfce27076e2af2e6
+ .quad 0xbfce598ed5a87e2f
+ .quad 0xbfce8c0252aa5a60
+ .quad 0xbfcebe61f4dd7b0b
+ .quad 0xbfcef0adcbdc5936
+ .quad 0xbfcf22e5e72f105d
+ .quad 0xbfcf550a564b7b37
+ .quad 0xbfcf871b28955045
+ .quad 0xbfcfb9186d5e3e2b
+ .quad 0xbfcfeb0233e607cc
+ .quad 0xbfd00e6c45ad501d
+ .quad 0xbfd0274dc16c232f
+ .quad 0xbfd0402594b4d041
+ .quad 0xbfd058f3c703ebc6
+ .quad 0xbfd071b85fcd590d
+ .quad 0xbfd08a73667c57af
+ .quad 0xbfd0a324e27390e3
+ .quad 0xbfd0bbccdb0d24bd
+ .quad 0xbfd0d46b579ab74b
+ .quad 0xbfd0ed005f657da4
+ .quad 0xbfd1058bf9ae4ad5
+ .quad 0xbfd11e0e2dad9cb7
+ .quad 0xbfd136870293a8b0
+ .quad 0xbfd14ef67f88685a
+ .quad 0xbfd1675cababa60e
+ .quad 0xbfd17fb98e15095d
+ .quad 0xbfd1980d2dd4236f
+ .quad 0xbfd1b05791f07b49
+ .quad 0xbfd1c898c16999fb
+ .quad 0xbfd1e0d0c33716be
+ .quad 0xbfd1f8ff9e48a2f3
+ .quad 0xbfd211255986160c
+ .quad 0xbfd22941fbcf7966
+ .quad 0xbfd241558bfd1404
+ .quad 0xbfd2596010df763a
+ .quad 0xbfd27161913f853d
+ .quad 0xbfd2895a13de86a3
+ .quad 0xbfd2a1499f762bc9
+ .quad 0xbfd2b9303ab89d25
+ .quad 0xbfd2d10dec508583
+ .quad 0xbfd2e8e2bae11d31
+ .quad 0xbfd300aead06350c
+ .quad 0xbfd31871c9544185
+ .quad 0xbfd3302c16586588
+ .quad 0xbfd347dd9a987d55
+ .quad 0xbfd35f865c93293e
+ .quad 0xbfd3772662bfd85b
+ .quad 0xbfd38ebdb38ed321
+ .quad 0xbfd3a64c556945ea
+ .quad 0xbfd3bdd24eb14b6a
+ .quad 0xbfd3d54fa5c1f710
+ .quad 0xbfd3ecc460ef5f50
+ .quad 0xbfd404308686a7e4
+ .quad 0xbfd41b941cce0bee
+ .quad 0xbfd432ef2a04e814
+ .quad 0xbfd44a41b463c47c
+ .quad 0xbfd4618bc21c5ec2
+ .quad 0xbfd478cd5959b3d9
+ .quad 0xbfd49006804009d1
+ .quad 0xbfd4a7373cecf997
+ .quad 0xbfd4be5f957778a1
+ .quad 0xbfd4d57f8fefe27f
+ .quad 0xbfd4ec973260026a
+ .quad 0xbfd503a682cb1cb3
+ .quad 0xbfd51aad872df82d
+ .quad 0xbfd531ac457ee77e
+ .quad 0xbfd548a2c3add263
+ .quad 0xbfd55f9107a43ee2
+ .quad 0xbfd5767717455a6c
+ .quad 0xbfd58d54f86e02f2
+ .quad 0xbfd5a42ab0f4cfe2
+ .quad 0xbfd5baf846aa1b19
+ .quad 0xbfd5d1bdbf5809ca
+ .quad 0xbfd5e87b20c2954a
+ .quad 0xbfd5ff3070a793d4
+ .quad 0xbfd615ddb4bec13c
+ .quad 0xbfd62c82f2b9c795
+ .quad 0x3fd61965cdb02c1f
+ .quad 0x3fd602d08af091ec
+ .quad 0x3fd5ec433d5c35ae
+ .quad 0x3fd5d5bddf595f30
+ .quad 0x3fd5bf406b543db2
+ .quad 0x3fd5a8cadbbedfa1
+ .quad 0x3fd5925d2b112a59
+ .quad 0x3fd57bf753c8d1fb
+ .quad 0x3fd565995069514c
+ .quad 0x3fd54f431b7be1a9
+ .quad 0x3fd538f4af8f72fe
+ .quad 0x3fd522ae0738a3d8
+ .quad 0x3fd50c6f1d11b97c
+ .quad 0x3fd4f637ebba9810
+ .quad 0x3fd4e0086dd8baca
+ .quad 0x3fd4c9e09e172c3c
+ .quad 0x3fd4b3c077267e9a
+ .quad 0x3fd49da7f3bcc41f
+ .quad 0x3fd487970e958770
+ .quad 0x3fd4718dc271c41b
+ .quad 0x3fd45b8c0a17df13
+ .quad 0x3fd44591e0539f49
+ .quad 0x3fd42f9f3ff62642
+ .quad 0x3fd419b423d5e8c7
+ .quad 0x3fd403d086cea79c
+ .quad 0x3fd3edf463c1683e
+ .quad 0x3fd3d81fb5946dba
+ .quad 0x3fd3c25277333184
+ .quad 0x3fd3ac8ca38e5c5f
+ .quad 0x3fd396ce359bbf54
+ .quad 0x3fd3811728564cb2
+ .quad 0x3fd36b6776be1117
+ .quad 0x3fd355bf1bd82c8b
+ .quad 0x3fd3401e12aecba1
+ .quad 0x3fd32a84565120a8
+ .quad 0x3fd314f1e1d35ce4
+ .quad 0x3fd2ff66b04ea9d4
+ .quad 0x3fd2e9e2bce12286
+ .quad 0x3fd2d46602adccee
+ .quad 0x3fd2bef07cdc9354
+ .quad 0x3fd2a982269a3dbf
+ .quad 0x3fd2941afb186b7c
+ .quad 0x3fd27ebaf58d8c9d
+ .quad 0x3fd269621134db92
+ .quad 0x3fd25410494e56c7
+ .quad 0x3fd23ec5991eba49
+ .quad 0x3fd22981fbef797b
+ .quad 0x3fd214456d0eb8d4
+ .quad 0x3fd1ff0fe7cf47a7
+ .quad 0x3fd1e9e1678899f4
+ .quad 0x3fd1d4b9e796c245
+ .quad 0x3fd1bf99635a6b95
+ .quad 0x3fd1aa7fd638d33f
+ .quad 0x3fd1956d3b9bc2fa
+ .quad 0x3fd180618ef18adf
+ .quad 0x3fd16b5ccbacfb73
+ .quad 0x3fd1565eed455fc3
+ .quad 0x3fd14167ef367783
+ .quad 0x3fd12c77cd00713b
+ .quad 0x3fd1178e8227e47c
+ .quad 0x3fd102ac0a35cc1c
+ .quad 0x3fd0edd060b78081
+ .quad 0x3fd0d8fb813eb1ef
+ .quad 0x3fd0c42d676162e3
+ .quad 0x3fd0af660eb9e279
+ .quad 0x3fd09aa572e6c6d4
+ .quad 0x3fd085eb8f8ae797
+ .quad 0x3fd07138604d5862
+ .quad 0x3fd05c8be0d9635a
+ .quad 0x3fd047e60cde83b8
+ .quad 0x3fd03346e0106062
+ .quad 0x3fd01eae5626c691
+ .quad 0x3fd00a1c6adda473
+ .quad 0x3fcfeb2233ea07cd
+ .quad 0x3fcfc218be620a5e
+ .quad 0x3fcf991c6cb3b379
+ .quad 0x3fcf702d36777df0
+ .quad 0x3fcf474b134df229
+ .quad 0x3fcf1e75fadf9bde
+ .quad 0x3fcef5ade4dcffe6
+ .quad 0x3fceccf2c8fe920a
+ .quad 0x3fcea4449f04aaf5
+ .quad 0x3fce7ba35eb77e2a
+ .quad 0x3fce530effe71012
+ .quad 0x3fce2a877a6b2c12
+ .quad 0x3fce020cc6235ab5
+ .quad 0x3fcdd99edaf6d7e9
+ .quad 0x3fcdb13db0d48940
+ .quad 0x3fcd88e93fb2f450
+ .quad 0x3fcd60a17f903515
+ .quad 0x3fcd38666871f465
+ .quad 0x3fcd1037f2655e7b
+ .quad 0x3fcce816157f1988
+ .quad 0x3fccc000c9db3c52
+ .quad 0x3fcc97f8079d44ec
+ .quad 0x3fcc6ffbc6f00f71
+ .quad 0x3fcc480c0005ccd1
+ .quad 0x3fcc2028ab17f9b4
+ .quad 0x3fcbf851c067555f
+ .quad 0x3fcbd087383bd8ad
+ .quad 0x3fcba8c90ae4ad19
+ .quad 0x3fcb811730b823d2
+ .quad 0x3fcb5971a213acdb
+ .quad 0x3fcb31d8575bce3d
+ .quad 0x3fcb0a4b48fc1b46
+ .quad 0x3fcae2ca6f672bd4
+ .quad 0x3fcabb55c31693ad
+ .quad 0x3fca93ed3c8ad9e3
+ .quad 0x3fca6c90d44b704e
+ .quad 0x3fca454082e6ab05
+ .quad 0x3fca1dfc40f1b7f1
+ .quad 0x3fc9f6c407089664
+ .quad 0x3fc9cf97cdce0ec3
+ .quad 0x3fc9a8778debaa38
+ .quad 0x3fc981634011aa75
+ .quad 0x3fc95a5adcf7017f
+ .quad 0x3fc9335e5d594989
+ .quad 0x3fc90c6db9fcbcd9
+ .quad 0x3fc8e588ebac2dbf
+ .quad 0x3fc8beafeb38fe8c
+ .quad 0x3fc897e2b17b19a5
+ .quad 0x3fc871213750e994
+ .quad 0x3fc84a6b759f512f
+ .quad 0x3fc823c16551a3c2
+ .quad 0x3fc7fd22ff599d4f
+ .quad 0x3fc7d6903caf5ad0
+ .quad 0x3fc7b0091651528c
+ .quad 0x3fc7898d85444c73
+ .quad 0x3fc7631d82935a86
+ .quad 0x3fc73cb9074fd14d
+ .quad 0x3fc716600c914054
+ .quad 0x3fc6f0128b756abc
+ .quad 0x3fc6c9d07d203fc7
+ .quad 0x3fc6a399dabbd383
+ .quad 0x3fc67d6e9d785771
+ .quad 0x3fc6574ebe8c133a
+ .quad 0x3fc6313a37335d76
+ .quad 0x3fc60b3100b09476
+ .quad 0x3fc5e533144c1719
+ .quad 0x3fc5bf406b543db2
+ .quad 0x3fc59958ff1d52f1
+ .quad 0x3fc5737cc9018cdd
+ .quad 0x3fc54dabc26105d2
+ .quad 0x3fc527e5e4a1b58d
+ .quad 0x3fc5022b292f6a45
+ .quad 0x3fc4dc7b897bc1c8
+ .quad 0x3fc4b6d6fefe22a4
+ .quad 0x3fc4913d8333b561
+ .quad 0x3fc46baf0f9f5db7
+ .quad 0x3fc4462b9dc9b3dc
+ .quad 0x3fc420b32740fdd4
+ .quad 0x3fc3fb45a59928cc
+ .quad 0x3fc3d5e3126bc27f
+ .quad 0x3fc3b08b6757f2a9
+ .quad 0x3fc38b3e9e027479
+ .quad 0x3fc365fcb0159016
+ .quad 0x3fc340c59741142e
+ .quad 0x3fc31b994d3a4f85
+ .quad 0x3fc2f677cbbc0a96
+ .quad 0x3fc2d1610c86813a
+ .quad 0x3fc2ac55095f5c59
+ .quad 0x3fc28753bc11aba5
+ .quad 0x3fc2625d1e6ddf57
+ .quad 0x3fc23d712a49c202
+ .quad 0x3fc2188fd9807263
+ .quad 0x3fc1f3b925f25d41
+ .quad 0x3fc1ceed09853752
+ .quad 0x3fc1aa2b7e23f72a
+ .quad 0x3fc185747dbecf34
+ .quad 0x3fc160c8024b27b1
+ .quad 0x3fc13c2605c398c3
+ .quad 0x3fc1178e8227e47c
+ .quad 0x3fc0f301717cf0fb
+ .quad 0x3fc0ce7ecdccc28d
+ .quad 0x3fc0aa06912675d5
+ .quad 0x3fc08598b59e3a07
+ .quad 0x3fc06135354d4b18
+ .quad 0x3fc03cdc0a51ec0d
+ .quad 0x3fc0188d2ecf6140
+ .quad 0x3fbfe89139dbd566
+ .quad 0x3fbfa01c9db57ce2
+ .quad 0x3fbf57bc7d9005db
+ .quad 0x3fbf0f70cdd992e3
+ .quad 0x3fbec739830a1120
+ .quad 0x3fbe7f1691a32d3e
+ .quad 0x3fbe3707ee30487b
+ .quad 0x3fbdef0d8d466db9
+ .quad 0x3fbda727638446a2
+ .quad 0x3fbd5f55659210e2
+ .quad 0x3fbd179788219364
+ .quad 0x3fbccfedbfee13a8
+ .quad 0x3fbc885801bc4b23
+ .quad 0x3fbc40d6425a5cb1
+ .quad 0x3fbbf968769fca11
+ .quad 0x3fbbb20e936d6974
+ .quad 0x3fbb6ac88dad5b1c
+ .quad 0x3fbb23965a52ff00
+ .quad 0x3fbadc77ee5aea8c
+ .quad 0x3fba956d3ecade63
+ .quad 0x3fba4e7640b1bc38
+ .quad 0x3fba0792e9277cac
+ .quad 0x3fb9c0c32d4d2548
+ .quad 0x3fb97a07024cbe74
+ .quad 0x3fb9335e5d594989
+ .quad 0x3fb8ecc933aeb6e8
+ .quad 0x3fb8a6477a91dc29
+ .quad 0x3fb85fd927506a48
+ .quad 0x3fb8197e2f40e3f0
+ .quad 0x3fb7d33687c293c9
+ .quad 0x3fb78d02263d82d3
+ .quad 0x3fb746e100226ed9
+ .quad 0x3fb700d30aeac0e1
+ .quad 0x3fb6bad83c1883b6
+ .quad 0x3fb674f089365a7a
+ .quad 0x3fb62f1be7d77743
+ .quad 0x3fb5e95a4d9791cb
+ .quad 0x3fb5a3abb01ade25
+ .quad 0x3fb55e10050e0384
+ .quad 0x3fb518874226130a
+ .quad 0x3fb4d3115d207eac
+ .quad 0x3fb48dae4bc31018
+ .quad 0x3fb4485e03dbdfad
+ .quad 0x3fb403207b414b7f
+ .quad 0x3fb3bdf5a7d1ee64
+ .quad 0x3fb378dd7f749714
+ .quad 0x3fb333d7f8183f4b
+ .quad 0x3fb2eee507b40301
+ .quad 0x3fb2aa04a44717a5
+ .quad 0x3fb26536c3d8c369
+ .quad 0x3fb2207b5c78549e
+ .quad 0x3fb1dbd2643d190b
+ .quad 0x3fb1973bd1465567
+ .quad 0x3fb152b799bb3cc9
+ .quad 0x3fb10e45b3cae831
+ .quad 0x3fb0c9e615ac4e17
+ .quad 0x3fb08598b59e3a07
+ .quad 0x3fb0415d89e74444
+ .quad 0x3faffa6911ab9301
+ .quad 0x3faf723b517fc523
+ .quad 0x3faeea31c006b87c
+ .quad 0x3fae624c4a0b5e1b
+ .quad 0x3fadda8adc67ee4e
+ .quad 0x3fad52ed6405d86f
+ .quad 0x3faccb73cdddb2cc
+ .quad 0x3fac441e06f72a9e
+ .quad 0x3fabbcebfc68f420
+ .quad 0x3fab35dd9b58baad
+ .quad 0x3faaaef2d0fb10fc
+ .quad 0x3faa282b8a936171
+ .quad 0x3fa9a187b573de7c
+ .quad 0x3fa91b073efd7314
+ .quad 0x3fa894aa149fb343
+ .quad 0x3fa80e7023d8ccc4
+ .quad 0x3fa788595a3577ba
+ .quad 0x3fa70265a550e777
+ .quad 0x3fa67c94f2d4bb58
+ .quad 0x3fa5f6e73078efb8
+ .quad 0x3fa5715c4c03ceef
+ .quad 0x3fa4ebf43349e26f
+ .quad 0x3fa466aed42de3ea
+ .quad 0x3fa3e18c1ca0ae92
+ .quad 0x3fa35c8bfaa1306b
+ .quad 0x3fa2d7ae5c3c5bae
+ .quad 0x3fa252f32f8d183f
+ .quad 0x3fa1ce5a62bc353a
+ .quad 0x3fa149e3e4005a8d
+ .quad 0x3fa0c58fa19dfaaa
+ .quad 0x3fa0415d89e74444
+ .quad 0x3f9f7a9b16782856
+ .quad 0x3f9e72bf2813ce51
+ .quad 0x3f9d6b2725979802
+ .quad 0x3f9c63d2ec14aaf2
+ .quad 0x3f9b5cc258b718e6
+ .quad 0x3f9a55f548c5c43f
+ .quad 0x3f994f6b99a24475
+ .quad 0x3f98492528c8cabf
+ .quad 0x3f974321d3d006d3
+ .quad 0x3f963d6178690bd6
+ .quad 0x3f9537e3f45f3565
+ .quad 0x3f9432a925980cc1
+ .quad 0x3f932db0ea132e22
+ .quad 0x3f9228fb1fea2e28
+ .quad 0x3f912487a5507f70
+ .quad 0x3f90205658935847
+ .quad 0x3f8e38ce3033310c
+ .quad 0x3f8c317384c75f06
+ .quad 0x3f8a2a9c6c170462
+ .quad 0x3f882448a388a2aa
+ .quad 0x3f861e77e8b53fc6
+ .quad 0x3f841929f96832f0
+ .quad 0x3f82145e939ef1e9
+ .quad 0x3f8010157588de71
+ .quad 0x3f7c189cbb0e27fb
+ .quad 0x3f78121214586b54
+ .quad 0x3f740c8a747878e2
+ .quad 0x3f70080559588b35
+ .quad 0x3f680904828985c0
+ .quad 0x3f60040155d5889e
+ .quad 0x3f50020055655889
+ .quad 0x0000000000000000
+ /*== poly_coeff[4] ==*/
+ .align 32
+ .quad 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A /* coeff4 */
+ .quad 0xbfd0000148058EE1, 0xbfd0000148058EE1, 0xbfd0000148058EE1, 0xbfd0000148058EE1 /* coeff3 */
+ .quad 0x3fd55555555543C5, 0x3fd55555555543C5, 0x3fd55555555543C5, 0x3fd55555555543C5 /* coeff2 */
+ .quad 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F /* coeff1 */
+ /*== ExpMask ==*/
+ .align 32
+ .quad 0x000fffffffffffff, 0x000fffffffffffff, 0x000fffffffffffff, 0x000fffffffffffff
+ /*== Two10 ==*/
+ .align 32
+ .quad 0x3f50000000000000, 0x3f50000000000000, 0x3f50000000000000, 0x3f50000000000000
+ /*== MinLog1p = -1+2^(-53) ==*/
+ .align 32
+ .quad 0xbfefffffffffffff, 0xbfefffffffffffff, 0xbfefffffffffffff, 0xbfefffffffffffff
+ /*== MaxLog1p ==*/
+ .align 32
+ .quad 0x7f3ffffffffff000, 0x7f3ffffffffff000, 0x7f3ffffffffff000, 0x7f3ffffffffff000
+ /*== One ==*/
+ .align 32
+ .quad 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000
+ /*== SgnMask ==*/
+ .align 32
+ .quad 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff
+ /*== XThreshold ==*/
+ .align 32
+ .quad 0x3e00000000000000, 0x3e00000000000000, 0x3e00000000000000, 0x3e00000000000000
+ /*== XhMask ==*/
+ .align 32
+ .quad 0xfffffffffffffc00, 0xfffffffffffffc00, 0xfffffffffffffc00, 0xfffffffffffffc00
+ /*== Threshold ==*/
+ .align 32
+ .quad 0x4086a00000000000, 0x4086a00000000000, 0x4086a00000000000, 0x4086a00000000000
+ /*== Bias ==*/
+ .align 32
+ .quad 0x408ff80000000000, 0x408ff80000000000, 0x408ff80000000000, 0x408ff80000000000
+ /*== Bias1 ==*/
+ .align 32
+ .quad 0x408ff00000000000, 0x408ff00000000000, 0x408ff00000000000, 0x408ff00000000000
+ /*== ExpMask ==*/
+ .align 32
+ .quad 0x7ff0000000000000, 0x7ff0000000000000, 0x7ff0000000000000, 0x7ff0000000000000
+ /*== ExpMask2 ==*/
+ .align 32
+ .quad 0x7f40000000000000, 0x7f40000000000000, 0x7f40000000000000, 0x7f40000000000000
+ /*== L2L ==*/
+ .align 32
+ .quad 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF
+ /*== dBigThreshold ==*/
+ .align 32
+ .quad 0x41D0000000000000, 0x41D0000000000000, 0x41D0000000000000, 0x41D0000000000000
+ /*== dC2 ==*/
+ .align 32
+ .quad 0x3FD8000000000000, 0x3FD8000000000000, 0x3FD8000000000000, 0x3FD8000000000000
+ /*== dC3 ==*/
+ .align 32
+ .quad 0x3FD4000000000000, 0x3FD4000000000000, 0x3FD4000000000000, 0x3FD4000000000000
+ /*== dC4 ==*/
+ .align 32
+ .quad 0x3FD1800000000000, 0x3FD1800000000000, 0x3FD1800000000000, 0x3FD1800000000000
+ /*== dC5 ==*/
+ .align 32
+ .quad 0x3FCF800000000000, 0x3FCF800000000000, 0x3FCF800000000000, 0x3FCF800000000000
+ /*== dHalf ==*/
+ .align 32
+ .quad 0x3FE0000000000000, 0x3FE0000000000000, 0x3FE0000000000000, 0x3FE0000000000000
+ /*== dLargestFinite ==*/
+ .align 32
+ .quad 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF
+ /*== dLittleThreshold ==*/
+ .align 32
+ .quad 0x3F60000000000000, 0x3F60000000000000, 0x3F60000000000000, 0x3F60000000000000
+ /*== dSign ==*/
+ .align 32
+ .quad 0x8000000000000000, 0x8000000000000000, 0x8000000000000000, 0x8000000000000000
+ /*== dThirtyOne ==*/
+ .align 32
+ .quad 0x403F000000000000, 0x403F000000000000, 0x403F000000000000, 0x403F000000000000
+ /*== dTopMask12 ==*/
+ .align 32
+ .quad 0xFFFFFE0000000000, 0xFFFFFE0000000000, 0xFFFFFE0000000000, 0xFFFFFE0000000000
+ /*== dTopMask29 ==*/
+ .align 32
+ .quad 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000
+ /*== XScale ==*/
+ .align 32
+ .quad 0x3E10000000000000, 0x3E10000000000000, 0x3E10000000000000, 0x3E10000000000000
+ .align 32
+ .type __svml_dasinh_data_internal,@object
+ .size __svml_dasinh_data_internal,.-__svml_dasinh_data_internal
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S
new file mode 100644
index 0000000..647c732
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S
@@ -0,0 +1,20 @@
+/* AVX2 version of vectorized asinh, vector length is 8.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define _ZGVeN8v_asinh _ZGVeN8v_asinh_avx2_wrapper
+#include "../svml_d_asinh8_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c
new file mode 100644
index 0000000..45e5ab7
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c
@@ -0,0 +1,27 @@
+/* Multiple versions of vectorized asinh, vector length is 8.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define SYMBOL_NAME _ZGVeN8v_asinh
+#include "ifunc-mathvec-avx512-skx.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVeN8v_asinh, __GI__ZGVeN8v_asinh, __redirect__ZGVeN8v_asinh)
+ __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S
new file mode 100644
index 0000000..8100e8a5
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S
@@ -0,0 +1,510 @@
+/* Function asinh vectorized with AVX-512.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ https://www.gnu.org/licenses/. */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ * Compute asinh(x) as log(x + sqrt(x*x + 1))
+ * using RSQRT instructions for starting the
+ * square root approximation, and small table lookups for log
+ * that map to AVX-512 permute instructions
+ *
+ * Special cases:
+ *
+ * asinh(NaN) = quiet NaN, and raise invalid exception
+ * asinh(INF) = that INF
+ * asinh(0) = that 0
+ *
+ */
+
+/* Offsets for data table __svml_dasinh_data_internal_avx512
+ */
+#define Log_tbl_H 0
+#define Log_tbl_L 128
+#define One 256
+#define AbsMask 320
+#define SmallThreshold 384
+#define Threshold 448
+#define LargeThreshold 512
+#define ca2 576
+#define ca1 640
+#define c4s 704
+#define c3s 768
+#define c2s 832
+#define c1s 896
+#define AddB5 960
+#define RcpBitMask 1024
+#define OneEighth 1088
+#define Four 1152
+#define poly_coeff9 1216
+#define poly_coeff8 1280
+#define poly_coeff7 1344
+#define poly_coeff6 1408
+#define poly_coeff5 1472
+#define poly_coeff4 1536
+#define poly_coeff3 1600
+#define poly_coeff2 1664
+#define poly_coeff1 1728
+#define L2H 1792
+#define L2L 1856
+
+#include <sysdep.h>
+
+ .text
+ .section .text.evex512,"ax",@progbits
+ENTRY(_ZGVeN8v_asinh_skx)
+ pushq %rbp
+ cfi_def_cfa_offset(16)
+ movq %rsp, %rbp
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+ andq $-64, %rsp
+ subq $192, %rsp
+ vmovaps %zmm0, %zmm3
+
+/* x^2 */
+ vmulpd {rn-sae}, %zmm3, %zmm3, %zmm14
+ vmovups One+__svml_dasinh_data_internal_avx512(%rip), %zmm9
+
+/* polynomial computation for small inputs */
+ vmovups ca2+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+ vmovups ca1+__svml_dasinh_data_internal_avx512(%rip), %zmm11
+
+/* not a very small input ? */
+ vmovups SmallThreshold+__svml_dasinh_data_internal_avx512(%rip), %zmm0
+
+/* A=max(x^2, 1); */
+ vmaxpd {sae}, %zmm14, %zmm9, %zmm4
+
+/* B=min(x^2, 1); */
+ vminpd {sae}, %zmm14, %zmm9, %zmm5
+ vfmadd231pd {rn-sae}, %zmm14, %zmm10, %zmm11
+
+/* 1+x^2 */
+ vaddpd {rn-sae}, %zmm9, %zmm14, %zmm8
+
+/* |input| */
+ vandpd AbsMask+__svml_dasinh_data_internal_avx512(%rip), %zmm3, %zmm1
+ vrsqrt14pd %zmm8, %zmm6
+ vcmppd $21, {sae}, %zmm0, %zmm1, %k2
+
+/* B_high */
+ vsubpd {rn-sae}, %zmm4, %zmm8, %zmm7
+
+/* sign bit */
+ vxorpd %zmm3, %zmm1, %zmm2
+ vmulpd {rn-sae}, %zmm14, %zmm11, %zmm4
+
+/* B_low */
+ vsubpd {rn-sae}, %zmm7, %zmm5, %zmm13
+ vmovups c2s+__svml_dasinh_data_internal_avx512(%rip), %zmm5
+ vmovups c1s+__svml_dasinh_data_internal_avx512(%rip), %zmm7
+
+/* polynomial computation for small inputs */
+ vfmadd213pd {rn-sae}, %zmm1, %zmm1, %zmm4
+
+/* (x^2)_low */
+ vmovaps %zmm3, %zmm15
+ vfmsub213pd {rn-sae}, %zmm14, %zmm3, %zmm15
+
+/* Sh ~sqrt(1+x^2) */
+ vmulpd {rn-sae}, %zmm6, %zmm8, %zmm14
+
+/* Yl = (x^2)_low + B_low */
+ vaddpd {rn-sae}, %zmm15, %zmm13, %zmm13
+
+/* very large inputs ? */
+ vmovups Threshold+__svml_dasinh_data_internal_avx512(%rip), %zmm15
+
+/* (Yh*R0)_low */
+ vfmsub213pd {rn-sae}, %zmm14, %zmm6, %zmm8
+ vcmppd $21, {sae}, %zmm15, %zmm1, %k1
+
+/* Sl = (Yh*R0)_low+(R0*Yl) */
+ vfmadd213pd {rn-sae}, %zmm8, %zmm6, %zmm13
+ vmovups LargeThreshold+__svml_dasinh_data_internal_avx512(%rip), %zmm8
+
+/* rel. error term: Eh=1-Sh*R0 */
+ vmovaps %zmm9, %zmm12
+ vfnmadd231pd {rn-sae}, %zmm14, %zmm6, %zmm12
+ vcmppd $22, {sae}, %zmm8, %zmm1, %k0
+
+/* rel. error term: Eh=(1-Sh*R0)-Sl*R0 */
+ vfnmadd231pd {rn-sae}, %zmm13, %zmm6, %zmm12
+
+/*
+ * sqrt(1+x^2) ~ Sh + Sl + Sh*Eh*poly_s
+ * poly_s = c1+c2*Eh+c3*Eh^2
+ */
+ vmovups c4s+__svml_dasinh_data_internal_avx512(%rip), %zmm6
+ vmovups c3s+__svml_dasinh_data_internal_avx512(%rip), %zmm8
+
+/* Sh*Eh */
+ vmulpd {rn-sae}, %zmm12, %zmm14, %zmm11
+ vfmadd231pd {rn-sae}, %zmm12, %zmm6, %zmm8
+
+/* Sh+x */
+ vaddpd {rn-sae}, %zmm1, %zmm14, %zmm6
+ kmovw %k0, %edx
+ vfmadd213pd {rn-sae}, %zmm5, %zmm12, %zmm8
+ vfmadd213pd {rn-sae}, %zmm7, %zmm12, %zmm8
+
+/* Xh */
+ vsubpd {rn-sae}, %zmm14, %zmm6, %zmm12
+
+/* Sl + Sh*Eh*poly_s */
+ vfmadd213pd {rn-sae}, %zmm13, %zmm8, %zmm11
+
+/* fixup for very large inputs */
+ vmovups OneEighth+__svml_dasinh_data_internal_avx512(%rip), %zmm8
+
+/* Xl */
+ vsubpd {rn-sae}, %zmm12, %zmm1, %zmm12
+
+/* Xin0+Sl+Sh*Eh*poly_s ~ x+sqrt(1+x^2) */
+ vaddpd {rn-sae}, %zmm11, %zmm6, %zmm10
+
+/* Sl_high */
+ vsubpd {rn-sae}, %zmm6, %zmm10, %zmm5
+ vmulpd {rn-sae}, %zmm8, %zmm1, %zmm10{%k1}
+
+/* Table lookups */
+ vmovups __svml_dasinh_data_internal_avx512(%rip), %zmm6
+
+/* Sl_l */
+ vsubpd {rn-sae}, %zmm5, %zmm11, %zmm7
+ vrcp14pd %zmm10, %zmm13
+
+/* Xin_low */
+ vaddpd {rn-sae}, %zmm12, %zmm7, %zmm14
+ vmovups Log_tbl_L+__svml_dasinh_data_internal_avx512(%rip), %zmm7
+ vmovups poly_coeff6+__svml_dasinh_data_internal_avx512(%rip), %zmm12
+
+/* round reciprocal to 1+4b mantissas */
+ vpaddq AddB5+__svml_dasinh_data_internal_avx512(%rip), %zmm13, %zmm11
+
+/* fixup for very large inputs */
+ vxorpd %zmm14, %zmm14, %zmm14{%k1}
+ vmovups poly_coeff5+__svml_dasinh_data_internal_avx512(%rip), %zmm13
+ vandpd RcpBitMask+__svml_dasinh_data_internal_avx512(%rip), %zmm11, %zmm15
+ vmovups poly_coeff7+__svml_dasinh_data_internal_avx512(%rip), %zmm11
+
+/* Prepare table index */
+ vpsrlq $48, %zmm15, %zmm5
+
+/* reduced argument for log(): (Rcp*Xin-1)+Rcp*Xin_low */
+ vfmsub231pd {rn-sae}, %zmm15, %zmm10, %zmm9
+
+/* exponents */
+ vgetexppd {sae}, %zmm15, %zmm8
+ vmovups Four+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+ vpermt2pd Log_tbl_H+64+__svml_dasinh_data_internal_avx512(%rip), %zmm5, %zmm6
+ vpermt2pd Log_tbl_L+64+__svml_dasinh_data_internal_avx512(%rip), %zmm5, %zmm7
+ vsubpd {rn-sae}, %zmm10, %zmm8, %zmm8{%k1}
+ vfmadd231pd {rn-sae}, %zmm15, %zmm14, %zmm9
+
+/* polynomials */
+ vmovups poly_coeff9+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+ vmovups poly_coeff8+__svml_dasinh_data_internal_avx512(%rip), %zmm5
+ vmovups poly_coeff4+__svml_dasinh_data_internal_avx512(%rip), %zmm14
+
+/* -K*L2H + Th */
+ vmovups L2H+__svml_dasinh_data_internal_avx512(%rip), %zmm15
+ vfmadd231pd {rn-sae}, %zmm9, %zmm10, %zmm5
+
+/* -K*L2L + Tl */
+ vmovups L2L+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+ vfnmadd231pd {rn-sae}, %zmm8, %zmm15, %zmm6
+ vfmadd213pd {rn-sae}, %zmm11, %zmm9, %zmm5
+ vfnmadd213pd {rn-sae}, %zmm7, %zmm10, %zmm8
+ vmovups poly_coeff3+__svml_dasinh_data_internal_avx512(%rip), %zmm7
+ vmovups poly_coeff1+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+
+/* R^2 */
+ vmulpd {rn-sae}, %zmm9, %zmm9, %zmm11
+ vfmadd213pd {rn-sae}, %zmm12, %zmm9, %zmm5
+ vfmadd213pd {rn-sae}, %zmm13, %zmm9, %zmm5
+ vfmadd213pd {rn-sae}, %zmm14, %zmm9, %zmm5
+ vfmadd213pd {rn-sae}, %zmm7, %zmm9, %zmm5
+ vmovups poly_coeff2+__svml_dasinh_data_internal_avx512(%rip), %zmm7
+ vfmadd213pd {rn-sae}, %zmm7, %zmm9, %zmm5
+ vfmadd213pd {rn-sae}, %zmm10, %zmm9, %zmm5
+
+/* Tl + R^2*Poly */
+ vfmadd213pd {rn-sae}, %zmm8, %zmm11, %zmm5
+
+/* R+Tl + R^2*Poly */
+ vaddpd {rn-sae}, %zmm9, %zmm5, %zmm9
+ vaddpd {rn-sae}, %zmm9, %zmm6, %zmm4{%k2}
+ vxorpd %zmm2, %zmm4, %zmm0
+ testl %edx, %edx
+
+/* Go to special inputs processing branch */
+ jne L(SPECIAL_VALUES_BRANCH)
+ # LOE rbx r12 r13 r14 r15 edx zmm0 zmm3
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+ movq %rbp, %rsp
+ popq %rbp
+ cfi_def_cfa(7, 8)
+ cfi_restore(6)
+ ret
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+ vmovups %zmm3, 64(%rsp)
+ vmovups %zmm0, 128(%rsp)
+ # LOE rbx r12 r13 r14 r15 edx zmm0
+
+ xorl %eax, %eax
+ # LOE rbx r12 r13 r14 r15 eax edx
+
+ vzeroupper
+ movq %r12, 16(%rsp)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
+ movl %eax, %r12d
+ movq %r13, 8(%rsp)
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
+ movl %edx, %r13d
+ movq %r14, (%rsp)
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+ btl %r12d, %r13d
+
+/* Call scalar math function */
+ jc L(SCALAR_MATH_CALL)
+ # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+ incl %r12d
+ cmpl $8, %r12d
+
+/* Check bits in range mask */
+ jl L(RANGEMASK_CHECK)
+ # LOE rbx r15 r12d r13d
+
+ movq 16(%rsp), %r12
+ cfi_restore(12)
+ movq 8(%rsp), %r13
+ cfi_restore(13)
+ movq (%rsp), %r14
+ cfi_restore(14)
+ vmovups 128(%rsp), %zmm0
+
+/* Go to exit */
+ jmp L(EXIT)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r12 r13 r14 r15 zmm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+ movl %r12d, %r14d
+ movsd 64(%rsp,%r14,8), %xmm0
+ call asinh@PLT
+ # LOE rbx r14 r15 r12d r13d xmm0
+
+ movsd %xmm0, 128(%rsp,%r14,8)
+
+/* Process special inputs in loop */
+ jmp L(SPECIAL_VALUES_LOOP)
+ # LOE rbx r15 r12d r13d
+END(_ZGVeN8v_asinh_skx)
+
+ .section .rodata, "a"
+ .align 64
+
+#ifdef __svml_dasinh_data_internal_avx512_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+ __declspec(align(64)) VUINT32 Log_tbl_H[16][2];
+ __declspec(align(64)) VUINT32 Log_tbl_L[16][2];
+ __declspec(align(64)) VUINT32 One[8][2];
+ __declspec(align(64)) VUINT32 AbsMask[8][2];
+ __declspec(align(64)) VUINT32 SmallThreshold[8][2];
+ __declspec(align(64)) VUINT32 Threshold[8][2];
+ __declspec(align(64)) VUINT32 LargeThreshold[8][2];
+ __declspec(align(64)) VUINT32 ca2[8][2];
+ __declspec(align(64)) VUINT32 ca1[8][2];
+ __declspec(align(64)) VUINT32 c4s[8][2];
+ __declspec(align(64)) VUINT32 c3s[8][2];
+ __declspec(align(64)) VUINT32 c2s[8][2];
+ __declspec(align(64)) VUINT32 c1s[8][2];
+ __declspec(align(64)) VUINT32 AddB5[8][2];
+ __declspec(align(64)) VUINT32 RcpBitMask[8][2];
+ __declspec(align(64)) VUINT32 OneEighth[8][2];
+ __declspec(align(64)) VUINT32 Four[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff9[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff8[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff7[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff6[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff5[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff4[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff3[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff2[8][2];
+ __declspec(align(64)) VUINT32 poly_coeff1[8][2];
+ __declspec(align(64)) VUINT32 L2H[8][2];
+ __declspec(align(64)) VUINT32 L2L[8][2];
+ } __svml_dasinh_data_internal_avx512;
+#endif
+__svml_dasinh_data_internal_avx512:
+ /*== Log_tbl_H ==*/
+ .quad 0x0000000000000000
+ .quad 0xbfaf0a30c0120000
+ .quad 0xbfbe27076e2b0000
+ .quad 0xbfc5ff3070a78000
+ .quad 0xbfcc8ff7c79a8000
+ .quad 0xbfd1675cababc000
+ .quad 0xbfd4618bc21c4000
+ .quad 0xbfd739d7f6bbc000
+ .quad 0xbfd9f323ecbf8000
+ .quad 0xbfdc8ff7c79a8000
+ .quad 0xbfdf128f5faf0000
+ .quad 0xbfe0be72e4252000
+ .quad 0xbfe1e85f5e704000
+ .quad 0xbfe307d7334f2000
+ .quad 0xbfe41d8fe8468000
+ .quad 0xbfe52a2d265bc000
+ /*== Log_tbl_L ==*/
+ .align 64
+ .quad 0x0000000000000000
+ .quad 0x3d53ab33d066d1d2
+ .quad 0x3d2a342c2af0003c
+ .quad 0xbd43d3c873e20a07
+ .quad 0xbd4a21ac25d81ef3
+ .quad 0x3d59f1fc63382a8f
+ .quad 0xbd5ec27d0b7b37b3
+ .quad 0xbd50069ce24c53fb
+ .quad 0xbd584bf2b68d766f
+ .quad 0xbd5a21ac25d81ef3
+ .quad 0xbd3bb2cd720ec44c
+ .quad 0xbd55056d312f7668
+ .quad 0xbd1a07bd8b34be7c
+ .quad 0x3d5e83c094debc15
+ .quad 0x3d5aa33736867a17
+ .quad 0xbd46abb9df22bc57
+ /*== One ==*/
+ .align 64
+ .quad 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000
+ /*== AbsMask ==*/
+ .align 64
+ .quad 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff
+ /*== SmallThreshold ==*/
+ .align 64
+ .quad 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000
+ /*== Threshold ==*/
+ .align 64
+ .quad 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000
+ /*== LargeThreshold ==*/
+ .align 64
+ .quad 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff
+ /*== ca2 ==*/
+ .align 64
+ .quad 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7
+ /*== ca1 ==*/
+ .align 64
+ .quad 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e
+ /*== c4s ==*/
+ .align 64
+ .quad 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612
+ /*== c3s ==*/
+ .align 64
+ .quad 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000
+ /*== c2s ==*/
+ .align 64
+ .quad 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000
+ /*== c1s ==*/
+ .align 64
+ .quad 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000
+ /*== AddB5 ==*/
+ .align 64
+ .quad 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000
+ /*== RcpBitMask ==*/
+ .align 64
+ .quad 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000
+ /*==OneEighth ==*/
+ .align 64
+ .quad 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000
+ /*== Four ==*/
+ .align 64
+ .quad 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000
+ /*== poly_coeff9 ==*/
+ .align 64
+ .quad 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368
+ /*== poly_coeff8 ==*/
+ .align 64
+ .quad 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778
+ /*== poly_coeff7 ==*/
+ .align 64
+ .quad 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9
+ /*== poly_coeff6 ==*/
+ .align 64
+ .quad 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1
+ /*== poly_coeff5 ==*/
+ .align 64
+ .quad 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736
+ /*== poly_coeff4 ==*/
+ .align 64
+ .quad 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af
+ /*== poly_coeff3 ==*/
+ .align 64
+ .quad 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65
+ /*== poly_coeff2 ==*/
+ .align 64
+ .quad 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1
+ /*== poly_coeff1 ==*/
+ .align 64
+ .quad 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000
+ /*== L2H = log(2)_high ==*/
+ .align 64
+ .quad 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000
+ /*== L2L = log(2)_low ==*/
+ .align 64
+ .quad 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000
+ .align 64
+ .type __svml_dasinh_data_internal_avx512,@object
+ .size __svml_dasinh_data_internal_avx512,.-__svml_dasinh_data_internal_avx512
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S
new file mode 100644
index 0000000..7dfd95e
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S
@@ -0,0 +1,20 @@
+/* AVX2 version of vectorized asinhf.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define _ZGVeN16v_asinhf _ZGVeN16v_asinhf_avx2_wrapper
+#include "../svml_s_asinhf16_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c
new file mode 100644
index 0000000..dc770a0
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c
@@ -0,0 +1,28 @@
+/* Multiple versions of vectorized asinhf, vector length is 16.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define SYMBOL_NAME _ZGVeN16v_asinhf
+#include "ifunc-mathvec-avx512-skx.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVeN16v_asinhf, __GI__ZGVeN16v_asinhf,
+ __redirect__ZGVeN16v_asinhf)
+ __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S
new file mode 100644
index 0000000..fc6a8e7
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S
@@ -0,0 +1,476 @@
+/* Function asinhf vectorized with AVX-512.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ https://www.gnu.org/licenses/. */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ * Compute asinh(x) as log(x + sqrt(x*x + 1))
+ * using RSQRT instructions for starting the
+ * square root approximation, and small table lookups for log
+ * that map to AVX-512 permute instructions
+ *
+ * Special cases:
+ *
+ * asinh(NaN) = quiet NaN, and raise invalid exception
+ * asinh(INF) = that INF
+ * asinh(0) = that 0
+ *
+ */
+
+/* Offsets for data table __svml_sasinh_data_internal_avx512
+ */
+#define Log_tbl_H 0
+#define Log_tbl_L 128
+#define One 256
+#define AbsMask 320
+#define SmallThreshold 384
+#define Threshold 448
+#define LargeThreshold 512
+#define ca1 576
+#define c2s 640
+#define c1s 704
+#define AddB5 768
+#define RcpBitMask 832
+#define OneEighth 896
+#define Four 960
+#define poly_coeff3 1024
+#define poly_coeff2 1088
+#define poly_coeff1 1152
+#define L2H 1216
+#define L2L 1280
+
+#include <sysdep.h>
+
+ .text
+ .section .text.exex512,"ax",@progbits
+ENTRY(_ZGVeN16v_asinhf_skx)
+ pushq %rbp
+ cfi_def_cfa_offset(16)
+ movq %rsp, %rbp
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+ andq $-64, %rsp
+ subq $192, %rsp
+ vmovaps %zmm0, %zmm10
+
+/* x^2 */
+ vmulps {rn-sae}, %zmm10, %zmm10, %zmm0
+ vmovups One+__svml_sasinh_data_internal_avx512(%rip), %zmm2
+
+/* polynomial computation for small inputs */
+ vmovups ca1+__svml_sasinh_data_internal_avx512(%rip), %zmm1
+
+/* not a very small input ? */
+ vmovups SmallThreshold+__svml_sasinh_data_internal_avx512(%rip), %zmm11
+
+/* 1+x^2 */
+ vaddps {rn-sae}, %zmm2, %zmm0, %zmm7
+
+/* |input| */
+ vandps AbsMask+__svml_sasinh_data_internal_avx512(%rip), %zmm10, %zmm12
+
+/* A=max(x^2, 1); */
+ vmaxps {sae}, %zmm0, %zmm2, %zmm14
+ vrsqrt14ps %zmm7, %zmm8
+
+/* B=min(x^2, 1); */
+ vminps {sae}, %zmm0, %zmm2, %zmm15
+ vcmpps $21, {sae}, %zmm11, %zmm12, %k2
+
+/* B_high */
+ vsubps {rn-sae}, %zmm14, %zmm7, %zmm9
+
+/* sign bit */
+ vxorps %zmm10, %zmm12, %zmm13
+
+/* Sh ~sqrt(1+x^2) */
+ vmulps {rn-sae}, %zmm8, %zmm7, %zmm6
+ vmovups LargeThreshold+__svml_sasinh_data_internal_avx512(%rip), %zmm14
+
+/* B_low */
+ vsubps {rn-sae}, %zmm9, %zmm15, %zmm3
+
+/* Sh+x */
+ vaddps {rn-sae}, %zmm12, %zmm6, %zmm15
+
+/* (Yh*R0)_low */
+ vfmsub213ps {rn-sae}, %zmm6, %zmm8, %zmm7
+ vmulps {rn-sae}, %zmm1, %zmm0, %zmm9
+ vcmpps $22, {sae}, %zmm14, %zmm12, %k0
+ vmovups c1s+__svml_sasinh_data_internal_avx512(%rip), %zmm1
+
+/* polynomial computation for small inputs */
+ vfmadd213ps {rn-sae}, %zmm12, %zmm12, %zmm9
+ kmovw %k0, %edx
+
+/* (x^2)_low */
+ vmovaps %zmm10, %zmm4
+ vfmsub213ps {rn-sae}, %zmm0, %zmm10, %zmm4
+
+/* Yl = (x^2)_low + B_low */
+ vaddps {rn-sae}, %zmm4, %zmm3, %zmm5
+
+/* rel. error term: Eh=1-Sh*R0 */
+ vmovaps %zmm2, %zmm0
+ vfnmadd231ps {rn-sae}, %zmm6, %zmm8, %zmm0
+
+/* Sl = (Yh*R0)_low+(R0*Yl) */
+ vfmadd213ps {rn-sae}, %zmm7, %zmm8, %zmm5
+
+/* very large inputs ? */
+ vmovups Threshold+__svml_sasinh_data_internal_avx512(%rip), %zmm7
+
+/* rel. error term: Eh=(1-Sh*R0)-Sl*R0 */
+ vfnmadd231ps {rn-sae}, %zmm5, %zmm8, %zmm0
+
+/* sqrt(1+x^2) ~ Sh + Sl + Sh*Eh*poly_s */
+ vmovups c2s+__svml_sasinh_data_internal_avx512(%rip), %zmm8
+ vcmpps $21, {sae}, %zmm7, %zmm12, %k1
+
+/* Sh*Eh */
+ vmulps {rn-sae}, %zmm0, %zmm6, %zmm4
+ vfmadd231ps {rn-sae}, %zmm0, %zmm8, %zmm1
+
+/* Sl + Sh*Eh*poly_s */
+ vfmadd213ps {rn-sae}, %zmm5, %zmm1, %zmm4
+
+/* Xh */
+ vsubps {rn-sae}, %zmm6, %zmm15, %zmm5
+
+/* fixup for very large inputs */
+ vmovups OneEighth+__svml_sasinh_data_internal_avx512(%rip), %zmm6
+
+/* Xin0+Sl+Sh*Eh*poly_s ~ x+sqrt(1+x^2) */
+ vaddps {rn-sae}, %zmm4, %zmm15, %zmm3
+
+/* Xl */
+ vsubps {rn-sae}, %zmm5, %zmm12, %zmm5
+
+/* Sl_high */
+ vsubps {rn-sae}, %zmm15, %zmm3, %zmm0
+ vmulps {rn-sae}, %zmm6, %zmm12, %zmm3{%k1}
+
+/* -K*L2H + Th */
+ vmovups L2H+__svml_sasinh_data_internal_avx512(%rip), %zmm15
+
+/* Sl_l */
+ vsubps {rn-sae}, %zmm0, %zmm4, %zmm1
+ vrcp14ps %zmm3, %zmm6
+
+/* Table lookups */
+ vmovups __svml_sasinh_data_internal_avx512(%rip), %zmm0
+
+/* Xin_low */
+ vaddps {rn-sae}, %zmm5, %zmm1, %zmm7
+
+/* round reciprocal to 1+4b mantissas */
+ vpaddd AddB5+__svml_sasinh_data_internal_avx512(%rip), %zmm6, %zmm4
+ vmovups poly_coeff1+__svml_sasinh_data_internal_avx512(%rip), %zmm5
+ vandps RcpBitMask+__svml_sasinh_data_internal_avx512(%rip), %zmm4, %zmm8
+
+/* fixup for very large inputs */
+ vxorps %zmm7, %zmm7, %zmm7{%k1}
+
+/* polynomial */
+ vmovups poly_coeff3+__svml_sasinh_data_internal_avx512(%rip), %zmm4
+
+/* reduced argument for log(): (Rcp*Xin-1)+Rcp*Xin_low */
+ vfmsub231ps {rn-sae}, %zmm8, %zmm3, %zmm2
+ vmovups Four+__svml_sasinh_data_internal_avx512(%rip), %zmm3
+
+/* exponents */
+ vgetexpps {sae}, %zmm8, %zmm1
+
+/* Prepare table index */
+ vpsrld $18, %zmm8, %zmm14
+ vfmadd231ps {rn-sae}, %zmm8, %zmm7, %zmm2
+ vmovups poly_coeff2+__svml_sasinh_data_internal_avx512(%rip), %zmm7
+ vsubps {rn-sae}, %zmm3, %zmm1, %zmm1{%k1}
+ vpermt2ps Log_tbl_H+64+__svml_sasinh_data_internal_avx512(%rip), %zmm14, %zmm0
+ vmovups Log_tbl_L+__svml_sasinh_data_internal_avx512(%rip), %zmm3
+ vfmadd231ps {rn-sae}, %zmm2, %zmm4, %zmm7
+ vfnmadd231ps {rn-sae}, %zmm1, %zmm15, %zmm0
+
+/* R^2 */
+ vmulps {rn-sae}, %zmm2, %zmm2, %zmm6
+ vfmadd213ps {rn-sae}, %zmm5, %zmm2, %zmm7
+ vpermt2ps Log_tbl_L+64+__svml_sasinh_data_internal_avx512(%rip), %zmm14, %zmm3
+
+/* -K*L2L + Tl */
+ vmovups L2L+__svml_sasinh_data_internal_avx512(%rip), %zmm14
+ vfnmadd213ps {rn-sae}, %zmm3, %zmm14, %zmm1
+
+/* Tl + R^2*Poly */
+ vfmadd213ps {rn-sae}, %zmm1, %zmm6, %zmm7
+
+/* R+Tl + R^2*Poly */
+ vaddps {rn-sae}, %zmm2, %zmm7, %zmm2
+ vaddps {rn-sae}, %zmm2, %zmm0, %zmm9{%k2}
+ vxorps %zmm13, %zmm9, %zmm0
+ testl %edx, %edx
+
+/* Go to special inputs processing branch */
+ jne L(SPECIAL_VALUES_BRANCH)
+ # LOE rbx r12 r13 r14 r15 edx zmm0 zmm10
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+ movq %rbp, %rsp
+ popq %rbp
+ cfi_def_cfa(7, 8)
+ cfi_restore(6)
+ ret
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+ vmovups %zmm10, 64(%rsp)
+ vmovups %zmm0, 128(%rsp)
+ # LOE rbx r12 r13 r14 r15 edx zmm0
+
+ xorl %eax, %eax
+ # LOE rbx r12 r13 r14 r15 eax edx
+
+ vzeroupper
+ movq %r12, 16(%rsp)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
+ movl %eax, %r12d
+ movq %r13, 8(%rsp)
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
+ movl %edx, %r13d
+ movq %r14, (%rsp)
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+ btl %r12d, %r13d
+
+/* Call scalar math function */
+ jc L(SCALAR_MATH_CALL)
+ # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+ incl %r12d
+ cmpl $16, %r12d
+
+/* Check bits in range mask */
+ jl L(RANGEMASK_CHECK)
+ # LOE rbx r15 r12d r13d
+
+ movq 16(%rsp), %r12
+ cfi_restore(12)
+ movq 8(%rsp), %r13
+ cfi_restore(13)
+ movq (%rsp), %r14
+ cfi_restore(14)
+ vmovups 128(%rsp), %zmm0
+
+/* Go to exit */
+ jmp L(EXIT)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r12 r13 r14 r15 zmm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+ movl %r12d, %r14d
+ movss 64(%rsp,%r14,4), %xmm0
+ call asinhf@PLT
+ # LOE rbx r14 r15 r12d r13d xmm0
+
+ movss %xmm0, 128(%rsp,%r14,4)
+
+/* Process special inputs in loop */
+ jmp L(SPECIAL_VALUES_LOOP)
+ # LOE rbx r15 r12d r13d
+END(_ZGVeN16v_asinhf_skx)
+
+ .section .rodata, "a"
+ .align 64
+
+#ifdef __svml_sasinh_data_internal_avx512_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+ __declspec(align(64)) VUINT32 Log_tbl_H[32][1];
+ __declspec(align(64)) VUINT32 Log_tbl_L[32][1];
+ __declspec(align(64)) VUINT32 One[16][1];
+ __declspec(align(64)) VUINT32 AbsMask[16][1];
+ __declspec(align(64)) VUINT32 SmallThreshold[16][1];
+ __declspec(align(64)) VUINT32 Threshold[16][1];
+ __declspec(align(64)) VUINT32 LargeThreshold[16][1];
+ __declspec(align(64)) VUINT32 ca1[16][1];
+ __declspec(align(64)) VUINT32 c2s[16][1];
+ __declspec(align(64)) VUINT32 c1s[16][1];
+ __declspec(align(64)) VUINT32 AddB5[16][1];
+ __declspec(align(64)) VUINT32 RcpBitMask[16][1];
+ __declspec(align(64)) VUINT32 OneEighth[16][1];
+ __declspec(align(64)) VUINT32 Four[16][1];
+ __declspec(align(64)) VUINT32 poly_coeff3[16][1];
+ __declspec(align(64)) VUINT32 poly_coeff2[16][1];
+ __declspec(align(64)) VUINT32 poly_coeff1[16][1];
+ __declspec(align(64)) VUINT32 L2H[16][1];
+ __declspec(align(64)) VUINT32 L2L[16][1];
+ } __svml_sasinh_data_internal_avx512;
+#endif
+__svml_sasinh_data_internal_avx512:
+ /*== Log_tbl_H ==*/
+ .long 0x00000000
+ .long 0xbcfc0000
+ .long 0xbd788000
+ .long 0xbdb78000
+ .long 0xbdf14000
+ .long 0xbe14a000
+ .long 0xbe300000
+ .long 0xbe4aa000
+ .long 0xbe648000
+ .long 0xbe7dc000
+ .long 0xbe8b4000
+ .long 0xbe974000
+ .long 0xbea31000
+ .long 0xbeae9000
+ .long 0xbeb9d000
+ .long 0xbec4d000
+ .long 0xbecfa000
+ .long 0xbeda2000
+ .long 0xbee48000
+ .long 0xbeeea000
+ .long 0xbef89000
+ .long 0xbf012800
+ .long 0xbf05f000
+ .long 0xbf0aa800
+ .long 0xbf0f4000
+ .long 0xbf13c800
+ .long 0xbf184000
+ .long 0xbf1ca000
+ .long 0xbf20f000
+ .long 0xbf252800
+ .long 0xbf295000
+ .long 0xbf2d6800
+ /*== Log_tbl_L ==*/
+ .align 64
+ .long 0x80000000
+ .long 0xb726c39e
+ .long 0x3839e7fe
+ .long 0xb7528ae5
+ .long 0x377891d5
+ .long 0xb8297c10
+ .long 0x37cf8f58
+ .long 0x3852b186
+ .long 0x35838656
+ .long 0xb80c36af
+ .long 0x38235454
+ .long 0xb862bae1
+ .long 0x37e87bc7
+ .long 0x37848150
+ .long 0x37202511
+ .long 0xb74e1b05
+ .long 0x385c1340
+ .long 0xb8777bcd
+ .long 0x36038656
+ .long 0xb7d40984
+ .long 0xb80f5faf
+ .long 0xb8254b4c
+ .long 0xb865c84a
+ .long 0x37f0b42d
+ .long 0xb83ebce1
+ .long 0xb83c2513
+ .long 0x37a332c4
+ .long 0x3779654f
+ .long 0x38602f73
+ .long 0x367449f8
+ .long 0xb7b4996f
+ .long 0xb800986b
+ /*== One ==*/
+ .align 64
+ .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
+ /*== AbsMask ==*/
+ .align 64
+ .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
+ /*== SmallThreshold ==*/
+ .align 64
+ .long 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000
+ /*== Threshold ==*/
+ .align 64
+ .long 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000
+ /*== LargeThreshold ==*/
+ .align 64
+ .long 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff
+ /*== ca1 ==*/
+ .align 64
+ .long 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE
+ /*== c2s ==*/
+ .align 64
+ .long 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000
+ /*== c1s ==*/
+ .align 64
+ .long 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000
+ /*== AddB5 ==*/
+ .align 64
+ .long 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000
+ /*== RcpBitMask ==*/
+ .align 64
+ .long 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000
+ /*==OneEighth ==*/
+ .align 64
+ .long 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000
+ /*== Four ==*/
+ .align 64
+ .long 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000
+ /*== poly_coeff3 ==*/
+ .align 64
+ .long 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810
+ /*== poly_coeff2 ==*/
+ .align 64
+ .long 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e
+ /*== poly_coeff1 ==*/
+ .align 64
+ .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000
+ /*== L2H = log(2)_high ==*/
+ .align 64
+ .long 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000
+ /*== L2L = log(2)_low ==*/
+ .align 64
+ .long 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4
+ .align 64
+ .type __svml_sasinh_data_internal_avx512,@object
+ .size __svml_sasinh_data_internal_avx512,.-__svml_sasinh_data_internal_avx512
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S
new file mode 100644
index 0000000..52e4d2f
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S
@@ -0,0 +1,20 @@
+/* SSE2 version of vectorized asinhf, vector length is 4.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define _ZGVbN4v_asinhf _ZGVbN4v_asinhf_sse2
+#include "../svml_s_asinhf4_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c
new file mode 100644
index 0000000..296d575
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c
@@ -0,0 +1,28 @@
+/* Multiple versions of vectorized asinhf, vector length is 4.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define SYMBOL_NAME _ZGVbN4v_asinhf
+#include "ifunc-mathvec-sse4_1.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVbN4v_asinhf, __GI__ZGVbN4v_asinhf,
+ __redirect__ZGVbN4v_asinhf)
+ __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S
new file mode 100644
index 0000000..1eeeb4f
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S
@@ -0,0 +1,509 @@
+/* Function asinhf vectorized with SSE4.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ https://www.gnu.org/licenses/. */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ * Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *
+ * Special cases:
+ *
+ * asinh(NaN) = quiet NaN, and raise invalid exception
+ * asinh(INF) = that INF
+ * asinh(0) = that 0
+ *
+ */
+
+/* Offsets for data table __svml_sasinh_data_internal
+ */
+#define SgnMask 0
+#define sOne 16
+#define sPoly 32
+#define iBrkValue 160
+#define iOffExpoMask 176
+#define sBigThreshold 192
+#define sC2 208
+#define sC3 224
+#define sHalf 240
+#define sLargestFinite 256
+#define sLittleThreshold 272
+#define sSign 288
+#define sThirtyOne 304
+#define sTopMask11 320
+#define sTopMask8 336
+#define XScale 352
+#define sLn2 368
+
+#include <sysdep.h>
+
+ .text
+ .section .text.sse4,"ax",@progbits
+ENTRY(_ZGVbN4v_asinhf_sse4)
+ subq $72, %rsp
+ cfi_def_cfa_offset(80)
+ movaps %xmm0, %xmm8
+
+/*
+ * Split X into high and low parts, XHi (<= 11 bits) and XLo (<= 13 bits)
+ * We could use either X or |X| here, but it doesn't seem to matter
+ */
+ movups sTopMask11+__svml_sasinh_data_internal(%rip), %xmm10
+ movaps %xmm8, %xmm2
+ andps %xmm8, %xmm10
+
+/*
+ * Compute X^2 = (XHi + XLo)^2 = XHi^2 + XLo * (X + XHi)
+ * The two parts are shifted off by around 11 bits. So even though
+ * the low bit will not in general be exact, it's near enough
+ */
+ movaps %xmm10, %xmm3
+ subps %xmm10, %xmm2
+ mulps %xmm10, %xmm3
+ addps %xmm8, %xmm10
+
+/* Load the constant 1 and a sign mask */
+ movups sOne+__svml_sasinh_data_internal(%rip), %xmm7
+
+/*
+ * Finally, express Y + W = X^2 + 1 accurately where Y has <= 8 bits.
+ * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
+ * as the dominant component in the compensated summation. Otherwise,
+ * if |X| >= 1, then since X2Hi only has 22 significant bits, the basic
+ * addition will be exact anyway until we get to |X| >= 2^24. But by
+ * that time the log function is well-conditioned enough that the
+ * rounding error doesn't matter. Hence we can treat 1 as dominant even
+ * if it literally isn't.
+ */
+ movaps %xmm7, %xmm11
+ movaps %xmm7, %xmm4
+ movups sTopMask8+__svml_sasinh_data_internal(%rip), %xmm12
+ addps %xmm3, %xmm11
+ mulps %xmm10, %xmm2
+ subps %xmm11, %xmm4
+ movaps %xmm12, %xmm0
+ addps %xmm3, %xmm4
+
+/*
+ * Unfortunately, we can still be in trouble if |X| <= 2^-5, since
+ * the absolute error 2^-(7+24)-ish in sqrt(1 + X^2) gets scaled up
+ * by 1/X and comes close to our threshold. Hence if |X| <= 2^-4,
+ * perform an alternative computation
+ * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
+ * X2 = X^2
+ */
+ addps %xmm2, %xmm3
+ addps %xmm2, %xmm4
+ andps %xmm11, %xmm0
+
+/*
+ * Compute R = 1/sqrt(Y + W) * (1 + d)
+ * Force R to <= 8 significant bits.
+ * This means that R * Y and R^2 * Y are exactly representable.
+ */
+ rsqrtps %xmm0, %xmm14
+ subps %xmm0, %xmm11
+ andps %xmm12, %xmm14
+ addps %xmm11, %xmm4
+
+/*
+ * Compute S = (Y/sqrt(Y + W)) * (1 + d)
+ * and T = (W/sqrt(Y + W)) * (1 + d)
+ * so that S + T = sqrt(Y + W) * (1 + d)
+ * S is exact, and the rounding error in T is OK.
+ */
+ mulps %xmm14, %xmm0
+ mulps %xmm14, %xmm4
+
+/*
+ * Get the absolute value of the input, since we will exploit antisymmetry
+ * and mostly assume X >= 0 in the core computation
+ */
+ movups SgnMask+__svml_sasinh_data_internal(%rip), %xmm6
+
+/*
+ * Compute e = -(2 * d + d^2)
+ * The first FMR is exact, and the rounding error in the other is acceptable
+ * since d and e are ~ 2^-8
+ */
+ movaps %xmm14, %xmm13
+ andps %xmm8, %xmm6
+
+/*
+ * Obtain sqrt(1 + X^2) - 1 in two pieces
+ * sqrt(1 + X^2) - 1
+ * = sqrt(Y + W) - 1
+ * = (S + T) * (1 + Corr) - 1
+ * = [S - 1] + [T + (S + T) * Corr]
+ * We need a compensated summation for the last part. We treat S - 1
+ * as the larger part; it certainly is until about X < 2^-4, and in that
+ * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
+ * Final sum is dTmp5 (hi) + dTmp7 (lo)
+ */
+ movaps %xmm0, %xmm1
+
+/*
+ * Check whether the input is finite, by checking |X| <= MaxFloat
+ * Otherwise set the rangemask so that the callout will get used.
+ * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
+ */
+ movaps %xmm6, %xmm9
+
+/*
+ * The following computation can go wrong for very large X, basically
+ * because X^2 overflows. But for large X we have
+ * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
+ * we can just later stick X back into the log and tweak up the exponent.
+ * Actually we scale X by 2^-30 and tweak the exponent up by 31,
+ * to stay in the safe range for the later log computation.
+ * Compute a flag now telling us when do do this.
+ */
+ movaps %xmm6, %xmm5
+ cmpnleps sLargestFinite+__svml_sasinh_data_internal(%rip), %xmm9
+ cmpltps sBigThreshold+__svml_sasinh_data_internal(%rip), %xmm5
+ mulps %xmm0, %xmm13
+ addps %xmm4, %xmm1
+ subps %xmm7, %xmm0
+ mulps %xmm4, %xmm14
+ movmskps %xmm9, %edx
+ movaps %xmm7, %xmm9
+
+/*
+ * Now 1 / (1 + d)
+ * = 1 / (1 + (sqrt(1 - e) - 1))
+ * = 1 / sqrt(1 - e)
+ * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 + ...
+ * So compute the first three nonconstant terms of that, so that
+ * we have a relative correction (1 + Corr) to apply to S etc.
+ * C1 = 1/2
+ * C2 = 3/8
+ * C3 = 5/16
+ */
+ movups sC3+__svml_sasinh_data_internal(%rip), %xmm15
+ subps %xmm13, %xmm9
+ movups sHalf+__svml_sasinh_data_internal(%rip), %xmm10
+ subps %xmm14, %xmm9
+
+/* sX2over2 = X^2/2 */
+ mulps %xmm10, %xmm3
+ mulps %xmm9, %xmm15
+
+/* sX46 = -X^4/4 + X^6/8 */
+ movaps %xmm3, %xmm2
+ movaps %xmm3, %xmm12
+
+/*
+ * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
+ * It's always safe to assume |X| is larger.
+ * This is the final 2-part argument to the log1p function
+ */
+ movaps %xmm6, %xmm14
+ addps sC2+__svml_sasinh_data_internal(%rip), %xmm15
+ mulps %xmm9, %xmm15
+ addps %xmm10, %xmm15
+ mulps %xmm15, %xmm9
+ mulps %xmm1, %xmm9
+
+/* Now multiplex to the case X = 2^-30 * input, Xl = sL = 0 in the "big" case. */
+ movups XScale+__svml_sasinh_data_internal(%rip), %xmm15
+ addps %xmm9, %xmm4
+ movaps %xmm4, %xmm11
+ addps %xmm0, %xmm11
+ subps %xmm11, %xmm0
+ addps %xmm0, %xmm4
+
+/* sX4over4 = X^4/4 */
+ movaps %xmm3, %xmm0
+ mulps %xmm3, %xmm0
+ mulps %xmm0, %xmm2
+ subps %xmm0, %xmm2
+
+/*
+ * Now we feed into the log1p code, using H in place of _VARG1 and
+ * also adding L into Xl.
+ * compute 1+x as high, low parts
+ */
+ movaps %xmm7, %xmm0
+
+/* sX46over2 = -X^4/8 + x^6/16 */
+ mulps %xmm2, %xmm10
+ movaps %xmm7, %xmm2
+ addps %xmm10, %xmm12
+ subps %xmm12, %xmm3
+ addps %xmm3, %xmm10
+
+/* Now multiplex the two possible computations */
+ movaps %xmm6, %xmm3
+ cmpleps sLittleThreshold+__svml_sasinh_data_internal(%rip), %xmm3
+ movaps %xmm3, %xmm13
+ andps %xmm3, %xmm12
+ andnps %xmm11, %xmm13
+ movaps %xmm3, %xmm1
+ orps %xmm12, %xmm13
+ andnps %xmm4, %xmm1
+ andps %xmm3, %xmm10
+ movaps %xmm6, %xmm4
+ orps %xmm10, %xmm1
+ addps %xmm13, %xmm14
+ mulps %xmm15, %xmm6
+ maxps %xmm14, %xmm0
+ minps %xmm14, %xmm2
+ subps %xmm14, %xmm4
+ movaps %xmm0, %xmm3
+ addps %xmm4, %xmm13
+ addps %xmm2, %xmm3
+ addps %xmm13, %xmm1
+ subps %xmm3, %xmm0
+ movaps %xmm5, %xmm4
+ andps %xmm5, %xmm3
+ andnps %xmm6, %xmm4
+ addps %xmm0, %xmm2
+
+/*
+ * Now resume the main code.
+ * reduction: compute r,n
+ */
+ movdqu iBrkValue+__svml_sasinh_data_internal(%rip), %xmm6
+ orps %xmm3, %xmm4
+ psubd %xmm6, %xmm4
+ movaps %xmm7, %xmm0
+ addps %xmm2, %xmm1
+ movdqu iOffExpoMask+__svml_sasinh_data_internal(%rip), %xmm2
+ pand %xmm4, %xmm2
+ psrad $23, %xmm4
+ cvtdq2ps %xmm4, %xmm3
+ pslld $23, %xmm4
+ andps %xmm5, %xmm1
+ paddd %xmm6, %xmm2
+ psubd %xmm4, %xmm0
+ mulps %xmm0, %xmm1
+
+/* polynomial evaluation */
+ subps %xmm7, %xmm2
+ movups sPoly+112+__svml_sasinh_data_internal(%rip), %xmm7
+ addps %xmm2, %xmm1
+ mulps %xmm1, %xmm7
+ movaps %xmm5, %xmm2
+
+/* Add 31 to the exponent in the "large" case to get log(2 * input) */
+ movups sThirtyOne+__svml_sasinh_data_internal(%rip), %xmm0
+ addps sPoly+96+__svml_sasinh_data_internal(%rip), %xmm7
+ addps %xmm3, %xmm0
+ mulps %xmm1, %xmm7
+ andnps %xmm0, %xmm2
+ andps %xmm5, %xmm3
+ orps %xmm3, %xmm2
+ addps sPoly+80+__svml_sasinh_data_internal(%rip), %xmm7
+
+/* final reconstruction */
+ mulps sLn2+__svml_sasinh_data_internal(%rip), %xmm2
+ mulps %xmm1, %xmm7
+
+/* Finally, reincorporate the original sign. */
+ movups sSign+__svml_sasinh_data_internal(%rip), %xmm0
+ andps %xmm8, %xmm0
+ addps sPoly+64+__svml_sasinh_data_internal(%rip), %xmm7
+ mulps %xmm1, %xmm7
+ addps sPoly+48+__svml_sasinh_data_internal(%rip), %xmm7
+ mulps %xmm1, %xmm7
+ addps sPoly+32+__svml_sasinh_data_internal(%rip), %xmm7
+ mulps %xmm1, %xmm7
+ addps sPoly+16+__svml_sasinh_data_internal(%rip), %xmm7
+ mulps %xmm1, %xmm7
+ addps sPoly+__svml_sasinh_data_internal(%rip), %xmm7
+ mulps %xmm1, %xmm7
+ mulps %xmm1, %xmm7
+ addps %xmm7, %xmm1
+ addps %xmm2, %xmm1
+ pxor %xmm1, %xmm0
+ testl %edx, %edx
+
+/* Go to special inputs processing branch */
+ jne L(SPECIAL_VALUES_BRANCH)
+ # LOE rbx rbp r12 r13 r14 r15 edx xmm0 xmm8
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+ addq $72, %rsp
+ cfi_def_cfa_offset(8)
+ ret
+ cfi_def_cfa_offset(80)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+ movups %xmm8, 32(%rsp)
+ movups %xmm0, 48(%rsp)
+ # LOE rbx rbp r12 r13 r14 r15 edx
+
+ xorl %eax, %eax
+ movq %r12, 16(%rsp)
+ cfi_offset(12, -64)
+ movl %eax, %r12d
+ movq %r13, 8(%rsp)
+ cfi_offset(13, -72)
+ movl %edx, %r13d
+ movq %r14, (%rsp)
+ cfi_offset(14, -80)
+ # LOE rbx rbp r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+ btl %r12d, %r13d
+
+/* Call scalar math function */
+ jc L(SCALAR_MATH_CALL)
+ # LOE rbx rbp r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+ incl %r12d
+ cmpl $4, %r12d
+
+/* Check bits in range mask */
+ jl L(RANGEMASK_CHECK)
+ # LOE rbx rbp r15 r12d r13d
+
+ movq 16(%rsp), %r12
+ cfi_restore(12)
+ movq 8(%rsp), %r13
+ cfi_restore(13)
+ movq (%rsp), %r14
+ cfi_restore(14)
+ movups 48(%rsp), %xmm0
+
+/* Go to exit */
+ jmp L(EXIT)
+ cfi_offset(12, -64)
+ cfi_offset(13, -72)
+ cfi_offset(14, -80)
+ # LOE rbx rbp r12 r13 r14 r15 xmm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+ movl %r12d, %r14d
+ movss 32(%rsp,%r14,4), %xmm0
+ call asinhf@PLT
+ # LOE rbx rbp r14 r15 r12d r13d xmm0
+
+ movss %xmm0, 48(%rsp,%r14,4)
+
+/* Process special inputs in loop */
+ jmp L(SPECIAL_VALUES_LOOP)
+ # LOE rbx rbp r15 r12d r13d
+END(_ZGVbN4v_asinhf_sse4)
+
+ .section .rodata, "a"
+ .align 16
+
+#ifdef __svml_sasinh_data_internal_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+ __declspec(align(16)) VUINT32 SgnMask[4][1];
+ __declspec(align(16)) VUINT32 sOne[4][1];
+ __declspec(align(16)) VUINT32 sPoly[8][4][1];
+ __declspec(align(16)) VUINT32 iBrkValue[4][1];
+ __declspec(align(16)) VUINT32 iOffExpoMask[4][1];
+ __declspec(align(16)) VUINT32 sBigThreshold[4][1];
+ __declspec(align(16)) VUINT32 sC2[4][1];
+ __declspec(align(16)) VUINT32 sC3[4][1];
+ __declspec(align(16)) VUINT32 sHalf[4][1];
+ __declspec(align(16)) VUINT32 sLargestFinite[4][1];
+ __declspec(align(16)) VUINT32 sLittleThreshold[4][1];
+ __declspec(align(16)) VUINT32 sSign[4][1];
+ __declspec(align(16)) VUINT32 sThirtyOne[4][1];
+ __declspec(align(16)) VUINT32 sTopMask11[4][1];
+ __declspec(align(16)) VUINT32 sTopMask8[4][1];
+ __declspec(align(16)) VUINT32 XScale[4][1];
+ __declspec(align(16)) VUINT32 sLn2[4][1];
+} __svml_sasinh_data_internal;
+#endif
+__svml_sasinh_data_internal:
+ /*== SgnMask ==*/
+ .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
+ /*== sOne = SP 1.0 ==*/
+ .align 16
+ .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
+ /*== sPoly[] = SP polynomial ==*/
+ .align 16
+ .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000 /* -5.0000000000000000000000000e-01 P0 */
+ .long 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94 /* 3.3333265781402587890625000e-01 P1 */
+ .long 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e /* -2.5004237890243530273437500e-01 P2 */
+ .long 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190 /* 2.0007920265197753906250000e-01 P3 */
+ .long 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37 /* -1.6472326219081878662109375e-01 P4 */
+ .long 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12 /* 1.4042308926582336425781250e-01 P5 */
+ .long 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3 /* -1.5122179687023162841796875e-01 P6 */
+ .long 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed /* 1.3820238411426544189453125e-01 P7 */
+ /*== iBrkValue = SP 2/3 ==*/
+ .align 16
+ .long 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab
+ /*== iOffExpoMask = SP significand mask ==*/
+ .align 16
+ .long 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff
+ /*== sBigThreshold ==*/
+ .align 16
+ .long 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000
+ /*== sC2 ==*/
+ .align 16
+ .long 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000
+ /*== sC3 ==*/
+ .align 16
+ .long 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000
+ /*== sHalf ==*/
+ .align 16
+ .long 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000
+ /*== sLargestFinite ==*/
+ .align 16
+ .long 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF
+ /*== sLittleThreshold ==*/
+ .align 16
+ .long 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000
+ /*== sSign ==*/
+ .align 16
+ .long 0x80000000, 0x80000000, 0x80000000, 0x80000000
+ /*== sThirtyOne ==*/
+ .align 16
+ .long 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000
+ /*== sTopMask11 ==*/
+ .align 16
+ .long 0xFFFFE000, 0xFFFFE000, 0xFFFFE000, 0xFFFFE000
+ /*== sTopMask8 ==*/
+ .align 16
+ .long 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000
+ /*== XScale ==*/
+ .align 16
+ .long 0x30800000, 0x30800000, 0x30800000, 0x30800000
+ /*== sLn2 = SP ln(2) ==*/
+ .align 16
+ .long 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218
+ .align 16
+ .type __svml_sasinh_data_internal,@object
+ .size __svml_sasinh_data_internal,.-__svml_sasinh_data_internal
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S
new file mode 100644
index 0000000..1a0e113
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S
@@ -0,0 +1,20 @@
+/* SSE version of vectorized asinhf, vector length is 8.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define _ZGVdN8v_asinhf _ZGVdN8v_asinhf_sse_wrapper
+#include "../svml_s_asinhf8_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c
new file mode 100644
index 0000000..d97097a
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c
@@ -0,0 +1,28 @@
+/* Multiple versions of vectorized asinhf, vector length is 8.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#define SYMBOL_NAME _ZGVdN8v_asinhf
+#include "ifunc-mathvec-avx2.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVdN8v_asinhf, __GI__ZGVdN8v_asinhf,
+ __redirect__ZGVdN8v_asinhf)
+ __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S
new file mode 100644
index 0000000..a966f53
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S
@@ -0,0 +1,457 @@
+/* Function asinhf vectorized with AVX2.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ https://www.gnu.org/licenses/. */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ * Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *
+ * Special cases:
+ *
+ * asinh(NaN) = quiet NaN, and raise invalid exception
+ * asinh(INF) = that INF
+ * asinh(0) = that 0
+ *
+ */
+
+/* Offsets for data table __svml_sasinh_data_internal
+ */
+#define SgnMask 0
+#define sOne 32
+#define sPoly 64
+#define iBrkValue 320
+#define iOffExpoMask 352
+#define sBigThreshold 384
+#define sC2 416
+#define sC3 448
+#define sHalf 480
+#define sLargestFinite 512
+#define sLittleThreshold 544
+#define sSign 576
+#define sThirtyOne 608
+#define sTopMask8 640
+#define XScale 672
+#define sLn2 704
+
+#include <sysdep.h>
+
+ .text
+ .section .text.avx2,"ax",@progbits
+ENTRY(_ZGVdN8v_asinhf_avx2)
+ pushq %rbp
+ cfi_def_cfa_offset(16)
+ movq %rsp, %rbp
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+ andq $-32, %rsp
+ subq $96, %rsp
+ vmovaps %ymm0, %ymm9
+
+/* Load the constant 1 and a sign mask */
+ vmovups sOne+__svml_sasinh_data_internal(%rip), %ymm8
+
+/* No need to split X when FMA is available in hardware. */
+ vmulps %ymm9, %ymm9, %ymm5
+ vmovups sTopMask8+__svml_sasinh_data_internal(%rip), %ymm1
+
+/*
+ * Finally, express Y + W = X^2 + 1 accurately where Y has <= 8 bits.
+ * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
+ * as the dominant component in the compensated summation. Otherwise,
+ * if |X| >= 1, then since X2Hi only has 22 significant bits, the basic
+ * addition will be exact anyway until we get to |X| >= 2^24. But by
+ * that time the log function is well-conditioned enough that the
+ * rounding error doesn't matter. Hence we can treat 1 as dominant even
+ * if it literally isn't.
+ */
+ vaddps %ymm5, %ymm8, %ymm13
+ vandps %ymm1, %ymm13, %ymm2
+ vmovaps %ymm9, %ymm4
+ vsubps %ymm13, %ymm8, %ymm11
+ vsubps %ymm2, %ymm13, %ymm15
+
+/*
+ * Compute R = 1/sqrt(Y + W) * (1 + d)
+ * Force R to <= 8 significant bits.
+ * This means that R * Y and R^2 * Y are exactly representable.
+ */
+ vrsqrtps %ymm2, %ymm0
+ vfmsub213ps %ymm5, %ymm9, %ymm4
+ vaddps %ymm11, %ymm5, %ymm12
+
+/*
+ * Get the absolute value of the input, since we will exploit antisymmetry
+ * and mostly assume X >= 0 in the core computation
+ */
+ vandps SgnMask+__svml_sasinh_data_internal(%rip), %ymm9, %ymm6
+
+/*
+ * Check whether the input is finite, by checking |X| <= MaxFloat
+ * Otherwise set the rangemask so that the callout will get used.
+ * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
+ */
+ vcmpnle_uqps sLargestFinite+__svml_sasinh_data_internal(%rip), %ymm6, %ymm10
+ vaddps %ymm12, %ymm4, %ymm14
+
+/*
+ * Unfortunately, we can still be in trouble if |X| <= 2^-5, since
+ * the absolute error 2^-(7+24)-ish in sqrt(1 + X^2) gets scaled up
+ * by 1/X and comes close to our threshold. Hence if |X| <= 2^-4,
+ * perform an alternative computation
+ * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
+ * X2 = X^2
+ */
+ vaddps %ymm4, %ymm5, %ymm4
+
+/*
+ * The following computation can go wrong for very large X, basically
+ * because X^2 overflows. But for large X we have
+ * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
+ * we can just later stick X back into the log and tweak up the exponent.
+ * Actually we scale X by 2^-30 and tweak the exponent up by 31,
+ * to stay in the safe range for the later log computation.
+ * Compute a flag now telling us when do do this.
+ */
+ vcmplt_oqps sBigThreshold+__svml_sasinh_data_internal(%rip), %ymm6, %ymm7
+ vaddps %ymm15, %ymm14, %ymm3
+
+/*
+ * Now 1 / (1 + d)
+ * = 1 / (1 + (sqrt(1 - e) - 1))
+ * = 1 / sqrt(1 - e)
+ * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 + ...
+ * So compute the first three nonconstant terms of that, so that
+ * we have a relative correction (1 + Corr) to apply to S etc.
+ * C1 = 1/2
+ * C2 = 3/8
+ * C3 = 5/16
+ */
+ vmovups sC3+__svml_sasinh_data_internal(%rip), %ymm12
+ vmovmskps %ymm10, %edx
+ vandps %ymm1, %ymm0, %ymm10
+
+/*
+ * Compute S = (Y/sqrt(Y + W)) * (1 + d)
+ * and T = (W/sqrt(Y + W)) * (1 + d)
+ * so that S + T = sqrt(Y + W) * (1 + d)
+ * S is exact, and the rounding error in T is OK.
+ */
+ vmulps %ymm10, %ymm2, %ymm15
+ vmulps %ymm3, %ymm10, %ymm14
+ vmovups sHalf+__svml_sasinh_data_internal(%rip), %ymm3
+ vsubps %ymm8, %ymm15, %ymm0
+
+/*
+ * Obtain sqrt(1 + X^2) - 1 in two pieces
+ * sqrt(1 + X^2) - 1
+ * = sqrt(Y + W) - 1
+ * = (S + T) * (1 + Corr) - 1
+ * = [S - 1] + [T + (S + T) * Corr]
+ * We need a compensated summation for the last part. We treat S - 1
+ * as the larger part; it certainly is until about X < 2^-4, and in that
+ * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
+ * Final sum is dTmp5 (hi) + dTmp7 (lo)
+ */
+ vaddps %ymm14, %ymm15, %ymm13
+
+/*
+ * Compute e = -(2 * d + d^2)
+ * The first FMR is exact, and the rounding error in the other is acceptable
+ * since d and e are ~ 2^-8
+ */
+ vmovaps %ymm8, %ymm11
+ vfnmadd231ps %ymm15, %ymm10, %ymm11
+ vfnmadd231ps %ymm14, %ymm10, %ymm11
+ vfmadd213ps sC2+__svml_sasinh_data_internal(%rip), %ymm11, %ymm12
+ vfmadd213ps %ymm3, %ymm11, %ymm12
+ vmulps %ymm12, %ymm11, %ymm1
+
+/* Now multiplex the two possible computations */
+ vcmple_oqps sLittleThreshold+__svml_sasinh_data_internal(%rip), %ymm6, %ymm11
+ vfmadd213ps %ymm14, %ymm13, %ymm1
+ vaddps %ymm0, %ymm1, %ymm2
+ vsubps %ymm2, %ymm0, %ymm10
+
+/* sX2over2 = X^2/2 */
+ vmulps %ymm4, %ymm3, %ymm0
+ vaddps %ymm10, %ymm1, %ymm1
+
+/* sX4over4 = X^4/4 */
+ vmulps %ymm0, %ymm0, %ymm5
+
+/* sX46 = -X^4/4 + X^6/8 */
+ vfmsub231ps %ymm0, %ymm5, %ymm5
+
+/* sX46over2 = -X^4/8 + x^6/16 */
+ vmulps %ymm5, %ymm3, %ymm3
+ vaddps %ymm3, %ymm0, %ymm5
+ vblendvps %ymm11, %ymm5, %ymm2, %ymm2
+ vsubps %ymm5, %ymm0, %ymm4
+
+/*
+ * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
+ * It's always safe to assume |X| is larger.
+ * This is the final 2-part argument to the log1p function
+ */
+ vaddps %ymm2, %ymm6, %ymm14
+
+/*
+ * Now resume the main code.
+ * reduction: compute r,n
+ */
+ vmovups iBrkValue+__svml_sasinh_data_internal(%rip), %ymm5
+ vaddps %ymm4, %ymm3, %ymm10
+
+/*
+ * Now we feed into the log1p code, using H in place of _VARG1 and
+ * also adding L into Xl.
+ * compute 1+x as high, low parts
+ */
+ vmaxps %ymm14, %ymm8, %ymm15
+ vminps %ymm14, %ymm8, %ymm0
+ vblendvps %ymm11, %ymm10, %ymm1, %ymm12
+ vsubps %ymm14, %ymm6, %ymm1
+ vaddps %ymm0, %ymm15, %ymm3
+
+/* Now multiplex to the case X = 2^-30 * input, Xl = sL = 0 in the "big" case. */
+ vmulps XScale+__svml_sasinh_data_internal(%rip), %ymm6, %ymm6
+ vaddps %ymm1, %ymm2, %ymm13
+ vsubps %ymm3, %ymm15, %ymm15
+ vaddps %ymm13, %ymm12, %ymm1
+ vaddps %ymm15, %ymm0, %ymm2
+ vblendvps %ymm7, %ymm3, %ymm6, %ymm0
+ vaddps %ymm2, %ymm1, %ymm4
+ vpsubd %ymm5, %ymm0, %ymm1
+ vpsrad $23, %ymm1, %ymm6
+ vpand iOffExpoMask+__svml_sasinh_data_internal(%rip), %ymm1, %ymm2
+ vmovups sPoly+224+__svml_sasinh_data_internal(%rip), %ymm1
+ vpslld $23, %ymm6, %ymm10
+ vpaddd %ymm5, %ymm2, %ymm13
+ vcvtdq2ps %ymm6, %ymm0
+ vpsubd %ymm10, %ymm8, %ymm12
+
+/* polynomial evaluation */
+ vsubps %ymm8, %ymm13, %ymm8
+
+/* Add 31 to the exponent in the "large" case to get log(2 * input) */
+ vaddps sThirtyOne+__svml_sasinh_data_internal(%rip), %ymm0, %ymm3
+ vandps %ymm7, %ymm4, %ymm11
+ vmulps %ymm12, %ymm11, %ymm14
+ vblendvps %ymm7, %ymm0, %ymm3, %ymm0
+ vaddps %ymm8, %ymm14, %ymm2
+ vfmadd213ps sPoly+192+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+ vfmadd213ps sPoly+160+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+ vfmadd213ps sPoly+128+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+ vfmadd213ps sPoly+96+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+ vfmadd213ps sPoly+64+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+ vfmadd213ps sPoly+32+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+ vfmadd213ps sPoly+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+ vmulps %ymm1, %ymm2, %ymm4
+ vfmadd213ps %ymm2, %ymm2, %ymm4
+
+/* final reconstruction */
+ vfmadd132ps sLn2+__svml_sasinh_data_internal(%rip), %ymm4, %ymm0
+
+/* Finally, reincorporate the original sign. */
+ vandps sSign+__svml_sasinh_data_internal(%rip), %ymm9, %ymm7
+ vxorps %ymm0, %ymm7, %ymm0
+ testl %edx, %edx
+
+/* Go to special inputs processing branch */
+ jne L(SPECIAL_VALUES_BRANCH)
+ # LOE rbx r12 r13 r14 r15 edx ymm0 ymm9
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+ movq %rbp, %rsp
+ popq %rbp
+ cfi_def_cfa(7, 8)
+ cfi_restore(6)
+ ret
+ cfi_def_cfa(6, 16)
+ cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+ vmovups %ymm9, 32(%rsp)
+ vmovups %ymm0, 64(%rsp)
+ # LOE rbx r12 r13 r14 r15 edx ymm0
+
+ xorl %eax, %eax
+ # LOE rbx r12 r13 r14 r15 eax edx
+
+ vzeroupper
+ movq %r12, 16(%rsp)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
+ movl %eax, %r12d
+ movq %r13, 8(%rsp)
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
+ movl %edx, %r13d
+ movq %r14, (%rsp)
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+ btl %r12d, %r13d
+
+/* Call scalar math function */
+ jc L(SCALAR_MATH_CALL)
+ # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+ incl %r12d
+ cmpl $8, %r12d
+
+/* Check bits in range mask */
+ jl L(RANGEMASK_CHECK)
+ # LOE rbx r15 r12d r13d
+
+ movq 16(%rsp), %r12
+ cfi_restore(12)
+ movq 8(%rsp), %r13
+ cfi_restore(13)
+ movq (%rsp), %r14
+ cfi_restore(14)
+ vmovups 64(%rsp), %ymm0
+
+/* Go to exit */
+ jmp L(EXIT)
+ /* DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
+ /* DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus) */
+ .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
+ # LOE rbx r12 r13 r14 r15 ymm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+ movl %r12d, %r14d
+ movss 32(%rsp,%r14,4), %xmm0
+ call asinhf@PLT
+ # LOE rbx r14 r15 r12d r13d xmm0
+
+ movss %xmm0, 64(%rsp,%r14,4)
+
+/* Process special inputs in loop */
+ jmp L(SPECIAL_VALUES_LOOP)
+ # LOE rbx r15 r12d r13d
+END(_ZGVdN8v_asinhf_avx2)
+
+ .section .rodata, "a"
+ .align 32
+
+#ifdef __svml_sasinh_data_internal_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+ __declspec(align(32)) VUINT32 SgnMask[8][1];
+ __declspec(align(32)) VUINT32 sOne[8][1];
+ __declspec(align(32)) VUINT32 sPoly[8][8][1];
+ __declspec(align(32)) VUINT32 iBrkValue[8][1];
+ __declspec(align(32)) VUINT32 iOffExpoMask[8][1];
+ __declspec(align(32)) VUINT32 sBigThreshold[8][1];
+ __declspec(align(32)) VUINT32 sC2[8][1];
+ __declspec(align(32)) VUINT32 sC3[8][1];
+ __declspec(align(32)) VUINT32 sHalf[8][1];
+ __declspec(align(32)) VUINT32 sLargestFinite[8][1];
+ __declspec(align(32)) VUINT32 sLittleThreshold[8][1];
+ __declspec(align(32)) VUINT32 sSign[8][1];
+ __declspec(align(32)) VUINT32 sThirtyOne[8][1];
+ __declspec(align(32)) VUINT32 sTopMask8[8][1];
+ __declspec(align(32)) VUINT32 XScale[8][1];
+ __declspec(align(32)) VUINT32 sLn2[8][1];
+} __svml_sasinh_data_internal;
+#endif
+__svml_sasinh_data_internal:
+ /*== SgnMask ==*/
+ .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
+ /*== sOne = SP 1.0 ==*/
+ .align 32
+ .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
+ /*== sPoly[] = SP polynomial ==*/
+ .align 32
+ .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000 /* -5.0000000000000000000000000e-01 P0 */
+ .long 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94 /* 3.3333265781402587890625000e-01 P1 */
+ .long 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e /* -2.5004237890243530273437500e-01 P2 */
+ .long 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190 /* 2.0007920265197753906250000e-01 P3 */
+ .long 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37 /* -1.6472326219081878662109375e-01 P4 */
+ .long 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12 /* 1.4042308926582336425781250e-01 P5 */
+ .long 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3 /* -1.5122179687023162841796875e-01 P6 */
+ .long 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed /* 1.3820238411426544189453125e-01 P7 */
+ /*== iBrkValue = SP 2/3 ==*/
+ .align 32
+ .long 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab
+ /*== iOffExpoMask = SP significand mask ==*/
+ .align 32
+ .long 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff
+ /*== sBigThreshold ==*/
+ .align 32
+ .long 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000
+ /*== sC2 ==*/
+ .align 32
+ .long 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000
+ /*== sC3 ==*/
+ .align 32
+ .long 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000
+ /*== sHalf ==*/
+ .align 32
+ .long 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000
+ /*== sLargestFinite ==*/
+ .align 32
+ .long 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF
+ /*== sLittleThreshold ==*/
+ .align 32
+ .long 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000
+ /*== sSign ==*/
+ .align 32
+ .long 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000
+ /*== sThirtyOne ==*/
+ .align 32
+ .long 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000
+ /*== sTopMask8 ==*/
+ .align 32
+ .long 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000
+ /*== XScale ==*/
+ .align 32
+ .long 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000
+ /*== sLn2 = SP ln(2) ==*/
+ .align 32
+ .long 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218
+ .align 32
+ .type __svml_sasinh_data_internal,@object
+ .size __svml_sasinh_data_internal,.-__svml_sasinh_data_internal
diff --git a/sysdeps/x86_64/fpu/svml_d_asinh2_core.S b/sysdeps/x86_64/fpu/svml_d_asinh2_core.S
new file mode 100644
index 0000000..60e3722
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_d_asinh2_core.S
@@ -0,0 +1,29 @@
+/* Function asinh vectorized with SSE2.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#include <sysdep.h>
+#include "svml_d_wrapper_impl.h"
+
+ .text
+ENTRY (_ZGVbN2v_asinh)
+WRAPPER_IMPL_SSE2 asinh
+END (_ZGVbN2v_asinh)
+
+#ifndef USE_MULTIARCH
+ libmvec_hidden_def (_ZGVbN2v_asinh)
+#endif
diff --git a/sysdeps/x86_64/fpu/svml_d_asinh4_core.S b/sysdeps/x86_64/fpu/svml_d_asinh4_core.S
new file mode 100644
index 0000000..c735001
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_d_asinh4_core.S
@@ -0,0 +1,29 @@
+/* Function asinh vectorized with AVX2, wrapper version.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#include <sysdep.h>
+#include "svml_d_wrapper_impl.h"
+
+ .text
+ENTRY (_ZGVdN4v_asinh)
+WRAPPER_IMPL_AVX _ZGVbN2v_asinh
+END (_ZGVdN4v_asinh)
+
+#ifndef USE_MULTIARCH
+ libmvec_hidden_def (_ZGVdN4v_asinh)
+#endif
diff --git a/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S b/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S
new file mode 100644
index 0000000..83aaa8c
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S
@@ -0,0 +1,25 @@
+/* Function asinh vectorized in AVX ISA as wrapper to SSE4 ISA version.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#include <sysdep.h>
+#include "svml_d_wrapper_impl.h"
+
+ .text
+ENTRY (_ZGVcN4v_asinh)
+WRAPPER_IMPL_AVX _ZGVbN2v_asinh
+END (_ZGVcN4v_asinh)
diff --git a/sysdeps/x86_64/fpu/svml_d_asinh8_core.S b/sysdeps/x86_64/fpu/svml_d_asinh8_core.S
new file mode 100644
index 0000000..9597975
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_d_asinh8_core.S
@@ -0,0 +1,25 @@
+/* Function asinh vectorized with AVX-512, wrapper to AVX2.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#include <sysdep.h>
+#include "svml_d_wrapper_impl.h"
+
+ .text
+ENTRY (_ZGVeN8v_asinh)
+WRAPPER_IMPL_AVX512 _ZGVdN4v_asinh
+END (_ZGVeN8v_asinh)
diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S
new file mode 100644
index 0000000..5b3d405
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S
@@ -0,0 +1,25 @@
+/* Function asinhf vectorized with AVX-512. Wrapper to AVX2 version.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#include <sysdep.h>
+#include "svml_s_wrapper_impl.h"
+
+ .text
+ENTRY (_ZGVeN16v_asinhf)
+WRAPPER_IMPL_AVX512 _ZGVdN8v_asinhf
+END (_ZGVeN16v_asinhf)
diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S
new file mode 100644
index 0000000..af44fa5
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S
@@ -0,0 +1,29 @@
+/* Function asinhf vectorized with SSE2, wrapper version.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#include <sysdep.h>
+#include "svml_s_wrapper_impl.h"
+
+ .text
+ENTRY (_ZGVbN4v_asinhf)
+WRAPPER_IMPL_SSE2 asinhf
+END (_ZGVbN4v_asinhf)
+
+#ifndef USE_MULTIARCH
+ libmvec_hidden_def (_ZGVbN4v_asinhf)
+#endif
diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S
new file mode 100644
index 0000000..3bd06d8
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S
@@ -0,0 +1,29 @@
+/* Function asinhf vectorized with AVX2, wrapper version.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#include <sysdep.h>
+#include "svml_s_wrapper_impl.h"
+
+ .text
+ENTRY (_ZGVdN8v_asinhf)
+WRAPPER_IMPL_AVX _ZGVbN4v_asinhf
+END (_ZGVdN8v_asinhf)
+
+#ifndef USE_MULTIARCH
+ libmvec_hidden_def (_ZGVdN8v_asinhf)
+#endif
diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S b/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S
new file mode 100644
index 0000000..f79616c
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S
@@ -0,0 +1,25 @@
+/* Function asinhf vectorized in AVX ISA as wrapper to SSE4 ISA version.
+ Copyright (C) 2021 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <https://www.gnu.org/licenses/>. */
+
+#include <sysdep.h>
+#include "svml_s_wrapper_impl.h"
+
+ .text
+ENTRY (_ZGVcN8v_asinhf)
+WRAPPER_IMPL_AVX _ZGVbN4v_asinhf
+END (_ZGVcN8v_asinhf)
diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c
new file mode 100644
index 0000000..da03528
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c
@@ -0,0 +1 @@
+#include "test-double-libmvec-asinh.c"
diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c
new file mode 100644
index 0000000..da03528
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c
@@ -0,0 +1 @@
+#include "test-double-libmvec-asinh.c"
diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c
new file mode 100644
index 0000000..da03528
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c
@@ -0,0 +1 @@
+#include "test-double-libmvec-asinh.c"
diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c
new file mode 100644
index 0000000..71e6b9f
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c
@@ -0,0 +1,3 @@
+#define LIBMVEC_TYPE double
+#define LIBMVEC_FUNC asinh
+#include "test-vector-abi-arg1.h"
diff --git a/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
index f53bb68..7611477 100644
--- a/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
@@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVbN2v_atanh)
VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVbN2v_acosh)
VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVbN2v_erf)
VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVbN2v_tanh)
+VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVbN2v_asinh)
#define VEC_INT_TYPE __m128i
diff --git a/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
index 0452c3d..1e0ee34 100644
--- a/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
@@ -48,6 +48,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVdN4v_atanh)
VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVdN4v_acosh)
VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVdN4v_erf)
VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVdN4v_tanh)
+VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVdN4v_asinh)
#ifndef __ILP32__
# define VEC_INT_TYPE __m256i
diff --git a/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
index 197d5af..17c43a7 100644
--- a/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
@@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVcN4v_atanh)
VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVcN4v_acosh)
VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVcN4v_erf)
VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVcN4v_tanh)
+VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVcN4v_asinh)
#define VEC_INT_TYPE __m128i
diff --git a/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
index e56ece6..1c6809e 100644
--- a/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
@@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVeN8v_atanh)
VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVeN8v_acosh)
VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVeN8v_erf)
VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVeN8v_tanh)
+VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVeN8v_asinh)
#ifndef __ILP32__
# define VEC_INT_TYPE __m512i
diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c
new file mode 100644
index 0000000..77e1838
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c
@@ -0,0 +1 @@
+#include "test-float-libmvec-asinhf.c"
diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c
new file mode 100644
index 0000000..77e1838
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c
@@ -0,0 +1 @@
+#include "test-float-libmvec-asinhf.c"
diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c
new file mode 100644
index 0000000..77e1838
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c
@@ -0,0 +1 @@
+#include "test-float-libmvec-asinhf.c"
diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c
new file mode 100644
index 0000000..3353754
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c
@@ -0,0 +1,3 @@
+#define LIBMVEC_TYPE float
+#define LIBMVEC_FUNC asinhf
+#include "test-vector-abi-arg1.h"
diff --git a/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
index abbebf9..e8ab188 100644
--- a/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
@@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVeN16v_atanhf)
VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVeN16v_acoshf)
VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVeN16v_erff)
VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVeN16v_tanhf)
+VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVeN16v_asinhf)
#define VEC_INT_TYPE __m512i
diff --git a/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
index ae1c8b9..a80c538 100644
--- a/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
@@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVbN4v_atanhf)
VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVbN4v_acoshf)
VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVbN4v_erff)
VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVbN4v_tanhf)
+VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVbN4v_asinhf)
#define VEC_INT_TYPE __m128i
diff --git a/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
index eb477a0..c3d1d59 100644
--- a/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
@@ -48,6 +48,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVdN8v_atanhf)
VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVdN8v_acoshf)
VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVdN8v_erff)
VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVdN8v_tanhf)
+VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVdN8v_asinhf)
/* Redefinition of wrapper to be compatible with _ZGVdN8vvv_sincosf. */
#undef VECTOR_WRAPPER_fFF
diff --git a/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
index 944f7f0..b7da0f5 100644
--- a/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
@@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVcN8v_atanhf)
VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVcN8v_acoshf)
VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVcN8v_erff)
VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVcN8v_tanhf)
+VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVcN8v_asinhf)
#define VEC_INT_TYPE __m128i