aboutsummaryrefslogtreecommitdiff
path: root/sysdeps/x86_64
diff options
context:
space:
mode:
authorNoah Goldstein <goldstein.w.n@gmail.com>2022-06-08 14:34:59 -0700
committerNoah Goldstein <goldstein.w.n@gmail.com>2022-06-08 17:07:34 -0700
commit2c9af8421d2b4a7fcce163e7bc81a118d22fd346 (patch)
tree4091d9f714fb5a2ae8e3bac3e7fe26268bfde82b /sysdeps/x86_64
parentaa13fd16183949bbc40b010552e2e42003ebee62 (diff)
downloadglibc-2c9af8421d2b4a7fcce163e7bc81a118d22fd346.zip
glibc-2c9af8421d2b4a7fcce163e7bc81a118d22fd346.tar.gz
glibc-2c9af8421d2b4a7fcce163e7bc81a118d22fd346.tar.bz2
x86: Fix page cross case in rawmemchr-avx2 [BZ #29234]
commit 6dcbb7d95dded20153b12d76d2f4e0ef0cda4f35 Author: Noah Goldstein <goldstein.w.n@gmail.com> Date: Mon Jun 6 21:11:33 2022 -0700 x86: Shrink code size of memchr-avx2.S Changed how the page cross case aligned string (rdi) in rawmemchr. This was incompatible with how `L(cross_page_continue)` expected the pointer to be aligned and would cause rawmemchr to read data start started before the beginning of the string. What it would read was in valid memory but could count CHAR matches resulting in an incorrect return value. This commit fixes that issue by essentially reverting the changes to the L(page_cross) case as they didn't really matter. Test cases added and all pass with the new code (and where confirmed to fail with the old code). Reviewed-by: H.J. Lu <hjl.tools@gmail.com>
Diffstat (limited to 'sysdeps/x86_64')
-rw-r--r--sysdeps/x86_64/multiarch/memchr-avx2.S16
1 files changed, 8 insertions, 8 deletions
diff --git a/sysdeps/x86_64/multiarch/memchr-avx2.S b/sysdeps/x86_64/multiarch/memchr-avx2.S
index 28a0128..c5a256e 100644
--- a/sysdeps/x86_64/multiarch/memchr-avx2.S
+++ b/sysdeps/x86_64/multiarch/memchr-avx2.S
@@ -409,19 +409,19 @@ L(cross_page_boundary):
computer return address if byte is found or adjusting length if it
is not and this is memchr. */
movq %rdi, %rcx
- /* Align data to VEC_SIZE. ALGN_PTR_REG is rcx for memchr and rdi for
- rawmemchr. */
- andq $-VEC_SIZE, %ALGN_PTR_REG
- VPCMPEQ (%ALGN_PTR_REG), %ymm0, %ymm1
+ /* Align data to VEC_SIZE - 1. ALGN_PTR_REG is rcx for memchr
+ and rdi for rawmemchr. */
+ orq $(VEC_SIZE - 1), %ALGN_PTR_REG
+ VPCMPEQ -(VEC_SIZE - 1)(%ALGN_PTR_REG), %ymm0, %ymm1
vpmovmskb %ymm1, %eax
# ifndef USE_AS_RAWMEMCHR
/* Calculate length until end of page (length checked for a match). */
- leal VEC_SIZE(%ALGN_PTR_REG), %esi
- subl %ERAW_PTR_REG, %esi
-# ifdef USE_AS_WMEMCHR
+ leaq 1(%ALGN_PTR_REG), %rsi
+ subq %RRAW_PTR_REG, %rsi
+# ifdef USE_AS_WMEMCHR
/* NB: Divide bytes by 4 to get wchar_t count. */
shrl $2, %esi
-# endif
+# endif
# endif
/* Remove the leading bytes. */
sarxl %ERAW_PTR_REG, %eax, %eax