diff options
author | Noah Goldstein <goldstein.w.n@gmail.com> | 2022-06-08 14:34:59 -0700 |
---|---|---|
committer | Noah Goldstein <goldstein.w.n@gmail.com> | 2022-06-08 17:07:34 -0700 |
commit | 2c9af8421d2b4a7fcce163e7bc81a118d22fd346 (patch) | |
tree | 4091d9f714fb5a2ae8e3bac3e7fe26268bfde82b /sysdeps/x86_64 | |
parent | aa13fd16183949bbc40b010552e2e42003ebee62 (diff) | |
download | glibc-2c9af8421d2b4a7fcce163e7bc81a118d22fd346.zip glibc-2c9af8421d2b4a7fcce163e7bc81a118d22fd346.tar.gz glibc-2c9af8421d2b4a7fcce163e7bc81a118d22fd346.tar.bz2 |
x86: Fix page cross case in rawmemchr-avx2 [BZ #29234]
commit 6dcbb7d95dded20153b12d76d2f4e0ef0cda4f35
Author: Noah Goldstein <goldstein.w.n@gmail.com>
Date: Mon Jun 6 21:11:33 2022 -0700
x86: Shrink code size of memchr-avx2.S
Changed how the page cross case aligned string (rdi) in
rawmemchr. This was incompatible with how
`L(cross_page_continue)` expected the pointer to be aligned and
would cause rawmemchr to read data start started before the
beginning of the string. What it would read was in valid memory
but could count CHAR matches resulting in an incorrect return
value.
This commit fixes that issue by essentially reverting the changes to
the L(page_cross) case as they didn't really matter.
Test cases added and all pass with the new code (and where confirmed
to fail with the old code).
Reviewed-by: H.J. Lu <hjl.tools@gmail.com>
Diffstat (limited to 'sysdeps/x86_64')
-rw-r--r-- | sysdeps/x86_64/multiarch/memchr-avx2.S | 16 |
1 files changed, 8 insertions, 8 deletions
diff --git a/sysdeps/x86_64/multiarch/memchr-avx2.S b/sysdeps/x86_64/multiarch/memchr-avx2.S index 28a0128..c5a256e 100644 --- a/sysdeps/x86_64/multiarch/memchr-avx2.S +++ b/sysdeps/x86_64/multiarch/memchr-avx2.S @@ -409,19 +409,19 @@ L(cross_page_boundary): computer return address if byte is found or adjusting length if it is not and this is memchr. */ movq %rdi, %rcx - /* Align data to VEC_SIZE. ALGN_PTR_REG is rcx for memchr and rdi for - rawmemchr. */ - andq $-VEC_SIZE, %ALGN_PTR_REG - VPCMPEQ (%ALGN_PTR_REG), %ymm0, %ymm1 + /* Align data to VEC_SIZE - 1. ALGN_PTR_REG is rcx for memchr + and rdi for rawmemchr. */ + orq $(VEC_SIZE - 1), %ALGN_PTR_REG + VPCMPEQ -(VEC_SIZE - 1)(%ALGN_PTR_REG), %ymm0, %ymm1 vpmovmskb %ymm1, %eax # ifndef USE_AS_RAWMEMCHR /* Calculate length until end of page (length checked for a match). */ - leal VEC_SIZE(%ALGN_PTR_REG), %esi - subl %ERAW_PTR_REG, %esi -# ifdef USE_AS_WMEMCHR + leaq 1(%ALGN_PTR_REG), %rsi + subq %RRAW_PTR_REG, %rsi +# ifdef USE_AS_WMEMCHR /* NB: Divide bytes by 4 to get wchar_t count. */ shrl $2, %esi -# endif +# endif # endif /* Remove the leading bytes. */ sarxl %ERAW_PTR_REG, %eax, %eax |