/* strcmp with SSSE3 Copyright (C) 2010-2023 Free Software Foundation, Inc. This file is part of the GNU C Library. The GNU C Library is free software; you can redistribute it and/or modify it under the terms of the GNU Lesser General Public License as published by the Free Software Foundation; either version 2.1 of the License, or (at your option) any later version. The GNU C Library is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for more details. You should have received a copy of the GNU Lesser General Public License along with the GNU C Library; if not, see . */ #if IS_IN (libc) #include #include "asm-syntax.h" #define CFI_PUSH(REG) \ cfi_adjust_cfa_offset (4); \ cfi_rel_offset (REG, 0) #define CFI_POP(REG) \ cfi_adjust_cfa_offset (-4); \ cfi_restore (REG) #define PUSH(REG) pushl REG; CFI_PUSH (REG) #define POP(REG) popl REG; CFI_POP (REG) #ifdef USE_AS_STRNCMP # ifndef STRCMP # define STRCMP __strncmp_ssse3 # endif # define STR1 8 # define STR2 STR1+4 # define CNT STR2+4 # define RETURN POP (REM); ret; .p2align 4; CFI_PUSH (REM) # define UPDATE_STRNCMP_COUNTER \ /* calculate left number to compare */ \ mov $16, %esi; \ sub %ecx, %esi; \ cmp %esi, REM; \ jbe L(more8byteseq); \ sub %esi, REM # define FLAGS %ebx # define REM %ebp #elif defined USE_AS_STRCASECMP_L # include "locale-defines.h" # ifndef STRCMP # define STRCMP __strcasecmp_l_ssse3 # endif # ifdef PIC # define STR1 8 # else # define STR1 4 # endif # define STR2 STR1+4 # define LOCALE 12 /* Loaded before the adjustment. */ # ifdef PIC # define RETURN POP (%ebx); ret; .p2align 4; CFI_PUSH (%ebx) # else # define RETURN ret; .p2align 4 # endif # define UPDATE_STRNCMP_COUNTER # define FLAGS (%esp) # define NONASCII __strcasecmp_nonascii #elif defined USE_AS_STRNCASECMP_L # include "locale-defines.h" # ifndef STRCMP # define STRCMP __strncasecmp_l_ssse3 # endif # ifdef PIC # define STR1 12 # else # define STR1 8 # endif # define STR2 STR1+4 # define CNT STR2+4 # define LOCALE 16 /* Loaded before the adjustment. */ # ifdef PIC # define RETURN POP (REM); POP (%ebx); ret; \ .p2align 4; CFI_PUSH (%ebx); CFI_PUSH (REM) # else # define RETURN POP (REM); ret; .p2align 4; CFI_PUSH (REM) # endif # define UPDATE_STRNCMP_COUNTER \ /* calculate left number to compare */ \ mov $16, %esi; \ sub %ecx, %esi; \ cmp %esi, REM; \ jbe L(more8byteseq); \ sub %esi, REM # define FLAGS (%esp) # define REM %ebp # define NONASCII __strncasecmp_nonascii #else # ifndef STRCMP # define STRCMP __strcmp_ssse3 # endif # define STR1 4 # define STR2 STR1+4 # define RETURN ret; .p2align 4 # define UPDATE_STRNCMP_COUNTER # define FLAGS %ebx #endif .section .text.ssse3,"ax",@progbits #ifdef USE_AS_STRCASECMP_L ENTRY (__strcasecmp_ssse3) # ifdef PIC PUSH (%ebx) LOAD_PIC_REG(bx) movl __libc_tsd_LOCALE@GOTNTPOFF(%ebx), %eax movl %gs:(%eax), %eax # else movl %gs:__libc_tsd_LOCALE@NTPOFF, %eax # endif # if LOCALE_T___LOCALES != 0 || LC_CTYPE != 0 movl LOCALE_T___LOCALES+LC_CTYPE*4(%eax), %eax # else movl (%eax), %eax # endif testl $1, LOCALE_DATA_VALUES+_NL_CTYPE_NONASCII_CASE*SIZEOF_VALUES(%eax) # ifdef PIC je L(ascii) POP (%ebx) jmp __strcasecmp_nonascii # else jne __strcasecmp_nonascii jmp L(ascii) # endif END (__strcasecmp_ssse3) #endif #ifdef USE_AS_STRNCASECMP_L ENTRY (__strncasecmp_ssse3) # ifdef PIC PUSH (%ebx) LOAD_PIC_REG(bx) movl __libc_tsd_LOCALE@GOTNTPOFF(%ebx), %eax movl %gs:(%eax), %eax # else movl %gs:__libc_tsd_LOCALE@NTPOFF, %eax # endif # if LOCALE_T___LOCALES != 0 || LC_CTYPE != 0 movl LOCALE_T___LOCALES+LC_CTYPE*4(%eax), %eax # else movl (%eax), %eax # endif testl $1, LOCALE_DATA_VALUES+_NL_CTYPE_NONASCII_CASE*SIZEOF_VALUES(%eax) # ifdef PIC je L(ascii) POP (%ebx) jmp __strncasecmp_nonascii # else jne __strncasecmp_nonascii jmp L(ascii) # endif END (__strncasecmp_ssse3) #endif ENTRY (STRCMP) #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L movl LOCALE(%esp), %eax # if LOCALE_T___LOCALES != 0 || LC_CTYPE != 0 movl LOCALE_T___LOCALES+LC_CTYPE*4(%eax), %eax # else movl (%eax), %eax # endif testl $1, LOCALE_DATA_VALUES+_NL_CTYPE_NONASCII_CASE*SIZEOF_VALUES(%eax) jne NONASCII # ifdef PIC PUSH (%ebx) LOAD_PIC_REG(bx) # endif L(ascii): .section .rodata.cst16,"aM",@progbits,16 .align 16 .Lbelowupper: .quad 0x4040404040404040 .quad 0x4040404040404040 .Ltopupper: .quad 0x5b5b5b5b5b5b5b5b .quad 0x5b5b5b5b5b5b5b5b .Ltouppermask: .quad 0x2020202020202020 .quad 0x2020202020202020 .previous # ifdef PIC # define UCLOW_reg .Lbelowupper@GOTOFF(%ebx) # define UCHIGH_reg .Ltopupper@GOTOFF(%ebx) # define LCQWORD_reg .Ltouppermask@GOTOFF(%ebx) # else # define UCLOW_reg .Lbelowupper # define UCHIGH_reg .Ltopupper # define LCQWORD_reg .Ltouppermask # endif #endif #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L PUSH (REM) #endif movl STR1(%esp), %edx movl STR2(%esp), %eax #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L movl CNT(%esp), REM cmp $16, REM jb L(less16bytes_sncmp) #elif !defined USE_AS_STRCASECMP_L movzbl (%eax), %ecx cmpb %cl, (%edx) jne L(neq) cmpl $0, %ecx je L(eq) movzbl 1(%eax), %ecx cmpb %cl, 1(%edx) jne L(neq) cmpl $0, %ecx je L(eq) movzbl 2(%eax), %ecx cmpb %cl, 2(%edx) jne L(neq) cmpl $0, %ecx je L(eq) movzbl 3(%eax), %ecx cmpb %cl, 3(%edx) jne L(neq) cmpl $0, %ecx je L(eq) movzbl 4(%eax), %ecx cmpb %cl, 4(%edx) jne L(neq) cmpl $0, %ecx je L(eq) movzbl 5(%eax), %ecx cmpb %cl, 5(%edx) jne L(neq) cmpl $0, %ecx je L(eq) movzbl 6(%eax), %ecx cmpb %cl, 6(%edx) jne L(neq) cmpl $0, %ecx je L(eq) movzbl 7(%eax), %ecx cmpb %cl, 7(%edx) jne L(neq) cmpl $0, %ecx je L(eq) add $8, %edx add $8, %eax #endif movl %edx, %ecx and $0xfff, %ecx cmp $0xff0, %ecx ja L(crosspage) mov %eax, %ecx and $0xfff, %ecx cmp $0xff0, %ecx ja L(crosspage) pxor %xmm0, %xmm0 movlpd (%eax), %xmm1 movlpd (%edx), %xmm2 movhpd 8(%eax), %xmm1 movhpd 8(%edx), %xmm2 #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # define TOLOWER(reg1, reg2) \ movdqa reg1, %xmm5; \ movdqa reg2, %xmm7; \ movdqa UCHIGH_reg, %xmm6; \ pcmpgtb UCLOW_reg, %xmm5; \ pcmpgtb UCLOW_reg, %xmm7; \ pcmpgtb reg1, %xmm6; \ pand %xmm6, %xmm5; \ movdqa UCHIGH_reg, %xmm6; \ pcmpgtb reg2, %xmm6; \ pand %xmm6, %xmm7; \ pand LCQWORD_reg, %xmm5; \ por %xmm5, reg1; \ pand LCQWORD_reg, %xmm7; \ por %xmm7, reg2 TOLOWER (%xmm1, %xmm2) #else # define TOLOWER(reg1, reg2) #endif pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %ecx sub $0xffff, %ecx jnz L(less16bytes) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(eq) #endif add $16, %eax add $16, %edx L(crosspage): #if !defined USE_AS_STRCASECMP_L && !defined USE_AS_STRNCASECMP_L PUSH (FLAGS) #endif PUSH (%edi) PUSH (%esi) #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L pushl $0 cfi_adjust_cfa_offset (4) #endif #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cfi_remember_state #endif movl %edx, %edi movl %eax, %ecx and $0xf, %ecx and $0xf, %edi xor %ecx, %eax xor %edi, %edx #if !defined USE_AS_STRCASECMP_L && !defined USE_AS_STRNCASECMP_L xor FLAGS, FLAGS #endif cmp %edi, %ecx je L(ashr_0) ja L(bigger) orl $0x20, FLAGS xchg %edx, %eax xchg %ecx, %edi L(bigger): lea 15(%edi), %edi sub %ecx, %edi cmp $8, %edi jle L(ashr_less_8) cmp $14, %edi je L(ashr_15) cmp $13, %edi je L(ashr_14) cmp $12, %edi je L(ashr_13) cmp $11, %edi je L(ashr_12) cmp $10, %edi je L(ashr_11) cmp $9, %edi je L(ashr_10) L(ashr_less_8): je L(ashr_9) cmp $7, %edi je L(ashr_8) cmp $6, %edi je L(ashr_7) cmp $5, %edi je L(ashr_6) cmp $4, %edi je L(ashr_5) cmp $3, %edi je L(ashr_4) cmp $2, %edi je L(ashr_3) cmp $1, %edi je L(ashr_2) cmp $0, %edi je L(ashr_1) /* * The following cases will be handled by ashr_0 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(0~15) n(0~15) 15(15+ n-n) ashr_0 */ .p2align 4 L(ashr_0): mov $0xffff, %esi movdqa (%eax), %xmm1 pxor %xmm0, %xmm0 pcmpeqb %xmm1, %xmm0 #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L movdqa (%edx), %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm2, %xmm1 #else pcmpeqb (%edx), %xmm1 #endif psubb %xmm0, %xmm1 pmovmskb %xmm1, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi mov %ecx, %edi jne L(less32bytes) UPDATE_STRNCMP_COUNTER movl $0x10, FLAGS mov $0x10, %ecx pxor %xmm0, %xmm0 .p2align 4 L(loop_ashr_0): movdqa (%eax, %ecx), %xmm1 #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L movdqa (%edx, %ecx), %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 #else pcmpeqb %xmm1, %xmm0 pcmpeqb (%edx, %ecx), %xmm1 #endif psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx jmp L(loop_ashr_0) /* * The following cases will be handled by ashr_1 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(15) n -15 0(15 +(n-15) - n) ashr_1 */ .p2align 4 L(ashr_1): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $15, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -15(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $1, FLAGS lea 1(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_1): add $16, %edi jg L(nibble_ashr_1) L(gobble_ashr_1): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $1, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_1) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $1, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_1) .p2align 4 L(nibble_ashr_1): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xfffe, %esi jnz L(ashr_1_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $15, REM jbe L(ashr_1_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_1) .p2align 4 L(ashr_1_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $1, %xmm0 psrldq $1, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_2 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(14~15) n -14 1(15 +(n-14) - n) ashr_2 */ .p2align 4 L(ashr_2): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $14, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -14(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $2, FLAGS lea 2(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_2): add $16, %edi jg L(nibble_ashr_2) L(gobble_ashr_2): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $2, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_2) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $2, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_2) .p2align 4 L(nibble_ashr_2): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xfffc, %esi jnz L(ashr_2_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $14, REM jbe L(ashr_2_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_2) .p2align 4 L(ashr_2_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $2, %xmm0 psrldq $2, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_3 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(13~15) n -13 2(15 +(n-13) - n) ashr_3 */ .p2align 4 L(ashr_3): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $13, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -13(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $3, FLAGS lea 3(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_3): add $16, %edi jg L(nibble_ashr_3) L(gobble_ashr_3): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $3, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_3) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $3, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_3) .p2align 4 L(nibble_ashr_3): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xfff8, %esi jnz L(ashr_3_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $13, REM jbe L(ashr_3_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_3) .p2align 4 L(ashr_3_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $3, %xmm0 psrldq $3, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_4 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(12~15) n -12 3(15 +(n-12) - n) ashr_4 */ .p2align 4 L(ashr_4): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $12, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -12(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $4, FLAGS lea 4(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_4): add $16, %edi jg L(nibble_ashr_4) L(gobble_ashr_4): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $4, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_4) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $4, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_4) .p2align 4 L(nibble_ashr_4): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xfff0, %esi jnz L(ashr_4_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $12, REM jbe L(ashr_4_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_4) .p2align 4 L(ashr_4_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $4, %xmm0 psrldq $4, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_5 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(11~15) n -11 4(15 +(n-11) - n) ashr_5 */ .p2align 4 L(ashr_5): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $11, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -11(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $5, FLAGS lea 5(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_5): add $16, %edi jg L(nibble_ashr_5) L(gobble_ashr_5): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $5, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_5) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $5, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_5) .p2align 4 L(nibble_ashr_5): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xffe0, %esi jnz L(ashr_5_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $11, REM jbe L(ashr_5_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_5) .p2align 4 L(ashr_5_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $5, %xmm0 psrldq $5, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_6 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(10~15) n -10 5(15 +(n-10) - n) ashr_6 */ .p2align 4 L(ashr_6): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $10, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -10(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $6, FLAGS lea 6(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_6): add $16, %edi jg L(nibble_ashr_6) L(gobble_ashr_6): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $6, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_6) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $6, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_6) .p2align 4 L(nibble_ashr_6): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xffc0, %esi jnz L(ashr_6_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $10, REM jbe L(ashr_6_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_6) .p2align 4 L(ashr_6_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $6, %xmm0 psrldq $6, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_7 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(9~15) n - 9 6(15 +(n-9) - n) ashr_7 */ .p2align 4 L(ashr_7): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $9, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -9(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $7, FLAGS lea 8(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_7): add $16, %edi jg L(nibble_ashr_7) L(gobble_ashr_7): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $7, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_7) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $7, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_7) .p2align 4 L(nibble_ashr_7): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xff80, %esi jnz L(ashr_7_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $9, REM jbe L(ashr_7_exittail) #endif pxor %xmm0, %xmm0 pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_7) .p2align 4 L(ashr_7_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $7, %xmm0 psrldq $7, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_8 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(8~15) n - 8 7(15 +(n-8) - n) ashr_8 */ .p2align 4 L(ashr_8): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $8, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -8(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $8, FLAGS lea 8(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_8): add $16, %edi jg L(nibble_ashr_8) L(gobble_ashr_8): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $8, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_8) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $8, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_8) .p2align 4 L(nibble_ashr_8): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xff00, %esi jnz L(ashr_8_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $8, REM jbe L(ashr_8_exittail) #endif pxor %xmm0, %xmm0 pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_8) .p2align 4 L(ashr_8_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $8, %xmm0 psrldq $8, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_9 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(7~15) n - 7 8(15 +(n-7) - n) ashr_9 */ .p2align 4 L(ashr_9): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $7, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -7(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $9, FLAGS lea 9(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_9): add $16, %edi jg L(nibble_ashr_9) L(gobble_ashr_9): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $9, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_9) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $9, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_9) .p2align 4 L(nibble_ashr_9): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xfe00, %esi jnz L(ashr_9_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $7, REM jbe L(ashr_9_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_9) .p2align 4 L(ashr_9_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $9, %xmm0 psrldq $9, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_10 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(6~15) n - 6 9(15 +(n-6) - n) ashr_10 */ .p2align 4 L(ashr_10): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $6, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -6(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $10, FLAGS lea 10(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_10): add $16, %edi jg L(nibble_ashr_10) L(gobble_ashr_10): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $10, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_10) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $10, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_10) .p2align 4 L(nibble_ashr_10): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xfc00, %esi jnz L(ashr_10_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $6, REM jbe L(ashr_10_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_10) .p2align 4 L(ashr_10_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $10, %xmm0 psrldq $10, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_11 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(5~15) n - 5 10(15 +(n-5) - n) ashr_11 */ .p2align 4 L(ashr_11): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $5, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -5(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $11, FLAGS lea 11(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_11): add $16, %edi jg L(nibble_ashr_11) L(gobble_ashr_11): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $11, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_11) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $11, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_11) .p2align 4 L(nibble_ashr_11): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xf800, %esi jnz L(ashr_11_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $5, REM jbe L(ashr_11_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_11) .p2align 4 L(ashr_11_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $11, %xmm0 psrldq $11, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_12 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(4~15) n - 4 11(15 +(n-4) - n) ashr_12 */ .p2align 4 L(ashr_12): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $4, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -4(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $12, FLAGS lea 12(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_12): add $16, %edi jg L(nibble_ashr_12) L(gobble_ashr_12): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $12, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_12) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $12, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_12) .p2align 4 L(nibble_ashr_12): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xf000, %esi jnz L(ashr_12_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $4, REM jbe L(ashr_12_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_12) .p2align 4 L(ashr_12_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $12, %xmm0 psrldq $12, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_13 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(3~15) n - 3 12(15 +(n-3) - n) ashr_13 */ .p2align 4 L(ashr_13): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -3(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $13, FLAGS lea 13(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_13): add $16, %edi jg L(nibble_ashr_13) L(gobble_ashr_13): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $13, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_13) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $13, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_13) .p2align 4 L(nibble_ashr_13): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xe000, %esi jnz L(ashr_13_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $3, REM jbe L(ashr_13_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_13) .p2align 4 L(ashr_13_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $13, %xmm0 psrldq $13, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_14 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(2~15) n - 2 13(15 +(n-2) - n) ashr_14 */ .p2align 4 L(ashr_14): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $2, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -2(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $14, FLAGS lea 14(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_14): add $16, %edi jg L(nibble_ashr_14) L(gobble_ashr_14): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $14, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_14) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $14, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_14) .p2align 4 L(nibble_ashr_14): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0xc000, %esi jnz L(ashr_14_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $2, REM jbe L(ashr_14_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_14) .p2align 4 L(ashr_14_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $14, %xmm0 psrldq $14, %xmm3 jmp L(aftertail) /* * The following cases will be handled by ashr_14 * ecx(offset of esi) eax(offset of edi) relative offset corresponding case * n(1~15) n - 1 14(15 +(n-1) - n) ashr_15 */ .p2align 4 L(ashr_15): mov $0xffff, %esi pxor %xmm0, %xmm0 movdqa (%edx), %xmm2 movdqa (%eax), %xmm1 pcmpeqb %xmm1, %xmm0 pslldq $1, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm2 psubb %xmm0, %xmm2 pmovmskb %xmm2, %edi shr %cl, %esi shr %cl, %edi sub %edi, %esi lea -1(%ecx), %edi jnz L(less32bytes) UPDATE_STRNCMP_COUNTER movdqa (%edx), %xmm3 pxor %xmm0, %xmm0 mov $16, %ecx orl $15, FLAGS lea 15(%edx), %edi and $0xfff, %edi sub $0x1000, %edi .p2align 4 L(loop_ashr_15): add $16, %edi jg L(nibble_ashr_15) L(gobble_ashr_15): movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $15, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 add $16, %edi jg L(nibble_ashr_15) movdqa (%eax, %ecx), %xmm1 movdqa (%edx, %ecx), %xmm2 movdqa %xmm2, %xmm4 palignr $15, %xmm3, %xmm2 TOLOWER (%xmm1, %xmm2) pcmpeqb %xmm1, %xmm0 pcmpeqb %xmm2, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi sub $0xffff, %esi jnz L(exit) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $16, REM lea -16(REM), REM jbe L(more8byteseq) #endif add $16, %ecx movdqa %xmm4, %xmm3 jmp L(loop_ashr_15) .p2align 4 L(nibble_ashr_15): pcmpeqb %xmm3, %xmm0 pmovmskb %xmm0, %esi test $0x8000, %esi jnz L(ashr_15_exittail) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $1, REM jbe L(ashr_15_exittail) #endif pxor %xmm0, %xmm0 sub $0x1000, %edi jmp L(gobble_ashr_15) .p2align 4 L(ashr_15_exittail): movdqa (%eax, %ecx), %xmm1 psrldq $15, %xmm0 psrldq $15, %xmm3 jmp L(aftertail) .p2align 4 L(aftertail): TOLOWER (%xmm1, %xmm3) pcmpeqb %xmm3, %xmm1 psubb %xmm0, %xmm1 pmovmskb %xmm1, %esi not %esi L(exit): mov FLAGS, %edi and $0x1f, %edi lea -16(%edi, %ecx), %edi L(less32bytes): add %edi, %edx add %ecx, %eax testl $0x20, FLAGS jz L(ret2) xchg %eax, %edx .p2align 4 L(ret2): mov %esi, %ecx #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L addl $4, %esp cfi_adjust_cfa_offset (-4) #endif POP (%esi) POP (%edi) #if !defined USE_AS_STRCASECMP_L && !defined USE_AS_STRNCASECMP_L POP (FLAGS) #endif L(less16bytes): test %cl, %cl jz L(2next_8_bytes) test $0x01, %cl jnz L(Byte0) test $0x02, %cl jnz L(Byte1) test $0x04, %cl jnz L(Byte2) test $0x08, %cl jnz L(Byte3) test $0x10, %cl jnz L(Byte4) test $0x20, %cl jnz L(Byte5) test $0x40, %cl jnz L(Byte6) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $7, REM jbe L(eq) #endif movzbl 7(%eax), %ecx movzbl 7(%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN L(Byte0): #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $0, REM jbe L(eq) #endif movzbl (%eax), %ecx movzbl (%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN L(Byte1): #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $1, REM jbe L(eq) #endif movzbl 1(%eax), %ecx movzbl 1(%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN L(Byte2): #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $2, REM jbe L(eq) #endif movzbl 2(%eax), %ecx movzbl 2(%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN L(Byte3): #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $3, REM jbe L(eq) #endif movzbl 3(%eax), %ecx movzbl 3(%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN L(Byte4): #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $4, REM jbe L(eq) #endif movzbl 4(%eax), %ecx movzbl 4(%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN L(Byte5): #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $5, REM jbe L(eq) #endif movzbl 5(%eax), %ecx movzbl 5(%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN L(Byte6): #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $6, REM jbe L(eq) #endif movzbl 6(%eax), %ecx movzbl 6(%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN L(2next_8_bytes): add $8, %eax add $8, %edx #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $8, REM lea -8(REM), REM jbe L(eq) #endif test $0x01, %ch jnz L(Byte0) test $0x02, %ch jnz L(Byte1) test $0x04, %ch jnz L(Byte2) test $0x08, %ch jnz L(Byte3) test $0x10, %ch jnz L(Byte4) test $0x20, %ch jnz L(Byte5) test $0x40, %ch jnz L(Byte6) #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L cmp $7, REM jbe L(eq) #endif movzbl 7(%eax), %ecx movzbl 7(%edx), %eax #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%eax,4), %eax # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%eax,4), %eax # endif #endif sub %ecx, %eax RETURN #ifdef USE_AS_STRNCMP L(neq_sncmp): #endif L(neq): mov $1, %eax ja L(neq_bigger) neg %eax L(neq_bigger): #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L addl $4, %esp cfi_adjust_cfa_offset (-4) #endif #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L POP (REM) #endif #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC POP (%ebx) # endif #endif ret #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L .p2align 4 cfi_restore_state L(more8byteseq): # ifdef USE_AS_STRNCASECMP_L addl $4, %esp cfi_adjust_cfa_offset (-4) # endif POP (%esi) POP (%edi) # ifdef USE_AS_STRNCMP POP (FLAGS) # endif #endif #ifdef USE_AS_STRNCMP L(eq_sncmp): #endif L(eq): #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L POP (REM) #endif #if defined USE_AS_STRCASECMP_L || defined USE_AS_STRNCASECMP_L # ifdef PIC POP (%ebx) # endif #endif xorl %eax, %eax ret #if defined USE_AS_STRNCMP || defined USE_AS_STRNCASECMP_L .p2align 4 # if defined USE_AS_STRNCASECMP_L && defined PIC CFI_PUSH (%ebx) # endif CFI_PUSH (REM) L(less16bytes_sncmp): # ifdef USE_AS_STRNCASECMP_L PUSH (%esi) # endif test REM, REM jz L(eq_sncmp) movzbl (%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl (%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, (%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $1, REM je L(eq_sncmp) movzbl 1(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 1(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 1(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $2, REM je L(eq_sncmp) movzbl 2(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 2(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 2(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $3, REM je L(eq_sncmp) movzbl 3(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 3(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 3(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $4, REM je L(eq_sncmp) movzbl 4(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 4(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 4(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $5, REM je L(eq_sncmp) movzbl 5(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 5(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 5(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $6, REM je L(eq_sncmp) movzbl 6(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 6(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 6(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $7, REM je L(eq_sncmp) movzbl 7(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 7(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 7(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $8, REM je L(eq_sncmp) movzbl 8(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 8(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 8(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $9, REM je L(eq_sncmp) movzbl 9(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 9(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 9(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $10, REM je L(eq_sncmp) movzbl 10(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 10(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 10(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $11, REM je L(eq_sncmp) movzbl 11(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 11(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 11(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $12, REM je L(eq_sncmp) movzbl 12(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 12(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 12(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $13, REM je L(eq_sncmp) movzbl 13(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 13(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 13(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $14, REM je L(eq_sncmp) movzbl 14(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 14(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 14(%edx) # endif jne L(neq_sncmp) test %cl, %cl je L(eq_sncmp) cmp $15, REM je L(eq_sncmp) movzbl 15(%eax), %ecx # ifdef USE_AS_STRNCASECMP_L movzbl 15(%edx), %esi # ifdef PIC movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower@GOTOFF+128*4(%ebx,%esi,4), %esi # else movl _nl_C_LC_CTYPE_tolower+128*4(,%ecx,4), %ecx movl _nl_C_LC_CTYPE_tolower+128*4(,%esi,4), %esi # endif cmpl %ecx, %esi # else cmpb %cl, 15(%edx) # endif jne L(neq_sncmp) # ifdef USE_AS_STRNCASECMP_L L(eq_sncmp): POP (%esi) # endif POP (REM) # if defined USE_AS_STRNCASECMP_L && defined PIC POP (%ebx) # endif xor %eax, %eax ret # ifdef USE_AS_STRNCASECMP_L .p2align 4 # ifdef PIC CFI_PUSH (%ebx) # endif CFI_PUSH (REM) CFI_PUSH (%esi) L(neq_sncmp): mov $1, %eax mov $-1, %edx cmovna %edx, %eax POP (%esi) POP (REM) # ifdef PIC POP (%ebx) # endif ret # endif #endif END (STRCMP) #endif