/* Optimized strstr implementation for PowerPC64/POWER7.
Copyright (C) 2015-2016 Free Software Foundation, Inc.
This file is part of the GNU C Library.
The GNU C Library is free software; you can redistribute it and/or
modify it under the terms of the GNU Lesser General Public
License as published by the Free Software Foundation; either
version 2.1 of the License, or (at your option) any later version.
The GNU C Library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
License along with the GNU C Library; if not, see
. */
#include
/* Char * [r3] strstr (char *s [r3], char * pat[r4]) */
/* The performance gain is obtained using aligned memory access, load
* doubleword and usage of cmpb instruction for quicker comparison. */
#define ITERATIONS 64
#ifndef STRLEN
/* For builds with no IFUNC support, local calls should be made to internal
GLIBC symbol (created by libc_hidden_builtin_def). */
# ifdef SHARED
# define STRLEN __GI_strlen
# else
# define STRLEN strlen
# endif
#endif
#ifndef STRNLEN
/* For builds with no IFUNC support, local calls should be made to internal
GLIBC symbol (created by libc_hidden_builtin_def). */
# ifdef SHARED
# define STRNLEN __GI_strnlen
# else
# define STRNLEN __strnlen
# endif
#endif
#ifndef STRCHR
# ifdef SHARED
# define STRCHR __GI_strchr
# else
# define STRCHR strchr
# endif
#endif
#define FRAMESIZE (FRAME_MIN_SIZE+32)
.machine power7
EALIGN (strstr, 4, 0)
CALL_MCOUNT 2
mflr r0 /* Load link register LR to r0. */
std r31, -8(r1) /* Save callers register r31. */
std r30, -16(r1) /* Save callers register r30. */
std r29, -24(r1) /* Save callers register r29. */
std r28, -32(r1) /* Save callers register r28. */
std r0, 16(r1) /* Store the link register. */
cfi_offset(r31, -8)
cfi_offset(r30, -16)
cfi_offset(r28, -32)
cfi_offset(r29, -24)
cfi_offset(lr, 16)
stdu r1, -FRAMESIZE(r1) /* Create the stack frame. */
cfi_adjust_cfa_offset(FRAMESIZE)
dcbt 0, r3
dcbt 0, r4
cmpdi cr7, r3, 0
beq cr7, L(retnull)
cmpdi cr7, r4, 0
beq cr7, L(retnull)
mr r29, r3
mr r30, r4
mr r3, r4
bl STRLEN
nop
cmpdi cr7, r3, 0 /* If search str is null. */
beq cr7, L(ret_r3)
mr r31, r3
mr r4, r3
mr r3, r29
bl STRNLEN
nop
cmpd cr7, r3, r31 /* If len(r3) < len(r4). */
blt cr7, L(retnull)
mr r3, r29
lbz r4, 0(r30)
bl STRCHR
nop
mr r11, r3
/* If first char of search str is not present. */
cmpdi cr7, r3, 0
ble cr7, L(end)
/* Reg r28 is used to count the number of iterations. */
li r28, 0
rldicl r8, r3, 0, 52 /* Page cross check. */
cmpldi cr7, r8, 4096-16
bgt cr7, L(bytebybyte)
rldicl r8, r30, 0, 52
cmpldi cr7, r8, 4096-16
bgt cr7, L(bytebybyte)
/* If len(r4) < 8 handle in a different way. */
/* Shift position based on null and use cmpb. */
cmpdi cr7, r31, 8
blt cr7, L(lessthan8)
/* Len(r4) >= 8 reaches here. */
mr r8, r3 /* Save r3 for future use. */
mr r4, r30 /* Restore r4. */
li r0, 0
rlwinm r10, r30, 3, 26, 28 /* Calculate padding in bits. */
clrrdi r4, r4, 3 /* Make r4 aligned to 8. */
ld r6, 0(r4)
addi r4, r4, 8
cmpdi cr7, r10, 0 /* Check if its already aligned? */
beq cr7, L(begin1)
#ifdef __LITTLE_ENDIAN__
srd r6, r6, r10 /* Discard unwanted bits. */
#else
sld r6, r6, r10
#endif
ld r9, 0(r4)
subfic r10, r10, 64
#ifdef __LITTLE_ENDIAN__
sld r9, r9, r10 /* Discard unwanted bits. */
#else
srd r9, r9, r10
#endif
or r6, r6, r9 /* Form complete search str. */
L(begin1):
mr r29, r6
rlwinm r10, r3, 3, 26, 28
clrrdi r3, r3, 3
ld r5, 0(r3)
cmpb r9, r0, r6 /* Check if input has null. */
cmpdi cr7, r9, 0
bne cr7, L(return3)
cmpb r9, r0, r5 /* Check if input has null. */
#ifdef __LITTLE_ENDIAN__
srd r9, r9, r10
#else
sld r9, r9, r10
#endif
cmpdi cr7, r9, 0
bne cr7, L(retnull)
li r12, -8 /* Shift values. */
li r11, 72 /* Shift values. */
cmpdi cr7, r10, 0
beq cr7, L(nextbyte1)
mr r12, r10
addi r12, r12, -8
subfic r11, r12, 64
L(nextbyte1):
ldu r7, 8(r3) /* Load next dw. */
addi r12, r12, 8 /* Shift one byte and compare. */
addi r11, r11, -8
#ifdef __LITTLE_ENDIAN__
srd r9, r5, r12 /* Rotate based on mask. */
sld r10, r7, r11
#else
sld r9, r5, r12
srd r10, r7, r11
#endif
/* Form single dw from few bytes on first load and second load. */
or r10, r9, r10
/* Check for null in the formed dw. */
cmpb r9, r0, r10
cmpdi cr7, r9, 0
bne cr7, L(retnull)
/* Cmpb search str and input str. */
cmpb r9, r10, r6
cmpdi cr7, r9, -1
beq cr7, L(match)
addi r8, r8, 1
b L(begin)
.align 4
L(match):
/* There is a match of 8 bytes, check next bytes. */
cmpdi cr7, r31, 8
beq cr7, L(return)
/* Update next starting point r8. */
srdi r9, r11, 3
subf r9, r9, r3
mr r8, r9
L(secondmatch):
mr r5, r7
rlwinm r10, r30, 3, 26, 28 /* Calculate padding in bits. */
ld r6, 0(r4)
addi r4, r4, 8
cmpdi cr7, r10, 0 /* Check if its already aligned? */
beq cr7, L(proceed3)
#ifdef __LITTLE_ENDIAN__
srd r6, r6, r10 /* Discard unwanted bits. */
cmpb r9, r0, r6
sld r9, r9, r10
#else
sld r6, r6, r10
cmpb r9, r0, r6
srd r9, r9, r10
#endif
cmpdi cr7, r9, 0
bne cr7, L(proceed3)
ld r9, 0(r4)
subfic r10, r10, 64
#ifdef __LITTLE_ENDIAN__
sld r9, r9, r10 /* Discard unwanted bits. */
#else
srd r9, r9, r10
#endif
or r6, r6, r9 /* Form complete search str. */
L(proceed3):
li r7, 0
addi r3, r3, 8
cmpb r9, r0, r5
cmpdi cr7, r9, 0
bne cr7, L(proceed4)
ld r7, 0(r3)
L(proceed4):
#ifdef __LITTLE_ENDIAN__
srd r9, r5, r12
sld r10, r7, r11
#else
sld r9, r5, r12
srd r10, r7, r11
#endif
/* Form single dw with few bytes from first and second load. */
or r10, r9, r10
cmpb r9, r0, r6
cmpdi cr7, r9, 0
bne cr7, L(return4)
/* Check for null in the formed dw. */
cmpb r9, r0, r10
cmpdi cr7, r9, 0
bne cr7, L(retnull)
/* If the next 8 bytes dont match, start search again. */
cmpb r9, r10, r6
cmpdi cr7, r9, -1
bne cr7, L(reset)
/* If the next 8 bytes match, load and compare next 8. */
b L(secondmatch)
.align 4
L(reset):
/* Start the search again. */
addi r8, r8, 1
b L(begin)
.align 4
L(return3):
/* Count leading zeros and compare partial dw. */
#ifdef __LITTLE_ENDIAN__
addi r7, r9, -1
andc r7, r7, r9
popcntd r7, r7
subfic r7, r7, 64
sld r10, r5, r7
sld r6, r6, r7
#else
cntlzd r7, r9
subfic r7, r7, 64
srd r10, r5, r7
srd r6, r6, r7
#endif
cmpb r9, r10, r6
cmpdi cr7, r9, -1
addi r8, r8, 1
/* Start search again if there is no match. */
bne cr7, L(begin)
/* If the words match, update return values. */
subfic r7, r7, 64
srdi r7, r7, 3
add r3, r3, r7
subf r3, r31, r3
b L(end)
.align 4
L(return4):
/* Count leading zeros and compare partial dw. */
#ifdef __LITTLE_ENDIAN__
addi r7, r9, -1
andc r7, r7, r9
popcntd r7, r7
subfic r7, r7, 64
sld r10, r10, r7
sld r6, r6, r7
#else
cntlzd r7, r9
subfic r7, r7, 64
srd r10, r10, r7
srd r6, r6, r7
#endif
cmpb r9, r10, r6
cmpdi cr7, r9, -1
addi r8, r8, 1
bne cr7, L(begin)
subfic r7, r7, 64
srdi r11, r11, 3
subf r3, r11, r3
srdi r7, r7, 3
add r3, r3, r7
subf r3, r31, r3
b L(end)
.align 4
L(begin):
mr r3, r8
/* When our iterations exceed ITERATIONS,fall back to default. */
addi r28, r28, 1
cmpdi cr7, r28, ITERATIONS
beq cr7, L(default)
lbz r4, 0(r30)
bl STRCHR
nop
/* If first char of search str is not present. */
cmpdi cr7, r3, 0
ble cr7, L(end)
mr r8, r3
mr r4, r30 /* Restore r4. */
li r0, 0
mr r6, r29
clrrdi r4, r4, 3
addi r4, r4, 8
b L(begin1)
/* Handle less than 8 search string. */
.align 4
L(lessthan8):
mr r4, r3
mr r9, r30
li r0, 0
rlwinm r10, r9, 3, 26, 28 /* Calculate padding in bits. */
srdi r8, r10, 3 /* Padding in bytes. */
clrrdi r9, r9, 3 /* Make r4 aligned to 8. */
ld r6, 0(r9)
cmpdi cr7, r10, 0 /* Check if its already aligned? */
beq cr7, L(proceed2)
#ifdef __LITTLE_ENDIAN__
srd r6, r6, r10 /* Discard unwanted bits. */
#else
sld r6, r6, r10
#endif
subfic r8, r8, 8
cmpd cr7, r8, r31 /* Next load needed? */
bge cr7, L(proceed2)
ld r7, 8(r9)
subfic r10, r10, 64
#ifdef __LITTLE_ENDIAN__
sld r7, r7, r10 /* Discard unwanted bits. */
#else
srd r7, r7, r10
#endif
or r6, r6, r7 /* Form complete search str. */
L(proceed2):
mr r29, r6
rlwinm r10, r3, 3, 26, 28
clrrdi r7, r3, 3 /* Make r3 aligned. */
ld r5, 0(r7)
sldi r8, r31, 3
subfic r8, r8, 64
#ifdef __LITTLE_ENDIAN__
sld r6, r6, r8
cmpb r9, r0, r5
srd r9, r9, r10
#else
srd r6, r6, r8
cmpb r9, r0, r5
sld r9, r9, r10
#endif
cmpdi cr7, r9, 0
bne cr7, L(noload)
cmpdi cr7, r10, 0
beq cr7, L(continue)
ld r7, 8(r7)
L(continue1):
mr r12, r10
addi r12, r12, -8
subfic r11, r12, 64
b L(nextbyte)
.align 4
L(continue):
ld r7, 8(r7)
li r12, -8 /* Shift values. */
li r11, 72 /* Shift values. */
L(nextbyte):
addi r12, r12, 8 /* Mask for rotation. */
addi r11, r11, -8
#ifdef __LITTLE_ENDIAN__
srd r9, r5, r12
sld r10, r7, r11
or r10, r9, r10
sld r10, r10, r8
cmpb r9, r0, r10
srd r9, r9, r8
#else
sld r9, r5, r12
srd r10, r7, r11
or r10, r9, r10
srd r10, r10, r8
cmpb r9, r0, r10
sld r9, r9, r8
#endif
cmpdi cr7, r9, 0
bne cr7, L(retnull)
cmpb r9, r10, r6
cmpdi cr7, r9, -1
beq cr7, L(end)
addi r3, r4, 1
/* When our iterations exceed ITERATIONS,fall back to default. */
addi r28, r28, 1
cmpdi cr7, r28, ITERATIONS
beq cr7, L(default)
lbz r4, 0(r30)
bl STRCHR
nop
/* If first char of search str is not present. */
cmpdi cr7, r3, 0
ble cr7, L(end)
mr r4, r3
mr r6, r29
li r0, 0
b L(proceed2)
.align 4
L(noload):
/* Reached null in r3, so skip next load. */
li r7, 0
b L(continue1)
.align 4
L(return):
/* Update return values. */
srdi r9, r11, 3
subf r3, r9, r3
b L(end)
/* Handling byte by byte. */
.align 4
L(bytebybyte):
mr r8, r3
addi r8, r8, -1
L(loop1):
addi r8, r8, 1
mr r3, r8
mr r4, r30
lbz r6, 0(r4)
cmpdi cr7, r6, 0
beq cr7, L(updater3)
L(loop):
lbz r5, 0(r3)
cmpdi cr7, r5, 0
beq cr7, L(retnull)
cmpld cr7, r6, r5
bne cr7, L(loop1)
addi r3, r3, 1
addi r4, r4, 1
lbz r6, 0(r4)
cmpdi cr7, r6, 0
beq cr7, L(updater3)
b L(loop)
/* Handling return values. */
.align 4
L(updater3):
subf r3, r31, r3 /* Reduce len of r4 from r3. */
b L(end)
.align 4
L(ret_r3):
mr r3, r29 /* Return r3. */
b L(end)
.align 4
L(retnull):
li r3, 0 /* Return NULL. */
b L(end)
.align 4
L(default):
mr r4, r30
bl __strstr_ppc
nop
.align 4
L(end):
addi r1, r1, FRAMESIZE /* Restore stack pointer. */
cfi_adjust_cfa_offset(-FRAMESIZE)
ld r0, 16(r1) /* Restore the saved link register. */
ld r28, -32(r1) /* Restore callers save register r28. */
ld r29, -24(r1) /* Restore callers save register r29. */
ld r30, -16(r1) /* Restore callers save register r30. */
ld r31, -8(r1) /* Restore callers save register r31. */
mtlr r0 /* Branch to link register. */
blr
END (strstr)
libc_hidden_builtin_def (strstr)