| /* Optimized strstr implementation for PowerPC64/POWER7. |
| Copyright (C) 2015-2018 Free Software Foundation, Inc. |
| This file is part of the GNU C Library. |
| |
| The GNU C Library is free software; you can redistribute it and/or |
| modify it under the terms of the GNU Lesser General Public |
| License as published by the Free Software Foundation; either |
| version 2.1 of the License, or (at your option) any later version. |
| |
| The GNU C Library is distributed in the hope that it will be useful, |
| but WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| Lesser General Public License for more details. |
| |
| You should have received a copy of the GNU Lesser General Public |
| License along with the GNU C Library; if not, see |
| <http://www.gnu.org/licenses/>. */ |
| |
| #include <sysdep.h> |
| |
| /* Char * [r3] strstr (char *s [r3], char * pat[r4]) */ |
| |
| /* The performance gain is obtained using aligned memory access, load |
| * doubleword and usage of cmpb instruction for quicker comparison. */ |
| |
| #define ITERATIONS 64 |
| |
| #ifndef STRSTR |
| # define STRSTR strstr |
| #endif |
| |
| #ifndef STRLEN |
| /* For builds with no IFUNC support, local calls should be made to internal |
| GLIBC symbol (created by libc_hidden_builtin_def). */ |
| # ifdef SHARED |
| # define STRLEN __GI_strlen |
| # define STRLEN_is_local |
| # else |
| # define STRLEN strlen |
| # endif |
| #endif |
| |
| #ifndef STRNLEN |
| /* For builds with no IFUNC support, local calls should be made to internal |
| GLIBC symbol (created by libc_hidden_builtin_def). */ |
| # ifdef SHARED |
| # define STRNLEN __GI_strnlen |
| # define STRNLEN_is_local |
| # else |
| # define STRNLEN __strnlen |
| # endif |
| #endif |
| |
| #ifndef STRCHR |
| # ifdef SHARED |
| # define STRCHR __GI_strchr |
| # define STRCHR_is_local |
| # else |
| # define STRCHR strchr |
| # endif |
| #endif |
| |
| #define FRAMESIZE (FRAME_MIN_SIZE+32) |
| .machine power7 |
| /* Can't be ENTRY_TOCLESS due to calling __strstr_ppc which uses r2. */ |
| ENTRY (STRSTR, 4) |
| CALL_MCOUNT 2 |
| mflr r0 /* Load link register LR to r0. */ |
| std r31, -8(r1) /* Save callers register r31. */ |
| std r30, -16(r1) /* Save callers register r30. */ |
| std r29, -24(r1) /* Save callers register r29. */ |
| std r28, -32(r1) /* Save callers register r28. */ |
| std r0, 16(r1) /* Store the link register. */ |
| cfi_offset(r31, -8) |
| cfi_offset(r30, -16) |
| cfi_offset(r28, -32) |
| cfi_offset(r29, -24) |
| cfi_offset(lr, 16) |
| stdu r1, -FRAMESIZE(r1) /* Create the stack frame. */ |
| cfi_adjust_cfa_offset(FRAMESIZE) |
| |
| dcbt 0, r3 |
| dcbt 0, r4 |
| cmpdi cr7, r3, 0 |
| beq cr7, L(retnull) |
| cmpdi cr7, r4, 0 |
| beq cr7, L(retnull) |
| |
| mr r29, r3 |
| mr r30, r4 |
| mr r3, r4 |
| bl STRLEN |
| #ifndef STRLEN_is_local |
| nop |
| #endif |
| |
| cmpdi cr7, r3, 0 /* If search str is null. */ |
| beq cr7, L(ret_r3) |
| |
| mr r31, r3 |
| mr r4, r3 |
| mr r3, r29 |
| bl STRNLEN |
| #ifndef STRNLEN_is_local |
| nop |
| #endif |
| |
| cmpd cr7, r3, r31 /* If len(r3) < len(r4). */ |
| blt cr7, L(retnull) |
| mr r3, r29 |
| lbz r4, 0(r30) |
| bl STRCHR |
| #ifndef STRCHR_is_local |
| nop |
| #endif |
| |
| mr r11, r3 |
| /* If first char of search str is not present. */ |
| cmpdi cr7, r3, 0 |
| ble cr7, L(end) |
| /* Reg r28 is used to count the number of iterations. */ |
| li r28, 0 |
| rldicl r8, r3, 0, 52 /* Page cross check. */ |
| cmpldi cr7, r8, 4096-16 |
| bgt cr7, L(bytebybyte) |
| |
| rldicl r8, r30, 0, 52 |
| cmpldi cr7, r8, 4096-16 |
| bgt cr7, L(bytebybyte) |
| |
| /* If len(r4) < 8 handle in a different way. */ |
| /* Shift position based on null and use cmpb. */ |
| cmpdi cr7, r31, 8 |
| blt cr7, L(lessthan8) |
| |
| /* Len(r4) >= 8 reaches here. */ |
| mr r8, r3 /* Save r3 for future use. */ |
| mr r4, r30 /* Restore r4. */ |
| li r0, 0 |
| rlwinm r10, r30, 3, 26, 28 /* Calculate padding in bits. */ |
| clrrdi r4, r4, 3 /* Make r4 aligned to 8. */ |
| ld r6, 0(r4) |
| addi r4, r4, 8 |
| cmpdi cr7, r10, 0 /* Check if its already aligned? */ |
| beq cr7, L(begin1) |
| #ifdef __LITTLE_ENDIAN__ |
| srd r6, r6, r10 /* Discard unwanted bits. */ |
| #else |
| sld r6, r6, r10 |
| #endif |
| ld r9, 0(r4) |
| subfic r10, r10, 64 |
| #ifdef __LITTLE_ENDIAN__ |
| sld r9, r9, r10 /* Discard unwanted bits. */ |
| #else |
| srd r9, r9, r10 |
| #endif |
| or r6, r6, r9 /* Form complete search str. */ |
| L(begin1): |
| mr r29, r6 |
| rlwinm r10, r3, 3, 26, 28 |
| clrrdi r3, r3, 3 |
| ld r5, 0(r3) |
| cmpb r9, r0, r6 /* Check if input has null. */ |
| cmpdi cr7, r9, 0 |
| bne cr7, L(return3) |
| cmpb r9, r0, r5 /* Check if input has null. */ |
| #ifdef __LITTLE_ENDIAN__ |
| srd r9, r9, r10 |
| #else |
| sld r9, r9, r10 |
| #endif |
| cmpdi cr7, r9, 0 |
| bne cr7, L(retnull) |
| |
| li r12, -8 /* Shift values. */ |
| li r11, 72 /* Shift values. */ |
| cmpdi cr7, r10, 0 |
| beq cr7, L(nextbyte1) |
| mr r12, r10 |
| addi r12, r12, -8 |
| subfic r11, r12, 64 |
| |
| L(nextbyte1): |
| ldu r7, 8(r3) /* Load next dw. */ |
| addi r12, r12, 8 /* Shift one byte and compare. */ |
| addi r11, r11, -8 |
| #ifdef __LITTLE_ENDIAN__ |
| srd r9, r5, r12 /* Rotate based on mask. */ |
| sld r10, r7, r11 |
| #else |
| sld r9, r5, r12 |
| srd r10, r7, r11 |
| #endif |
| /* Form single dw from few bytes on first load and second load. */ |
| or r10, r9, r10 |
| /* Check for null in the formed dw. */ |
| cmpb r9, r0, r10 |
| cmpdi cr7, r9, 0 |
| bne cr7, L(retnull) |
| /* Cmpb search str and input str. */ |
| cmpb r9, r10, r6 |
| cmpdi cr7, r9, -1 |
| beq cr7, L(match) |
| addi r8, r8, 1 |
| b L(begin) |
| |
| .align 4 |
| L(match): |
| /* There is a match of 8 bytes, check next bytes. */ |
| cmpdi cr7, r31, 8 |
| beq cr7, L(return) |
| /* Update next starting point r8. */ |
| srdi r9, r11, 3 |
| subf r9, r9, r3 |
| mr r8, r9 |
| |
| L(secondmatch): |
| mr r5, r7 |
| rlwinm r10, r30, 3, 26, 28 /* Calculate padding in bits. */ |
| ld r6, 0(r4) |
| addi r4, r4, 8 |
| cmpdi cr7, r10, 0 /* Check if its already aligned? */ |
| beq cr7, L(proceed3) |
| #ifdef __LITTLE_ENDIAN__ |
| srd r6, r6, r10 /* Discard unwanted bits. */ |
| cmpb r9, r0, r6 |
| sld r9, r9, r10 |
| #else |
| sld r6, r6, r10 |
| cmpb r9, r0, r6 |
| srd r9, r9, r10 |
| #endif |
| cmpdi cr7, r9, 0 |
| bne cr7, L(proceed3) |
| ld r9, 0(r4) |
| subfic r10, r10, 64 |
| #ifdef __LITTLE_ENDIAN__ |
| sld r9, r9, r10 /* Discard unwanted bits. */ |
| #else |
| srd r9, r9, r10 |
| #endif |
| or r6, r6, r9 /* Form complete search str. */ |
| |
| L(proceed3): |
| li r7, 0 |
| addi r3, r3, 8 |
| cmpb r9, r0, r5 |
| cmpdi cr7, r9, 0 |
| bne cr7, L(proceed4) |
| ld r7, 0(r3) |
| L(proceed4): |
| #ifdef __LITTLE_ENDIAN__ |
| srd r9, r5, r12 |
| sld r10, r7, r11 |
| #else |
| sld r9, r5, r12 |
| srd r10, r7, r11 |
| #endif |
| /* Form single dw with few bytes from first and second load. */ |
| or r10, r9, r10 |
| cmpb r9, r0, r6 |
| cmpdi cr7, r9, 0 |
| bne cr7, L(return4) |
| /* Check for null in the formed dw. */ |
| cmpb r9, r0, r10 |
| cmpdi cr7, r9, 0 |
| bne cr7, L(retnull) |
| /* If the next 8 bytes dont match, start search again. */ |
| cmpb r9, r10, r6 |
| cmpdi cr7, r9, -1 |
| bne cr7, L(reset) |
| /* If the next 8 bytes match, load and compare next 8. */ |
| b L(secondmatch) |
| |
| .align 4 |
| L(reset): |
| /* Start the search again. */ |
| addi r8, r8, 1 |
| b L(begin) |
| |
| .align 4 |
| L(return3): |
| /* Count leading zeros and compare partial dw. */ |
| #ifdef __LITTLE_ENDIAN__ |
| addi r7, r9, -1 |
| andc r7, r7, r9 |
| popcntd r7, r7 |
| subfic r7, r7, 64 |
| sld r10, r5, r7 |
| sld r6, r6, r7 |
| #else |
| cntlzd r7, r9 |
| subfic r7, r7, 64 |
| srd r10, r5, r7 |
| srd r6, r6, r7 |
| #endif |
| cmpb r9, r10, r6 |
| cmpdi cr7, r9, -1 |
| addi r8, r8, 1 |
| /* Start search again if there is no match. */ |
| bne cr7, L(begin) |
| /* If the words match, update return values. */ |
| subfic r7, r7, 64 |
| srdi r7, r7, 3 |
| add r3, r3, r7 |
| subf r3, r31, r3 |
| b L(end) |
| |
| .align 4 |
| L(return4): |
| /* Count leading zeros and compare partial dw. */ |
| #ifdef __LITTLE_ENDIAN__ |
| addi r7, r9, -1 |
| andc r7, r7, r9 |
| popcntd r7, r7 |
| subfic r7, r7, 64 |
| sld r10, r10, r7 |
| sld r6, r6, r7 |
| #else |
| cntlzd r7, r9 |
| subfic r7, r7, 64 |
| srd r10, r10, r7 |
| srd r6, r6, r7 |
| #endif |
| cmpb r9, r10, r6 |
| cmpdi cr7, r9, -1 |
| addi r8, r8, 1 |
| bne cr7, L(begin) |
| subfic r7, r7, 64 |
| srdi r11, r11, 3 |
| subf r3, r11, r3 |
| srdi r7, r7, 3 |
| add r3, r3, r7 |
| subf r3, r31, r3 |
| b L(end) |
| |
| .align 4 |
| L(begin): |
| mr r3, r8 |
| /* When our iterations exceed ITERATIONS,fall back to default. */ |
| addi r28, r28, 1 |
| cmpdi cr7, r28, ITERATIONS |
| beq cr7, L(default) |
| lbz r4, 0(r30) |
| bl STRCHR |
| #ifndef STRCHR_is_local |
| nop |
| #endif |
| /* If first char of search str is not present. */ |
| cmpdi cr7, r3, 0 |
| ble cr7, L(end) |
| mr r8, r3 |
| mr r4, r30 /* Restore r4. */ |
| li r0, 0 |
| mr r6, r29 |
| clrrdi r4, r4, 3 |
| addi r4, r4, 8 |
| b L(begin1) |
| |
| /* Handle less than 8 search string. */ |
| .align 4 |
| L(lessthan8): |
| mr r4, r3 |
| mr r9, r30 |
| li r0, 0 |
| |
| rlwinm r10, r9, 3, 26, 28 /* Calculate padding in bits. */ |
| srdi r8, r10, 3 /* Padding in bytes. */ |
| clrrdi r9, r9, 3 /* Make r4 aligned to 8. */ |
| ld r6, 0(r9) |
| cmpdi cr7, r10, 0 /* Check if its already aligned? */ |
| beq cr7, L(proceed2) |
| #ifdef __LITTLE_ENDIAN__ |
| srd r6, r6, r10 /* Discard unwanted bits. */ |
| #else |
| sld r6, r6, r10 |
| #endif |
| subfic r8, r8, 8 |
| cmpd cr7, r8, r31 /* Next load needed? */ |
| bge cr7, L(proceed2) |
| ld r7, 8(r9) |
| subfic r10, r10, 64 |
| #ifdef __LITTLE_ENDIAN__ |
| sld r7, r7, r10 /* Discard unwanted bits. */ |
| #else |
| srd r7, r7, r10 |
| #endif |
| or r6, r6, r7 /* Form complete search str. */ |
| L(proceed2): |
| mr r29, r6 |
| rlwinm r10, r3, 3, 26, 28 |
| clrrdi r7, r3, 3 /* Make r3 aligned. */ |
| ld r5, 0(r7) |
| sldi r8, r31, 3 |
| subfic r8, r8, 64 |
| #ifdef __LITTLE_ENDIAN__ |
| sld r6, r6, r8 |
| cmpb r9, r0, r5 |
| srd r9, r9, r10 |
| #else |
| srd r6, r6, r8 |
| cmpb r9, r0, r5 |
| sld r9, r9, r10 |
| #endif |
| cmpdi cr7, r9, 0 |
| bne cr7, L(noload) |
| cmpdi cr7, r10, 0 |
| beq cr7, L(continue) |
| ld r7, 8(r7) |
| L(continue1): |
| mr r12, r10 |
| addi r12, r12, -8 |
| subfic r11, r12, 64 |
| b L(nextbyte) |
| |
| .align 4 |
| L(continue): |
| ld r7, 8(r7) |
| li r12, -8 /* Shift values. */ |
| li r11, 72 /* Shift values. */ |
| L(nextbyte): |
| addi r12, r12, 8 /* Mask for rotation. */ |
| addi r11, r11, -8 |
| #ifdef __LITTLE_ENDIAN__ |
| srd r9, r5, r12 |
| sld r10, r7, r11 |
| or r10, r9, r10 |
| sld r10, r10, r8 |
| cmpb r9, r0, r10 |
| srd r9, r9, r8 |
| #else |
| sld r9, r5, r12 |
| srd r10, r7, r11 |
| or r10, r9, r10 |
| srd r10, r10, r8 |
| cmpb r9, r0, r10 |
| sld r9, r9, r8 |
| #endif |
| cmpdi cr7, r9, 0 |
| bne cr7, L(retnull) |
| cmpb r9, r10, r6 |
| cmpdi cr7, r9, -1 |
| beq cr7, L(end) |
| addi r3, r4, 1 |
| /* When our iterations exceed ITERATIONS,fall back to default. */ |
| addi r28, r28, 1 |
| cmpdi cr7, r28, ITERATIONS |
| beq cr7, L(default) |
| lbz r4, 0(r30) |
| bl STRCHR |
| #ifndef STRCHR_is_local |
| nop |
| #endif |
| /* If first char of search str is not present. */ |
| cmpdi cr7, r3, 0 |
| ble cr7, L(end) |
| mr r4, r3 |
| mr r6, r29 |
| li r0, 0 |
| b L(proceed2) |
| |
| .align 4 |
| L(noload): |
| /* Reached null in r3, so skip next load. */ |
| li r7, 0 |
| b L(continue1) |
| |
| .align 4 |
| L(return): |
| /* Update return values. */ |
| srdi r9, r11, 3 |
| subf r3, r9, r3 |
| b L(end) |
| |
| /* Handling byte by byte. */ |
| .align 4 |
| L(bytebybyte): |
| mr r8, r3 |
| addi r8, r8, -1 |
| L(loop1): |
| addi r8, r8, 1 |
| mr r3, r8 |
| mr r4, r30 |
| lbz r6, 0(r4) |
| cmpdi cr7, r6, 0 |
| beq cr7, L(updater3) |
| L(loop): |
| lbz r5, 0(r3) |
| cmpdi cr7, r5, 0 |
| beq cr7, L(retnull) |
| cmpld cr7, r6, r5 |
| bne cr7, L(loop1) |
| addi r3, r3, 1 |
| addi r4, r4, 1 |
| lbz r6, 0(r4) |
| cmpdi cr7, r6, 0 |
| beq cr7, L(updater3) |
| b L(loop) |
| |
| /* Handling return values. */ |
| .align 4 |
| L(updater3): |
| subf r3, r31, r3 /* Reduce len of r4 from r3. */ |
| b L(end) |
| |
| .align 4 |
| L(ret_r3): |
| mr r3, r29 /* Return r3. */ |
| b L(end) |
| |
| .align 4 |
| L(retnull): |
| li r3, 0 /* Return NULL. */ |
| b L(end) |
| |
| .align 4 |
| L(default): |
| mr r4, r30 |
| bl __strstr_ppc |
| nop |
| |
| .align 4 |
| L(end): |
| addi r1, r1, FRAMESIZE /* Restore stack pointer. */ |
| cfi_adjust_cfa_offset(-FRAMESIZE) |
| ld r0, 16(r1) /* Restore the saved link register. */ |
| ld r28, -32(r1) /* Restore callers save register r28. */ |
| ld r29, -24(r1) /* Restore callers save register r29. */ |
| ld r30, -16(r1) /* Restore callers save register r30. */ |
| ld r31, -8(r1) /* Restore callers save register r31. */ |
| mtlr r0 /* Branch to link register. */ |
| blr |
| END (STRSTR) |
| libc_hidden_builtin_def (strstr) |