glibc/sysdeps/powerpc/powerpc32/power7/memset.S

431 lines
6.8 KiB
ArmAsm
Raw Normal View History

2010-05-20 11:35:05 +00:00
/* Optimized memset implementation for PowerPC32/POWER7.
Copyright (C) 2010-2021 Free Software Foundation, Inc.
2010-05-20 11:35:05 +00:00
This file is part of the GNU C Library.
The GNU C Library is free software; you can redistribute it and/or
modify it under the terms of the GNU Lesser General Public
License as published by the Free Software Foundation; either
version 2.1 of the License, or (at your option) any later version.
The GNU C Library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
License along with the GNU C Library; if not, see
Prefer https to http for gnu.org and fsf.org URLs Also, change sources.redhat.com to sourceware.org. This patch was automatically generated by running the following shell script, which uses GNU sed, and which avoids modifying files imported from upstream: sed -ri ' s,(http|ftp)(://(.*\.)?(gnu|fsf|sourceware)\.org($|[^.]|\.[^a-z])),https\2,g s,(http|ftp)(://(.*\.)?)sources\.redhat\.com($|[^.]|\.[^a-z]),https\2sourceware.org\4,g ' \ $(find $(git ls-files) -prune -type f \ ! -name '*.po' \ ! -name 'ChangeLog*' \ ! -path COPYING ! -path COPYING.LIB \ ! -path manual/fdl-1.3.texi ! -path manual/lgpl-2.1.texi \ ! -path manual/texinfo.tex ! -path scripts/config.guess \ ! -path scripts/config.sub ! -path scripts/install-sh \ ! -path scripts/mkinstalldirs ! -path scripts/move-if-change \ ! -path INSTALL ! -path locale/programs/charmap-kw.h \ ! -path po/libc.pot ! -path sysdeps/gnu/errlist.c \ ! '(' -name configure \ -execdir test -f configure.ac -o -f configure.in ';' ')' \ ! '(' -name preconfigure \ -execdir test -f preconfigure.ac ';' ')' \ -print) and then by running 'make dist-prepare' to regenerate files built from the altered files, and then executing the following to cleanup: chmod a+x sysdeps/unix/sysv/linux/riscv/configure # Omit irrelevant whitespace and comment-only changes, # perhaps from a slightly-different Autoconf version. git checkout -f \ sysdeps/csky/configure \ sysdeps/hppa/configure \ sysdeps/riscv/configure \ sysdeps/unix/sysv/linux/csky/configure # Omit changes that caused a pre-commit check to fail like this: # remote: *** error: sysdeps/powerpc/powerpc64/ppc-mcount.S: trailing lines git checkout -f \ sysdeps/powerpc/powerpc64/ppc-mcount.S \ sysdeps/unix/sysv/linux/s390/s390-64/syscall.S # Omit change that caused a pre-commit check to fail like this: # remote: *** error: sysdeps/sparc/sparc64/multiarch/memcpy-ultra3.S: last line does not end in newline git checkout -f sysdeps/sparc/sparc64/multiarch/memcpy-ultra3.S
2019-09-07 05:40:42 +00:00
<https://www.gnu.org/licenses/>. */
2010-05-20 11:35:05 +00:00
#include <sysdep.h>
Do not use __ptr_t. sys/cdefs.h has a macro __ptr_t, which a few places in glibc use instead of void *. void * is a well-understood standard type for that purpose and in a post-C89 context there is no need for a macro for it; this patch changes those places to use void * directly instead. Unlike __long_double_t, __ptr_t is widely used outside glibc (or at least has many hits on codesearch.debian.net). I don't know how many of those uses would break if sys/cdefs.h ceased to define the macro, but there's enough risk that this patch leaves the definition and just removes the uses within glibc; removal of the definition can be considered separately if desired. Tested for x86_64, and with build-many-glibcs.py. * malloc/mcheck.c (old_free_hook): Use void * instead of __ptr_t. (old_malloc_hook): Likewise. (old_memalign_hook): Likewise. (old_realloc_hook): Likewise. (struct hdr): Likewise. (flood): Likewise. (freehook): Likewise. (mallochook): Likewise. (memalignhook): Likewise. (reallochook): Likewise. (mprobe): Likewise. * malloc/mtrace.c (mallwatch): Likewise. (tr_old_free_hook): Likewise. (tr_old_malloc_hook): Likewise. (tr_old_realloc_hook): Likewise. (tr_old_memalign_hook): Likewise. (tr_where): Likewise. (lock_and_info): Likewise. (tr_freehook): Likewise. (tr_mallochook): Likewise. (tr_reallochook): Likewise. (tr_memalignhook): Likewise. * misc/err.h [!__GNUC_VA_LIST] (__gnuc_va_list): Likewise. * misc/mmap.c (__mmap): Likewise. * misc/mmap64.c (__mmap64): Likewise. * misc/mprotect.c (__mprotect): Likewise. * misc/msync.c (msync): Likewise. * misc/munmap.c (__munmap): Likewise. * posix/posix_madvise.c (posix_madvise): Likewise. * socket/send.c (__send): Likewise. * socket/sendto.c (__sendto): Likewise. * socket/setsockopt.c (__setsockopt): Likewise. * string/memcmp.c (__ptr_t): Remove macro. (MEMCMP): Use void * instead of ptr_t. * string/memrchr.c (__ptr_t): Remove macro. (__memrchr): Use void * instead of ptr_t. * sysdeps/mach/hurd/dl-sysdep.c (__mmap): Likewise. * sysdeps/mach/hurd/mmap.c (__mmap): Likewise. * sysdeps/mach/hurd/mmap64.c (__mmap64): Likewise. * sysdeps/mach/mprotect.c (__mprotect): Likewise. * sysdeps/mach/msync.c (msync): Likewise. * sysdeps/mach/munmap.c (__munmap): Likewise. * sysdeps/mips/bits/setjmp.h (struct __jmp_buf_internal_tag): Likewise. * sysdeps/posix/getcwd.c (__getcwd): Likewise. * sysdeps/powerpc/powerpc32/memset.S (memset): Likewise. * sysdeps/powerpc/powerpc32/power4/memcpy.S (memcpy): Likewise. * sysdeps/powerpc/powerpc32/power4/memset.S (memset): Likewise. * sysdeps/powerpc/powerpc32/power6/memcpy.S (memcpy): Likewise. * sysdeps/powerpc/powerpc32/power6/memset.S (memset): Likewise. * sysdeps/powerpc/powerpc32/power7/memcpy.S (memcpy): Likewise. * sysdeps/powerpc/powerpc32/power7/mempcpy.S (__mempcpy): Likewise. * sysdeps/powerpc/powerpc32/power7/memset.S (memset): Likewise. * sysdeps/powerpc/powerpc64/memcpy.S (memcpy): Likewise. * sysdeps/powerpc/powerpc64/memset.S (memset): Likewise. * sysdeps/powerpc/powerpc64/power4/memcpy.S (memcpy): Likewise. * sysdeps/powerpc/powerpc64/power4/memset.S (memset): Likewise. * sysdeps/powerpc/powerpc64/power6/memcpy.S (memcpy): Likewise. * sysdeps/powerpc/powerpc64/power6/memset.S (memset): Likewise. * sysdeps/powerpc/powerpc64/power7/memcpy.S (memcpy): Likewise. * sysdeps/powerpc/powerpc64/power7/mempcpy.S (__mempcpy): Likewise. * sysdeps/powerpc/powerpc64/power7/memset.S (memset): Likewise. * sysdeps/powerpc/powerpc64/power8/memset.S (memset): Likewise. * sysdeps/tile/memcmp.c (__ptr_t): Remove macro. (MEMCMP): Use void * instead of ptr_t. * sysdeps/unix/sysv/linux/alpha/oldglob.c (old_glob_t): Likewise. * sysdeps/unix/sysv/linux/mmap.c (__mmap): Likewise.
2017-08-08 17:14:49 +00:00
/* void * [r3] memset (void *s [r3], int c [r4], size_t n [r5]));
2010-05-20 11:35:05 +00:00
Returns 's'. */
.machine power7
EALIGN (memset, 5, 0)
2010-05-20 11:35:05 +00:00
CALL_MCOUNT
.align 4
L(_memset):
cmplwi cr7,5,31
cmplwi cr6,5,8
mr 10,3 /* Save original argument for later. */
mr 7,1 /* Save original r1 for later. */
cfi_offset(31,-8)
/* Replicate byte to word. */
insrwi 4,4,8,16
insrwi 4,4,16,0
2010-05-20 11:35:05 +00:00
ble cr6,L(small) /* If length <= 8, use short copy code. */
neg 0,3
ble cr7,L(medium) /* If length < 32, use medium copy code. */
/* Save our word twice to create a doubleword that we will later
copy to a FPR. */
stwu 1,-32(1)
andi. 11,10,7 /* Check alignment of DST. */
mr 12,5
stw 4,24(1)
stw 4,28(1)
beq L(big_aligned)
clrlwi 0,0,29
mtocrf 0x01,0
subf 5,0,5
/* Get DST aligned to 8 bytes. */
1: bf 31,2f
stb 4,0(10)
addi 10,10,1
2: bf 30,4f
sth 4,0(10)
addi 10,10,2
4: bf 29,L(big_aligned)
stw 4,0(10)
addi 10,10,4
.align 4
L(big_aligned):
cmplwi cr5,5,255
li 0,32
cmplwi cr1,5,160
dcbtst 0,10
cmplwi cr6,4,0
srwi 9,5,3 /* Number of full doublewords remaining. */
crand 27,26,21
mtocrf 0x01,9
bt 27,L(huge)
/* From this point on, we'll copy 32+ bytes and the value
isn't 0 (so we can't use dcbz). */
srwi 8,5,5
clrlwi 11,5,29
cmplwi cr6,11,0
cmplwi cr1,9,4
mtctr 8
/* Copy 1~3 doublewords so the main loop starts
at a multiple of 32 bytes. */
bf 30,1f
stw 4,0(10)
stw 4,4(10)
stw 4,8(10)
stw 4,12(10)
addi 10,10,16
bf 31,L(big_loop)
stw 4,0(10)
stw 4,4(10)
addi 10,10,8
mr 12,10
blt cr1,L(tail_bytes)
b L(big_loop)
.align 4
1: /* Copy 1 doubleword. */
bf 31,L(big_loop)
stw 4,0(10)
stw 4,4(10)
addi 10,10,8
/* First use a 32-bytes loop with stw's to try and avoid the LHS due
to the lfd we will do next. Also, ping-pong through r10 and r12
to avoid AGEN delays. */
.align 4
L(big_loop):
addi 12,10,32
stw 4,0(10)
stw 4,4(10)
stw 4,8(10)
stw 4,12(10)
stw 4,16(10)
stw 4,20(10)
stw 4,24(10)
stw 4,28(10)
bdz L(tail_bytes)
addi 10,10,64
stw 4,0(12)
stw 4,4(12)
stw 4,8(12)
stw 4,12(12)
stw 4,16(12)
stw 4,20(12)
stw 4,24(12)
stw 4,28(12)
bdnz L(big_loop_fast_setup)
mr 12,10
b L(tail_bytes)
/* Now that we're probably past the LHS window, use the VSX to
speed up the loop. */
L(big_loop_fast_setup):
li 11,24
li 6,16
lxvdsx 4,1,11
.align 4
L(big_loop_fast):
addi 12,10,32
2010-05-25 00:56:44 +00:00
stxvd2x 4,0,10
2010-05-20 11:35:05 +00:00
stxvd2x 4,10,6
bdz L(tail_bytes)
addi 10,10,64
2010-05-25 00:56:44 +00:00
stxvd2x 4,0,12
2010-05-20 11:35:05 +00:00
stxvd2x 4,12,6
bdnz L(big_loop_fast)
mr 12,10
.align 4
L(tail_bytes):
/* Check for tail bytes. */
mr 1,7 /* Restore r1. */
beqlr cr6
clrlwi 0,5,29
mtocrf 0x01,0
/* At this point we have a tail of 0-7 bytes and we know that the
destination is doubleword-aligned. */
4: /* Copy 4 bytes. */
bf 29,2f
stw 4,0(12)
addi 12,12,4
2: /* Copy 2 bytes. */
bf 30,1f
sth 4,0(12)
addi 12,12,2
1: /* Copy 1 byte. */
bflr 31
stb 4,0(12)
blr
/* Special case when value is 0 and we have a long length to deal
with. Use dcbz to zero out 128-bytes at a time. Before using
dcbz though, we need to get the destination 128-bytes aligned. */
.align 4
L(huge):
lfd 4,24(1)
andi. 11,10,127
neg 0,10
beq L(huge_aligned)
clrlwi 0,0,25
subf 5,0,5
srwi 0,0,3
mtocrf 0x01,0
/* Get DST aligned to 128 bytes. */
8: bf 28,4f
stfd 4,0(10)
stfd 4,8(10)
stfd 4,16(10)
stfd 4,24(10)
stfd 4,32(10)
stfd 4,40(10)
stfd 4,48(10)
stfd 4,56(10)
addi 10,10,64
.align 4
4: bf 29,2f
stfd 4,0(10)
stfd 4,8(10)
stfd 4,16(10)
stfd 4,24(10)
addi 10,10,32
.align 4
2: bf 30,1f
stfd 4,0(10)
stfd 4,8(10)
addi 10,10,16
.align 4
1: bf 31,L(huge_aligned)
stfd 4,0(10)
addi 10,10,8
L(huge_aligned):
srwi 8,5,7
clrlwi 11,5,25
cmplwi cr6,11,0
mtctr 8
/* Copies 128-bytes at a time. */
.align 4
L(huge_loop):
dcbz 0,10
addi 10,10,128
bdnz L(huge_loop)
/* We have a tail of 0~127 bytes to handle. */
mr 1,7 /* Restore r1. */
beqlr cr6
subf 9,3,10
subf 5,9,12
srwi 8,5,3
cmplwi cr6,8,0
mtocrf 0x01,8
/* We have a tail o 1~127 bytes. Copy up to 15 doublewords for
speed. We'll handle the resulting tail bytes later. */
beq cr6,L(tail)
8: bf 28,4f
stfd 4,0(10)
stfd 4,8(10)
stfd 4,16(10)
stfd 4,24(10)
stfd 4,32(10)
stfd 4,40(10)
stfd 4,48(10)
stfd 4,56(10)
addi 10,10,64
.align 4
4: bf 29,2f
stfd 4,0(10)
stfd 4,8(10)
stfd 4,16(10)
stfd 4,24(10)
addi 10,10,32
.align 4
2: bf 30,1f
stfd 4,0(10)
stfd 4,8(10)
addi 10,10,16
.align 4
1: bf 31,L(tail)
stfd 4,0(10)
addi 10,10,8
/* Handle the rest of the tail bytes here. */
L(tail):
mtocrf 0x01,5
.align 4
4: bf 29,2f
stw 4,0(10)
addi 10,10,4
.align 4
2: bf 30,1f
sth 4,0(10)
addi 10,10,2
.align 4
1: bflr 31
stb 4,0(10)
blr
/* Expanded tree to copy tail bytes without increments. */
.align 4
L(copy_tail):
bf 29,L(FXX)
stw 4,0(10)
bf 30,L(TFX)
sth 4,4(10)
bflr 31
stb 4,6(10)
blr
.align 4
L(FXX): bf 30,L(FFX)
sth 4,0(10)
bflr 31
stb 4,2(10)
blr
.align 4
L(TFX): bflr 31
stb 4,4(10)
blr
.align 4
L(FFX): bflr 31
stb 4,0(10)
blr
/* Handle copies of 9~31 bytes. */
.align 4
L(medium):
/* At least 9 bytes to go. */
andi. 11,10,3
clrlwi 0,0,30
beq L(medium_aligned)
/* Force 4-bytes alignment for DST. */
mtocrf 0x01,0
subf 5,0,5
1: /* Copy 1 byte. */
bf 31,2f
stb 4,0(10)
addi 10,10,1
2: /* Copy 2 bytes. */
bf 30,L(medium_aligned)
sth 4,0(10)
addi 10,10,2
.align 4
L(medium_aligned):
/* At least 6 bytes to go, and DST is word-aligned. */
cmplwi cr1,5,16
mtocrf 0x01,5
blt cr1,8f
/* Copy 16 bytes. */
stw 4,0(10)
stw 4,4(10)
stw 4,8(10)
stw 4,12(10)
addi 10,10,16
8: /* Copy 8 bytes. */
bf 28,4f
stw 4,0(10)
stw 4,4(10)
addi 10,10,8
4: /* Copy 4 bytes. */
bf 29,2f
stw 4,0(10)
addi 10,10,4
2: /* Copy 2-3 bytes. */
bf 30,1f
sth 4,0(10)
addi 10,10,2
1: /* Copy 1 byte. */
bflr 31
stb 4,0(10)
blr
/* Handles copies of 0~8 bytes. */
.align 4
L(small):
mtocrf 0x01,5
bne cr6,L(copy_tail)
stw 4,0(10)
stw 4,4(10)
blr
END (memset)
2010-05-20 11:35:05 +00:00
libc_hidden_builtin_def (memset)