mirror of
https://sourceware.org/git/glibc.git
synced 2024-11-30 00:31:08 +00:00
04277e02d7
* All files with FSF copyright notices: Update copyright dates using scripts/update-copyrights. * locale/programs/charmap-kw.h: Regenerated. * locale/programs/locfile-kw.h: Likewise.
908 lines
22 KiB
ArmAsm
908 lines
22 KiB
ArmAsm
/* Optimized memcpy implementation for PowerPC32 on POWER6.
|
|
Copyright (C) 2003-2019 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include <sysdep.h>
|
|
|
|
/* void * [r3] memcpy (void *dst [r3], void *src [r4], size_t len [r5]);
|
|
Returns 'dst'.
|
|
|
|
Memcpy handles short copies (< 32-bytes) using a binary move blocks
|
|
(no loops) of lwz/stw. The tail (remaining 1-3) bytes is handled
|
|
with the appropriate combination of byte and halfword load/stores.
|
|
There is minimal effort to optimize the alignment of short moves.
|
|
|
|
Longer moves (>= 32-bytes) justify the effort to get at least the
|
|
destination word (4-byte) aligned. Further optimization is
|
|
possible when both source and destination are word aligned.
|
|
Each case has an optimized unrolled loop. */
|
|
|
|
.machine power6
|
|
EALIGN (memcpy, 5, 0)
|
|
CALL_MCOUNT
|
|
|
|
stwu 1,-32(1)
|
|
cfi_adjust_cfa_offset(32)
|
|
cmplwi cr1,5,31 /* check for short move. */
|
|
neg 0,3
|
|
cmplwi cr1,5,31
|
|
clrlwi 10,4,30 /* check alignment of src. */
|
|
andi. 11,3,3 /* check alignment of dst. */
|
|
clrlwi 0,0,30 /* Number of bytes until the 1st word of dst. */
|
|
ble- cr1,L(word_unaligned_short) /* If move < 32 bytes. */
|
|
cmplw cr6,10,11
|
|
stw 31,24(1)
|
|
stw 30,20(1)
|
|
cfi_offset(31,(24-32))
|
|
cfi_offset(30,(20-32))
|
|
mr 30,3
|
|
beq .L0
|
|
mtcrf 0x01,0
|
|
subf 31,0,5 /* Length after alignment. */
|
|
add 12,4,0 /* Compute src addr after alignment. */
|
|
/* Move 0-3 bytes as needed to get the destination word aligned. */
|
|
1: bf 31,2f
|
|
lbz 6,0(4)
|
|
bf 30,3f
|
|
lhz 7,1(4)
|
|
stb 6,0(3)
|
|
sth 7,1(3)
|
|
addi 3,3,3
|
|
b 0f
|
|
3:
|
|
stb 6,0(3)
|
|
addi 3,3,1
|
|
b 0f
|
|
2: bf 30,0f
|
|
lhz 6,0(4)
|
|
sth 6,0(3)
|
|
addi 3,3,2
|
|
0:
|
|
clrlwi 10,12,30 /* check alignment of src again. */
|
|
srwi 9,31,2 /* Number of full words remaining. */
|
|
bne- cr6,L(wdu) /* If source is not word aligned. .L6 */
|
|
clrlwi 11,31,30 /* calculate the number of tail bytes */
|
|
b L(word_aligned)
|
|
/* Copy words from source to destination, assuming the destination is
|
|
aligned on a word boundary.
|
|
|
|
At this point we know there are at least 29 bytes left (32-3) to copy.
|
|
The next step is to determine if the source is also word aligned.
|
|
If not branch to the unaligned move code at .L6. which uses
|
|
a load, shift, store strategy.
|
|
|
|
Otherwise source and destination are word aligned, and we can use
|
|
the optimized word copy loop. */
|
|
.align 4
|
|
.L0:
|
|
mr 31,5
|
|
mr 12,4
|
|
bne- cr6,L(wdu) /* If source is not word aligned. .L6 */
|
|
srwi 9,5,2 /* Number of full words remaining. */
|
|
clrlwi 11,5,30 /* calculate the number of tail bytes */
|
|
|
|
/* Move words where destination and source are word aligned.
|
|
Use an unrolled loop to copy 4 words (16-bytes) per iteration.
|
|
If the copy is not an exact multiple of 16 bytes, 1-3
|
|
words are copied as needed to set up the main loop. After
|
|
the main loop exits there may be a tail of 1-3 bytes. These bytes are
|
|
copied a halfword/byte at a time as needed to preserve alignment. */
|
|
L(word_aligned):
|
|
mtcrf 0x01,9
|
|
srwi 8,31,4 /* calculate the 16 byte loop count */
|
|
cmplwi cr1,9,4
|
|
cmplwi cr6,11,0
|
|
mr 11,12
|
|
|
|
bf 30,1f
|
|
lwz 6,0(12)
|
|
lwz 7,4(12)
|
|
addi 11,12,8
|
|
mtctr 8
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
addi 10,3,8
|
|
bf 31,4f
|
|
lwz 0,8(12)
|
|
stw 0,8(3)
|
|
blt cr1,3f
|
|
addi 11,12,12
|
|
addi 10,3,12
|
|
b 4f
|
|
.align 4
|
|
1:
|
|
mr 10,3
|
|
mtctr 8
|
|
bf 31,4f
|
|
lwz 6,0(12)
|
|
addi 11,12,4
|
|
stw 6,0(3)
|
|
addi 10,3,4
|
|
|
|
.align 4
|
|
4:
|
|
lwz 6,0(11)
|
|
lwz 7,4(11)
|
|
lwz 8,8(11)
|
|
lwz 0,12(11)
|
|
stw 6,0(10)
|
|
stw 7,4(10)
|
|
stw 8,8(10)
|
|
stw 0,12(10)
|
|
addi 11,11,16
|
|
addi 10,10,16
|
|
bdnz 4b
|
|
3:
|
|
clrrwi 0,31,2
|
|
mtcrf 0x01,31
|
|
beq cr6,0f
|
|
.L9:
|
|
add 3,3,0
|
|
add 12,12,0
|
|
|
|
/* At this point we have a tail of 0-3 bytes and we know that the
|
|
destination is word aligned. */
|
|
2: bf 30,1f
|
|
lhz 6,0(12)
|
|
addi 12,12,2
|
|
sth 6,0(3)
|
|
addi 3,3,2
|
|
1: bf 31,0f
|
|
lbz 6,0(12)
|
|
stb 6,0(3)
|
|
0:
|
|
/* Return original dst pointer. */
|
|
mr 3,30
|
|
lwz 30,20(1)
|
|
lwz 31,24(1)
|
|
addi 1,1,32
|
|
blr
|
|
|
|
/* Copy up to 31 bytes. This divided into two cases 0-8 bytes and 9-31
|
|
bytes. Each case is handled without loops, using binary (1,2,4,8)
|
|
tests.
|
|
|
|
In the short (0-8 byte) case no attempt is made to force alignment
|
|
of either source or destination. The hardware will handle the
|
|
unaligned load/stores with small delays for crossing 32- 128-byte,
|
|
and 4096-byte boundaries. Since these short moves are unlikely to be
|
|
unaligned or cross these boundaries, the overhead to force
|
|
alignment is not justified.
|
|
|
|
The longer (9-31 byte) move is more likely to cross 32- or 128-byte
|
|
boundaries. Since only loads are sensitive to the 32-/128-byte
|
|
boundaries it is more important to align the source then the
|
|
destination. If the source is not already word aligned, we first
|
|
move 1-3 bytes as needed. Since we are only word aligned we don't
|
|
use double word load/stores to insure that all loads are aligned.
|
|
While the destination and stores may still be unaligned, this
|
|
is only an issue for page (4096 byte boundary) crossing, which
|
|
should be rare for these short moves. The hardware handles this
|
|
case automatically with a small (~20 cycle) delay. */
|
|
.align 4
|
|
|
|
cfi_same_value (31)
|
|
cfi_same_value (30)
|
|
L(word_unaligned_short):
|
|
mtcrf 0x01,5
|
|
cmplwi cr6,5,8
|
|
neg 8,4
|
|
clrrwi 9,4,2
|
|
andi. 0,8,3
|
|
beq cr6,L(wus_8) /* Handle moves of 8 bytes. */
|
|
/* At least 9 bytes left. Get the source word aligned. */
|
|
cmplwi cr1,5,16
|
|
mr 12,4
|
|
ble cr6,L(wus_4) /* Handle moves of 0-8 bytes. */
|
|
mr 11,3
|
|
mr 10,5
|
|
cmplwi cr6,0,2
|
|
beq L(wus_tail) /* If the source is already word aligned skip this. */
|
|
/* Copy 1-3 bytes to get source address word aligned. */
|
|
lwz 6,0(9)
|
|
subf 10,0,5
|
|
add 12,4,0
|
|
blt cr6,5f
|
|
srwi 7,6,16
|
|
bgt cr6,3f
|
|
#ifdef __LITTLE_ENDIAN__
|
|
sth 7,0(3)
|
|
#else
|
|
sth 6,0(3)
|
|
#endif
|
|
b 7f
|
|
.align 4
|
|
3:
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rotlwi 6,6,24
|
|
stb 6,0(3)
|
|
sth 7,1(3)
|
|
#else
|
|
stb 7,0(3)
|
|
sth 6,1(3)
|
|
#endif
|
|
b 7f
|
|
.align 4
|
|
5:
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rotlwi 6,6,8
|
|
#endif
|
|
stb 6,0(3)
|
|
7:
|
|
cmplwi cr1,10,16
|
|
add 11,3,0
|
|
mtcrf 0x01,10
|
|
.align 4
|
|
L(wus_tail):
|
|
/* At least 6 bytes left and the source is word aligned. This allows
|
|
some speculative loads up front. */
|
|
/* We need to special case the fall-through because the biggest delays
|
|
are due to address computation not being ready in time for the
|
|
AGEN. */
|
|
lwz 6,0(12)
|
|
lwz 7,4(12)
|
|
blt cr1,L(wus_tail8)
|
|
cmplwi cr0,10,24
|
|
L(wus_tail16): /* Move 16 bytes. */
|
|
stw 6,0(11)
|
|
stw 7,4(11)
|
|
lwz 6,8(12)
|
|
lwz 7,12(12)
|
|
stw 6,8(11)
|
|
stw 7,12(11)
|
|
/* Move 8 bytes more. */
|
|
bf 28,L(wus_tail16p8)
|
|
cmplwi cr1,10,28
|
|
lwz 6,16(12)
|
|
lwz 7,20(12)
|
|
stw 6,16(11)
|
|
stw 7,20(11)
|
|
/* Move 4 bytes more. */
|
|
bf 29,L(wus_tail16p4)
|
|
lwz 6,24(12)
|
|
stw 6,24(11)
|
|
addi 12,12,28
|
|
addi 11,11,28
|
|
bgt cr1,L(wus_tail2)
|
|
/* exactly 28 bytes. Return original dst pointer and exit. */
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wus_tail16p8): /* less than 8 bytes left. */
|
|
beq cr1,L(wus_tailX) /* exactly 16 bytes, early exit. */
|
|
cmplwi cr1,10,20
|
|
bf 29,L(wus_tail16p2)
|
|
/* Move 4 bytes more. */
|
|
lwz 6,16(12)
|
|
stw 6,16(11)
|
|
addi 12,12,20
|
|
addi 11,11,20
|
|
bgt cr1,L(wus_tail2)
|
|
/* exactly 20 bytes. Return original dst pointer and exit. */
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wus_tail16p4): /* less than 4 bytes left. */
|
|
addi 12,12,24
|
|
addi 11,11,24
|
|
bgt cr0,L(wus_tail2)
|
|
/* exactly 24 bytes. Return original dst pointer and exit. */
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wus_tail16p2): /* 16 bytes moved, less than 4 bytes left. */
|
|
addi 12,12,16
|
|
addi 11,11,16
|
|
b L(wus_tail2)
|
|
|
|
.align 4
|
|
L(wus_tail8): /* Move 8 bytes. */
|
|
/* r6, r7 already loaded speculatively. */
|
|
cmplwi cr1,10,8
|
|
cmplwi cr0,10,12
|
|
bf 28,L(wus_tail4)
|
|
stw 6,0(11)
|
|
stw 7,4(11)
|
|
/* Move 4 bytes more. */
|
|
bf 29,L(wus_tail8p4)
|
|
lwz 6,8(12)
|
|
stw 6,8(11)
|
|
addi 12,12,12
|
|
addi 11,11,12
|
|
bgt cr0,L(wus_tail2)
|
|
/* exactly 12 bytes. Return original dst pointer and exit. */
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wus_tail8p4): /* less than 4 bytes left. */
|
|
addi 12,12,8
|
|
addi 11,11,8
|
|
bgt cr1,L(wus_tail2)
|
|
/* exactly 8 bytes. Return original dst pointer and exit. */
|
|
addi 1,1,32
|
|
blr
|
|
|
|
.align 4
|
|
L(wus_tail4): /* Move 4 bytes. */
|
|
/* r6 already loaded speculatively. If we are here we know there is
|
|
more than 4 bytes left. So there is no need to test. */
|
|
addi 12,12,4
|
|
stw 6,0(11)
|
|
addi 11,11,4
|
|
L(wus_tail2): /* Move 2-3 bytes. */
|
|
bf 30,L(wus_tail1)
|
|
lhz 6,0(12)
|
|
sth 6,0(11)
|
|
bf 31,L(wus_tailX)
|
|
lbz 7,2(12)
|
|
stb 7,2(11)
|
|
addi 1,1,32
|
|
blr
|
|
L(wus_tail1): /* Move 1 byte. */
|
|
bf 31,L(wus_tailX)
|
|
lbz 6,0(12)
|
|
stb 6,0(11)
|
|
L(wus_tailX):
|
|
/* Return original dst pointer. */
|
|
addi 1,1,32
|
|
blr
|
|
|
|
/* Special case to copy 0-8 bytes. */
|
|
.align 4
|
|
L(wus_8):
|
|
lwz 6,0(4)
|
|
lwz 7,4(4)
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
/* Return original dst pointer. */
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wus_4):
|
|
bf 29,L(wus_2)
|
|
lwz 6,0(4)
|
|
stw 6,0(3)
|
|
bf 30,L(wus_5)
|
|
lhz 7,4(4)
|
|
sth 7,4(3)
|
|
bf 31,L(wus_0)
|
|
lbz 8,6(4)
|
|
stb 8,6(3)
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wus_5):
|
|
bf 31,L(wus_0)
|
|
lbz 6,4(4)
|
|
stb 6,4(3)
|
|
/* Return original dst pointer. */
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wus_2): /* Move 2-3 bytes. */
|
|
bf 30,L(wus_1)
|
|
lhz 6,0(4)
|
|
sth 6,0(3)
|
|
bf 31,L(wus_0)
|
|
lbz 7,2(4)
|
|
stb 7,2(3)
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wus_1): /* Move 1 byte. */
|
|
bf 31,L(wus_0)
|
|
lbz 6,0(4)
|
|
stb 6,0(3)
|
|
.align 3
|
|
L(wus_0):
|
|
/* Return original dst pointer. */
|
|
addi 1,1,32
|
|
blr
|
|
|
|
.align 4
|
|
cfi_offset(31,(24-32))
|
|
cfi_offset(30,(20-32))
|
|
L(wdu):
|
|
|
|
/* Copy words where the destination is aligned but the source is
|
|
not. For power4, power5 and power6 machines there is penalty for
|
|
unaligned loads (src) that cross 32-byte, cacheline, or page
|
|
boundaries. So we want to use simple (unaligned) loads where
|
|
possible but avoid them where we know the load would span a 32-byte
|
|
boundary.
|
|
|
|
At this point we know we have at least 29 (32-3) bytes to copy
|
|
the src is unaligned. and we may cross at least one 32-byte
|
|
boundary. Also we have the following register values:
|
|
r3 == adjusted dst, word aligned
|
|
r4 == unadjusted src
|
|
r5 == unadjusted len
|
|
r9 == adjusted Word length
|
|
r10 == src alignment (1-3)
|
|
r12 == adjusted src, not aligned
|
|
r31 == adjusted len
|
|
|
|
First we need to copy word up to but not crossing the next 32-byte
|
|
boundary. Then perform aligned loads just before and just after
|
|
the boundary and use shifts and or to generate the next aligned
|
|
word for dst. If more than 32 bytes remain we copy (unaligned src)
|
|
the next 7 words and repeat the loop until less than 32-bytes
|
|
remain.
|
|
|
|
Then if more than 4 bytes remain we again use aligned loads,
|
|
shifts and or to generate the next dst word. We then process the
|
|
remaining words using unaligned loads as needed. Finally we check
|
|
if there are more than 0 bytes (1-3) bytes remaining and use
|
|
halfword and or byte load/stores to complete the copy.
|
|
*/
|
|
mr 4,12 /* restore unaligned adjusted src ptr */
|
|
clrlwi 0,12,27 /* Find dist from previous 32-byte boundary. */
|
|
slwi 10,10,3 /* calculate number of bits to shift 1st word left */
|
|
cmplwi cr5,0,16
|
|
subfic 8,0,32 /* Number of bytes to next 32-byte boundary. */
|
|
|
|
mtcrf 0x01,8
|
|
cmplwi cr1,10,16
|
|
subfic 9,10,32 /* number of bits to shift 2nd word right */
|
|
/* This test is reversed because the timing to compare the bytes to
|
|
32-byte boundary could not be meet. So we compare the bytes from
|
|
previous 32-byte boundary and invert the test. */
|
|
bge cr5,L(wdu_h32_8)
|
|
.align 4
|
|
lwz 6,0(4)
|
|
lwz 7,4(4)
|
|
addi 12,4,16 /* generate alternate pointers to avoid agen */
|
|
addi 11,3,16 /* timing issues downstream. */
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
subi 31,31,16
|
|
lwz 6,8(4)
|
|
lwz 7,12(4)
|
|
addi 4,4,16
|
|
stw 6,8(3)
|
|
stw 7,12(3)
|
|
addi 3,3,16
|
|
bf 28,L(wdu_h32_4)
|
|
lwz 6,0(12)
|
|
lwz 7,4(12)
|
|
subi 31,31,8
|
|
addi 4,4,8
|
|
stw 6,0(11)
|
|
stw 7,4(11)
|
|
addi 3,3,8
|
|
bf 29,L(wdu_h32_0)
|
|
lwz 6,8(12)
|
|
addi 4,4,4
|
|
subi 31,31,4
|
|
stw 6,8(11)
|
|
addi 3,3,4
|
|
b L(wdu_h32_0)
|
|
.align 4
|
|
L(wdu_h32_8):
|
|
bf 28,L(wdu_h32_4)
|
|
lwz 6,0(4)
|
|
lwz 7,4(4)
|
|
subi 31,31,8
|
|
bf 29,L(wdu_h32_8x)
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
lwz 6,8(4)
|
|
addi 4,4,12
|
|
subi 31,31,4
|
|
stw 6,8(3)
|
|
addi 3,3,12
|
|
b L(wdu_h32_0)
|
|
.align 4
|
|
L(wdu_h32_8x):
|
|
addi 4,4,8
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
addi 3,3,8
|
|
b L(wdu_h32_0)
|
|
.align 4
|
|
L(wdu_h32_4):
|
|
bf 29,L(wdu_h32_0)
|
|
lwz 6,0(4)
|
|
subi 31,31,4
|
|
addi 4,4,4
|
|
stw 6,0(3)
|
|
addi 3,3,4
|
|
.align 4
|
|
L(wdu_h32_0):
|
|
/* set up for 32-byte boundary crossing word move and possibly 32-byte
|
|
move loop. */
|
|
clrrwi 12,4,2
|
|
cmplwi cr5,31,32
|
|
bge cr1,L(wdu2_32)
|
|
#if 0
|
|
b L(wdu1_32)
|
|
/*
|
|
cmplwi cr1,10,8
|
|
beq cr1,L(wdu1_32)
|
|
cmplwi cr1,10,16
|
|
beq cr1,L(wdu2_32)
|
|
cmplwi cr1,10,24
|
|
beq cr1,L(wdu3_32)
|
|
*/
|
|
L(wdu_32):
|
|
lwz 6,0(12)
|
|
cmplwi cr6,31,4
|
|
srwi 8,31,5 /* calculate the 32 byte loop count */
|
|
slw 0,6,10
|
|
clrlwi 31,31,27 /* The remaining bytes, < 32. */
|
|
blt cr5,L(wdu_32tail)
|
|
mtctr 8
|
|
cmplwi cr6,31,4
|
|
.align 4
|
|
L(wdu_loop32):
|
|
/* copy 32 bytes at a time */
|
|
lwz 8,4(12)
|
|
addi 12,12,32
|
|
lwz 7,4(4)
|
|
srw 8,8,9
|
|
or 0,0,8
|
|
stw 0,0(3)
|
|
stw 7,4(3)
|
|
lwz 6,8(4)
|
|
lwz 7,12(4)
|
|
stw 6,8(3)
|
|
stw 7,12(3)
|
|
lwz 6,16(4)
|
|
lwz 7,20(4)
|
|
stw 6,16(3)
|
|
stw 7,20(3)
|
|
lwz 6,24(4)
|
|
lwz 7,28(4)
|
|
lwz 8,0(12)
|
|
addi 4,4,32
|
|
stw 6,24(3)
|
|
stw 7,28(3)
|
|
addi 3,3,32
|
|
slw 0,8,10
|
|
bdnz+ L(wdu_loop32)
|
|
|
|
L(wdu_32tail):
|
|
mtcrf 0x01,31
|
|
cmplwi cr5,31,16
|
|
blt cr6,L(wdu_4tail)
|
|
/* calculate and store the final word */
|
|
lwz 8,4(12)
|
|
srw 8,8,9
|
|
or 6,0,8
|
|
b L(wdu_32tailx)
|
|
#endif
|
|
.align 4
|
|
L(wdu1_32):
|
|
lwz 6,-1(4)
|
|
cmplwi cr6,31,4
|
|
srwi 8,31,5 /* calculate the 32 byte loop count */
|
|
#ifdef __LITTLE_ENDIAN__
|
|
srwi 6,6,8
|
|
#else
|
|
slwi 6,6,8
|
|
#endif
|
|
clrlwi 31,31,27 /* The remaining bytes, < 32. */
|
|
blt cr5,L(wdu1_32tail)
|
|
mtctr 8
|
|
cmplwi cr6,31,4
|
|
|
|
lwz 8,3(4)
|
|
lwz 7,4(4)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,24,32
|
|
#else
|
|
/* Equivalent to: srwi 8,8,32-8; or 6,6,8 */
|
|
rlwimi 6,8,8,(32-8),31
|
|
#endif
|
|
b L(wdu1_loop32x)
|
|
.align 4
|
|
L(wdu1_loop32):
|
|
/* copy 32 bytes at a time */
|
|
lwz 8,3(4)
|
|
lwz 7,4(4)
|
|
stw 10,-8(3)
|
|
stw 11,-4(3)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,24,32
|
|
#else
|
|
/* Equivalent to srwi 8,8,32-8; or 6,6,8 */
|
|
rlwimi 6,8,8,(32-8),31
|
|
#endif
|
|
L(wdu1_loop32x):
|
|
lwz 10,8(4)
|
|
lwz 11,12(4)
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
lwz 6,16(4)
|
|
lwz 7,20(4)
|
|
stw 10,8(3)
|
|
stw 11,12(3)
|
|
lwz 10,24(4)
|
|
lwz 11,28(4)
|
|
lwz 8,32-1(4)
|
|
addi 4,4,32
|
|
stw 6,16(3)
|
|
stw 7,20(3)
|
|
addi 3,3,32
|
|
#ifdef __LITTLE_ENDIAN__
|
|
srwi 6,8,8
|
|
#else
|
|
slwi 6,8,8
|
|
#endif
|
|
bdnz+ L(wdu1_loop32)
|
|
stw 10,-8(3)
|
|
stw 11,-4(3)
|
|
|
|
L(wdu1_32tail):
|
|
mtcrf 0x01,31
|
|
cmplwi cr5,31,16
|
|
blt cr6,L(wdu_4tail)
|
|
/* calculate and store the final word */
|
|
lwz 8,3(4)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,24,32
|
|
#else
|
|
/* Equivalent to: srwi 8,8,32-8; or 6,6,8 */
|
|
rlwimi 6,8,8,(32-8),31
|
|
#endif
|
|
b L(wdu_32tailx)
|
|
|
|
L(wdu2_32):
|
|
bgt cr1,L(wdu3_32)
|
|
lwz 6,-2(4)
|
|
cmplwi cr6,31,4
|
|
srwi 8,31,5 /* calculate the 32 byte loop count */
|
|
#ifdef __LITTLE_ENDIAN__
|
|
srwi 6,6,16
|
|
#else
|
|
slwi 6,6,16
|
|
#endif
|
|
clrlwi 31,31,27 /* The remaining bytes, < 32. */
|
|
blt cr5,L(wdu2_32tail)
|
|
mtctr 8
|
|
cmplwi cr6,31,4
|
|
|
|
lwz 8,2(4)
|
|
lwz 7,4(4)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,16,32
|
|
#else
|
|
rlwimi 6,8,16,(32-16),31
|
|
#endif
|
|
b L(wdu2_loop32x)
|
|
.align 4
|
|
L(wdu2_loop32):
|
|
/* copy 32 bytes at a time */
|
|
lwz 8,2(4)
|
|
lwz 7,4(4)
|
|
stw 10,-8(3)
|
|
stw 11,-4(3)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,16,32
|
|
#else
|
|
rlwimi 6,8,16,(32-16),31
|
|
#endif
|
|
L(wdu2_loop32x):
|
|
lwz 10,8(4)
|
|
lwz 11,12(4)
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
lwz 6,16(4)
|
|
lwz 7,20(4)
|
|
stw 10,8(3)
|
|
stw 11,12(3)
|
|
lwz 10,24(4)
|
|
lwz 11,28(4)
|
|
/* lwz 8,0(12) */
|
|
lwz 8,32-2(4)
|
|
addi 4,4,32
|
|
stw 6,16(3)
|
|
stw 7,20(3)
|
|
addi 3,3,32
|
|
#ifdef __LITTLE_ENDIAN__
|
|
srwi 6,8,16
|
|
#else
|
|
slwi 6,8,16
|
|
#endif
|
|
bdnz+ L(wdu2_loop32)
|
|
stw 10,-8(3)
|
|
stw 11,-4(3)
|
|
|
|
L(wdu2_32tail):
|
|
mtcrf 0x01,31
|
|
cmplwi cr5,31,16
|
|
blt cr6,L(wdu_4tail)
|
|
/* calculate and store the final word */
|
|
lwz 8,2(4)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,16,32
|
|
#else
|
|
rlwimi 6,8,16,(32-16),31
|
|
#endif
|
|
b L(wdu_32tailx)
|
|
|
|
L(wdu3_32):
|
|
/* lwz 6,0(12) */
|
|
lwz 6,-3(4)
|
|
cmplwi cr6,31,4
|
|
srwi 8,31,5 /* calculate the 32 byte loop count */
|
|
#ifdef __LITTLE_ENDIAN__
|
|
srwi 6,6,24
|
|
#else
|
|
slwi 6,6,24
|
|
#endif
|
|
clrlwi 31,31,27 /* The remaining bytes, < 32. */
|
|
blt cr5,L(wdu3_32tail)
|
|
mtctr 8
|
|
cmplwi cr6,31,4
|
|
|
|
lwz 8,1(4)
|
|
lwz 7,4(4)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,8,32
|
|
#else
|
|
rlwimi 6,8,24,(32-24),31
|
|
#endif
|
|
b L(wdu3_loop32x)
|
|
.align 4
|
|
L(wdu3_loop32):
|
|
/* copy 32 bytes at a time */
|
|
lwz 8,1(4)
|
|
lwz 7,4(4)
|
|
stw 10,-8(3)
|
|
stw 11,-4(3)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,8,32
|
|
#else
|
|
rlwimi 6,8,24,(32-24),31
|
|
#endif
|
|
L(wdu3_loop32x):
|
|
lwz 10,8(4)
|
|
lwz 11,12(4)
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
lwz 6,16(4)
|
|
lwz 7,20(4)
|
|
stw 10,8(3)
|
|
stw 11,12(3)
|
|
lwz 10,24(4)
|
|
lwz 11,28(4)
|
|
lwz 8,32-3(4)
|
|
addi 4,4,32
|
|
stw 6,16(3)
|
|
stw 7,20(3)
|
|
addi 3,3,32
|
|
#ifdef __LITTLE_ENDIAN__
|
|
srwi 6,8,24
|
|
#else
|
|
slwi 6,8,24
|
|
#endif
|
|
bdnz+ L(wdu3_loop32)
|
|
stw 10,-8(3)
|
|
stw 11,-4(3)
|
|
|
|
L(wdu3_32tail):
|
|
mtcrf 0x01,31
|
|
cmplwi cr5,31,16
|
|
blt cr6,L(wdu_4tail)
|
|
/* calculate and store the final word */
|
|
lwz 8,1(4)
|
|
#ifdef __LITTLE_ENDIAN__
|
|
rldimi 6,8,8,32
|
|
#else
|
|
rlwimi 6,8,24,(32-24),31
|
|
#endif
|
|
b L(wdu_32tailx)
|
|
.align 4
|
|
L(wdu_32tailx):
|
|
blt cr5,L(wdu_t32_8)
|
|
lwz 7,4(4)
|
|
addi 12,4,16 /* generate alternate pointers to avoid agen */
|
|
addi 11,3,16 /* timing issues downstream. */
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
subi 31,31,16
|
|
lwz 6,8(4)
|
|
lwz 7,12(4)
|
|
addi 4,4,16
|
|
stw 6,8(3)
|
|
stw 7,12(3)
|
|
addi 3,3,16
|
|
bf 28,L(wdu_t32_4x)
|
|
lwz 6,0(12)
|
|
lwz 7,4(12)
|
|
addi 4,4,8
|
|
subi 31,31,8
|
|
stw 6,0(11)
|
|
stw 7,4(11)
|
|
addi 3,3,8
|
|
bf 29,L(wdu_t32_0)
|
|
lwz 6,8(12)
|
|
addi 4,4,4
|
|
subi 31,31,4
|
|
stw 6,8(11)
|
|
addi 3,3,4
|
|
b L(wdu_t32_0)
|
|
.align 4
|
|
L(wdu_t32_4x):
|
|
bf 29,L(wdu_t32_0)
|
|
lwz 6,0(4)
|
|
addi 4,4,4
|
|
subi 31,31,4
|
|
stw 6,0(3)
|
|
addi 3,3,4
|
|
b L(wdu_t32_0)
|
|
.align 4
|
|
L(wdu_t32_8):
|
|
bf 28,L(wdu_t32_4)
|
|
lwz 7,4(4)
|
|
subi 31,31,8
|
|
bf 29,L(wdu_t32_8x)
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
lwz 6,8(4)
|
|
subi 31,31,4
|
|
addi 4,4,12
|
|
stw 6,8(3)
|
|
addi 3,3,12
|
|
b L(wdu_t32_0)
|
|
.align 4
|
|
L(wdu_t32_8x):
|
|
addi 4,4,8
|
|
stw 6,0(3)
|
|
stw 7,4(3)
|
|
addi 3,3,8
|
|
b L(wdu_t32_0)
|
|
.align 4
|
|
L(wdu_t32_4):
|
|
subi 31,31,4
|
|
stw 6,0(3)
|
|
addi 4,4,4
|
|
addi 3,3,4
|
|
.align 4
|
|
L(wdu_t32_0):
|
|
L(wdu_4tail):
|
|
cmplwi cr6,31,0
|
|
beq cr6,L(wdus_0) /* If the tail is 0 bytes we are done! */
|
|
bf 30,L(wdus_3)
|
|
lhz 7,0(4)
|
|
sth 7,0(3)
|
|
bf 31,L(wdus_0)
|
|
lbz 8,2(4)
|
|
stb 8,2(3)
|
|
mr 3,30
|
|
lwz 30,20(1)
|
|
lwz 31,24(1)
|
|
addi 1,1,32
|
|
blr
|
|
.align 4
|
|
L(wdus_3):
|
|
bf 31,L(wus_0)
|
|
lbz 6,0(4)
|
|
stb 6,0(3)
|
|
.align 4
|
|
L(wdus_0):
|
|
/* Return original dst pointer. */
|
|
mr 3,30
|
|
lwz 30,20(1)
|
|
lwz 31,24(1)
|
|
addi 1,1,32
|
|
blr
|
|
END (memcpy)
|
|
|
|
libc_hidden_builtin_def (memcpy)
|