mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-23 03:10:05 +00:00
436 lines
11 KiB
ArmAsm
436 lines
11 KiB
ArmAsm
/* Optimized version of the standard memcpy() function.
|
|
This file is part of the GNU C Library.
|
|
Copyright (C) 2000, 2001, 2003 Free Software Foundation, Inc.
|
|
Contributed by Dan Pop for Itanium <Dan.Pop@cern.ch>.
|
|
Rewritten for McKinley by Sverre Jarp, HP Labs/CERN <Sverre.Jarp@cern.ch>
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
/* Return: dest
|
|
|
|
Inputs:
|
|
in0: dest
|
|
in1: src
|
|
in2: byte count
|
|
|
|
An assembly implementation of the algorithm used by the generic C
|
|
version from glibc. The case when source and sest are aligned is
|
|
treated separately, for extra performance.
|
|
|
|
In this form, memcpy assumes little endian mode. For big endian mode,
|
|
sh1 must be computed using an extra instruction: sub sh1 = 64, sh1
|
|
and the order of r[MEMLAT] and r[MEMLAT+1] must be reverted in the
|
|
shrp instruction. */
|
|
|
|
#define USE_LFETCH
|
|
#define USE_FLP
|
|
#include <sysdep.h>
|
|
#undef ret
|
|
|
|
#define LFETCH_DIST 500
|
|
|
|
#define ALIGN_UNROLL_no 4 // no. of elements
|
|
#define ALIGN_UNROLL_sh 2 // (shift amount)
|
|
|
|
#define MEMLAT 8
|
|
#define Nrot ((4*(MEMLAT+2) + 7) & ~7)
|
|
|
|
#define OP_T_THRES 16
|
|
#define OPSIZ 8
|
|
|
|
#define loopcnt r14
|
|
#define elemcnt r15
|
|
#define saved_pr r16
|
|
#define saved_lc r17
|
|
#define adest r18
|
|
#define dest r19
|
|
#define asrc r20
|
|
#define src r21
|
|
#define len r22
|
|
#define tmp2 r23
|
|
#define tmp3 r24
|
|
#define tmp4 r25
|
|
#define ptable r26
|
|
#define ploop56 r27
|
|
#define loopaddr r28
|
|
#define sh1 r29
|
|
#define ptr1 r30
|
|
#define ptr2 r31
|
|
|
|
#define movi0 mov
|
|
|
|
#define p_scr p6
|
|
#define p_xtr p7
|
|
#define p_nxtr p8
|
|
#define p_few p9
|
|
|
|
#if defined(USE_FLP)
|
|
#define load ldf8
|
|
#define store stf8
|
|
#define tempreg f6
|
|
#define the_r fr
|
|
#define the_s fs
|
|
#define the_t ft
|
|
#define the_q fq
|
|
#define the_w fw
|
|
#define the_x fx
|
|
#define the_y fy
|
|
#define the_z fz
|
|
#elif defined(USE_INT)
|
|
#define load ld8
|
|
#define store st8
|
|
#define tempreg tmp2
|
|
#define the_r r
|
|
#define the_s s
|
|
#define the_t t
|
|
#define the_q q
|
|
#define the_w w
|
|
#define the_x x
|
|
#define the_y y
|
|
#define the_z z
|
|
#endif
|
|
|
|
#ifdef GAS_ALIGN_BREAKS_UNWIND_INFO
|
|
/* Manually force proper loop-alignment. Note: be sure to
|
|
double-check the code-layout after making any changes to
|
|
this routine! */
|
|
# define ALIGN(n) { nop 0 }
|
|
#else
|
|
# define ALIGN(n) .align n
|
|
#endif
|
|
|
|
#if defined(USE_LFETCH)
|
|
#define LOOP(shift) \
|
|
ALIGN(32); \
|
|
.loop##shift##: \
|
|
{ .mmb \
|
|
(p[0]) ld8.nt1 r[0] = [asrc], 8 ; \
|
|
(p[0]) lfetch.nt1 [ptr1], 16 ; \
|
|
nop.b 0 ; \
|
|
} { .mib \
|
|
(p[MEMLAT+1]) st8 [dest] = tmp3, 8 ; \
|
|
(p[MEMLAT]) shrp tmp3 = r[MEMLAT], s[MEMLAT+1], shift ; \
|
|
nop.b 0 ;; \
|
|
} { .mmb \
|
|
(p[0]) ld8.nt1 s[0] = [asrc], 8 ; \
|
|
(p[0]) lfetch.nt1 [ptr2], 16 ; \
|
|
nop.b 0 ; \
|
|
} { .mib \
|
|
(p[MEMLAT+1]) st8 [dest] = tmp4, 8 ; \
|
|
(p[MEMLAT]) shrp tmp4 = s[MEMLAT], r[MEMLAT], shift ; \
|
|
br.ctop.sptk.many .loop##shift \
|
|
;; } \
|
|
{ .mib \
|
|
br.cond.sptk.many .copy_bytes ; /* deal with the remaining bytes */ \
|
|
}
|
|
#else
|
|
#define LOOP(shift) \
|
|
ALIGN(32); \
|
|
.loop##shift##: \
|
|
{ .mmb \
|
|
(p[0]) ld8.nt1 r[0] = [asrc], 8 ; \
|
|
nop.b 0 ; \
|
|
} { .mib \
|
|
(p[MEMLAT+1]) st8 [dest] = tmp3, 8 ; \
|
|
(p[MEMLAT]) shrp tmp3 = r[MEMLAT], s[MEMLAT+1], shift ; \
|
|
nop.b 0 ;; \
|
|
} { .mmb \
|
|
(p[0]) ld8.nt1 s[0] = [asrc], 8 ; \
|
|
nop.b 0 ; \
|
|
} { .mib \
|
|
(p[MEMLAT+1]) st8 [dest] = tmp4, 8 ; \
|
|
(p[MEMLAT]) shrp tmp4 = s[MEMLAT], r[MEMLAT], shift ; \
|
|
br.ctop.sptk.many .loop##shift \
|
|
;; } \
|
|
{ .mib \
|
|
br.cond.sptk.many .copy_bytes ; /* deal with the remaining bytes */ \
|
|
}
|
|
#endif
|
|
|
|
|
|
ENTRY(memcpy)
|
|
{ .mmi
|
|
.prologue
|
|
alloc r2 = ar.pfs, 3, Nrot - 3, 0, Nrot
|
|
.rotr r[MEMLAT+1], s[MEMLAT+2], q[MEMLAT+1], t[MEMLAT+1]
|
|
.rotp p[MEMLAT+2]
|
|
.rotf fr[MEMLAT+1], fq[MEMLAT+1], fs[MEMLAT+1], ft[MEMLAT+1]
|
|
mov ret0 = in0 // return tmp2 = dest
|
|
.save pr, saved_pr
|
|
movi0 saved_pr = pr // save the predicate registers
|
|
} { .mmi
|
|
and tmp4 = 7, in0 // check if destination is aligned
|
|
mov dest = in0 // dest
|
|
mov src = in1 // src
|
|
;; }
|
|
{ .mii
|
|
cmp.eq p_scr, p0 = in2, r0 // if (len == 0)
|
|
.save ar.lc, saved_lc
|
|
movi0 saved_lc = ar.lc // save the loop counter
|
|
.body
|
|
cmp.ge p_few, p0 = OP_T_THRES, in2 // is len <= OP_T_THRESH
|
|
} { .mbb
|
|
mov len = in2 // len
|
|
(p_scr) br.cond.dpnt.few .restore_and_exit // Branch no. 1: return dest
|
|
(p_few) br.cond.dpnt.many .copy_bytes // Branch no. 2: copy byte by byte
|
|
;; }
|
|
{ .mmi
|
|
#if defined(USE_LFETCH)
|
|
lfetch.nt1 [dest] //
|
|
lfetch.nt1 [src] //
|
|
#endif
|
|
shr.u elemcnt = len, 3 // elemcnt = len / 8
|
|
} { .mib
|
|
cmp.eq p_scr, p0 = tmp4, r0 // is destination aligned?
|
|
sub loopcnt = 7, tmp4 //
|
|
(p_scr) br.cond.dptk.many .dest_aligned
|
|
;; }
|
|
{ .mmi
|
|
ld1 tmp2 = [src], 1 //
|
|
sub len = len, loopcnt, 1 // reduce len
|
|
movi0 ar.lc = loopcnt //
|
|
} { .mib
|
|
cmp.ne p_scr, p0 = 0, loopcnt // avoid loading beyond end-point
|
|
;; }
|
|
|
|
.l0: // ---------------------------- // L0: Align src on 8-byte boundary
|
|
{ .mmi
|
|
st1 [dest] = tmp2, 1 //
|
|
(p_scr) ld1 tmp2 = [src], 1 //
|
|
} { .mib
|
|
cmp.lt p_scr, p0 = 1, loopcnt // avoid load beyond end-point
|
|
add loopcnt = -1, loopcnt
|
|
br.cloop.dptk.few .l0 //
|
|
;; }
|
|
|
|
.dest_aligned:
|
|
{ .mmi
|
|
and tmp4 = 7, src // ready for alignment check
|
|
shr.u elemcnt = len, 3 // elemcnt = len / 8
|
|
;; }
|
|
{ .mib
|
|
cmp.ne p_scr, p0 = tmp4, r0 // is source also aligned
|
|
tbit.nz p_xtr, p_nxtr = src, 3 // prepare a separate move if src
|
|
} { .mib // is not 16B aligned
|
|
add ptr2 = LFETCH_DIST, dest // prefetch address
|
|
add ptr1 = LFETCH_DIST, src
|
|
(p_scr) br.cond.dptk.many .src_not_aligned
|
|
;; }
|
|
|
|
// The optimal case, when dest, and src are aligned
|
|
|
|
.both_aligned:
|
|
{ .mmi
|
|
.pred.rel "mutex",p_xtr,p_nxtr
|
|
(p_xtr) cmp.gt p_scr, p0 = ALIGN_UNROLL_no+1, elemcnt // Need N + 1 to qualify
|
|
(p_nxtr) cmp.gt p_scr, p0 = ALIGN_UNROLL_no, elemcnt // Need only N to qualify
|
|
movi0 pr.rot = 1 << 16 // set rotating predicates
|
|
} { .mib
|
|
(p_scr) br.cond.dpnt.many .copy_full_words
|
|
;; }
|
|
|
|
{ .mmi
|
|
(p_xtr) load tempreg = [src], 8
|
|
(p_xtr) add elemcnt = -1, elemcnt
|
|
movi0 ar.ec = MEMLAT + 1 // set the epilog counter
|
|
;; }
|
|
{ .mmi
|
|
(p_xtr) add len = -8, len //
|
|
add asrc = 16, src // one bank apart (for USE_INT)
|
|
shr.u loopcnt = elemcnt, ALIGN_UNROLL_sh // cater for unrolling
|
|
;;}
|
|
{ .mmi
|
|
add loopcnt = -1, loopcnt
|
|
(p_xtr) store [dest] = tempreg, 8 // copy the "extra" word
|
|
nop.i 0
|
|
;; }
|
|
{ .mib
|
|
add adest = 16, dest
|
|
movi0 ar.lc = loopcnt // set the loop counter
|
|
;; }
|
|
|
|
#ifdef GAS_ALIGN_BREAKS_UNWIND_INFO
|
|
{ nop 0 }
|
|
#else
|
|
.align 32
|
|
#endif
|
|
#if defined(USE_FLP)
|
|
.l1: // ------------------------------- // L1: Everything a multiple of 8
|
|
{ .mmi
|
|
#if defined(USE_LFETCH)
|
|
(p[0]) lfetch.nt1 [ptr2],32
|
|
#endif
|
|
(p[0]) ldfp8 the_r[0],the_q[0] = [src], 16
|
|
(p[0]) add len = -32, len
|
|
} {.mmb
|
|
(p[MEMLAT]) store [dest] = the_r[MEMLAT], 8
|
|
(p[MEMLAT]) store [adest] = the_s[MEMLAT], 8
|
|
;; }
|
|
{ .mmi
|
|
#if defined(USE_LFETCH)
|
|
(p[0]) lfetch.nt1 [ptr1],32
|
|
#endif
|
|
(p[0]) ldfp8 the_s[0], the_t[0] = [src], 16
|
|
} {.mmb
|
|
(p[MEMLAT]) store [dest] = the_q[MEMLAT], 24
|
|
(p[MEMLAT]) store [adest] = the_t[MEMLAT], 24
|
|
br.ctop.dptk.many .l1
|
|
;; }
|
|
#elif defined(USE_INT)
|
|
.l1: // ------------------------------- // L1: Everything a multiple of 8
|
|
{ .mmi
|
|
(p[0]) load the_r[0] = [src], 8
|
|
(p[0]) load the_q[0] = [asrc], 8
|
|
(p[0]) add len = -32, len
|
|
} {.mmb
|
|
(p[MEMLAT]) store [dest] = the_r[MEMLAT], 8
|
|
(p[MEMLAT]) store [adest] = the_q[MEMLAT], 8
|
|
;; }
|
|
{ .mmi
|
|
(p[0]) load the_s[0] = [src], 24
|
|
(p[0]) load the_t[0] = [asrc], 24
|
|
} {.mmb
|
|
(p[MEMLAT]) store [dest] = the_s[MEMLAT], 24
|
|
(p[MEMLAT]) store [adest] = the_t[MEMLAT], 24
|
|
#if defined(USE_LFETCH)
|
|
;; }
|
|
{ .mmb
|
|
(p[0]) lfetch.nt1 [ptr2],32
|
|
(p[0]) lfetch.nt1 [ptr1],32
|
|
#endif
|
|
br.ctop.dptk.many .l1
|
|
;; }
|
|
#endif
|
|
|
|
.copy_full_words:
|
|
{ .mib
|
|
cmp.gt p_scr, p0 = 8, len //
|
|
shr.u elemcnt = len, 3 //
|
|
(p_scr) br.cond.dpnt.many .copy_bytes
|
|
;; }
|
|
{ .mii
|
|
load tempreg = [src], 8
|
|
add loopcnt = -1, elemcnt //
|
|
;; }
|
|
{ .mii
|
|
cmp.ne p_scr, p0 = 0, loopcnt //
|
|
mov ar.lc = loopcnt //
|
|
;; }
|
|
|
|
.l2: // ------------------------------- // L2: Max 4 words copied separately
|
|
{ .mmi
|
|
store [dest] = tempreg, 8
|
|
(p_scr) load tempreg = [src], 8 //
|
|
add len = -8, len
|
|
} { .mib
|
|
cmp.lt p_scr, p0 = 1, loopcnt // avoid load beyond end-point
|
|
add loopcnt = -1, loopcnt
|
|
br.cloop.dptk.few .l2
|
|
;; }
|
|
|
|
.copy_bytes:
|
|
{ .mib
|
|
cmp.eq p_scr, p0 = len, r0 // is len == 0 ?
|
|
add loopcnt = -1, len // len--;
|
|
(p_scr) br.cond.spnt .restore_and_exit
|
|
;; }
|
|
{ .mii
|
|
ld1 tmp2 = [src], 1
|
|
movi0 ar.lc = loopcnt
|
|
cmp.ne p_scr, p0 = 0, loopcnt // avoid load beyond end-point
|
|
;; }
|
|
|
|
.l3: // ------------------------------- // L3: Final byte move
|
|
{ .mmi
|
|
st1 [dest] = tmp2, 1
|
|
(p_scr) ld1 tmp2 = [src], 1
|
|
} { .mib
|
|
cmp.lt p_scr, p0 = 1, loopcnt // avoid load beyond end-point
|
|
add loopcnt = -1, loopcnt
|
|
br.cloop.dptk.few .l3
|
|
;; }
|
|
|
|
.restore_and_exit:
|
|
{ .mmi
|
|
movi0 pr = saved_pr, -1 // restore the predicate registers
|
|
;; }
|
|
{ .mib
|
|
movi0 ar.lc = saved_lc // restore the loop counter
|
|
br.ret.sptk.many b0
|
|
;; }
|
|
|
|
|
|
.src_not_aligned:
|
|
{ .mmi
|
|
cmp.gt p_scr, p0 = 16, len
|
|
and sh1 = 7, src // sh1 = src % 8
|
|
shr.u loopcnt = len, 4 // element-cnt = len / 16
|
|
} { .mib
|
|
add tmp4 = @ltoff(.table), gp
|
|
add tmp3 = @ltoff(.loop56), gp
|
|
(p_scr) br.cond.dpnt.many .copy_bytes // do byte by byte if too few
|
|
;; }
|
|
{ .mmi
|
|
and asrc = -8, src // asrc = (-8) -- align src for loop
|
|
add loopcnt = -1, loopcnt // loopcnt--
|
|
shl sh1 = sh1, 3 // sh1 = 8 * (src % 8)
|
|
} { .mmi
|
|
ld8 ptable = [tmp4] // ptable = &table
|
|
ld8 ploop56 = [tmp3] // ploop56 = &loop56
|
|
and tmp2 = -16, len // tmp2 = len & -OPSIZ
|
|
;; }
|
|
{ .mmi
|
|
add tmp3 = ptable, sh1 // tmp3 = &table + sh1
|
|
add src = src, tmp2 // src += len & (-16)
|
|
movi0 ar.lc = loopcnt // set LC
|
|
;; }
|
|
{ .mmi
|
|
ld8 tmp4 = [tmp3] // tmp4 = loop offset
|
|
sub len = len, tmp2 // len -= len & (-16)
|
|
movi0 ar.ec = MEMLAT + 2 // one more pass needed
|
|
;; }
|
|
{ .mmi
|
|
ld8 s[1] = [asrc], 8 // preload
|
|
sub loopaddr = ploop56,tmp4 // loopadd = &loop56 - loop offset
|
|
movi0 pr.rot = 1 << 16 // set rotating predicates
|
|
;; }
|
|
{ .mib
|
|
nop.m 0
|
|
movi0 b6 = loopaddr
|
|
br b6 // jump to the appropriate loop
|
|
;; }
|
|
|
|
LOOP(8)
|
|
LOOP(16)
|
|
LOOP(24)
|
|
LOOP(32)
|
|
LOOP(40)
|
|
LOOP(48)
|
|
LOOP(56)
|
|
END(memcpy)
|
|
libc_hidden_builtin_def (memcpy)
|
|
|
|
.rodata
|
|
.align 8
|
|
.table:
|
|
data8 0 // dummy entry
|
|
data8 .loop56 - .loop8
|
|
data8 .loop56 - .loop16
|
|
data8 .loop56 - .loop24
|
|
data8 .loop56 - .loop32
|
|
data8 .loop56 - .loop40
|
|
data8 .loop56 - .loop48
|
|
data8 .loop56 - .loop56
|