mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-02 01:40:07 +00:00
743b9c2a98
This patch removes the ununsed ARM code path for armv6t2 memchr and strlen and armv7 memch and strcmp. In all implementation, the ARM code is not used in any possible build (unless glibc is explicit build with the non-documented NO_THUMB compiler flag) and for armv7 the resulting code either produces wrong results (memchr) and throw build error (strcmp). Checked on arm-linux-gnueabihf built targeting both armv6 and armv7. * sysdeps/arm/armv6t2/memchr.S (memchr): Remove ARM code path. * sysdeps/arm/armv6t2/strlen.S (memchr): Likewise. * sysdeps/arm/armv7/multiarch/memchr_neon.S (memchr): Likewise. * sysdeps/arm/armv7/strcmp.S (strcmp): Likewise.
203 lines
5.4 KiB
ArmAsm
203 lines
5.4 KiB
ArmAsm
/* memchr implemented using NEON.
|
|
Copyright (C) 2011-2018 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library. If not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include <sysdep.h>
|
|
|
|
/* For __ARM_NEON__ this file defines memchr. */
|
|
#ifndef __ARM_NEON__
|
|
# define memchr __memchr_neon
|
|
# undef libc_hidden_builtin_def
|
|
# define libc_hidden_builtin_def(a)
|
|
#endif
|
|
|
|
.arch armv7-a
|
|
.fpu neon
|
|
|
|
|
|
/* Arguments */
|
|
#define srcin r0
|
|
#define chrin r1
|
|
#define cntin r2
|
|
|
|
/* Retval */
|
|
#define result r0 /* Live range does not overlap with srcin */
|
|
|
|
/* Working registers */
|
|
#define src r1 /* Live range does not overlap with chrin */
|
|
#define tmp r3
|
|
#define synd r0 /* No overlap with srcin or result */
|
|
#define soff r12
|
|
|
|
/* Working NEON registers */
|
|
#define vrepchr q0
|
|
#define vdata0 q1
|
|
#define vdata0_0 d2 /* Lower half of vdata0 */
|
|
#define vdata0_1 d3 /* Upper half of vdata0 */
|
|
#define vdata1 q2
|
|
#define vdata1_0 d4 /* Lower half of vhas_chr0 */
|
|
#define vdata1_1 d5 /* Upper half of vhas_chr0 */
|
|
#define vrepmask q3
|
|
#define vrepmask0 d6
|
|
#define vrepmask1 d7
|
|
#define vend q4
|
|
#define vend0 d8
|
|
#define vend1 d9
|
|
|
|
/*
|
|
* Core algorithm:
|
|
*
|
|
* For each 32-byte chunk we calculate a 32-bit syndrome value, with one bit per
|
|
* byte. Each bit is set if the relevant byte matched the requested character
|
|
* and cleared otherwise. Since the bits in the syndrome reflect exactly the
|
|
* order in which things occur in the original string, counting trailing zeros
|
|
* allows to identify exactly which byte has matched.
|
|
*/
|
|
|
|
.thumb_func
|
|
.p2align 4,,15
|
|
|
|
ENTRY(memchr)
|
|
/* Use a simple loop if there are less than 8 bytes to search. */
|
|
cmp cntin, #7
|
|
bhi .Llargestr
|
|
and chrin, chrin, #0xff
|
|
|
|
.Lsmallstr:
|
|
subs cntin, cntin, #1
|
|
blo .Lnotfound /* Return not found if reached end. */
|
|
ldrb tmp, [srcin], #1
|
|
cmp tmp, chrin
|
|
bne .Lsmallstr /* Loop again if not found. */
|
|
/* Otherwise fixup address and return. */
|
|
sub result, srcin, #1
|
|
bx lr
|
|
|
|
|
|
.Llargestr:
|
|
vdup.8 vrepchr, chrin /* Duplicate char across all lanes. */
|
|
/*
|
|
* Magic constant 0x8040201008040201 allows us to identify which lane
|
|
* matches the requested byte.
|
|
*/
|
|
movw tmp, #0x0201
|
|
movt tmp, #0x0804
|
|
lsl soff, tmp, #4
|
|
vmov vrepmask0, tmp, soff
|
|
vmov vrepmask1, tmp, soff
|
|
/* Work with aligned 32-byte chunks */
|
|
bic src, srcin, #31
|
|
ands soff, srcin, #31
|
|
beq .Lloopintro /* Go straight to main loop if it's aligned. */
|
|
|
|
/*
|
|
* Input string is not 32-byte aligned. We calculate the syndrome
|
|
* value for the aligned 32 bytes block containing the first bytes
|
|
* and mask the irrelevant part.
|
|
*/
|
|
vld1.8 {vdata0, vdata1}, [src:256]!
|
|
sub tmp, soff, #32
|
|
adds cntin, cntin, tmp
|
|
vceq.i8 vdata0, vdata0, vrepchr
|
|
vceq.i8 vdata1, vdata1, vrepchr
|
|
vand vdata0, vdata0, vrepmask
|
|
vand vdata1, vdata1, vrepmask
|
|
vpadd.i8 vdata0_0, vdata0_0, vdata0_1
|
|
vpadd.i8 vdata1_0, vdata1_0, vdata1_1
|
|
vpadd.i8 vdata0_0, vdata0_0, vdata1_0
|
|
vpadd.i8 vdata0_0, vdata0_0, vdata0_0
|
|
vmov synd, vdata0_0[0]
|
|
|
|
/* Clear the soff lower bits */
|
|
lsr synd, synd, soff
|
|
lsl synd, synd, soff
|
|
/* The first block can also be the last */
|
|
bls .Lmasklast
|
|
/* Have we found something already? */
|
|
cbnz synd, .Ltail
|
|
|
|
|
|
.Lloopintro:
|
|
vpush {vend}
|
|
/* 264/265 correspond to d8/d9 for q4 */
|
|
cfi_adjust_cfa_offset (16)
|
|
cfi_rel_offset (264, 0)
|
|
cfi_rel_offset (265, 8)
|
|
.p2align 3,,7
|
|
.Lloop:
|
|
vld1.8 {vdata0, vdata1}, [src:256]!
|
|
subs cntin, cntin, #32
|
|
vceq.i8 vdata0, vdata0, vrepchr
|
|
vceq.i8 vdata1, vdata1, vrepchr
|
|
/* If we're out of data we finish regardless of the result. */
|
|
bls .Lend
|
|
/* Use a fast check for the termination condition. */
|
|
vorr vend, vdata0, vdata1
|
|
vorr vend0, vend0, vend1
|
|
vmov synd, tmp, vend0
|
|
orrs synd, synd, tmp
|
|
/* We're not out of data, loop if we haven't found the character. */
|
|
beq .Lloop
|
|
|
|
.Lend:
|
|
vpop {vend}
|
|
cfi_adjust_cfa_offset (-16)
|
|
cfi_restore (264)
|
|
cfi_restore (265)
|
|
|
|
/* Termination condition found, let's calculate the syndrome value. */
|
|
vand vdata0, vdata0, vrepmask
|
|
vand vdata1, vdata1, vrepmask
|
|
vpadd.i8 vdata0_0, vdata0_0, vdata0_1
|
|
vpadd.i8 vdata1_0, vdata1_0, vdata1_1
|
|
vpadd.i8 vdata0_0, vdata0_0, vdata1_0
|
|
vpadd.i8 vdata0_0, vdata0_0, vdata0_0
|
|
vmov synd, vdata0_0[0]
|
|
cbz synd, .Lnotfound
|
|
bhi .Ltail /* Uses the condition code from
|
|
subs cntin, cntin, #32 above. */
|
|
|
|
|
|
.Lmasklast:
|
|
/* Clear the (-cntin) upper bits to avoid out-of-bounds matches. */
|
|
neg cntin, cntin
|
|
lsl synd, synd, cntin
|
|
lsrs synd, synd, cntin
|
|
it eq
|
|
moveq src, #0 /* If no match, set src to 0 so the retval is 0. */
|
|
|
|
|
|
.Ltail:
|
|
/* Count the trailing zeros using bit reversing */
|
|
rbit synd, synd
|
|
/* Compensate the last post-increment */
|
|
sub src, src, #32
|
|
/* Count the leading zeros */
|
|
clz synd, synd
|
|
/* Compute the potential result and return */
|
|
add result, src, synd
|
|
bx lr
|
|
|
|
|
|
.Lnotfound:
|
|
/* Set result to NULL if not found and return */
|
|
mov result, #0
|
|
bx lr
|
|
|
|
END(memchr)
|
|
libc_hidden_builtin_def (memchr)
|