mirror of
https://sourceware.org/git/glibc.git
synced 2024-11-28 07:41:05 +00:00
30891f35fa
We stopped adding "Contributed by" or similar lines in sources in 2012 in favour of git logs and keeping the Contributors section of the glibc manual up to date. Removing these lines makes the license header a bit more consistent across files and also removes the possibility of error in attribution when license blocks or files are copied across since the contributed-by lines don't actually reflect reality in those cases. Move all "Contributed by" and similar lines (Written by, Test by, etc.) into a new file CONTRIBUTED-BY to retain record of these contributions. These contributors are also mentioned in manual/contrib.texi, so we just maintain this additional record as a courtesy to the earlier developers. The following scripts were used to filter a list of files to edit in place and to clean up the CONTRIBUTED-BY file respectively. These were not added to the glibc sources because they're not expected to be of any use in future given that this is a one time task: https://gist.github.com/siddhesh/b5ecac94eabfd72ed2916d6d8157e7dc https://gist.github.com/siddhesh/15ea1f5e435ace9774f485030695ee02 Reviewed-by: Carlos O'Donell <carlos@redhat.com>
194 lines
5.7 KiB
ArmAsm
194 lines
5.7 KiB
ArmAsm
/* Copyright (C) 1996-2021 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library. If not, see
|
|
<https://www.gnu.org/licenses/>. */
|
|
|
|
/* Bytewise compare two null-terminated strings. */
|
|
|
|
#include <sysdep.h>
|
|
|
|
.set noat
|
|
.set noreorder
|
|
|
|
.text
|
|
|
|
ENTRY(strcmp)
|
|
#ifdef PROF
|
|
ldgp gp, 0(pv)
|
|
lda AT, _mcount
|
|
jmp AT, (AT), _mcount
|
|
.prologue 1
|
|
#else
|
|
.prologue 0
|
|
#endif
|
|
|
|
ldq_u t0, 0(a0) # e0 : give cache time to catch up
|
|
xor a0, a1, t2 # .. e1 : are s1 and s2 co-aligned?
|
|
ldq_u t1, 0(a1) # e0 :
|
|
and t2, 7, t2 # .. e1 :
|
|
lda t3, -1 # e0 :
|
|
bne t2, $unaligned # .. e1 :
|
|
|
|
/* On entry to this basic block:
|
|
t0 == the first destination word for masking back in
|
|
t1 == the first source word.
|
|
t3 == -1. */
|
|
|
|
$aligned:
|
|
mskqh t3, a0, t3 # e0 :
|
|
nop # .. e1 :
|
|
ornot t1, t3, t1 # e0 :
|
|
ornot t0, t3, t0 # .. e1 :
|
|
cmpbge zero, t1, t7 # e0 : bits set iff null found
|
|
bne t7, $eos # e1 (zdb)
|
|
|
|
/* Aligned compare main loop.
|
|
On entry to this basic block:
|
|
t0 == an s1 word.
|
|
t1 == an s2 word not containing a null. */
|
|
|
|
$a_loop:
|
|
xor t0, t1, t2 # e0 :
|
|
bne t2, $wordcmp # .. e1 (zdb)
|
|
ldq_u t1, 8(a1) # e0 :
|
|
ldq_u t0, 8(a0) # .. e1 :
|
|
addq a1, 8, a1 # e0 :
|
|
addq a0, 8, a0 # .. e1 :
|
|
cmpbge zero, t1, t7 # e0 :
|
|
beq t7, $a_loop # .. e1 (zdb)
|
|
br $eos # e1 :
|
|
|
|
/* The two strings are not co-aligned. Align s1 and cope. */
|
|
|
|
$unaligned:
|
|
and a0, 7, t4 # e0 : find s1 misalignment
|
|
and a1, 7, t5 # .. e1 : find s2 misalignment
|
|
subq a1, t4, a1 # e0 :
|
|
|
|
/* If s2 misalignment is larger than s2 misalignment, we need
|
|
extra startup checks to avoid SEGV. */
|
|
|
|
cmplt t4, t5, t8 # .. e1 :
|
|
beq t8, $u_head # e1 :
|
|
|
|
mskqh t3, t5, t3 # e0 :
|
|
ornot t1, t3, t3 # e0 :
|
|
cmpbge zero, t3, t7 # e1 : is there a zero?
|
|
beq t7, $u_head # e1 :
|
|
|
|
/* We've found a zero in the first partial word of s2. Align
|
|
our current s1 and s2 words and compare what we've got. */
|
|
|
|
extql t1, t5, t1 # e0 :
|
|
extql t0, a0, t0 # e0 :
|
|
cmpbge zero, t1, t7 # .. e1 : find that zero again
|
|
br $eos # e1 : and finish up
|
|
|
|
.align 3
|
|
$u_head:
|
|
/* We know just enough now to be able to assemble the first
|
|
full word of s2. We can still find a zero at the end of it.
|
|
|
|
On entry to this basic block:
|
|
t0 == first word of s1
|
|
t1 == first partial word of s2. */
|
|
|
|
ldq_u t2, 8(a1) # e0 : load second partial s2 word
|
|
lda t3, -1 # .. e1 : create leading garbage mask
|
|
extql t1, a1, t1 # e0 : create first s2 word
|
|
mskqh t3, a0, t3 # e0 :
|
|
extqh t2, a1, t4 # e0 :
|
|
ornot t0, t3, t0 # .. e1 : kill s1 garbage
|
|
or t1, t4, t1 # e0 : s2 word now complete
|
|
cmpbge zero, t0, t7 # .. e1 : find zero in first s1 word
|
|
ornot t1, t3, t1 # e0 : kill s2 garbage
|
|
lda t3, -1 # .. e1 :
|
|
mskql t3, a1, t3 # e0 : mask for s2[1] bits we have seen
|
|
bne t7, $eos # .. e1 :
|
|
xor t0, t1, t4 # e0 : compare aligned words
|
|
bne t4, $wordcmp # .. e1 (zdb)
|
|
or t2, t3, t3 # e0 :
|
|
cmpbge zero, t3, t7 # e1 :
|
|
bne t7, $u_final # e1 :
|
|
|
|
/* Unaligned copy main loop. In order to avoid reading too much,
|
|
the loop is structured to detect zeros in aligned words from s2.
|
|
This has, unfortunately, effectively pulled half of a loop
|
|
iteration out into the head and half into the tail, but it does
|
|
prevent nastiness from accumulating in the very thing we want
|
|
to run as fast as possible.
|
|
|
|
On entry to this basic block:
|
|
t2 == the unshifted low-bits from the next s2 word. */
|
|
|
|
.align 3
|
|
$u_loop:
|
|
extql t2, a1, t3 # e0 :
|
|
ldq_u t2, 16(a1) # .. e1 : load next s2 high bits
|
|
ldq_u t0, 8(a0) # e0 : load next s1 word
|
|
addq a1, 8, a1 # .. e1 :
|
|
addq a0, 8, a0 # e0 :
|
|
nop # .. e1 :
|
|
extqh t2, a1, t1 # e0 :
|
|
cmpbge zero, t0, t7 # .. e1 : find zero in current s1 word
|
|
or t1, t3, t1 # e0 :
|
|
bne t7, $eos # .. e1 :
|
|
xor t0, t1, t4 # e0 : compare the words
|
|
bne t4, $wordcmp # .. e1 (zdb)
|
|
cmpbge zero, t2, t4 # e0 : find zero in next low bits
|
|
beq t4, $u_loop # .. e1 (zdb)
|
|
|
|
/* We've found a zero in the low bits of the last s2 word. Get
|
|
the next s1 word and align them. */
|
|
$u_final:
|
|
ldq_u t0, 8(a0) # e1 :
|
|
extql t2, a1, t1 # .. e0 :
|
|
cmpbge zero, t1, t7 # e0 :
|
|
|
|
/* We've found a zero somewhere in a word we just read.
|
|
On entry to this basic block:
|
|
t0 == s1 word
|
|
t1 == s2 word
|
|
t7 == cmpbge mask containing the zero. */
|
|
|
|
.align 3
|
|
$eos:
|
|
negq t7, t6 # e0 : create bytemask of valid data
|
|
and t6, t7, t8 # e1 :
|
|
subq t8, 1, t6 # e0 :
|
|
or t6, t8, t7 # e1 :
|
|
zapnot t0, t7, t0 # e0 : kill the garbage
|
|
zapnot t1, t7, t1 # .. e1 :
|
|
xor t0, t1, v0 # e0 : and compare
|
|
beq v0, $done # .. e1 :
|
|
|
|
/* Here we have two differing co-aligned words in t0 & t1.
|
|
Bytewise compare them and return (t0 > t1 ? 1 : -1). */
|
|
$wordcmp:
|
|
cmpbge t0, t1, t2 # e0 : comparison yields bit mask of ge
|
|
cmpbge t1, t0, t3 # .. e1 :
|
|
xor t2, t3, t0 # e0 : bits set iff t0/t1 bytes differ
|
|
negq t0, t1 # e1 : clear all but least bit
|
|
and t0, t1, t0 # e0 :
|
|
lda v0, -1 # .. e1 :
|
|
and t0, t2, t1 # e0 : was bit set in t0 > t1?
|
|
cmovne t1, 1, v0 # .. e1 (zdb)
|
|
|
|
$done:
|
|
ret # e1 :
|
|
|
|
END(strcmp)
|
|
libc_hidden_builtin_def (strcmp)
|