mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-04 02:40:06 +00:00
ceabdcd130
1. Add default ISA level selection in non-multiarch/rtld implementations. 2. Add ISA level build guards to different implementations. - I.e strcmp-avx2.S which is ISA level 3 will only build if compiled ISA level <= 3. Otherwise there is no reason to include it as we will always use one of the ISA level 4 implementations (strcmp-evex.S). 3. Refactor the ifunc selector and ifunc implementation list to use the ISA level aware wrapper macros that allow functions below the compiled ISA level (with a guranteed replacement) to be skipped. Tested with and without multiarch on x86_64 for ISA levels: {generic, x86-64-v2, x86-64-v3, x86-64-v4} And m32 with and without multiarch.
243 lines
3.8 KiB
ArmAsm
243 lines
3.8 KiB
ArmAsm
/* wcslen optimized with SSE2.
|
|
Copyright (C) 2017-2022 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<https://www.gnu.org/licenses/>. */
|
|
|
|
#include <isa-level.h>
|
|
|
|
#if ISA_SHOULD_BUILD (1)
|
|
|
|
# include <sysdep.h>
|
|
|
|
#ifndef WCSLEN
|
|
# define WCSLEN __wcslen_sse2
|
|
#endif
|
|
|
|
|
|
.text
|
|
ENTRY (WCSLEN)
|
|
cmpl $0, (%rdi)
|
|
jz L(exit_tail0)
|
|
cmpl $0, 4(%rdi)
|
|
jz L(exit_tail1)
|
|
cmpl $0, 8(%rdi)
|
|
jz L(exit_tail2)
|
|
cmpl $0, 12(%rdi)
|
|
jz L(exit_tail3)
|
|
cmpl $0, 16(%rdi)
|
|
jz L(exit_tail4)
|
|
cmpl $0, 20(%rdi)
|
|
jz L(exit_tail5)
|
|
cmpl $0, 24(%rdi)
|
|
jz L(exit_tail6)
|
|
cmpl $0, 28(%rdi)
|
|
jz L(exit_tail7)
|
|
|
|
pxor %xmm0, %xmm0
|
|
|
|
lea 32(%rdi), %rax
|
|
addq $16, %rdi
|
|
and $-16, %rax
|
|
|
|
pcmpeqd (%rax), %xmm0
|
|
pmovmskb %xmm0, %edx
|
|
pxor %xmm1, %xmm1
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm1
|
|
pmovmskb %xmm1, %edx
|
|
pxor %xmm2, %xmm2
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm2
|
|
pmovmskb %xmm2, %edx
|
|
pxor %xmm3, %xmm3
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm3
|
|
pmovmskb %xmm3, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm0
|
|
pmovmskb %xmm0, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm1
|
|
pmovmskb %xmm1, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm2
|
|
pmovmskb %xmm2, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm3
|
|
pmovmskb %xmm3, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm0
|
|
pmovmskb %xmm0, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm1
|
|
pmovmskb %xmm1, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm2
|
|
pmovmskb %xmm2, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd (%rax), %xmm3
|
|
pmovmskb %xmm3, %edx
|
|
addq $16, %rax
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
and $-0x40, %rax
|
|
|
|
.p2align 4
|
|
L(aligned_64_loop):
|
|
movaps (%rax), %xmm0
|
|
movaps 16(%rax), %xmm1
|
|
movaps 32(%rax), %xmm2
|
|
movaps 48(%rax), %xmm6
|
|
|
|
pminub %xmm1, %xmm0
|
|
pminub %xmm6, %xmm2
|
|
pminub %xmm0, %xmm2
|
|
pcmpeqd %xmm3, %xmm2
|
|
pmovmskb %xmm2, %edx
|
|
addq $64, %rax
|
|
test %edx, %edx
|
|
jz L(aligned_64_loop)
|
|
|
|
pcmpeqd -64(%rax), %xmm3
|
|
pmovmskb %xmm3, %edx
|
|
addq $48, %rdi
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd %xmm1, %xmm3
|
|
pmovmskb %xmm3, %edx
|
|
addq $-16, %rdi
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd -32(%rax), %xmm3
|
|
pmovmskb %xmm3, %edx
|
|
addq $-16, %rdi
|
|
test %edx, %edx
|
|
jnz L(exit)
|
|
|
|
pcmpeqd %xmm6, %xmm3
|
|
pmovmskb %xmm3, %edx
|
|
addq $-16, %rdi
|
|
test %edx, %edx
|
|
jz L(aligned_64_loop)
|
|
|
|
.p2align 4
|
|
L(exit):
|
|
sub %rdi, %rax
|
|
shr $2, %rax
|
|
test %dl, %dl
|
|
jz L(exit_high)
|
|
|
|
andl $15, %edx
|
|
jz L(exit_1)
|
|
ret
|
|
|
|
/* No align here. Naturally aligned % 16 == 1. */
|
|
L(exit_high):
|
|
andl $(15 << 8), %edx
|
|
jz L(exit_3)
|
|
add $2, %rax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_1):
|
|
add $1, %rax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_3):
|
|
add $3, %rax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_tail0):
|
|
xorl %eax, %eax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_tail1):
|
|
movl $1, %eax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_tail2):
|
|
movl $2, %eax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_tail3):
|
|
movl $3, %eax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_tail4):
|
|
movl $4, %eax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_tail5):
|
|
movl $5, %eax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_tail6):
|
|
movl $6, %eax
|
|
ret
|
|
|
|
.p2align 3
|
|
L(exit_tail7):
|
|
movl $7, %eax
|
|
ret
|
|
|
|
END (WCSLEN)
|
|
|
|
#endif
|