mirror of
https://sourceware.org/git/glibc.git
synced 2024-11-22 13:00:06 +00:00
13efa86ece
Although the Enhanced REP MOVSB/STOSB (ERMS) implementations of memmove, memcpy, mempcpy and memset aren't used by the current processors, this patch adds Prefer_ERMS check in memmove, memcpy, mempcpy and memset so that they can be used in the future. * sysdeps/x86/cpu-features.h (bit_arch_Prefer_ERMS): New. (index_arch_Prefer_ERMS): Likewise. * sysdeps/x86_64/multiarch/memcpy.S (__new_memcpy): Return __memcpy_erms for Prefer_ERMS. * sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S (__memmove_erms): Enabled for libc.a. * ysdeps/x86_64/multiarch/memmove.S (__libc_memmove): Return __memmove_erms or Prefer_ERMS. * sysdeps/x86_64/multiarch/mempcpy.S (__mempcpy): Return __mempcpy_erms for Prefer_ERMS. * sysdeps/x86_64/multiarch/memset.S (memset): Return __memset_erms for Prefer_ERMS.
76 lines
2.3 KiB
ArmAsm
76 lines
2.3 KiB
ArmAsm
/* Multiple versions of memset
|
|
All versions must be listed in ifunc-impl-list.c.
|
|
Copyright (C) 2014-2016 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include <sysdep.h>
|
|
#include <shlib-compat.h>
|
|
#include <init-arch.h>
|
|
|
|
/* Define multiple versions only for the definition in lib. */
|
|
#if IS_IN (libc)
|
|
ENTRY(memset)
|
|
.type memset, @gnu_indirect_function
|
|
LOAD_RTLD_GLOBAL_RO_RDX
|
|
lea __memset_erms(%rip), %RAX_LP
|
|
HAS_ARCH_FEATURE (Prefer_ERMS)
|
|
jnz 2f
|
|
lea __memset_sse2_unaligned_erms(%rip), %RAX_LP
|
|
HAS_CPU_FEATURE (ERMS)
|
|
jnz 1f
|
|
lea __memset_sse2_unaligned(%rip), %RAX_LP
|
|
1:
|
|
HAS_ARCH_FEATURE (AVX2_Usable)
|
|
jz 2f
|
|
lea __memset_avx2_unaligned_erms(%rip), %RAX_LP
|
|
HAS_CPU_FEATURE (ERMS)
|
|
jnz L(AVX512F)
|
|
lea __memset_avx2_unaligned(%rip), %RAX_LP
|
|
L(AVX512F):
|
|
# ifdef HAVE_AVX512_ASM_SUPPORT
|
|
HAS_ARCH_FEATURE (AVX512F_Usable)
|
|
jz 2f
|
|
lea __memset_avx512_no_vzeroupper(%rip), %RAX_LP
|
|
HAS_ARCH_FEATURE (Prefer_No_VZEROUPPER)
|
|
jnz 2f
|
|
lea __memset_avx512_unaligned_erms(%rip), %RAX_LP
|
|
HAS_CPU_FEATURE (ERMS)
|
|
jnz 2f
|
|
lea __memset_avx512_unaligned(%rip), %RAX_LP
|
|
# endif
|
|
2: ret
|
|
END(memset)
|
|
#endif
|
|
|
|
#if IS_IN (libc)
|
|
# define MEMSET_SYMBOL(p,s) p##_sse2_##s
|
|
|
|
# ifdef SHARED
|
|
# undef libc_hidden_builtin_def
|
|
/* It doesn't make sense to send libc-internal memset calls through a PLT.
|
|
The speedup we get from using SSE2 instructions is likely eaten away
|
|
by the indirect call in the PLT. */
|
|
# define libc_hidden_builtin_def(name) \
|
|
.globl __GI_memset; __GI_memset = __memset_sse2_unaligned
|
|
# endif
|
|
|
|
# undef strong_alias
|
|
# define strong_alias(original, alias)
|
|
#endif
|
|
|
|
#include "../memset.S"
|