mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-23 19:30:10 +00:00
e41b395523
On AMD processors, memcpy optimized with unaligned SSE load is slower than emcpy optimized with aligned SSSE3 while other string functions are faster with unaligned SSE load. A feature bit, Fast_Unaligned_Copy, is added to select memcpy optimized with unaligned SSE load. [BZ #19583] * sysdeps/x86/cpu-features.c (init_cpu_features): Set Fast_Unaligned_Copy with Fast_Unaligned_Load for Intel processors. Set Fast_Copy_Backward for AMD Excavator processors. * sysdeps/x86/cpu-features.h (bit_arch_Fast_Unaligned_Copy): New. (index_arch_Fast_Unaligned_Copy): Likewise. * sysdeps/x86_64/multiarch/memcpy.S (__new_memcpy): Check Fast_Unaligned_Copy instead of Fast_Unaligned_Load. |
||
---|---|---|
.. | ||
bits | ||
fpu | ||
cpu-features-offsets.sym | ||
cpu-features.c | ||
cpu-features.h | ||
dl-get-cpu-features.c | ||
elide.h | ||
fpu_control.h | ||
init-arch.h | ||
libc-start.c | ||
linkmap.h | ||
Makefile | ||
rtld-global-offsets.sym | ||
string_private.h | ||
tininess.h | ||
tst-get-cpu-features-static.c | ||
tst-get-cpu-features.c | ||
Versions |