mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-04 10:50:07 +00:00
fb1cf10811
For ia32 PIC, the first thing of many syscalls does is to call __x86.get_pc_thunk.reg to load PC into reg in case there is an error, which is required for setting errno. In most cases, there are no errors. But we still call __x86.get_pc_thunk.reg. This patch adds INLINE_SYSCALL_ERROR_RETURN_VALUE so that i386 can optimize setting errno by branching to the internal __syscall_error without PLT. With i386 INLINE_SYSCALL_ERROR_RETURN_VALUE and i386 syscall inlining optimization for GCC 5, for sysdeps/unix/sysv/linux/fchmodat.c with -O2 -march=i686 -mtune=generic, GCC 5.2 now generates: <fchmodat>: 0: push %ebx 1: mov 0x14(%esp),%eax 5: mov 0x8(%esp),%ebx 9: mov 0xc(%esp),%ecx d: mov 0x10(%esp),%edx 11: test $0xfffffeff,%eax 16: jne 38 <fchmodat+0x38> 18: test $0x1,%ah 1b: jne 48 <fchmodat+0x48> 1d: mov $0x132,%eax 22: call *%gs:0x10 29: cmp $0xfffff000,%eax 2e: ja 58 <fchmodat+0x58> 30: pop %ebx 31: ret 32: lea 0x0(%esi),%esi 38: pop %ebx 39: mov $0xffffffea,%eax 3e: jmp 3f <fchmodat+0x3f> 3f: R_386_PC32 __syscall_error 43: nop 44: lea 0x0(%esi,%eiz,1),%esi 48: pop %ebx 49: mov $0xffffffa1,%eax 4e: jmp 4f <fchmodat+0x4f> 4f: R_386_PC32 __syscall_error 53: nop 54: lea 0x0(%esi,%eiz,1),%esi 58: pop %ebx 59: jmp 5a <fchmodat+0x5a> 5a: R_386_PC32 __syscall_error instead of <fchmodat>: 0: sub $0x8,%esp 3: mov 0x18(%esp),%eax 7: mov %ebx,(%esp) a: call b <fchmodat+0xb> b: R_386_PC32 __x86.get_pc_thunk.bx f: add $0x2,%ebx 11: R_386_GOTPC _GLOBAL_OFFSET_TABLE_ 15: mov %edi,0x4(%esp) 19: test $0xfffffeff,%eax 1e: jne 70 <fchmodat+0x70> 20: test $0x1,%ah 23: jne 88 <fchmodat+0x88> 25: mov 0x14(%esp),%edx 29: mov 0x10(%esp),%ecx 2d: mov 0xc(%esp),%edi 31: xchg %ebx,%edi 33: mov $0x132,%eax 38: call *%gs:0x10 3f: xchg %edi,%ebx 41: cmp $0xfffff000,%eax 46: ja 58 <fchmodat+0x58> 48: mov (%esp),%ebx 4b: mov 0x4(%esp),%edi 4f: add $0x8,%esp 52: ret 53: nop 54: lea 0x0(%esi,%eiz,1),%esi 58: mov 0x0(%ebx),%edx 5a: R_386_TLS_GOTIE __libc_errno 5e: neg %eax 60: mov %eax,%gs:(%edx) 63: mov $0xffffffff,%eax 68: jmp 48 <fchmodat+0x48> 6a: lea 0x0(%esi),%esi 70: mov 0x0(%ebx),%eax 72: R_386_TLS_GOTIE __libc_errno 76: movl $0x16,%gs:(%eax) 7d: mov $0xffffffff,%eax 82: jmp 48 <fchmodat+0x48> 84: lea 0x0(%esi,%eiz,1),%esi 88: mov 0x0(%ebx),%eax 8a: R_386_TLS_GOTIE __libc_errno 8e: movl $0x5f,%gs:(%eax) 95: mov $0xffffffff,%eax 9a: jmp 48 <fchmodat+0x48> * sysdeps/unix/sysv/linux/sysdep.h: New file. * sysdeps/unix/sysv/linux/i386/sysdep.c: Likewise. * sysdeps/unix/sysv/linux/alpha/sysdep.h: Include <sysdeps/unix/sysv/linux/sysdep.h>. * sysdeps/unix/sysv/linux/arm/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/generic/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/hppa/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/ia64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/m68k/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/microblaze/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/mips/mips32/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/mips/mips64/n32/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/mips/mips64/n64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/powerpc/powerpc32/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/powerpc/powerpc64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/s390/s390-32/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/s390/s390-64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/sh/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/sparc/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/x86_64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/i386/Makefile [$(subdir) == csu] (sysdep-dl-routines): Add sysdep. [$(subdir) == nptl] (libpthread-routines): Likewise. [$(subdir) == rt] (librt-routines): Likewise. * sysdeps/unix/sysv/linux/i386/clone.S (__clone): Don't check PIC when branching to SYSCALL_ERROR_LABEL. * sysdeps/unix/sysv/linux/i386/sysdep.S: Removed. * sysdeps/unix/sysv/linux/i386/sysdep.h: Include <sysdeps/unix/sysv/linux/sysdep.h>. (SYSCALL_ERROR_LABEL): Changed to __syscall_error. (SYSCALL_ERROR_ERRNO): Removed. (SYSCALL_ERROR_HANDLER): Changed to empty. (SYSCALL_ERROR_HANDLER_TLS_STORE): Likewise. (__syscall_error): New prototype. [IS_IN (libc)] (INLINE_SYSCALL): New macro. (INLINE_SYSCALL_ERROR_RETURN_VALUE): Likewise.
258 lines
9.3 KiB
C
258 lines
9.3 KiB
C
/* Copyright (C) 1992-2015 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
/* Alan Modra <amodra@bigpond.net.au> rewrote the INLINE_SYSCALL macro */
|
|
|
|
#ifndef _LINUX_POWERPC_SYSDEP_H
|
|
#define _LINUX_POWERPC_SYSDEP_H 1
|
|
|
|
#include <sysdeps/unix/sysv/linux/sysdep.h>
|
|
#include <sysdeps/unix/powerpc/sysdep.h>
|
|
#include <tls.h>
|
|
|
|
/* Define __set_errno() for INLINE_SYSCALL macro below. */
|
|
#ifndef __ASSEMBLER__
|
|
#include <errno.h>
|
|
#endif
|
|
|
|
/* Some systen calls got renamed over time, but retained the same semantics.
|
|
Handle them here so they can be catched by both C and assembler stubs in
|
|
glibc. */
|
|
|
|
#ifdef __NR_pread64
|
|
# ifdef __NR_pread
|
|
# error "__NR_pread and __NR_pread64 both defined???"
|
|
# endif
|
|
# define __NR_pread __NR_pread64
|
|
#endif
|
|
|
|
#ifdef __NR_pwrite64
|
|
# ifdef __NR_pwrite
|
|
# error "__NR_pwrite and __NR_pwrite64 both defined???"
|
|
# endif
|
|
# define __NR_pwrite __NR_pwrite64
|
|
#endif
|
|
|
|
/* For Linux we can use the system call table in the header file
|
|
/usr/include/asm/unistd.h
|
|
of the kernel. But these symbols do not follow the SYS_* syntax
|
|
so we have to redefine the `SYS_ify' macro here. */
|
|
#undef SYS_ify
|
|
#define SYS_ify(syscall_name) __NR_##syscall_name
|
|
|
|
#ifdef __ASSEMBLER__
|
|
|
|
/* This seems to always be the case on PPC. */
|
|
# define ALIGNARG(log2) log2
|
|
# define ASM_SIZE_DIRECTIVE(name) .size name,.-name
|
|
|
|
#endif /* __ASSEMBLER__ */
|
|
|
|
/* This version is for internal uses when there is no desire
|
|
to set errno */
|
|
#define INTERNAL_VSYSCALL_NO_SYSCALL_FALLBACK(name, err, type, nr, args...) \
|
|
({ \
|
|
type sc_ret = ENOSYS; \
|
|
\
|
|
__typeof (__vdso_##name) vdsop = __vdso_##name; \
|
|
PTR_DEMANGLE (vdsop); \
|
|
if (vdsop != NULL) \
|
|
sc_ret = \
|
|
INTERNAL_VSYSCALL_CALL_TYPE (vdsop, err, type, nr, ##args); \
|
|
else \
|
|
err = 1 << 28; \
|
|
sc_ret; \
|
|
})
|
|
|
|
/* List of system calls which are supported as vsyscalls. */
|
|
#define HAVE_CLOCK_GETRES_VSYSCALL 1
|
|
#define HAVE_CLOCK_GETTIME_VSYSCALL 1
|
|
#define HAVE_GETCPU_VSYSCALL 1
|
|
|
|
/* Define a macro which expands inline into the wrapper code for a system
|
|
call. This use is for internal calls that do not need to handle errors
|
|
normally. It will never touch errno. This returns just what the kernel
|
|
gave back in the non-error (CR0.SO cleared) case, otherwise (CR0.SO set)
|
|
the negation of the return value in the kernel gets reverted. */
|
|
|
|
#define INTERNAL_VSYSCALL_CALL_TYPE(funcptr, err, type, nr, args...) \
|
|
({ \
|
|
register void *r0 __asm__ ("r0"); \
|
|
register long int r3 __asm__ ("r3"); \
|
|
register long int r4 __asm__ ("r4"); \
|
|
register long int r5 __asm__ ("r5"); \
|
|
register long int r6 __asm__ ("r6"); \
|
|
register long int r7 __asm__ ("r7"); \
|
|
register long int r8 __asm__ ("r8"); \
|
|
register type rval __asm__ ("r3"); \
|
|
LOADARGS_##nr (funcptr, args); \
|
|
__asm__ __volatile__ \
|
|
("mtctr %0\n\t" \
|
|
"bctrl\n\t" \
|
|
"mfcr %0\n\t" \
|
|
"0:" \
|
|
: "+r" (r0), "+r" (r3), "+r" (r4), "+r" (r5), "+r" (r6), \
|
|
"+r" (r7), "+r" (r8) \
|
|
: : "r9", "r10", "r11", "r12", "cr0", "ctr", "lr", "memory"); \
|
|
err = (long int) r0; \
|
|
__asm__ __volatile__ ("" : "=r" (rval) : "r" (r3)); \
|
|
rval; \
|
|
})
|
|
|
|
#define INTERNAL_VSYSCALL_CALL(funcptr, err, nr, args...) \
|
|
INTERNAL_VSYSCALL_CALL_TYPE(funcptr, err, long int, nr, args)
|
|
|
|
#undef INLINE_SYSCALL
|
|
|
|
/* This version is for kernels that implement system calls that
|
|
behave like function calls as far as register saving. */
|
|
#define INLINE_SYSCALL(name, nr, args...) \
|
|
({ \
|
|
INTERNAL_SYSCALL_DECL (sc_err); \
|
|
long int sc_ret = INTERNAL_SYSCALL (name, sc_err, nr, args); \
|
|
if (INTERNAL_SYSCALL_ERROR_P (sc_ret, sc_err)) \
|
|
{ \
|
|
__set_errno (INTERNAL_SYSCALL_ERRNO (sc_ret, sc_err)); \
|
|
sc_ret = -1L; \
|
|
} \
|
|
sc_ret; \
|
|
})
|
|
|
|
/* Define a macro which expands inline into the wrapper code for a system
|
|
call. This use is for internal calls that do not need to handle errors
|
|
normally. It will never touch errno. This returns just what the kernel
|
|
gave back in the non-error (CR0.SO cleared) case, otherwise (CR0.SO set)
|
|
the negation of the return value in the kernel gets reverted. */
|
|
|
|
#undef INTERNAL_SYSCALL
|
|
#define INTERNAL_SYSCALL_NCS(name, err, nr, args...) \
|
|
({ \
|
|
register long int r0 __asm__ ("r0"); \
|
|
register long int r3 __asm__ ("r3"); \
|
|
register long int r4 __asm__ ("r4"); \
|
|
register long int r5 __asm__ ("r5"); \
|
|
register long int r6 __asm__ ("r6"); \
|
|
register long int r7 __asm__ ("r7"); \
|
|
register long int r8 __asm__ ("r8"); \
|
|
LOADARGS_##nr (name, ##args); \
|
|
ABORT_TRANSACTION; \
|
|
__asm__ __volatile__ \
|
|
("sc\n\t" \
|
|
"mfcr %0\n\t" \
|
|
"0:" \
|
|
: "=&r" (r0), \
|
|
"=&r" (r3), "=&r" (r4), "=&r" (r5), \
|
|
"=&r" (r6), "=&r" (r7), "=&r" (r8) \
|
|
: ASM_INPUT_##nr \
|
|
: "r9", "r10", "r11", "r12", \
|
|
"cr0", "ctr", "memory"); \
|
|
err = r0; \
|
|
r3; \
|
|
})
|
|
#define INTERNAL_SYSCALL(name, err, nr, args...) \
|
|
INTERNAL_SYSCALL_NCS (__NR_##name, err, nr, args)
|
|
|
|
#undef INTERNAL_SYSCALL_DECL
|
|
#define INTERNAL_SYSCALL_DECL(err) long int err __attribute__ ((unused))
|
|
|
|
#undef INTERNAL_SYSCALL_ERROR_P
|
|
#define INTERNAL_SYSCALL_ERROR_P(val, err) \
|
|
((void) (val), __builtin_expect ((err) & (1 << 28), 0))
|
|
|
|
#undef INTERNAL_SYSCALL_ERRNO
|
|
#define INTERNAL_SYSCALL_ERRNO(val, err) (val)
|
|
|
|
#define LOADARGS_0(name, dummy) \
|
|
r0 = name
|
|
#define LOADARGS_1(name, __arg1) \
|
|
long int arg1 = (long int) (__arg1); \
|
|
LOADARGS_0(name, 0); \
|
|
extern void __illegally_sized_syscall_arg1 (void); \
|
|
if (__builtin_classify_type (__arg1) != 5 && sizeof (__arg1) > 8) \
|
|
__illegally_sized_syscall_arg1 (); \
|
|
r3 = arg1
|
|
#define LOADARGS_2(name, __arg1, __arg2) \
|
|
long int arg2 = (long int) (__arg2); \
|
|
LOADARGS_1(name, __arg1); \
|
|
extern void __illegally_sized_syscall_arg2 (void); \
|
|
if (__builtin_classify_type (__arg2) != 5 && sizeof (__arg2) > 8) \
|
|
__illegally_sized_syscall_arg2 (); \
|
|
r4 = arg2
|
|
#define LOADARGS_3(name, __arg1, __arg2, __arg3) \
|
|
long int arg3 = (long int) (__arg3); \
|
|
LOADARGS_2(name, __arg1, __arg2); \
|
|
extern void __illegally_sized_syscall_arg3 (void); \
|
|
if (__builtin_classify_type (__arg3) != 5 && sizeof (__arg3) > 8) \
|
|
__illegally_sized_syscall_arg3 (); \
|
|
r5 = arg3
|
|
#define LOADARGS_4(name, __arg1, __arg2, __arg3, __arg4) \
|
|
long int arg4 = (long int) (__arg4); \
|
|
LOADARGS_3(name, __arg1, __arg2, __arg3); \
|
|
extern void __illegally_sized_syscall_arg4 (void); \
|
|
if (__builtin_classify_type (__arg4) != 5 && sizeof (__arg4) > 8) \
|
|
__illegally_sized_syscall_arg4 (); \
|
|
r6 = arg4
|
|
#define LOADARGS_5(name, __arg1, __arg2, __arg3, __arg4, __arg5) \
|
|
long int arg5 = (long int) (__arg5); \
|
|
LOADARGS_4(name, __arg1, __arg2, __arg3, __arg4); \
|
|
extern void __illegally_sized_syscall_arg5 (void); \
|
|
if (__builtin_classify_type (__arg5) != 5 && sizeof (__arg5) > 8) \
|
|
__illegally_sized_syscall_arg5 (); \
|
|
r7 = arg5
|
|
#define LOADARGS_6(name, __arg1, __arg2, __arg3, __arg4, __arg5, __arg6) \
|
|
long int arg6 = (long int) (__arg6); \
|
|
LOADARGS_5(name, __arg1, __arg2, __arg3, __arg4, __arg5); \
|
|
extern void __illegally_sized_syscall_arg6 (void); \
|
|
if (__builtin_classify_type (__arg6) != 5 && sizeof (__arg6) > 8) \
|
|
__illegally_sized_syscall_arg6 (); \
|
|
r8 = arg6
|
|
|
|
#define ASM_INPUT_0 "0" (r0)
|
|
#define ASM_INPUT_1 ASM_INPUT_0, "1" (r3)
|
|
#define ASM_INPUT_2 ASM_INPUT_1, "2" (r4)
|
|
#define ASM_INPUT_3 ASM_INPUT_2, "3" (r5)
|
|
#define ASM_INPUT_4 ASM_INPUT_3, "4" (r6)
|
|
#define ASM_INPUT_5 ASM_INPUT_4, "5" (r7)
|
|
#define ASM_INPUT_6 ASM_INPUT_5, "6" (r8)
|
|
|
|
|
|
/* Pointer mangling support. */
|
|
#if IS_IN (rtld)
|
|
/* We cannot use the thread descriptor because in ld.so we use setjmp
|
|
earlier than the descriptor is initialized. */
|
|
#else
|
|
# ifdef __ASSEMBLER__
|
|
# define PTR_MANGLE(reg, tmpreg) \
|
|
ld tmpreg,POINTER_GUARD(r13); \
|
|
xor reg,tmpreg,reg
|
|
# define PTR_MANGLE2(reg, tmpreg) \
|
|
xor reg,tmpreg,reg
|
|
# define PTR_MANGLE3(destreg, reg, tmpreg) \
|
|
ld tmpreg,POINTER_GUARD(r13); \
|
|
xor destreg,tmpreg,reg
|
|
# define PTR_DEMANGLE(reg, tmpreg) PTR_MANGLE (reg, tmpreg)
|
|
# define PTR_DEMANGLE2(reg, tmpreg) PTR_MANGLE2 (reg, tmpreg)
|
|
# define PTR_DEMANGLE3(destreg, reg, tmpreg) PTR_MANGLE3 (destreg, reg, tmpreg)
|
|
# else
|
|
# define PTR_MANGLE(var) \
|
|
(var) = (__typeof (var)) ((uintptr_t) (var) ^ THREAD_GET_POINTER_GUARD ())
|
|
# define PTR_DEMANGLE(var) PTR_MANGLE (var)
|
|
# endif
|
|
#endif
|
|
|
|
#endif /* linux/powerpc/powerpc64/sysdep.h */
|