mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-12 22:30:12 +00:00
fb1cf10811
For ia32 PIC, the first thing of many syscalls does is to call __x86.get_pc_thunk.reg to load PC into reg in case there is an error, which is required for setting errno. In most cases, there are no errors. But we still call __x86.get_pc_thunk.reg. This patch adds INLINE_SYSCALL_ERROR_RETURN_VALUE so that i386 can optimize setting errno by branching to the internal __syscall_error without PLT. With i386 INLINE_SYSCALL_ERROR_RETURN_VALUE and i386 syscall inlining optimization for GCC 5, for sysdeps/unix/sysv/linux/fchmodat.c with -O2 -march=i686 -mtune=generic, GCC 5.2 now generates: <fchmodat>: 0: push %ebx 1: mov 0x14(%esp),%eax 5: mov 0x8(%esp),%ebx 9: mov 0xc(%esp),%ecx d: mov 0x10(%esp),%edx 11: test $0xfffffeff,%eax 16: jne 38 <fchmodat+0x38> 18: test $0x1,%ah 1b: jne 48 <fchmodat+0x48> 1d: mov $0x132,%eax 22: call *%gs:0x10 29: cmp $0xfffff000,%eax 2e: ja 58 <fchmodat+0x58> 30: pop %ebx 31: ret 32: lea 0x0(%esi),%esi 38: pop %ebx 39: mov $0xffffffea,%eax 3e: jmp 3f <fchmodat+0x3f> 3f: R_386_PC32 __syscall_error 43: nop 44: lea 0x0(%esi,%eiz,1),%esi 48: pop %ebx 49: mov $0xffffffa1,%eax 4e: jmp 4f <fchmodat+0x4f> 4f: R_386_PC32 __syscall_error 53: nop 54: lea 0x0(%esi,%eiz,1),%esi 58: pop %ebx 59: jmp 5a <fchmodat+0x5a> 5a: R_386_PC32 __syscall_error instead of <fchmodat>: 0: sub $0x8,%esp 3: mov 0x18(%esp),%eax 7: mov %ebx,(%esp) a: call b <fchmodat+0xb> b: R_386_PC32 __x86.get_pc_thunk.bx f: add $0x2,%ebx 11: R_386_GOTPC _GLOBAL_OFFSET_TABLE_ 15: mov %edi,0x4(%esp) 19: test $0xfffffeff,%eax 1e: jne 70 <fchmodat+0x70> 20: test $0x1,%ah 23: jne 88 <fchmodat+0x88> 25: mov 0x14(%esp),%edx 29: mov 0x10(%esp),%ecx 2d: mov 0xc(%esp),%edi 31: xchg %ebx,%edi 33: mov $0x132,%eax 38: call *%gs:0x10 3f: xchg %edi,%ebx 41: cmp $0xfffff000,%eax 46: ja 58 <fchmodat+0x58> 48: mov (%esp),%ebx 4b: mov 0x4(%esp),%edi 4f: add $0x8,%esp 52: ret 53: nop 54: lea 0x0(%esi,%eiz,1),%esi 58: mov 0x0(%ebx),%edx 5a: R_386_TLS_GOTIE __libc_errno 5e: neg %eax 60: mov %eax,%gs:(%edx) 63: mov $0xffffffff,%eax 68: jmp 48 <fchmodat+0x48> 6a: lea 0x0(%esi),%esi 70: mov 0x0(%ebx),%eax 72: R_386_TLS_GOTIE __libc_errno 76: movl $0x16,%gs:(%eax) 7d: mov $0xffffffff,%eax 82: jmp 48 <fchmodat+0x48> 84: lea 0x0(%esi,%eiz,1),%esi 88: mov 0x0(%ebx),%eax 8a: R_386_TLS_GOTIE __libc_errno 8e: movl $0x5f,%gs:(%eax) 95: mov $0xffffffff,%eax 9a: jmp 48 <fchmodat+0x48> * sysdeps/unix/sysv/linux/sysdep.h: New file. * sysdeps/unix/sysv/linux/i386/sysdep.c: Likewise. * sysdeps/unix/sysv/linux/alpha/sysdep.h: Include <sysdeps/unix/sysv/linux/sysdep.h>. * sysdeps/unix/sysv/linux/arm/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/generic/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/hppa/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/ia64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/m68k/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/microblaze/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/mips/mips32/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/mips/mips64/n32/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/mips/mips64/n64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/powerpc/powerpc32/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/powerpc/powerpc64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/s390/s390-32/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/s390/s390-64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/sh/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/sparc/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/x86_64/sysdep.h: Likewise. * sysdeps/unix/sysv/linux/i386/Makefile [$(subdir) == csu] (sysdep-dl-routines): Add sysdep. [$(subdir) == nptl] (libpthread-routines): Likewise. [$(subdir) == rt] (librt-routines): Likewise. * sysdeps/unix/sysv/linux/i386/clone.S (__clone): Don't check PIC when branching to SYSCALL_ERROR_LABEL. * sysdeps/unix/sysv/linux/i386/sysdep.S: Removed. * sysdeps/unix/sysv/linux/i386/sysdep.h: Include <sysdeps/unix/sysv/linux/sysdep.h>. (SYSCALL_ERROR_LABEL): Changed to __syscall_error. (SYSCALL_ERROR_ERRNO): Removed. (SYSCALL_ERROR_HANDLER): Changed to empty. (SYSCALL_ERROR_HANDLER_TLS_STORE): Likewise. (__syscall_error): New prototype. [IS_IN (libc)] (INLINE_SYSCALL): New macro. (INLINE_SYSCALL_ERROR_RETURN_VALUE): Likewise.
301 lines
9.5 KiB
C
301 lines
9.5 KiB
C
/* Copyright (C) 2000-2015 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library. If not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#ifndef _LINUX_MIPS_SYSDEP_H
|
|
#define _LINUX_MIPS_SYSDEP_H 1
|
|
|
|
/* There is some commonality. */
|
|
#include <sysdeps/unix/sysv/linux/sysdep.h>
|
|
#include <sysdeps/unix/mips/mips64/n64/sysdep.h>
|
|
|
|
#include <tls.h>
|
|
|
|
/* In order to get __set_errno() definition in INLINE_SYSCALL. */
|
|
#ifndef __ASSEMBLER__
|
|
#include <errno.h>
|
|
#endif
|
|
|
|
/* For Linux we can use the system call table in the header file
|
|
/usr/include/asm/unistd.h
|
|
of the kernel. But these symbols do not follow the SYS_* syntax
|
|
so we have to redefine the `SYS_ify' macro here. */
|
|
#undef SYS_ify
|
|
#define SYS_ify(syscall_name) __NR_##syscall_name
|
|
|
|
#ifdef __ASSEMBLER__
|
|
|
|
/* We don't want the label for the error handler to be visible in the symbol
|
|
table when we define it here. */
|
|
# define SYSCALL_ERROR_LABEL 99b
|
|
|
|
#else /* ! __ASSEMBLER__ */
|
|
|
|
/* Define a macro which expands into the inline wrapper code for a system
|
|
call. */
|
|
#undef INLINE_SYSCALL
|
|
#define INLINE_SYSCALL(name, nr, args...) \
|
|
({ INTERNAL_SYSCALL_DECL (_sc_err); \
|
|
long result_var = INTERNAL_SYSCALL (name, _sc_err, nr, args); \
|
|
if ( INTERNAL_SYSCALL_ERROR_P (result_var, _sc_err) ) \
|
|
{ \
|
|
__set_errno (INTERNAL_SYSCALL_ERRNO (result_var, _sc_err)); \
|
|
result_var = -1L; \
|
|
} \
|
|
result_var; })
|
|
|
|
#undef INTERNAL_SYSCALL_DECL
|
|
#define INTERNAL_SYSCALL_DECL(err) long err __attribute__ ((unused))
|
|
|
|
#undef INTERNAL_SYSCALL_ERROR_P
|
|
#define INTERNAL_SYSCALL_ERROR_P(val, err) ((void) (val), (long) (err))
|
|
|
|
#undef INTERNAL_SYSCALL_ERRNO
|
|
#define INTERNAL_SYSCALL_ERRNO(val, err) ((void) (err), val)
|
|
|
|
/* Note that the original Linux syscall restart convention required the
|
|
instruction immediately preceding SYSCALL to initialize $v0 with the
|
|
syscall number. Then if a restart triggered, $v0 would have been
|
|
clobbered by the syscall interrupted, and needed to be reinititalized.
|
|
The kernel would decrement the PC by 4 before switching back to the
|
|
user mode so that $v0 had been reloaded before SYSCALL was executed
|
|
again. This implied the place $v0 was loaded from must have been
|
|
preserved across a syscall, e.g. an immediate, static register, stack
|
|
slot, etc.
|
|
|
|
The convention was relaxed in Linux with a change applied to the kernel
|
|
GIT repository as commit 96187fb0bc30cd7919759d371d810e928048249d, that
|
|
first appeared in the 2.6.36 release. Since then the kernel has had
|
|
code that reloads $v0 upon syscall restart and resumes right at the
|
|
SYSCALL instruction, so no special arrangement is needed anymore.
|
|
|
|
For backwards compatibility with existing kernel binaries we support
|
|
the old convention by choosing the instruction preceding SYSCALL
|
|
carefully. This also means we have to force a 32-bit encoding of the
|
|
microMIPS MOVE instruction if one is used. */
|
|
|
|
#ifdef __mips_micromips
|
|
# define MOVE32 "move32"
|
|
#else
|
|
# define MOVE32 "move"
|
|
#endif
|
|
|
|
#undef INTERNAL_SYSCALL
|
|
#define INTERNAL_SYSCALL(name, err, nr, args...) \
|
|
internal_syscall##nr ("li\t%0, %2\t\t\t# " #name "\n\t", \
|
|
"IK" (SYS_ify (name)), \
|
|
0, err, args)
|
|
|
|
#undef INTERNAL_SYSCALL_NCS
|
|
#define INTERNAL_SYSCALL_NCS(number, err, nr, args...) \
|
|
internal_syscall##nr (MOVE32 "\t%0, %2\n\t", \
|
|
"r" (__s0), \
|
|
number, err, args)
|
|
|
|
#define internal_syscall0(v0_init, input, number, err, dummy...) \
|
|
({ \
|
|
long _sys_result; \
|
|
\
|
|
{ \
|
|
register long __s0 asm ("$16") __attribute__ ((unused)) \
|
|
= (number); \
|
|
register long __v0 asm ("$2"); \
|
|
register long __a3 asm ("$7"); \
|
|
__asm__ volatile ( \
|
|
".set\tnoreorder\n\t" \
|
|
v0_init \
|
|
"syscall\n\t" \
|
|
".set reorder" \
|
|
: "=r" (__v0), "=r" (__a3) \
|
|
: input \
|
|
: __SYSCALL_CLOBBERS); \
|
|
err = __a3; \
|
|
_sys_result = __v0; \
|
|
} \
|
|
_sys_result; \
|
|
})
|
|
|
|
#define internal_syscall1(v0_init, input, number, err, arg1) \
|
|
({ \
|
|
long _sys_result; \
|
|
\
|
|
{ \
|
|
register long __s0 asm ("$16") __attribute__ ((unused)) \
|
|
= (number); \
|
|
register long __v0 asm ("$2"); \
|
|
register long __a0 asm ("$4") = (long) (arg1); \
|
|
register long __a3 asm ("$7"); \
|
|
__asm__ volatile ( \
|
|
".set\tnoreorder\n\t" \
|
|
v0_init \
|
|
"syscall\n\t" \
|
|
".set reorder" \
|
|
: "=r" (__v0), "=r" (__a3) \
|
|
: input, "r" (__a0) \
|
|
: __SYSCALL_CLOBBERS); \
|
|
err = __a3; \
|
|
_sys_result = __v0; \
|
|
} \
|
|
_sys_result; \
|
|
})
|
|
|
|
#define internal_syscall2(v0_init, input, number, err, arg1, arg2) \
|
|
({ \
|
|
long _sys_result; \
|
|
\
|
|
{ \
|
|
register long __s0 asm ("$16") __attribute__ ((unused)) \
|
|
= (number); \
|
|
register long __v0 asm ("$2"); \
|
|
register long __a0 asm ("$4") = (long) (arg1); \
|
|
register long __a1 asm ("$5") = (long) (arg2); \
|
|
register long __a3 asm ("$7"); \
|
|
__asm__ volatile ( \
|
|
".set\tnoreorder\n\t" \
|
|
v0_init \
|
|
"syscall\n\t" \
|
|
".set\treorder" \
|
|
: "=r" (__v0), "=r" (__a3) \
|
|
: input, "r" (__a0), "r" (__a1) \
|
|
: __SYSCALL_CLOBBERS); \
|
|
err = __a3; \
|
|
_sys_result = __v0; \
|
|
} \
|
|
_sys_result; \
|
|
})
|
|
|
|
#define internal_syscall3(v0_init, input, number, err, \
|
|
arg1, arg2, arg3) \
|
|
({ \
|
|
long _sys_result; \
|
|
\
|
|
{ \
|
|
register long __s0 asm ("$16") __attribute__ ((unused)) \
|
|
= (number); \
|
|
register long __v0 asm ("$2"); \
|
|
register long __a0 asm ("$4") = (long) (arg1); \
|
|
register long __a1 asm ("$5") = (long) (arg2); \
|
|
register long __a2 asm ("$6") = (long) (arg3); \
|
|
register long __a3 asm ("$7"); \
|
|
__asm__ volatile ( \
|
|
".set\tnoreorder\n\t" \
|
|
v0_init \
|
|
"syscall\n\t" \
|
|
".set\treorder" \
|
|
: "=r" (__v0), "=r" (__a3) \
|
|
: input, "r" (__a0), "r" (__a1), "r" (__a2) \
|
|
: __SYSCALL_CLOBBERS); \
|
|
err = __a3; \
|
|
_sys_result = __v0; \
|
|
} \
|
|
_sys_result; \
|
|
})
|
|
|
|
#define internal_syscall4(v0_init, input, number, err, \
|
|
arg1, arg2, arg3, arg4) \
|
|
({ \
|
|
long _sys_result; \
|
|
\
|
|
{ \
|
|
register long __s0 asm ("$16") __attribute__ ((unused)) \
|
|
= (number); \
|
|
register long __v0 asm ("$2"); \
|
|
register long __a0 asm ("$4") = (long) (arg1); \
|
|
register long __a1 asm ("$5") = (long) (arg2); \
|
|
register long __a2 asm ("$6") = (long) (arg3); \
|
|
register long __a3 asm ("$7") = (long) (arg4); \
|
|
__asm__ volatile ( \
|
|
".set\tnoreorder\n\t" \
|
|
v0_init \
|
|
"syscall\n\t" \
|
|
".set\treorder" \
|
|
: "=r" (__v0), "+r" (__a3) \
|
|
: input, "r" (__a0), "r" (__a1), "r" (__a2) \
|
|
: __SYSCALL_CLOBBERS); \
|
|
err = __a3; \
|
|
_sys_result = __v0; \
|
|
} \
|
|
_sys_result; \
|
|
})
|
|
|
|
#define internal_syscall5(v0_init, input, number, err, \
|
|
arg1, arg2, arg3, arg4, arg5) \
|
|
({ \
|
|
long _sys_result; \
|
|
\
|
|
{ \
|
|
register long __s0 asm ("$16") __attribute__ ((unused)) \
|
|
= (number); \
|
|
register long __v0 asm ("$2"); \
|
|
register long __a0 asm ("$4") = (long) (arg1); \
|
|
register long __a1 asm ("$5") = (long) (arg2); \
|
|
register long __a2 asm ("$6") = (long) (arg3); \
|
|
register long __a3 asm ("$7") = (long) (arg4); \
|
|
register long __a4 asm ("$8") = (long) (arg5); \
|
|
__asm__ volatile ( \
|
|
".set\tnoreorder\n\t" \
|
|
v0_init \
|
|
"syscall\n\t" \
|
|
".set\treorder" \
|
|
: "=r" (__v0), "+r" (__a3) \
|
|
: input, "r" (__a0), "r" (__a1), "r" (__a2), "r" (__a4) \
|
|
: __SYSCALL_CLOBBERS); \
|
|
err = __a3; \
|
|
_sys_result = __v0; \
|
|
} \
|
|
_sys_result; \
|
|
})
|
|
|
|
#define internal_syscall6(v0_init, input, number, err, \
|
|
arg1, arg2, arg3, arg4, arg5, arg6) \
|
|
({ \
|
|
long _sys_result; \
|
|
\
|
|
{ \
|
|
register long __s0 asm ("$16") __attribute__ ((unused)) \
|
|
= (number); \
|
|
register long __v0 asm ("$2"); \
|
|
register long __a0 asm ("$4") = (long) (arg1); \
|
|
register long __a1 asm ("$5") = (long) (arg2); \
|
|
register long __a2 asm ("$6") = (long) (arg3); \
|
|
register long __a3 asm ("$7") = (long) (arg4); \
|
|
register long __a4 asm ("$8") = (long) (arg5); \
|
|
register long __a5 asm ("$9") = (long) (arg6); \
|
|
__asm__ volatile ( \
|
|
".set\tnoreorder\n\t" \
|
|
v0_init \
|
|
"syscall\n\t" \
|
|
".set\treorder" \
|
|
: "=r" (__v0), "+r" (__a3) \
|
|
: input, "r" (__a0), "r" (__a1), "r" (__a2), "r" (__a4), \
|
|
"r" (__a5) \
|
|
: __SYSCALL_CLOBBERS); \
|
|
err = __a3; \
|
|
_sys_result = __v0; \
|
|
} \
|
|
_sys_result; \
|
|
})
|
|
|
|
#define __SYSCALL_CLOBBERS "$1", "$3", "$10", "$11", "$12", "$13", \
|
|
"$14", "$15", "$24", "$25", "hi", "lo", "memory"
|
|
#endif /* __ASSEMBLER__ */
|
|
|
|
/* Pointer mangling is not yet supported for MIPS. */
|
|
#define PTR_MANGLE(var) (void) (var)
|
|
#define PTR_DEMANGLE(var) (void) (var)
|
|
|
|
#endif /* linux/mips/sysdep.h */
|