mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-02 09:40:13 +00:00
82fd7314c7
The powerpc optimization to provide a fast stacktrace requires some ad-hoc code to handle Linux signal frames and the change is fragile once the kernel decides to slight change its execution sequence [1]. The generic implementation work as-is and it should be future proof since the kernel provides the expected CFI directives in vDSO shared page. Checked on powerpc-linux-gnu, powerpc64le-linux-gnu, and powerpc64-linux-gnu. [1] https://sourceware.org/pipermail/libc-alpha/2021-January/122027.html
259 lines
9.1 KiB
C
259 lines
9.1 KiB
C
/* Syscall definitions, Linux PowerPC generic version.
|
|
Copyright (C) 2019-2021 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#ifndef _LINUX_POWERPC_SYSDEP_H
|
|
#define _LINUX_POWERPC_SYSDEP_H 1
|
|
|
|
#include <sysdeps/unix/sysv/linux/sysdep.h>
|
|
#include <sysdeps/unix/powerpc/sysdep.h>
|
|
#include <tls.h>
|
|
/* Define __set_errno() for INLINE_SYSCALL macro below. */
|
|
#include <errno.h>
|
|
|
|
/* For Linux we can use the system call table in the header file
|
|
/usr/include/asm/unistd.h
|
|
of the kernel. But these symbols do not follow the SYS_* syntax
|
|
so we have to redefine the `SYS_ify' macro here. */
|
|
#undef SYS_ify
|
|
#define SYS_ify(syscall_name) __NR_##syscall_name
|
|
|
|
/* Define a macro which expands inline into the wrapper code for a system
|
|
call. This use is for internal calls that do not need to handle errors
|
|
normally. It will never touch errno. This returns just what the kernel
|
|
gave back in the non-error (CR0.SO cleared) case, otherwise (CR0.SO set)
|
|
the negation of the return value in the kernel gets reverted. */
|
|
|
|
#define INTERNAL_VSYSCALL_CALL_TYPE(funcptr, type, nr, args...) \
|
|
({ \
|
|
register void *r0 __asm__ ("r0"); \
|
|
register long int r3 __asm__ ("r3"); \
|
|
register long int r4 __asm__ ("r4"); \
|
|
register long int r5 __asm__ ("r5"); \
|
|
register long int r6 __asm__ ("r6"); \
|
|
register long int r7 __asm__ ("r7"); \
|
|
register long int r8 __asm__ ("r8"); \
|
|
register type rval __asm__ ("r3"); \
|
|
LOADARGS_##nr (funcptr, args); \
|
|
__asm__ __volatile__ \
|
|
("mtctr %0\n\t" \
|
|
"bctrl\n\t" \
|
|
"mfcr %0\n\t" \
|
|
"0:" \
|
|
: "+r" (r0), "+r" (r3), "+r" (r4), "+r" (r5), "+r" (r6), \
|
|
"+r" (r7), "+r" (r8) \
|
|
: : "r9", "r10", "r11", "r12", \
|
|
"cr0", "cr1", "cr5", "cr6", "cr7", \
|
|
"xer", "lr", "ctr", "memory"); \
|
|
__asm__ __volatile__ ("" : "=r" (rval) : "r" (r3)); \
|
|
(long int) r0 & (1 << 28) ? -rval : rval; \
|
|
})
|
|
|
|
#define INTERNAL_VSYSCALL_CALL(funcptr, nr, args...) \
|
|
INTERNAL_VSYSCALL_CALL_TYPE(funcptr, long int, nr, args)
|
|
|
|
#define DECLARE_REGS \
|
|
register long int r0 __asm__ ("r0"); \
|
|
register long int r3 __asm__ ("r3"); \
|
|
register long int r4 __asm__ ("r4"); \
|
|
register long int r5 __asm__ ("r5"); \
|
|
register long int r6 __asm__ ("r6"); \
|
|
register long int r7 __asm__ ("r7"); \
|
|
register long int r8 __asm__ ("r8");
|
|
|
|
#define SYSCALL_SCV(nr) \
|
|
({ \
|
|
__asm__ __volatile__ \
|
|
(".machine \"push\"\n\t" \
|
|
".machine \"power9\"\n\t" \
|
|
"scv 0\n\t" \
|
|
".machine \"pop\"\n\t" \
|
|
"0:" \
|
|
: "=&r" (r0), \
|
|
"=&r" (r3), "=&r" (r4), "=&r" (r5), \
|
|
"=&r" (r6), "=&r" (r7), "=&r" (r8) \
|
|
: ASM_INPUT_##nr \
|
|
: "r9", "r10", "r11", "r12", \
|
|
"cr0", "cr1", "cr5", "cr6", "cr7", \
|
|
"xer", "lr", "ctr", "memory"); \
|
|
r3; \
|
|
})
|
|
|
|
#define SYSCALL_SC(nr) \
|
|
({ \
|
|
__asm__ __volatile__ \
|
|
("sc\n\t" \
|
|
"mfcr %0\n\t" \
|
|
"0:" \
|
|
: "=&r" (r0), \
|
|
"=&r" (r3), "=&r" (r4), "=&r" (r5), \
|
|
"=&r" (r6), "=&r" (r7), "=&r" (r8) \
|
|
: ASM_INPUT_##nr \
|
|
: "r9", "r10", "r11", "r12", \
|
|
"xer", "cr0", "ctr", "memory"); \
|
|
r0 & (1 << 28) ? -r3 : r3; \
|
|
})
|
|
|
|
/* This will only be non-empty for 64-bit systems, see below. */
|
|
#define TRY_SYSCALL_SCV(nr)
|
|
|
|
#if defined(__PPC64__) || defined(__powerpc64__)
|
|
# define SYSCALL_ARG_SIZE 8
|
|
|
|
/* For the static case, unlike the dynamic loader, there is no compile-time way
|
|
to check if we are inside startup code. So we need to check if the thread
|
|
pointer has already been setup before trying to access the TLS. */
|
|
# ifndef SHARED
|
|
# define CHECK_THREAD_POINTER (__thread_register != 0)
|
|
# else
|
|
# define CHECK_THREAD_POINTER (1)
|
|
# endif
|
|
|
|
/* When inside the dynamic loader, the thread pointer may not have been
|
|
initialized yet, so don't check for scv support in that case. */
|
|
# if defined(USE_PPC_SCV) && !IS_IN(rtld)
|
|
# undef TRY_SYSCALL_SCV
|
|
# define TRY_SYSCALL_SCV(nr) \
|
|
CHECK_THREAD_POINTER && THREAD_GET_HWCAP() & PPC_FEATURE2_SCV ? \
|
|
SYSCALL_SCV(nr) :
|
|
# endif
|
|
|
|
#else
|
|
# define SYSCALL_ARG_SIZE 4
|
|
#endif
|
|
|
|
# define INTERNAL_SYSCALL_NCS(name, nr, args...) \
|
|
({ \
|
|
DECLARE_REGS; \
|
|
LOADARGS_##nr (name, ##args); \
|
|
TRY_SYSCALL_SCV(nr) \
|
|
SYSCALL_SC(nr); \
|
|
})
|
|
|
|
#undef INTERNAL_SYSCALL
|
|
#define INTERNAL_SYSCALL(name, nr, args...) \
|
|
INTERNAL_SYSCALL_NCS (__NR_##name, nr, args)
|
|
|
|
#define LOADARGS_0(name, dummy) \
|
|
r0 = name
|
|
#define LOADARGS_1(name, __arg1) \
|
|
long int _arg1 = (long int) (__arg1); \
|
|
LOADARGS_0(name, 0); \
|
|
extern void __illegally_sized_syscall_arg1 (void); \
|
|
if (__builtin_classify_type (__arg1) != 5 \
|
|
&& sizeof (__arg1) > SYSCALL_ARG_SIZE) \
|
|
__illegally_sized_syscall_arg1 (); \
|
|
r3 = _arg1
|
|
#define LOADARGS_2(name, __arg1, __arg2) \
|
|
long int _arg2 = (long int) (__arg2); \
|
|
LOADARGS_1(name, __arg1); \
|
|
extern void __illegally_sized_syscall_arg2 (void); \
|
|
if (__builtin_classify_type (__arg2) != 5 \
|
|
&& sizeof (__arg2) > SYSCALL_ARG_SIZE) \
|
|
__illegally_sized_syscall_arg2 (); \
|
|
r4 = _arg2
|
|
#define LOADARGS_3(name, __arg1, __arg2, __arg3) \
|
|
long int _arg3 = (long int) (__arg3); \
|
|
LOADARGS_2(name, __arg1, __arg2); \
|
|
extern void __illegally_sized_syscall_arg3 (void); \
|
|
if (__builtin_classify_type (__arg3) != 5 \
|
|
&& sizeof (__arg3) > SYSCALL_ARG_SIZE) \
|
|
__illegally_sized_syscall_arg3 (); \
|
|
r5 = _arg3
|
|
#define LOADARGS_4(name, __arg1, __arg2, __arg3, __arg4) \
|
|
long int _arg4 = (long int) (__arg4); \
|
|
LOADARGS_3(name, __arg1, __arg2, __arg3); \
|
|
extern void __illegally_sized_syscall_arg4 (void); \
|
|
if (__builtin_classify_type (__arg4) != 5 \
|
|
&& sizeof (__arg4) > SYSCALL_ARG_SIZE) \
|
|
__illegally_sized_syscall_arg4 (); \
|
|
r6 = _arg4
|
|
#define LOADARGS_5(name, __arg1, __arg2, __arg3, __arg4, __arg5) \
|
|
long int _arg5 = (long int) (__arg5); \
|
|
LOADARGS_4(name, __arg1, __arg2, __arg3, __arg4); \
|
|
extern void __illegally_sized_syscall_arg5 (void); \
|
|
if (__builtin_classify_type (__arg5) != 5 \
|
|
&& sizeof (__arg5) > SYSCALL_ARG_SIZE) \
|
|
__illegally_sized_syscall_arg5 (); \
|
|
r7 = _arg5
|
|
#define LOADARGS_6(name, __arg1, __arg2, __arg3, __arg4, __arg5, __arg6) \
|
|
long int _arg6 = (long int) (__arg6); \
|
|
LOADARGS_5(name, __arg1, __arg2, __arg3, __arg4, __arg5); \
|
|
extern void __illegally_sized_syscall_arg6 (void); \
|
|
if (__builtin_classify_type (__arg6) != 5 \
|
|
&& sizeof (__arg6) > SYSCALL_ARG_SIZE) \
|
|
__illegally_sized_syscall_arg6 (); \
|
|
r8 = _arg6
|
|
|
|
#define ASM_INPUT_0 "0" (r0)
|
|
#define ASM_INPUT_1 ASM_INPUT_0, "1" (r3)
|
|
#define ASM_INPUT_2 ASM_INPUT_1, "2" (r4)
|
|
#define ASM_INPUT_3 ASM_INPUT_2, "3" (r5)
|
|
#define ASM_INPUT_4 ASM_INPUT_3, "4" (r6)
|
|
#define ASM_INPUT_5 ASM_INPUT_4, "5" (r7)
|
|
#define ASM_INPUT_6 ASM_INPUT_5, "6" (r8)
|
|
|
|
|
|
/* Pointer mangling support. */
|
|
#if IS_IN (rtld)
|
|
/* We cannot use the thread descriptor because in ld.so we use setjmp
|
|
earlier than the descriptor is initialized. */
|
|
#else
|
|
# ifdef __ASSEMBLER__
|
|
# if defined(__PPC64__) || defined(__powerpc64__)
|
|
# define LOAD ld
|
|
# define TPREG r13
|
|
# else
|
|
# define LOAD lwz
|
|
# define TPREG r2
|
|
# endif
|
|
# define PTR_MANGLE(reg, tmpreg) \
|
|
LOAD tmpreg,POINTER_GUARD(TPREG); \
|
|
xor reg,tmpreg,reg
|
|
# define PTR_MANGLE2(reg, tmpreg) \
|
|
xor reg,tmpreg,reg
|
|
# define PTR_MANGLE3(destreg, reg, tmpreg) \
|
|
LOAD tmpreg,POINTER_GUARD(TPREG); \
|
|
xor destreg,tmpreg,reg
|
|
# define PTR_DEMANGLE(reg, tmpreg) PTR_MANGLE (reg, tmpreg)
|
|
# define PTR_DEMANGLE2(reg, tmpreg) PTR_MANGLE2 (reg, tmpreg)
|
|
# define PTR_DEMANGLE3(destreg, reg, tmpreg) PTR_MANGLE3 (destreg, reg, tmpreg)
|
|
# else
|
|
# define PTR_MANGLE(var) \
|
|
(var) = (__typeof (var)) ((uintptr_t) (var) ^ THREAD_GET_POINTER_GUARD ())
|
|
# define PTR_DEMANGLE(var) PTR_MANGLE (var)
|
|
# endif
|
|
#endif
|
|
|
|
/* List of system calls which are supported as vsyscalls. */
|
|
#define VDSO_NAME "LINUX_2.6.15"
|
|
#define VDSO_HASH 123718565
|
|
|
|
#if defined(__PPC64__) || defined(__powerpc64__)
|
|
#define HAVE_CLOCK_GETRES64_VSYSCALL "__kernel_clock_getres"
|
|
#define HAVE_CLOCK_GETTIME64_VSYSCALL "__kernel_clock_gettime"
|
|
#else
|
|
#define HAVE_CLOCK_GETRES_VSYSCALL "__kernel_clock_getres"
|
|
#define HAVE_CLOCK_GETTIME_VSYSCALL "__kernel_clock_gettime"
|
|
#endif
|
|
#define HAVE_GETCPU_VSYSCALL "__kernel_getcpu"
|
|
#define HAVE_TIME_VSYSCALL "__kernel_time"
|
|
#define HAVE_GETTIMEOFDAY_VSYSCALL "__kernel_gettimeofday"
|
|
#define HAVE_GET_TBFREQ "__kernel_get_tbfreq"
|
|
|
|
#endif /* _LINUX_POWERPC_SYSDEP_H */
|