mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-26 20:51:11 +00:00
403b4feb22
The race leads either to pthread_mutex_destroy returning EBUSY or triggering an assertion (See description in bugzilla). This patch is fixing the race by ensuring that the elision path is used in all cases if elision is enabled by the GLIBC_TUNABLES framework. The __kind variable in struct __pthread_mutex_s is accessed concurrently. Therefore we are now using the atomic macros. The new testcase tst-mutex10 is triggering the race on s390x and intel. Presumably also on power, but I don't have access to a power machine with lock-elision. At least the code for power is the same as on the other two architectures. ChangeLog: [BZ #23275] * nptl/tst-mutex10.c: New File. * nptl/Makefile (tests): Add tst-mutex10. (tst-mutex10-ENV): New variable. * sysdeps/unix/sysv/linux/s390/force-elision.h: (FORCE_ELISION): Ensure that elision path is used if elision is available. * sysdeps/unix/sysv/linux/powerpc/force-elision.h (FORCE_ELISION): Likewise. * sysdeps/unix/sysv/linux/x86/force-elision.h: (FORCE_ELISION): Likewise. * nptl/pthreadP.h (PTHREAD_MUTEX_TYPE, PTHREAD_MUTEX_TYPE_ELISION) (PTHREAD_MUTEX_PSHARED): Use atomic_load_relaxed. * nptl/pthread_mutex_consistent.c (pthread_mutex_consistent): Likewise. * nptl/pthread_mutex_getprioceiling.c (pthread_mutex_getprioceiling): Likewise. * nptl/pthread_mutex_lock.c (__pthread_mutex_lock_full) (__pthread_mutex_cond_lock_adjust): Likewise. * nptl/pthread_mutex_setprioceiling.c (pthread_mutex_setprioceiling): Likewise. * nptl/pthread_mutex_timedlock.c (__pthread_mutex_timedlock): Likewise. * nptl/pthread_mutex_trylock.c (__pthread_mutex_trylock): Likewise. * nptl/pthread_mutex_unlock.c (__pthread_mutex_unlock_full): Likewise. * sysdeps/nptl/bits/thread-shared-types.h (struct __pthread_mutex_s): Add comments. * nptl/pthread_mutex_destroy.c (__pthread_mutex_destroy): Use atomic_load_relaxed and atomic_store_relaxed. * nptl/pthread_mutex_init.c (__pthread_mutex_init): Use atomic_store_relaxed.
167 lines
5.3 KiB
C
167 lines
5.3 KiB
C
/* Copyright (C) 2002-2018 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
Contributed by Ulrich Drepper <drepper@redhat.com>, 2002.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include <assert.h>
|
|
#include <errno.h>
|
|
#include <stdbool.h>
|
|
#include <string.h>
|
|
#include <kernel-features.h>
|
|
#include "pthreadP.h"
|
|
#include <atomic.h>
|
|
#include <pthread-offsets.h>
|
|
|
|
#include <stap-probe.h>
|
|
|
|
static const struct pthread_mutexattr default_mutexattr =
|
|
{
|
|
/* Default is a normal mutex, not shared between processes. */
|
|
.mutexkind = PTHREAD_MUTEX_NORMAL
|
|
};
|
|
|
|
|
|
static bool
|
|
prio_inherit_missing (void)
|
|
{
|
|
#ifdef __NR_futex
|
|
static int tpi_supported;
|
|
if (__glibc_unlikely (tpi_supported == 0))
|
|
{
|
|
int lock = 0;
|
|
INTERNAL_SYSCALL_DECL (err);
|
|
int ret = INTERNAL_SYSCALL (futex, err, 4, &lock, FUTEX_UNLOCK_PI, 0, 0);
|
|
assert (INTERNAL_SYSCALL_ERROR_P (ret, err));
|
|
tpi_supported = INTERNAL_SYSCALL_ERRNO (ret, err) == ENOSYS ? -1 : 1;
|
|
}
|
|
return __glibc_unlikely (tpi_supported < 0);
|
|
#endif
|
|
return true;
|
|
}
|
|
|
|
int
|
|
__pthread_mutex_init (pthread_mutex_t *mutex,
|
|
const pthread_mutexattr_t *mutexattr)
|
|
{
|
|
const struct pthread_mutexattr *imutexattr;
|
|
|
|
ASSERT_TYPE_SIZE (pthread_mutex_t, __SIZEOF_PTHREAD_MUTEX_T);
|
|
|
|
ASSERT_PTHREAD_INTERNAL_OFFSET (pthread_mutex_t, __data.__nusers,
|
|
__PTHREAD_MUTEX_NUSERS_OFFSET);
|
|
ASSERT_PTHREAD_INTERNAL_OFFSET (pthread_mutex_t, __data.__kind,
|
|
__PTHREAD_MUTEX_KIND_OFFSET);
|
|
ASSERT_PTHREAD_INTERNAL_OFFSET (pthread_mutex_t, __data.__spins,
|
|
__PTHREAD_MUTEX_SPINS_OFFSET);
|
|
#if __PTHREAD_MUTEX_LOCK_ELISION
|
|
ASSERT_PTHREAD_INTERNAL_OFFSET (pthread_mutex_t, __data.__elision,
|
|
__PTHREAD_MUTEX_ELISION_OFFSET);
|
|
#endif
|
|
ASSERT_PTHREAD_INTERNAL_OFFSET (pthread_mutex_t, __data.__list,
|
|
__PTHREAD_MUTEX_LIST_OFFSET);
|
|
|
|
imutexattr = ((const struct pthread_mutexattr *) mutexattr
|
|
?: &default_mutexattr);
|
|
|
|
/* Sanity checks. */
|
|
switch (__builtin_expect (imutexattr->mutexkind
|
|
& PTHREAD_MUTEXATTR_PROTOCOL_MASK,
|
|
PTHREAD_PRIO_NONE
|
|
<< PTHREAD_MUTEXATTR_PROTOCOL_SHIFT))
|
|
{
|
|
case PTHREAD_PRIO_NONE << PTHREAD_MUTEXATTR_PROTOCOL_SHIFT:
|
|
break;
|
|
|
|
case PTHREAD_PRIO_INHERIT << PTHREAD_MUTEXATTR_PROTOCOL_SHIFT:
|
|
if (__glibc_unlikely (prio_inherit_missing ()))
|
|
return ENOTSUP;
|
|
break;
|
|
|
|
default:
|
|
/* XXX: For now we don't support robust priority protected mutexes. */
|
|
if (imutexattr->mutexkind & PTHREAD_MUTEXATTR_FLAG_ROBUST)
|
|
return ENOTSUP;
|
|
break;
|
|
}
|
|
|
|
/* Clear the whole variable. */
|
|
memset (mutex, '\0', __SIZEOF_PTHREAD_MUTEX_T);
|
|
|
|
/* Copy the values from the attribute. */
|
|
int mutex_kind = imutexattr->mutexkind & ~PTHREAD_MUTEXATTR_FLAG_BITS;
|
|
|
|
if ((imutexattr->mutexkind & PTHREAD_MUTEXATTR_FLAG_ROBUST) != 0)
|
|
{
|
|
#ifndef __ASSUME_SET_ROBUST_LIST
|
|
if ((imutexattr->mutexkind & PTHREAD_MUTEXATTR_FLAG_PSHARED) != 0
|
|
&& __set_robust_list_avail < 0)
|
|
return ENOTSUP;
|
|
#endif
|
|
|
|
mutex_kind |= PTHREAD_MUTEX_ROBUST_NORMAL_NP;
|
|
}
|
|
|
|
switch (imutexattr->mutexkind & PTHREAD_MUTEXATTR_PROTOCOL_MASK)
|
|
{
|
|
case PTHREAD_PRIO_INHERIT << PTHREAD_MUTEXATTR_PROTOCOL_SHIFT:
|
|
mutex_kind |= PTHREAD_MUTEX_PRIO_INHERIT_NP;
|
|
break;
|
|
|
|
case PTHREAD_PRIO_PROTECT << PTHREAD_MUTEXATTR_PROTOCOL_SHIFT:
|
|
mutex_kind |= PTHREAD_MUTEX_PRIO_PROTECT_NP;
|
|
|
|
int ceiling = (imutexattr->mutexkind
|
|
& PTHREAD_MUTEXATTR_PRIO_CEILING_MASK)
|
|
>> PTHREAD_MUTEXATTR_PRIO_CEILING_SHIFT;
|
|
if (! ceiling)
|
|
{
|
|
/* See __init_sched_fifo_prio. */
|
|
if (atomic_load_relaxed (&__sched_fifo_min_prio) == -1)
|
|
__init_sched_fifo_prio ();
|
|
if (ceiling < atomic_load_relaxed (&__sched_fifo_min_prio))
|
|
ceiling = atomic_load_relaxed (&__sched_fifo_min_prio);
|
|
}
|
|
mutex->__data.__lock = ceiling << PTHREAD_MUTEX_PRIO_CEILING_SHIFT;
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/* The kernel when waking robust mutexes on exit never uses
|
|
FUTEX_PRIVATE_FLAG FUTEX_WAKE. */
|
|
if ((imutexattr->mutexkind & (PTHREAD_MUTEXATTR_FLAG_PSHARED
|
|
| PTHREAD_MUTEXATTR_FLAG_ROBUST)) != 0)
|
|
mutex_kind |= PTHREAD_MUTEX_PSHARED_BIT;
|
|
|
|
/* See concurrency notes regarding __kind in struct __pthread_mutex_s
|
|
in sysdeps/nptl/bits/thread-shared-types.h. */
|
|
atomic_store_relaxed (&(mutex->__data.__kind), mutex_kind);
|
|
|
|
/* Default values: mutex not used yet. */
|
|
// mutex->__count = 0; already done by memset
|
|
// mutex->__owner = 0; already done by memset
|
|
// mutex->__nusers = 0; already done by memset
|
|
// mutex->__spins = 0; already done by memset
|
|
// mutex->__next = NULL; already done by memset
|
|
|
|
LIBC_PROBE (mutex_init, 1, mutex);
|
|
|
|
return 0;
|
|
}
|
|
weak_alias (__pthread_mutex_init, pthread_mutex_init)
|
|
hidden_def (__pthread_mutex_init)
|