mirror of
https://sourceware.org/git/glibc.git
synced 2024-11-14 01:00:07 +00:00
145 lines
5.7 KiB
C
145 lines
5.7 KiB
C
/* Copyright (C) 2003-2015 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
Contributed by Jakub Jelinek <jakub@redhat.com>, 2003.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include "pthreadP.h"
|
|
#include <lowlevellock.h>
|
|
#include <atomic.h>
|
|
|
|
|
|
unsigned long int __fork_generation attribute_hidden;
|
|
|
|
|
|
static void
|
|
clear_once_control (void *arg)
|
|
{
|
|
pthread_once_t *once_control = (pthread_once_t *) arg;
|
|
|
|
/* Reset to the uninitialized state here. We don't need a stronger memory
|
|
order because we do not need to make any other of our writes visible to
|
|
other threads that see this value: This function will be called if we
|
|
get interrupted (see __pthread_once), so all we need to relay to other
|
|
threads is the state being reset again. */
|
|
atomic_store_relaxed (once_control, 0);
|
|
lll_futex_wake (once_control, INT_MAX, LLL_PRIVATE);
|
|
}
|
|
|
|
|
|
/* This is similar to a lock implementation, but we distinguish between three
|
|
states: not yet initialized (0), initialization in progress
|
|
(__fork_generation | __PTHREAD_ONCE_INPROGRESS), and initialization
|
|
finished (__PTHREAD_ONCE_DONE); __fork_generation does not use the bits
|
|
that are used for __PTHREAD_ONCE_INPROGRESS and __PTHREAD_ONCE_DONE (which
|
|
is what __PTHREAD_ONCE_FORK_GEN_INCR is used for). If in the first state,
|
|
threads will try to run the initialization by moving to the second state;
|
|
the first thread to do so via a CAS on once_control runs init_routine,
|
|
other threads block.
|
|
When forking the process, some threads can be interrupted during the second
|
|
state; they won't be present in the forked child, so we need to restart
|
|
initialization in the child. To distinguish an in-progress initialization
|
|
from an interrupted initialization (in which case we need to reclaim the
|
|
lock), we look at the fork generation that's part of the second state: We
|
|
can reclaim iff it differs from the current fork generation.
|
|
XXX: This algorithm has an ABA issue on the fork generation: If an
|
|
initialization is interrupted, we then fork 2^30 times (30 bits of
|
|
once_control are used for the fork generation), and try to initialize
|
|
again, we can deadlock because we can't distinguish the in-progress and
|
|
interrupted cases anymore.
|
|
XXX: We split out this slow path because current compilers do not generate
|
|
as efficient code when the fast path in __pthread_once below is not in a
|
|
separate function. */
|
|
static int
|
|
__attribute__ ((noinline))
|
|
__pthread_once_slow (pthread_once_t *once_control, void (*init_routine) (void))
|
|
{
|
|
while (1)
|
|
{
|
|
int val, newval;
|
|
|
|
/* We need acquire memory order for this load because if the value
|
|
signals that initialization has finished, we need to see any
|
|
data modifications done during initialization. */
|
|
val = atomic_load_acquire (once_control);
|
|
do
|
|
{
|
|
/* Check if the initialization has already been done. */
|
|
if (__glibc_likely ((val & __PTHREAD_ONCE_DONE) != 0))
|
|
return 0;
|
|
|
|
/* We try to set the state to in-progress and having the current
|
|
fork generation. We don't need atomic accesses for the fork
|
|
generation because it's immutable in a particular process, and
|
|
forked child processes start with a single thread that modified
|
|
the generation. */
|
|
newval = __fork_generation | __PTHREAD_ONCE_INPROGRESS;
|
|
/* We need acquire memory order here for the same reason as for the
|
|
load from once_control above. */
|
|
}
|
|
while (__glibc_unlikely (!atomic_compare_exchange_weak_acquire (
|
|
once_control, &val, newval)));
|
|
|
|
/* Check if another thread already runs the initializer. */
|
|
if ((val & __PTHREAD_ONCE_INPROGRESS) != 0)
|
|
{
|
|
/* Check whether the initializer execution was interrupted by a
|
|
fork. We know that for both values, __PTHREAD_ONCE_INPROGRESS
|
|
is set and __PTHREAD_ONCE_DONE is not. */
|
|
if (val == newval)
|
|
{
|
|
/* Same generation, some other thread was faster. Wait. */
|
|
lll_futex_wait (once_control, newval, LLL_PRIVATE);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
/* This thread is the first here. Do the initialization.
|
|
Register a cleanup handler so that in case the thread gets
|
|
interrupted the initialization can be restarted. */
|
|
pthread_cleanup_push (clear_once_control, once_control);
|
|
|
|
init_routine ();
|
|
|
|
pthread_cleanup_pop (0);
|
|
|
|
|
|
/* Mark *once_control as having finished the initialization. We need
|
|
release memory order here because we need to synchronize with other
|
|
threads that want to use the initialized data. */
|
|
atomic_store_release (once_control, __PTHREAD_ONCE_DONE);
|
|
|
|
/* Wake up all other threads. */
|
|
lll_futex_wake (once_control, INT_MAX, LLL_PRIVATE);
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
__pthread_once (pthread_once_t *once_control, void (*init_routine) (void))
|
|
{
|
|
/* Fast path. See __pthread_once_slow. */
|
|
int val;
|
|
val = atomic_load_acquire (once_control);
|
|
if (__glibc_likely ((val & __PTHREAD_ONCE_DONE) != 0))
|
|
return 0;
|
|
else
|
|
return __pthread_once_slow (once_control, init_routine);
|
|
}
|
|
weak_alias (__pthread_once, pthread_once)
|
|
hidden_def (__pthread_once)
|