mirror of
https://sourceware.org/git/glibc.git
synced 2024-11-22 13:00:06 +00:00
8bd336a00a
And make it an installed header. This addresses a few aliasing violations (which do not seem to result in miscompilation due to the use of atomics), and also enables use of wide counters in other parts of the library. The debug output in nptl/tst-cond22 has been adjusted to print the 32-bit values instead because it avoids a big-endian/little-endian difference. Reviewed-by: Adhemerval Zanella <adhemerval.zanella@linaro.org>
90 lines
2.8 KiB
C
90 lines
2.8 KiB
C
/* Monotonically increasing wide counters (at least 62 bits).
|
|
Copyright (C) 2016-2021 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<https://www.gnu.org/licenses/>. */
|
|
|
|
#ifndef _ATOMIC_WIDE_COUNTER_H
|
|
#define _ATOMIC_WIDE_COUNTER_H
|
|
|
|
#include <atomic.h>
|
|
#include <bits/atomic_wide_counter.h>
|
|
|
|
#if __HAVE_64B_ATOMICS
|
|
|
|
static inline uint64_t
|
|
__atomic_wide_counter_load_relaxed (__atomic_wide_counter *c)
|
|
{
|
|
return atomic_load_relaxed (&c->__value64);
|
|
}
|
|
|
|
static inline uint64_t
|
|
__atomic_wide_counter_fetch_add_relaxed (__atomic_wide_counter *c,
|
|
unsigned int val)
|
|
{
|
|
return atomic_fetch_add_relaxed (&c->__value64, val);
|
|
}
|
|
|
|
static inline uint64_t
|
|
__atomic_wide_counter_fetch_add_acquire (__atomic_wide_counter *c,
|
|
unsigned int val)
|
|
{
|
|
return atomic_fetch_add_acquire (&c->__value64, val);
|
|
}
|
|
|
|
static inline void
|
|
__atomic_wide_counter_add_relaxed (__atomic_wide_counter *c,
|
|
unsigned int val)
|
|
{
|
|
atomic_store_relaxed (&c->__value64,
|
|
atomic_load_relaxed (&c->__value64) + val);
|
|
}
|
|
|
|
static uint64_t __attribute__ ((unused))
|
|
__atomic_wide_counter_fetch_xor_release (__atomic_wide_counter *c,
|
|
unsigned int val)
|
|
{
|
|
return atomic_fetch_xor_release (&c->__value64, val);
|
|
}
|
|
|
|
#else /* !__HAVE_64B_ATOMICS */
|
|
|
|
uint64_t __atomic_wide_counter_load_relaxed (__atomic_wide_counter *c)
|
|
attribute_hidden;
|
|
|
|
uint64_t __atomic_wide_counter_fetch_add_relaxed (__atomic_wide_counter *c,
|
|
unsigned int op)
|
|
attribute_hidden;
|
|
|
|
static inline uint64_t
|
|
__atomic_wide_counter_fetch_add_acquire (__atomic_wide_counter *c,
|
|
unsigned int val)
|
|
{
|
|
uint64_t r = __atomic_wide_counter_fetch_add_relaxed (c, val);
|
|
atomic_thread_fence_acquire ();
|
|
return r;
|
|
}
|
|
|
|
static inline void
|
|
__atomic_wide_counter_add_relaxed (__atomic_wide_counter *c,
|
|
unsigned int val)
|
|
{
|
|
__atomic_wide_counter_fetch_add_relaxed (c, val);
|
|
}
|
|
|
|
#endif /* !__HAVE_64B_ATOMICS */
|
|
|
|
#endif /* _ATOMIC_WIDE_COUNTER_H */
|