/* Spin locks for communication between threads and signal handlers.
Copyright (C) 2020-2024 Free Software Foundation, Inc.
This file is free software: you can redistribute it and/or modify
it under the terms of the GNU Lesser General Public License as
published by the Free Software Foundation; either version 2.1 of the
License, or (at your option) any later version.
This file is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public License
along with this program. If not, see . */
/* Written by Bruno Haible , 2020. */
#include
/* Specification. */
#include "asyncsafe-spin.h"
#include
#if defined _AIX
# include
#endif
#if 0x590 <= __SUNPRO_C && __STDC__
# define asm __asm
#endif
#if defined _WIN32 && ! defined __CYGWIN__
/* Use Windows threads. */
void
asyncsafe_spin_init (asyncsafe_spinlock_t *lock)
{
glwthread_spin_init (lock);
}
static inline void
do_lock (asyncsafe_spinlock_t *lock)
{
glwthread_spin_lock (lock);
}
static inline void
do_unlock (asyncsafe_spinlock_t *lock)
{
if (glwthread_spin_unlock (lock))
abort ();
}
void
asyncsafe_spin_destroy (asyncsafe_spinlock_t *lock)
{
glwthread_spin_destroy (lock);
}
#else
# if HAVE_PTHREAD_H
/* Use POSIX threads. */
/* We don't use semaphores (although sem_post() is allowed in signal handlers),
because it would require to link with -lrt on HP-UX 11, OSF/1, Solaris 10,
and also because on macOS only named semaphores work.
We don't use the C11 (available in GCC >= 4.9) because it would
require to link with -latomic. */
# if (__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 7) \
|| __clang_major__ > 3 || (__clang_major__ == 3 && __clang_minor__ >= 1)) \
&& !defined __ibmxl__
/* Use GCC built-ins (available in GCC >= 4.7 and clang >= 3.1) that operate on
the first byte of the lock.
Documentation:
*/
# if 1
/* An implementation that verifies the unlocks. */
void
asyncsafe_spin_init (asyncsafe_spinlock_t *lock)
{
__atomic_store_n (lock, 0, __ATOMIC_SEQ_CST);
}
static inline void
do_lock (asyncsafe_spinlock_t *lock)
{
/* Wait until *lock becomes 0, then replace it with 1. */
asyncsafe_spinlock_t zero;
while (!(zero = 0,
__atomic_compare_exchange_n (lock, &zero, 1, false,
__ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST)))
;
}
static inline void
do_unlock (asyncsafe_spinlock_t *lock)
{
/* If *lock is 1, then replace it with 0. */
asyncsafe_spinlock_t one = 1;
if (!__atomic_compare_exchange_n (lock, &one, 0, false,
__ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST))
abort ();
}
# else
/* An implementation that is a little bit more optimized, but does not verify
the unlocks. */
void
asyncsafe_spin_init (asyncsafe_spinlock_t *lock)
{
__atomic_clear (lock, __ATOMIC_SEQ_CST);
}
static inline void
do_lock (asyncsafe_spinlock_t *lock)
{
while (__atomic_test_and_set (lock, __ATOMIC_SEQ_CST))
;
}
static inline void
do_unlock (asyncsafe_spinlock_t *lock)
{
__atomic_clear (lock, __ATOMIC_SEQ_CST);
}
# endif
# elif (((__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 1)) \
|| __clang_major__ >= 3) \
&& HAVE_ATOMIC_COMPARE_AND_SWAP_GCC41)
/* Use GCC built-ins (available on many platforms with GCC >= 4.1 or
clang >= 3.0).
Documentation:
*/
void
asyncsafe_spin_init (asyncsafe_spinlock_t *lock)
{
volatile unsigned int *vp = lock;
*vp = 0;
__sync_synchronize ();
}
static inline void
do_lock (asyncsafe_spinlock_t *lock)
{
/* Wait until *lock becomes 0, then replace it with 1. */
while (__sync_val_compare_and_swap (lock, 0, 1) != 0)
;
}
static inline void
do_unlock (asyncsafe_spinlock_t *lock)
{
/* If *lock is 1, then replace it with 0. */
if (__sync_val_compare_and_swap (lock, 1, 0) != 1)
abort ();
}
# elif defined _AIX
/* AIX */
void
asyncsafe_spin_init (asyncsafe_spinlock_t *lock)
{
atomic_p vp = (int *) lock;
_clear_lock (vp, 0);
}
static inline void
do_lock (asyncsafe_spinlock_t *lock)
{
atomic_p vp = (int *) lock;
while (_check_lock (vp, 0, 1))
;
}
static inline void
do_unlock (asyncsafe_spinlock_t *lock)
{
atomic_p vp = (int *) lock;
if (_check_lock (vp, 1, 0))
abort ();
}
# elif ((defined __GNUC__ || defined __clang__ || defined __SUNPRO_C) && (defined __sparc || defined __i386 || defined __x86_64__)) || (defined __TINYC__ && (defined __i386 || defined __x86_64__))
/* For older versions of GCC or clang, use inline assembly.
GCC, clang, and the Oracle Studio C 12 compiler understand GCC's extended
asm syntax, but the plain Oracle Studio C 11 compiler understands only
simple asm. */
/* An implementation that verifies the unlocks. */
static void
memory_barrier (void)
{
# if defined __GNUC__ || defined __clang__ || __SUNPRO_C >= 0x590 || defined __TINYC__
# if defined __i386 || defined __x86_64__
# if defined __TINYC__ && defined __i386
/* Cannot use the SSE instruction "mfence" with this compiler. */
asm volatile ("lock orl $0,(%esp)");
# else
asm volatile ("mfence");
# endif
# endif
# if defined __sparc
asm volatile ("membar 2");
# endif
# else
# if defined __i386 || defined __x86_64__
asm ("mfence");
# endif
# if defined __sparc
asm ("membar 2");
# endif
# endif
}
/* Store NEWVAL in *VP if the old value *VP is == CMP.
Return the old value. */
static unsigned int
atomic_compare_and_swap (volatile unsigned int *vp, unsigned int cmp,
unsigned int newval)
{
# if defined __GNUC__ || defined __clang__ || __SUNPRO_C >= 0x590 || defined __TINYC__
unsigned int oldval;
# if defined __i386 || defined __x86_64__
asm volatile (" lock\n cmpxchgl %3,(%1)"
: "=a" (oldval) : "r" (vp), "a" (cmp), "r" (newval) : "memory");
# endif
# if defined __sparc
asm volatile (" cas [%1],%2,%3\n"
" mov %3,%0"
: "=r" (oldval) : "r" (vp), "r" (cmp), "r" (newval) : "memory");
# endif
return oldval;
# else /* __SUNPRO_C */
# if defined __x86_64__
asm (" movl %esi,%eax\n"
" lock\n cmpxchgl %edx,(%rdi)");
# elif defined __i386
asm (" movl 16(%ebp),%ecx\n"
" movl 12(%ebp),%eax\n"
" movl 8(%ebp),%edx\n"
" lock\n cmpxchgl %ecx,(%edx)");
# endif
# if defined __sparc
asm (" cas [%i0],%i1,%i2\n"
" mov %i2,%i0");
# endif
# endif
}
void
asyncsafe_spin_init (asyncsafe_spinlock_t *lock)
{
volatile unsigned int *vp = lock;
*vp = 0;
memory_barrier ();
}
static inline void
do_lock (asyncsafe_spinlock_t *lock)
{
volatile unsigned int *vp = lock;
while (atomic_compare_and_swap (vp, 0, 1) != 0)
;
}
static inline void
do_unlock (asyncsafe_spinlock_t *lock)
{
volatile unsigned int *vp = lock;
if (atomic_compare_and_swap (vp, 1, 0) != 1)
abort ();
}
# else
/* Fallback code. It has some race conditions. */
void
asyncsafe_spin_init (asyncsafe_spinlock_t *lock)
{
volatile unsigned int *vp = lock;
*vp = 0;
}
static inline void
do_lock (asyncsafe_spinlock_t *lock)
{
volatile unsigned int *vp = lock;
while (*vp)
;
*vp = 1;
}
static inline void
do_unlock (asyncsafe_spinlock_t *lock)
{
volatile unsigned int *vp = lock;
*vp = 0;
}
# endif
# else
/* Provide a dummy implementation for single-threaded applications. */
void
asyncsafe_spin_init (asyncsafe_spinlock_t *lock)
{
}
static inline void
do_lock (asyncsafe_spinlock_t *lock)
{
}
static inline void
do_unlock (asyncsafe_spinlock_t *lock)
{
}
# endif
void
asyncsafe_spin_destroy (asyncsafe_spinlock_t *lock)
{
}
#endif
void
asyncsafe_spin_lock (asyncsafe_spinlock_t *lock,
const sigset_t *mask, sigset_t *saved_mask)
{
sigprocmask (SIG_BLOCK, mask, saved_mask); /* equivalent to pthread_sigmask */
do_lock (lock);
}
void
asyncsafe_spin_unlock (asyncsafe_spinlock_t *lock, const sigset_t *saved_mask)
{
do_unlock (lock);
sigprocmask (SIG_SETMASK, saved_mask, NULL); /* equivalent to pthread_sigmask */
}