Commit d055e28f authored by Sergey Vojtovich's avatar Sergey Vojtovich

MDEV-10813 - Clean-up InnoDB atomics, memory barriers and mutexes

Simplified InnoDB mutex implementations, corrected memory barriers usage, use
server atomic builtins.
parent cc49f009
...@@ -31,6 +31,7 @@ Created 2013-03-26 Sunny Bains. ...@@ -31,6 +31,7 @@ Created 2013-03-26 Sunny Bains.
#include "ut0ut.h" #include "ut0ut.h"
#include "ut0rnd.h" #include "ut0rnd.h"
#include "os0event.h" #include "os0event.h"
#include "sync0arr.h"
/** OS mutex for tracking lock/unlock for debugging */ /** OS mutex for tracking lock/unlock for debugging */
template <template <typename> class Policy = NoPolicy> template <template <typename> class Policy = NoPolicy>
...@@ -67,8 +68,6 @@ struct OSTrackMutex { ...@@ -67,8 +68,6 @@ struct OSTrackMutex {
m_mutex.init(); m_mutex.init();
ut_d(m_freed = false); ut_d(m_freed = false);
m_policy.init(*this, id, filename, line);
} }
/** Destroy the mutex */ /** Destroy the mutex */
...@@ -80,8 +79,6 @@ struct OSTrackMutex { ...@@ -80,8 +79,6 @@ struct OSTrackMutex {
m_mutex.destroy(); m_mutex.destroy();
ut_d(m_freed = true); ut_d(m_freed = true);
m_policy.destroy();
} }
/** Release the mutex. */ /** Release the mutex. */
...@@ -129,15 +126,6 @@ struct OSTrackMutex { ...@@ -129,15 +126,6 @@ struct OSTrackMutex {
return(locked); return(locked);
} }
#ifdef UNIV_DEBUG
/** @return true if the thread owns the mutex. */
bool is_owned() const
UNIV_NOTHROW
{
return(m_locked && m_policy.is_owned());
}
#endif /* UNIV_DEBUG */
/** @return non-const version of the policy */ /** @return non-const version of the policy */
MutexPolicy& policy() MutexPolicy& policy()
UNIV_NOTHROW UNIV_NOTHROW
...@@ -208,7 +196,6 @@ struct TTASFutexMutex { ...@@ -208,7 +196,6 @@ struct TTASFutexMutex {
UNIV_NOTHROW UNIV_NOTHROW
{ {
ut_a(m_lock_word == MUTEX_STATE_UNLOCKED); ut_a(m_lock_word == MUTEX_STATE_UNLOCKED);
m_policy.init(*this, id, filename, line);
} }
/** Destroy the mutex. */ /** Destroy the mutex. */
...@@ -216,7 +203,6 @@ struct TTASFutexMutex { ...@@ -216,7 +203,6 @@ struct TTASFutexMutex {
{ {
/* The destructor can be called at shutdown. */ /* The destructor can be called at shutdown. */
ut_a(m_lock_word == MUTEX_STATE_UNLOCKED); ut_a(m_lock_word == MUTEX_STATE_UNLOCKED);
m_policy.destroy();
} }
/** Acquire the mutex. /** Acquire the mutex.
...@@ -230,29 +216,28 @@ struct TTASFutexMutex { ...@@ -230,29 +216,28 @@ struct TTASFutexMutex {
const char* filename, const char* filename,
uint32_t line) UNIV_NOTHROW uint32_t line) UNIV_NOTHROW
{ {
uint32_t n_spins; uint32_t n_spins, n_waits;
lock_word_t lock = ttas(max_spins, max_delay, n_spins);
/* If there were no waiters when this thread tried
to acquire the mutex then set the waiters flag now.
Additionally, when this thread set the waiters flag it is
possible that the mutex had already been released
by then. In this case the thread can assume it
was granted the mutex. */
uint32_t n_waits; for (n_spins= 0; n_spins < max_spins; n_spins++) {
if (try_lock()) {
if (lock != MUTEX_STATE_UNLOCKED) { m_policy.add(n_spins, 0);
return;
}
if (lock != MUTEX_STATE_LOCKED || !set_waiters()) { ut_delay(ut_rnd_interval(0, max_delay));
}
n_waits = wait(); for (n_waits= 0;; n_waits++) {
} else { if (my_atomic_fas32_explicit(&m_lock_word,
n_waits = 0; MUTEX_STATE_WAITERS,
MY_MEMORY_ORDER_ACQUIRE)
== MUTEX_STATE_UNLOCKED) {
break;
} }
} else { syscall(SYS_futex, &m_lock_word,
n_waits = 0; FUTEX_WAIT_PRIVATE, MUTEX_STATE_WAITERS,
0, 0, 0);
} }
m_policy.add(n_spins, n_waits); m_policy.add(n_spins, n_waits);
...@@ -261,53 +246,26 @@ struct TTASFutexMutex { ...@@ -261,53 +246,26 @@ struct TTASFutexMutex {
/** Release the mutex. */ /** Release the mutex. */
void exit() UNIV_NOTHROW void exit() UNIV_NOTHROW
{ {
/* If there are threads waiting then we have to wake if (my_atomic_fas32_explicit(&m_lock_word,
them up. Reset the lock state to unlocked so that waiting MUTEX_STATE_UNLOCKED,
threads can test for success. */ MY_MEMORY_ORDER_RELEASE)
== MUTEX_STATE_WAITERS) {
os_rmb; syscall(SYS_futex, &m_lock_word, FUTEX_WAKE_PRIVATE,
1, 0, 0, 0);
if (state() == MUTEX_STATE_WAITERS) {
m_lock_word = MUTEX_STATE_UNLOCKED;
} else if (unlock() == MUTEX_STATE_LOCKED) {
/* No threads waiting, no need to signal a wakeup. */
return;
} }
signal();
}
/** Try and lock the mutex.
@return the old state of the mutex */
lock_word_t trylock() UNIV_NOTHROW
{
return(CAS(&m_lock_word,
MUTEX_STATE_UNLOCKED, MUTEX_STATE_LOCKED));
} }
/** Try and lock the mutex. /** Try and lock the mutex.
@return true if successful */ @return true if successful */
bool try_lock() UNIV_NOTHROW bool try_lock() UNIV_NOTHROW
{ {
return(trylock() == MUTEX_STATE_UNLOCKED); int32 oldval = MUTEX_STATE_UNLOCKED;
return(my_atomic_cas32_strong_explicit(&m_lock_word, &oldval,
MUTEX_STATE_LOCKED,
MY_MEMORY_ORDER_ACQUIRE,
MY_MEMORY_ORDER_RELAXED));
} }
/** @return true if mutex is unlocked */
bool is_locked() const UNIV_NOTHROW
{
return(state() != MUTEX_STATE_UNLOCKED);
}
#ifdef UNIV_DEBUG
/** @return true if the thread owns the mutex. */
bool is_owned() const UNIV_NOTHROW
{
return(is_locked() && m_policy.is_owned());
}
#endif /* UNIV_DEBUG */
/** @return non-const version of the policy */ /** @return non-const version of the policy */
MutexPolicy& policy() UNIV_NOTHROW MutexPolicy& policy() UNIV_NOTHROW
{ {
...@@ -320,105 +278,12 @@ struct TTASFutexMutex { ...@@ -320,105 +278,12 @@ struct TTASFutexMutex {
return(m_policy); return(m_policy);
} }
private: private:
/** @return the lock state. */
lock_word_t state() const UNIV_NOTHROW
{
return(m_lock_word);
}
/** Release the mutex.
@return the new state of the mutex */
lock_word_t unlock() UNIV_NOTHROW
{
return(TAS(&m_lock_word, MUTEX_STATE_UNLOCKED));
}
/** Note that there are threads waiting and need to be woken up.
@return true if state was MUTEX_STATE_UNLOCKED (ie. granted) */
bool set_waiters() UNIV_NOTHROW
{
return(TAS(&m_lock_word, MUTEX_STATE_WAITERS)
== MUTEX_STATE_UNLOCKED);
}
/** Set the waiters flag, only if the mutex is locked
@return true if succesful. */
bool try_set_waiters() UNIV_NOTHROW
{
return(CAS(&m_lock_word,
MUTEX_STATE_LOCKED, MUTEX_STATE_WAITERS)
!= MUTEX_STATE_UNLOCKED);
}
/** Wait if the lock is contended.
@return the number of waits */
uint32_t wait() UNIV_NOTHROW
{
uint32_t n_waits = 0;
/* Use FUTEX_WAIT_PRIVATE because our mutexes are
not shared between processes. */
do {
++n_waits;
syscall(SYS_futex, &m_lock_word,
FUTEX_WAIT_PRIVATE, MUTEX_STATE_WAITERS,
0, 0, 0);
// Since we are retrying the operation the return
// value doesn't matter.
} while (!set_waiters());
return(n_waits);
}
/** Wakeup a waiting thread */
void signal() UNIV_NOTHROW
{
syscall(SYS_futex, &m_lock_word, FUTEX_WAKE_PRIVATE,
MUTEX_STATE_LOCKED, 0, 0, 0);
}
/** Poll waiting for mutex to be unlocked.
@param[in] max_spins max spins
@param[in] max_delay max delay per spin
@param[out] n_spins retries before acquire
@return value of lock word before locking. */
lock_word_t ttas(
uint32_t max_spins,
uint32_t max_delay,
uint32_t& n_spins) UNIV_NOTHROW
{
os_rmb;
for (n_spins = 0; n_spins < max_spins; ++n_spins) {
if (!is_locked()) {
lock_word_t lock = trylock();
if (lock == MUTEX_STATE_UNLOCKED) {
/* Lock successful */
return(lock);
}
}
ut_delay(ut_rnd_interval(0, max_delay));
}
return(trylock());
}
private:
/** Policy data */ /** Policy data */
MutexPolicy m_policy; MutexPolicy m_policy;
/** lock_word is the target of the atomic test-and-set instruction /** lock_word is the target of the atomic test-and-set instruction
when atomic operations are enabled. */ when atomic operations are enabled. */
lock_word_t m_lock_word MY_ALIGNED(MY_ALIGNOF(ulint)); int32 m_lock_word;
}; };
#endif /* HAVE_IB_LINUX_FUTEX */ #endif /* HAVE_IB_LINUX_FUTEX */
...@@ -453,7 +318,6 @@ struct TTASMutex { ...@@ -453,7 +318,6 @@ struct TTASMutex {
UNIV_NOTHROW UNIV_NOTHROW
{ {
ut_ad(m_lock_word == MUTEX_STATE_UNLOCKED); ut_ad(m_lock_word == MUTEX_STATE_UNLOCKED);
m_policy.init(*this, id, filename, line);
} }
/** Destroy the mutex. */ /** Destroy the mutex. */
...@@ -461,45 +325,25 @@ struct TTASMutex { ...@@ -461,45 +325,25 @@ struct TTASMutex {
{ {
/* The destructor can be called at shutdown. */ /* The destructor can be called at shutdown. */
ut_ad(m_lock_word == MUTEX_STATE_UNLOCKED); ut_ad(m_lock_word == MUTEX_STATE_UNLOCKED);
m_policy.destroy();
}
/**
Try and acquire the lock using TestAndSet.
@return true if lock succeeded */
bool tas_lock() UNIV_NOTHROW
{
return(TAS(&m_lock_word, MUTEX_STATE_LOCKED)
== MUTEX_STATE_UNLOCKED);
}
/** In theory __sync_lock_release should be used to release the lock.
Unfortunately, it does not work properly alone. The workaround is
that more conservative __sync_lock_test_and_set is used instead. */
void tas_unlock() UNIV_NOTHROW
{
#ifdef UNIV_DEBUG
ut_ad(state() == MUTEX_STATE_LOCKED);
lock_word_t lock =
#endif /* UNIV_DEBUG */
TAS(&m_lock_word, MUTEX_STATE_UNLOCKED);
ut_ad(lock == MUTEX_STATE_LOCKED);
} }
/** Try and lock the mutex. /** Try and lock the mutex.
@return true on success */ @return true on success */
bool try_lock() UNIV_NOTHROW bool try_lock() UNIV_NOTHROW
{ {
return(tas_lock()); int32 oldval = MUTEX_STATE_UNLOCKED;
return(my_atomic_cas32_strong_explicit(&m_lock_word, &oldval,
MUTEX_STATE_LOCKED,
MY_MEMORY_ORDER_ACQUIRE,
MY_MEMORY_ORDER_RELAXED));
} }
/** Release the mutex. */ /** Release the mutex. */
void exit() UNIV_NOTHROW void exit() UNIV_NOTHROW
{ {
tas_unlock(); ut_ad(m_lock_word == MUTEX_STATE_LOCKED);
my_atomic_store32_explicit(&m_lock_word, MUTEX_STATE_UNLOCKED,
MY_MEMORY_ORDER_RELEASE);
} }
/** Acquire the mutex. /** Acquire the mutex.
...@@ -513,34 +357,19 @@ struct TTASMutex { ...@@ -513,34 +357,19 @@ struct TTASMutex {
const char* filename, const char* filename,
uint32_t line) UNIV_NOTHROW uint32_t line) UNIV_NOTHROW
{ {
if (!try_lock()) { const uint32_t step = max_spins;
uint32_t n_spins = 0;
uint32_t n_spins = ttas(max_spins, max_delay);
/* No OS waits for spin mutexes */ while (!try_lock()) {
m_policy.add(n_spins, 0); ut_delay(ut_rnd_interval(0, max_delay));
if (++n_spins == max_spins) {
os_thread_yield();
max_spins+= step;
}
} }
}
/** @return the lock state. */ m_policy.add(n_spins, 0);
lock_word_t state() const UNIV_NOTHROW
{
return(m_lock_word);
}
/** @return true if locked by some thread */
bool is_locked() const UNIV_NOTHROW
{
return(m_lock_word != MUTEX_STATE_UNLOCKED);
}
#ifdef UNIV_DEBUG
/** @return true if the calling thread owns the mutex. */
bool is_owned() const UNIV_NOTHROW
{
return(is_locked() && m_policy.is_owned());
} }
#endif /* UNIV_DEBUG */
/** @return non-const version of the policy */ /** @return non-const version of the policy */
MutexPolicy& policy() UNIV_NOTHROW MutexPolicy& policy() UNIV_NOTHROW
...@@ -554,43 +383,6 @@ struct TTASMutex { ...@@ -554,43 +383,6 @@ struct TTASMutex {
return(m_policy); return(m_policy);
} }
private:
/** Spin and try to acquire the lock.
@param[in] max_spins max spins
@param[in] max_delay max delay per spin
@return number spins before acquire */
uint32_t ttas(
uint32_t max_spins,
uint32_t max_delay)
UNIV_NOTHROW
{
uint32_t i = 0;
const uint32_t step = max_spins;
os_rmb;
do {
while (is_locked()) {
ut_delay(ut_rnd_interval(0, max_delay));
++i;
if (i >= max_spins) {
max_spins += step;
os_thread_yield();
break;
}
}
} while (!try_lock());
return(i);
}
private: private:
// Disable copying // Disable copying
TTASMutex(const TTASMutex&); TTASMutex(const TTASMutex&);
...@@ -601,7 +393,7 @@ struct TTASMutex { ...@@ -601,7 +393,7 @@ struct TTASMutex {
/** lock_word is the target of the atomic test-and-set instruction /** lock_word is the target of the atomic test-and-set instruction
when atomic operations are enabled. */ when atomic operations are enabled. */
lock_word_t m_lock_word; int32 m_lock_word;
}; };
template <template <typename> class Policy = NoPolicy> template <template <typename> class Policy = NoPolicy>
...@@ -613,7 +405,6 @@ struct TTASEventMutex { ...@@ -613,7 +405,6 @@ struct TTASEventMutex {
UNIV_NOTHROW UNIV_NOTHROW
: :
m_lock_word(MUTEX_STATE_UNLOCKED), m_lock_word(MUTEX_STATE_UNLOCKED),
m_waiters(),
m_event() m_event()
{ {
/* Check that lock_word is aligned. */ /* Check that lock_word is aligned. */
...@@ -641,8 +432,6 @@ struct TTASEventMutex { ...@@ -641,8 +432,6 @@ struct TTASEventMutex {
ut_a(m_lock_word == MUTEX_STATE_UNLOCKED); ut_a(m_lock_word == MUTEX_STATE_UNLOCKED);
m_event = os_event_create(sync_latch_get_name(id)); m_event = os_event_create(sync_latch_get_name(id));
m_policy.init(*this, id, filename, line);
} }
/** This is the real desctructor. This mutex can be created in BSS and /** This is the real desctructor. This mutex can be created in BSS and
...@@ -656,8 +445,6 @@ struct TTASEventMutex { ...@@ -656,8 +445,6 @@ struct TTASEventMutex {
/* We have to free the event before InnoDB shuts down. */ /* We have to free the event before InnoDB shuts down. */
os_event_destroy(m_event); os_event_destroy(m_event);
m_event = 0; m_event = 0;
m_policy.destroy();
} }
/** Try and lock the mutex. Note: POSIX returns 0 on success. /** Try and lock the mutex. Note: POSIX returns 0 on success.
...@@ -665,29 +452,23 @@ struct TTASEventMutex { ...@@ -665,29 +452,23 @@ struct TTASEventMutex {
bool try_lock() bool try_lock()
UNIV_NOTHROW UNIV_NOTHROW
{ {
return(tas_lock()); int32 oldval = MUTEX_STATE_UNLOCKED;
return(my_atomic_cas32_strong_explicit(&m_lock_word, &oldval,
MUTEX_STATE_LOCKED,
MY_MEMORY_ORDER_ACQUIRE,
MY_MEMORY_ORDER_RELAXED));
} }
/** Release the mutex. */ /** Release the mutex. */
void exit() void exit()
UNIV_NOTHROW UNIV_NOTHROW
{ {
/* A problem: we assume that mutex_reset_lock word if (my_atomic_fas32_explicit(&m_lock_word,
is a memory barrier, that is when we read the waiters MUTEX_STATE_UNLOCKED,
field next, the read must be serialized in memory MY_MEMORY_ORDER_RELEASE)
after the reset. A speculative processor might == MUTEX_STATE_WAITERS) {
perform the read first, which could leave a waiting os_event_set(m_event);
thread hanging indefinitely. sync_array_object_signalled();
Our current solution call every second
sync_arr_wake_threads_if_sema_free()
to wake up possible hanging threads if they are missed
in mutex_signal_object. */
tas_unlock();
if (m_waiters != 0) {
signal();
} }
} }
...@@ -703,13 +484,46 @@ struct TTASEventMutex { ...@@ -703,13 +484,46 @@ struct TTASEventMutex {
uint32_t line) uint32_t line)
UNIV_NOTHROW UNIV_NOTHROW
{ {
if (!try_lock()) { uint32_t n_spins = 0;
spin_and_try_lock(max_spins, max_delay, filename, line); uint32_t n_waits = 0;
const uint32_t step = max_spins;
while (!try_lock()) {
if (n_spins++ == max_spins) {
max_spins += step;
n_waits++;
os_thread_yield();
sync_cell_t* cell;
sync_array_t *sync_arr = sync_array_get_and_reserve_cell(
this,
(m_policy.get_id() == LATCH_ID_BUF_BLOCK_MUTEX
|| m_policy.get_id() == LATCH_ID_BUF_POOL_ZIP)
? SYNC_BUF_BLOCK
: SYNC_MUTEX,
filename, line, &cell);
int32 oldval = MUTEX_STATE_LOCKED;
my_atomic_cas32_strong_explicit(&m_lock_word, &oldval,
MUTEX_STATE_WAITERS,
MY_MEMORY_ORDER_RELAXED,
MY_MEMORY_ORDER_RELAXED);
if (oldval == MUTEX_STATE_UNLOCKED) {
sync_array_free_cell(sync_arr, cell);
} else {
sync_array_wait_event(sync_arr, cell);
}
} else {
ut_delay(ut_rnd_interval(0, max_delay));
}
} }
m_policy.add(n_spins, n_waits);
} }
/** @return the lock state. */ /** @return the lock state. */
lock_word_t state() const int32 state() const
UNIV_NOTHROW UNIV_NOTHROW
{ {
return(m_lock_word); return(m_lock_word);
...@@ -723,22 +537,6 @@ struct TTASEventMutex { ...@@ -723,22 +537,6 @@ struct TTASEventMutex {
return(m_event); return(m_event);
} }
/** @return true if locked by some thread */
bool is_locked() const
UNIV_NOTHROW
{
return(m_lock_word != MUTEX_STATE_UNLOCKED);
}
#ifdef UNIV_DEBUG
/** @return true if the calling thread owns the mutex. */
bool is_owned() const
UNIV_NOTHROW
{
return(is_locked() && m_policy.is_owned());
}
#endif /* UNIV_DEBUG */
/** @return non-const version of the policy */ /** @return non-const version of the policy */
MutexPolicy& policy() MutexPolicy& policy()
UNIV_NOTHROW UNIV_NOTHROW
...@@ -753,153 +551,6 @@ struct TTASEventMutex { ...@@ -753,153 +551,6 @@ struct TTASEventMutex {
return(m_policy); return(m_policy);
} }
private:
/** Wait in the sync array.
@param[in] filename from where it was called
@param[in] line line number in file
@param[in] spin retry this many times again
@return true if the mutex acquisition was successful. */
bool wait(
const char* filename,
uint32_t line,
uint32_t spin)
UNIV_NOTHROW;
/** Spin and wait for the mutex to become free.
@param[in] max_spins max spins
@param[in] max_delay max delay per spin
@param[in,out] n_spins spin start index
@return true if unlocked */
bool is_free(
uint32_t max_spins,
uint32_t max_delay,
uint32_t& n_spins) const
UNIV_NOTHROW
{
ut_ad(n_spins <= max_spins);
/* Spin waiting for the lock word to become zero. Note
that we do not have to assume that the read access to
the lock word is atomic, as the actual locking is always
committed with atomic test-and-set. In reality, however,
all processors probably have an atomic read of a memory word. */
do {
if (!is_locked()) {
return(true);
}
ut_delay(ut_rnd_interval(0, max_delay));
++n_spins;
} while (n_spins < max_spins);
return(false);
}
/** Spin while trying to acquire the mutex
@param[in] max_spins max number of spins
@param[in] max_delay max delay per spin
@param[in] filename from where called
@param[in] line within filename */
void spin_and_try_lock(
uint32_t max_spins,
uint32_t max_delay,
const char* filename,
uint32_t line)
UNIV_NOTHROW
{
uint32_t n_spins = 0;
uint32_t n_waits = 0;
const uint32_t step = max_spins;
os_rmb;
for (;;) {
/* If the lock was free then try and acquire it. */
if (is_free(max_spins, max_delay, n_spins)) {
if (try_lock()) {
break;
} else {
continue;
}
} else {
max_spins = n_spins + step;
}
++n_waits;
os_thread_yield();
/* The 4 below is a heuristic that has existed for a
very long time now. It is unclear if changing this
value will make a difference.
NOTE: There is a delay that happens before the retry,
finding a free slot in the sync arary and the yield
above. Otherwise we could have simply done the extra
spin above. */
if (wait(filename, line, 4)) {
n_spins += 4;
break;
}
}
/* Waits and yields will be the same number in our
mutex design */
m_policy.add(n_spins, n_waits);
}
/** @return the value of the m_waiters flag */
lock_word_t waiters() UNIV_NOTHROW
{
return(m_waiters);
}
/** Note that there are threads waiting on the mutex */
void set_waiters() UNIV_NOTHROW
{
m_waiters = 1;
os_wmb;
}
/** Note that there are no threads waiting on the mutex */
void clear_waiters() UNIV_NOTHROW
{
m_waiters = 0;
os_wmb;
}
/** Try and acquire the lock using TestAndSet.
@return true if lock succeeded */
bool tas_lock() UNIV_NOTHROW
{
return(TAS(&m_lock_word, MUTEX_STATE_LOCKED)
== MUTEX_STATE_UNLOCKED);
}
/** In theory __sync_lock_release should be used to release the lock.
Unfortunately, it does not work properly alone. The workaround is
that more conservative __sync_lock_test_and_set is used instead. */
void tas_unlock() UNIV_NOTHROW
{
TAS(&m_lock_word, MUTEX_STATE_UNLOCKED);
}
/** Wakeup any waiting thread(s). */
void signal() UNIV_NOTHROW;
private: private:
/** Disable copying */ /** Disable copying */
TTASEventMutex(const TTASEventMutex&); TTASEventMutex(const TTASEventMutex&);
...@@ -907,11 +558,7 @@ struct TTASEventMutex { ...@@ -907,11 +558,7 @@ struct TTASEventMutex {
/** lock_word is the target of the atomic test-and-set instruction /** lock_word is the target of the atomic test-and-set instruction
when atomic operations are enabled. */ when atomic operations are enabled. */
lock_word_t m_lock_word; int32 m_lock_word;
/** Set to 0 or 1. 1 if there are (or may be) threads waiting
in the global wait array for this mutex to be released. */
lock_word_t m_waiters;
/** Used by sync0arr.cc for the wait queue */ /** Used by sync0arr.cc for the wait queue */
os_event_t m_event; os_event_t m_event;
...@@ -1031,7 +678,7 @@ struct PolicyMutex ...@@ -1031,7 +678,7 @@ struct PolicyMutex
/** @return true if the thread owns the mutex. */ /** @return true if the thread owns the mutex. */
bool is_owned() const UNIV_NOTHROW bool is_owned() const UNIV_NOTHROW
{ {
return(m_impl.is_owned()); return(policy().is_owned());
} }
#endif /* UNIV_DEBUG */ #endif /* UNIV_DEBUG */
...@@ -1052,6 +699,7 @@ struct PolicyMutex ...@@ -1052,6 +699,7 @@ struct PolicyMutex
#endif /* UNIV_PFS_MUTEX */ #endif /* UNIV_PFS_MUTEX */
m_impl.init(id, filename, line); m_impl.init(id, filename, line);
policy().init(m_impl, id, filename, line);
} }
/** Free resources (if any) */ /** Free resources (if any) */
...@@ -1061,6 +709,7 @@ struct PolicyMutex ...@@ -1061,6 +709,7 @@ struct PolicyMutex
pfs_del(); pfs_del();
#endif /* UNIV_PFS_MUTEX */ #endif /* UNIV_PFS_MUTEX */
m_impl.destroy(); m_impl.destroy();
policy().destroy();
} }
/** Required for os_event_t */ /** Required for os_event_t */
......
...@@ -28,6 +28,7 @@ Created 9/5/1995 Heikki Tuuri ...@@ -28,6 +28,7 @@ Created 9/5/1995 Heikki Tuuri
#include <vector> #include <vector>
#include <iostream> #include <iostream>
#include <my_atomic.h>
#include "ut0new.h" #include "ut0new.h"
#include "ut0counter.h" #include "ut0counter.h"
......
...@@ -97,8 +97,6 @@ typedef BlockSyncArrayMutex ib_bpmutex_t; ...@@ -97,8 +97,6 @@ typedef BlockSyncArrayMutex ib_bpmutex_t;
#error "ib_mutex_t type is unknown" #error "ib_mutex_t type is unknown"
#endif /* MUTEX_FUTEX */ #endif /* MUTEX_FUTEX */
#include "ut0mutex.ic"
extern ulong srv_spin_wait_delay; extern ulong srv_spin_wait_delay;
extern ulong srv_n_spin_wait_rounds; extern ulong srv_n_spin_wait_rounds;
......
/*****************************************************************************
Copyright (c) 2013, 2015, Oracle and/or its affiliates. All Rights Reserved.
Portions of this file contain modifications contributed and copyrighted by
Google, Inc. Those modifications are gratefully acknowledged and are described
briefly in the InnoDB documentation. The contributions by Google are
incorporated with their permission, and subject to the conditions contained in
the file COPYING.Google.
This program is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free Software
Foundation; version 2 of the License.
This program is distributed in the hope that it will be useful, but WITHOUT
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
You should have received a copy of the GNU General Public License along with
this program; if not, write to the Free Software Foundation, Inc.,
51 Franklin Street, Suite 500, Boston, MA 02110-1335 USA
*****************************************************************************/
/**************************************************//**
@file include/ut0mutex.ic
Mutex implementation include file
Created 2012/08/21 Sunny Bains
*******************************************************/
#include "sync0arr.h"
#include "sync0debug.h"
/**
Wait in the sync array.
@return true if the mutex acquisition was successful. */
template <template <typename> class Policy>
bool
TTASEventMutex<Policy>::wait(
const char* filename,
uint32_t line,
uint32_t spin)
UNIV_NOTHROW
{
sync_cell_t* cell;
sync_array_t* sync_arr;
sync_arr = sync_array_get_and_reserve_cell(
this,
(m_policy.get_id() == LATCH_ID_BUF_BLOCK_MUTEX
|| m_policy.get_id() == LATCH_ID_BUF_POOL_ZIP)
? SYNC_BUF_BLOCK
: SYNC_MUTEX,
filename, line, &cell);
/* The memory order of the array reservation and
the change in the waiters field is important: when
we suspend a thread, we first reserve the cell and
then set waiters field to 1. When threads are released
in mutex_exit, the waiters field is first set to zero
and then the event is set to the signaled state. */
set_waiters();
/* Try to reserve still a few times. */
for (uint32_t i = 0; i < spin; ++i) {
if (try_lock()) {
sync_array_free_cell(sync_arr, cell);
/* Note that in this case we leave
the waiters field set to 1. We cannot
reset it to zero, as we do not know if
there are other waiters. */
return(true);
}
}
/* Now we know that there has been some thread
holding the mutex after the change in the wait
array and the waiters field was made. Now there
is no risk of infinite wait on the event. */
sync_array_wait_event(sync_arr, cell);
return(false);
}
/** Wakeup any waiting thread(s). */
template <template <typename> class Policy>
void
TTASEventMutex<Policy>::signal() UNIV_NOTHROW
{
clear_waiters();
/* The memory order of resetting the waiters field and
signaling the object is important. See LEMMA 1 above. */
os_event_set(m_event);
sync_array_object_signalled();
}
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment