| /* Copyright (c) 2006, Google Inc. | 
 |  * All rights reserved. | 
 |  *  | 
 |  * Redistribution and use in source and binary forms, with or without | 
 |  * modification, are permitted provided that the following conditions are | 
 |  * met: | 
 |  *  | 
 |  *     * Redistributions of source code must retain the above copyright | 
 |  * notice, this list of conditions and the following disclaimer. | 
 |  *     * Redistributions in binary form must reproduce the above | 
 |  * copyright notice, this list of conditions and the following disclaimer | 
 |  * in the documentation and/or other materials provided with the | 
 |  * distribution. | 
 |  *     * Neither the name of Google Inc. nor the names of its | 
 |  * contributors may be used to endorse or promote products derived from | 
 |  * this software without specific prior written permission. | 
 |  *  | 
 |  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS | 
 |  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT | 
 |  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR | 
 |  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT | 
 |  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, | 
 |  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT | 
 |  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, | 
 |  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY | 
 |  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | 
 |  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | 
 |  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 
 |  * | 
 |  * --- | 
 |  * Author: Sanjay Ghemawat | 
 |  */ | 
 |  | 
 | // | 
 | // Fast spinlocks (at least on x86, a lock/unlock pair is approximately | 
 | // half the cost of a Mutex because the unlock just does a store instead | 
 | // of a compare-and-swap which is expensive). | 
 |  | 
 | // SpinLock is async signal safe. | 
 | // If used within a signal handler, all lock holders | 
 | // should block the signal even outside the signal handler. | 
 |  | 
 | #ifndef BASE_SPINLOCK_H_ | 
 | #define BASE_SPINLOCK_H_ | 
 |  | 
 | #include <config.h> | 
 | #include "base/atomicops.h" | 
 | #include "base/basictypes.h" | 
 | #include "base/dynamic_annotations.h" | 
 | #include "base/thread_annotations.h" | 
 |  | 
 | class LOCKABLE SpinLock { | 
 |  public: | 
 |   SpinLock() : lockword_(kSpinLockFree) { } | 
 |  | 
 |   // Special constructor for use with static SpinLock objects.  E.g., | 
 |   // | 
 |   //    static SpinLock lock(base::LINKER_INITIALIZED); | 
 |   // | 
 |   // When intialized using this constructor, we depend on the fact | 
 |   // that the linker has already initialized the memory appropriately. | 
 |   // A SpinLock constructed like this can be freely used from global | 
 |   // initializers without worrying about the order in which global | 
 |   // initializers run. | 
 |   explicit SpinLock(base::LinkerInitialized /*x*/) { | 
 |     // Does nothing; lockword_ is already initialized | 
 |   } | 
 |  | 
 |   // Acquire this SpinLock. | 
 |   // TODO(csilvers): uncomment the annotation when we figure out how to | 
 |   //                 support this macro with 0 args (see thread_annotations.h) | 
 |   inline void Lock() /*EXCLUSIVE_LOCK_FUNCTION()*/ { | 
 |     if (base::subtle::Acquire_CompareAndSwap(&lockword_, kSpinLockFree, | 
 |                                              kSpinLockHeld) != kSpinLockFree) { | 
 |       SlowLock(); | 
 |     } | 
 |     ANNOTATE_RWLOCK_ACQUIRED(this, 1); | 
 |   } | 
 |  | 
 |   // Try to acquire this SpinLock without blocking and return true if the | 
 |   // acquisition was successful.  If the lock was not acquired, false is | 
 |   // returned.  If this SpinLock is free at the time of the call, TryLock | 
 |   // will return true with high probability. | 
 |   inline bool TryLock() EXCLUSIVE_TRYLOCK_FUNCTION(true) { | 
 |     bool res = | 
 |         (base::subtle::Acquire_CompareAndSwap(&lockword_, kSpinLockFree, | 
 |                                               kSpinLockHeld) == kSpinLockFree); | 
 |     if (res) { | 
 |       ANNOTATE_RWLOCK_ACQUIRED(this, 1); | 
 |     } | 
 |     return res; | 
 |   } | 
 |  | 
 |   // Release this SpinLock, which must be held by the calling thread. | 
 |   // TODO(csilvers): uncomment the annotation when we figure out how to | 
 |   //                 support this macro with 0 args (see thread_annotations.h) | 
 |   inline void Unlock() /*UNLOCK_FUNCTION()*/ { | 
 |     uint64 wait_cycles = | 
 |         static_cast<uint64>(base::subtle::NoBarrier_Load(&lockword_)); | 
 |     ANNOTATE_RWLOCK_RELEASED(this, 1); | 
 |     base::subtle::Release_Store(&lockword_, kSpinLockFree); | 
 |     if (wait_cycles != kSpinLockHeld) { | 
 |       // Collect contentionz profile info, and speed the wakeup of any waiter. | 
 |       // The wait_cycles value indicates how long this thread spent waiting | 
 |       // for the lock. | 
 |       SlowUnlock(wait_cycles); | 
 |     } | 
 |   } | 
 |  | 
 |   // Determine if the lock is held.  When the lock is held by the invoking | 
 |   // thread, true will always be returned. Intended to be used as | 
 |   // CHECK(lock.IsHeld()). | 
 |   inline bool IsHeld() const { | 
 |     return base::subtle::NoBarrier_Load(&lockword_) != kSpinLockFree; | 
 |   } | 
 |  | 
 |   static const base::LinkerInitialized LINKER_INITIALIZED;  // backwards compat | 
 |  private: | 
 |   enum { kSpinLockFree = 0 }; | 
 |   enum { kSpinLockHeld = 1 }; | 
 |   enum { kSpinLockSleeper = 2 }; | 
 |  | 
 |   volatile Atomic32 lockword_; | 
 |  | 
 |   void SlowLock(); | 
 |   void SlowUnlock(uint64 wait_cycles); | 
 |   Atomic32 SpinLoop(int64 initial_wait_timestamp, Atomic32* wait_cycles); | 
 |   inline int32 CalculateWaitCycles(int64 wait_start_time); | 
 |  | 
 |   DISALLOW_COPY_AND_ASSIGN(SpinLock); | 
 | }; | 
 |  | 
 | // Corresponding locker object that arranges to acquire a spinlock for | 
 | // the duration of a C++ scope. | 
 | class SCOPED_LOCKABLE SpinLockHolder { | 
 |  private: | 
 |   SpinLock* lock_; | 
 |  public: | 
 |   inline explicit SpinLockHolder(SpinLock* l) EXCLUSIVE_LOCK_FUNCTION(l) | 
 |       : lock_(l) { | 
 |     l->Lock(); | 
 |   } | 
 |   // TODO(csilvers): uncomment the annotation when we figure out how to | 
 |   //                 support this macro with 0 args (see thread_annotations.h) | 
 |   inline ~SpinLockHolder() /*UNLOCK_FUNCTION()*/ { lock_->Unlock(); } | 
 | }; | 
 | // Catch bug where variable name is omitted, e.g. SpinLockHolder (&lock); | 
 | #define SpinLockHolder(x) COMPILE_ASSERT(0, spin_lock_decl_missing_var_name) | 
 |  | 
 |  | 
 | #endif  // BASE_SPINLOCK_H_ |