mirror of
https://github.com/beefytech/Beef.git
synced 2025-06-23 01:58:00 +02:00
275 lines
9.7 KiB
C++
275 lines
9.7 KiB
C++
/* Copyright (c) 2006, Google Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are
|
|
* met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following disclaimer
|
|
* in the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Google Inc. nor the names of its
|
|
* contributors may be used to endorse or promote products derived from
|
|
* this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* ---
|
|
* Author: Sanjay Ghemawat
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include "base/logging.h"
|
|
#include "base/atomicops.h"
|
|
|
|
#define GG_ULONGLONG(x) static_cast<uint64>(x)
|
|
|
|
template <class AtomicType>
|
|
static void TestAtomicIncrement(AtomicType (*atomic_increment_func)
|
|
(volatile AtomicType*, AtomicType)) {
|
|
// For now, we just test single threaded execution
|
|
|
|
// use a guard value to make sure the atomic_increment_func doesn't go
|
|
// outside the expected address bounds. This is in particular to
|
|
// test that some future change to the asm code doesn't cause the
|
|
// 32-bit atomic_increment_func doesn't do the wrong thing on 64-bit
|
|
// machines.
|
|
struct {
|
|
AtomicType prev_word;
|
|
AtomicType count;
|
|
AtomicType next_word;
|
|
} s;
|
|
|
|
AtomicType prev_word_value, next_word_value;
|
|
memset(&prev_word_value, 0xFF, sizeof(AtomicType));
|
|
memset(&next_word_value, 0xEE, sizeof(AtomicType));
|
|
|
|
s.prev_word = prev_word_value;
|
|
s.count = 0;
|
|
s.next_word = next_word_value;
|
|
|
|
ASSERT_EQ(1, (*atomic_increment_func)(&s.count, 1));
|
|
ASSERT_EQ(1, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
|
|
ASSERT_EQ(3, (*atomic_increment_func)(&s.count, 2));
|
|
ASSERT_EQ(3, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
|
|
ASSERT_EQ(6, (*atomic_increment_func)(&s.count, 3));
|
|
ASSERT_EQ(6, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
|
|
ASSERT_EQ(3, (*atomic_increment_func)(&s.count, -3));
|
|
ASSERT_EQ(3, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
|
|
ASSERT_EQ(1, (*atomic_increment_func)(&s.count, -2));
|
|
ASSERT_EQ(1, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
|
|
ASSERT_EQ(0, (*atomic_increment_func)(&s.count, -1));
|
|
ASSERT_EQ(0, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
|
|
ASSERT_EQ(-1, (*atomic_increment_func)(&s.count, -1));
|
|
ASSERT_EQ(-1, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
|
|
ASSERT_EQ(-5, (*atomic_increment_func)(&s.count, -4));
|
|
ASSERT_EQ(-5, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
|
|
ASSERT_EQ(0, (*atomic_increment_func)(&s.count, 5));
|
|
ASSERT_EQ(0, s.count);
|
|
ASSERT_EQ(prev_word_value, s.prev_word);
|
|
ASSERT_EQ(next_word_value, s.next_word);
|
|
}
|
|
|
|
|
|
#define NUM_BITS(T) (sizeof(T) * 8)
|
|
|
|
|
|
template <class AtomicType>
|
|
static void TestCompareAndSwap(AtomicType (*compare_and_swap_func)
|
|
(volatile AtomicType*, AtomicType, AtomicType)) {
|
|
AtomicType value = 0;
|
|
AtomicType prev = (*compare_and_swap_func)(&value, 0, 1);
|
|
ASSERT_EQ(1, value);
|
|
ASSERT_EQ(0, prev);
|
|
|
|
// Use test value that has non-zero bits in both halves, more for testing
|
|
// 64-bit implementation on 32-bit platforms.
|
|
const AtomicType k_test_val = (GG_ULONGLONG(1) <<
|
|
(NUM_BITS(AtomicType) - 2)) + 11;
|
|
value = k_test_val;
|
|
prev = (*compare_and_swap_func)(&value, 0, 5);
|
|
ASSERT_EQ(k_test_val, value);
|
|
ASSERT_EQ(k_test_val, prev);
|
|
|
|
value = k_test_val;
|
|
prev = (*compare_and_swap_func)(&value, k_test_val, 5);
|
|
ASSERT_EQ(5, value);
|
|
ASSERT_EQ(k_test_val, prev);
|
|
}
|
|
|
|
|
|
template <class AtomicType>
|
|
static void TestAtomicExchange(AtomicType (*atomic_exchange_func)
|
|
(volatile AtomicType*, AtomicType)) {
|
|
AtomicType value = 0;
|
|
AtomicType new_value = (*atomic_exchange_func)(&value, 1);
|
|
ASSERT_EQ(1, value);
|
|
ASSERT_EQ(0, new_value);
|
|
|
|
// Use test value that has non-zero bits in both halves, more for testing
|
|
// 64-bit implementation on 32-bit platforms.
|
|
const AtomicType k_test_val = (GG_ULONGLONG(1) <<
|
|
(NUM_BITS(AtomicType) - 2)) + 11;
|
|
value = k_test_val;
|
|
new_value = (*atomic_exchange_func)(&value, k_test_val);
|
|
ASSERT_EQ(k_test_val, value);
|
|
ASSERT_EQ(k_test_val, new_value);
|
|
|
|
value = k_test_val;
|
|
new_value = (*atomic_exchange_func)(&value, 5);
|
|
ASSERT_EQ(5, value);
|
|
ASSERT_EQ(k_test_val, new_value);
|
|
}
|
|
|
|
|
|
template <class AtomicType>
|
|
static void TestAtomicIncrementBounds(AtomicType (*atomic_increment_func)
|
|
(volatile AtomicType*, AtomicType)) {
|
|
// Test increment at the half-width boundary of the atomic type.
|
|
// It is primarily for testing at the 32-bit boundary for 64-bit atomic type.
|
|
AtomicType test_val = GG_ULONGLONG(1) << (NUM_BITS(AtomicType) / 2);
|
|
AtomicType value = test_val - 1;
|
|
AtomicType new_value = (*atomic_increment_func)(&value, 1);
|
|
ASSERT_EQ(test_val, value);
|
|
ASSERT_EQ(value, new_value);
|
|
|
|
(*atomic_increment_func)(&value, -1);
|
|
ASSERT_EQ(test_val - 1, value);
|
|
}
|
|
|
|
// This is a simple sanity check that values are correct. Not testing
|
|
// atomicity
|
|
template <class AtomicType>
|
|
static void TestStore() {
|
|
const AtomicType kVal1 = static_cast<AtomicType>(0xa5a5a5a5a5a5a5a5LL);
|
|
const AtomicType kVal2 = static_cast<AtomicType>(-1);
|
|
|
|
AtomicType value;
|
|
|
|
base::subtle::NoBarrier_Store(&value, kVal1);
|
|
ASSERT_EQ(kVal1, value);
|
|
base::subtle::NoBarrier_Store(&value, kVal2);
|
|
ASSERT_EQ(kVal2, value);
|
|
|
|
base::subtle::Acquire_Store(&value, kVal1);
|
|
ASSERT_EQ(kVal1, value);
|
|
base::subtle::Acquire_Store(&value, kVal2);
|
|
ASSERT_EQ(kVal2, value);
|
|
|
|
base::subtle::Release_Store(&value, kVal1);
|
|
ASSERT_EQ(kVal1, value);
|
|
base::subtle::Release_Store(&value, kVal2);
|
|
ASSERT_EQ(kVal2, value);
|
|
}
|
|
|
|
// This is a simple sanity check that values are correct. Not testing
|
|
// atomicity
|
|
template <class AtomicType>
|
|
static void TestLoad() {
|
|
const AtomicType kVal1 = static_cast<AtomicType>(0xa5a5a5a5a5a5a5a5LL);
|
|
const AtomicType kVal2 = static_cast<AtomicType>(-1);
|
|
|
|
AtomicType value;
|
|
|
|
value = kVal1;
|
|
ASSERT_EQ(kVal1, base::subtle::NoBarrier_Load(&value));
|
|
value = kVal2;
|
|
ASSERT_EQ(kVal2, base::subtle::NoBarrier_Load(&value));
|
|
|
|
value = kVal1;
|
|
ASSERT_EQ(kVal1, base::subtle::Acquire_Load(&value));
|
|
value = kVal2;
|
|
ASSERT_EQ(kVal2, base::subtle::Acquire_Load(&value));
|
|
|
|
value = kVal1;
|
|
ASSERT_EQ(kVal1, base::subtle::Release_Load(&value));
|
|
value = kVal2;
|
|
ASSERT_EQ(kVal2, base::subtle::Release_Load(&value));
|
|
}
|
|
|
|
template <class AtomicType>
|
|
static void TestAtomicOps() {
|
|
TestCompareAndSwap<AtomicType>(base::subtle::NoBarrier_CompareAndSwap);
|
|
TestCompareAndSwap<AtomicType>(base::subtle::Acquire_CompareAndSwap);
|
|
TestCompareAndSwap<AtomicType>(base::subtle::Release_CompareAndSwap);
|
|
|
|
TestAtomicExchange<AtomicType>(base::subtle::NoBarrier_AtomicExchange);
|
|
TestAtomicExchange<AtomicType>(base::subtle::Acquire_AtomicExchange);
|
|
TestAtomicExchange<AtomicType>(base::subtle::Release_AtomicExchange);
|
|
|
|
TestAtomicIncrementBounds<AtomicType>(
|
|
base::subtle::NoBarrier_AtomicIncrement);
|
|
TestAtomicIncrementBounds<AtomicType>(
|
|
base::subtle::Barrier_AtomicIncrement);
|
|
|
|
TestStore<AtomicType>();
|
|
TestLoad<AtomicType>();
|
|
}
|
|
|
|
int main(int argc, char** argv) {
|
|
TestAtomicIncrement<AtomicWord>(base::subtle::NoBarrier_AtomicIncrement);
|
|
TestAtomicIncrement<AtomicWord>(base::subtle::Barrier_AtomicIncrement);
|
|
TestAtomicIncrement<Atomic32>(base::subtle::NoBarrier_AtomicIncrement);
|
|
TestAtomicIncrement<Atomic32>(base::subtle::Barrier_AtomicIncrement);
|
|
|
|
TestAtomicOps<AtomicWord>();
|
|
TestAtomicOps<Atomic32>();
|
|
|
|
// I've commented the Atomic64 tests out for now, because Atomic64
|
|
// doesn't work on x86 systems that are not compiled to support mmx
|
|
// registers. Since I want this project to be as portable as
|
|
// possible -- that is, not to assume we've compiled for mmx or even
|
|
// that the processor supports it -- and we don't actually use
|
|
// Atomic64 anywhere, I've commented it out of the test for now.
|
|
// (Luckily, if we ever do use Atomic64 by accident, we'll get told
|
|
// via a compiler error rather than some obscure runtime failure, so
|
|
// this course of action is safe.)
|
|
// If we ever *do* want to enable this, try adding -msse (or -mmmx?)
|
|
// to the CXXFLAGS in Makefile.am.
|
|
#if 0 and defined(BASE_HAS_ATOMIC64)
|
|
TestAtomicIncrement<base::subtle::Atomic64>(
|
|
base::subtle::NoBarrier_AtomicIncrement);
|
|
TestAtomicIncrement<base::subtle::Atomic64>(
|
|
base::subtle::Barrier_AtomicIncrement);
|
|
#endif
|
|
|
|
printf("PASS\n");
|
|
return 0;
|
|
}
|