1 /* 2 * Copyright (C) 2014 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #include <gtest/gtest.h> 18 // Fool stdatomic.h into not using <atomic>. 19 #undef _USING_LIBCXX 20 #include <stdatomic.h> 21 #include <pthread.h> 22 #include <stdint.h> 23 24 TEST(stdatomic, LOCK_FREE) { 25 ASSERT_TRUE(ATOMIC_BOOL_LOCK_FREE); 26 ASSERT_TRUE(ATOMIC_CHAR16_T_LOCK_FREE); 27 ASSERT_TRUE(ATOMIC_CHAR32_T_LOCK_FREE); 28 ASSERT_TRUE(ATOMIC_CHAR_LOCK_FREE); 29 ASSERT_TRUE(ATOMIC_INT_LOCK_FREE); 30 ASSERT_TRUE(ATOMIC_LLONG_LOCK_FREE); 31 ASSERT_TRUE(ATOMIC_LONG_LOCK_FREE); 32 ASSERT_TRUE(ATOMIC_POINTER_LOCK_FREE); 33 ASSERT_TRUE(ATOMIC_SHORT_LOCK_FREE); 34 ASSERT_TRUE(ATOMIC_WCHAR_T_LOCK_FREE); 35 } 36 37 TEST(stdatomic, init) { 38 atomic_int v = ATOMIC_VAR_INIT(123); 39 ASSERT_EQ(123, atomic_load(&v)); 40 41 atomic_init(&v, 456); 42 ASSERT_EQ(456, atomic_load(&v)); 43 44 atomic_flag f = ATOMIC_FLAG_INIT; 45 ASSERT_FALSE(atomic_flag_test_and_set(&f)); 46 } 47 48 TEST(stdatomic, atomic_thread_fence) { 49 atomic_thread_fence(memory_order_relaxed); 50 atomic_thread_fence(memory_order_consume); 51 atomic_thread_fence(memory_order_acquire); 52 atomic_thread_fence(memory_order_release); 53 atomic_thread_fence(memory_order_acq_rel); 54 atomic_thread_fence(memory_order_seq_cst); 55 } 56 57 TEST(stdatomic, atomic_signal_fence) { 58 atomic_signal_fence(memory_order_relaxed); 59 atomic_signal_fence(memory_order_consume); 60 atomic_signal_fence(memory_order_acquire); 61 atomic_signal_fence(memory_order_release); 62 atomic_signal_fence(memory_order_acq_rel); 63 atomic_signal_fence(memory_order_seq_cst); 64 } 65 66 TEST(stdatomic, atomic_is_lock_free) { 67 atomic_char small; 68 ASSERT_TRUE(atomic_is_lock_free(&small)); 69 atomic_intmax_t big; 70 // atomic_intmax_t(size = 64) is not lock free on mips32. 71 #if defined(__mips__) && !defined(__LP64__) 72 ASSERT_FALSE(atomic_is_lock_free(&big)); 73 #else 74 ASSERT_TRUE(atomic_is_lock_free(&big)); 75 #endif 76 } 77 78 TEST(stdatomic, atomic_flag) { 79 atomic_flag f = ATOMIC_FLAG_INIT; 80 ASSERT_FALSE(atomic_flag_test_and_set(&f)); 81 ASSERT_TRUE(atomic_flag_test_and_set(&f)); 82 83 atomic_flag_clear(&f); 84 85 ASSERT_FALSE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed)); 86 ASSERT_TRUE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed)); 87 88 atomic_flag_clear_explicit(&f, memory_order_relaxed); 89 ASSERT_FALSE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed)); 90 } 91 92 TEST(stdatomic, atomic_store) { 93 atomic_int i; 94 atomic_store(&i, 123); 95 ASSERT_EQ(123, atomic_load(&i)); 96 atomic_store_explicit(&i, 123, memory_order_relaxed); 97 ASSERT_EQ(123, atomic_load_explicit(&i, memory_order_relaxed)); 98 } 99 100 TEST(stdatomic, atomic_exchange) { 101 atomic_int i; 102 atomic_store(&i, 123); 103 ASSERT_EQ(123, atomic_exchange(&i, 456)); 104 ASSERT_EQ(456, atomic_exchange_explicit(&i, 123, memory_order_relaxed)); 105 } 106 107 TEST(stdatomic, atomic_compare_exchange) { 108 atomic_int i; 109 int expected; 110 111 atomic_store(&i, 123); 112 expected = 123; 113 ASSERT_TRUE(atomic_compare_exchange_strong(&i, &expected, 456)); 114 ASSERT_FALSE(atomic_compare_exchange_strong(&i, &expected, 456)); 115 ASSERT_EQ(456, expected); 116 117 atomic_store(&i, 123); 118 expected = 123; 119 ASSERT_TRUE(atomic_compare_exchange_strong_explicit(&i, &expected, 456, memory_order_relaxed, 120 memory_order_relaxed)); 121 ASSERT_FALSE(atomic_compare_exchange_strong_explicit(&i, &expected, 456, memory_order_relaxed, 122 memory_order_relaxed)); 123 ASSERT_EQ(456, expected); 124 125 atomic_store(&i, 123); 126 expected = 123; 127 int iter_count = 0; 128 do { 129 ++iter_count; 130 ASSERT_LT(iter_count, 100); // Arbitrary limit on spurious compare_exchange failures. 131 ASSERT_EQ(expected, 123); 132 } while(!atomic_compare_exchange_weak(&i, &expected, 456)); 133 ASSERT_FALSE(atomic_compare_exchange_weak(&i, &expected, 456)); 134 ASSERT_EQ(456, expected); 135 136 atomic_store(&i, 123); 137 expected = 123; 138 iter_count = 0; 139 do { 140 ++iter_count; 141 ASSERT_LT(iter_count, 100); 142 ASSERT_EQ(expected, 123); 143 } while(!atomic_compare_exchange_weak_explicit(&i, &expected, 456, memory_order_relaxed, 144 memory_order_relaxed)); 145 ASSERT_FALSE(atomic_compare_exchange_weak_explicit(&i, &expected, 456, memory_order_relaxed, 146 memory_order_relaxed)); 147 ASSERT_EQ(456, expected); 148 } 149 150 TEST(stdatomic, atomic_fetch_add) { 151 atomic_int i = ATOMIC_VAR_INIT(123); 152 ASSERT_EQ(123, atomic_fetch_add(&i, 1)); 153 ASSERT_EQ(124, atomic_fetch_add_explicit(&i, 1, memory_order_relaxed)); 154 ASSERT_EQ(125, atomic_load(&i)); 155 } 156 157 TEST(stdatomic, atomic_fetch_sub) { 158 atomic_int i = ATOMIC_VAR_INIT(123); 159 ASSERT_EQ(123, atomic_fetch_sub(&i, 1)); 160 ASSERT_EQ(122, atomic_fetch_sub_explicit(&i, 1, memory_order_relaxed)); 161 ASSERT_EQ(121, atomic_load(&i)); 162 } 163 164 TEST(stdatomic, atomic_fetch_or) { 165 atomic_int i = ATOMIC_VAR_INIT(0x100); 166 ASSERT_EQ(0x100, atomic_fetch_or(&i, 0x020)); 167 ASSERT_EQ(0x120, atomic_fetch_or_explicit(&i, 0x003, memory_order_relaxed)); 168 ASSERT_EQ(0x123, atomic_load(&i)); 169 } 170 171 TEST(stdatomic, atomic_fetch_xor) { 172 atomic_int i = ATOMIC_VAR_INIT(0x100); 173 ASSERT_EQ(0x100, atomic_fetch_xor(&i, 0x120)); 174 ASSERT_EQ(0x020, atomic_fetch_xor_explicit(&i, 0x103, memory_order_relaxed)); 175 ASSERT_EQ(0x123, atomic_load(&i)); 176 } 177 178 TEST(stdatomic, atomic_fetch_and) { 179 atomic_int i = ATOMIC_VAR_INIT(0x123); 180 ASSERT_EQ(0x123, atomic_fetch_and(&i, 0x00f)); 181 ASSERT_EQ(0x003, atomic_fetch_and_explicit(&i, 0x2, memory_order_relaxed)); 182 ASSERT_EQ(0x002, atomic_load(&i)); 183 } 184 185 // And a rudimentary test of acquire-release memory ordering: 186 187 constexpr static uint_least32_t BIG = 10000000ul; // Assumed even below. 188 189 struct three_atomics { 190 atomic_uint_least32_t x; 191 char a[123]; // Everything in different cache lines, 192 // increase chance of compiler getting alignment wrong. 193 atomic_uint_least32_t y; 194 char b[4013]; 195 atomic_uint_least32_t z; 196 }; 197 198 // Very simple acquire/release memory ordering sanity check. 199 static void* writer(void* arg) { 200 three_atomics* a = reinterpret_cast<three_atomics*>(arg); 201 for (uint_least32_t i = 0; i <= BIG; i+=2) { 202 atomic_store_explicit(&a->x, i, memory_order_relaxed); 203 atomic_store_explicit(&a->z, i, memory_order_relaxed); 204 atomic_store_explicit(&a->y, i, memory_order_release); 205 atomic_store_explicit(&a->x, i+1, memory_order_relaxed); 206 atomic_store_explicit(&a->z, i+1, memory_order_relaxed); 207 atomic_store_explicit(&a->y, i+1, memory_order_release); 208 } 209 return 0; 210 } 211 212 static void* reader(void* arg) { 213 three_atomics* a = reinterpret_cast<three_atomics*>(arg); 214 uint_least32_t xval = 0, yval = 0, zval = 0; 215 size_t repeat = 0; 216 size_t repeat_limit = 1000; 217 while (yval != BIG + 1) { 218 yval = atomic_load_explicit(&a->y, memory_order_acquire); 219 zval = atomic_load_explicit(&a->z, memory_order_relaxed); 220 xval = atomic_load_explicit(&a->x, memory_order_relaxed); 221 // If we see a given value of y, the immediately preceding 222 // stores to z and x, or later ones, should also be visible. 223 if (zval < yval) { 224 // Cant just ASSERT, since we are in a non-void function. 225 ADD_FAILURE() << "acquire-release ordering violation: " 226 << zval << " < " << yval << ", " << xval << "\n"; 227 return 0; // Only report once. 228 } 229 if (xval < yval) { 230 // Cant just ASSERT, since we are in a non-void function. 231 ADD_FAILURE() << "acquire-release ordering violation: " 232 << xval << " < " << yval << ", " << zval << "\n"; 233 return 0; // Only report once. 234 } 235 if (repeat < repeat_limit) ++repeat; 236 } 237 // The following assertion is not technically guaranteed to hold. 238 // But if it fails to hold, this test was useless, and we have a 239 // serious scheduling issue that we should probably know about. 240 EXPECT_EQ(repeat, repeat_limit); 241 return 0; 242 } 243 244 TEST(stdatomic, ordering) { 245 // Run a memory ordering sanity test. 246 void* result; 247 three_atomics a; 248 atomic_init(&a.x, 0ul); 249 atomic_init(&a.y, 0ul); 250 atomic_init(&a.z, 0ul); 251 pthread_t t1,t2; 252 ASSERT_EQ(0, pthread_create(&t1, 0, reader, &a)); 253 ASSERT_EQ(0, pthread_create(&t2, 0, writer, &a)); 254 ASSERT_EQ(0, pthread_join(t1, &result)); 255 EXPECT_EQ(0, result); 256 ASSERT_EQ(0, pthread_join(t2, &result)); 257 EXPECT_EQ(0, result); 258 EXPECT_EQ(atomic_load_explicit(&a.x, memory_order_consume), BIG + 1); 259 EXPECT_EQ(atomic_load_explicit(&a.y, memory_order_seq_cst), BIG + 1); 260 EXPECT_EQ(atomic_load(&a.z), BIG + 1); 261 } 262