Home | History | Annotate | Download | only in include
      1 /*-
      2  * Copyright (c) 2011 Ed Schouten <ed (at) FreeBSD.org>
      3  *                    David Chisnall <theraven (at) FreeBSD.org>
      4  * All rights reserved.
      5  *
      6  * Redistribution and use in source and binary forms, with or without
      7  * modification, are permitted provided that the following conditions
      8  * are met:
      9  * 1. Redistributions of source code must retain the above copyright
     10  *    notice, this list of conditions and the following disclaimer.
     11  * 2. Redistributions in binary form must reproduce the above copyright
     12  *    notice, this list of conditions and the following disclaimer in the
     13  *    documentation and/or other materials provided with the distribution.
     14  *
     15  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
     16  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     17  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     18  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
     19  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     20  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     21  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     22  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     23  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     24  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     25  * SUCH DAMAGE.
     26  *
     27  * $FreeBSD$
     28  */
     29 
     30 #ifndef _STDATOMIC_H_
     31 #define	_STDATOMIC_H_
     32 
     33 #include <sys/cdefs.h>
     34 
     35 
     36 #if defined(__cplusplus) && defined(_USING_LIBCXX)
     37 # ifdef __clang__
     38 #  if __has_feature(cxx_atomic)
     39 #   define _STDATOMIC_HAVE_ATOMIC
     40 #  endif
     41 # else /* gcc */
     42 #  if __GNUC_PREREQ(4, 7)
     43 #   define _STDATOMIC_HAVE_ATOMIC
     44 #  endif
     45 # endif
     46 #endif
     47 
     48 #ifdef _STDATOMIC_HAVE_ATOMIC
     49 
     50 /* We have a usable C++ <atomic>; use it instead.  */
     51 
     52 #include <atomic>
     53 
     54 #undef _Atomic
     55         /* Also defined by <atomic> for gcc.  But not used in macros. */
     56         /* Also a clang intrinsic.                                    */
     57         /* Should not be used by client code before this file is      */
     58         /* included.  The definitions in <atomic> themselves see      */
     59         /* the old definition, as they should.                        */
     60         /* Client code sees the following definition.                 */
     61 
     62 #define _Atomic(t) std::atomic<t>
     63 
     64 using std::atomic_is_lock_free;
     65 using std::atomic_init;
     66 using std::atomic_store;
     67 using std::atomic_store_explicit;
     68 using std::atomic_load;
     69 using std::atomic_load_explicit;
     70 using std::atomic_exchange;
     71 using std::atomic_exchange_explicit;
     72 using std::atomic_compare_exchange_strong;
     73 using std::atomic_compare_exchange_strong_explicit;
     74 using std::atomic_compare_exchange_weak;
     75 using std::atomic_compare_exchange_weak_explicit;
     76 using std::atomic_fetch_add;
     77 using std::atomic_fetch_add_explicit;
     78 using std::atomic_fetch_sub;
     79 using std::atomic_fetch_sub_explicit;
     80 using std::atomic_fetch_or;
     81 using std::atomic_fetch_or_explicit;
     82 using std::atomic_fetch_xor;
     83 using std::atomic_fetch_xor_explicit;
     84 using std::atomic_fetch_and;
     85 using std::atomic_fetch_and_explicit;
     86 using std::atomic_thread_fence;
     87 using std::atomic_signal_fence;
     88 
     89 using std::memory_order;
     90 using std::memory_order_relaxed;
     91 using std::memory_order_consume;
     92 using std::memory_order_release;
     93 using std::memory_order_acq_rel;
     94 using std::memory_order_seq_cst;
     95 
     96 using std::atomic_bool;
     97 using std::atomic_char;
     98 using std::atomic_schar;
     99 using std::atomic_uchar;
    100 using std::atomic_short;
    101 using std::atomic_ushort;
    102 using std::atomic_int;
    103 using std::atomic_uint;
    104 using std::atomic_long;
    105 using std::atomic_ulong;
    106 using std::atomic_llong;
    107 using std::atomic_ullong;
    108 using std::atomic_char16_t;
    109 using std::atomic_char32_t;
    110 using std::atomic_wchar_t;
    111 using std::atomic_int_least8_t;
    112 using std::atomic_uint_least8_t;
    113 using std::atomic_int_least16_t;
    114 using std::atomic_uint_least16_t;
    115 using std::atomic_int_least32_t;
    116 using std::atomic_uint_least32_t;
    117 using std::atomic_int_least64_t;
    118 using std::atomic_uint_least64_t;
    119 using std::atomic_int_fast8_t;
    120 using std::atomic_uint_fast8_t;
    121 using std::atomic_int_fast16_t;
    122 using std::atomic_uint_fast16_t;
    123 using std::atomic_int_fast32_t;
    124 using std::atomic_uint_fast32_t;
    125 using std::atomic_int_fast64_t;
    126 using std::atomic_uint_fast64_t;
    127 using std::atomic_intptr_t;
    128 using std::atomic_uintptr_t;
    129 using std::atomic_size_t;
    130 using std::atomic_ptrdiff_t;
    131 using std::atomic_intmax_t;
    132 using std::atomic_uintmax_t;
    133 
    134 #else /* <atomic> unavailable, possibly because this is C, not C++ */
    135 
    136 #include <sys/types.h>
    137 #include <stdbool.h>
    138 
    139 /*
    140  * C: Do it ourselves.
    141  * Note that the runtime representation defined here should be compatible
    142  * with the C++ one, i.e. an _Atomic(T) needs to contain the same
    143  * bits as a T.
    144  */
    145 
    146 #include <stddef.h>  /* For ptrdiff_t.                          */
    147 #include <stdint.h>  /* TODO: Should pollute namespace less.    */
    148 #if __STDC_VERSION__ >= 201112L
    149 # include <uchar.h>  /* For char16_t and char32_t.              */
    150 #endif
    151 
    152 #ifdef __clang__
    153 # if __has_extension(c_atomic) || __has_extension(cxx_atomic)
    154 #  define       __CLANG_ATOMICS
    155 # else
    156 #  error "stdatomic.h does not support your compiler"
    157 # endif
    158 # if __has_builtin(__sync_swap)
    159 #  define __HAS_BUILTIN_SYNC_SWAP
    160 # endif
    161 #else
    162 # if __GNUC_PREREQ(4, 7)
    163 #  define	__GNUC_ATOMICS
    164 # else
    165 #  define	__SYNC_ATOMICS
    166 #  ifdef __cplusplus
    167 #   define       __ATOMICS_AVOID_DOT_INIT
    168 #  endif
    169 # endif
    170 #endif
    171 
    172 /*
    173  * 7.17.1 Atomic lock-free macros.
    174  */
    175 
    176 #ifdef __GCC_ATOMIC_BOOL_LOCK_FREE
    177 #define	ATOMIC_BOOL_LOCK_FREE		__GCC_ATOMIC_BOOL_LOCK_FREE
    178 #elif defined(__SYNC_ATOMICS)
    179 #define	ATOMIC_BOOL_LOCK_FREE           2 /* For all modern platforms */
    180 #endif
    181 #ifdef __GCC_ATOMIC_CHAR_LOCK_FREE
    182 #define	ATOMIC_CHAR_LOCK_FREE		__GCC_ATOMIC_CHAR_LOCK_FREE
    183 #elif defined(__SYNC_ATOMICS)
    184 #define	ATOMIC_CHAR_LOCK_FREE           2
    185 #endif
    186 #ifdef __GCC_ATOMIC_CHAR16_T_LOCK_FREE
    187 #define	ATOMIC_CHAR16_T_LOCK_FREE	__GCC_ATOMIC_CHAR16_T_LOCK_FREE
    188 #elif defined(__SYNC_ATOMICS)
    189 #define	ATOMIC_CHAR16_T_LOCK_FREE       2
    190 #endif
    191 #ifdef __GCC_ATOMIC_CHAR32_T_LOCK_FREE
    192 #define	ATOMIC_CHAR32_T_LOCK_FREE	__GCC_ATOMIC_CHAR32_T_LOCK_FREE
    193 #elif defined(__SYNC_ATOMICS)
    194 #define	ATOMIC_CHAR32_T_LOCK_FREE       2
    195 #endif
    196 #ifdef __GCC_ATOMIC_WCHAR_T_LOCK_FREE
    197 #define	ATOMIC_WCHAR_T_LOCK_FREE	__GCC_ATOMIC_WCHAR_T_LOCK_FREE
    198 #elif defined(__SYNC_ATOMICS)
    199 #define	ATOMIC_WCHAR_T_LOCK_FREE        2
    200 #endif
    201 #ifdef __GCC_ATOMIC_SHORT_LOCK_FREE
    202 #define	ATOMIC_SHORT_LOCK_FREE		__GCC_ATOMIC_SHORT_LOCK_FREE
    203 #elif defined(__SYNC_ATOMICS)
    204 #define	ATOMIC_SHORT_LOCK_FREE          2
    205 #endif
    206 #ifdef __GCC_ATOMIC_INT_LOCK_FREE
    207 #define	ATOMIC_INT_LOCK_FREE		__GCC_ATOMIC_INT_LOCK_FREE
    208 #elif defined(__SYNC_ATOMICS)
    209 #define	ATOMIC_INT_LOCK_FREE            2
    210 #endif
    211 #ifdef __GCC_ATOMIC_LONG_LOCK_FREE
    212 #define	ATOMIC_LONG_LOCK_FREE		__GCC_ATOMIC_LONG_LOCK_FREE
    213 #elif defined(__SYNC_ATOMICS)
    214 #define	ATOMIC_LONG_LOCK_FREE           2
    215 #endif
    216 #ifdef __GCC_ATOMIC_LLONG_LOCK_FREE
    217 #define	ATOMIC_LLONG_LOCK_FREE		__GCC_ATOMIC_LLONG_LOCK_FREE
    218 #elif defined(__SYNC_ATOMICS)
    219 #define	ATOMIC_LLONG_LOCK_FREE          1 /* maybe */
    220 #endif
    221 #ifdef __GCC_ATOMIC_POINTER_LOCK_FREE
    222 #define	ATOMIC_POINTER_LOCK_FREE	__GCC_ATOMIC_POINTER_LOCK_FREE
    223 #elif defined(__SYNC_ATOMICS)
    224 #define	ATOMIC_POINTER_LOCK_FREE        2
    225 #endif
    226 
    227 /*
    228  * 7.17.2 Initialization.
    229  */
    230 
    231 #if defined(__CLANG_ATOMICS)
    232 #define	ATOMIC_VAR_INIT(value)		(value)
    233 #define	atomic_init(obj, value)		__c11_atomic_init(obj, value)
    234 #else
    235 #ifdef __ATOMICS_AVOID_DOT_INIT
    236 #define	ATOMIC_VAR_INIT(value)		{ value }
    237 #else
    238 #define	ATOMIC_VAR_INIT(value)		{ .__val = (value) }
    239 #endif
    240 #define	atomic_init(obj, value)		((void)((obj)->__val = (value)))
    241 #endif
    242 
    243 /*
    244  * Clang and recent GCC both provide predefined macros for the memory
    245  * orderings.  If we are using a compiler that doesn't define them, use the
    246  * clang values - these will be ignored in the fallback path.
    247  */
    248 
    249 #ifndef __ATOMIC_RELAXED
    250 #define __ATOMIC_RELAXED		0
    251 #endif
    252 #ifndef __ATOMIC_CONSUME
    253 #define __ATOMIC_CONSUME		1
    254 #endif
    255 #ifndef __ATOMIC_ACQUIRE
    256 #define __ATOMIC_ACQUIRE		2
    257 #endif
    258 #ifndef __ATOMIC_RELEASE
    259 #define __ATOMIC_RELEASE		3
    260 #endif
    261 #ifndef __ATOMIC_ACQ_REL
    262 #define __ATOMIC_ACQ_REL		4
    263 #endif
    264 #ifndef __ATOMIC_SEQ_CST
    265 #define __ATOMIC_SEQ_CST		5
    266 #endif
    267 
    268 /*
    269  * 7.17.3 Order and consistency.
    270  *
    271  * The memory_order_* constants that denote the barrier behaviour of the
    272  * atomic operations.
    273  * The enum values must be identical to those used by the
    274  * C++ <atomic> header.
    275  */
    276 
    277 typedef enum {
    278 	memory_order_relaxed = __ATOMIC_RELAXED,
    279 	memory_order_consume = __ATOMIC_CONSUME,
    280 	memory_order_acquire = __ATOMIC_ACQUIRE,
    281 	memory_order_release = __ATOMIC_RELEASE,
    282 	memory_order_acq_rel = __ATOMIC_ACQ_REL,
    283 	memory_order_seq_cst = __ATOMIC_SEQ_CST
    284 } memory_order;
    285 
    286 /*
    287  * 7.17.4 Fences.
    288  */
    289 
    290 static __inline void
    291 atomic_thread_fence(memory_order __order __attribute__((unused)))
    292 {
    293 
    294 #ifdef __CLANG_ATOMICS
    295 	__c11_atomic_thread_fence(__order);
    296 #elif defined(__GNUC_ATOMICS)
    297 	__atomic_thread_fence(__order);
    298 #else
    299 	__sync_synchronize();
    300 #endif
    301 }
    302 
    303 static __inline void
    304 atomic_signal_fence(memory_order __order __attribute__((unused)))
    305 {
    306 
    307 #ifdef __CLANG_ATOMICS
    308 	__c11_atomic_signal_fence(__order);
    309 #elif defined(__GNUC_ATOMICS)
    310 	__atomic_signal_fence(__order);
    311 #else
    312 	__asm volatile ("" ::: "memory");
    313 #endif
    314 }
    315 
    316 /*
    317  * 7.17.5 Lock-free property.
    318  */
    319 
    320 #if defined(_KERNEL)
    321 /* Atomics in kernelspace are always lock-free. */
    322 #define	atomic_is_lock_free(obj) \
    323 	((void)(obj), (_Bool)1)
    324 #elif defined(__CLANG_ATOMICS)
    325 #define	atomic_is_lock_free(obj) \
    326 	__c11_atomic_is_lock_free(sizeof(*(obj)))
    327 #elif defined(__GNUC_ATOMICS)
    328 #define	atomic_is_lock_free(obj) \
    329 	__atomic_is_lock_free(sizeof((obj)->__val), &(obj)->__val)
    330 #else
    331 #define	atomic_is_lock_free(obj) \
    332 	((void)(obj), sizeof((obj)->__val) <= sizeof(void *))
    333 #endif
    334 
    335 /*
    336  * 7.17.6 Atomic integer types.
    337  */
    338 
    339 #ifndef __CLANG_ATOMICS
    340 /*
    341  * No native support for _Atomic(). Place object in structure to prevent
    342  * most forms of direct non-atomic access.
    343  */
    344 #define _Atomic(T)              struct { T volatile __val; }
    345 #endif
    346 
    347 typedef _Atomic(bool)			atomic_bool;
    348 typedef _Atomic(char)			atomic_char;
    349 typedef _Atomic(signed char)		atomic_schar;
    350 typedef _Atomic(unsigned char)		atomic_uchar;
    351 typedef _Atomic(short)			atomic_short;
    352 typedef _Atomic(unsigned short)		atomic_ushort;
    353 typedef _Atomic(int)			atomic_int;
    354 typedef _Atomic(unsigned int)		atomic_uint;
    355 typedef _Atomic(long)			atomic_long;
    356 typedef _Atomic(unsigned long)		atomic_ulong;
    357 typedef _Atomic(long long)		atomic_llong;
    358 typedef _Atomic(unsigned long long)	atomic_ullong;
    359 #if __STDC_VERSION__ >= 201112L || __cplusplus >= 201103L
    360   typedef _Atomic(char16_t)		atomic_char16_t;
    361   typedef _Atomic(char32_t)		atomic_char32_t;
    362 #endif
    363 typedef _Atomic(wchar_t)		atomic_wchar_t;
    364 typedef _Atomic(int_least8_t)		atomic_int_least8_t;
    365 typedef _Atomic(uint_least8_t)	atomic_uint_least8_t;
    366 typedef _Atomic(int_least16_t)	atomic_int_least16_t;
    367 typedef _Atomic(uint_least16_t)	atomic_uint_least16_t;
    368 typedef _Atomic(int_least32_t)	atomic_int_least32_t;
    369 typedef _Atomic(uint_least32_t)	atomic_uint_least32_t;
    370 typedef _Atomic(int_least64_t)	atomic_int_least64_t;
    371 typedef _Atomic(uint_least64_t)	atomic_uint_least64_t;
    372 typedef _Atomic(int_fast8_t)		atomic_int_fast8_t;
    373 typedef _Atomic(uint_fast8_t)		atomic_uint_fast8_t;
    374 typedef _Atomic(int_fast16_t)		atomic_int_fast16_t;
    375 typedef _Atomic(uint_fast16_t)	atomic_uint_fast16_t;
    376 typedef _Atomic(int_fast32_t)		atomic_int_fast32_t;
    377 typedef _Atomic(uint_fast32_t)	atomic_uint_fast32_t;
    378 typedef _Atomic(int_fast64_t)		atomic_int_fast64_t;
    379 typedef _Atomic(uint_fast64_t)	atomic_uint_fast64_t;
    380 typedef _Atomic(intptr_t)		atomic_intptr_t;
    381 typedef _Atomic(uintptr_t)		atomic_uintptr_t;
    382 typedef _Atomic(size_t)		atomic_size_t;
    383 typedef _Atomic(ptrdiff_t)		atomic_ptrdiff_t;
    384 typedef _Atomic(intmax_t)		atomic_intmax_t;
    385 typedef _Atomic(uintmax_t)		atomic_uintmax_t;
    386 
    387 /*
    388  * 7.17.7 Operations on atomic types.
    389  */
    390 
    391 /*
    392  * Compiler-specific operations.
    393  */
    394 
    395 #if defined(__CLANG_ATOMICS)
    396 #define	atomic_compare_exchange_strong_explicit(object, expected,	\
    397     desired, success, failure)						\
    398 	__c11_atomic_compare_exchange_strong(object, expected, desired,	\
    399 	    success, failure)
    400 #define	atomic_compare_exchange_weak_explicit(object, expected,		\
    401     desired, success, failure)						\
    402 	__c11_atomic_compare_exchange_weak(object, expected, desired,	\
    403 	    success, failure)
    404 #define	atomic_exchange_explicit(object, desired, order)		\
    405 	__c11_atomic_exchange(object, desired, order)
    406 #define	atomic_fetch_add_explicit(object, operand, order)		\
    407 	__c11_atomic_fetch_add(object, operand, order)
    408 #define	atomic_fetch_and_explicit(object, operand, order)		\
    409 	__c11_atomic_fetch_and(object, operand, order)
    410 #define	atomic_fetch_or_explicit(object, operand, order)		\
    411 	__c11_atomic_fetch_or(object, operand, order)
    412 #define	atomic_fetch_sub_explicit(object, operand, order)		\
    413 	__c11_atomic_fetch_sub(object, operand, order)
    414 #define	atomic_fetch_xor_explicit(object, operand, order)		\
    415 	__c11_atomic_fetch_xor(object, operand, order)
    416 #define	atomic_load_explicit(object, order)				\
    417 	__c11_atomic_load(object, order)
    418 #define	atomic_store_explicit(object, desired, order)			\
    419 	__c11_atomic_store(object, desired, order)
    420 #elif defined(__GNUC_ATOMICS)
    421 #define	atomic_compare_exchange_strong_explicit(object, expected,	\
    422     desired, success, failure)						\
    423 	__atomic_compare_exchange_n(&(object)->__val, expected,		\
    424 	    desired, 0, success, failure)
    425 #define	atomic_compare_exchange_weak_explicit(object, expected,		\
    426     desired, success, failure)						\
    427 	__atomic_compare_exchange_n(&(object)->__val, expected,		\
    428 	    desired, 1, success, failure)
    429 #define	atomic_exchange_explicit(object, desired, order)		\
    430 	__atomic_exchange_n(&(object)->__val, desired, order)
    431 #define	atomic_fetch_add_explicit(object, operand, order)		\
    432 	__atomic_fetch_add(&(object)->__val, operand, order)
    433 #define	atomic_fetch_and_explicit(object, operand, order)		\
    434 	__atomic_fetch_and(&(object)->__val, operand, order)
    435 #define	atomic_fetch_or_explicit(object, operand, order)		\
    436 	__atomic_fetch_or(&(object)->__val, operand, order)
    437 #define	atomic_fetch_sub_explicit(object, operand, order)		\
    438 	__atomic_fetch_sub(&(object)->__val, operand, order)
    439 #define	atomic_fetch_xor_explicit(object, operand, order)		\
    440 	__atomic_fetch_xor(&(object)->__val, operand, order)
    441 #define	atomic_load_explicit(object, order)				\
    442 	__atomic_load_n(&(object)->__val, order)
    443 #define	atomic_store_explicit(object, desired, order)			\
    444 	__atomic_store_n(&(object)->__val, desired, order)
    445 #else
    446 #define	__atomic_apply_stride(object, operand) \
    447 	(((__typeof__((object)->__val))0) + (operand))
    448 #define	atomic_compare_exchange_strong_explicit(object, expected,	\
    449     desired, success, failure)	__extension__ ({			\
    450 	__typeof__(expected) __ep = (expected);				\
    451 	__typeof__(*__ep) __e = *__ep;					\
    452 	(void)(success); (void)(failure);				\
    453 	(bool)((*__ep = __sync_val_compare_and_swap(&(object)->__val,	\
    454 	    __e, desired)) == __e);					\
    455 })
    456 #define	atomic_compare_exchange_weak_explicit(object, expected,		\
    457     desired, success, failure)						\
    458 	atomic_compare_exchange_strong_explicit(object, expected,	\
    459 		desired, success, failure)
    460 #ifdef __HAS_BUILTIN_SYNC_SWAP
    461 /* Clang provides a full-barrier atomic exchange - use it if available. */
    462 #define	atomic_exchange_explicit(object, desired, order)		\
    463 	((void)(order), __sync_swap(&(object)->__val, desired))
    464 #else
    465 /*
    466  * __sync_lock_test_and_set() is only an acquire barrier in theory (although in
    467  * practice it is usually a full barrier) so we need an explicit barrier before
    468  * it.
    469  */
    470 #define	atomic_exchange_explicit(object, desired, order)		\
    471 __extension__ ({							\
    472 	__typeof__(object) __o = (object);				\
    473 	__typeof__(desired) __d = (desired);				\
    474 	(void)(order);							\
    475 	__sync_synchronize();						\
    476 	__sync_lock_test_and_set(&(__o)->__val, __d);			\
    477 })
    478 #endif
    479 #define	atomic_fetch_add_explicit(object, operand, order)		\
    480 	((void)(order), __sync_fetch_and_add(&(object)->__val,		\
    481 	    __atomic_apply_stride(object, operand)))
    482 #define	atomic_fetch_and_explicit(object, operand, order)		\
    483 	((void)(order), __sync_fetch_and_and(&(object)->__val, operand))
    484 #define	atomic_fetch_or_explicit(object, operand, order)		\
    485 	((void)(order), __sync_fetch_and_or(&(object)->__val, operand))
    486 #define	atomic_fetch_sub_explicit(object, operand, order)		\
    487 	((void)(order), __sync_fetch_and_sub(&(object)->__val,		\
    488 	    __atomic_apply_stride(object, operand)))
    489 #define	atomic_fetch_xor_explicit(object, operand, order)		\
    490 	((void)(order), __sync_fetch_and_xor(&(object)->__val, operand))
    491 #define	atomic_load_explicit(object, order)				\
    492 	((void)(order), __sync_fetch_and_add(&(object)->__val, 0))
    493 #define	atomic_store_explicit(object, desired, order)			\
    494 	((void)atomic_exchange_explicit(object, desired, order))
    495 #endif
    496 
    497 /*
    498  * Convenience functions.
    499  *
    500  * Don't provide these in kernel space. In kernel space, we should be
    501  * disciplined enough to always provide explicit barriers.
    502  */
    503 
    504 #ifndef _KERNEL
    505 #define	atomic_compare_exchange_strong(object, expected, desired)	\
    506 	atomic_compare_exchange_strong_explicit(object, expected,	\
    507 	    desired, memory_order_seq_cst, memory_order_seq_cst)
    508 #define	atomic_compare_exchange_weak(object, expected, desired)		\
    509 	atomic_compare_exchange_weak_explicit(object, expected,		\
    510 	    desired, memory_order_seq_cst, memory_order_seq_cst)
    511 #define	atomic_exchange(object, desired)				\
    512 	atomic_exchange_explicit(object, desired, memory_order_seq_cst)
    513 #define	atomic_fetch_add(object, operand)				\
    514 	atomic_fetch_add_explicit(object, operand, memory_order_seq_cst)
    515 #define	atomic_fetch_and(object, operand)				\
    516 	atomic_fetch_and_explicit(object, operand, memory_order_seq_cst)
    517 #define	atomic_fetch_or(object, operand)				\
    518 	atomic_fetch_or_explicit(object, operand, memory_order_seq_cst)
    519 #define	atomic_fetch_sub(object, operand)				\
    520 	atomic_fetch_sub_explicit(object, operand, memory_order_seq_cst)
    521 #define	atomic_fetch_xor(object, operand)				\
    522 	atomic_fetch_xor_explicit(object, operand, memory_order_seq_cst)
    523 #define	atomic_load(object)						\
    524 	atomic_load_explicit(object, memory_order_seq_cst)
    525 #define	atomic_store(object, desired)					\
    526 	atomic_store_explicit(object, desired, memory_order_seq_cst)
    527 #endif /* !_KERNEL */
    528 
    529 /*
    530  * 7.17.8 Atomic flag type and operations.
    531  *
    532  * XXX: Assume atomic_bool can be used as an atomic_flag. Is there some
    533  * kind of compiler built-in type we could use?
    534  */
    535 
    536 typedef struct {
    537 	atomic_bool	__flag;
    538 } atomic_flag;
    539 
    540 #define	ATOMIC_FLAG_INIT		{ ATOMIC_VAR_INIT(false) }
    541 
    542 static __inline bool
    543 atomic_flag_test_and_set_explicit(volatile atomic_flag *__object,
    544     memory_order __order)
    545 {
    546 	return (atomic_exchange_explicit(&__object->__flag, 1, __order));
    547 }
    548 
    549 static __inline void
    550 atomic_flag_clear_explicit(volatile atomic_flag *__object, memory_order __order)
    551 {
    552 
    553 	atomic_store_explicit(&__object->__flag, 0, __order);
    554 }
    555 
    556 #ifndef _KERNEL
    557 static __inline bool
    558 atomic_flag_test_and_set(volatile atomic_flag *__object)
    559 {
    560 
    561 	return (atomic_flag_test_and_set_explicit(__object,
    562 	    memory_order_seq_cst));
    563 }
    564 
    565 static __inline void
    566 atomic_flag_clear(volatile atomic_flag *__object)
    567 {
    568 
    569 	atomic_flag_clear_explicit(__object, memory_order_seq_cst);
    570 }
    571 #endif /* !_KERNEL */
    572 
    573 #endif /* <atomic> unavailable */
    574 
    575 #endif /* !_STDATOMIC_H_ */
    576