Home | History | Annotate | Download | only in sanitizer_common
      1 //===-- sanitizer_atomic_msvc.h ---------------------------------*- C++ -*-===//
      2 //
      3 //                     The LLVM Compiler Infrastructure
      4 //
      5 // This file is distributed under the University of Illinois Open Source
      6 // License. See LICENSE.TXT for details.
      7 //
      8 //===----------------------------------------------------------------------===//
      9 //
     10 // This file is a part of ThreadSanitizer/AddressSanitizer runtime.
     11 // Not intended for direct inclusion. Include sanitizer_atomic.h.
     12 //
     13 //===----------------------------------------------------------------------===//
     14 
     15 #ifndef SANITIZER_ATOMIC_MSVC_H
     16 #define SANITIZER_ATOMIC_MSVC_H
     17 
     18 extern "C" void _ReadWriteBarrier();
     19 #pragma intrinsic(_ReadWriteBarrier)
     20 extern "C" void _mm_mfence();
     21 #pragma intrinsic(_mm_mfence)
     22 extern "C" void _mm_pause();
     23 #pragma intrinsic(_mm_pause)
     24 extern "C" char _InterlockedExchange8(   // NOLINT
     25     char volatile *Addend, char Value);  // NOLINT
     26 #pragma intrinsic(_InterlockedExchange8)
     27 extern "C" short _InterlockedExchange16(   // NOLINT
     28     short volatile *Addend, short Value);  // NOLINT
     29 #pragma intrinsic(_InterlockedExchange16)
     30 extern "C" long _InterlockedExchange(    // NOLINT
     31     long volatile *Addend, long Value);  // NOLINT
     32 #pragma intrinsic(_InterlockedExchange)
     33 extern "C" long _InterlockedExchangeAdd(  // NOLINT
     34     long volatile * Addend, long Value);  // NOLINT
     35 #pragma intrinsic(_InterlockedExchangeAdd)
     36 extern "C" short _InterlockedCompareExchange16(  // NOLINT
     37     short volatile *Destination,                 // NOLINT
     38     short Exchange, short Comparand);            // NOLINT
     39 #pragma intrinsic(_InterlockedCompareExchange16)
     40 extern "C"
     41 long long _InterlockedCompareExchange64(  // NOLINT
     42     long long volatile *Destination,              // NOLINT
     43     long long Exchange, long long Comparand);     // NOLINT
     44 #pragma intrinsic(_InterlockedCompareExchange64)
     45 extern "C" void *_InterlockedCompareExchangePointer(
     46     void *volatile *Destination,
     47     void *Exchange, void *Comparand);
     48 #pragma intrinsic(_InterlockedCompareExchangePointer)
     49 extern "C"
     50 long __cdecl _InterlockedCompareExchange(  // NOLINT
     51     long volatile *Destination,            // NOLINT
     52     long Exchange, long Comparand);        // NOLINT
     53 #pragma intrinsic(_InterlockedCompareExchange)
     54 
     55 #ifdef _WIN64
     56 extern "C" long long _InterlockedExchangeAdd64(     // NOLINT
     57     long long volatile * Addend, long long Value);  // NOLINT
     58 #pragma intrinsic(_InterlockedExchangeAdd64)
     59 #endif
     60 
     61 namespace __sanitizer {
     62 
     63 INLINE void atomic_signal_fence(memory_order) {
     64   _ReadWriteBarrier();
     65 }
     66 
     67 INLINE void atomic_thread_fence(memory_order) {
     68   _mm_mfence();
     69 }
     70 
     71 INLINE void proc_yield(int cnt) {
     72   for (int i = 0; i < cnt; i++)
     73     _mm_pause();
     74 }
     75 
     76 template<typename T>
     77 INLINE typename T::Type atomic_load(
     78     const volatile T *a, memory_order mo) {
     79   DCHECK(mo & (memory_order_relaxed | memory_order_consume
     80       | memory_order_acquire | memory_order_seq_cst));
     81   DCHECK(!((uptr)a % sizeof(*a)));
     82   typename T::Type v;
     83   // FIXME(dvyukov): 64-bit load is not atomic on 32-bits.
     84   if (mo == memory_order_relaxed) {
     85     v = a->val_dont_use;
     86   } else {
     87     atomic_signal_fence(memory_order_seq_cst);
     88     v = a->val_dont_use;
     89     atomic_signal_fence(memory_order_seq_cst);
     90   }
     91   return v;
     92 }
     93 
     94 template<typename T>
     95 INLINE void atomic_store(volatile T *a, typename T::Type v, memory_order mo) {
     96   DCHECK(mo & (memory_order_relaxed | memory_order_release
     97       | memory_order_seq_cst));
     98   DCHECK(!((uptr)a % sizeof(*a)));
     99   // FIXME(dvyukov): 64-bit store is not atomic on 32-bits.
    100   if (mo == memory_order_relaxed) {
    101     a->val_dont_use = v;
    102   } else {
    103     atomic_signal_fence(memory_order_seq_cst);
    104     a->val_dont_use = v;
    105     atomic_signal_fence(memory_order_seq_cst);
    106   }
    107   if (mo == memory_order_seq_cst)
    108     atomic_thread_fence(memory_order_seq_cst);
    109 }
    110 
    111 INLINE u32 atomic_fetch_add(volatile atomic_uint32_t *a,
    112     u32 v, memory_order mo) {
    113   (void)mo;
    114   DCHECK(!((uptr)a % sizeof(*a)));
    115   return (u32)_InterlockedExchangeAdd(
    116       (volatile long*)&a->val_dont_use, (long)v);  // NOLINT
    117 }
    118 
    119 INLINE uptr atomic_fetch_add(volatile atomic_uintptr_t *a,
    120     uptr v, memory_order mo) {
    121   (void)mo;
    122   DCHECK(!((uptr)a % sizeof(*a)));
    123 #ifdef _WIN64
    124   return (uptr)_InterlockedExchangeAdd64(
    125       (volatile long long*)&a->val_dont_use, (long long)v);  // NOLINT
    126 #else
    127   return (uptr)_InterlockedExchangeAdd(
    128       (volatile long*)&a->val_dont_use, (long)v);  // NOLINT
    129 #endif
    130 }
    131 
    132 INLINE u32 atomic_fetch_sub(volatile atomic_uint32_t *a,
    133     u32 v, memory_order mo) {
    134   (void)mo;
    135   DCHECK(!((uptr)a % sizeof(*a)));
    136   return (u32)_InterlockedExchangeAdd(
    137       (volatile long*)&a->val_dont_use, -(long)v);  // NOLINT
    138 }
    139 
    140 INLINE uptr atomic_fetch_sub(volatile atomic_uintptr_t *a,
    141     uptr v, memory_order mo) {
    142   (void)mo;
    143   DCHECK(!((uptr)a % sizeof(*a)));
    144 #ifdef _WIN64
    145   return (uptr)_InterlockedExchangeAdd64(
    146       (volatile long long*)&a->val_dont_use, -(long long)v);  // NOLINT
    147 #else
    148   return (uptr)_InterlockedExchangeAdd(
    149       (volatile long*)&a->val_dont_use, -(long)v);  // NOLINT
    150 #endif
    151 }
    152 
    153 INLINE u8 atomic_exchange(volatile atomic_uint8_t *a,
    154     u8 v, memory_order mo) {
    155   (void)mo;
    156   DCHECK(!((uptr)a % sizeof(*a)));
    157   return (u8)_InterlockedExchange8((volatile char*)&a->val_dont_use, v);
    158 }
    159 
    160 INLINE u16 atomic_exchange(volatile atomic_uint16_t *a,
    161     u16 v, memory_order mo) {
    162   (void)mo;
    163   DCHECK(!((uptr)a % sizeof(*a)));
    164   return (u16)_InterlockedExchange16((volatile short*)&a->val_dont_use, v);
    165 }
    166 
    167 INLINE u32 atomic_exchange(volatile atomic_uint32_t *a,
    168     u32 v, memory_order mo) {
    169   (void)mo;
    170   DCHECK(!((uptr)a % sizeof(*a)));
    171   return (u32)_InterlockedExchange((volatile long*)&a->val_dont_use, v);
    172 }
    173 
    174 #ifndef _WIN64
    175 
    176 INLINE bool atomic_compare_exchange_strong(volatile atomic_uint8_t *a,
    177                                            u8 *cmp,
    178                                            u8 xchgv,
    179                                            memory_order mo) {
    180   (void)mo;
    181   DCHECK(!((uptr)a % sizeof(*a)));
    182   u8 cmpv = *cmp;
    183   u8 prev;
    184   __asm {
    185     mov al, cmpv
    186     mov ecx, a
    187     mov dl, xchgv
    188     lock cmpxchg [ecx], dl
    189     mov prev, al
    190   }
    191   if (prev == cmpv)
    192     return true;
    193   *cmp = prev;
    194   return false;
    195 }
    196 
    197 #endif
    198 
    199 INLINE bool atomic_compare_exchange_strong(volatile atomic_uintptr_t *a,
    200                                            uptr *cmp,
    201                                            uptr xchg,
    202                                            memory_order mo) {
    203   uptr cmpv = *cmp;
    204   uptr prev = (uptr)_InterlockedCompareExchangePointer(
    205       (void*volatile*)&a->val_dont_use, (void*)xchg, (void*)cmpv);
    206   if (prev == cmpv)
    207     return true;
    208   *cmp = prev;
    209   return false;
    210 }
    211 
    212 INLINE bool atomic_compare_exchange_strong(volatile atomic_uint16_t *a,
    213                                            u16 *cmp,
    214                                            u16 xchg,
    215                                            memory_order mo) {
    216   u16 cmpv = *cmp;
    217   u16 prev = (u16)_InterlockedCompareExchange16(
    218       (volatile short*)&a->val_dont_use, (short)xchg, (short)cmpv);
    219   if (prev == cmpv)
    220     return true;
    221   *cmp = prev;
    222   return false;
    223 }
    224 
    225 INLINE bool atomic_compare_exchange_strong(volatile atomic_uint32_t *a,
    226                                            u32 *cmp,
    227                                            u32 xchg,
    228                                            memory_order mo) {
    229   u32 cmpv = *cmp;
    230   u32 prev = (u32)_InterlockedCompareExchange(
    231       (volatile long*)&a->val_dont_use, (long)xchg, (long)cmpv);
    232   if (prev == cmpv)
    233     return true;
    234   *cmp = prev;
    235   return false;
    236 }
    237 
    238 INLINE bool atomic_compare_exchange_strong(volatile atomic_uint64_t *a,
    239                                            u64 *cmp,
    240                                            u64 xchg,
    241                                            memory_order mo) {
    242   u64 cmpv = *cmp;
    243   u64 prev = (u64)_InterlockedCompareExchange64(
    244       (volatile long long*)&a->val_dont_use, (long long)xchg, (long long)cmpv);
    245   if (prev == cmpv)
    246     return true;
    247   *cmp = prev;
    248   return false;
    249 }
    250 
    251 template<typename T>
    252 INLINE bool atomic_compare_exchange_weak(volatile T *a,
    253                                          typename T::Type *cmp,
    254                                          typename T::Type xchg,
    255                                          memory_order mo) {
    256   return atomic_compare_exchange_strong(a, cmp, xchg, mo);
    257 }
    258 
    259 }  // namespace __sanitizer
    260 
    261 #endif  // SANITIZER_ATOMIC_CLANG_H
    262