1 /* 2 * Copyright (C) 2011 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 #ifndef BIONIC_ATOMIC_X86_H 17 #define BIONIC_ATOMIC_X86_H 18 19 /* Define a full memory barrier, this is only needed if we build the 20 * platform for a multi-core device. 21 */ 22 #if defined(ANDROID_SMP) && ANDROID_SMP == 1 23 __ATOMIC_INLINE__ void 24 __bionic_memory_barrier() 25 { 26 __asm__ __volatile__ ( "mfence" : : : "memory" ); 27 } 28 #else 29 __ATOMIC_INLINE__ void 30 __bionic_memory_barrier() 31 { 32 /* A simple compiler barrier */ 33 __asm__ __volatile__ ( "" : : : "memory" ); 34 } 35 #endif 36 37 /* Compare-and-swap, without any explicit barriers. Note that this function 38 * returns 0 on success, and 1 on failure. The opposite convention is typically 39 * used on other platforms. 40 */ 41 __ATOMIC_INLINE__ int 42 __bionic_cmpxchg(int32_t old_value, int32_t new_value, volatile int32_t* ptr) 43 { 44 int32_t prev; 45 __asm__ __volatile__ ("lock; cmpxchgl %1, %2" 46 : "=a" (prev) 47 : "q" (new_value), "m" (*ptr), "0" (old_value) 48 : "memory"); 49 return prev != old_value; 50 } 51 52 53 /* Swap, without any explicit barriers */ 54 __ATOMIC_INLINE__ int32_t 55 __bionic_swap(int32_t new_value, volatile int32_t *ptr) 56 { 57 __asm__ __volatile__ ("xchgl %1, %0" 58 : "=r" (new_value) 59 : "m" (*ptr), "0" (new_value) 60 : "memory"); 61 return new_value; 62 } 63 64 /* Atomic increment, without explicit barriers */ 65 __ATOMIC_INLINE__ int32_t 66 __bionic_atomic_inc(volatile int32_t *ptr) 67 { 68 int increment = 1; 69 __asm__ __volatile__ ("lock; xaddl %0, %1" 70 : "+r" (increment), "+m" (*ptr) 71 : : "memory"); 72 /* increment now holds the old value of *ptr */ 73 return increment; 74 } 75 76 /* Atomic decrement, without explicit barriers */ 77 __ATOMIC_INLINE__ int32_t 78 __bionic_atomic_dec(volatile int32_t *ptr) 79 { 80 int increment = -1; 81 __asm__ __volatile__ ("lock; xaddl %0, %1" 82 : "+r" (increment), "+m" (*ptr) 83 : : "memory"); 84 /* increment now holds the old value of *ptr */ 85 return increment; 86 } 87 88 #endif /* BIONIC_ATOMIC_X86_H */ 89