Home | History | Annotate | Download | only in include

Lines Matching refs:mb

210 #define mb() 	__asm__ __volatile__ ("lock; addl $0,0(%%esp)": : :"memory")
211 #define rmb() mb()
214 #define mb() asm volatile("mfence":::"memory")
358 * mb(): wmb()/rmb() combo, i.e., all previous memory
363 * Note: "mb()" and its variants cannot be used as a fence to order
369 #define mb() __asm__ __volatile__ ("mf" ::: "memory")
370 #define rmb() mb()
371 #define wmb() mb()
448 " mb\n"
466 " mb\n"
484 #define mb() \
485 __asm__ __volatile__("mb": : :"memory")
488 __asm__ __volatile__("mb": : :"memory")
669 * mb() prevents loads and stores being reordered across this point.
675 * stronger but slower sync instruction for mb and rmb.
677 #define mb() __asm__ __volatile__ ("sync" : : : "memory")
876 #define mb() \
886 #define rmb() mb()
887 #define wmb() mb()
1005 #define mb() __asm__ __volatile__ ("" : : : "memory")
1006 #define rmb() mb()
1007 #define wmb() mb()
1074 #define mb() __asm__ __volatile__ ("" : : : "memory")
1075 #define rmb() mb()
1076 #define wmb() mb()
1099 #define mb()