Home | History | Annotate | Download | only in rtl

Lines Matching refs:morder

57 } morder;
59 static bool IsLoadOrder(morder mo) {
64 static bool IsStoreOrder(morder mo) {
68 static bool IsReleaseOrder(morder mo) {
72 static bool IsAcquireOrder(morder mo) {
77 static bool IsAcqRelOrder(morder mo) {
220 static memory_order to_mo(morder mo) {
234 static T NoTsanAtomicLoad(const volatile T *a, morder mo) {
239 static a128 NoTsanAtomicLoad(const volatile a128 *a, morder mo) {
247 morder mo) {
264 static void NoTsanAtomicStore(volatile T *a, T v, morder mo) {
269 static void NoTsanAtomicStore(volatile a128 *a, a128 v, morder mo) {
277 morder mo) {
299 static T AtomicRMW(ThreadState *thr, uptr pc, volatile T *a, T v, morder mo) {
321 static T NoTsanAtomicExchange(volatile T *a, T v, morder mo) {
326 static T NoTsanAtomicFetchAdd(volatile T *a, T v, morder mo) {
331 static T NoTsanAtomicFetchSub(volatile T *a, T v, morder mo) {
336 static T NoTsanAtomicFetchAnd(volatile T *a, T v, morder mo) {
341 static T NoTsanAtomicFetchOr(volatile T *a, T v, morder mo) {
346 static T NoTsanAtomicFetchXor(volatile T *a, T v, morder mo) {
351 static T NoTsanAtomicFetchNand(volatile T *a, T v, morder mo) {
357 morder mo) {
363 morder mo) {
369 morder mo) {
375 morder mo) {
381 morder mo) {
387 morder mo) {
393 morder mo) {
398 static bool NoTsanAtomicCAS(volatile T *a, T *c, T v, morder mo, morder fmo) {
404 morder mo, morder fmo) {
415 static T NoTsanAtomicCAS(volatile T *a, T c, T v, morder mo, morder fmo) {
422 volatile T *a, T *c, T v, morder mo, morder fmo) {
455 volatile T *a, T c, T v, morder mo, morder fmo) {
461 static void NoTsanAtomicFence(morder mo) {
465 static void AtomicFence(ThreadState *thr, uptr pc, morder mo) {
479 mo = flags()->force_seq_cst_atomics ? (morder)mo_seq_cst : mo; \
491 morder mo, const char *func)
504 static void AtomicStatInc(ThreadState *thr, uptr size, morder mo, StatType t) {
522 a8 __tsan_atomic8_load(const volatile a8 *a, morder mo) {
527 a16 __tsan_atomic16_load(const volatile a16 *a, morder mo) {
532 a32 __tsan_atomic32_load(const volatile a32 *a, morder mo) {
537 a64 __tsan_atomic64_load(const volatile a64 *a, morder mo) {
543 a128 __tsan_atomic128_load(const volatile a128 *a, morder mo) {
549 void __tsan_atomic8_store(volatile a8 *a, a8 v, morder mo) {
554 void __tsan_atomic16_store(volatile a16 *a, a16 v, morder mo) {
559 void __tsan_atomic32_store(volatile a32 *a, a32 v, morder mo) {
564 void __tsan_atomic64_store(volatile a64 *a, a64 v, morder mo) {
570 void __tsan_atomic128_store(volatile a128 *a, a128 v, morder mo) {
576 a8 __tsan_atomic8_exchange(volatile a8 *a, a8 v, morder mo) {
581 a16 __tsan_atomic16_exchange(volatile a16 *a, a16 v, morder mo) {
586 a32 __tsan_atomic32_exchange(volatile a32 *a, a32 v, morder mo) {
591 a64 __tsan_atomic64_exchange(volatile a64 *a, a64 v, morder mo) {
597 morder mo) {
603 a8 __tsan_atomic8_fetch_add(volatile a8 *a, a8 v, morder mo) {
608 a16 __tsan_atomic16_fetch_add(volatile a16 *a, a16 v, morder mo) {
613 a32 __tsan_atomic32_fetch_add(volatile a32 *a, a32 v, morder mo) {
618 a64 __tsan_atomic64_fetch_add(volatile a64 *a, a64 v, morder mo) {
624 a128 __tsan_atomic128_fetch_add(volatile a128 *a, a128 v, morder mo) {
630 a8 __tsan_atomic8_fetch_sub(volatile a8 *a, a8 v, morder mo) {
635 a16 __tsan_atomic16_fetch_sub(volatile a16 *a, a16 v, morder mo) {
640 a32 __tsan_atomic32_fetch_sub(volatile a32 *a, a32 v, morder mo) {
645 a64 __tsan_atomic64_fetch_sub(volatile a64 *a, a64 v, morder mo) {
651 a128 __tsan_atomic128_fetch_sub(volatile a128 *a, a128 v, morder mo) {
657 a8 __tsan_atomic8_fetch_and(volatile a8 *a, a8 v, morder mo) {
662 a16 __tsan_atomic16_fetch_and(volatile a16 *a, a16 v, morder mo) {
667 a32 __tsan_atomic32_fetch_and(volatile a32 *a, a32 v, morder mo) {
672 a64 __tsan_atomic64_fetch_and(volatile a64 *a, a64 v, morder mo) {
678 a128 __tsan_atomic128_fetch_and(volatile a128 *a, a128 v, morder mo) {
684 a8 __tsan_atomic8_fetch_or(volatile a8 *a, a8 v, morder mo) {
689 a16 __tsan_atomic16_fetch_or(volatile a16 *a, a16 v, morder mo) {
694 a32 __tsan_atomic32_fetch_or(volatile a32 *a, a32 v, morder mo) {
699 a64 __tsan_atomic64_fetch_or(volatile a64 *a, a64 v, morder mo) {
705 a128 __tsan_atomic128_fetch_or(volatile a128 *a, a128 v, morder mo) {
711 a8 __tsan_atomic8_fetch_xor(volatile a8 *a, a8 v, morder mo) {
716 a16 __tsan_atomic16_fetch_xor(volatile a16 *a, a16 v, morder mo) {
721 a32 __tsan_atomic32_fetch_xor(volatile a32 *a, a32 v, morder mo) {
726 a64 __tsan_atomic64_fetch_xor(volatile a64 *a, a64 v, morder mo) {
732 a128 __tsan_atomic128_fetch_xor(volatile a128 *a, a128 v, morder mo) {
738 a8 __tsan_atomic8_fetch_nand(volatile a8 *a, a8 v, morder mo) {
743 a16 __tsan_atomic16_fetch_nand(volatile a16 *a, a16 v, morder mo) {
748 a32 __tsan_atomic32_fetch_nand(volatile a32 *a, a32 v, morder mo) {
753 a64 __tsan_atomic64_fetch_nand(volatile a64 *a, a64 v, morder mo) {
759 a128 __tsan_atomic128_fetch_nand(volatile a128 *a, a128 v, morder mo) {
766 morder mo, morder fmo) {
772 morder mo, morder fmo) {
778 morder mo, morder fmo) {
784 morder mo, morder fmo) {
791 morder mo, morder fmo) {
798 morder mo, morder fmo) {
804 morder mo, morder fmo) {
810 morder mo, morder fmo) {
816 morder mo, morder fmo) {
823 morder mo, morder fmo) {
830 morder mo, morder fmo) {
836 morder mo, morder fmo) {
842 morder mo, morder fmo) {
848 morder mo, morder fmo) {
855 morder mo, morder fmo) {
861 void __tsan_atomic_thread_fence(morder mo) {
867 void __tsan_atomic_signal_fence(morder mo) {