Lines Matching refs:morder
36 static bool IsLoadOrder(morder mo) { in IsLoadOrder()
41 static bool IsStoreOrder(morder mo) { in IsStoreOrder()
45 static bool IsReleaseOrder(morder mo) { in IsReleaseOrder()
49 static bool IsAcquireOrder(morder mo) { in IsAcquireOrder()
54 static bool IsAcqRelOrder(morder mo) { in IsAcqRelOrder()
197 static memory_order to_mo(morder mo) { in to_mo()
211 static T NoTsanAtomicLoad(const volatile T *a, morder mo) { in NoTsanAtomicLoad()
216 static a128 NoTsanAtomicLoad(const volatile a128 *a, morder mo) { in NoTsanAtomicLoad()
224 morder mo) { in AtomicLoad()
241 static void NoTsanAtomicStore(volatile T *a, T v, morder mo) { in NoTsanAtomicStore()
246 static void NoTsanAtomicStore(volatile a128 *a, a128 v, morder mo) { in NoTsanAtomicStore()
254 morder mo) { in AtomicStore()
276 static T AtomicRMW(ThreadState *thr, uptr pc, volatile T *a, T v, morder mo) { in AtomicRMW()
298 static T NoTsanAtomicExchange(volatile T *a, T v, morder mo) { in NoTsanAtomicExchange()
303 static T NoTsanAtomicFetchAdd(volatile T *a, T v, morder mo) { in NoTsanAtomicFetchAdd()
308 static T NoTsanAtomicFetchSub(volatile T *a, T v, morder mo) { in NoTsanAtomicFetchSub()
313 static T NoTsanAtomicFetchAnd(volatile T *a, T v, morder mo) { in NoTsanAtomicFetchAnd()
318 static T NoTsanAtomicFetchOr(volatile T *a, T v, morder mo) { in NoTsanAtomicFetchOr()
323 static T NoTsanAtomicFetchXor(volatile T *a, T v, morder mo) { in NoTsanAtomicFetchXor()
328 static T NoTsanAtomicFetchNand(volatile T *a, T v, morder mo) { in NoTsanAtomicFetchNand()
334 morder mo) { in AtomicExchange()
340 morder mo) { in AtomicFetchAdd()
346 morder mo) { in AtomicFetchSub()
352 morder mo) { in AtomicFetchAnd()
358 morder mo) { in AtomicFetchOr()
364 morder mo) { in AtomicFetchXor()
370 morder mo) { in AtomicFetchNand()
375 static bool NoTsanAtomicCAS(volatile T *a, T *c, T v, morder mo, morder fmo) { in NoTsanAtomicCAS()
381 morder mo, morder fmo) { in NoTsanAtomicCAS()
392 static T NoTsanAtomicCAS(volatile T *a, T c, T v, morder mo, morder fmo) { in NoTsanAtomicCAS()
399 volatile T *a, T *c, T v, morder mo, morder fmo) { in AtomicCAS()
432 volatile T *a, T c, T v, morder mo, morder fmo) { in AtomicCAS()
438 static void NoTsanAtomicFence(morder mo) { in NoTsanAtomicFence()
442 static void AtomicFence(ThreadState *thr, uptr pc, morder mo) { in AtomicFence()
456 mo = flags()->force_seq_cst_atomics ? (morder)mo_seq_cst : mo; \
468 morder mo, const char *func) in ScopedAtomic()
481 static void AtomicStatInc(ThreadState *thr, uptr size, morder mo, StatType t) { in AtomicStatInc()
499 a8 __tsan_atomic8_load(const volatile a8 *a, morder mo) { in __tsan_atomic8_load()
504 a16 __tsan_atomic16_load(const volatile a16 *a, morder mo) { in __tsan_atomic16_load()
509 a32 __tsan_atomic32_load(const volatile a32 *a, morder mo) { in __tsan_atomic32_load()
514 a64 __tsan_atomic64_load(const volatile a64 *a, morder mo) { in __tsan_atomic64_load()
520 a128 __tsan_atomic128_load(const volatile a128 *a, morder mo) { in __tsan_atomic128_load()
526 void __tsan_atomic8_store(volatile a8 *a, a8 v, morder mo) { in __tsan_atomic8_store()
531 void __tsan_atomic16_store(volatile a16 *a, a16 v, morder mo) { in __tsan_atomic16_store()
536 void __tsan_atomic32_store(volatile a32 *a, a32 v, morder mo) { in __tsan_atomic32_store()
541 void __tsan_atomic64_store(volatile a64 *a, a64 v, morder mo) { in __tsan_atomic64_store()
547 void __tsan_atomic128_store(volatile a128 *a, a128 v, morder mo) { in __tsan_atomic128_store()
553 a8 __tsan_atomic8_exchange(volatile a8 *a, a8 v, morder mo) { in __tsan_atomic8_exchange()
558 a16 __tsan_atomic16_exchange(volatile a16 *a, a16 v, morder mo) { in __tsan_atomic16_exchange()
563 a32 __tsan_atomic32_exchange(volatile a32 *a, a32 v, morder mo) { in __tsan_atomic32_exchange()
568 a64 __tsan_atomic64_exchange(volatile a64 *a, a64 v, morder mo) { in __tsan_atomic64_exchange()
574 a128 __tsan_atomic128_exchange(volatile a128 *a, a128 v, morder mo) { in __tsan_atomic128_exchange()
580 a8 __tsan_atomic8_fetch_add(volatile a8 *a, a8 v, morder mo) { in __tsan_atomic8_fetch_add()
585 a16 __tsan_atomic16_fetch_add(volatile a16 *a, a16 v, morder mo) { in __tsan_atomic16_fetch_add()
590 a32 __tsan_atomic32_fetch_add(volatile a32 *a, a32 v, morder mo) { in __tsan_atomic32_fetch_add()
595 a64 __tsan_atomic64_fetch_add(volatile a64 *a, a64 v, morder mo) { in __tsan_atomic64_fetch_add()
601 a128 __tsan_atomic128_fetch_add(volatile a128 *a, a128 v, morder mo) { in __tsan_atomic128_fetch_add()
607 a8 __tsan_atomic8_fetch_sub(volatile a8 *a, a8 v, morder mo) { in __tsan_atomic8_fetch_sub()
612 a16 __tsan_atomic16_fetch_sub(volatile a16 *a, a16 v, morder mo) { in __tsan_atomic16_fetch_sub()
617 a32 __tsan_atomic32_fetch_sub(volatile a32 *a, a32 v, morder mo) { in __tsan_atomic32_fetch_sub()
622 a64 __tsan_atomic64_fetch_sub(volatile a64 *a, a64 v, morder mo) { in __tsan_atomic64_fetch_sub()
628 a128 __tsan_atomic128_fetch_sub(volatile a128 *a, a128 v, morder mo) { in __tsan_atomic128_fetch_sub()
634 a8 __tsan_atomic8_fetch_and(volatile a8 *a, a8 v, morder mo) { in __tsan_atomic8_fetch_and()
639 a16 __tsan_atomic16_fetch_and(volatile a16 *a, a16 v, morder mo) { in __tsan_atomic16_fetch_and()
644 a32 __tsan_atomic32_fetch_and(volatile a32 *a, a32 v, morder mo) { in __tsan_atomic32_fetch_and()
649 a64 __tsan_atomic64_fetch_and(volatile a64 *a, a64 v, morder mo) { in __tsan_atomic64_fetch_and()
655 a128 __tsan_atomic128_fetch_and(volatile a128 *a, a128 v, morder mo) { in __tsan_atomic128_fetch_and()
661 a8 __tsan_atomic8_fetch_or(volatile a8 *a, a8 v, morder mo) { in __tsan_atomic8_fetch_or()
666 a16 __tsan_atomic16_fetch_or(volatile a16 *a, a16 v, morder mo) { in __tsan_atomic16_fetch_or()
671 a32 __tsan_atomic32_fetch_or(volatile a32 *a, a32 v, morder mo) { in __tsan_atomic32_fetch_or()
676 a64 __tsan_atomic64_fetch_or(volatile a64 *a, a64 v, morder mo) { in __tsan_atomic64_fetch_or()
682 a128 __tsan_atomic128_fetch_or(volatile a128 *a, a128 v, morder mo) { in __tsan_atomic128_fetch_or()
688 a8 __tsan_atomic8_fetch_xor(volatile a8 *a, a8 v, morder mo) { in __tsan_atomic8_fetch_xor()
693 a16 __tsan_atomic16_fetch_xor(volatile a16 *a, a16 v, morder mo) { in __tsan_atomic16_fetch_xor()
698 a32 __tsan_atomic32_fetch_xor(volatile a32 *a, a32 v, morder mo) { in __tsan_atomic32_fetch_xor()
703 a64 __tsan_atomic64_fetch_xor(volatile a64 *a, a64 v, morder mo) { in __tsan_atomic64_fetch_xor()
709 a128 __tsan_atomic128_fetch_xor(volatile a128 *a, a128 v, morder mo) { in __tsan_atomic128_fetch_xor()
715 a8 __tsan_atomic8_fetch_nand(volatile a8 *a, a8 v, morder mo) { in __tsan_atomic8_fetch_nand()
720 a16 __tsan_atomic16_fetch_nand(volatile a16 *a, a16 v, morder mo) { in __tsan_atomic16_fetch_nand()
725 a32 __tsan_atomic32_fetch_nand(volatile a32 *a, a32 v, morder mo) { in __tsan_atomic32_fetch_nand()
730 a64 __tsan_atomic64_fetch_nand(volatile a64 *a, a64 v, morder mo) { in __tsan_atomic64_fetch_nand()
736 a128 __tsan_atomic128_fetch_nand(volatile a128 *a, a128 v, morder mo) { in __tsan_atomic128_fetch_nand()
743 morder mo, morder fmo) { in __tsan_atomic8_compare_exchange_strong()
749 morder mo, morder fmo) { in __tsan_atomic16_compare_exchange_strong()
755 morder mo, morder fmo) { in __tsan_atomic32_compare_exchange_strong()
761 morder mo, morder fmo) { in __tsan_atomic64_compare_exchange_strong()
768 morder mo, morder fmo) { in __tsan_atomic128_compare_exchange_strong()
775 morder mo, morder fmo) { in __tsan_atomic8_compare_exchange_weak()
781 morder mo, morder fmo) { in __tsan_atomic16_compare_exchange_weak()
787 morder mo, morder fmo) { in __tsan_atomic32_compare_exchange_weak()
793 morder mo, morder fmo) { in __tsan_atomic64_compare_exchange_weak()
800 morder mo, morder fmo) { in __tsan_atomic128_compare_exchange_weak()
807 morder mo, morder fmo) { in __tsan_atomic8_compare_exchange_val()
813 morder mo, morder fmo) { in __tsan_atomic16_compare_exchange_val()
819 morder mo, morder fmo) { in __tsan_atomic32_compare_exchange_val()
825 morder mo, morder fmo) { in __tsan_atomic64_compare_exchange_val()
832 morder mo, morder fmo) { in __tsan_atomic128_compare_exchange_val()
838 void __tsan_atomic_thread_fence(morder mo) { in __tsan_atomic_thread_fence()
844 void __tsan_atomic_signal_fence(morder mo) { in __tsan_atomic_signal_fence()