/external/llvm/test/CodeGen/X86/ |
D | atomic64.ll | 9 %t1 = atomicrmw add i64* @sc64, i64 1 acquire 12 %t2 = atomicrmw add i64* @sc64, i64 3 acquire 15 %t3 = atomicrmw add i64* @sc64, i64 5 acquire 18 %t4 = atomicrmw add i64* @sc64, i64 %t3 acquire 28 %t1 = atomicrmw sub i64* @sc64, i64 1 acquire 31 %t2 = atomicrmw sub i64* @sc64, i64 3 acquire 34 %t3 = atomicrmw sub i64* @sc64, i64 5 acquire 37 %t4 = atomicrmw sub i64* @sc64, i64 %t3 acquire 47 %t1 = atomicrmw and i64* @sc64, i64 3 acquire 50 %t2 = atomicrmw and i64* @sc64, i64 5 acquire [all …]
|
D | atomic6432.ll | 9 %t1 = atomicrmw add i64* @sc64, i64 1 acquire 14 %t2 = atomicrmw add i64* @sc64, i64 3 acquire 19 %t3 = atomicrmw add i64* @sc64, i64 5 acquire 24 %t4 = atomicrmw add i64* @sc64, i64 %t3 acquire 36 %t1 = atomicrmw sub i64* @sc64, i64 1 acquire 41 %t2 = atomicrmw sub i64* @sc64, i64 3 acquire 46 %t3 = atomicrmw sub i64* @sc64, i64 5 acquire 51 %t4 = atomicrmw sub i64* @sc64, i64 %t3 acquire 63 %t1 = atomicrmw and i64* @sc64, i64 3 acquire 68 %t2 = atomicrmw and i64* @sc64, i64 4294967297 acquire [all …]
|
D | atomic8.ll | 11 %t1 = atomicrmw add i8* @sc8, i8 1 acquire 16 %t2 = atomicrmw add i8* @sc8, i8 3 acquire 21 %t3 = atomicrmw add i8* @sc8, i8 5 acquire 26 %t4 = atomicrmw add i8* @sc8, i8 %t3 acquire 39 %t1 = atomicrmw sub i8* @sc8, i8 1 acquire 44 %t2 = atomicrmw sub i8* @sc8, i8 3 acquire 49 %t3 = atomicrmw sub i8* @sc8, i8 5 acquire 54 %t4 = atomicrmw sub i8* @sc8, i8 %t3 acquire 67 %t1 = atomicrmw and i8* @sc8, i8 3 acquire 72 %t2 = atomicrmw and i8* @sc8, i8 5 acquire [all …]
|
D | atomic16.ll | 11 %t1 = atomicrmw add i16* @sc16, i16 1 acquire 16 %t2 = atomicrmw add i16* @sc16, i16 3 acquire 21 %t3 = atomicrmw add i16* @sc16, i16 5 acquire 26 %t4 = atomicrmw add i16* @sc16, i16 %t3 acquire 39 %t1 = atomicrmw sub i16* @sc16, i16 1 acquire 44 %t2 = atomicrmw sub i16* @sc16, i16 3 acquire 49 %t3 = atomicrmw sub i16* @sc16, i16 5 acquire 54 %t4 = atomicrmw sub i16* @sc16, i16 %t3 acquire 67 %t1 = atomicrmw and i16* @sc16, i16 3 acquire 72 %t2 = atomicrmw and i16* @sc16, i16 5 acquire [all …]
|
D | atomic32.ll | 12 %t1 = atomicrmw add i32* @sc32, i32 1 acquire 17 %t2 = atomicrmw add i32* @sc32, i32 3 acquire 22 %t3 = atomicrmw add i32* @sc32, i32 5 acquire 27 %t4 = atomicrmw add i32* @sc32, i32 %t3 acquire 40 %t1 = atomicrmw sub i32* @sc32, i32 1 acquire 45 %t2 = atomicrmw sub i32* @sc32, i32 3 acquire 50 %t3 = atomicrmw sub i32* @sc32, i32 5 acquire 55 %t4 = atomicrmw sub i32* @sc32, i32 %t3 acquire 68 %t1 = atomicrmw and i32* @sc32, i32 3 acquire 73 %t2 = atomicrmw and i32* @sc32, i32 5 acquire [all …]
|
D | pr5145.ll | 6 %1 = atomicrmw max i8* @sc8, i8 5 acquire 13 %2 = atomicrmw min i8* @sc8, i8 6 acquire 20 %3 = atomicrmw umax i8* @sc8, i8 7 acquire 27 %4 = atomicrmw umin i8* @sc8, i8 8 acquire
|
D | atomic-minmax-i6432.ll | 8 %1 = atomicrmw max i64* @sc64, i64 5 acquire 17 %2 = atomicrmw min i64* @sc64, i64 6 acquire 26 %3 = atomicrmw umax i64* @sc64, i64 7 acquire 35 %4 = atomicrmw umin i64* @sc64, i64 8 acquire
|
/external/llvm/test/Bitcode/ |
D | memInstructions.3.2.ll | 91 ; CHECK-NEXT: %res3 = load atomic i8* %ptr1 acquire, align 1 92 %res3 = load atomic i8* %ptr1 acquire, align 1 103 ; CHECK-NEXT: %res7 = load atomic volatile i8* %ptr1 acquire, align 1 104 %res7 = load atomic volatile i8* %ptr1 acquire, align 1 115 ; CHECK-NEXT: %res11 = load atomic i8* %ptr1 singlethread acquire, align 1 116 %res11 = load atomic i8* %ptr1 singlethread acquire, align 1 127 ; CHECK-NEXT: %res15 = load atomic volatile i8* %ptr1 singlethread acquire, align 1 128 %res15 = load atomic volatile i8* %ptr1 singlethread acquire, align 1 243 ; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new acquire acquire 245 %res5 = cmpxchg i32* %ptr, i32 %cmp, i32 %new acquire acquire [all …]
|
D | cmpxchg-upgrade.ll | 10 cmpxchg i32* %addr, i32 42, i32 0 acquire 11 ; CHECK: cmpxchg i32* %addr, i32 42, i32 0 acquire acquire 17 ; CHECK: cmpxchg i32* %addr, i32 42, i32 0 acq_rel acquire
|
/external/chromium_org/third_party/webrtc/base/ |
D | linked_ptr.h | 48 {acquire(r);} in throw() 53 acquire(r); 61 {acquire(r);} in linked_ptr() 66 acquire(r); 82 void acquire(const linked_ptr& r) throw() in acquire() function 96 template <class Y> void acquire(const linked_ptr<Y>& r) throw() in acquire() function
|
/external/compiler-rt/lib/tsan/tests/unit/ |
D | tsan_clock_test.cc | 41 vector.acquire(&chunked); in TEST() 59 vector2.acquire(&chunked); in TEST() 76 thr2.acquire(&sync); in TEST() 77 thr2.acquire(&sync); in TEST() 88 vector.acquire(&chunked); in TEST() 96 vector.acquire(&chunked); in TEST() 125 vector2.acquire(&chunked); in TEST() 131 vector1.acquire(&chunked); in TEST() 187 void acquire(const SimpleSyncClock *src) { in acquire() function 202 acquire(dst); in acq_rel() [all …]
|
/external/llvm/include/llvm/Support/ |
D | Mutex.h | 52 bool acquire(); 96 bool acquire() { in acquire() function 98 return MutexImpl::acquire(); in acquire() 143 mtx.acquire(); in SmartScopedLock()
|
/external/chromium_org/third_party/tlslite/tlslite/ |
D | basedb.py | 57 self.lock.acquire() 71 self.lock.acquire() 83 self.lock.acquire() 105 self.lock.acquire() 124 self.lock.acquire()
|
/external/deqp/framework/delibs/decpp/ |
D | deSharedPtr.hpp | 136 void acquire (void); 178 void acquire (void); 238 acquire(); in SharedPtr() 253 acquire(); in SharedPtr() 301 acquire(); in operator =() 327 acquire(); in operator =() 465 inline void SharedPtr<T, Deleter, threadSafe>::acquire (void) in acquire() function in de::SharedPtr 543 acquire(); in WeakPtr() 555 acquire(); in WeakPtr() 583 acquire(); in operator =() [all …]
|
/external/robolectric/src/test/java/com/xtremelabs/robolectric/shadows/ |
D | PowerManagerTest.java | 44 lock.acquire(); in shouldAcquireAndReleaseReferenceCountedLock() 46 lock.acquire(); in shouldAcquireAndReleaseReferenceCountedLock() 62 lock.acquire(); in shouldAcquireAndReleaseNonReferenceCountedLock() 64 lock.acquire(); in shouldAcquireAndReleaseNonReferenceCountedLock()
|
D | WifiManagerTest.java | 146 lock.acquire(); in shouldAcquireAndReleaseWifilockRefCounted() 147 lock.acquire(); in shouldAcquireAndReleaseWifilockRefCounted() 159 lock.acquire(); in shouldAcquireAndReleaseWifilockNonRefCounted() 161 lock.acquire(); in shouldAcquireAndReleaseWifilockNonRefCounted() 176 for (int i=0; i<ShadowWifiManager.ShadowWifiLock.MAX_ACTIVE_LOCKS; i++) lock.acquire(); in shouldThrowUnsupportedOperationIfLockisOverlocked()
|
/external/compiler-rt/lib/ubsan/ |
D | ubsan_handlers.cc | 31 Location Loc = Data->Loc.acquire(); in handleTypeMismatchImpl() 69 SourceLocation Loc = Data->Loc.acquire(); in HandleIntegerOverflow() 114 SourceLocation Loc = Data->Loc.acquire(); in __ubsan_handle_negate_overflow() 136 SourceLocation Loc = Data->Loc.acquire(); in __ubsan_handle_divrem_overflow() 159 SourceLocation Loc = Data->Loc.acquire(); in __ubsan_handle_shift_out_of_bounds() 188 SourceLocation Loc = Data->Loc.acquire(); in __ubsan_handle_out_of_bounds() 216 SourceLocation Loc = Data->Loc.acquire(); in __ubsan_handle_vla_bound_not_positive() 249 SourceLocation Loc = Data->Loc.acquire(); in __ubsan_handle_load_invalid_value()
|
/external/robolectric/src/main/java/com/xtremelabs/robolectric/shadows/ |
D | ShadowPowerManager.java | 37 public void acquire() { in acquire() method in ShadowPowerManager.ShadowWakeLock 38 acquire(0); in acquire() 43 public synchronized void acquire(long timeout) { in acquire() method in ShadowPowerManager.ShadowWakeLock
|
/external/llvm/test/Assembler/ |
D | atomic.ll | 7 ; CHECK: load atomic volatile i32* %x singlethread acquire, align 4 8 load atomic volatile i32* %x singlethread acquire, align 4 15 ; CHECK: cmpxchg volatile i32* %x, i32 0, i32 1 acq_rel acquire 16 cmpxchg volatile i32* %x, i32 0, i32 1 acq_rel acquire
|
/external/llvm/test/Instrumentation/ThreadSanitizer/ |
D | atomic.ll | 23 %0 = load atomic i8* %a acquire, align 1 127 atomicrmw xchg i8* %a, i8 0 acquire 135 atomicrmw add i8* %a, i8 0 acquire 143 atomicrmw sub i8* %a, i8 0 acquire 151 atomicrmw and i8* %a, i8 0 acquire 159 atomicrmw or i8* %a, i8 0 acquire 167 atomicrmw xor i8* %a, i8 0 acquire 175 atomicrmw nand i8* %a, i8 0 acquire 359 cmpxchg i8* %a, i8 0, i8 1 acquire acquire 375 cmpxchg i8* %a, i8 0, i8 1 acq_rel acquire [all …]
|
/external/chromium-trace/trace-viewer/third_party/pywebsocket/src/mod_pywebsocket/ |
D | mux.py | 546 self._write_condition.acquire() 572 self._write_condition.acquire() 590 self._read_condition.acquire() 608 self._read_condition.acquire() 641 self._read_condition.acquire() 690 self._write_inner_frame_semaphore.acquire() 693 self._send_quota_condition.acquire() 734 self._send_quota_condition.acquire() 888 self._deque_condition.acquire() 912 self._deque_condition.acquire() [all …]
|
/external/llvm/lib/Support/Unix/ |
D | RWMutex.inc | 26 // will therefore deadlock if a thread tries to acquire a read lock 36 return static_cast<Mutex *>(data_)->acquire(); 44 return static_cast<Mutex *>(data_)->acquire();
|
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/thirdparty/mod_pywebsocket/ |
D | mux.py | 587 self._write_condition.acquire() 614 self._write_condition.acquire() 628 self._write_condition.acquire() 643 self._read_condition.acquire() 661 self._read_condition.acquire() 694 self._read_condition.acquire() 868 self._write_inner_frame_semaphore.acquire() 873 self._send_condition.acquire() 888 self._send_condition.acquire() 933 self._send_condition.acquire() [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | swift-atomics.ll | 17 ; followed by an acquire does not get reordered. In that case a "dmb ishst" is 35 ; Also, pure acquire operations should definitely not have an ishst barrier. 43 %val = load atomic i32* %addr acquire, align 4
|
/external/llvm/test/Instrumentation/MemorySanitizer/ |
D | atomics.ll | 93 %0 = load atomic i32* %p acquire, align 16 98 ; CHECK: load atomic i32* {{.*}} acquire, align 16 104 ; atomic load monotonic: bump up to load acquire 113 ; CHECK: load atomic i32* {{.*}} acquire, align 16 119 ; atomic load unordered: bump up to load acquire 128 ; CHECK: load atomic i32* {{.*}} acquire, align 16
|