Searched refs:rmw (Results 1 – 12 of 12) sorted by relevance
/external/swiftshader/third_party/subzero/tests_lit/llvm2ice_tests/ |
D | rmw.ll | 12 %rmw = add i32 %val, %var 13 store i32 %rmw, i32* %addr, align 1 24 %rmw = add i32 %val, 19 25 store i32 %rmw, i32* %addr, align 1 36 %rmw = add i32 %val, %var 37 store i32 %rmw, i32* %addr, align 1 38 ret i32 %rmw 48 %rmw = add i16 %val, %var 49 store i16 %rmw, i16* %addr, align 1 60 %rmw = add i16 %val, 19 [all …]
|
D | nacl-atomic-errors.ll | 13 declare i8 @llvm.nacl.atomic.rmw.i8(i32, i8*, i8, i32) 14 declare i16 @llvm.nacl.atomic.rmw.i16(i32, i16*, i16, i32) 15 declare i32 @llvm.nacl.atomic.rmw.i32(i32, i32*, i32, i32) 16 declare i64 @llvm.nacl.atomic.rmw.i64(i32, i64*, i64, i32) 88 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr, i8 %trunc, i32 1) 97 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 7) 106 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 0, i16* %ptr, i16 %trunc, i32 6) 115 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 7, i32* %ptr, i32 %v, i32 6) 123 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4294967295, i32* %ptr, i32 %v, i32 6) 202 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr, i8 %trunc, i32 %iptr)
|
D | nacl-atomic-intrinsics.ll | 49 declare i8 @llvm.nacl.atomic.rmw.i8(i32, i8*, i8, i32) 50 declare i16 @llvm.nacl.atomic.rmw.i16(i32, i16*, i16, i32) 51 declare i32 @llvm.nacl.atomic.rmw.i32(i32, i32*, i32, i32) 52 declare i64 @llvm.nacl.atomic.rmw.i64(i32, i64*, i64, i32) 339 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr, i8 %trunc, i32 6) 380 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i16* %ptr, i16 %trunc, i32 6) 420 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) 444 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) 479 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) 506 %old = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) [all …]
|
D | abi-atomics.ll | 17 declare i8 @llvm.nacl.atomic.rmw.i8(i32, i8*, i8, i32) 18 declare i16 @llvm.nacl.atomic.rmw.i16(i32, i16*, i16, i32) 19 declare i32 @llvm.nacl.atomic.rmw.i32(i32, i32*, i32, i32) 20 declare i64 @llvm.nacl.atomic.rmw.i64(i32, i64*, i64, i32) 148 ; rmw 152 %1 = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 0, i32 7) 159 %1 = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 0, i32 0) 166 %1 = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 0, i32 6) 173 %1 = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 0, i32 5) 180 %1 = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 0, i32 4) [all …]
|
/external/swiftshader/third_party/subzero/tests_lit/reader_tests/ |
D | nacl-atomic-intrinsics.ll | 15 declare i8 @llvm.nacl.atomic.rmw.i8(i32, i8*, i8, i32) 16 declare i16 @llvm.nacl.atomic.rmw.i16(i32, i16*, i16, i32) 17 declare i32 @llvm.nacl.atomic.rmw.i32(i32, i32*, i32, i32) 18 declare i64 @llvm.nacl.atomic.rmw.i64(i32, i64*, i64, i32) 166 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr, i8 %trunc, i32 6) 174 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i32 %iptr, i8 %trunc, i32 6) 183 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i16* %ptr, i16 %trunc, i32 6) 191 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i32 %iptr, i16 %trunc, i32 6) 199 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) 205 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32 %iptr, i32 %v, i32 6) [all …]
|
/external/swiftshader/third_party/subzero/tests_lit/assembler/arm32/ |
D | ldrex-strex.ll | 21 declare i8 @llvm.nacl.atomic.rmw.i8(i32, i8*, i8, i32) 23 declare i16 @llvm.nacl.atomic.rmw.i16(i32, i16*, i16, i32) 25 declare i32 @llvm.nacl.atomic.rmw.i32(i32, i32*, i32, i32) #0 27 declare i64 @llvm.nacl.atomic.rmw.i64(i32, i64*, i64, i32) #0 38 %v = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr.asptr, 78 %v = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i16* %ptr.asptr, 107 %v = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr.asptr, 136 %v = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr.asptr,
|
/external/webrtc/webrtc/modules/audio_processing/beamformer/ |
D | nonlinear_beamformer.cc | 435 complex_f rmw = abs(ConjugateDotProduct(delay_sum_masks_[i], eig_m_)); in ProcessAudioBlock() local 436 rmw *= rmw; in ProcessAudioBlock() 437 float rmw_r = rmw.real(); in ProcessAudioBlock()
|
/external/llvm/test/CodeGen/X86/ |
D | atomic_idempotent.ll | 4 ; On x86, an atomic rmw operation that does not modify the value in memory
|
/external/swiftshader/third_party/subzero/tests_lit/assembler/x86/ |
D | jump_encodings.ll | 11 declare i32 @llvm.nacl.atomic.rmw.i32(i32, i32*, i32, i32) 191 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32* %ptr, i32 %val, i32 6)
|
/external/swiftshader/third_party/subzero/src/ |
D | IceClFlags.def | 374 clEnumValN(Ice::IceV_RMW, "rmw", "ReadModifyWrite optimization"), \
|
/external/llvm/lib/Target/AVR/ |
D | AVR.td | 122 def FeatureRMW : SubtargetFeature<"rmw", "m_supportsRMW", "true",
|
/external/llvm/docs/ |
D | Atomics.rst | 461 * atomic rmw -> loop with cmpxchg or load-linked/store-conditional
|