/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/SystemZ/ |
D | int-sub-06.ll | 7 define void @f1(i128 *%aptr, i32 %b) { 12 %a = load i128, i128 *%aptr 16 store i128 %sub, i128 *%aptr 21 define void @f2(i128 *%aptr, i64 %b) { 26 %a = load i128, i128 *%aptr 31 store i128 %sub, i128 *%aptr 37 define void @f3(i128 *%aptr, i64 %b) { 42 %a = load i128, i128 *%aptr 47 store i128 %sub, i128 *%aptr 52 define void @f4(i128 *%aptr, i32 *%bsrc) { [all …]
|
D | int-add-10.ll | 7 define void @f1(i128 *%aptr, i32 %b) { 12 %a = load i128, i128 *%aptr 16 store i128 %add, i128 *%aptr 21 define void @f2(i128 *%aptr, i64 %b) { 26 %a = load i128, i128 *%aptr 31 store i128 %add, i128 *%aptr 37 define void @f3(i128 *%aptr, i64 %b) { 42 %a = load i128, i128 *%aptr 47 store i128 %add, i128 *%aptr 52 define void @f4(i128 *%aptr, i32 *%bsrc) { [all …]
|
D | int-add-09.ll | 7 define void @f1(i128 *%aptr) { 12 %a = load i128, i128 *%aptr 15 store i128 %add, i128 *%aptr 20 define void @f2(i128 *%aptr) { 25 %a = load i128, i128 *%aptr 28 store i128 %add, i128 *%aptr 33 define void @f3(i128 *%aptr) { 38 %a = load i128, i128 *%aptr 41 store i128 %add, i128 *%aptr 46 define void @f4(i128 *%aptr) { [all …]
|
D | int-add-08.ll | 22 define void @f2(i128 *%aptr, i64 %addr) { 28 %a = load volatile i128, i128 *%aptr 31 store i128 %add, i128 *%aptr 36 define void @f3(i128 *%aptr, i64 %base) { 43 %a = load volatile i128, i128 *%aptr 46 store i128 %add, i128 *%aptr 51 define void @f4(i128 *%aptr, i64 %base) { 60 %a = load volatile i128, i128 *%aptr 63 store i128 %add, i128 *%aptr 70 define void @f5(i128 *%aptr, i64 %base) { [all …]
|
D | int-sub-05.ll | 31 %aptr = getelementptr i128, i128 *%bptr, i64 -8 32 %a = load i128, i128 *%aptr 35 store i128 %sub, i128 *%aptr 47 %aptr = getelementptr i128, i128 *%bptr, i64 -8 48 %a = load i128, i128 *%aptr 51 store i128 %sub, i128 *%aptr 65 %aptr = getelementptr i128, i128 *%bptr, i64 -8 66 %a = load i128, i128 *%aptr 69 store i128 %sub, i128 *%aptr 83 %aptr = getelementptr i128, i128 *%bptr, i64 -8 [all …]
|
/external/llvm/test/CodeGen/SystemZ/ |
D | int-add-10.ll | 7 define void @f1(i128 *%aptr, i32 %b) { 12 %a = load i128 , i128 *%aptr 16 store i128 %add, i128 *%aptr 21 define void @f2(i128 *%aptr, i64 %b) { 26 %a = load i128 , i128 *%aptr 31 store i128 %add, i128 *%aptr 37 define void @f3(i128 *%aptr, i64 %b) { 42 %a = load i128 , i128 *%aptr 47 store i128 %add, i128 *%aptr 52 define void @f4(i128 *%aptr, i32 *%bsrc) { [all …]
|
D | int-sub-06.ll | 7 define void @f1(i128 *%aptr, i32 %b) { 12 %a = load i128 , i128 *%aptr 16 store i128 %sub, i128 *%aptr 21 define void @f2(i128 *%aptr, i64 %b) { 26 %a = load i128 , i128 *%aptr 31 store i128 %sub, i128 *%aptr 37 define void @f3(i128 *%aptr, i64 %b) { 42 %a = load i128 , i128 *%aptr 47 store i128 %sub, i128 *%aptr 52 define void @f4(i128 *%aptr, i32 *%bsrc) { [all …]
|
D | int-add-09.ll | 7 define void @f1(i128 *%aptr) { 12 %a = load i128 , i128 *%aptr 15 store i128 %add, i128 *%aptr 20 define void @f2(i128 *%aptr) { 25 %a = load i128 , i128 *%aptr 28 store i128 %add, i128 *%aptr 33 define void @f3(i128 *%aptr) { 38 %a = load i128 , i128 *%aptr 41 store i128 %add, i128 *%aptr 46 define void @f4(i128 *%aptr) { [all …]
|
D | int-add-08.ll | 22 define void @f2(i128 *%aptr, i64 %addr) { 28 %a = load volatile i128 , i128 *%aptr 31 store i128 %add, i128 *%aptr 36 define void @f3(i128 *%aptr, i64 %base) { 43 %a = load volatile i128 , i128 *%aptr 46 store i128 %add, i128 *%aptr 51 define void @f4(i128 *%aptr, i64 %base) { 60 %a = load volatile i128 , i128 *%aptr 63 store i128 %add, i128 *%aptr 70 define void @f5(i128 *%aptr, i64 %base) { [all …]
|
D | int-sub-05.ll | 31 %aptr = getelementptr i128, i128 *%bptr, i64 -8 32 %a = load i128 , i128 *%aptr 35 store i128 %sub, i128 *%aptr 47 %aptr = getelementptr i128, i128 *%bptr, i64 -8 48 %a = load i128 , i128 *%aptr 51 store i128 %sub, i128 *%aptr 65 %aptr = getelementptr i128, i128 *%bptr, i64 -8 66 %a = load i128 , i128 *%aptr 69 store i128 %sub, i128 *%aptr 83 %aptr = getelementptr i128, i128 *%bptr, i64 -8 [all …]
|
/external/webrtc/webrtc/common_audio/signal_processing/ |
D | refl_coef_to_lpc.c | 23 int16_t *aptr, *aptr2, *anyptr; in WebRtcSpl_ReflCoefToLpc() local 35 aptr = a; in WebRtcSpl_ReflCoefToLpc() 36 aptr++; in WebRtcSpl_ReflCoefToLpc() 44 *anyptr = *aptr + (int16_t)((*aptr2 * *kptr) >> 15); in WebRtcSpl_ReflCoefToLpc() 46 aptr++; in WebRtcSpl_ReflCoefToLpc() 50 aptr = a; in WebRtcSpl_ReflCoefToLpc() 54 *aptr = *anyptr; in WebRtcSpl_ReflCoefToLpc() 55 aptr++; in WebRtcSpl_ReflCoefToLpc()
|
/external/boringssl/src/crypto/fipsmodule/bn/asm/ |
D | x86_64-mont5.pl | 1072 my $aptr="%rsi"; # const BN_ULONG *aptr, 1133 sub %r11,%rbp # align with $aptr 1197 mov $aptr,$rptr 1306 lea ($aptr,$num),$aptr # end of a[] buffer, ($aptr,$i)=&ap[2] 1311 mov -32($aptr,$i),$a0 # a[0] 1313 mov -24($aptr,$i),%rax # a[1] 1315 mov -16($aptr,$i),$ai # a[2] 1332 mov -8($aptr,$i),$ai # a[3] 1351 mov ($aptr,$j),$ai # a[4] 1361 mov 8($aptr,$j),$ai # a[5] [all …]
|
D | armv4-mont.pl | 304 my ($rptr,$aptr,$bptr,$nptr,$n0,$num)=map("r$_",(0..5)); 329 vld1.32 {$A0-$A3}, [$aptr]! @ can't specify :32 :-( 462 vld1.32 {$A0-$A3},[$aptr]! 532 vld1.32 {$A0-$A3},[$aptr]! 600 subeq $aptr,$aptr,$num,lsl#2 @ rewind 604 vld1.32 {$A0-$A3},[$aptr]! 682 subs $aptr,sp,#0 @ clear carry flag 686 ldmia $aptr!, {r4-r7} 692 teq $aptr,$bptr @ preserves carry 696 ldr r10, [$aptr] @ load top-most bit [all …]
|
D | x86_64-mont.pl | 810 my $aptr="%rsi"; # const BN_ULONG *aptr, 854 # ensure that stack frame doesn't alias with $aptr modulo 861 sub $aptr,%r11 865 sub %r11,%rbp # align with $aptr 1097 my ($aptr, $bptr, $nptr, $tptr, $mi, $bi, $zero, $num)= 1106 mulx 0*8($aptr),$mi,%rax # a[0]*b[0] 1107 mulx 1*8($aptr),%r11,%r14 # a[1]*b[0] 1110 mulx 2*8($aptr),%r12,%r13 # ... 1118 mulx 3*8($aptr),%rax,%r14 1120 lea 4*8($aptr),$aptr [all …]
|
/external/libcxx/test/std/utilities/smartptr/unique.ptr/unique.ptr.class/unique.ptr.asgn/ |
D | move_convert.single.pass.cpp | 26 void testAssign(APtr& aptr, BPtr& bptr) { in testAssign() argument 29 aptr = std::move(bptr); in testAssign() 30 assert(aptr.get() == p); in testAssign() 121 std::unique_ptr<A> aptr(new A); in main() local 122 testAssign(aptr, bptr); in main() 129 std::unique_ptr<A, Deleter<A> > aptr(new A); in main() local 130 testAssign(aptr, bptr); in main() 131 checkDeleter(aptr, bptr, 42, 0); in main() 139 std::unique_ptr<A, CDeleter<A>&> aptr(new A, adel); in main() local 140 testAssign(aptr, bptr); in main() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | clamp.ll | 8 define amdgpu_kernel void @v_clamp_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #0 { 10 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 23 define amdgpu_kernel void @v_clamp_neg_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #0 { 25 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 39 define amdgpu_kernel void @v_clamp_negabs_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) … 41 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 58 define amdgpu_kernel void @v_clamp_negzero_f32(float addrspace(1)* %out, float addrspace(1)* %aptr)… 60 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 74 …pu_kernel void @v_clamp_multi_use_max_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #0 { 76 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid [all …]
|
D | and.ll | 98 define amdgpu_kernel void @v_and_i32_vgpr_vgpr(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32… 100 %gep.a = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 128 define amdgpu_kernel void @v_and_i32_vgpr_sgpr(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32… 130 %gep.a = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 140 define amdgpu_kernel void @v_and_constant_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr) { 142 %gep = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 151 define amdgpu_kernel void @v_and_inline_imm_64_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr)… 153 %gep = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 162 … amdgpu_kernel void @v_and_inline_imm_neg_16_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr) { 164 %gep = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid [all …]
|
D | clamp-modifier.ll | 9 define amdgpu_kernel void @v_clamp_add_src_f32(float addrspace(1)* %out, float addrspace(1)* %aptr)… 11 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 25 …pu_kernel void @v_clamp_multi_use_src_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #0 { 27 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 42 …dgpu_kernel void @v_clamp_dbg_use_src_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #0 { 44 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 59 …dgpu_kernel void @v_clamp_add_neg_src_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #0 { 61 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 76 define amdgpu_kernel void @v_non_clamp_max_f32(float addrspace(1)* %out, float addrspace(1)* %aptr)… 78 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid [all …]
|
D | fmed3.ll | 12 …l void @v_test_nnan_input_fmed3_r_i_i_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #1 { 14 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 30 …mdgpu_kernel void @v_test_fmed3_r_i_i_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #1 { 32 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 48 …nel void @v_test_fmed3_r_i_i_commute0_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #1 { 50 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 66 …nel void @v_test_fmed3_r_i_i_commute1_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #1 { 68 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 82 …id @v_test_fmed3_r_i_i_constant_order_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #1 { 84 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid [all …]
|
D | min3.ll | 7 define amdgpu_kernel void @v_test_imin3_slt_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i3… 9 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 26 define amdgpu_kernel void @v_test_umin3_ult_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i3… 28 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 46 define amdgpu_kernel void @v_test_umin_umin_umin(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i… 49 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 53 %gep3 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid2 80 define amdgpu_kernel void @v_test_umin3_2_uses(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32… 83 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 87 %gep3 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid2 [all …]
|
D | spill-offset-calculation.ll | 15 %aptr = getelementptr i32, i32 addrspace(5)* %buf, i32 1 17 %a = load volatile i32, i32 addrspace(5)* %aptr 36 %aptr = getelementptr i32, i32 addrspace(5)* %buf, i32 1 41 %a = load volatile i32, i32 addrspace(5)* %aptr 64 %aptr = getelementptr <2 x i32>, <2 x i32> addrspace(5)* %bufv2, i32 1 65 %a = load volatile <2 x i32>, <2 x i32> addrspace(5)* %aptr 95 %aptr = getelementptr <2 x i32>, <2 x i32> addrspace(5)* %bufv2, i32 1 96 %a = load volatile <2 x i32>, <2 x i32> addrspace(5)* %aptr 119 %aptr = getelementptr i32, i32 addrspace(5)* %buf, i32 1 121 %a = load volatile i32, i32 addrspace(5)* %aptr [all …]
|
D | max.i16.ll | 7 define amdgpu_kernel void @v_test_imax_sge_i16(i16 addrspace(1)* %out, i16 addrspace(1)* %aptr, i16… 9 %gep0 = getelementptr i16, i16 addrspace(1)* %aptr, i32 %tid 26 …est_imax_sge_v2i16(<2 x i16> addrspace(1)* %out, <2 x i16> addrspace(1)* %aptr, <2 x i16> addrspac… 28 %gep0 = getelementptr <2 x i16>, <2 x i16> addrspace(1)* %aptr, i32 %tid 48 …est_imax_sge_v3i16(<3 x i16> addrspace(1)* %out, <3 x i16> addrspace(1)* %aptr, <3 x i16> addrspac… 50 %gep0 = getelementptr <3 x i16>, <3 x i16> addrspace(1)* %aptr, i32 %tid 70 …est_imax_sge_v4i16(<4 x i16> addrspace(1)* %out, <4 x i16> addrspace(1)* %aptr, <4 x i16> addrspac… 72 %gep0 = getelementptr <4 x i16>, <4 x i16> addrspace(1)* %aptr, i32 %tid 86 define amdgpu_kernel void @v_test_imax_sgt_i16(i16 addrspace(1)* %out, i16 addrspace(1)* %aptr, i16… 88 %gep0 = getelementptr i16, i16 addrspace(1)* %aptr, i32 %tid [all …]
|
/external/webrtc/webrtc/modules/audio_coding/codecs/cng/ |
D | cng_helpfuns.c | 20 int16_t *aptr, *aptr2, *anyptr; in WebRtcCng_K2a16() local 30 aptr = a; in WebRtcCng_K2a16() 31 aptr++; in WebRtcCng_K2a16() 38 *anyptr++ = (*aptr++) + in WebRtcCng_K2a16() 42 aptr = a; in WebRtcCng_K2a16() 45 *aptr++ = *anyptr++; in WebRtcCng_K2a16()
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | and.ll | 96 define void @v_and_i32_vgpr_vgpr(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 98 %gep.a = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 126 define void @v_and_i32_vgpr_sgpr(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 %b) { 128 %gep.a = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 138 define void @v_and_constant_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr) { 139 %a = load i32, i32 addrspace(1)* %aptr, align 4 147 define void @v_and_inline_imm_64_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr) { 148 %a = load i32, i32 addrspace(1)* %aptr, align 4 156 define void @v_and_inline_imm_neg_16_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr) { 157 %a = load i32, i32 addrspace(1)* %aptr, align 4 [all …]
|
D | fmed3.ll | 15 define void @v_test_fmed3_r_i_i_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #1 { 17 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 33 define void @v_test_fmed3_r_i_i_commute0_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #… 35 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 51 define void @v_test_fmed3_r_i_i_commute1_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #… 53 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 67 …id @v_test_fmed3_r_i_i_constant_order_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) #1 { 69 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid 84 define void @v_test_fmed3_r_i_i_multi_use_f32(float addrspace(1)* %out, float addrspace(1)* %aptr) … 86 %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid [all …]
|