/external/llvm/test/CodeGen/SystemZ/ |
D | int-add-10.ll | 7 define void @f1(i128 *%aptr, i32 %b) { 12 %a = load i128 , i128 *%aptr 16 store i128 %add, i128 *%aptr 21 define void @f2(i128 *%aptr, i64 %b) { 26 %a = load i128 , i128 *%aptr 31 store i128 %add, i128 *%aptr 37 define void @f3(i128 *%aptr, i64 %b) { 42 %a = load i128 , i128 *%aptr 47 store i128 %add, i128 *%aptr 52 define void @f4(i128 *%aptr, i32 *%bsrc) { [all …]
|
D | int-sub-06.ll | 7 define void @f1(i128 *%aptr, i32 %b) { 12 %a = load i128 , i128 *%aptr 16 store i128 %sub, i128 *%aptr 21 define void @f2(i128 *%aptr, i64 %b) { 26 %a = load i128 , i128 *%aptr 31 store i128 %sub, i128 *%aptr 37 define void @f3(i128 *%aptr, i64 %b) { 42 %a = load i128 , i128 *%aptr 47 store i128 %sub, i128 *%aptr 52 define void @f4(i128 *%aptr, i32 *%bsrc) { [all …]
|
D | int-add-09.ll | 7 define void @f1(i128 *%aptr) { 12 %a = load i128 , i128 *%aptr 15 store i128 %add, i128 *%aptr 20 define void @f2(i128 *%aptr) { 25 %a = load i128 , i128 *%aptr 28 store i128 %add, i128 *%aptr 33 define void @f3(i128 *%aptr) { 38 %a = load i128 , i128 *%aptr 41 store i128 %add, i128 *%aptr 46 define void @f4(i128 *%aptr) { [all …]
|
D | int-add-08.ll | 22 define void @f2(i128 *%aptr, i64 %addr) { 28 %a = load volatile i128 , i128 *%aptr 31 store i128 %add, i128 *%aptr 36 define void @f3(i128 *%aptr, i64 %base) { 43 %a = load volatile i128 , i128 *%aptr 46 store i128 %add, i128 *%aptr 51 define void @f4(i128 *%aptr, i64 %base) { 60 %a = load volatile i128 , i128 *%aptr 63 store i128 %add, i128 *%aptr 70 define void @f5(i128 *%aptr, i64 %base) { [all …]
|
D | int-sub-05.ll | 31 %aptr = getelementptr i128, i128 *%bptr, i64 -8 32 %a = load i128 , i128 *%aptr 35 store i128 %sub, i128 *%aptr 47 %aptr = getelementptr i128, i128 *%bptr, i64 -8 48 %a = load i128 , i128 *%aptr 51 store i128 %sub, i128 *%aptr 65 %aptr = getelementptr i128, i128 *%bptr, i64 -8 66 %a = load i128 , i128 *%aptr 69 store i128 %sub, i128 *%aptr 83 %aptr = getelementptr i128, i128 *%bptr, i64 -8 [all …]
|
D | fp-copysign-01.ll | 81 define void @f7(fp128 *%cptr, fp128 *%aptr, float %bf) { 89 %a = load volatile fp128 , fp128 *%aptr 97 define void @f8(fp128 *%cptr, fp128 *%aptr, double %bd) { 105 %a = load volatile fp128 , fp128 *%aptr 114 define void @f9(fp128 *%cptr, fp128 *%aptr, fp128 *%bptr) { 123 %a = load volatile fp128 , fp128 *%aptr
|
/external/boringssl/src/crypto/bn/asm/ |
D | x86_64-mont5.pl | 403 # ensure that stack frame doesn't alias with $aptr+4*$num 907 my $aptr="%rsi"; # const BN_ULONG *aptr, 953 # ensure that stack frame doesn't alias with $aptr+4*$num 959 sub $aptr,%r11 963 sub %r11,%rsp # align with $aptr 1005 mov $aptr,$rptr 1104 lea ($aptr,$num),$aptr # end of a[] buffer, ($aptr,$i)=&ap[2] 1109 mov -32($aptr,$i),$a0 # a[0] 1111 mov -24($aptr,$i),%rax # a[1] 1113 mov -16($aptr,$i),$ai # a[2] [all …]
|
D | armv4-mont.pl | 272 my ($rptr,$aptr,$bptr,$nptr,$n0,$num)=map("r$_",(0..5)); 291 vld1.32 {$A0-$A3}, [$aptr]! @ can't specify :32 :-( 407 vld1.32 {$A0-$A3}, [$aptr]! 439 sub $aptr,$aptr,$num,lsl#2 @ rewind $aptr 468 vld1.32 {$A0-$A3}, [$aptr]! 497 vld1.32 {$A0-$A3}, [$aptr]! 533 sub $aptr,$aptr,$num,lsl#2 @ rewind $aptr 627 subs $aptr,sp,#0 @ clear carry flag 631 ldmia $aptr!, {r4-r7} 637 teq $aptr,$bptr @ preserves carry [all …]
|
D | x86_64-mont.pl | 733 my $aptr="%rsi"; # const BN_ULONG *aptr, 768 # ensure that stack frame doesn't alias with $aptr modulo 774 sub $aptr,%r11 778 sub %r11,%rsp # align with $aptr 930 my ($aptr, $bptr, $nptr, $tptr, $mi, $bi, $zero, $num)= 939 mulx 0*8($aptr),$mi,%rax # a[0]*b[0] 940 mulx 1*8($aptr),%r11,%r14 # a[1]*b[0] 943 mulx 2*8($aptr),%r12,%r13 # ... 951 mulx 3*8($aptr),%rax,%r14 953 lea 4*8($aptr),$aptr [all …]
|
/external/webrtc/src/common_audio/signal_processing/ |
D | refl_coef_to_lpc.c | 23 WebRtc_Word16 *aptr, *aptr2, *anyptr; in WebRtcSpl_ReflCoefToLpc() local 35 aptr = a; in WebRtcSpl_ReflCoefToLpc() 36 aptr++; in WebRtcSpl_ReflCoefToLpc() 44 *anyptr = (*aptr) in WebRtcSpl_ReflCoefToLpc() 47 aptr++; in WebRtcSpl_ReflCoefToLpc() 51 aptr = a; in WebRtcSpl_ReflCoefToLpc() 55 *aptr = *anyptr; in WebRtcSpl_ReflCoefToLpc() 56 aptr++; in WebRtcSpl_ReflCoefToLpc()
|
/external/llvm/test/CodeGen/R600/ |
D | and.ll | 59 define void @v_and_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)* %bptr) { 60 %a = load i32, i32 addrspace(1)* %aptr, align 4 69 define void @v_and_constant_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr) { 70 %a = load i32, i32 addrspace(1)* %aptr, align 4 78 define void @v_and_inline_imm_64_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr) { 79 %a = load i32, i32 addrspace(1)* %aptr, align 4 87 define void @v_and_inline_imm_neg_16_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr) { 88 %a = load i32, i32 addrspace(1)* %aptr, align 4 122 define void @v_and_i64(i64 addrspace(1)* %out, i64 addrspace(1)* %aptr, i64 addrspace(1)* %bptr) { 123 %a = load i64, i64 addrspace(1)* %aptr, align 8 [all …]
|
D | min3.ll | 7 define void @v_test_imin3_slt_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)… 9 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 26 define void @v_test_umin3_ult_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)… 28 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 46 define void @v_test_umin_umin_umin(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1… 49 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 53 %gep3 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid2 80 define void @v_test_umin3_2_uses(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 83 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 87 %gep3 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid2
|
D | min.ll | 7 define void @v_test_imin_sle_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 9 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 31 define void @v_test_imin_slt_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 33 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 73 define void @v_test_umin_ule_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 75 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 97 define void @v_test_umin_ult_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 99 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 125 …i_use(i32 addrspace(1)* %out0, i1 addrspace(1)* %out1, i32 addrspace(1)* %aptr, i32 addrspace(1)* … 127 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid
|
D | fmax3.f64.ll | 14 define void @test_fmax3_f64(double addrspace(1)* %out, double addrspace(1)* %aptr) nounwind { 15 %bptr = getelementptr double, double addrspace(1)* %aptr, i32 1 16 %cptr = getelementptr double, double addrspace(1)* %aptr, i32 2 17 %a = load double, double addrspace(1)* %aptr, align 8
|
D | llvm.AMDGPU.trig_preop.ll | 12 define void @test_trig_preop_f64(double addrspace(1)* %out, double addrspace(1)* %aptr, i32 addrspa… 13 %a = load double, double addrspace(1)* %aptr, align 8 25 define void @test_trig_preop_f64_imm_segment(double addrspace(1)* %out, double addrspace(1)* %aptr)… 26 %a = load double, double addrspace(1)* %aptr, align 8
|
D | max.ll | 7 define void @v_test_imax_sge_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 9 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 49 define void @v_test_imax_sgt_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 51 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 73 define void @v_test_umax_uge_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 75 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 97 define void @v_test_umax_ugt_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)*… 99 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid
|
D | fmax3.ll | 13 define void @test_fmax3_olt_0(float addrspace(1)* %out, float addrspace(1)* %aptr, float addrspace(… 14 %a = load float, float addrspace(1)* %aptr, align 4 31 define void @test_fmax3_olt_1(float addrspace(1)* %out, float addrspace(1)* %aptr, float addrspace(… 32 %a = load float, float addrspace(1)* %aptr, align 4
|
D | fmin3.ll | 14 define void @test_fmin3_olt_0(float addrspace(1)* %out, float addrspace(1)* %aptr, float addrspace(… 15 %a = load float, float addrspace(1)* %aptr, align 4 32 define void @test_fmin3_olt_1(float addrspace(1)* %out, float addrspace(1)* %aptr, float addrspace(… 33 %a = load float, float addrspace(1)* %aptr, align 4
|
D | max3.ll | 7 define void @v_test_imax3_sgt_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)… 9 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid 26 define void @v_test_umax3_ugt_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %aptr, i32 addrspace(1)… 28 %gep0 = getelementptr i32, i32 addrspace(1)* %aptr, i32 %tid
|
D | select64.ll | 43 define void @v_select_trunc_i64_2(i32 addrspace(1)* %out, i32 %cond, i64 addrspace(1)* %aptr, i64 a… 45 %a = load i64, i64 addrspace(1)* %aptr, align 8 61 define void @v_select_i64_split_imm(i64 addrspace(1)* %out, i32 %cond, i64 addrspace(1)* %aptr, i64… 63 %a = load i64, i64 addrspace(1)* %aptr, align 8
|
D | saddo.ll | 30 …32(i32 addrspace(1)* %out, i1 addrspace(1)* %carryout, i32 addrspace(1)* %aptr, i32 addrspace(1)* … 31 %a = load i32, i32 addrspace(1)* %aptr, align 4 54 …64(i64 addrspace(1)* %out, i1 addrspace(1)* %carryout, i64 addrspace(1)* %aptr, i64 addrspace(1)* … 55 %a = load i64, i64 addrspace(1)* %aptr, align 4
|
/external/libedit/src/ |
D | readline.c | 640 char *tmp, *search = NULL, *aptr; in _history_expand_command() local 647 aptr = NULL; in _history_expand_command() 668 if ((aptr = el_malloc((offs + 1) * sizeof(*aptr))) in _history_expand_command() 671 (void)strncpy(aptr, command, offs); in _history_expand_command() 672 aptr[offs] = '\0'; in _history_expand_command() 683 if (ptr == NULL && aptr == NULL) in _history_expand_command() 687 *result = strdup(aptr ? aptr : ptr); in _history_expand_command() 688 if (aptr) in _history_expand_command() 689 el_free(aptr); in _history_expand_command() 732 tmp = history_arg_extract(start, end, aptr? aptr:ptr); in _history_expand_command() [all …]
|
/external/llvm/test/Transforms/SROA/ |
D | address-spaces.ll | 17 %aptr = bitcast <2 x i64> addrspace(1)* %a to i8 addrspace(1)* 19 call void @llvm.memcpy.p0i8.p1i8.i32(i8* %aaptr, i8 addrspace(1)* %aptr, i32 16, i32 2, i1 false) 31 %aptr = bitcast <2 x i64> addrspace(1)* %a to i8 addrspace(1)* 33 call void @llvm.memcpy.p0i8.p1i8.i32(i8* %aaptr, i8 addrspace(1)* %aptr, i32 16, i32 2, i1 false) 45 %aptr = bitcast <2 x i64>* %a to i8* 47 call void @llvm.memcpy.p0i8.p0i8.i32(i8* %aaptr, i8* %aptr, i32 16, i32 2, i1 false)
|
/external/llvm/test/Transforms/BBVectorize/ |
D | simple-ldstr-ptrs.ll | 21 %aptr = getelementptr inbounds double, double* %ptr, i64 2 23 %av = load double, double* %aptr, align 16 37 ; CHECK: %aptr = getelementptr inbounds double, <2 x double*> %ptr, <2 x i64> <i64 2, i64 3> 38 ; CHECK: %aptr.v.r1 = extractelement <2 x double*> %aptr, i32 0 39 ; CHECK: %aptr.v.r2 = extractelement <2 x double*> %aptr, i32 1 40 ; CHECK: %av = load double, double* %aptr.v.r1, align 16 41 ; CHECK: %av5 = load double, double* %aptr.v.r2, align 16
|
/external/clang/test/CodeGenObjC/ |
D | objc2-write-barrier-2.m | 49 void funct2(AStruct *aptr) { 51 aptr->alfred = 0; 52 **ppptr = aptr->alfred;
|