/external/openssl/crypto/sha/asm/ |
D | sha1-armv4-large.s | 20 ldrb r10,[r1,#-1] 25 orr r9,r10,r9,lsl#24 26 eor r10,r5,r6 @ F_xx_xx 29 and r10,r4,r10,ror#2 31 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 33 add r7,r7,r10 @ E+=F_00_19(B,C,D) 35 ldrb r10,[r1,#-1] 40 orr r9,r10,r9,lsl#24 41 eor r10,r4,r5 @ F_xx_xx 44 and r10,r3,r10,ror#2 [all …]
|
D | sha512-armv4.s | 60 ldr r10, [r0,#48+0] 65 str r10, [sp,#48+4] 73 ldr r10, [r0,#16+0] 79 str r10, [sp,#16+4] 90 ldrb r10, [r1,#5] 96 orr r3,r3,r10,lsl#16 97 ldrb r10, [r1],#8 101 orr r4,r4,r10,lsl#24 110 mov r10,r8,lsr#14 112 eor r10,r10,r7,lsl#18 [all …]
|
D | sha256-armv4.s | 31 ldmia r0,{r4,r5,r6,r7,r8,r9,r10,r11} 47 eor r2,r9,r10 51 eor r2,r2,r10 @ Ch(e,f,g) 82 add r3,r3,r10 83 mov r10,r11,ror#2 85 eor r10,r10,r11,ror#13 87 eor r10,r10,r11,ror#22 @ Sigma0(a) 91 add r10,r10,r3 94 add r10,r10,r0 113 mov r9,r10,ror#2 [all …]
|
/external/libvpx/vp8/common/arm/armv6/ |
D | loopfilter_v6.asm | 76 ldr r10, [src], pstep ; p2 89 uqsub8 r6, r9, r10 ; p3 - p2 90 uqsub8 r7, r10, r9 ; p2 - p3 91 uqsub8 r8, r10, r11 ; p2 - p1 92 uqsub8 r10, r11, r10 ; p1 - p2 95 orr r8, r8, r10 ; abs (p2-p1) 102 ldr r10, [src], pstep ; q1 108 uqsub8 r6, r11, r10 ; p1 - q1 109 uqsub8 r7, r10, r11 ; q1 - p1 118 uqsub8 r7, r9, r10 ; q0 - q1 [all …]
|
D | filter_v6.asm | 59 ldrb r10, [r0], #2 66 pkhbt r8, r9, r10, lsl #16 ; r10 | r9 71 pkhbt r10, r10, r11, lsl #16 ; r11 | r10 75 smlad lr, r10, r5, lr 76 ldrb r10, [r0, #1] 82 pkhbt r9, r9, r10, lsl #16 ; r10 | r9 83 pkhbt r10, r10, r11, lsl #16 ; r11 | r10 86 smlad r11, r10, r6, r8 88 ands r10, r7, #0xff ; test loop counter 99 ldrneb r10, [r0], #2 [all …]
|
D | idct_v6.asm | 12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14 71 …smulwb r10, r5, r12 ; ([4] * sinpi8sqrt2) >> 16 1, r10 un 2, r12/r5 ^1 … 76 rsb r11, r11, r10 ; c1 = temp1 - temp2 1 c1 78 …smulwb r10, r4, r12 ; ([4] * cospi8sqrt2minus1) >> 16 1, r10 un 2, r12/r4 ^1 … 82 add r10, r12, r10 ; temp1 1 83 add r3, r10, r3 ; d1 = temp1 + temp2 1 d1 84 add r10, r9, r3 ; a1 + d1 1 a+d 91 strh r10, [r1], #0x2 ; out[0] = a1+d1 1 ++ 104 …smulwb r10, r4, r8 ; ([3] * cospi8sqrt2minus1) >> 16 1, r10 un 2, r4/r8 ^1 … 107 add r10, r8, r10 ; temp2 1 [all …]
|
D | iwalsh_v6.asm | 34 qadd16 r10, r2, r8 ; a1 [1+13 | 0+12] 39 qadd16 r2, r10, r11 ; a1 + b1 [1 | 0] 41 qsub16 r6, r10, r11 ; a1 - b1 [9 | 8] 44 qadd16 r10, r3, r9 ; a1 [3+15 | 2+14] 49 qadd16 r3, r10, r11 ; a1 + b1 [3 | 2] 51 qsub16 r7, r10, r11 ; a1 - b1 [11 | 10] 56 qsubaddx r10, r2, r3 ; [c1|a1] [1-2 | 0+3] 61 qaddsubx r2, r10, r11 ; [b2|c2] [c1+d1 | a1-b1] 62 qaddsubx r3, r11, r10 ; [a2|d2] [b1+a1 | d1-c1] 63 ldr r10, c0x00030003 [all …]
|
D | sixtappredict8x4_v6.asm | 54 ldrb r10, [r0, #-1] 62 pkhbt r9, r9, r10, lsl #16 ; r10 | r9 74 pkhbt r10, r10, r6, lsl #16 ; r10 | r9 75 pkhbt r6, r6, r7, lsl #16 ; r11 | r10 76 smlad r11, r10, r5, r11 ; vp8_filter[4], vp8_filter[5] 95 movne r8, r10 97 movne r10, r12 144 smulbt r10, r3, r6 149 smladx r10, r12, r7, r10 154 smladx r10, r11, r8, r10 [all …]
|
/external/tremolo/Tremolo/ |
D | bitwiseARM.s | 45 STMFD r13!,{r10,r11,r14} 53 LDR r10,[r3] @ r10= ptr[0] 57 MOV r10,r10,LSR r14 @ r10= ptr[0]>>(32-bitsLeftInWord) 60 ORRLT r10,r10,r11,LSL r14 @ r10= Next 32 bits. 63 AND r0,r10,r14 64 LDMFD r13!,{r10,r11,PC} 68 ADDS r10,r2,r1 @ r10= bitsLeftInSegment + bits (i.e. 70 @ r10 = bitsLeftInSegment (initial) 73 MOV r5,r10 @ r5 = bitsLeftInSegment (initial) 75 BEQ look_next_segment @ r10= r12 = 0, if we branch [all …]
|
D | mdctARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 189 @ r10= wR 198 LDR r6, [r10,#-4]! @ r6 = *--wR 226 LDMFD r12,{r8,r9,r10} @ r8 = step 228 @ r10= wR 237 LDR r6, [r10,#-4]! @ r6 = *--wR 319 LDR r10,[r5],r2,LSL #2 @ r10= T[0] T += step 325 SMLAL r8, r9, r6, r10 @ (r8, r9) += s0*T[0] 328 SMULL r8, r12,r7, r10 @ (r8, r12) = s2*T[0] 341 LDR r10,[r5,#4] @ r10= T[1] [all …]
|
D | mdctLARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 189 @ r10= wR 197 LDRB r6, [r10,#-1]! @ r6 = *--wR 227 LDMFD r12,{r8,r9,r10} @ r8 = step 229 @ r10= wR 238 LDRB r6, [r10,#-1]! @ r6 = *--wR 322 LDRB r10,[r5],r2 @ r10= T[0] T += step 327 MUL r9, r6, r10 @ r9 = s0*T[0] 331 MUL r12,r7, r10 @ r12 = s2*T[0] 340 LDRB r10,[r5,#1] @ r10= T[1] [all …]
|
/external/libffi/src/x86/ |
D | unix64.S | 48 movq (%rsp), %r10 /* Load return address. */ 53 movq %r10, 24(%rax) /* Relocate return address. */ 56 movq %rdi, %r10 /* Save a copy of the register area. */ 61 movq (%r10), %rdi 62 movq 8(%r10), %rsi 63 movq 16(%r10), %rdx 64 movq 24(%r10), %rcx 65 movq 32(%r10), %r8 66 movq 40(%r10), %r9 72 leaq 176(%r10), %rsp [all …]
|
D | darwin64.S | 48 movq (%rsp), %r10 /* Load return address. */ 53 movq %r10, 24(%rax) /* Relocate return address. */ 56 movq %rdi, %r10 /* Save a copy of the register area. */ 61 movq (%r10), %rdi 62 movq 8(%r10), %rsi 63 movq 16(%r10), %rdx 64 movq 24(%r10), %rcx 65 movq 32(%r10), %r8 66 movq 40(%r10), %r9 72 leaq 176(%r10), %rsp [all …]
|
/external/webrtc/src/common_audio/signal_processing_library/main/source/ |
D | webrtc_fft_4oiq14_gcc_android.s | 24 ldrsh r10, [r12], +r2 29 sub r12, r8, r10 31 add r10, r8, r10 33 sub r9, r4, r10 35 add r4, r4, r10 37 add r10, r6, lr 48 sub r12, r10, r11 50 sub r12, r12, r10, lsl #1 51 mul r10, r12, lr 58 ldrsh r10, [r4], +r2 [all …]
|
D | webrtc_fft_4ofq14_gcc_android.s | 24 ldrsh r10, [r12], +r2 29 sub r12, r8, r10 31 add r10, r8, r10 33 sub r9, r4, r10 35 add r4, r4, r10 37 sub r10, r6, lr 48 sub r12, r11, r10 49 mul r10, r12, lr 58 ldrsh r10, [r4], +r2 69 sub r4, r10, r12 [all …]
|
/external/openssl/crypto/aes/asm/ |
D | aes-armv4.s | 120 sub r10,r3,#AES_encrypt-AES_Te @ Te 207 ldr r4,[r10,r7,lsl#2] @ Te3[s0>>0] 209 ldr r5,[r10,r8,lsl#2] @ Te2[s0>>8] 211 ldr r6,[r10,r9,lsl#2] @ Te1[s0>>16] 213 ldr r0,[r10,r0,lsl#2] @ Te0[s0>>24] 216 ldr r7,[r10,r7,lsl#2] @ Te1[s1>>16] 217 ldr r8,[r10,r8,lsl#2] @ Te3[s1>>0] 218 ldr r9,[r10,r9,lsl#2] @ Te2[s1>>8] 220 ldr r1,[r10,r1,lsl#2] @ Te0[s1>>24] 227 ldr r7,[r10,r7,lsl#2] @ Te2[s2>>8] [all …]
|
/external/libvpx/vp8/decoder/arm/armv6/ |
D | dequant_idct_v6.asm | 70 smulwt r10, r4, r6 74 pkhbt r8, r8, r10, lsl #16 78 smulwb r10, r4, r12 82 pkhbt r10, r10, r7, lsl #16 85 uadd16 r6, r6, r10 86 uadd16 r10, r11, r14 88 uadd16 r9, r10, r6 89 usub16 r10, r10, r6 95 str r10, [r1, #24] 109 smulwt r10, r4, r8 [all …]
|
D | dequant_dc_idct_v6.asm | 92 smulwt r10, r4, r6 96 pkhbt r8, r8, r10, lsl #16 100 smulwb r10, r4, r12 104 pkhbt r10, r10, r7, lsl #16 107 uadd16 r6, r6, r10 108 uadd16 r10, r11, r14 110 uadd16 r9, r10, r6 111 usub16 r10, r10, r6 117 str r10, [r1, #24] 131 smulwt r10, r4, r8 [all …]
|
/external/libvpx/vp8/encoder/arm/armv6/ |
D | vp8_subtract_armv6.asm | 94 uxtb16 r10, r6, ror #8 ; [s3 | s1] (A) 98 usub16 r7, r10, r11 ; [d3 | d1] (A) 100 ldr r10, [r1, #4] ; src (B) 107 uxtb16 r8, r10 ; [s2 | s0] (B) 111 uxtb16 r10, r10, ror #8 ; [s3 | s1] (B) 115 usub16 r7, r10, r11 ; [d3 | d1] (B) 137 uxtb16 r10, r6, ror #8 ; [s3 | s1] (A) 141 usub16 r7, r10, r11 ; [d3 | d1] (A) 143 ldr r10, [r2, #4] ; src (B) 150 uxtb16 r8, r10 ; [s2 | s0] (B) [all …]
|
D | walsh_v6.asm | 34 qsubaddx r10, r2, r3 ; [c1|a1] [1-2 | 0+3] 39 qaddsubx r2, r10, r11 ; [1 | 2] [c1+d1 | a1-b1] 40 qaddsubx r3, r11, r10 ; [0 | 3] [b1+a1 | d1-c1] 44 qsubaddx r10, r6, r7 ; [c1|a1] [9-10 | 8+11] 49 qaddsubx r6, r10, r11 ; [9 |10] [c1+d1 | a1-b1] 50 qaddsubx r7, r11, r10 ; [8 |11] [b1+a1 | d1-c1] 56 qadd16 r10, r3, r9 ; a1 [0+12 | 3+15] 61 qadd16 r3, r10, r11 ; a2 [a1+b1] [0 | 3] 63 qsub16 r7, r10, r11 ; c2 [a1-b1] [8 |11] 66 qadd16 r10, r2, r8 ; a1 [1+13 | 2+14] [all …]
|
/external/valgrind/main/coregrind/m_mach/ |
D | mach_traps-amd64-darwin.S | 40 movq %rcx, %r10 61 movq %rcx, %r10 71 movq %rcx, %r10 81 movq %rcx, %r10 91 movq %rcx, %r10 101 movq %rcx, %r10 111 movq %rcx, %r10 121 movq %rcx, %r10 131 movq %rcx, %r10
|
/external/libvpx/vp8/common/ppc/ |
D | recon_altivec.asm | 65 stvx v2, 0, r10 ;# 2 rows to dst from buf 66 lwz r0, 0(r10) 72 lwz r0, 4(r10) 74 lwz r0, 8(r10) 76 lwz r0, 12(r10) 95 la r10, -48(r1) ;# buf 110 stw r0, 0(r10) 112 stw r0, 4(r10) 114 stw r0, 8(r10) 116 stw r0, 12(r10) [all …]
|
/external/libvpx/vp8/encoder/arm/armv5te/ |
D | vp8_packtokens_partitions_armv5.asm | 66 ldr r10, [sp, #40] ; ptr 81 str r10, [r0, #vp8_writer_buffer] 112 ldr r10, [sp, #88] ; vp8_coef_tree 133 ldrsb lr, [r10, lr] ; i = vp8_coef_tree[i+bb] 158 mov r10, #0 159 strb r10, [r7, r4] ; w->buffer[x] =(unsigned char)0 169 ldrb r10, [r7, r4] ; w->buffer[x] 170 add r10, r10, #1 171 strb r10, [r7, r4] ; w->buffer[x] + 1 174 ldr r10, [r0, #vp8_writer_buffer] [all …]
|
D | vp8_packtokens_armv5.asm | 60 ldr r10, [sp, #52] ; vp8_coef_tree 81 ldrsb lr, [r10, lr] ; i = vp8_coef_tree[i+bb] 106 mov r10, #0 107 strb r10, [r7, r4] ; w->buffer[x] =(unsigned char)0 117 ldrb r10, [r7, r4] ; w->buffer[x] 118 add r10, r10, #1 119 strb r10, [r7, r4] ; w->buffer[x] + 1 122 ldr r10, [r0, #vp8_writer_buffer] 131 strb r7, [r10, r4] ; w->buffer[w->pos++] 133 ; r10 is used earlier in the loop, but r10 is used as [all …]
|
/external/llvm/test/MC/ARM/ |
D | basic-arm-instructions.s | 190 and r10, r1, #0xf 191 and r10, r1, r6 192 and r10, r1, r6, lsl #10 193 and r10, r1, r6, lsr #10 194 and r10, r1, r6, lsr #10 195 and r10, r1, r6, asr #10 196 and r10, r1, r6, ror #10 201 and r10, r1, r6, rrx 205 and r10, r1 206 and r10, r1, lsl #10 [all …]
|