/external/boringssl/linux-x86_64/crypto/sha/ |
D | sha512-x86_64.S | 24 pushq %r12 54 movq 0(%rsi),%r12 57 bswapq %r12 65 movq %r12,0(%rsp) 70 addq %r11,%r12 75 addq %r15,%r12 78 addq (%rbp),%r12 87 addq %r13,%r12 90 addq %r12,%rdx 91 addq %r12,%r11 [all …]
|
/external/boringssl/mac-x86_64/crypto/sha/ |
D | sha512-x86_64.S | 23 pushq %r12 53 movq 0(%rsi),%r12 56 bswapq %r12 64 movq %r12,0(%rsp) 69 addq %r11,%r12 74 addq %r15,%r12 77 addq (%rbp),%r12 86 addq %r13,%r12 89 addq %r12,%rdx 90 addq %r12,%r11 [all …]
|
/external/boringssl/win-x86_64/crypto/sha/ |
D | sha512-x86_64.asm | 35 push r12 65 mov r12,QWORD[rsi] 68 bswap r12 76 mov QWORD[rsp],r12 81 add r12,r11 86 add r12,r15 89 add r12,QWORD[rbp] 98 add r12,r13 101 add rdx,r12 102 add r11,r12 [all …]
|
/external/llvm/test/MC/PowerPC/ |
D | ppc64-encoding-6xx.s | 8 mfibatu %r12, 0 11 mfibatl %r12, 0 14 mfibatu %r12, 1 17 mfibatl %r12, 1 20 mfibatu %r12, 2 23 mfibatl %r12, 2 26 mfibatu %r12, 3 29 mfibatl %r12, 3 32 mtibatu 0, %r12 35 mtibatl 0, %r12 [all …]
|
/external/boringssl/src/crypto/bn/asm/ |
D | rsaz-x86_64.pl | 123 push %r12 169 movq %rdx, %r12 170 adcq \$0, %r12 173 addq %rax, %r12 220 addq %rax, %r12 223 addq %rbx, %r12 270 addq %rax, %r12 292 movq %r12, %r10 293 lea (%rbx,%r12,2), %r12 #shld \$1, %rbx, %r12 311 leaq (%r10,%r13,2), %r13 #shld \$1, %r12, %r13 [all …]
|
/external/boringssl/linux-arm/crypto/aes/ |
D | aes-armv4.S | 172 stmdb sp!,{r1,r4-r12,lr} 178 mov r12,r0 @ inp 181 ldrb r0,[r12,#3] @ load input data in endian-neutral 182 ldrb r4,[r12,#2] @ manner... 183 ldrb r5,[r12,#1] 184 ldrb r6,[r12,#0] 186 ldrb r1,[r12,#7] 188 ldrb r4,[r12,#6] 190 ldrb r5,[r12,#5] 191 ldrb r6,[r12,#4] [all …]
|
/external/tremolo/Tremolo/ |
D | mdctLARM.s | 61 LDMDB r2!,{r5,r6,r7,r12} 66 MOV r12,r12,ASR #9 @ r12= (*--r)>>9 68 MOV r14,r12,ASR #15 70 EORNE r12,r4, r14,ASR #31 71 STRH r12,[r0], r3 121 LDR r12,[r2],#8 126 RSB r12,r12,#0 131 MOV r12, r12,ASR #9 @ r12= (-*l)>>9 136 MOV r14,r12,ASR #15 138 EORNE r12,r4, r14,ASR #31 [all …]
|
D | mdctARM.s | 63 LDMDB r2!,{r5,r6,r7,r12} 68 MOV r12,r12,ASR #9 @ r12= (*--r)>>9 70 MOV r14,r12,ASR #15 72 EORNE r12,r4, r14,ASR #31 73 STRH r12,[r0], r3 123 LDR r12,[r2],#8 128 RSB r12,r12,#0 133 MOV r12, r12,ASR #9 @ r12= (-*l)>>9 138 MOV r14,r12,ASR #15 140 EORNE r12,r4, r14,ASR #31 [all …]
|
D | bitwiseARM.s | 46 LDMIA r0,{r2,r3,r12} 49 @ r12= bitsLeftInWord 54 RSB r14,r12,#32 @ r14= 32-bitsLeftInWord 55 SUBS r12,r12,r1 @ r12= bitsLeftInWord -= bits 58 ADDLE r12,r12,#32 @ r12= bitsLeftInWord += 32 71 @ r12 = bitsLeftInWord 72 RSB r14,r12,#32 @ r14= 32-bitsLeftInWord 75 BEQ look_next_segment @ r10= r12 = 0, if we branch 76 CMP r12,r10 @ If bitsLeftInWord < bitsLeftInSeg 82 ORRLT r10,r10,r6,LSL r12 @ r10= first bitsLeftInSeg bits+crap [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_intra_pred_chroma_horz.s | 99 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 105 add r12,r0,r6 @*pu1_ref[four_nt] 115 sub r12,r12,#16 @move to 16th value pointer 119 vld1.16 {q0},[r12] @load 16 values. d1[7] will have the 1st value. 120 sub r12,r12,#16 121 vld1.16 {q5},[r12] @load 16 values. d1[7] will have the 1st value. 177 sub r12,r12,#16 @move to 16th value pointer 190 ldmfd sp!,{r4-r12,r15} @reload the registers from sp 194 ldrb lr,[r12],#1 @pu1_ref[two_nt] 195 @vld1.8 {q15},[r12] @pu1_ref[two_nt + 1 + col] [all …]
|
D | ihevc_inter_pred_chroma_horz.s | 108 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 121 sub r12,r0,#2 @pu1_src - 2 123 add r4,r12,r2 @pu1_src_tmp2_8 = pu1_src + src_strd 152 add r4,r12,r2 154 and r0, r12, #31 156 pld [r12, r2, lsl #1] 161 vld1.u32 {q0},[r12],r11 @vector load pu1_src 163 vld1.u32 {q1},[r12],r11 @vector load pu1_src 165 vld1.u32 {q2},[r12],r11 @vector load pu1_src 167 vld1.u32 {q3},[r12],r9 @vector load pu1_src [all …]
|
D | ihevc_inter_pred_chroma_horz_w16out.s | 107 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 124 sub r12,r0,#2 @pu1_src - 2 126 add r4,r12,r2 @pu1_src_tmp2_8 = pu1_src + src_strd 166 add r4,r12,r2 169 and r0, r12, #31 170 pld [r12, r2, lsl #1] 178 vld1.u32 {q0},[r12],r11 @vector load pu1_src 181 vld1.u32 {q1},[r12],r11 @vector load pu1_src 184 vld1.u32 {q2},[r12],r11 @vector load pu1_src 187 vld1.u32 {q3},[r12],r9 @vector load pu1_src [all …]
|
D | ihevc_inter_pred_luma_horz_w16out.s | 108 @r12 - src_ptr1 124 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 136 sub r12,r0,#3 @pu1_src - 3 138 add r4,r12,r2 @pu1_src_tmp2_8 = pu1_src + src_strd 185 sub r12,r0,#3 @pu1_src - 3 189 add r12,#8 196 add r4,r12,r2 @pu1_src + src_strd 202 vld1.u32 {d0},[r12],r11 @vector load pu1_src 203 vld1.u32 {d1},[r12],r11 204 vld1.u32 {d2},[r12],r11 [all …]
|
D | ihevc_intra_pred_luma_horz.s | 99 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 106 add r12,r0,r6 @*pu1_ref[two_nt] 115 sub r12,r12,#16 @move to 16th value pointer 119 vld1.8 {q0},[r12] @load 16 values. d1[7] will have the 1st value. 175 sub r12,r12,#16 @move to 16th value pointer 188 ldmfd sp!,{r4-r12,r15} @reload the registers from sp 192 ldrb lr,[r12],#1 @pu1_ref[two_nt] 193 vld1.8 {q15},[r12] @pu1_ref[two_nt + 1 + col] 196 sub r12,r12,#17 197 vld1.8 {q0},[r12] [all …]
|
D | ihevc_sao_edge_offset_class0_chroma.s | 75 STMFD sp!, {r4-r12, r14} @stack stores the values of the arguments 83 LDRH r12,[r11,#-2] @pu1_src_top[wd - 1] 87 STRH r12,[r4] @*pu1_src_top_left = pu1_src_top[wd - 1] 105 MOV r12,r9 @Move wd to r12 for loop count 109 SUBS r12,r12,#8 @Decrement the loop counter by 8 125 LDRB r12,[r7] @pu1_avail[0] 126 VMOV.8 D8[0],r12 @vsetq_lane_s8(pu1_avail[0], au1_mask, 0) 127 VMOV.8 D8[1],r12 @vsetq_lane_s8(pu1_avail[0], au1_mask, 1) 131 MOV r12,#-1 @move -1 to r12 132 VMOV.16 D8[0],r12 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0) [all …]
|
D | ihevc_inter_pred_filters_luma_horz.s | 118 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 139 sub r12,r0,#3 @pu1_src - 3 141 add r4,r12,r2 @pu1_src_tmp2_8 = pu1_src + src_strd 173 sub r12,r0,#3 @pu1_src - 3 177 add r12,#16 185 add r4,r12,r2 @pu1_src + src_strd 191 vld1.u32 {d0},[r12],r11 @vector load pu1_src 192 vld1.u32 {d1},[r12],r11 193 vld1.u32 {d2},[r12],r11 194 vld1.u32 {d3},[r12],r11 [all …]
|
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
D | loopfilter_v6.asm | 88 ldr r12, [src], pstep ; p0 99 uqsub8 r6, r11, r12 ; p1 - p0 101 uqsub8 r7, r12, r11 ; p0 - p1 111 uqsub8 r11, r12, r9 ; p0 - q0 112 uqsub8 r12, r9, r12 ; q0 - p0 115 orr r12, r11, r12 ; abs (p0-q0) 117 uqadd8 r12, r12, r12 ; abs (p0-q0) * 2 120 uqadd8 r12, r12, r6 ; abs (p0-q0)*2 + abs (p1-q1)/2 122 uqsub8 r12, r12, r4 ; compare to flimit 125 orr lr, lr, r12 [all …]
|
D | dequant_idct_v6.asm | 29 mov r12, #4 44 subs r12, r12, #1 63 ldr r12, [r0, #24] 70 smulwt r11, r3, r12 73 smulwt r7, r4, r12 74 smulwb r9, r3, r12 75 smulwb r10, r4, r12 80 uadd16 r7, r12, r9 104 smulwt r12, r4, r6 109 pkhbt r12, r10, r12, lsl #16 [all …]
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | idct16x16_1_add_neon.asm | 29 mov r12, #0x2d00 30 add r12, #0x41 33 mul r0, r0, r12 ; input[0] * cospi_16_64 38 mul r0, r0, r12 ; out * cospi_16_64 39 mov r12, r1 ; save dest 69 vst1.64 {d2}, [r12], r0 70 vst1.64 {d3}, [r12], r2 71 vst1.64 {d30}, [r12], r0 72 vst1.64 {d31}, [r12], r2 82 vst1.64 {d2}, [r12], r0 [all …]
|
/external/llvm/test/MC/X86/ |
D | x86_64-bmi-encoding.s | 69 bextrq %r12, (%rax), %r10 73 bextrq %r12, %r11, %r10 85 bzhiq %r12, (%rax), %r10 89 bzhiq %r12, %r11, %r10 101 pextq %r12, %r11, %r10 117 pdepq %r12, %r11, %r10 133 mulxq %r12, %r11, %r10 149 rorxq $1, %r12, %r10 165 shlxq %r12, (%rax), %r10 169 shlxq %r12, %r11, %r10 [all …]
|
/external/boringssl/mac-x86_64/crypto/bn/ |
D | rsaz-x86_64.S | 13 pushq %r12 51 movq %rdx,%r12 52 adcq $0,%r12 55 addq %rax,%r12 102 addq %rax,%r12 105 addq %rbx,%r12 152 addq %rax,%r12 174 movq %r12,%r10 175 leaq (%rbx,%r12,2),%r12 197 adcq %rdx,%r12 [all …]
|
/external/boringssl/linux-x86_64/crypto/bn/ |
D | rsaz-x86_64.S | 14 pushq %r12 52 movq %rdx,%r12 53 adcq $0,%r12 56 addq %rax,%r12 103 addq %rax,%r12 106 addq %rbx,%r12 153 addq %rax,%r12 175 movq %r12,%r10 176 leaq (%rbx,%r12,2),%r12 198 adcq %rdx,%r12 [all …]
|
/external/boringssl/linux-x86_64/crypto/ec/ |
D | p256-x86_64-asm.S | 23 pushq %r12 43 movq %r11,%r12 52 cmovzq %r12,%r11 57 popq %r12 68 pushq %r12 90 movq %r11,%r12 99 cmovzq %r12,%r11 104 popq %r12 121 pushq %r12 130 movq 24(%rsi),%r12 [all …]
|
/external/boringssl/mac-x86_64/crypto/ec/ |
D | p256-x86_64-asm.S | 22 pushq %r12 42 movq %r11,%r12 51 cmovzq %r12,%r11 56 popq %r12 67 pushq %r12 89 movq %r11,%r12 98 cmovzq %r12,%r11 103 popq %r12 120 pushq %r12 129 movq 24(%rsi),%r12 [all …]
|
/external/libavc/encoder/arm/ |
D | ih264e_fmt_conv.s | 69 stmfd sp!, {r4-r12, lr} 171 ldmfd sp!, {r4-r12, pc} 243 stmfd sp!, {r4-r12, lr} @// Back the register which are used 258 sub r12, r4, r7 @// u2_offset1 = u4_stride_y - u4_width 271 add r4, r12, r4 @// u2_offset1 = u2_offset1 + u4_stride_y 290 mov r12, r7 @// Inner loop count = u4_width / 16 295 sub r12, r12, #16 305 cmp r12, #15 307 cmp r12, #0 314 rsb r12, r12, #16 [all …]
|