/third_party/openh264/codec/encoder/core/arm/ |
D | intra_pred_neon.S | 54 GET_8BYTE_DATA d0, r3, r2 65 vadd.u16 d0, d0, d1 66 vpaddl.u16 d0, d0 67 vpaddl.u32 d0, d0 70 vrshr.u16 d0, d0, #5 71 vdup.8 q0, d0[0] 76 vst1.8 {d0,d1}, [r0]! 77 vst1.8 {d0,d1}, [r0]! 78 vst1.8 {d0,d1}, [r0]! 79 vst1.8 {d0,d1}, [r0]! [all …]
|
D | pixel_neon.S | 40 vsubl.u8 q8, d0, d2 52 vsubl.u8 q12, d0, d2 126 vld1.64 {d0}, [r0,:64], r1 130 vsubl.u8 q8, d0, d1 189 vabal.u8 q10, d0, d2 202 vld1.64 {d0}, [r0, :64], r1 205 vabal.u8 q10, d0, d1 228 vabdl.u8 q10, d0, d2 256 vadd.u16 d0, d30, d31 257 vpaddl.u16 d0, d0 [all …]
|
D | svc_motion_estimation.S | 38 vld1.64 {d0}, [r0], r1 52 vpadd.i32 d0, d1 53 vpadd.i32 d0, d0 54 vmov r0, r1, d0 66 vpadd.i32 d0, d1 67 vpadd.i32 d0, d0 68 vmov r0, r1, d0 86 vld1.64 {d0}, [r0], r3 100 vpadd.i32 d0, d1 101 vpadd.i32 d0, d0 [all …]
|
/third_party/openh264/codec/decoder/core/arm/ |
D | intra_pred_neon.S | 53 vldm r2, {d0, d1} 59 vst1.8 {d0,d1}, [r2], r1 60 vst1.8 {d0,d1}, [r2], r1 61 vst1.8 {d0,d1}, [r2], r1 62 vst1.8 {d0,d1}, [r2], r1 75 vld1.8 {d0[],d1[]}, [r2], r1 81 vst1.8 {d0,d1}, [r0], r1 96 GET_8BYTE_DATA d0, r2, r1 107 vadd.u16 d0, d0, d1 108 vpaddl.u16 d0, d0 [all …]
|
/third_party/ffmpeg/libavcodec/arm/ |
D | h264pred_neon.S | 53 add16x8 q0, d0, d1, d0, d1 54 vrshrn.u16 d0, q0, #4 55 vdup.8 q0, d0[0] 61 ldcol.8 d0, r2, r1 63 add16x8 q0, d0, d1, d0, d1 64 vrshrn.u16 d0, q0, #4 65 vdup.8 q0, d0[0] 75 vaddl.u8 q0, d0, d1 78 vadd.u16 d0, d0, d1 79 vpadd.u16 d0, d0, d0 [all …]
|
D | sbrdsp_neon.S | 53 vadd.f32 d0, d0, d1 54 vpadd.f32 d0, d0, d0 55 NOVFP vmov.32 r0, d0[0] 82 vld1.32 {d0}, [r0,:64]! 83 vst1.32 {d0}, [r2,:64]! 95 vswp d0, d1 111 vswp d0, d1 136 vst2.32 {d0,d2}, [r0,:128]! 157 vld2.32 {d0,d1}, [r1,:128], r3 158 veor d0, d0, d2 [all …]
|
D | sbcdsp_neon.S | 67 vpadd.s32 d0, d0, d1 70 vrshrn.s32 d0, q0, SBC_PROTO_FIXED_SCALE 74 vdup.i32 d1, d0[1] /* TODO: can be eliminated */ 75 vdup.i32 d0, d0[0] /* TODO: can be eliminated */ 77 vmull.s16 q3, d2, d0 78 vmull.s16 q4, d3, d0 82 vpadd.s32 d0, d6, d7 /* TODO: can be eliminated */ 85 vst1.32 {d0, d1}, [r1, :128] 141 vpadd.s32 d0, d12, d13 148 vmovn.s32 d0, q0 [all …]
|
/third_party/musl/arch/m68k/ |
D | syscall_arch.h | 8 register unsigned long d0 __asm__("d0") = n; in __syscall0() 9 __asm__ __volatile__ ("trap #0" : "+r"(d0) in __syscall0() 12 return d0; in __syscall0() 17 register unsigned long d0 __asm__("d0") = n; in __syscall1() 19 __asm__ __volatile__ ("trap #0" : "+r"(d0) in __syscall1() 22 return d0; in __syscall1() 27 register unsigned long d0 __asm__("d0") = n; in __syscall2() 30 __asm__ __volatile__ ("trap #0" : "+r"(d0) in __syscall2() 33 return d0; in __syscall2() 38 register unsigned long d0 __asm__("d0") = n; in __syscall3() [all …]
|
/third_party/ffmpeg/libswresample/arm/ |
D | resample.S | 30 …vpadd.f32 d0, d0, d1 @ pair adding of the 4x32-bit a… 31 …vpadd.f32 d0, d0, d0 @ pair adding of the 4x32-bit a… 32 vst1.32 {d0[0]}, [r0] @ write accumulator 46 …vpadd.f32 d0, d0, d1 @ pair adding of the 4x32-bit a… 47 …vpadd.f32 d0, d0, d0 @ pair adding of the 4x32-bit a… 48 vst1.32 {d0[0]}, [r0] @ write accumulator 59 …vpadd.s32 d0, d0, d1 @ pair adding of the 4x32-bit a… 60 …vpadd.s32 d0, d0, d0 @ pair adding of the 4x32-bit a… 61 vst1.32 {d0[0]}, [r0] @ write accumulator 73 …vpadd.s32 d0, d0, d1 @ pair adding of the 4x32-bit a… [all …]
|
/third_party/ffmpeg/libavresample/arm/ |
D | resample_neon.S | 103 veor.32 d0, d0 104 vld1.32 {d0[0]}, [r0]! /* load filter */ 108 vld1.32 {d0}, [r0]! /* load filter */ 112 vld1.32 {d0,d1}, [r0]! /* load filter */ 116 vmla.f32 d16, d0, d4[0] 119 vmla.f32 d16, d0, d4 122 vmla.f32 d16, d0, d4 126 vmul.f32 d16, d0, d4 143 veor.32 d0, d0 144 vld1.32 {d0[0]}, [r0]! /* load filter */ [all …]
|
/third_party/libjpeg-turbo/ |
D | jdcol565.c | 115 JLONG d0 = dither_matrix[cinfo->output_scanline & DITHER_MASK]; in LOCAL() local 131 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 134 SCALEBITS)), d0)]; in LOCAL() 135 b = range_limit[DITHER_565_B(y + Cbbtab[cb], d0)]; in LOCAL() 145 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 148 SCALEBITS)), d0)]; in LOCAL() 149 b = range_limit[DITHER_565_B(y + Cbbtab[cb], d0)]; in LOCAL() 150 d0 = DITHER_ROTATE(d0); in LOCAL() 156 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 159 SCALEBITS)), d0)]; in LOCAL() [all …]
|
/third_party/skia/third_party/externals/libjpeg-turbo/ |
D | jdcol565.c | 115 JLONG d0 = dither_matrix[cinfo->output_scanline & DITHER_MASK]; in LOCAL() local 131 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 134 SCALEBITS)), d0)]; in LOCAL() 135 b = range_limit[DITHER_565_B(y + Cbbtab[cb], d0)]; in LOCAL() 145 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 148 SCALEBITS)), d0)]; in LOCAL() 149 b = range_limit[DITHER_565_B(y + Cbbtab[cb], d0)]; in LOCAL() 150 d0 = DITHER_ROTATE(d0); in LOCAL() 156 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 159 SCALEBITS)), d0)]; in LOCAL() [all …]
|
/third_party/flutter/skia/third_party/externals/libjpeg-turbo/ |
D | jdcol565.c | 115 JLONG d0 = dither_matrix[cinfo->output_scanline & DITHER_MASK]; in LOCAL() local 131 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 134 SCALEBITS)), d0)]; in LOCAL() 135 b = range_limit[DITHER_565_B(y + Cbbtab[cb], d0)]; in LOCAL() 145 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 148 SCALEBITS)), d0)]; in LOCAL() 149 b = range_limit[DITHER_565_B(y + Cbbtab[cb], d0)]; in LOCAL() 150 d0 = DITHER_ROTATE(d0); in LOCAL() 156 r = range_limit[DITHER_565_R(y + Crrtab[cr], d0)]; in LOCAL() 159 SCALEBITS)), d0)]; in LOCAL() [all …]
|
/third_party/libffi/src/m68k/ |
D | sysv.S | 91 move.l %d0,%a1 120 move.l %d0,(%a1) 126 move.l %d0,(%a1) 136 move.l %d0,(%a1) 146 move.l %d0,(%a1)+ 157 move.l %d0,(%a1)+ 167 move.l %d0,(%a1) 176 move.b %d0,(%a1) 182 move.w %d0,(%a1) 190 ext.w %d0 [all …]
|
/third_party/optimized-routines/string/arm/ |
D | memcpy.S | 88 vstr d0, [dst, #\base + 8] 89 vldr d0, [src, #\base + 8] 96 vstr d0, [dst, #\base + 40] 97 vldr d0, [src, #\base + 40] 107 vstr d0, [dst, #\base + 8] 108 vldr d0, [src, #\base + 8] 114 vstr d0, [dst, #\base + 40] 115 vldr d0, [src, #\base + 40] 136 vld1.8 {d0}, [src]! /* 14 words to go. */ 137 vst1.8 {d0}, [dst]! [all …]
|
/third_party/openssl/crypto/des/ |
D | cfb_enc.c | 28 register DES_LONG d0, d1, v0, v1; in DES_cfb_encrypt() local 57 c2ln(in, d0, d1, n); in DES_cfb_encrypt() 59 d0 ^= ti[0]; in DES_cfb_encrypt() 61 l2cn(d0, d1, out, n); in DES_cfb_encrypt() 69 v1 = d0; in DES_cfb_encrypt() 71 v0 = d0; in DES_cfb_encrypt() 78 l2c(d0, iv); in DES_cfb_encrypt() 81 sh[0] = v0, sh[1] = v1, sh[2] = d0, sh[3] = d1; in DES_cfb_encrypt() 104 c2ln(in, d0, d1, n); in DES_cfb_encrypt() 112 v1 = d0; in DES_cfb_encrypt() [all …]
|
/third_party/openssl/crypto/poly1305/asm/ |
D | poly1305-ppc.pl | 92 my ($h0,$h1,$h2,$d0,$d1,$d2, $r0,$r1,$s1, $t0,$t1) = map("r$_",(7..12,27..31)); 107 ld $d0,0($inp) # load key material 112 lwbrx $d0,0,$inp # load key material 118 insrdi $d0,$h0,32,0 127 and $d0,$d0,$h0 130 std $d0,32($ctx) # store key 176 li $d0,4 179 lwbrx $d0,$d0,$inp 183 insrdi $t0,$d0,32,0 192 mulld $d0,$h0,$r0 # h0*r0 [all …]
|
/third_party/cmsis/CMSIS/DSP/Source/ComplexMathFunctions/ |
D | arm_cmplx_dot_prod_q15.c | 168 q15_t a0,b0,c0,d0; in arm_cmplx_dot_prod_q15() local 179 d0 = *pSrcB++; in arm_cmplx_dot_prod_q15() 182 imag_sum += (q31_t)a0 * d0; in arm_cmplx_dot_prod_q15() 183 real_sum -= (q31_t)b0 * d0; in arm_cmplx_dot_prod_q15() 189 d0 = *pSrcB++; in arm_cmplx_dot_prod_q15() 192 imag_sum += (q31_t)a0 * d0; in arm_cmplx_dot_prod_q15() 193 real_sum -= (q31_t)b0 * d0; in arm_cmplx_dot_prod_q15() 199 d0 = *pSrcB++; in arm_cmplx_dot_prod_q15() 202 imag_sum += (q31_t)a0 * d0; in arm_cmplx_dot_prod_q15() 203 real_sum -= (q31_t)b0 * d0; in arm_cmplx_dot_prod_q15() [all …]
|
D | arm_cmplx_dot_prod_f16.c | 199 _Float16 a0,b0,c0,d0; in arm_cmplx_dot_prod_f16() local 211 d0 = *pSrcB++; in arm_cmplx_dot_prod_f16() 214 imag_sum += a0 * d0; in arm_cmplx_dot_prod_f16() 215 real_sum -= b0 * d0; in arm_cmplx_dot_prod_f16() 221 d0 = *pSrcB++; in arm_cmplx_dot_prod_f16() 224 imag_sum += a0 * d0; in arm_cmplx_dot_prod_f16() 225 real_sum -= b0 * d0; in arm_cmplx_dot_prod_f16() 231 d0 = *pSrcB++; in arm_cmplx_dot_prod_f16() 234 imag_sum += a0 * d0; in arm_cmplx_dot_prod_f16() 235 real_sum -= b0 * d0; in arm_cmplx_dot_prod_f16() [all …]
|
D | arm_cmplx_dot_prod_q31.c | 172 q31_t a0,b0,c0,d0; in arm_cmplx_dot_prod_q31() local 184 d0 = *pSrcB++; in arm_cmplx_dot_prod_q31() 187 imag_sum += ((q63_t)a0 * d0) >> 14; in arm_cmplx_dot_prod_q31() 188 real_sum -= ((q63_t)b0 * d0) >> 14; in arm_cmplx_dot_prod_q31() 194 d0 = *pSrcB++; in arm_cmplx_dot_prod_q31() 197 imag_sum += ((q63_t)a0 * d0) >> 14; in arm_cmplx_dot_prod_q31() 198 real_sum -= ((q63_t)b0 * d0) >> 14; in arm_cmplx_dot_prod_q31() 204 d0 = *pSrcB++; in arm_cmplx_dot_prod_q31() 207 imag_sum += ((q63_t)a0 * d0) >> 14; in arm_cmplx_dot_prod_q31() 208 real_sum -= ((q63_t)b0 * d0) >> 14; in arm_cmplx_dot_prod_q31() [all …]
|
/third_party/skia/third_party/externals/libpng/arm/ |
D | filter_neon.S | 70 vadd.u8 d0, d3, d4 71 vadd.u8 d1, d0, d5 74 vst4.32 {d0[0],d1[0],d2[0],d3[0]},[r1,:128]! 90 vadd.u8 d0, d3, d22 92 vadd.u8 d1, d0, d5 95 vst1.32 {d0[0]}, [r1,:32], r2 126 vhadd.u8 d0, d3, d16 127 vadd.u8 d0, d0, d4 128 vhadd.u8 d1, d0, d17 134 vst4.32 {d0[0],d1[0],d2[0],d3[0]},[r1,:128]! [all …]
|
/third_party/libpng/arm/ |
D | filter_neon.S | 70 vadd.u8 d0, d3, d4 71 vadd.u8 d1, d0, d5 74 vst4.32 {d0[0],d1[0],d2[0],d3[0]},[r1,:128]! 90 vadd.u8 d0, d3, d22 92 vadd.u8 d1, d0, d5 95 vst1.32 {d0[0]}, [r1,:32], r2 126 vhadd.u8 d0, d3, d16 127 vadd.u8 d0, d0, d4 128 vhadd.u8 d1, d0, d17 134 vst4.32 {d0[0],d1[0],d2[0],d3[0]},[r1,:128]! [all …]
|
/third_party/flutter/skia/third_party/externals/libpng/arm/ |
D | filter_neon.S | 70 vadd.u8 d0, d3, d4 71 vadd.u8 d1, d0, d5 74 vst4.32 {d0[0],d1[0],d2[0],d3[0]},[r1,:128]! 90 vadd.u8 d0, d3, d22 92 vadd.u8 d1, d0, d5 95 vst1.32 {d0[0]}, [r1,:32], r2 126 vhadd.u8 d0, d3, d16 127 vadd.u8 d0, d0, d4 128 vhadd.u8 d1, d0, d17 134 vst4.32 {d0[0],d1[0],d2[0],d3[0]},[r1,:128]! [all …]
|
/third_party/openssl/crypto/chacha/asm/ |
D | chacha-c64xplus.pl | 124 my ($d0,$d1,$d2,$d3) = (12..15); 136 || XOR @X[$a0],@X[$d0],@X[$d0] 142 || SWAP2 @X[$d0],@X[$d0] 147 || ADD @X[$d0],@X[$c0],@X[$c0] 172 XOR @Y[$a0],@Y[$d0],@Y[$d0] 178 XOR @X[$a0],@X[$d0],@X[$d0] 184 || SWAP2 @Y[$d0],@Y[$d0] 186 ROTL @X[$d0],8,@X[$d0] 190 || ADD @Y[$d0],@Y[$c0],@Y[$c0] 200 ADD @X[$d0],@X[$c0],@X[$c0] [all …]
|
/third_party/openh264/codec/common/arm/ |
D | mc_neon.S | 155 vld1.u8 {d0,d1,d2}, [r0], r1 //only use 21(16+5); q0=src[-2] 165 FILTER_6TAG_8BITS d0, d4, d6, d16, d18, d20, d2, q14, q15 187 vld1.u8 {d0,d1}, [r0], r1 //only use 13(8+5); q0=src[-2] 190 vext.8 d2, d0, d1, #1 //d2=src[-1] 191 vext.8 d3, d0, d1, #2 //d3=src[0] 192 vext.8 d4, d0, d1, #3 //d4=src[1] 193 vext.8 d5, d0, d1, #4 //d5=src[2] 194 vext.8 d6, d0, d1, #5 //d6=src[3] 196 FILTER_6TAG_8BITS d0, d2, d3, d4, d5, d6, d1, q14, q15 216 vld1.u8 {d0, d1}, [r0], r1 //only use 9(4+5);d0: 1st row src[-2:5] [all …]
|