/external/libvpx/vp8/common/arm/neon/ |
D | loopfiltersimplehorizontaledge_neon.asm | 59 ;vqsub.s8 q2, q7, q6 ; ( qs0 - ps0) 60 vsubl.s8 q2, d14, d12 ; ( qs0 - ps0) 65 ;vmul.i8 q2, q2, q10 ; 3 * ( qs0 - ps0) 66 vadd.s16 q11, q2, q2 ; 3 * ( qs0 - ps0) 71 vadd.s16 q2, q2, q11 74 vaddw.s8 q2, q2, d8 ; vp8_filter + 3 * ( qs0 - ps0) 77 …;vqadd.s8 q4, q4, q2 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs… 78 …vqmovn.s16 d8, q2 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs… 84 vqadd.s8 q2, q4, q10 ; Filter2 = vp8_signed_char_clamp(vp8_filter+3) 86 vshr.s8 q2, q2, #3 ; Filter2 >>= 3 [all …]
|
D | loopfiltersimpleverticaledge_neon.asm | 79 ;vqsub.s8 q2, q5, q4 ; ( qs0 - ps0) 80 vsubl.s8 q2, d8, d10 ; ( qs0 - ps0) 85 …;vmul.i8 q2, q2, q11 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs… 86 vadd.s16 q10, q2, q2 ; 3 * ( qs0 - ps0) 88 vadd.s16 q2, q2, q10 91 ;vqadd.s8 q1, q1, q2 92 vaddw.s8 q2, q2, d2 ; vp8_filter + 3 * ( qs0 - ps0) 95 …vqmovn.s16 d2, q2 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs… 104 vqadd.s8 q2, q1, q11 ; Filter2 = vp8_signed_char_clamp(vp8_filter+3) 106 vshr.s8 q2, q2, #3 ; Filter2 >>= 3 [all …]
|
D | buildintrapredictorsmby_neon.asm | 60 vpaddl.u8 q2, q1 61 vpaddl.u16 q3, q2 180 vdup.u8 q2, r5 184 vst1.u8 {q2}, [r1]! 193 vdup.u8 q2, r5 197 vst1.u8 {q2}, [r1]! 207 vdup.u8 q2, r5 211 vst1.u8 {q2}, [r1]! 220 vdup.u8 q2, r5 224 vst1.u8 {q2}, [r1]! [all …]
|
D | mbloopfilter_neon.asm | 49 vld1.u8 {q9}, [r0], r1 ; q2 101 vld1.u8 {d18}, [r0], r1 ; q2 102 vld1.u8 {d19}, [r3], r1 ; q2 364 ; q2 thresh 371 ; q9 q2 382 vabd.u8 q3, q9, q8 ; abs(q2 - q1) 383 vabd.u8 q0, q10, q9 ; abs(q3 - q2) 393 vcgt.u8 q13, q13, q2 ; (abs(p1 - p0) > thresh) * -1 394 vcgt.u8 q14, q14, q2 ; (abs(q1 - q0) > thresh) * -1 401 vadd.u8 q2, q2, q2 ; flimit * 2 [all …]
|
D | loopfilter_neon.asm | 49 vld1.u8 {q9}, [r2], r1 ; q2 88 vld1.u8 {d18}, [r3], r1 ; q2 100 vld1.u8 {d19}, [r12], r1 ; q2 301 ; q2 thresh 308 ; q9 q2 318 vabd.u8 q3, q9, q8 ; abs(q2 - q1) 319 vabd.u8 q4, q10, q9 ; abs(q3 - q2) 328 vcgt.u8 q13, q13, q2 ; (abs(p1 - p0) > thresh)*-1 329 vcgt.u8 q14, q14, q2 ; (abs(q1 - q0) > thresh)*-1 336 vabd.u8 q2, q5, q8 ; a = abs(p1 - q1) [all …]
|
D | recon16x16mb_neon.asm | 37 vmovl.u8 q2, d26 52 vadd.s16 q2, q2, q10 61 vqmovun.s16 d2, q2 69 vst1.u8 {q2}, [r2], r3 88 vmovl.u8 q2, d26 99 vadd.s16 q2, q2, q10 108 vqmovun.s16 d2, q2
|
D | recon4b_neon.asm | 32 vmovl.u8 q2, d26 43 vadd.s16 q2, q2, q10 52 vqmovun.s16 d2, q2 63 vst1.u8 {q2}, [r0]
|
D | iwalsh_neon.asm | 59 vmov.i16 q2, #3 60 vadd.i16 q0, q0, q2 ;e/f += 3 61 vadd.i16 q1, q1, q2 ;g/h += 3 85 vmov.i16 q2, #3 86 vadd.i16 q3, q1, q2
|
D | recon2b_neon.asm | 31 vmovl.u8 q2, d18 36 vadd.s16 q2, q2, q6 41 vqmovun.s16 d2, q2
|
D | bilinearpredict16x16_neon.asm | 139 vmull.u8 q2, d12, d0 152 vmlal.u8 q2, d12, d1 ;(src_ptr[0] * vp8_filter[1]) 162 vqrshrn.u16 d17, q2, #7 187 vmull.u8 q2, d23, d0 199 vmlal.u8 q2, d25, d1 210 vqrshrn.u16 d3, q2, #7 311 vmull.u8 q2, d23, d0 323 vmlal.u8 q2, d25, d1 332 vqrshrn.u16 d3, q2, #7
|
/external/libvpx/vp8/decoder/arm/neon/ |
D | dequant_idct_neon.asm | 41 vmul.i16 q2, q4, q6 45 vswp d3, d4 ;q2(vp[4] vp[12]) 47 vqdmulh.s16 q3, q2, d0[2] 48 vqdmulh.s16 q4, q2, d0[0] 56 vqadd.s16 q3, q3, q2 57 vqadd.s16 q4, q4, q2 76 vqdmulh.s16 q3, q2, d0[2] 77 vqdmulh.s16 q4, q2, d0[0] 87 vqadd.s16 q3, q3, q2 88 vqadd.s16 q4, q4, q2 [all …]
|
D | idct_dequant_dc_full_2x_neon.asm | 28 vld1.16 {q2, q3}, [r0] ; l q 50 vmul.i16 q2, q2, q0 61 ; q2: l0r0 q3: l8r8 76 vqadd.s16 q10, q2, q3 ; a1 = 0 + 8 77 vqsub.s16 q11, q2, q3 ; b1 = 0 - 8 97 vqsub.s16 q2, q6, q5 105 vqadd.s16 q5, q11, q2 106 vqsub.s16 q6, q11, q2 129 vqadd.s16 q2, q4, q6 ; a1 = 0 + 2 150 vqadd.s16 q4, q2, q9 [all …]
|
D | idct_dequant_full_2x_neon.asm | 28 vld1.16 {q2, q3}, [r0] ; l q 46 vmul.i16 q2, q2, q0 53 ; q2: l0r0 q3: l8r8 68 vqadd.s16 q10, q2, q3 ; a1 = 0 + 8 69 vqsub.s16 q11, q2, q3 ; b1 = 0 - 8 89 vqsub.s16 q2, q6, q5 97 vqadd.s16 q5, q11, q2 98 vqsub.s16 q6, q11, q2 121 vqadd.s16 q2, q4, q6 ; a1 = 0 + 2 142 vqadd.s16 q4, q2, q9 [all …]
|
/external/libvpx/vpx_scale/arm/neon/ |
D | vp8_vpxyv12_copyframeyonly_neon.asm | 50 vld1.8 {q2, q3}, [r8]! 61 vst1.8 {q2, q3}, [r9]! 120 vmov q3, q2 131 vst1.8 {q2, q3}, [r6], lr 155 vld1.8 {q2, q3}, [r1]! 169 vst1.8 {q2, q3}, [r5]! 201 vld1.8 {q2}, [r2]! 209 vst1.8 {q2}, [r6], lr 211 vst1.8 {q2}, [r6], lr 213 vst1.8 {q2}, [r6], lr [all …]
|
D | vp8_vpxyv12_extendframeborders_neon.asm | 62 vmov q3, q2 73 vst1.8 {q2, q3}, [r6], lr 97 vld1.8 {q2, q3}, [r1]! 111 vst1.8 {q2, q3}, [r5]! 179 vst1.8 {q2}, [r5], lr 207 vld1.8 {q2, q3}, [r1]! 217 vst1.8 {q2, q3}, [r5]! 249 vld1.8 {q2}, [r2]! 257 vst1.8 {q2}, [r6], lr 259 vst1.8 {q2}, [r6], lr [all …]
|
D | vp8_vpxyv12_copysrcframe_func_neon.asm | 54 vld1.8 {q2, q3}, [r2]! 64 vst1.8 {q2, q3}, [r3]! 110 vld1.8 {q2, q3}, [r2]! 116 vst1.8 {q2, q3}, [r3]! 168 vld1.8 {q2, q3}, [r2]! 174 vst1.8 {q2, q3}, [r3]! 216 vld1.8 {q2, q3}, [r2]! 220 vst1.8 {q2, q3}, [r3]!
|
/external/libvpx/vp8/encoder/arm/neon/ |
D | vp8_subpixelvariance16x16s_neon.asm | 56 vext.8 q3, q2, q3, #1 62 vrhadd.u8 q1, q2, q3 63 vrhadd.u8 q2, q4, q5 72 vsubl.u8 q2, d6, d28 97 vpadal.s16 q8, q2 145 vld1.u8 {q2}, [r0], r1 153 vrhadd.u8 q0, q0, q2 155 vrhadd.u8 q2, q2, q4 165 vsubl.u8 q2, d12, d14 190 vpadal.s16 q8, q2 [all …]
|
D | vp8_shortwalsh4x4_neon.asm | 62 vcgt.s16 q4, q2, #0 65 vsub.s16 q2, q2, q4 68 vshr.s16 q2, q2, #1 70 vst1.16 {q1, q2}, [r1]
|
D | fastfdct8x4_neon.asm | 33 vld1.16 {q2}, [r0], r2 62 vshl.i16 q2, q12, #1 72 vqdmulh.s16 q15, q2, d0[0] 74 vqdmulh.s16 q14, q2, d0[2] 83 …vadd.s16 q15, q2, q15 ;d30:temp2 = ((c1 * x_c1)>>16) + c1; d31:temp2 = ((d1 * x… 146 vclt.s16 q6, q2, #0 151 vsub.s16 q2, q2, q6 156 vshr.s16 q2, q2, #1 160 vst1.16 {q1, q2}, [r1]!
|
D | sad16_neon.asm | 36 vld1.8 {q2}, [r0], r1 61 vld1.8 {q2}, [r0], r1 86 vld1.8 {q2}, [r0], r1 111 vld1.8 {q2}, [r0], r1 155 vld1.8 {q2}, [r0], r1 179 vld1.8 {q2}, [r0], r1
|
D | fastfdct4x4_neon.asm | 100 vclt.s16 q4, q2, #0 103 vsub.s16 q2, q2, q4 106 vshr.s16 q2, q2, #1 108 vst1.16 {q1, q2}, [r1]
|
/external/bluetooth/glib/tests/ |
D | queue-test.c | 749 GQueue *q, *q2; in main() local 891 q2 = g_queue_copy (q); in main() 893 check_integrity (q2); in main() 895 g_assert (g_list_length (q2->head) == 0); in main() 897 check_integrity (q2); in main() 899 g_queue_sort (q2, compare_int, NULL); in main() 900 check_integrity (q2); in main() 908 check_integrity (q2); in main() 915 check_integrity (q2); in main() 924 check_integrity (q2); in main() [all …]
|
/external/libvpx/vp8/common/x86/ |
D | loopfilter_mmx.asm | 51 movq mm1, [rsi+2*rax] ; q2 52 movq mm6, mm1 ; q2 53 psubusb mm1, mm2 ; q2-=q3 54 psubusb mm2, mm6 ; q3-=q2 55 por mm1, mm2 ; abs(q3-q2) 61 psubusb mm4, mm6 ; q1-=q2 62 psubusb mm6, mm3 ; q2-=q1 63 por mm4, mm6 ; abs(q2-q1) 639 movq mm1, [rsi+2*rax] ; q2 640 movq mm6, mm1 ; q2 [all …]
|
D | loopfilter_sse2.asm | 20 movdqa xmm1, [rsi+2*rax] ; q2 26 movlps xmm1, [rsi + rcx] ; q2 38 movdqa XMMWORD PTR [rsp], xmm1 ; store q2 42 movdqa xmm6, xmm1 ; q2 45 psubusb xmm1, xmm2 ; q2-=q3 46 psubusb xmm2, xmm6 ; q3-=q2 48 psubusb xmm4, xmm6 ; q1-=q2 49 psubusb xmm6, xmm3 ; q2-=q1 51 por xmm4, xmm6 ; abs(q2-q1) 52 por xmm1, xmm2 ; abs(q3-q2) [all …]
|
/external/libvpx/vp8/common/arm/armv6/ |
D | loopfilter_v6.asm | 115 ldr r11, [src], pstep ; q2 122 uqsub8 r9, r11, r10 ; q2 - q1 127 uqsub8 r10, r10, r11 ; q1 - q2 129 orr r10, r9, r10 ; abs (q2-q1) 136 uqsub8 r10, r12, r11 ; q3 - q2 137 uqsub8 r9, r11, r12 ; q2 - q3 141 orr r10, r10, r9 ; abs (q3-q2) 326 ldr r11, [src], pstep ; q2 333 uqsub8 r9, r11, r10 ; q2 - q1 339 uqsub8 r10, r10, r11 ; q1 - q2 [all …]
|