/external/libvpx/vp8/common/ |
D | loopfilter_filters.c | 51 static __inline void vp8_filter(signed char mask, signed char hev, uc *op1, uc *op0, uc *oq0, uc *o… in vp8_filter() function 56 signed char vp8_filter, Filter1, Filter2; in vp8_filter() local 65 vp8_filter = vp8_signed_char_clamp(ps1 - qs1); in vp8_filter() 66 vp8_filter &= hev; in vp8_filter() 69 vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * (qs0 - ps0)); in vp8_filter() 70 vp8_filter &= mask; in vp8_filter() 76 Filter1 = vp8_signed_char_clamp(vp8_filter + 4); in vp8_filter() 77 Filter2 = vp8_signed_char_clamp(vp8_filter + 3); in vp8_filter() 84 vp8_filter = Filter1; in vp8_filter() 87 vp8_filter += 1; in vp8_filter() [all …]
|
D | filter.c | 49 const short *vp8_filter in filter_block2d_first_pass() argument 59 Temp = ((int)src_ptr[-2 * (int)pixel_step] * vp8_filter[0]) + in filter_block2d_first_pass() 60 ((int)src_ptr[-1 * (int)pixel_step] * vp8_filter[1]) + in filter_block2d_first_pass() 61 ((int)src_ptr[0] * vp8_filter[2]) + in filter_block2d_first_pass() 62 ((int)src_ptr[pixel_step] * vp8_filter[3]) + in filter_block2d_first_pass() 63 ((int)src_ptr[2*pixel_step] * vp8_filter[4]) + in filter_block2d_first_pass() 64 ((int)src_ptr[3*pixel_step] * vp8_filter[5]) + in filter_block2d_first_pass() 94 const short *vp8_filter in filter_block2d_second_pass() argument 105 Temp = ((int)src_ptr[-2 * (int)pixel_step] * vp8_filter[0]) + in filter_block2d_second_pass() 106 ((int)src_ptr[-1 * (int)pixel_step] * vp8_filter[1]) + in filter_block2d_second_pass() [all …]
|
/external/libvpx/vp8/common/arm/neon/ |
D | sixtappredict4x4_neon.asm | 86 vmull.u8 q7, d18, d5 ;(src_ptr[3] * vp8_filter[5]) 96 vmlal.u8 q7, d6, d0 ;+(src_ptr[-2] * vp8_filter[0]) 103 vmlsl.u8 q7, d18, d1 ;-(src_ptr[-1] * vp8_filter[1]) 110 vmlsl.u8 q7, d6, d4 ;-(src_ptr[2] * vp8_filter[4]) 117 vmlal.u8 q7, d18, d2 ;(src_ptr[0] * vp8_filter[2]) 122 vmull.u8 q9, d6, d3 ;(src_ptr[1] * vp8_filter[3]) 151 vmull.u8 q7, d18, d5 ;(src_ptr[3] * vp8_filter[5]) 153 vmull.u8 q12, d31, d5 ;(src_ptr[3] * vp8_filter[5]) 163 vmlal.u8 q7, d6, d0 ;+(src_ptr[-2] * vp8_filter[0]) 165 vmlal.u8 q12, d22, d0 ;(src_ptr[-2] * vp8_filter[0]) [all …]
|
D | sixtappredict16x16_neon.asm | 90 vmull.u8 q8, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 101 vmlsl.u8 q8, d28, d1 ;-(src_ptr[-1] * vp8_filter[1]) 109 vmlsl.u8 q9, d28, d1 ;-(src_ptr[-1] * vp8_filter[1]) 117 vmlsl.u8 q8, d28, d4 ;-(src_ptr[2] * vp8_filter[4]) 125 vmlsl.u8 q9, d28, d4 ;-(src_ptr[2] * vp8_filter[4]) 133 vmlal.u8 q8, d28, d5 ;(src_ptr[3] * vp8_filter[5]) 141 vmlal.u8 q9, d28, d5 ;(src_ptr[3] * vp8_filter[5]) 149 vmlal.u8 q8, d28, d2 ;(src_ptr[0] * vp8_filter[2]) 157 vmlal.u8 q9, d28, d2 ;(src_ptr[0] * vp8_filter[2]) 169 vmull.u8 q4, d28, d3 ;(src_ptr[1] * vp8_filter[3]) [all …]
|
D | sixtappredict8x4_neon.asm | 80 vmull.u8 q7, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 90 vmlsl.u8 q7, d28, d1 ;-(src_ptr[-1] * vp8_filter[1]) 100 vmlsl.u8 q7, d28, d4 ;-(src_ptr[2] * vp8_filter[4]) 110 vmlal.u8 q7, d28, d2 ;(src_ptr[0] * vp8_filter[2]) 120 vmlal.u8 q7, d28, d5 ;(src_ptr[3] * vp8_filter[5]) 130 vmull.u8 q3, d28, d3 ;(src_ptr[1] * vp8_filter[3]) 157 vmull.u8 q8, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 169 vmlsl.u8 q8, d27, d1 ;-(src_ptr[-1] * vp8_filter[1]) 181 vmlsl.u8 q8, d27, d4 ;-(src_ptr[2] * vp8_filter[4]) 193 vmlal.u8 q8, d27, d2 ;(src_ptr[0] * vp8_filter[2]) [all …]
|
D | loopfiltersimplehorizontaledge_neon.asm | 48 ;vp8_filter() function 63 vqsub.s8 q4, q5, q8 ; q4: vp8_filter = vp8_signed_char_clamp(ps1-qs1) 74 vaddw.s8 q2, q2, d8 ; vp8_filter + 3 * ( qs0 - ps0) 77 …;vqadd.s8 q4, q4, q2 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs… 78 …vqmovn.s16 d8, q2 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs… 82 vand q4, q4, q15 ; vp8_filter &= mask 84 vqadd.s8 q2, q4, q10 ; Filter2 = vp8_signed_char_clamp(vp8_filter+3) 85 vqadd.s8 q4, q4, q9 ; Filter1 = vp8_signed_char_clamp(vp8_filter+4)
|
D | sixtappredict8x8_neon.asm | 72 vmull.u8 q7, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 82 vmlsl.u8 q7, d28, d1 ;-(src_ptr[-1] * vp8_filter[1]) 92 vmlsl.u8 q7, d28, d4 ;-(src_ptr[2] * vp8_filter[4]) 102 vmlal.u8 q7, d28, d2 ;(src_ptr[0] * vp8_filter[2]) 112 vmlal.u8 q7, d28, d5 ;(src_ptr[3] * vp8_filter[5]) 122 vmull.u8 q3, d28, d3 ;(src_ptr[1] * vp8_filter[3]) 158 vmull.u8 q8, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 170 vmlsl.u8 q8, d27, d1 ;-(src_ptr[-1] * vp8_filter[1]) 182 vmlsl.u8 q8, d27, d4 ;-(src_ptr[2] * vp8_filter[4]) 194 vmlal.u8 q8, d27, d2 ;(src_ptr[0] * vp8_filter[2]) [all …]
|
D | loopfiltersimpleverticaledge_neon.asm | 77 ;vp8_filter() function 83 vqsub.s8 q1, q3, q6 ; vp8_filter = vp8_signed_char_clamp(ps1-qs1) 85 …;vmul.i8 q2, q2, q11 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs… 92 vaddw.s8 q2, q2, d2 ; vp8_filter + 3 * ( qs0 - ps0) 95 …vqmovn.s16 d2, q2 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs… 102 vand q1, q1, q15 ; vp8_filter &= mask 104 vqadd.s8 q2, q1, q11 ; Filter2 = vp8_signed_char_clamp(vp8_filter+3) 105 vqadd.s8 q1, q1, q12 ; Filter1 = vp8_signed_char_clamp(vp8_filter+4)
|
D | bilinearpredict16x16_neon.asm | 62 vmull.u8 q7, d2, d0 ;(src_ptr[0] * vp8_filter[0]) 76 vmlal.u8 q7, d2, d1 ;(src_ptr[0] * vp8_filter[1]) 86 vmlal.u8 q8, d3, d1 ;(src_ptr[0] * vp8_filter[1]) 115 vmull.u8 q9, d2, d0 ;(src_ptr[0] * vp8_filter[0]) 126 vmlal.u8 q9, d2, d1 ;(src_ptr[0] * vp8_filter[1]) 134 vmlal.u8 q10, d3, d1 ;(src_ptr[0] * vp8_filter[1]) 146 vmlal.u8 q1, d11, d1 ;(src_ptr[0] * vp8_filter[1]) 152 vmlal.u8 q2, d12, d1 ;(src_ptr[0] * vp8_filter[1]) 185 vmull.u8 q1, d22, d0 ;(src_ptr[0] * vp8_filter[0]) 198 vmlal.u8 q1, d24, d1 ;(src_ptr[pixel_step] * vp8_filter[1]) [all …]
|
D | bilinearpredict8x8_neon.asm | 46 vmull.u8 q6, d2, d0 ;(src_ptr[0] * vp8_filter[0]) 56 vmlal.u8 q6, d3, d1 ;(src_ptr[1] * vp8_filter[1]) 73 vmull.u8 q6, d2, d0 ;(src_ptr[0] * vp8_filter[0]) 85 vmlal.u8 q6, d3, d1 ;(src_ptr[1] * vp8_filter[1]) 111 vmull.u8 q1, d22, d0 ;(src_ptr[0] * vp8_filter[0]) 120 vmlal.u8 q1, d23, d1 ;(src_ptr[pixel_step] * vp8_filter[1])
|
D | loopfilter_neon.asm | 344 ; vp8_filter() function 358 vqsub.s8 q1, q5, q8 ; vp8_filter = clamp(ps1-qs1) 364 vand q1, q1, q14 ; vp8_filter &= hev 372 ; vp8_filter = clamp(vp8_filter + 3 * ( qs0 - ps0)) 375 vand q1, q1, q15 ; vp8_filter &= mask 377 vqadd.s8 q2, q1, q10 ; Filter2 = clamp(vp8_filter+3) 378 vqadd.s8 q1, q1, q9 ; Filter1 = clamp(vp8_filter+4) 385 ; outer tap adjustments: ++vp8_filter >> 1 387 vbic q1, q1, q14 ; vp8_filter &= ~hev 389 vqadd.s8 q13, q5, q1 ; u = clamp(ps1 + vp8_filter) [all …]
|
D | bilinearpredict8x4_neon.asm | 46 vmull.u8 q6, d2, d0 ;(src_ptr[0] * vp8_filter[0]) 59 vmlal.u8 q6, d3, d1 ;(src_ptr[1] * vp8_filter[1]) 85 vmull.u8 q1, d22, d0 ;(src_ptr[0] * vp8_filter[0]) 90 vmlal.u8 q1, d23, d1 ;(src_ptr[pixel_step] * vp8_filter[1])
|
D | mbloopfilter_neon.asm | 411 ; vp8_filter 425 vqsub.s8 q1, q5, q8 ; vp8_filter = clamp(ps1-qs1) 436 vaddw.s8 q2, q2, d2 ; vp8_filter + 3 * ( qs0 - ps0) 441 ; vp8_filter = clamp(vp8_filter + 3 * ( qs0 - ps0)) 445 vand q1, q1, q15 ; vp8_filter &= mask 470 vbic q1, q1, q14 ; vp8_filter &= ~hev
|
D | bilinearpredict4x4_neon.asm | 57 vmull.u8 q7, d2, d0 ;(src_ptr[0] * vp8_filter[0]) 61 vmlal.u8 q7, d8, d1 ;(src_ptr[1] * vp8_filter[1])
|
/external/libvpx/vp8/common/arm/ |
D | filter_arm.c | 25 const short *vp8_filter 34 const short *vp8_filter 43 const short *vp8_filter 53 const short *vp8_filter 64 const short *vp8_filter
|
D | bilinearfilter_arm.h | 22 const short *vp8_filter 32 const short *vp8_filter
|
/external/libvpx/vp8/common/x86/ |
D | vp8_asm_stubs.c | 27 const short *vp8_filter 38 const short *vp8_filter 48 const short *vp8_filter 58 const short *vp8_filter 69 const short *vp8_filter 80 const short *vp8_filter 97 const short *vp8_filter 106 const short *vp8_filter 115 const short *vp8_filter
|
D | subpixel_mmx.asm | 28 ; short * vp8_filter 40 mov rdx, arg(6) ;vp8_filter 125 ; short * vp8_filter 139 mov rbx, arg(7) ;vp8_filter
|
D | subpixel_sse2.asm | 33 ; short *vp8_filter 46 mov rdx, arg(6) ;vp8_filter 147 ; short *vp8_filter 166 mov rdx, arg(6) ;vp8_filter 325 ; short * vp8_filter 342 mov rax, arg(7) ;vp8_filter 420 ; const short *vp8_filter 437 mov rax, arg(7) ;vp8_filter 533 ; const short *vp8_filter 547 mov rdx, arg(5) ;vp8_filter [all …]
|
D | loopfilter_mmx.asm | 1174 ; mm1 = vp8_filter, mm4=hev, mm6=ps0, mm3=qs0 1175 movq mm2, mm1 ; vp8_filter 1176 pand mm2, mm4; ; Filter2 = vp8_filter & hev 1202 ; mm0= filter2 mm1 = vp8_filter, mm3 =qs0 mm5=s mm4 =hev mm6=ps0 1206 ; mm1=vp8_filter, mm3=qs0, mm4 =hev mm6=ps0 1207 ; vp8_filter &= ~hev; 1208 ; Filter2 = vp8_filter; 1209 pandn mm4, mm1 ; vp8_filter&=~hev
|
/external/libvpx/vp8/encoder/ |
D | variance_c.c | 233 const short *vp8_filter in var_filter_block2d_bil_first_pass() argument 243 output_ptr[j] = (((int)src_ptr[0] * vp8_filter[0]) + in var_filter_block2d_bil_first_pass() 244 ((int)src_ptr[pixel_step] * vp8_filter[1]) + in var_filter_block2d_bil_first_pass() 291 const short *vp8_filter in var_filter_block2d_bil_second_pass() argument 302 Temp = ((int)src_ptr[0] * vp8_filter[0]) + in var_filter_block2d_bil_second_pass() 303 ((int)src_ptr[pixel_step] * vp8_filter[1]) + in var_filter_block2d_bil_second_pass()
|
/external/libvpx/vp8/common/arm/armv6/ |
D | loopfilter_v6.asm | 165 ;vp8_filter() function 182 qsub8 r8, r9, r8 ; vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0)) 184 and r7, r7, r6 ; vp8_filter (r7) &= hev 193 and r7, r7, lr ; vp8_filter &= mask; 195 ;modify code for vp8 -- Filter1 = vp8_filter (r7) 196 qadd8 r8 , r7 , r9 ; Filter2 (r8) = vp8_signed_char_clamp(vp8_filter+3) 197 qadd8 r7 , r7 , r10 ; vp8_filter = vp8_signed_char_clamp(vp8_filter+4) 201 shadd8 r7 , r7 , r9 ; vp8_filter >>= 3 214 ;qadd8 lr, r8, r7 ; u = vp8_signed_char_clamp(s + vp8_filter) 221 qsub8 r8 ,r8, r7 ; u = vp8_signed_char_clamp(qs0 - vp8_filter) [all …]
|
D | sixtappredict8x4_v6.asm | 88 smuad r11, r6, r3 ; vp8_filter[0], vp8_filter[1] 93 smlad r11, r8, r4, r11 ; vp8_filter[2], vp8_filter[3] 99 smlad r11, r10, r5, r11 ; vp8_filter[4], vp8_filter[5]
|
D | filter_v6.asm | 25 ; stack const short *vp8_filter 27 ; vp8_filter the input and put in the output array. Apply the 6 tap FIR filter with 32 ldr r11, [sp, #40] ; vp8_filter address 128 ; stack const short *vp8_filter 133 ldr r11, [sp, #36] ; vp8_filter address 201 ; stack const short *vp8_filter 206 ldr r11, [sp, #36] ; vp8_filter address 260 ; stack const short *vp8_filter 358 ; stack const short *vp8_filter
|
D | bilinearfilter_v6.asm | 23 ; stack const short *vp8_filter 29 ldr r11, [sp, #40] ; vp8_filter address 147 ; stack const short *vp8_filter 152 ldr r11, [sp, #40] ; vp8_filter address
|