Home
last modified time | relevance | path

Searched refs:cospi_8_64 (Results 1 – 25 of 34) sorted by relevance

12

/external/libvpx/libvpx/vpx_dsp/mips/
Ditrans32_cols_dspr2.c348 [cospi_8_64] "r" (cospi_8_64), [cospi_24_64] "r" (cospi_24_64) in vpx_idct32_cols_add_blk_dspr2()
405 [cospi_8_64] "r" (cospi_8_64), [cospi_24_64] "r" (cospi_24_64) in vpx_idct32_cols_add_blk_dspr2()
468 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in vpx_idct32_cols_add_blk_dspr2()
471 temp21 = step2_18 * cospi_24_64 + step2_29 * cospi_8_64; in vpx_idct32_cols_add_blk_dspr2()
487 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in vpx_idct32_cols_add_blk_dspr2()
490 temp21 = step2_19 * cospi_24_64 + step2_28 * cospi_8_64; in vpx_idct32_cols_add_blk_dspr2()
511 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in vpx_idct32_cols_add_blk_dspr2()
514 temp21 = -step2_20 * cospi_8_64 + step2_27 * cospi_24_64; in vpx_idct32_cols_add_blk_dspr2()
530 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in vpx_idct32_cols_add_blk_dspr2()
533 temp21 = -step2_21 * cospi_8_64 + step2_26 * cospi_24_64; in vpx_idct32_cols_add_blk_dspr2()
[all …]
Ditrans32_dspr2.c406 [cospi_8_64] "r" (cospi_8_64), [cospi_24_64] "r" (cospi_24_64) in idct32_rows_dspr2()
468 [cospi_8_64] "r" (cospi_8_64), [cospi_24_64] "r" (cospi_24_64) in idct32_rows_dspr2()
536 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct32_rows_dspr2()
539 temp21 = step2_18 * cospi_24_64 + step2_29 * cospi_8_64; in idct32_rows_dspr2()
555 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct32_rows_dspr2()
558 temp21 = step2_19 * cospi_24_64 + step2_28 * cospi_8_64; in idct32_rows_dspr2()
579 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct32_rows_dspr2()
582 temp21 = -step2_20 * cospi_8_64 + step2_27 * cospi_24_64; in idct32_rows_dspr2()
598 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct32_rows_dspr2()
601 temp21 = -step2_21 * cospi_8_64 + step2_26 * cospi_24_64; in idct32_rows_dspr2()
[all …]
Ditrans16_dspr2.c75 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64), in idct16_rows_dspr2()
138 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct16_rows_dspr2()
200 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct16_rows_dspr2()
471 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64), in idct16_cols_add_blk_dspr2()
534 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct16_cols_add_blk_dspr2()
596 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct16_cols_add_blk_dspr2()
1159 s4 = x4 * cospi_8_64 + x5 * cospi_24_64; in iadst16_dspr2()
1160 s5 = x4 * cospi_24_64 - x5 * cospi_8_64; in iadst16_dspr2()
1161 s6 = - x6 * cospi_24_64 + x7 * cospi_8_64; in iadst16_dspr2()
1162 s7 = x6 * cospi_8_64 + x7 * cospi_24_64; in iadst16_dspr2()
[all …]
Ditrans8_dspr2.c187 [cospi_20_64] "r" (cospi_20_64), [cospi_8_64] "r" (cospi_8_64), in idct8_rows_dspr2()
436 [cospi_20_64] "r" (cospi_20_64), [cospi_8_64] "r" (cospi_8_64), in idct8_columns_add_blk_dspr2()
634 s4 = cospi_8_64 * x4 + cospi_24_64 * x5; in iadst8_dspr2()
635 s5 = cospi_24_64 * x4 - cospi_8_64 * x5; in iadst8_dspr2()
636 s6 = -cospi_24_64 * x6 + cospi_8_64 * x7; in iadst8_dspr2()
637 s7 = cospi_8_64 * x6 + cospi_24_64 * x7; in iadst8_dspr2()
Ditrans4_dspr2.c92 [cospi_8_64] "r" (cospi_8_64), [cospi_16_64] "r" (cospi_16_64), in vpx_idct4_rows_dspr2()
215 [cospi_8_64] "r" (cospi_8_64), [cospi_16_64] "r" (cospi_16_64), in vpx_idct4_columns_add_blk_dspr2()
Dinv_txfm_msa.h24 v8i16 coeff1_m = { cospi_8_64, -cospi_8_64, cospi_16_64, \
122 c2_m = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); \
123 c3_m = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); \
237 k2_m = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); \
238 k3_m = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); \
254 cospi_6_64, -cospi_26_64, cospi_8_64, cospi_24_64, -cospi_8_64 }; \
255 v8i16 mask3_m = { -cospi_24_64, cospi_8_64, cospi_16_64, \
385 k0_m = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); \
386 k1_m = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); \
387 k2_m = VP9_SET_COSPI_PAIR(-cospi_24_64, cospi_8_64); \
Dfwd_dct32x32_msa.c82 DOTP_CONST_PAIR(in0, in1, cospi_24_64, cospi_8_64, temp1, temp0); in fdct8x32_1d_column_even_store()
104 DOTP_CONST_PAIR(vec6, vec1, cospi_24_64, cospi_8_64, in1, in3); in fdct8x32_1d_column_even_store()
118 DOTP_CONST_PAIR((-vec2), vec5, cospi_24_64, cospi_8_64, in2, in1); in fdct8x32_1d_column_even_store()
182 DOTP_CONST_PAIR(in26, in21, cospi_24_64, cospi_8_64, in18, in29); in fdct8x32_1d_column_odd_store()
183 DOTP_CONST_PAIR(in27, in20, cospi_24_64, cospi_8_64, in19, in28); in fdct8x32_1d_column_odd_store()
216 DOTP_CONST_PAIR(-in16, in27, cospi_24_64, cospi_8_64, in20, in27); in fdct8x32_1d_column_odd_store()
217 DOTP_CONST_PAIR(-in17, in26, cospi_24_64, cospi_8_64, in21, in26); in fdct8x32_1d_column_odd_store()
339 cospi_8_64, vec4_r, tmp3_w, vec6_r, vec3_r); in fdct8x32_1d_row_even_4x()
366 DOTP_CONST_PAIR(vec6, vec1, cospi_24_64, cospi_8_64, in1, in3); in fdct8x32_1d_row_even_4x()
380 DOTP_CONST_PAIR((-vec2), vec5, cospi_24_64, cospi_8_64, in2, in1); in fdct8x32_1d_row_even_4x()
[all …]
Didct32x32_msa.c59 DOTP_CONST_PAIR(reg2, reg6, cospi_24_64, cospi_8_64, reg2, reg6); in idct32x8_row_even_process_store()
87 DOTP_CONST_PAIR(reg7, reg0, cospi_24_64, cospi_8_64, reg0, reg7); in idct32x8_row_even_process_store()
88 DOTP_CONST_PAIR((-reg6), reg1, cospi_24_64, cospi_8_64, reg6, reg1); in idct32x8_row_even_process_store()
160 DOTP_CONST_PAIR(vec1, vec0, cospi_24_64, cospi_8_64, vec0, vec1); in idct32x8_row_odd_process_store()
169 DOTP_CONST_PAIR(vec2, vec3, cospi_24_64, cospi_8_64, vec2, vec3); in idct32x8_row_odd_process_store()
197 DOTP_CONST_PAIR(vec3, vec2, -cospi_8_64, cospi_24_64, vec0, vec1); in idct32x8_row_odd_process_store()
207 DOTP_CONST_PAIR(reg3, reg2, -cospi_8_64, cospi_24_64, reg0, reg1); in idct32x8_row_odd_process_store()
375 DOTP_CONST_PAIR(reg2, reg6, cospi_24_64, cospi_8_64, reg2, reg6); in idct8x32_column_even_process_store()
405 DOTP_CONST_PAIR(reg7, reg0, cospi_24_64, cospi_8_64, reg0, reg7); in idct8x32_column_even_process_store()
406 DOTP_CONST_PAIR((-reg6), reg1, cospi_24_64, cospi_8_64, reg6, reg1); in idct8x32_column_even_process_store()
[all …]
Didct16x16_msa.c32 DOTP_CONST_PAIR(reg4, reg12, cospi_24_64, cospi_8_64, reg4, reg12); in vpx_idct16_1d_rows_msa()
65 DOTP_CONST_PAIR(reg7, reg9, cospi_24_64, cospi_8_64, reg7, reg9); in vpx_idct16_1d_rows_msa()
66 DOTP_CONST_PAIR((-reg5), (-reg11), cospi_8_64, cospi_24_64, reg5, reg11); in vpx_idct16_1d_rows_msa()
124 DOTP_CONST_PAIR(reg4, reg12, cospi_24_64, cospi_8_64, reg4, reg12); in vpx_idct16_1d_columns_addblk_msa()
162 DOTP_CONST_PAIR(reg7, reg9, cospi_24_64, cospi_8_64, reg7, reg9); in vpx_idct16_1d_columns_addblk_msa()
163 DOTP_CONST_PAIR((-reg5), (-reg11), cospi_8_64, cospi_24_64, reg5, reg11); in vpx_idct16_1d_columns_addblk_msa()
413 k0 = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); in vpx_iadst16_1d_columns_addblk_msa()
414 k1 = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); in vpx_iadst16_1d_columns_addblk_msa()
415 k2 = VP9_SET_COSPI_PAIR(-cospi_24_64, cospi_8_64); in vpx_iadst16_1d_columns_addblk_msa()
Dfwd_txfm_msa.h36 v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, \
37 cospi_24_64, -cospi_8_64, 0, 0, 0 }; \
74 v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, \
134 v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, cospi_24_64, \
198 v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, \
199 cospi_24_64, -cospi_8_64, -cospi_24_64, \
Didct8x8_msa.c68 k2 = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); in vpx_idct8x8_12_add_msa()
69 k3 = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); in vpx_idct8x8_12_add_msa()
/external/libvpx/libvpx/vpx_dsp/
Dfwd_txfm.c57 temp1 = step[2] * cospi_24_64 + step[3] * cospi_8_64; in vpx_fdct4x4_c()
58 temp2 = -step[2] * cospi_8_64 + step[3] * cospi_24_64; in vpx_fdct4x4_c()
136 t2 = x2 * cospi_24_64 + x3 * cospi_8_64; in vpx_fdct8x8_c()
137 t3 = -x2 * cospi_8_64 + x3 * cospi_24_64; in vpx_fdct8x8_c()
272 t2 = x3 * cospi_8_64 + x2 * cospi_24_64; in vpx_fdct16x16_c()
273 t3 = x3 * cospi_24_64 - x2 * cospi_8_64; in vpx_fdct16x16_c()
322 temp1 = step3[1] * -cospi_8_64 + step3[6] * cospi_24_64; in vpx_fdct16x16_c()
323 temp2 = step3[2] * cospi_24_64 + step3[5] * cospi_8_64; in vpx_fdct16x16_c()
326 temp1 = step3[2] * cospi_8_64 - step3[5] * cospi_24_64; in vpx_fdct16x16_c()
327 temp2 = step3[1] * cospi_24_64 + step3[6] * cospi_8_64; in vpx_fdct16x16_c()
[all …]
Dinv_txfm.c102 temp1 = input[1] * cospi_24_64 - input[3] * cospi_8_64; in idct4_c()
103 temp2 = input[1] * cospi_8_64 + input[3] * cospi_24_64; in idct4_c()
178 temp1 = step1[1] * cospi_24_64 - step1[3] * cospi_8_64; in idct8_c()
179 temp2 = step1[1] * cospi_8_64 + step1[3] * cospi_24_64; in idct8_c()
327 s4 = (int)(cospi_8_64 * x4 + cospi_24_64 * x5); in iadst8_c()
328 s5 = (int)(cospi_24_64 * x4 - cospi_8_64 * x5); in iadst8_c()
329 s6 = (int)(-cospi_24_64 * x6 + cospi_8_64 * x7); in iadst8_c()
330 s7 = (int)(cospi_8_64 * x6 + cospi_24_64 * x7); in iadst8_c()
469 temp1 = step1[2] * cospi_24_64 - step1[3] * cospi_8_64; in idct16_c()
470 temp2 = step1[2] * cospi_8_64 + step1[3] * cospi_24_64; in idct16_c()
[all …]
Dtxfm_common.h35 static const tran_high_t cospi_8_64 = 15137; variable
/external/libvpx/libvpx/vp9/encoder/
Dvp9_dct.c36 temp1 = step[2] * cospi_24_64 + step[3] * cospi_8_64; in fdct4()
37 temp2 = -step[2] * cospi_8_64 + step[3] * cospi_24_64; in fdct4()
64 t2 = x2 * cospi_24_64 + x3 * cospi_8_64; in fdct8()
65 t3 = -x2 * cospi_8_64 + x3 * cospi_24_64; in fdct8()
143 t2 = x3 * cospi_8_64 + x2 * cospi_24_64; in fdct16()
144 t3 = x3 * cospi_24_64 - x2 * cospi_8_64; in fdct16()
194 temp1 = step3[1] * -cospi_8_64 + step3[6] * cospi_24_64; in fdct16()
195 temp2 = step3[2] * cospi_24_64 + step3[5] * cospi_8_64; in fdct16()
198 temp1 = step3[2] * cospi_8_64 - step3[5] * cospi_24_64; in fdct16()
199 temp2 = step3[1] * cospi_24_64 + step3[6] * cospi_8_64; in fdct16()
[all …]
/external/libvpx/libvpx/vpx_dsp/arm/
Didct4x4_add_neon.asm39 ; cospi_8_64 = 15137 = 0x3b21
58 vdup.16 d20, r0 ; replicate cospi_8_64
80 vmull.s16 q1, d17, d20 ; input[1] * cospi_8_64
87 ; input[1] * cospi_24_64 - input[3] * cospi_8_64;
88 ; input[1] * cospi_8_64 + input[3] * cospi_24_64;
131 vmull.s16 q1, d17, d20 ; input[1] * cospi_8_64
138 ; input[1] * cospi_24_64 - input[3] * cospi_8_64;
139 ; input[1] * cospi_8_64 + input[3] * cospi_24_64;
Didct16x16_add_neon.asm158 ; generate cospi_8_64 = 15137
163 vdup.16 d31, r3 ; duplicate cospi_8_64
181 ; step1[2] * cospi_24_64 - step1[3] * cospi_8_64;
182 ; step1[2] * cospi_8_64
190 ; temp2 = input[1] * cospi_8_64 + input[3] * cospi_24_64
194 ; temp1 = input[1] * cospi_24_64 - input[3] * cospi_8_64
473 ; generate cospi_8_64 = 15137
477 ; -step1[9] * cospi_8_64 + step1[14] * cospi_24_64
478 vdup.16 d30, r12 ; duplicate cospi_8_64
489 ; temp2 = step1[9] * cospi_24_64 + step1[14] * cospi_8_64
[all …]
Didct8x8_add_neon.asm99 vdup.16 d1, r9 ; duplicate cospi_8_64
109 ; input[1] * cospi_24_64 - input[3] * cospi_8_64
114 ; input[1] * cospi_8_64
118 ; input[1] * cospi_24_64 - input[3] * cospi_8_64
122 ; input[1] * cospi_8_64 + input[3] * cospi_24_64
242 ; generate cospi_8_64 = 15137
354 ; generate cospi_8_64 = 15137
399 vdup.16 q0, r12 ; duplicate cospi_8_64*2
404 ; dct_const_round_shift(input[1] * cospi_8_64)
Didct32x32_add_neon.asm20 cospi_8_64 EQU 15137 define
547 ;temp1 = step1b[29][i] * cospi_24_64 - step1b[18][i] * cospi_8_64;
548 ;temp2 = step1b[29][i] * cospi_8_64 + step1b[18][i] * cospi_24_64;
551 DO_BUTTERFLY_STD cospi_24_64, cospi_8_64, d0, d1, d2, d3
561 ;temp1 = step1b[28][i] * cospi_24_64 - step1b[19][i] * cospi_8_64;
562 ;temp2 = step1b[28][i] * cospi_8_64 + step1b[19][i] * cospi_24_64;
565 DO_BUTTERFLY_STD cospi_24_64, cospi_8_64, d8, d9, d12, d13
709 ;temp1 = step1b[20][i] * (-cospi_8_64) - step1b[27][i] * (-cospi_24_64);
710 ;temp2 = step1b[20][i] * (-cospi_24_64) + step1b[27][i] * (-cospi_8_64);
713 DO_BUTTERFLY_STD (-cospi_8_64), (-cospi_24_64), d10, d11, d12, d13
[all …]
Dfwd_txfm_neon.c51 v_t2_lo = vmlal_n_s16(v_t2_lo, vget_low_s16(v_x3), (int16_t)cospi_8_64); in vpx_fdct8x8_neon()
52 v_t2_hi = vmlal_n_s16(v_t2_hi, vget_high_s16(v_x3), (int16_t)cospi_8_64); in vpx_fdct8x8_neon()
53 v_t3_lo = vmlsl_n_s16(v_t3_lo, vget_low_s16(v_x2), (int16_t)cospi_8_64); in vpx_fdct8x8_neon()
54 v_t3_hi = vmlsl_n_s16(v_t3_hi, vget_high_s16(v_x2), (int16_t)cospi_8_64); in vpx_fdct8x8_neon()
Didct32x32_add_neon.c519 DO_BUTTERFLY_STD(cospi_24_64, cospi_8_64, &q0s16, &q1s16) in vpx_idct32x32_1024_add_neon()
525 DO_BUTTERFLY_STD(cospi_24_64, cospi_8_64, &q4s16, &q6s16) in vpx_idct32x32_1024_add_neon()
586 DO_BUTTERFLY_STD(-cospi_8_64, -cospi_24_64, &q5s16, &q6s16); in vpx_idct32x32_1024_add_neon()
589 DO_BUTTERFLY_STD(-cospi_8_64, -cospi_24_64, &q0s16, &q1s16); in vpx_idct32x32_1024_add_neon()
625 DO_BUTTERFLY_STD(cospi_24_64, cospi_8_64, &q1s16, &q3s16) in vpx_idct32x32_1024_add_neon()
639 DO_BUTTERFLY_STD(-cospi_8_64, -cospi_24_64, &q4s16, &q7s16) in vpx_idct32x32_1024_add_neon()
679 DO_BUTTERFLY_STD(cospi_24_64, cospi_8_64, &q14s16, &q6s16) in vpx_idct32x32_1024_add_neon()
Didct4x4_add_neon.c27 int16_t cospi_8_64 = 15137; in vpx_idct4x4_16_add_neon() local
46 d20s16 = vdup_n_s16(cospi_8_64); in vpx_idct4x4_16_add_neon()
/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_txfm_impl_sse2.h54 const __m128i k__cospi_C = octa_set_epi16(cospi_8_64, cospi_24_64, in FDCT4x4_2D()
55 cospi_8_64, cospi_24_64, in FDCT4x4_2D()
56 cospi_24_64, -cospi_8_64, in FDCT4x4_2D()
57 cospi_24_64, -cospi_8_64); in FDCT4x4_2D()
58 const __m128i k__cospi_D = octa_set_epi16(cospi_24_64, -cospi_8_64, in FDCT4x4_2D()
59 cospi_24_64, -cospi_8_64, in FDCT4x4_2D()
60 cospi_8_64, cospi_24_64, in FDCT4x4_2D()
61 cospi_8_64, cospi_24_64); in FDCT4x4_2D()
70 const __m128i k__cospi_G = octa_set_epi16(cospi_8_64, cospi_24_64, in FDCT4x4_2D()
71 cospi_8_64, cospi_24_64, in FDCT4x4_2D()
[all …]
Dinv_txfm_sse2.c30 (int16_t)-cospi_16_64, (int16_t)cospi_24_64, (int16_t)-cospi_8_64, in vpx_idct4x4_16_add_sse2()
31 (int16_t)cospi_8_64, (int16_t)cospi_24_64); in vpx_idct4x4_16_add_sse2()
184 const __m128i k__cospi_p24_m08 = pair_set_epi16(cospi_24_64, -cospi_8_64); in idct4_sse2()
185 const __m128i k__cospi_p08_p24 = pair_set_epi16(cospi_8_64, cospi_24_64); in idct4_sse2()
465 const __m128i stg2_2 = pair_set_epi16(cospi_24_64, -cospi_8_64); in vpx_idct8x8_64_add_sse2()
466 const __m128i stg2_3 = pair_set_epi16(cospi_8_64, cospi_24_64); in vpx_idct8x8_64_add_sse2()
554 const __m128i stg2_2 = pair_set_epi16(cospi_24_64, -cospi_8_64); in idct8_sse2()
555 const __m128i stg2_3 = pair_set_epi16(cospi_8_64, cospi_24_64); in idct8_sse2()
580 const __m128i k__cospi_p08_p24 = pair_set_epi16(cospi_8_64, cospi_24_64); in iadst8_sse2()
581 const __m128i k__cospi_p24_m08 = pair_set_epi16(cospi_24_64, -cospi_8_64); in iadst8_sse2()
[all …]
/external/libvpx/libvpx/vp9/encoder/mips/msa/
Dvp9_fdct_msa.h24 v8i16 coeff1_m = { cospi_8_64, -cospi_8_64, cospi_16_64, -cospi_16_64, \

12