/external/libvpx/libvpx/vpx_dsp/mips/ |
D | itrans32_cols_dspr2.c | 348 [cospi_8_64] "r" (cospi_8_64), [cospi_24_64] "r" (cospi_24_64) in vpx_idct32_cols_add_blk_dspr2() 405 [cospi_8_64] "r" (cospi_8_64), [cospi_24_64] "r" (cospi_24_64) in vpx_idct32_cols_add_blk_dspr2() 468 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in vpx_idct32_cols_add_blk_dspr2() 471 temp21 = step2_18 * cospi_24_64 + step2_29 * cospi_8_64; in vpx_idct32_cols_add_blk_dspr2() 487 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in vpx_idct32_cols_add_blk_dspr2() 490 temp21 = step2_19 * cospi_24_64 + step2_28 * cospi_8_64; in vpx_idct32_cols_add_blk_dspr2() 511 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in vpx_idct32_cols_add_blk_dspr2() 514 temp21 = -step2_20 * cospi_8_64 + step2_27 * cospi_24_64; in vpx_idct32_cols_add_blk_dspr2() 530 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in vpx_idct32_cols_add_blk_dspr2() 533 temp21 = -step2_21 * cospi_8_64 + step2_26 * cospi_24_64; in vpx_idct32_cols_add_blk_dspr2() [all …]
|
D | itrans32_dspr2.c | 406 [cospi_8_64] "r" (cospi_8_64), [cospi_24_64] "r" (cospi_24_64) in idct32_rows_dspr2() 468 [cospi_8_64] "r" (cospi_8_64), [cospi_24_64] "r" (cospi_24_64) in idct32_rows_dspr2() 536 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct32_rows_dspr2() 539 temp21 = step2_18 * cospi_24_64 + step2_29 * cospi_8_64; in idct32_rows_dspr2() 555 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct32_rows_dspr2() 558 temp21 = step2_19 * cospi_24_64 + step2_28 * cospi_8_64; in idct32_rows_dspr2() 579 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct32_rows_dspr2() 582 temp21 = -step2_20 * cospi_8_64 + step2_27 * cospi_24_64; in idct32_rows_dspr2() 598 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct32_rows_dspr2() 601 temp21 = -step2_21 * cospi_8_64 + step2_26 * cospi_24_64; in idct32_rows_dspr2() [all …]
|
D | itrans16_dspr2.c | 75 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64), in idct16_rows_dspr2() 138 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct16_rows_dspr2() 200 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct16_rows_dspr2() 471 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64), in idct16_cols_add_blk_dspr2() 534 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct16_cols_add_blk_dspr2() 596 [cospi_24_64] "r" (cospi_24_64), [cospi_8_64] "r" (cospi_8_64) in idct16_cols_add_blk_dspr2() 1159 s4 = x4 * cospi_8_64 + x5 * cospi_24_64; in iadst16_dspr2() 1160 s5 = x4 * cospi_24_64 - x5 * cospi_8_64; in iadst16_dspr2() 1161 s6 = - x6 * cospi_24_64 + x7 * cospi_8_64; in iadst16_dspr2() 1162 s7 = x6 * cospi_8_64 + x7 * cospi_24_64; in iadst16_dspr2() [all …]
|
D | itrans8_dspr2.c | 187 [cospi_20_64] "r" (cospi_20_64), [cospi_8_64] "r" (cospi_8_64), in idct8_rows_dspr2() 436 [cospi_20_64] "r" (cospi_20_64), [cospi_8_64] "r" (cospi_8_64), in idct8_columns_add_blk_dspr2() 634 s4 = cospi_8_64 * x4 + cospi_24_64 * x5; in iadst8_dspr2() 635 s5 = cospi_24_64 * x4 - cospi_8_64 * x5; in iadst8_dspr2() 636 s6 = -cospi_24_64 * x6 + cospi_8_64 * x7; in iadst8_dspr2() 637 s7 = cospi_8_64 * x6 + cospi_24_64 * x7; in iadst8_dspr2()
|
D | itrans4_dspr2.c | 92 [cospi_8_64] "r" (cospi_8_64), [cospi_16_64] "r" (cospi_16_64), in vpx_idct4_rows_dspr2() 215 [cospi_8_64] "r" (cospi_8_64), [cospi_16_64] "r" (cospi_16_64), in vpx_idct4_columns_add_blk_dspr2()
|
D | inv_txfm_msa.h | 24 v8i16 coeff1_m = { cospi_8_64, -cospi_8_64, cospi_16_64, \ 122 c2_m = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); \ 123 c3_m = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); \ 237 k2_m = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); \ 238 k3_m = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); \ 254 cospi_6_64, -cospi_26_64, cospi_8_64, cospi_24_64, -cospi_8_64 }; \ 255 v8i16 mask3_m = { -cospi_24_64, cospi_8_64, cospi_16_64, \ 385 k0_m = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); \ 386 k1_m = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); \ 387 k2_m = VP9_SET_COSPI_PAIR(-cospi_24_64, cospi_8_64); \
|
D | fwd_dct32x32_msa.c | 82 DOTP_CONST_PAIR(in0, in1, cospi_24_64, cospi_8_64, temp1, temp0); in fdct8x32_1d_column_even_store() 104 DOTP_CONST_PAIR(vec6, vec1, cospi_24_64, cospi_8_64, in1, in3); in fdct8x32_1d_column_even_store() 118 DOTP_CONST_PAIR((-vec2), vec5, cospi_24_64, cospi_8_64, in2, in1); in fdct8x32_1d_column_even_store() 182 DOTP_CONST_PAIR(in26, in21, cospi_24_64, cospi_8_64, in18, in29); in fdct8x32_1d_column_odd_store() 183 DOTP_CONST_PAIR(in27, in20, cospi_24_64, cospi_8_64, in19, in28); in fdct8x32_1d_column_odd_store() 216 DOTP_CONST_PAIR(-in16, in27, cospi_24_64, cospi_8_64, in20, in27); in fdct8x32_1d_column_odd_store() 217 DOTP_CONST_PAIR(-in17, in26, cospi_24_64, cospi_8_64, in21, in26); in fdct8x32_1d_column_odd_store() 339 cospi_8_64, vec4_r, tmp3_w, vec6_r, vec3_r); in fdct8x32_1d_row_even_4x() 366 DOTP_CONST_PAIR(vec6, vec1, cospi_24_64, cospi_8_64, in1, in3); in fdct8x32_1d_row_even_4x() 380 DOTP_CONST_PAIR((-vec2), vec5, cospi_24_64, cospi_8_64, in2, in1); in fdct8x32_1d_row_even_4x() [all …]
|
D | idct32x32_msa.c | 59 DOTP_CONST_PAIR(reg2, reg6, cospi_24_64, cospi_8_64, reg2, reg6); in idct32x8_row_even_process_store() 87 DOTP_CONST_PAIR(reg7, reg0, cospi_24_64, cospi_8_64, reg0, reg7); in idct32x8_row_even_process_store() 88 DOTP_CONST_PAIR((-reg6), reg1, cospi_24_64, cospi_8_64, reg6, reg1); in idct32x8_row_even_process_store() 160 DOTP_CONST_PAIR(vec1, vec0, cospi_24_64, cospi_8_64, vec0, vec1); in idct32x8_row_odd_process_store() 169 DOTP_CONST_PAIR(vec2, vec3, cospi_24_64, cospi_8_64, vec2, vec3); in idct32x8_row_odd_process_store() 197 DOTP_CONST_PAIR(vec3, vec2, -cospi_8_64, cospi_24_64, vec0, vec1); in idct32x8_row_odd_process_store() 207 DOTP_CONST_PAIR(reg3, reg2, -cospi_8_64, cospi_24_64, reg0, reg1); in idct32x8_row_odd_process_store() 375 DOTP_CONST_PAIR(reg2, reg6, cospi_24_64, cospi_8_64, reg2, reg6); in idct8x32_column_even_process_store() 405 DOTP_CONST_PAIR(reg7, reg0, cospi_24_64, cospi_8_64, reg0, reg7); in idct8x32_column_even_process_store() 406 DOTP_CONST_PAIR((-reg6), reg1, cospi_24_64, cospi_8_64, reg6, reg1); in idct8x32_column_even_process_store() [all …]
|
D | idct16x16_msa.c | 32 DOTP_CONST_PAIR(reg4, reg12, cospi_24_64, cospi_8_64, reg4, reg12); in vpx_idct16_1d_rows_msa() 65 DOTP_CONST_PAIR(reg7, reg9, cospi_24_64, cospi_8_64, reg7, reg9); in vpx_idct16_1d_rows_msa() 66 DOTP_CONST_PAIR((-reg5), (-reg11), cospi_8_64, cospi_24_64, reg5, reg11); in vpx_idct16_1d_rows_msa() 124 DOTP_CONST_PAIR(reg4, reg12, cospi_24_64, cospi_8_64, reg4, reg12); in vpx_idct16_1d_columns_addblk_msa() 162 DOTP_CONST_PAIR(reg7, reg9, cospi_24_64, cospi_8_64, reg7, reg9); in vpx_idct16_1d_columns_addblk_msa() 163 DOTP_CONST_PAIR((-reg5), (-reg11), cospi_8_64, cospi_24_64, reg5, reg11); in vpx_idct16_1d_columns_addblk_msa() 413 k0 = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); in vpx_iadst16_1d_columns_addblk_msa() 414 k1 = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); in vpx_iadst16_1d_columns_addblk_msa() 415 k2 = VP9_SET_COSPI_PAIR(-cospi_24_64, cospi_8_64); in vpx_iadst16_1d_columns_addblk_msa()
|
D | fwd_txfm_msa.h | 36 v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, \ 37 cospi_24_64, -cospi_8_64, 0, 0, 0 }; \ 74 v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, \ 134 v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, cospi_24_64, \ 198 v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, \ 199 cospi_24_64, -cospi_8_64, -cospi_24_64, \
|
D | idct8x8_msa.c | 68 k2 = VP9_SET_COSPI_PAIR(cospi_24_64, -cospi_8_64); in vpx_idct8x8_12_add_msa() 69 k3 = VP9_SET_COSPI_PAIR(cospi_8_64, cospi_24_64); in vpx_idct8x8_12_add_msa()
|
/external/libvpx/libvpx/vpx_dsp/ |
D | fwd_txfm.c | 57 temp1 = step[2] * cospi_24_64 + step[3] * cospi_8_64; in vpx_fdct4x4_c() 58 temp2 = -step[2] * cospi_8_64 + step[3] * cospi_24_64; in vpx_fdct4x4_c() 136 t2 = x2 * cospi_24_64 + x3 * cospi_8_64; in vpx_fdct8x8_c() 137 t3 = -x2 * cospi_8_64 + x3 * cospi_24_64; in vpx_fdct8x8_c() 272 t2 = x3 * cospi_8_64 + x2 * cospi_24_64; in vpx_fdct16x16_c() 273 t3 = x3 * cospi_24_64 - x2 * cospi_8_64; in vpx_fdct16x16_c() 322 temp1 = step3[1] * -cospi_8_64 + step3[6] * cospi_24_64; in vpx_fdct16x16_c() 323 temp2 = step3[2] * cospi_24_64 + step3[5] * cospi_8_64; in vpx_fdct16x16_c() 326 temp1 = step3[2] * cospi_8_64 - step3[5] * cospi_24_64; in vpx_fdct16x16_c() 327 temp2 = step3[1] * cospi_24_64 + step3[6] * cospi_8_64; in vpx_fdct16x16_c() [all …]
|
D | inv_txfm.c | 102 temp1 = input[1] * cospi_24_64 - input[3] * cospi_8_64; in idct4_c() 103 temp2 = input[1] * cospi_8_64 + input[3] * cospi_24_64; in idct4_c() 178 temp1 = step1[1] * cospi_24_64 - step1[3] * cospi_8_64; in idct8_c() 179 temp2 = step1[1] * cospi_8_64 + step1[3] * cospi_24_64; in idct8_c() 327 s4 = (int)(cospi_8_64 * x4 + cospi_24_64 * x5); in iadst8_c() 328 s5 = (int)(cospi_24_64 * x4 - cospi_8_64 * x5); in iadst8_c() 329 s6 = (int)(-cospi_24_64 * x6 + cospi_8_64 * x7); in iadst8_c() 330 s7 = (int)(cospi_8_64 * x6 + cospi_24_64 * x7); in iadst8_c() 469 temp1 = step1[2] * cospi_24_64 - step1[3] * cospi_8_64; in idct16_c() 470 temp2 = step1[2] * cospi_8_64 + step1[3] * cospi_24_64; in idct16_c() [all …]
|
D | txfm_common.h | 35 static const tran_high_t cospi_8_64 = 15137; variable
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_dct.c | 36 temp1 = step[2] * cospi_24_64 + step[3] * cospi_8_64; in fdct4() 37 temp2 = -step[2] * cospi_8_64 + step[3] * cospi_24_64; in fdct4() 64 t2 = x2 * cospi_24_64 + x3 * cospi_8_64; in fdct8() 65 t3 = -x2 * cospi_8_64 + x3 * cospi_24_64; in fdct8() 143 t2 = x3 * cospi_8_64 + x2 * cospi_24_64; in fdct16() 144 t3 = x3 * cospi_24_64 - x2 * cospi_8_64; in fdct16() 194 temp1 = step3[1] * -cospi_8_64 + step3[6] * cospi_24_64; in fdct16() 195 temp2 = step3[2] * cospi_24_64 + step3[5] * cospi_8_64; in fdct16() 198 temp1 = step3[2] * cospi_8_64 - step3[5] * cospi_24_64; in fdct16() 199 temp2 = step3[1] * cospi_24_64 + step3[6] * cospi_8_64; in fdct16() [all …]
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | idct4x4_add_neon.asm | 39 ; cospi_8_64 = 15137 = 0x3b21 58 vdup.16 d20, r0 ; replicate cospi_8_64 80 vmull.s16 q1, d17, d20 ; input[1] * cospi_8_64 87 ; input[1] * cospi_24_64 - input[3] * cospi_8_64; 88 ; input[1] * cospi_8_64 + input[3] * cospi_24_64; 131 vmull.s16 q1, d17, d20 ; input[1] * cospi_8_64 138 ; input[1] * cospi_24_64 - input[3] * cospi_8_64; 139 ; input[1] * cospi_8_64 + input[3] * cospi_24_64;
|
D | idct16x16_add_neon.asm | 158 ; generate cospi_8_64 = 15137 163 vdup.16 d31, r3 ; duplicate cospi_8_64 181 ; step1[2] * cospi_24_64 - step1[3] * cospi_8_64; 182 ; step1[2] * cospi_8_64 190 ; temp2 = input[1] * cospi_8_64 + input[3] * cospi_24_64 194 ; temp1 = input[1] * cospi_24_64 - input[3] * cospi_8_64 473 ; generate cospi_8_64 = 15137 477 ; -step1[9] * cospi_8_64 + step1[14] * cospi_24_64 478 vdup.16 d30, r12 ; duplicate cospi_8_64 489 ; temp2 = step1[9] * cospi_24_64 + step1[14] * cospi_8_64 [all …]
|
D | idct8x8_add_neon.asm | 99 vdup.16 d1, r9 ; duplicate cospi_8_64 109 ; input[1] * cospi_24_64 - input[3] * cospi_8_64 114 ; input[1] * cospi_8_64 118 ; input[1] * cospi_24_64 - input[3] * cospi_8_64 122 ; input[1] * cospi_8_64 + input[3] * cospi_24_64 242 ; generate cospi_8_64 = 15137 354 ; generate cospi_8_64 = 15137 399 vdup.16 q0, r12 ; duplicate cospi_8_64*2 404 ; dct_const_round_shift(input[1] * cospi_8_64)
|
D | idct32x32_add_neon.asm | 20 cospi_8_64 EQU 15137 define 547 ;temp1 = step1b[29][i] * cospi_24_64 - step1b[18][i] * cospi_8_64; 548 ;temp2 = step1b[29][i] * cospi_8_64 + step1b[18][i] * cospi_24_64; 551 DO_BUTTERFLY_STD cospi_24_64, cospi_8_64, d0, d1, d2, d3 561 ;temp1 = step1b[28][i] * cospi_24_64 - step1b[19][i] * cospi_8_64; 562 ;temp2 = step1b[28][i] * cospi_8_64 + step1b[19][i] * cospi_24_64; 565 DO_BUTTERFLY_STD cospi_24_64, cospi_8_64, d8, d9, d12, d13 709 ;temp1 = step1b[20][i] * (-cospi_8_64) - step1b[27][i] * (-cospi_24_64); 710 ;temp2 = step1b[20][i] * (-cospi_24_64) + step1b[27][i] * (-cospi_8_64); 713 DO_BUTTERFLY_STD (-cospi_8_64), (-cospi_24_64), d10, d11, d12, d13 [all …]
|
D | fwd_txfm_neon.c | 51 v_t2_lo = vmlal_n_s16(v_t2_lo, vget_low_s16(v_x3), (int16_t)cospi_8_64); in vpx_fdct8x8_neon() 52 v_t2_hi = vmlal_n_s16(v_t2_hi, vget_high_s16(v_x3), (int16_t)cospi_8_64); in vpx_fdct8x8_neon() 53 v_t3_lo = vmlsl_n_s16(v_t3_lo, vget_low_s16(v_x2), (int16_t)cospi_8_64); in vpx_fdct8x8_neon() 54 v_t3_hi = vmlsl_n_s16(v_t3_hi, vget_high_s16(v_x2), (int16_t)cospi_8_64); in vpx_fdct8x8_neon()
|
D | idct32x32_add_neon.c | 519 DO_BUTTERFLY_STD(cospi_24_64, cospi_8_64, &q0s16, &q1s16) in vpx_idct32x32_1024_add_neon() 525 DO_BUTTERFLY_STD(cospi_24_64, cospi_8_64, &q4s16, &q6s16) in vpx_idct32x32_1024_add_neon() 586 DO_BUTTERFLY_STD(-cospi_8_64, -cospi_24_64, &q5s16, &q6s16); in vpx_idct32x32_1024_add_neon() 589 DO_BUTTERFLY_STD(-cospi_8_64, -cospi_24_64, &q0s16, &q1s16); in vpx_idct32x32_1024_add_neon() 625 DO_BUTTERFLY_STD(cospi_24_64, cospi_8_64, &q1s16, &q3s16) in vpx_idct32x32_1024_add_neon() 639 DO_BUTTERFLY_STD(-cospi_8_64, -cospi_24_64, &q4s16, &q7s16) in vpx_idct32x32_1024_add_neon() 679 DO_BUTTERFLY_STD(cospi_24_64, cospi_8_64, &q14s16, &q6s16) in vpx_idct32x32_1024_add_neon()
|
D | idct4x4_add_neon.c | 27 int16_t cospi_8_64 = 15137; in vpx_idct4x4_16_add_neon() local 46 d20s16 = vdup_n_s16(cospi_8_64); in vpx_idct4x4_16_add_neon()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_txfm_impl_sse2.h | 54 const __m128i k__cospi_C = octa_set_epi16(cospi_8_64, cospi_24_64, in FDCT4x4_2D() 55 cospi_8_64, cospi_24_64, in FDCT4x4_2D() 56 cospi_24_64, -cospi_8_64, in FDCT4x4_2D() 57 cospi_24_64, -cospi_8_64); in FDCT4x4_2D() 58 const __m128i k__cospi_D = octa_set_epi16(cospi_24_64, -cospi_8_64, in FDCT4x4_2D() 59 cospi_24_64, -cospi_8_64, in FDCT4x4_2D() 60 cospi_8_64, cospi_24_64, in FDCT4x4_2D() 61 cospi_8_64, cospi_24_64); in FDCT4x4_2D() 70 const __m128i k__cospi_G = octa_set_epi16(cospi_8_64, cospi_24_64, in FDCT4x4_2D() 71 cospi_8_64, cospi_24_64, in FDCT4x4_2D() [all …]
|
D | inv_txfm_sse2.c | 30 (int16_t)-cospi_16_64, (int16_t)cospi_24_64, (int16_t)-cospi_8_64, in vpx_idct4x4_16_add_sse2() 31 (int16_t)cospi_8_64, (int16_t)cospi_24_64); in vpx_idct4x4_16_add_sse2() 184 const __m128i k__cospi_p24_m08 = pair_set_epi16(cospi_24_64, -cospi_8_64); in idct4_sse2() 185 const __m128i k__cospi_p08_p24 = pair_set_epi16(cospi_8_64, cospi_24_64); in idct4_sse2() 465 const __m128i stg2_2 = pair_set_epi16(cospi_24_64, -cospi_8_64); in vpx_idct8x8_64_add_sse2() 466 const __m128i stg2_3 = pair_set_epi16(cospi_8_64, cospi_24_64); in vpx_idct8x8_64_add_sse2() 554 const __m128i stg2_2 = pair_set_epi16(cospi_24_64, -cospi_8_64); in idct8_sse2() 555 const __m128i stg2_3 = pair_set_epi16(cospi_8_64, cospi_24_64); in idct8_sse2() 580 const __m128i k__cospi_p08_p24 = pair_set_epi16(cospi_8_64, cospi_24_64); in iadst8_sse2() 581 const __m128i k__cospi_p24_m08 = pair_set_epi16(cospi_24_64, -cospi_8_64); in iadst8_sse2() [all …]
|
/external/libvpx/libvpx/vp9/encoder/mips/msa/ |
D | vp9_fdct_msa.h | 24 v8i16 coeff1_m = { cospi_8_64, -cospi_8_64, cospi_16_64, -cospi_16_64, \
|