/external/libvpx/libvpx/vpx_dsp/mips/ |
D | itrans8_dspr2.c | 186 [cospi_4_64] "r" (cospi_4_64), [cospi_12_64] "r" (cospi_12_64), in idct8_rows_dspr2() 435 [cospi_4_64] "r" (cospi_4_64), [cospi_12_64] "r" (cospi_12_64), in idct8_columns_add_blk_dspr2()
|
D | itrans16_dspr2.c | 264 [cospi_4_64] "r" (cospi_4_64), [cospi_28_64] "r" (cospi_28_64), in idct16_rows_dspr2() 661 [cospi_4_64] "r" (cospi_4_64), [cospi_28_64] "r" (cospi_28_64), in idct16_cols_add_blk_dspr2() 1128 s8 = x8 * cospi_4_64 + x9 * cospi_28_64; in iadst16_dspr2() 1129 s9 = x8 * cospi_28_64 - x9 * cospi_4_64; in iadst16_dspr2() 1132 s12 = - x12 * cospi_28_64 + x13 * cospi_4_64; in iadst16_dspr2() 1133 s13 = x12 * cospi_4_64 + x13 * cospi_28_64; in iadst16_dspr2()
|
D | itrans32_cols_dspr2.c | 111 [cospi_4_64] "r" (cospi_4_64), [cospi_17_64] "r" (cospi_17_64), in vpx_idct32_cols_add_blk_dspr2() 172 [cospi_4_64] "r" (cospi_4_64), [cospi_7_64] "r" (cospi_7_64), in vpx_idct32_cols_add_blk_dspr2() 660 [cospi_4_64] "r" (cospi_4_64), [cospi_28_64] "r" (cospi_28_64), in vpx_idct32_cols_add_blk_dspr2()
|
D | itrans32_dspr2.c | 157 [cospi_4_64] "r" (cospi_4_64), [cospi_17_64] "r" (cospi_17_64), in idct32_rows_dspr2() 219 [cospi_4_64] "r" (cospi_4_64), [cospi_7_64] "r" (cospi_7_64), in idct32_rows_dspr2() 735 [cospi_4_64] "r" (cospi_4_64), [cospi_28_64] "r" (cospi_28_64), in idct32_rows_dspr2()
|
D | idct8x8_msa.c | 54 k0 = VP9_SET_COSPI_PAIR(cospi_28_64, -cospi_4_64); in vpx_idct8x8_12_add_msa() 55 k1 = VP9_SET_COSPI_PAIR(cospi_4_64, cospi_28_64); in vpx_idct8x8_12_add_msa()
|
D | inv_txfm_msa.h | 218 v8i16 mask_m = { cospi_28_64, cospi_4_64, cospi_20_64, cospi_12_64, \ 219 cospi_16_64, -cospi_4_64, -cospi_20_64, -cospi_16_64 }; \ 369 k0_m = VP9_SET_COSPI_PAIR(cospi_4_64, cospi_28_64); \ 370 k1_m = VP9_SET_COSPI_PAIR(cospi_28_64, -cospi_4_64); \ 371 k2_m = VP9_SET_COSPI_PAIR(-cospi_28_64, cospi_4_64); \
|
D | idct16x16_msa.c | 27 DOTP_CONST_PAIR(reg2, reg14, cospi_28_64, cospi_4_64, reg2, reg14); in vpx_idct16_1d_rows_msa() 119 DOTP_CONST_PAIR(reg2, reg14, cospi_28_64, cospi_4_64, reg2, reg14); in vpx_idct16_1d_columns_addblk_msa() 359 k0 = VP9_SET_COSPI_PAIR(cospi_4_64, cospi_28_64); in vpx_iadst16_1d_columns_addblk_msa() 360 k1 = VP9_SET_COSPI_PAIR(cospi_28_64, -cospi_4_64); in vpx_iadst16_1d_columns_addblk_msa() 361 k2 = VP9_SET_COSPI_PAIR(-cospi_28_64, cospi_4_64); in vpx_iadst16_1d_columns_addblk_msa()
|
D | fwd_dct32x32_msa.c | 90 DOTP_CONST_PAIR(vec1, vec0, cospi_28_64, cospi_4_64, temp1, temp0); in fdct8x32_1d_column_even_store() 185 DOTP_CONST_PAIR(in21, in22, cospi_28_64, cospi_4_64, in26, in24); in fdct8x32_1d_column_odd_store() 199 DOTP_CONST_PAIR((-in23), in20, cospi_28_64, cospi_4_64, in27, in25); in fdct8x32_1d_column_odd_store() 351 DOTP_CONST_PAIR(vec1, vec0, cospi_28_64, cospi_4_64, in5, in4); in fdct8x32_1d_row_even_4x() 425 DOTP_CONST_PAIR(vec1, vec0, cospi_28_64, cospi_4_64, temp1, temp0); in fdct8x32_1d_row_even() 523 DOTP_CONST_PAIR(in21, in22, cospi_28_64, cospi_4_64, in26, in24); in fdct8x32_1d_row_odd() 539 DOTP_CONST_PAIR((-in23), in20, cospi_28_64, cospi_4_64, in27, in25); in fdct8x32_1d_row_odd() 741 DOTP_CONST_PAIR(vec1, vec0, cospi_28_64, cospi_4_64, temp1, temp0); in fdct8x32_1d_row_even_rd() 844 DOTP_CONST_PAIR(in21, in22, cospi_28_64, cospi_4_64, in26, in24); in fdct8x32_1d_row_odd_rd() 856 DOTP_CONST_PAIR((-in23), in20, cospi_28_64, cospi_4_64, in27, in25); in fdct8x32_1d_row_odd_rd()
|
D | idct32x32_msa.c | 50 DOTP_CONST_PAIR(reg1, reg7, cospi_28_64, cospi_4_64, reg1, reg7); in idct32x8_row_even_process_store() 164 DOTP_CONST_PAIR(reg7, reg0, cospi_28_64, cospi_4_64, reg0, reg7); in idct32x8_row_odd_process_store() 165 DOTP_CONST_PAIR(reg6, reg1, -cospi_4_64, cospi_28_64, reg1, reg6); in idct32x8_row_odd_process_store() 366 DOTP_CONST_PAIR(reg1, reg7, cospi_28_64, cospi_4_64, reg1, reg7); in idct8x32_column_even_process_store() 474 DOTP_CONST_PAIR(reg7, reg0, cospi_28_64, cospi_4_64, reg0, reg7); in idct8x32_column_odd_process_store() 475 DOTP_CONST_PAIR(reg6, reg1, -cospi_4_64, cospi_28_64, reg1, reg6); in idct8x32_column_odd_process_store()
|
D | fwd_txfm_msa.h | 75 cospi_24_64, cospi_4_64, cospi_28_64, \ 135 cospi_4_64, cospi_28_64, cospi_12_64, cospi_20_64 }; \
|
/external/libvpx/libvpx/vpx_dsp/ |
D | fwd_txfm.c | 156 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in vpx_fdct8x8_c() 159 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in vpx_fdct8x8_c() 292 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in vpx_fdct16x16_c() 295 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in vpx_fdct16x16_c() 614 output[4] = dct_32_round(step[4] * cospi_28_64 + step[7] * cospi_4_64); in vpx_fdct32() 617 output[7] = dct_32_round(step[7] * cospi_28_64 + step[4] * -cospi_4_64); in vpx_fdct32() 628 output[17] = dct_32_round(step[17] * -cospi_4_64 + step[30] * cospi_28_64); in vpx_fdct32() 629 output[18] = dct_32_round(step[18] * -cospi_28_64 + step[29] * -cospi_4_64); in vpx_fdct32() 640 output[29] = dct_32_round(step[29] * cospi_28_64 + step[18] * -cospi_4_64); in vpx_fdct32() 641 output[30] = dct_32_round(step[30] * cospi_4_64 + step[17] * cospi_28_64); in vpx_fdct32()
|
D | inv_txfm.c | 164 temp1 = input[1] * cospi_28_64 - input[7] * cospi_4_64; in idct8_c() 165 temp2 = input[1] * cospi_4_64 + input[7] * cospi_28_64; in idct8_c() 446 temp1 = step2[4] * cospi_28_64 - step2[7] * cospi_4_64; in idct16_c() 447 temp2 = step2[4] * cospi_4_64 + step2[7] * cospi_28_64; in idct16_c() 653 s8 = x8 * cospi_4_64 + x9 * cospi_28_64; in iadst16_c() 654 s9 = x8 * cospi_28_64 - x9 * cospi_4_64; in iadst16_c() 657 s12 = - x12 * cospi_28_64 + x13 * cospi_4_64; in iadst16_c() 658 s13 = x12 * cospi_4_64 + x13 * cospi_28_64; in iadst16_c() 906 temp1 = step2[4] * cospi_28_64 - step2[7] * cospi_4_64; in idct32_c() 907 temp2 = step2[4] * cospi_4_64 + step2[7] * cospi_28_64; in idct32_c() [all …]
|
D | txfm_common.h | 31 static const tran_high_t cospi_4_64 = 16069; variable
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_dct.c | 84 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in fdct8() 87 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in fdct8() 163 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in fdct16() 166 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in fdct16() 411 s8 = x8 * cospi_4_64 + x9 * cospi_28_64; in fadst16() 412 s9 = x8 * cospi_28_64 - x9 * cospi_4_64; in fadst16() 415 s12 = - x12 * cospi_28_64 + x13 * cospi_4_64; in fadst16() 416 s13 = x12 * cospi_4_64 + x13 * cospi_28_64; in fadst16() 623 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in vp9_fdct8x8_quant_c() 626 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in vp9_fdct8x8_quant_c()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | fwd_txfm_neon.c | 94 v_t0_lo = vmull_n_s16(vget_low_s16(v_x3), (int16_t)cospi_4_64); in vpx_fdct8x8_neon() 95 v_t0_hi = vmull_n_s16(vget_high_s16(v_x3), (int16_t)cospi_4_64); in vpx_fdct8x8_neon() 108 v_t3_lo = vmlsl_n_s16(v_t3_lo, vget_low_s16(v_x0), (int16_t)cospi_4_64); in vpx_fdct8x8_neon() 109 v_t3_hi = vmlsl_n_s16(v_t3_hi, vget_high_s16(v_x0), (int16_t)cospi_4_64); in vpx_fdct8x8_neon()
|
D | idct8x8_add_neon.asm | 27 vdup.16 d1, r4 ; duplicate cospi_4_64 39 ; input[1]*cospi_28_64-input[7]*cospi_4_64 55 ; input[1] * cospi_4_64 63 ; input[1]*cospi_4_64+input[7]*cospi_28_64 222 ; generate cospi_4_64 = 16069 334 ; generate cospi_4_64 = 16069 367 vdup.16 q1, r12 ; duplicate cospi_4_64*2 376 ; dct_const_round_shift(input[1] * cospi_4_64)
|
D | idct32x32_add_neon.asm | 16 cospi_4_64 EQU 16069 define 487 ;temp1 = step1b[30][i] * cospi_28_64 - step1b[17][i] * cospi_4_64; 488 ;temp2 = step1b[30][i] * cospi_4_64 - step1b[17][i] * cospi_28_64; 491 DO_BUTTERFLY_STD cospi_28_64, cospi_4_64, d10, d11, d14, d15 522 ;temp1 = step1b[18][i] * (-cospi_4_64) - step1b[29][i] * (-cospi_28_64); 523 ;temp2 = step1b[18][i] * (-cospi_28_64) + step1b[29][i] * (-cospi_4_64); 526 DO_BUTTERFLY_STD (-cospi_4_64), (-cospi_28_64), d2, d3, d0, d1 891 ;temp1 = input[4 * 32] * cospi_28_64 - input[28 * 32] * cospi_4_64; 892 ;temp2 = input[4 * 32] * cospi_4_64 + input[28 * 32] * cospi_28_64; 896 DO_BUTTERFLY_STD cospi_28_64, cospi_4_64, d0, d1, d4, d5
|
D | idct16x16_add_neon.asm | 67 ; generate cospi_4_64 = 16069 76 vdup.16 d1, r12 ; duplicate cospi_4_64 91 ; step2[4] * cospi_4_64 95 ; temp1 = step2[4] * cospi_28_64 - step2[7] * cospi_4_64 99 ; temp2 = step2[4] * cospi_4_64 + step2[7] * cospi_28_64 817 ; generate cospi_4_64*2 = 32138 826 vdup.16 q1, r12 ; duplicate cospi_4_64*2 840 ; dct_const_round_shift(step2[4] * cospi_4_64);
|
D | idct32x32_add_neon.c | 494 DO_BUTTERFLY_STD(cospi_28_64, cospi_4_64, &q5s16, &q7s16) in vpx_idct32x32_1024_add_neon() 508 DO_BUTTERFLY_STD(-cospi_4_64, -cospi_28_64, &q1s16, &q0s16) in vpx_idct32x32_1024_add_neon() 663 DO_BUTTERFLY_STD(cospi_28_64, cospi_4_64, &q0s16, &q2s16) in vpx_idct32x32_1024_add_neon()
|
D | idct8x8_add_neon.c | 103 d1s16 = vdup_n_s16(cospi_4_64); in IDCT8x8_1D() 383 q1s16 = vdupq_n_s16(cospi_4_64 * 2); in vpx_idct8x8_12_add_neon()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | inv_txfm_sse2.c | 459 const __m128i stg1_0 = pair_set_epi16(cospi_28_64, -cospi_4_64); in vpx_idct8x8_64_add_sse2() 460 const __m128i stg1_1 = pair_set_epi16(cospi_4_64, cospi_28_64); in vpx_idct8x8_64_add_sse2() 548 const __m128i stg1_0 = pair_set_epi16(cospi_28_64, -cospi_4_64); in idct8_sse2() 549 const __m128i stg1_1 = pair_set_epi16(cospi_4_64, cospi_28_64); in idct8_sse2() 804 const __m128i stg1_0 = pair_set_epi16(cospi_28_64, -cospi_4_64); in vpx_idct8x8_12_add_sse2() 805 const __m128i stg1_1 = pair_set_epi16(cospi_4_64, cospi_28_64); in vpx_idct8x8_12_add_sse2() 1192 const __m128i stg3_0 = pair_set_epi16(cospi_28_64, -cospi_4_64); in vpx_idct16x16_256_add_sse2() 1193 const __m128i stg3_1 = pair_set_epi16(cospi_4_64, cospi_28_64); in vpx_idct16x16_256_add_sse2() 1354 const __m128i k__cospi_p04_p28 = pair_set_epi16(cospi_4_64, cospi_28_64); in iadst16_8col() 1355 const __m128i k__cospi_p28_m04 = pair_set_epi16(cospi_28_64, -cospi_4_64); in iadst16_8col() [all …]
|
D | fwd_txfm_impl_sse2.h | 277 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in FDCT8x8_2D() 278 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in FDCT8x8_2D() 599 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in FDCT16x16_2D() 600 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in FDCT16x16_2D()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_dct_ssse3.c | 44 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in vp9_fdct8x8_quant_ssse3() 45 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in vp9_fdct8x8_quant_ssse3()
|
D | vp9_dct_sse2.c | 205 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in vp9_fdct8x8_quant_sse2() 206 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in vp9_fdct8x8_quant_sse2() 774 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in fdct8_sse2() 775 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in fdct8_sse2() 1230 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in fdct16_8col() 1231 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in fdct16_8col() 1562 const __m128i k__cospi_p04_p28 = pair_set_epi16(cospi_4_64, cospi_28_64); in fadst16_8col() 1563 const __m128i k__cospi_p28_m04 = pair_set_epi16(cospi_28_64, -cospi_4_64); in fadst16_8col() 1566 const __m128i k__cospi_m28_p04 = pair_set_epi16(-cospi_28_64, cospi_4_64); in fadst16_8col()
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_iht8x8_add_neon.c | 19 static int16_t cospi_4_64 = 16069; variable 121 d1s16 = vdup_n_s16(cospi_4_64); in IDCT8x8_1D()
|