Home
last modified time | relevance | path

Searched refs:cospi_4_64 (Results 1 – 25 of 28) sorted by relevance

12

/external/libvpx/libvpx/vpx_dsp/mips/
Ditrans16_dspr2.c264 [cospi_4_64] "r" (cospi_4_64), [cospi_28_64] "r" (cospi_28_64), in idct16_rows_dspr2()
661 [cospi_4_64] "r" (cospi_4_64), [cospi_28_64] "r" (cospi_28_64), in idct16_cols_add_blk_dspr2()
1128 s8 = x8 * cospi_4_64 + x9 * cospi_28_64; in iadst16_dspr2()
1129 s9 = x8 * cospi_28_64 - x9 * cospi_4_64; in iadst16_dspr2()
1132 s12 = - x12 * cospi_28_64 + x13 * cospi_4_64; in iadst16_dspr2()
1133 s13 = x12 * cospi_4_64 + x13 * cospi_28_64; in iadst16_dspr2()
Ditrans8_dspr2.c186 [cospi_4_64] "r" (cospi_4_64), [cospi_12_64] "r" (cospi_12_64), in idct8_rows_dspr2()
435 [cospi_4_64] "r" (cospi_4_64), [cospi_12_64] "r" (cospi_12_64), in idct8_columns_add_blk_dspr2()
Ditrans32_cols_dspr2.c111 [cospi_4_64] "r" (cospi_4_64), [cospi_17_64] "r" (cospi_17_64), in vpx_idct32_cols_add_blk_dspr2()
172 [cospi_4_64] "r" (cospi_4_64), [cospi_7_64] "r" (cospi_7_64), in vpx_idct32_cols_add_blk_dspr2()
660 [cospi_4_64] "r" (cospi_4_64), [cospi_28_64] "r" (cospi_28_64), in vpx_idct32_cols_add_blk_dspr2()
Ditrans32_dspr2.c157 [cospi_4_64] "r" (cospi_4_64), [cospi_17_64] "r" (cospi_17_64), in idct32_rows_dspr2()
219 [cospi_4_64] "r" (cospi_4_64), [cospi_7_64] "r" (cospi_7_64), in idct32_rows_dspr2()
735 [cospi_4_64] "r" (cospi_4_64), [cospi_28_64] "r" (cospi_28_64), in idct32_rows_dspr2()
Didct8x8_msa.c54 k0 = VP9_SET_COSPI_PAIR(cospi_28_64, -cospi_4_64); in vpx_idct8x8_12_add_msa()
55 k1 = VP9_SET_COSPI_PAIR(cospi_4_64, cospi_28_64); in vpx_idct8x8_12_add_msa()
Dinv_txfm_msa.h218 v8i16 mask_m = { cospi_28_64, cospi_4_64, cospi_20_64, cospi_12_64, \
219 cospi_16_64, -cospi_4_64, -cospi_20_64, -cospi_16_64 }; \
369 k0_m = VP9_SET_COSPI_PAIR(cospi_4_64, cospi_28_64); \
370 k1_m = VP9_SET_COSPI_PAIR(cospi_28_64, -cospi_4_64); \
371 k2_m = VP9_SET_COSPI_PAIR(-cospi_28_64, cospi_4_64); \
Dfwd_dct32x32_msa.c90 DOTP_CONST_PAIR(vec1, vec0, cospi_28_64, cospi_4_64, temp1, temp0); in fdct8x32_1d_column_even_store()
185 DOTP_CONST_PAIR(in21, in22, cospi_28_64, cospi_4_64, in26, in24); in fdct8x32_1d_column_odd_store()
199 DOTP_CONST_PAIR((-in23), in20, cospi_28_64, cospi_4_64, in27, in25); in fdct8x32_1d_column_odd_store()
351 DOTP_CONST_PAIR(vec1, vec0, cospi_28_64, cospi_4_64, in5, in4); in fdct8x32_1d_row_even_4x()
425 DOTP_CONST_PAIR(vec1, vec0, cospi_28_64, cospi_4_64, temp1, temp0); in fdct8x32_1d_row_even()
523 DOTP_CONST_PAIR(in21, in22, cospi_28_64, cospi_4_64, in26, in24); in fdct8x32_1d_row_odd()
539 DOTP_CONST_PAIR((-in23), in20, cospi_28_64, cospi_4_64, in27, in25); in fdct8x32_1d_row_odd()
741 DOTP_CONST_PAIR(vec1, vec0, cospi_28_64, cospi_4_64, temp1, temp0); in fdct8x32_1d_row_even_rd()
844 DOTP_CONST_PAIR(in21, in22, cospi_28_64, cospi_4_64, in26, in24); in fdct8x32_1d_row_odd_rd()
856 DOTP_CONST_PAIR((-in23), in20, cospi_28_64, cospi_4_64, in27, in25); in fdct8x32_1d_row_odd_rd()
Didct16x16_msa.c27 DOTP_CONST_PAIR(reg2, reg14, cospi_28_64, cospi_4_64, reg2, reg14); in vpx_idct16_1d_rows_msa()
119 DOTP_CONST_PAIR(reg2, reg14, cospi_28_64, cospi_4_64, reg2, reg14); in vpx_idct16_1d_columns_addblk_msa()
359 k0 = VP9_SET_COSPI_PAIR(cospi_4_64, cospi_28_64); in vpx_iadst16_1d_columns_addblk_msa()
360 k1 = VP9_SET_COSPI_PAIR(cospi_28_64, -cospi_4_64); in vpx_iadst16_1d_columns_addblk_msa()
361 k2 = VP9_SET_COSPI_PAIR(-cospi_28_64, cospi_4_64); in vpx_iadst16_1d_columns_addblk_msa()
Didct32x32_msa.c50 DOTP_CONST_PAIR(reg1, reg7, cospi_28_64, cospi_4_64, reg1, reg7); in idct32x8_row_even_process_store()
164 DOTP_CONST_PAIR(reg7, reg0, cospi_28_64, cospi_4_64, reg0, reg7); in idct32x8_row_odd_process_store()
165 DOTP_CONST_PAIR(reg6, reg1, -cospi_4_64, cospi_28_64, reg1, reg6); in idct32x8_row_odd_process_store()
366 DOTP_CONST_PAIR(reg1, reg7, cospi_28_64, cospi_4_64, reg1, reg7); in idct8x32_column_even_process_store()
474 DOTP_CONST_PAIR(reg7, reg0, cospi_28_64, cospi_4_64, reg0, reg7); in idct8x32_column_odd_process_store()
475 DOTP_CONST_PAIR(reg6, reg1, -cospi_4_64, cospi_28_64, reg1, reg6); in idct8x32_column_odd_process_store()
Dfwd_txfm_msa.h75 cospi_24_64, cospi_4_64, cospi_28_64, \
135 cospi_4_64, cospi_28_64, cospi_12_64, cospi_20_64 }; \
/external/libvpx/libvpx/vpx_dsp/
Dfwd_txfm.c156 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in vpx_fdct8x8_c()
159 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in vpx_fdct8x8_c()
292 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in vpx_fdct16x16_c()
295 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in vpx_fdct16x16_c()
614 output[4] = dct_32_round(step[4] * cospi_28_64 + step[7] * cospi_4_64); in vpx_fdct32()
617 output[7] = dct_32_round(step[7] * cospi_28_64 + step[4] * -cospi_4_64); in vpx_fdct32()
628 output[17] = dct_32_round(step[17] * -cospi_4_64 + step[30] * cospi_28_64); in vpx_fdct32()
629 output[18] = dct_32_round(step[18] * -cospi_28_64 + step[29] * -cospi_4_64); in vpx_fdct32()
640 output[29] = dct_32_round(step[29] * cospi_28_64 + step[18] * -cospi_4_64); in vpx_fdct32()
641 output[30] = dct_32_round(step[30] * cospi_4_64 + step[17] * cospi_28_64); in vpx_fdct32()
Dinv_txfm.c164 temp1 = input[1] * cospi_28_64 - input[7] * cospi_4_64; in idct8_c()
165 temp2 = input[1] * cospi_4_64 + input[7] * cospi_28_64; in idct8_c()
446 temp1 = step2[4] * cospi_28_64 - step2[7] * cospi_4_64; in idct16_c()
447 temp2 = step2[4] * cospi_4_64 + step2[7] * cospi_28_64; in idct16_c()
653 s8 = x8 * cospi_4_64 + x9 * cospi_28_64; in iadst16_c()
654 s9 = x8 * cospi_28_64 - x9 * cospi_4_64; in iadst16_c()
657 s12 = - x12 * cospi_28_64 + x13 * cospi_4_64; in iadst16_c()
658 s13 = x12 * cospi_4_64 + x13 * cospi_28_64; in iadst16_c()
906 temp1 = step2[4] * cospi_28_64 - step2[7] * cospi_4_64; in idct32_c()
907 temp2 = step2[4] * cospi_4_64 + step2[7] * cospi_28_64; in idct32_c()
[all …]
Dtxfm_common.h31 static const tran_high_t cospi_4_64 = 16069; variable
/external/libvpx/libvpx/vp9/encoder/
Dvp9_dct.c84 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in fdct8()
87 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in fdct8()
163 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in fdct16()
166 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in fdct16()
411 s8 = x8 * cospi_4_64 + x9 * cospi_28_64; in fadst16()
412 s9 = x8 * cospi_28_64 - x9 * cospi_4_64; in fadst16()
415 s12 = - x12 * cospi_28_64 + x13 * cospi_4_64; in fadst16()
416 s13 = x12 * cospi_4_64 + x13 * cospi_28_64; in fadst16()
623 t0 = x0 * cospi_28_64 + x3 * cospi_4_64; in vp9_fdct8x8_quant_c()
626 t3 = x3 * cospi_28_64 + x0 * -cospi_4_64; in vp9_fdct8x8_quant_c()
/external/libvpx/libvpx/vpx_dsp/arm/
Didct8x8_add_neon.asm27 vdup.16 d1, r4 ; duplicate cospi_4_64
39 ; input[1]*cospi_28_64-input[7]*cospi_4_64
55 ; input[1] * cospi_4_64
63 ; input[1]*cospi_4_64+input[7]*cospi_28_64
222 ; generate cospi_4_64 = 16069
334 ; generate cospi_4_64 = 16069
367 vdup.16 q1, r12 ; duplicate cospi_4_64*2
376 ; dct_const_round_shift(input[1] * cospi_4_64)
Dfwd_txfm_neon.c94 v_t0_lo = vmull_n_s16(vget_low_s16(v_x3), (int16_t)cospi_4_64); in vpx_fdct8x8_neon()
95 v_t0_hi = vmull_n_s16(vget_high_s16(v_x3), (int16_t)cospi_4_64); in vpx_fdct8x8_neon()
108 v_t3_lo = vmlsl_n_s16(v_t3_lo, vget_low_s16(v_x0), (int16_t)cospi_4_64); in vpx_fdct8x8_neon()
109 v_t3_hi = vmlsl_n_s16(v_t3_hi, vget_high_s16(v_x0), (int16_t)cospi_4_64); in vpx_fdct8x8_neon()
Didct32x32_add_neon.asm16 cospi_4_64 EQU 16069 define
487 ;temp1 = step1b[30][i] * cospi_28_64 - step1b[17][i] * cospi_4_64;
488 ;temp2 = step1b[30][i] * cospi_4_64 - step1b[17][i] * cospi_28_64;
491 DO_BUTTERFLY_STD cospi_28_64, cospi_4_64, d10, d11, d14, d15
522 ;temp1 = step1b[18][i] * (-cospi_4_64) - step1b[29][i] * (-cospi_28_64);
523 ;temp2 = step1b[18][i] * (-cospi_28_64) + step1b[29][i] * (-cospi_4_64);
526 DO_BUTTERFLY_STD (-cospi_4_64), (-cospi_28_64), d2, d3, d0, d1
891 ;temp1 = input[4 * 32] * cospi_28_64 - input[28 * 32] * cospi_4_64;
892 ;temp2 = input[4 * 32] * cospi_4_64 + input[28 * 32] * cospi_28_64;
896 DO_BUTTERFLY_STD cospi_28_64, cospi_4_64, d0, d1, d4, d5
Didct16x16_add_neon.asm67 ; generate cospi_4_64 = 16069
76 vdup.16 d1, r12 ; duplicate cospi_4_64
91 ; step2[4] * cospi_4_64
95 ; temp1 = step2[4] * cospi_28_64 - step2[7] * cospi_4_64
99 ; temp2 = step2[4] * cospi_4_64 + step2[7] * cospi_28_64
817 ; generate cospi_4_64*2 = 32138
826 vdup.16 q1, r12 ; duplicate cospi_4_64*2
840 ; dct_const_round_shift(step2[4] * cospi_4_64);
Didct32x32_add_neon.c494 DO_BUTTERFLY_STD(cospi_28_64, cospi_4_64, &q5s16, &q7s16) in vpx_idct32x32_1024_add_neon()
508 DO_BUTTERFLY_STD(-cospi_4_64, -cospi_28_64, &q1s16, &q0s16) in vpx_idct32x32_1024_add_neon()
663 DO_BUTTERFLY_STD(cospi_28_64, cospi_4_64, &q0s16, &q2s16) in vpx_idct32x32_1024_add_neon()
Didct8x8_add_neon.c103 d1s16 = vdup_n_s16(cospi_4_64); in IDCT8x8_1D()
383 q1s16 = vdupq_n_s16(cospi_4_64 * 2); in vpx_idct8x8_12_add_neon()
/external/libvpx/libvpx/vpx_dsp/x86/
Dinv_txfm_sse2.c459 const __m128i stg1_0 = pair_set_epi16(cospi_28_64, -cospi_4_64); in vpx_idct8x8_64_add_sse2()
460 const __m128i stg1_1 = pair_set_epi16(cospi_4_64, cospi_28_64); in vpx_idct8x8_64_add_sse2()
548 const __m128i stg1_0 = pair_set_epi16(cospi_28_64, -cospi_4_64); in idct8_sse2()
549 const __m128i stg1_1 = pair_set_epi16(cospi_4_64, cospi_28_64); in idct8_sse2()
804 const __m128i stg1_0 = pair_set_epi16(cospi_28_64, -cospi_4_64); in vpx_idct8x8_12_add_sse2()
805 const __m128i stg1_1 = pair_set_epi16(cospi_4_64, cospi_28_64); in vpx_idct8x8_12_add_sse2()
1192 const __m128i stg3_0 = pair_set_epi16(cospi_28_64, -cospi_4_64); in vpx_idct16x16_256_add_sse2()
1193 const __m128i stg3_1 = pair_set_epi16(cospi_4_64, cospi_28_64); in vpx_idct16x16_256_add_sse2()
1354 const __m128i k__cospi_p04_p28 = pair_set_epi16(cospi_4_64, cospi_28_64); in iadst16_8col()
1355 const __m128i k__cospi_p28_m04 = pair_set_epi16(cospi_28_64, -cospi_4_64); in iadst16_8col()
[all …]
Dfwd_txfm_impl_sse2.h277 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in FDCT8x8_2D()
278 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in FDCT8x8_2D()
599 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in FDCT16x16_2D()
600 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in FDCT16x16_2D()
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_dct_ssse3.c44 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in vp9_fdct8x8_quant_ssse3()
45 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in vp9_fdct8x8_quant_ssse3()
Dvp9_dct_sse2.c205 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in vp9_fdct8x8_quant_sse2()
206 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in vp9_fdct8x8_quant_sse2()
774 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in fdct8_sse2()
775 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in fdct8_sse2()
1230 const __m128i k__cospi_p28_p04 = pair_set_epi16(cospi_28_64, cospi_4_64); in fdct16_8col()
1231 const __m128i k__cospi_m04_p28 = pair_set_epi16(-cospi_4_64, cospi_28_64); in fdct16_8col()
1562 const __m128i k__cospi_p04_p28 = pair_set_epi16(cospi_4_64, cospi_28_64); in fadst16_8col()
1563 const __m128i k__cospi_p28_m04 = pair_set_epi16(cospi_28_64, -cospi_4_64); in fadst16_8col()
1566 const __m128i k__cospi_m28_p04 = pair_set_epi16(-cospi_28_64, cospi_4_64); in fadst16_8col()
/external/libvpx/libvpx/vp9/common/arm/neon/
Dvp9_iht8x8_add_neon.c19 static int16_t cospi_4_64 = 16069; variable
121 d1s16 = vdup_n_s16(cospi_4_64); in IDCT8x8_1D()

12