/external/libvpx/libvpx/vpx_dsp/x86/ |
D | quantize_avx.c | 34 __m128i coeff0, coeff1; in vpx_quantize_b_avx() local 51 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_avx() 54 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_avx() 84 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_avx() 95 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in vpx_quantize_b_avx() 98 store_tran_low(coeff1, dqcoeff_ptr + 8); in vpx_quantize_b_avx() 100 eob = scan_for_eob(&coeff0, &coeff1, cmp_mask0, cmp_mask1, iscan_ptr, 0, in vpx_quantize_b_avx() 107 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_avx() 110 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_avx() 130 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_avx() [all …]
|
D | quantize_ssse3.c | 30 __m128i coeff0, coeff1; in vpx_quantize_b_ssse3() local 44 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_ssse3() 47 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_ssse3() 61 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_ssse3() 72 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in vpx_quantize_b_ssse3() 75 store_tran_low(coeff1, dqcoeff_ptr + 8); in vpx_quantize_b_ssse3() 78 scan_for_eob(&coeff0, &coeff1, cmp_mask0, cmp_mask1, iscan_ptr, 0, zero); in vpx_quantize_b_ssse3() 83 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_ssse3() 86 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_ssse3() 95 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_ssse3() [all …]
|
D | quantize_sse2.c | 31 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in vpx_quantize_b_sse2() local 46 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_sse2() 50 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2() 52 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in vpx_quantize_b_sse2() 79 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in vpx_quantize_b_sse2() 82 store_tran_low(coeff1, dqcoeff_ptr + 8); in vpx_quantize_b_sse2() 85 scan_for_eob(&coeff0, &coeff1, cmp_mask0, cmp_mask1, iscan_ptr, 0, zero); in vpx_quantize_b_sse2() 90 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_sse2() 93 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2() 95 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in vpx_quantize_b_sse2() [all …]
|
D | quantize_x86.h | 51 static INLINE __m128i scan_for_eob(__m128i *coeff0, __m128i *coeff1, in scan_for_eob() argument 57 const __m128i zero_coeff1 = _mm_cmpeq_epi16(*coeff1, zero); in scan_for_eob()
|
D | avg_intrin_avx2.c | 151 const __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in vpx_hadamard_16x16_avx2() local 155 __m256i b0 = _mm256_add_epi16(coeff0, coeff1); in vpx_hadamard_16x16_avx2() 156 __m256i b1 = _mm256_sub_epi16(coeff0, coeff1); in vpx_hadamard_16x16_avx2()
|
D | avg_intrin_sse2.c | 260 __m128i coeff1 = load_tran_low(coeff + 64); in vpx_hadamard_16x16_sse2() local 264 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in vpx_hadamard_16x16_sse2() 265 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in vpx_hadamard_16x16_sse2() 275 coeff1 = _mm_add_epi16(b1, b3); in vpx_hadamard_16x16_sse2() 277 store_tran_low(coeff1, coeff + 64); in vpx_hadamard_16x16_sse2()
|
/external/libvpx/libvpx/vp8/encoder/mips/msa/ |
D | encodeopt_msa.c | 18 v8i16 coeff, dq_coeff, coeff0, coeff1; in vp8_block_error_msa() local 26 ILVRL_H2_SH(coeff, dq_coeff, coeff0, coeff1); in vp8_block_error_msa() 27 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_block_error_msa() 47 v8i16 coeff, coeff0, coeff1, coeff2, coeff3, coeff4; in vp8_mbblock_error_msa() local 79 ILVRL_H2_SH(coeff, dq_coeff, coeff0, coeff1); in vp8_mbblock_error_msa() 80 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa() 83 ILVRL_H2_SH(coeff2, dq_coeff2, coeff0, coeff1); in vp8_mbblock_error_msa() 84 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa() 91 ILVRL_H2_SH(coeff3, dq_coeff3, coeff0, coeff1); in vp8_mbblock_error_msa() 92 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa() [all …]
|
D | quantize_msa.c | 24 v8i16 coeff0, coeff1, z0, z1; in fast_quantize_b_msa() local 35 LD_SH2(coeff_ptr, 8, coeff0, coeff1); in fast_quantize_b_msa() 36 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, z0, in fast_quantize_b_msa() 38 LD_SH2(round, 8, coeff0, coeff1); in fast_quantize_b_msa() 39 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, round0, in fast_quantize_b_msa() 41 LD_SH2(quant, 8, coeff0, coeff1); in fast_quantize_b_msa() 42 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, quant0, in fast_quantize_b_msa() 93 v8i16 coeff0, coeff1, z0, z1; in exact_regular_quantize_b_msa() local 105 LD_SH2(coeff_ptr, 8, coeff0, coeff1); in exact_regular_quantize_b_msa() 106 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, z0, in exact_regular_quantize_b_msa() [all …]
|
D | denoising_msa.c | 33 v16u8 coeff0, coeff1; in vp8_denoiser_filter_msa() local 77 ILVRL_B2_UB(mc_running_avg_y0, sig0, coeff0, coeff1); in vp8_denoiser_filter_msa() 78 HSUB_UB2_SH(coeff0, coeff1, diff0, diff1); in vp8_denoiser_filter_msa() 125 ILVRL_B2_UB(mc_running_avg_y1, sig1, coeff0, coeff1); in vp8_denoiser_filter_msa() 126 HSUB_UB2_SH(coeff0, coeff1, diff0, diff1); in vp8_denoiser_filter_msa() 204 ILVRL_B2_UB(mc_running_avg_y0, sig0, coeff0, coeff1); in vp8_denoiser_filter_msa() 205 HSUB_UB2_SH(coeff0, coeff1, diff0, diff1); in vp8_denoiser_filter_msa() 240 ILVRL_B2_UB(mc_running_avg_y1, sig1, coeff0, coeff1); in vp8_denoiser_filter_msa() 241 HSUB_UB2_SH(coeff0, coeff1, diff0, diff1); in vp8_denoiser_filter_msa()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_quantize_sse2.c | 44 __m128i coeff0, coeff1; in vp9_quantize_fp_sse2() local 59 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2() 63 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_quantize_fp_sse2() 65 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); in vp9_quantize_fp_sse2() 87 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); in vp9_quantize_fp_sse2() 90 store_tran_low(coeff1, dqcoeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2() 100 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); in vp9_quantize_fp_sse2() 119 __m128i coeff0, coeff1; in vp9_quantize_fp_sse2() local 126 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2() 130 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_quantize_fp_sse2() [all …]
|
D | vp9_dct_ssse3.c | 294 __m128i coeff0, coeff1; in vp9_fdct8x8_quant_ssse3() local 309 coeff1 = *in[1]; in vp9_fdct8x8_quant_ssse3() 313 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_fdct8x8_quant_ssse3() 315 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); in vp9_fdct8x8_quant_ssse3() 337 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); in vp9_fdct8x8_quant_ssse3() 340 store_tran_low(coeff1, dqcoeff_ptr + n_coeffs + 8); in vp9_fdct8x8_quant_ssse3() 350 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); in vp9_fdct8x8_quant_ssse3() 369 __m128i coeff0, coeff1; in vp9_fdct8x8_quant_ssse3() local 377 coeff1 = *in[index + 1]; in vp9_fdct8x8_quant_ssse3() 381 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_fdct8x8_quant_ssse3() [all …]
|
D | vp9_dct_intrin_sse2.c | 462 __m128i coeff0, coeff1; in vp9_fdct8x8_quant_sse2() local 477 coeff1 = *in[1]; in vp9_fdct8x8_quant_sse2() 481 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_fdct8x8_quant_sse2() 483 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); in vp9_fdct8x8_quant_sse2() 505 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); in vp9_fdct8x8_quant_sse2() 508 _mm_store_si128((__m128i *)(dqcoeff_ptr + n_coeffs) + 1, coeff1); in vp9_fdct8x8_quant_sse2() 518 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); in vp9_fdct8x8_quant_sse2() 536 __m128i coeff0, coeff1; in vp9_fdct8x8_quant_sse2() local 544 coeff1 = *in[index + 1]; in vp9_fdct8x8_quant_sse2() 548 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_fdct8x8_quant_sse2() [all …]
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/ode/nonstiff/ |
D | MidpointStepInterpolator.java | 86 final double coeff1 = oneMinusThetaH * theta; in computeInterpolatedStateAndDerivatives() local 94 interpolatedState[i] = currentState[i] + coeff1 * yDot1 - coeff2 * yDot2; in computeInterpolatedStateAndDerivatives()
|
D | ClassicalRungeKuttaStepInterpolator.java | 92 final double coeff1 = s * ((-fourTheta + 5) * theta - 1); in computeInterpolatedStateAndDerivatives() local 103 currentState[i] + coeff1 * yDot1 + coeff23 * yDot23 + coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
|
D | ThreeEighthesStepInterpolator.java | 93 final double coeff1 = s * (1 - 7 * theta + 2 * fourTheta2); in computeInterpolatedStateAndDerivatives() local 108 currentState[i] - coeff1 * yDot1 - coeff2 * yDot2 - coeff3 * yDot3 - coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
|
D | GillStepInterpolator.java | 102 final double coeff1 = soMt * (1 - fourTheta); in computeInterpolatedStateAndDerivatives() local 118 currentState[i] - coeff1 * yDot1 - coeff2 * yDot2 - coeff3 * yDot3 - coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
|
/external/webp/src/dsp/ |
D | upsampling_neon.c | 145 const int16x8_t Y1 = vqdmulhq_lane_s16(Y0, coeff1, 0); \ 146 const int16x8_t R0 = vqdmulhq_lane_s16(V0, coeff1, 1); \ 147 const int16x8_t G0 = vqdmulhq_lane_s16(U0, coeff1, 2); \ 148 const int16x8_t G1 = vqdmulhq_lane_s16(V0, coeff1, 3); \ 210 const int16x4_t coeff1 = vld1_s16(kCoeffs1); \
|
D | enc_msa.c | 235 v8i16 coeff0, coeff1; in CollectHistogram_MSA() local 238 LD_SH2(&out[0], 8, coeff0, coeff1); in CollectHistogram_MSA() 240 coeff1 = __msa_add_a_h(coeff1, zero); in CollectHistogram_MSA() 241 SRAI_H2_SH(coeff0, coeff1, 3); in CollectHistogram_MSA() 243 coeff1 = __msa_min_s_h(coeff1, max_coeff_thr); in CollectHistogram_MSA() 244 ST_SH2(coeff0, coeff1, &out[0], 8); in CollectHistogram_MSA()
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/estimation/ |
D | LevenbergMarquardtEstimator.java | 394 double coeff1 = 0; in estimate() local 396 coeff1 += work1[j] * work1[j]; in estimate() 399 coeff1 = coeff1 / pc2; in estimate() 401 double preRed = coeff1 + 2 * coeff2; in estimate() 402 double dirDer = -(coeff1 + coeff2); in estimate()
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/optimization/general/ |
D | LevenbergMarquardtOptimizer.java | 395 double coeff1 = 0; in doOptimize() local 397 coeff1 += work1[j] * work1[j]; in doOptimize() 400 coeff1 = coeff1 / pc2; in doOptimize() 402 double preRed = coeff1 + 2 * coeff2; in doOptimize() 403 double dirDer = -(coeff1 + coeff2); in doOptimize()
|
/external/libxcam/cl_kernel/ |
D | kernel_gauss_lap_pyramid.cl | 78 float coeff1 = coeffs[i_ver + COEFF_MID - 2]; 81 result_pre[1] += tmp_data * coeff1; 84 result_cur[1] += tmp_data * coeff1; 86 result_next[1] += tmp_data * coeff1; 569 float coeff0, coeff1; 575 coeff1 = mask_coeffs[i_line + MASK_COEFF_MID - 1]; 578 result_pre[1] += tmp_data * coeff1; 582 result_cur[1] += tmp_data * coeff1; 584 result_next[1] += tmp_data * coeff1;
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | fwd_txfm_msa.c | 38 v8i16 coeff1 = { cospi_2_64, cospi_30_64, cospi_14_64, cospi_18_64, in fdct8x16_1d_column() local 88 SPLATI_H2_SH(coeff1, 0, 1, cnst0, cnst1); in fdct8x16_1d_column() 100 SPLATI_H2_SH(coeff1, 2, 3, cnst0, cnst1); in fdct8x16_1d_column() 124 SPLATI_H2_SH(coeff1, 4, 5, cnst0, cnst1); in fdct8x16_1d_column() 136 SPLATI_H2_SH(coeff1, 6, 7, cnst0, cnst1); in fdct8x16_1d_column()
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | vp8_macros_msa.h | 1706 #define DPADD_SH3_SH(in0, in1, in2, coeff0, coeff1, coeff2) \ argument 1712 out0_m = __msa_dpadd_s_h(out0_m, (v16i8)in1, (v16i8)coeff1); \
|