/external/libvpx/libvpx/vpx_dsp/mips/ |
D | variance_msa.c | 186 v8i16 avg0 = { 0 }; in sse_diff_32x64_msa() local 195 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_32x64_msa() 202 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_32x64_msa() 209 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_32x64_msa() 216 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_32x64_msa() 220 vec = __msa_hadd_s_w(avg0, avg0); in sse_diff_32x64_msa() 233 v8i16 avg0 = { 0 }; in sse_diff_64x32_msa() local 242 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_64x32_msa() 243 CALC_MSE_AVG_B(src2, ref2, var, avg0); in sse_diff_64x32_msa() 251 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_64x32_msa() [all …]
|
D | sub_pixel_variance_msa.c | 229 v8i16 avg0 = { 0 }; in avg_sse_diff_32x64_msa() local 241 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_32x64_msa() 251 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_32x64_msa() 261 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_32x64_msa() 271 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_32x64_msa() 275 vec = __msa_hadd_s_w(avg0, avg0); in avg_sse_diff_32x64_msa() 291 v8i16 avg0 = { 0 }; in avg_sse_diff_64x32_msa() local 304 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_64x32_msa() 305 CALC_MSE_AVG_B(src2, ref2, var, avg0); in avg_sse_diff_64x32_msa() 317 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_64x32_msa() [all …]
|
/external/libaom/libaom/aom_dsp/mips/ |
D | variance_msa.c | 187 v8i16 avg0 = { 0 }; in sse_diff_32x64_msa() local 196 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_32x64_msa() 203 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_32x64_msa() 210 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_32x64_msa() 217 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_32x64_msa() 221 vec = __msa_hadd_s_w(avg0, avg0); in sse_diff_32x64_msa() 234 v8i16 avg0 = { 0 }; in sse_diff_64x32_msa() local 243 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_64x32_msa() 244 CALC_MSE_AVG_B(src2, ref2, var, avg0); in sse_diff_64x32_msa() 252 CALC_MSE_AVG_B(src0, ref0, var, avg0); in sse_diff_64x32_msa() [all …]
|
D | sub_pixel_variance_msa.c | 226 v8i16 avg0 = { 0 }; in avg_sse_diff_32x64_msa() local 238 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_32x64_msa() 248 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_32x64_msa() 258 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_32x64_msa() 268 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_32x64_msa() 272 vec = __msa_hadd_s_w(avg0, avg0); in avg_sse_diff_32x64_msa() 288 v8i16 avg0 = { 0 }; in avg_sse_diff_64x32_msa() local 301 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_64x32_msa() 302 CALC_MSE_AVG_B(src2, ref2, var, avg0); in avg_sse_diff_64x32_msa() 314 CALC_MSE_AVG_B(src0, ref0, var, avg0); in avg_sse_diff_64x32_msa() [all …]
|
/external/tensorflow/tensorflow/python/training/ |
D | moving_averages_test.py | 187 avg0 = ema.average(var0) 193 self.assertNotIn(avg0, variables.trainable_variables()) 198 self.assertEqual("v0/ExponentialMovingAverage:0", avg0.name) 208 self.assertAllClose(tens, self.evaluate(avg0)) 219 self.assertAllClose(expected, self.evaluate(avg0)) 229 self.assertAllClose(expected, self.evaluate(avg0))
|
/external/webp/src/dsp/ |
D | lossless_neon.c | 185 const uint8x8_t avg0 = Average2_u8_NEON(a0, a2); in Average3_NEON() local 187 const uint32_t avg = GET_U8_AS_U32(vhadd_u8(avg0, A1)); in Average3_NEON()
|