/external/XNNPACK/test/ |
D | qs8-vaddc-minmax.cc | 114 TEST(QS8_VADDC_MINMAX__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QS8_VADDC_MINMAX__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QS8_VADDC_MINMAX__NEON_LD64_X24, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 531 TEST(QS8_VADDC_MINMAX__NEON_LD64_X32, b_scale) { in TEST() argument 534 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 670 TEST(QS8_VADDC_MINMAX__NEON_LD128_X16, b_scale) { in TEST() argument 673 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
D | qs8-vmulc-minmax-fp32.cc | 114 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD128_X16, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 531 TEST(QS8_VMULC_MINMAX_FP32__NEONV8_LD64_X8, b_scale) { in TEST() argument 534 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 670 TEST(QS8_VMULC_MINMAX_FP32__NEONV8_LD64_X16, b_scale) { in TEST() argument 673 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
D | qu8-vmulc-minmax-fp32.cc | 114 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD128_X16, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 531 TEST(QU8_VMULC_MINMAX_FP32__NEONV8_LD64_X8, b_scale) { in TEST() argument 534 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 670 TEST(QU8_VMULC_MINMAX_FP32__NEONV8_LD64_X16, b_scale) { in TEST() argument 673 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
D | qu8-vaddc-minmax.cc | 114 TEST(QU8_VADDC_MINMAX__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QU8_VADDC_MINMAX__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QU8_VADDC_MINMAX__NEON_LD64_X32, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 531 TEST(QU8_VADDC_MINMAX__NEON_LD128_X16, b_scale) { in TEST() argument 534 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 670 TEST(QU8_VADDC_MINMAX__SSE2_MUL16_LD64_X8, b_scale) { in TEST() argument 673 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
D | qs8-vadd-minmax.cc | 135 TEST(QS8_VADD_MINMAX__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QS8_VADD_MINMAX__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QS8_VADD_MINMAX__NEON_LD64_X24, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 615 TEST(QS8_VADD_MINMAX__NEON_LD64_X32, b_scale) { in TEST() argument 618 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 775 TEST(QS8_VADD_MINMAX__NEON_LD128_X16, b_scale) { in TEST() argument 778 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
D | qu8-vadd-minmax.cc | 135 TEST(QU8_VADD_MINMAX__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QU8_VADD_MINMAX__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QU8_VADD_MINMAX__NEON_LD64_X32, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 615 TEST(QU8_VADD_MINMAX__NEON_LD128_X16, b_scale) { in TEST() argument 618 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 775 TEST(QU8_VADD_MINMAX__SSE2_MUL16_LD64_X8, b_scale) { in TEST() argument 778 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
D | vaddc-microkernel-tester.h | 69 inline VAddCMicrokernelTester& b_scale(float b_scale) { in b_scale() argument 76 inline float b_scale() const { in b_scale() function
|
D | qu8-vmul-minmax-fp32.cc | 135 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD128_X16, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 615 TEST(QU8_VMUL_MINMAX_FP32__NEONV8_LD64_X8, b_scale) { in TEST() argument 618 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 775 TEST(QU8_VMUL_MINMAX_FP32__NEONV8_LD64_X16, b_scale) { in TEST() argument 778 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
D | qs8-vmul-minmax-fp32.cc | 135 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD128_X16, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 615 TEST(QS8_VMUL_MINMAX_FP32__NEONV8_LD64_X8, b_scale) { in TEST() argument 618 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 775 TEST(QS8_VMUL_MINMAX_FP32__NEONV8_LD64_X16, b_scale) { in TEST() argument 778 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
D | vadd-microkernel-tester.h | 78 inline VAddMicrokernelTester& b_scale(float b_scale) { in b_scale() argument 85 inline float b_scale() const { in b_scale() function
|
D | vmulc-microkernel-tester.h | 66 inline VMulCMicrokernelTester& b_scale(float b_scale) { in b_scale() function 73 inline float b_scale() const { in b_scale() function
|
D | qu8-vmulc-minmax-rndnu.cc | 114 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD128_X16, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local
|
D | qs8-vmulc-minmax-rndnu.cc | 114 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD128_X16, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local
|
D | vmul-microkernel-tester.h | 75 inline VMulMicrokernelTester& b_scale(float b_scale) { in b_scale() argument 82 inline float b_scale() const { in b_scale() function
|
D | qu8-vmul-minmax-rndnu.cc | 135 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD128_X16, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local
|
D | qs8-vmul-minmax-rndnu.cc | 135 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD128_X16, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local
|
/external/libyuv/files/source/ |
D | row_common.cc | 855 const uint32_t b_scale = REPEAT8(value & 0xff); in ARGBShadeRow_C() local 890 const uint32_t b_scale = src_argb1[0]; in ARGBMultiplyRow_C() local
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | row_common.cc | 855 const uint32_t b_scale = REPEAT8(value & 0xff); in ARGBShadeRow_C() local 890 const uint32_t b_scale = src_argb1[0]; in ARGBMultiplyRow_C() local
|
/external/libaom/libaom/third_party/libyuv/source/ |
D | row_common.cc | 1124 const uint32_t b_scale = REPEAT8(value & 0xff); in ARGBShadeRow_C() local 1159 const uint32_t b_scale = src_argb1[0]; in ARGBMultiplyRow_C() local
|