Searched refs:KeepDims (Results 1 – 7 of 7) sorted by relevance
/external/tensorflow/tensorflow/lite/kernels/ |
D | reduce_test.cc | 247 TEST(ConstFloatMeanOpTest, KeepDims) { in TEST() argument 345 TEST(DynamicFloatMeanOpTest, KeepDims) { in TEST() argument 387 TEST(ConstUint8MeanOpTest, KeepDims) { in TEST() argument 450 TEST(DynamicUint8MeanOpTest, KeepDims) { in TEST() argument 508 TEST(ConstFloatSumOpTest, KeepDims) { in TEST() argument 547 TEST(DynamicFloatSumOpTest, KeepDims) { in TEST() argument 600 TEST(ConstUint8SumOpTest, KeepDims) { in TEST() argument 629 TEST(DynamicUint8SumOpTest, KeepDims) { in TEST() argument 661 TEST(ConstFloatProdOpTest, KeepDims) { in TEST() argument 692 TEST(DynamicFloatProdOpTest, KeepDims) { in TEST() argument [all …]
|
/external/tensorflow/tensorflow/core/grappler/optimizers/ |
D | shape_optimizer_test.cc | 37 Output e = ops::ReduceProd(s.WithOpName("e"), c, d, attrs.KeepDims(false)); in TEST_F() 38 Output f = ops::ReduceProd(s.WithOpName("f"), c, d, attrs.KeepDims(true)); in TEST_F()
|
D | layout_optimizer.cc | 1890 if (KeepDims()) { in AddLayoutTransposeToOutputs() 1898 return KeepDims() || ((IsAlongAllFourDims() || IsAlongHWC() || in IsReduceAxisSupported() 1900 !KeepDims()); in IsReduceAxisSupported() 1935 bool KeepDims() const { return node_->attr().at("keep_dims").b(); } in KeepDims() function in tensorflow::grappler::__anonbd4a17fb0111::ReduceProcessor
|
D | constant_folding_test.cc | 2434 attr = attr.KeepDims(true); in TEST_F() 2490 attr = attr.KeepDims(false); in TEST_F() 2493 one_axis, attr.KeepDims(false)); in TEST_F() 2495 multiple_axes, attr.KeepDims(false)); in TEST_F() 2498 one_axis, attr.KeepDims(false)); in TEST_F() 2501 variable_axis, attr.KeepDims(false)); in TEST_F() 2504 multiple_axes, attr.KeepDims(true)); in TEST_F()
|
/external/tensorflow/tensorflow/tools/graph_transforms/ |
D | quantize_nodes_test.cc | 528 dequantize_reduction_dims, Min::Attrs().KeepDims(false)); in TestRemoveRedundantQuantization() 532 dequantize_reduction_dims, Max::Attrs().KeepDims(false)); in TestRemoveRedundantQuantization() 627 dequantize_reduction_dims, Min::Attrs().KeepDims(false)); in TestRemoveRedundantQuantizationWithBiasAdd() 631 dequantize_reduction_dims, Max::Attrs().KeepDims(false)); in TestRemoveRedundantQuantizationWithBiasAdd() 753 dequantize_reduction_dims, Min::Attrs().KeepDims(false)); in TestRemoveRedundantQuantizationWithMultipleOutputs() 757 dequantize_reduction_dims, Max::Attrs().KeepDims(false)); in TestRemoveRedundantQuantizationWithMultipleOutputs() 1167 quantize_reduction_dims1, Min::Attrs().KeepDims(false)); in TestMergeDuplicateQuantizes() 1171 quantize_reduction_dims1, Max::Attrs().KeepDims(false)); in TestMergeDuplicateQuantizes() 1194 quantize_reduction_dims2, Min::Attrs().KeepDims(false)); in TestMergeDuplicateQuantizes() 1198 quantize_reduction_dims2, Max::Attrs().KeepDims(false)); in TestMergeDuplicateQuantizes()
|
/external/tensorflow/tensorflow/cc/gradients/ |
D | nn_grad.cc | 120 auto sum = Sum(scope, grad_inputs[0], {1}, Sum::KeepDims(true)); in LogSoftmaxGrad()
|
/external/tensorflow/tensorflow/lite/toco/graph_transformations/ |
D | propagate_fixed_sizes.cc | 532 bool KeepDims(const Operator& op) { in KeepDims() function 563 const bool keep_dims = KeepDims(*op); in ProcessTensorFlowReductionOperator()
|