/external/tensorflow/tensorflow/core/grappler/optimizers/ |
D | remapper_test.cc | 152 const auto fused_ops = node.attr().at("fused_ops").list().s(); in TEST_F() local 153 EXPECT_EQ(1, fused_ops.size()); in TEST_F() 154 EXPECT_EQ("BiasAdd", fused_ops[0]); in TEST_F() 216 const auto fused_ops = node.attr().at("fused_ops").list().s(); in TEST_F() local 217 ASSERT_EQ(2, fused_ops.size()); in TEST_F() 218 EXPECT_EQ("BiasAdd", fused_ops[0]); in TEST_F() 219 EXPECT_EQ("Relu", fused_ops[1]); in TEST_F() 294 const auto fused_ops = node.attr().at("fused_ops").list().s(); in TEST_F() local 295 EXPECT_EQ(1, fused_ops.size()); in TEST_F() 296 EXPECT_EQ("FusedBatchNorm", fused_ops[0]); in TEST_F() [all …]
|
D | remapper.cc | 479 NodeDef* fused_conv2d, const absl::Span<const absl::string_view> fused_ops, in SetFusedConv2DAttributes() argument 482 SetAttrValue(fused_ops, &(*attr)["fused_ops"]); in SetFusedConv2DAttributes()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | conv_ops_fused_impl.h | 840 std::vector<string> fused_ops; 841 OP_REQUIRES_OK(context, context->GetAttr("fused_ops", &fused_ops)); 842 OP_REQUIRES(context, !fused_ops.empty(), 853 if (FusedOpsMatchAndSupportedOnDevice(fused_ops, {"BiasAdd"}, 856 } else if (FusedOpsMatchAndSupportedOnDevice(fused_ops, {"BiasAdd", "Relu"}, 859 } else if (FusedOpsMatchAndSupportedOnDevice(fused_ops, {"FusedBatchNorm"}, 862 } else if (FusedOpsMatchAndSupportedOnDevice(fused_ops, 869 absl::StrJoin(fused_ops, ","), "]")); 939 bool FusedOpsMatchAndSupportedOnDevice(const std::vector<string>& fused_ops, 945 return fused_ops == expected;
|
D | mkl_fused_ops_test.cc | 173 const std::vector<string>& fused_ops, Tensor* output, in RunMklFusedConv2DOp() argument 189 .Attr("fused_ops", fused_ops) in RunMklFusedConv2DOp() 597 const std::vector<string>& fused_ops, in RunMklFusedConv2DWithPadOp() argument 619 .Attr("fused_ops", fused_ops) in RunMklFusedConv2DWithPadOp()
|
D | conv_ops_test.cc | 709 const std::vector<string>& fused_ops, Tensor* output, in RunFusedConv2DOp() argument 737 .Attr("fused_ops", fused_ops) in RunFusedConv2DOp() 1233 const std::vector<string>& fused_ops = {}) { in FusedConv2DWithBias() argument 1255 .Attr("fused_ops", fused_ops) 1265 int out_depth, const std::vector<string>& fused_ops = {}) { in FusedConv2DWithBatchNorm() argument 1293 .Attr("fused_ops", fused_ops)
|
D | mkl_conv_ops.cc | 1465 std::vector<string> fused_ops; in MklFusedConvOp() local 1466 OP_REQUIRES_OK(context, context->GetAttr("fused_ops", &fused_ops)); in MklFusedConvOp() 1470 OP_REQUIRES(context, !fused_ops.empty(), in MklFusedConvOp() 1474 if (fused_ops == std::vector<string>{"BiasAdd"}) { in MklFusedConvOp() 1479 } else if (fused_ops == std::vector<string>{"Relu"}) { in MklFusedConvOp() 1481 } else if (fused_ops == std::vector<string>{"BiasAdd", "Relu"}) { in MklFusedConvOp() 1490 str_util::Join(fused_ops, ","), "]")); in MklFusedConvOp()
|
/external/tensorflow/tensorflow/core/graph/ |
D | mkl_layout_pass.cc | 1321 std::vector<string> fused_ops; in FusedConv2DRewrite() local 1322 TF_CHECK_OK(GetNodeAttr(n->def(), "fused_ops", &fused_ops)); in FusedConv2DRewrite() 1323 return (fused_ops == std::vector<string>{"BiasAdd"} || in FusedConv2DRewrite() 1324 fused_ops == std::vector<string>{"Relu"} || in FusedConv2DRewrite() 1325 fused_ops == std::vector<string>{"BiasAdd", "Relu"}); in FusedConv2DRewrite() 2183 std::vector<string> fused_ops; in CopyAttrsFromPadAndFusedConv2D() local 2193 TF_CHECK_OK(GetNodeAttr(fused_conv2d->def(), "fused_ops", &fused_ops)); in CopyAttrsFromPadAndFusedConv2D() 2206 nb->Attr("fused_ops", fused_ops); in CopyAttrsFromPadAndFusedConv2D() 2591 std::vector<string> fused_ops; in CopyAttrsFusedConv2D() local 2600 TF_CHECK_OK(GetNodeAttr(orig_node->def(), "fused_ops", &fused_ops)); in CopyAttrsFusedConv2D() [all …]
|