Home
last modified time | relevance | path

Searched refs:op_def (Results 1 – 25 of 277) sorted by relevance

12345678910>>...12

/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/
Dsimple_selectors.cc49 std::unique_ptr<GPUOperation> SelectLSTM(const OperationDef& op_def, in SelectLSTM() argument
51 return absl::make_unique<GPUOperation>(CreateLSTM(op_def, gpu_info)); in SelectLSTM()
55 const OperationDef& op_def) { in SelectReLU() argument
56 return absl::make_unique<GPUOperation>(CreateReLU(op_def, attr)); in SelectReLU()
61 const OperationDef& op_def) { in SelectPReLU() argument
62 return absl::make_unique<GPUOperation>(CreatePReLU(gpu_info, op_def, attr)); in SelectPReLU()
66 const OperationDef& op_def) { in SelectPooling() argument
67 return absl::make_unique<GPUOperation>(CreatePooling(op_def, attr)); in SelectPooling()
71 const MaxUnpooling2DAttributes& attr, const OperationDef& op_def) { in SelectMaxUnpooling() argument
72 return absl::make_unique<GPUOperation>(CreateMaxUnpooling(op_def, attr)); in SelectMaxUnpooling()
[all …]
Dsimple_selectors.h30 std::unique_ptr<GPUOperation> SelectLSTM(const OperationDef& op_def,
34 const OperationDef& op_def);
38 const OperationDef& op_def);
41 const OperationDef& op_def);
44 const MaxUnpooling2DAttributes& attr, const OperationDef& op_def);
46 void SelectAdd(const OperationDef& op_def, const std::vector<int>& channels,
50 const OperationDef& op_def,
55 const OperationDef& op_def, const GpuInfo& gpu_info,
60 const OperationDef& op_def);
63 const OperationDef& op_def,
[all …]
Doperation_selector.cc82 const OperationDef& op_def, ModelHints hints, in WinogradFromNode() argument
99 gpu_info, shape_0, op_def.src_tensors[0].storage_type, in WinogradFromNode()
100 op_def.src_tensors[0].data_type, op_def.src_tensors[0].layout); in WinogradFromNode()
101 td_0.data_type = op_def.src_tensors[0].data_type; in WinogradFromNode()
102 td_0.layout = op_def.src_tensors[0].layout; in WinogradFromNode()
105 gpu_info, shape_1, op_def.src_tensors[0].storage_type, in WinogradFromNode()
106 op_def.src_tensors[0].data_type, op_def.src_tensors[0].layout); in WinogradFromNode()
107 td_1.data_type = op_def.src_tensors[0].data_type; in WinogradFromNode()
108 td_1.layout = op_def.src_tensors[0].layout; in WinogradFromNode()
114 winograd_up_def.precision = op_def.precision; in WinogradFromNode()
[all …]
/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/default/
Dconvolution_selector.cc38 const GpuInfo& gpu_info, const OperationDef& op_def, in SelectConvolutionAdreno() argument
40 if (IsConvConstantsSupported(gpu_info, op_def, attr)) { in SelectConvolutionAdreno()
41 GPUOperation conv = CreateConvConstants(gpu_info, op_def, attr); in SelectConvolutionAdreno()
44 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionAdreno()
51 const GpuInfo& gpu_info, const OperationDef& op_def, in SelectConvolutionWinogradAdreno() argument
54 CreateConvPowerVRWino4x4To6x6(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionWinogradAdreno()
61 const OperationDef& op_def, ModelHints hints, in SelectConvolutionDynamicWeightsAdreno() argument
64 gpu_info, op_def, attr, weights_shape, &dst_shape); in SelectConvolutionDynamicWeightsAdreno()
71 const GpuInfo& gpu_info, const OperationDef& op_def) { in SelectConvolutionNVidia() argument
72 if (IsConvConstantsSupported(gpu_info, op_def, attr)) { in SelectConvolutionNVidia()
[all …]
Dconvolution_transposed_selector.cc30 const OperationDef& op_def) { in SelectConvolutionTransposedAdreno() argument
33 CreateConvolutionTransposedThin(gpu_info, op_def, attr); in SelectConvolutionTransposedAdreno()
37 CreateConvolutionTransposed3x3Thin(gpu_info, op_def, attr); in SelectConvolutionTransposedAdreno()
41 CreateConvolutionTransposed(gpu_info, op_def, attr); in SelectConvolutionTransposedAdreno()
48 const OperationDef& op_def) { in SelectConvolutionTransposedPowerVR() argument
51 CreateConvolutionTransposedThin(gpu_info, op_def, attr); in SelectConvolutionTransposedPowerVR()
55 CreateConvolutionTransposed3x3Thin(gpu_info, op_def, attr); in SelectConvolutionTransposedPowerVR()
57 } else if (IsConvolutionTransposed3x3Supported(op_def, attr)) { in SelectConvolutionTransposedPowerVR()
59 CreateConvolutionTransposed3x3(gpu_info, op_def, attr); in SelectConvolutionTransposedPowerVR()
61 } else if (IsConvolutionTransposed4x4Supported(op_def, attr)) { in SelectConvolutionTransposedPowerVR()
[all …]
Dfully_connected_selector.cc29 const OperationDef& op_def, int batch_size) { in SelectFullyConnectedGeneric() argument
30 if (op_def.IsBatchSupported()) { in SelectFullyConnectedGeneric()
32 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedGeneric()
35 FullyConnected fc = CreateFullyConnected(gpu_info, op_def, attr); in SelectFullyConnectedGeneric()
42 const OperationDef& op_def, int batch_size) { in SelectFullyConnectedAdreno() argument
43 if (op_def.IsBatchSupported()) { in SelectFullyConnectedAdreno()
45 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedAdreno()
48 FullyConnected fc = CreateFullyConnected(gpu_info, op_def, attr); in SelectFullyConnectedAdreno()
55 const OperationDef& op_def, int batch_size) { in SelectFullyConnectedPowerVR() argument
56 if (op_def.IsBatchSupported()) { in SelectFullyConnectedPowerVR()
[all …]
Ddw_convolution_selector.cc28 const OperationDef& op_def) { in SelectDWConvolutionAdreno() argument
31 CreateDepthwiseConv3x3(gpu_info, op_def, attr)); in SelectDWConvolutionAdreno()
34 CreateDepthwiseConvolution2D(gpu_info, op_def, attr)); in SelectDWConvolutionAdreno()
40 const OperationDef& op_def) { in SelectDWConvolutionPowerVR() argument
43 CreateDepthwiseConv3x3(gpu_info, op_def, attr)); in SelectDWConvolutionPowerVR()
46 CreateDepthwiseConvolution2D(gpu_info, op_def, attr)); in SelectDWConvolutionPowerVR()
52 const OperationDef& op_def) { in SelectDWConvolutionMali() argument
53 const auto storage_type = op_def.src_tensors[0].storage_type; in SelectDWConvolutionMali()
58 !buffer_type && op_def.precision != CalculationsPrecision::F32) { in SelectDWConvolutionMali()
60 CreateDepthwiseConv3x3(gpu_info, op_def, attr)); in SelectDWConvolutionMali()
[all …]
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/
Delementwise_test_util.cc35 OperationDef op_def; in AbsTest() local
36 op_def.precision = precision; in AbsTest()
38 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AbsTest()
39 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in AbsTest()
42 env->GetGpuInfo(), op_def, OperationType::ABS); in AbsTest()
62 OperationDef op_def; in CosTest() local
63 op_def.precision = precision; in CosTest()
65 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in CosTest()
66 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in CosTest()
69 env->GetGpuInfo(), op_def, OperationType::COS); in CosTest()
[all …]
Dpadding_test_util.cc40 OperationDef op_def; in PaddingAppendWidthTest() local
41 op_def.precision = precision; in PaddingAppendWidthTest()
43 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingAppendWidthTest()
44 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingAppendWidthTest()
46 GPUOperation operation = CreatePadding(op_def, attr); in PaddingAppendWidthTest()
70 OperationDef op_def; in PaddingPrependWidthTest() local
71 op_def.precision = precision; in PaddingPrependWidthTest()
73 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingPrependWidthTest()
74 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingPrependWidthTest()
76 GPUOperation operation = CreatePadding(op_def, attr); in PaddingPrependWidthTest()
[all …]
Dresize_test_util.cc41 OperationDef op_def; in ResizeBilinearAlignedTest() local
42 op_def.precision = precision; in ResizeBilinearAlignedTest()
44 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearAlignedTest()
45 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearAlignedTest()
47 Resize operation = CreateResize(op_def, attr); in ResizeBilinearAlignedTest()
73 OperationDef op_def; in ResizeBilinearNonAlignedTest() local
74 op_def.precision = precision; in ResizeBilinearNonAlignedTest()
76 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearNonAlignedTest()
77 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearNonAlignedTest()
79 Resize operation = CreateResize(op_def, attr); in ResizeBilinearNonAlignedTest()
[all …]
Drelu_test_util.cc40 OperationDef op_def; in ReLUNoClipNoAlphaTest() local
41 op_def.precision = precision; in ReLUNoClipNoAlphaTest()
43 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReLUNoClipNoAlphaTest()
44 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ReLUNoClipNoAlphaTest()
46 GPUOperation operation = CreateReLU(op_def, attr); in ReLUNoClipNoAlphaTest()
69 OperationDef op_def; in ReLUClipTest() local
70 op_def.precision = precision; in ReLUClipTest()
72 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReLUClipTest()
73 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ReLUClipTest()
75 GPUOperation operation = CreateReLU(op_def, attr); in ReLUClipTest()
[all …]
Dreduce_test_util.cc37 OperationDef op_def; in MeanHWTest() local
38 op_def.precision = precision; in MeanHWTest()
40 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in MeanHWTest()
41 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in MeanHWTest()
44 CreateReduce(axis, src_tensor.shape, OperationType::MEAN, op_def, in MeanHWTest()
64 OperationDef op_def; in ReduceSumChannelsTest() local
65 op_def.precision = precision; in ReduceSumChannelsTest()
67 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReduceSumChannelsTest()
68 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ReduceSumChannelsTest()
72 op_def, env->GetGpuInfo()); in ReduceSumChannelsTest()
[all …]
Dconcat_test_util.cc42 OperationDef op_def; in ConcatWidthTest() local
43 op_def.precision = precision; in ConcatWidthTest()
45 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatWidthTest()
46 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatWidthTest()
47 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatWidthTest()
49 GPUOperation operation = CreateConcatXY(op_def, attr); in ConcatWidthTest()
75 OperationDef op_def; in ConcatHeightTest() local
76 op_def.precision = precision; in ConcatHeightTest()
78 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatHeightTest()
79 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatHeightTest()
[all …]
Dquantize_and_dequantize_test_util.cc45 OperationDef op_def; in QuantAndDequant_Dim2Bits8Test() local
46 op_def.precision = precision; in QuantAndDequant_Dim2Bits8Test()
48 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in QuantAndDequant_Dim2Bits8Test()
49 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in QuantAndDequant_Dim2Bits8Test()
51 GPUOperation operation = CreateQuantizeAndDequantize(op_def, attr); in QuantAndDequant_Dim2Bits8Test()
82 OperationDef op_def; in QuantAndDequant_Dim3Bits8_NegativeRangeTest() local
83 op_def.precision = precision; in QuantAndDequant_Dim3Bits8_NegativeRangeTest()
85 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in QuantAndDequant_Dim3Bits8_NegativeRangeTest()
86 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in QuantAndDequant_Dim3Bits8_NegativeRangeTest()
88 GPUOperation operation = CreateQuantizeAndDequantize(op_def, attr); in QuantAndDequant_Dim3Bits8_NegativeRangeTest()
[all …]
Dconv_weights_converter_test_util.cc32 const OperationDef& op_def) { in ConvolutionWeightsConverterTest() argument
93 auto converter = ConverterToConvWeights(op_def, weight_desc); in ConvolutionWeightsConverterTest()
126 OperationDef op_def; in ConverterToConvWeights1x1OutX4Test() local
127 op_def.precision = precision; in ConverterToConvWeights1x1OutX4Test()
129 op_def.src_tensors.push_back({data_type, storage, Layout::BHWC}); in ConverterToConvWeights1x1OutX4Test()
130 op_def.dst_tensors.push_back( in ConverterToConvWeights1x1OutX4Test()
133 weights, conv_weight_desc, env, op_def)); in ConverterToConvWeights1x1OutX4Test()
160 OperationDef op_def; in ConverterToConvWeights1x1OutX4UnalignedTest() local
161 op_def.precision = precision; in ConverterToConvWeights1x1OutX4UnalignedTest()
163 op_def.src_tensors.push_back({data_type, storage, Layout::BHWC}); in ConverterToConvWeights1x1OutX4UnalignedTest()
[all …]
Dadd_test_util.cc37 OperationDef op_def; in AddTwoEqualTensorsTest() local
38 op_def.precision = precision; in AddTwoEqualTensorsTest()
40 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddTwoEqualTensorsTest()
41 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddTwoEqualTensorsTest()
42 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in AddTwoEqualTensorsTest()
44 GPUOperation operation = CreateAdd(op_def, channels, channels[0]); in AddTwoEqualTensorsTest()
67 OperationDef op_def; in AddFirstTensorHasMoreChannelsThanSecondTest() local
68 op_def.precision = precision; in AddFirstTensorHasMoreChannelsThanSecondTest()
70 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddFirstTensorHasMoreChannelsThanSecondTest()
71 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddFirstTensorHasMoreChannelsThanSecondTest()
[all …]
Dpooling.cc27 std::string GetAveragePoolingKernelCode(const OperationDef& op_def, in GetAveragePoolingKernelCode() argument
30 auto src_desc = op_def.src_tensors[0]; in GetAveragePoolingKernelCode()
32 if (op_def.IsBatchSupported()) { in GetAveragePoolingKernelCode()
36 auto dst_desc = op_def.dst_tensors[0]; in GetAveragePoolingKernelCode()
37 if (op_def.IsBatchSupported()) { in GetAveragePoolingKernelCode()
55 if (op_def.dst_tensors[0].HasAxis(axis)) { in GetAveragePoolingKernelCode()
58 if (op_def.src_tensors[0].HasAxis(axis)) { in GetAveragePoolingKernelCode()
72 op_def.src_tensors[0].storage_type == TensorStorageType::BUFFER || in GetAveragePoolingKernelCode()
73 op_def.src_tensors[0].storage_type == TensorStorageType::IMAGE_BUFFER; in GetAveragePoolingKernelCode()
78 if (op_def.dst_tensors[0].HasAxis(Axis::DEPTH)) { in GetAveragePoolingKernelCode()
[all …]
Dspace_to_depth_test_util.cc36 OperationDef op_def; in SpaceToDepthTensorShape1x2x2x1BlockSize2Test() local
37 op_def.precision = precision; in SpaceToDepthTensorShape1x2x2x1BlockSize2Test()
39 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in SpaceToDepthTensorShape1x2x2x1BlockSize2Test()
40 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in SpaceToDepthTensorShape1x2x2x1BlockSize2Test()
42 GPUOperation operation = CreateSpaceToDepth(op_def, attr); in SpaceToDepthTensorShape1x2x2x1BlockSize2Test()
63 OperationDef op_def; in SpaceToDepthTensorShape1x2x2x2BlockSize2Test() local
64 op_def.precision = precision; in SpaceToDepthTensorShape1x2x2x2BlockSize2Test()
66 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in SpaceToDepthTensorShape1x2x2x2BlockSize2Test()
67 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in SpaceToDepthTensorShape1x2x2x2BlockSize2Test()
69 GPUOperation operation = CreateSpaceToDepth(op_def, attr); in SpaceToDepthTensorShape1x2x2x2BlockSize2Test()
[all …]
Dpooling_test_util.cc43 OperationDef op_def; in AveragePoolingTest() local
44 op_def.precision = precision; in AveragePoolingTest()
46 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AveragePoolingTest()
47 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in AveragePoolingTest()
49 GPUOperation operation = CreatePooling(op_def, attr); in AveragePoolingTest()
74 OperationDef op_def; in AveragePoolingNonEmptyPaddingTest() local
75 op_def.precision = precision; in AveragePoolingNonEmptyPaddingTest()
77 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AveragePoolingNonEmptyPaddingTest()
78 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in AveragePoolingNonEmptyPaddingTest()
80 GPUOperation operation = CreatePooling(op_def, attr); in AveragePoolingNonEmptyPaddingTest()
[all …]
Dsoftmax_test_util.cc38 OperationDef op_def; in SoftmaxTest() local
39 op_def.precision = precision; in SoftmaxTest()
41 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in SoftmaxTest()
42 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in SoftmaxTest()
44 GPUOperation operation = CreateSoftmax(op_def); in SoftmaxTest()
78 OperationDef op_def; in SoftmaxBigNumberTest() local
79 op_def.precision = precision; in SoftmaxBigNumberTest()
81 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in SoftmaxBigNumberTest()
82 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in SoftmaxBigNumberTest()
84 GPUOperation operation = CreateSoftmax(op_def); in SoftmaxBigNumberTest()
[all …]
Dconv_powervr_test_util.cc46 OperationDef op_def; in ConvPowerVR1x1SimpleWeightsTest() local
47 op_def.precision = precision; in ConvPowerVR1x1SimpleWeightsTest()
49 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConvPowerVR1x1SimpleWeightsTest()
50 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ConvPowerVR1x1SimpleWeightsTest()
53 CreateConvPowerVR(env->GetGpuInfo(), op_def, attr); in ConvPowerVR1x1SimpleWeightsTest()
83 OperationDef op_def; in ConvPowerVR1x1Test() local
84 op_def.precision = precision; in ConvPowerVR1x1Test()
86 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConvPowerVR1x1Test()
87 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ConvPowerVR1x1Test()
90 CreateConvPowerVR(env->GetGpuInfo(), op_def, attr); in ConvPowerVR1x1Test()
[all …]
/external/tensorflow/tensorflow/core/framework/
Dop_compatibility_test.cc46 const OpDef* op_def; in RegisteredOpDef() local
47 TF_CHECK_OK(OpRegistry::Global()->LookUpOpDef(node_def()->op(), &op_def)); in RegisteredOpDef()
48 return op_def; in RegisteredOpDef()
225 TF_ASSERT_OK(NodeDefBuilder("add_attr", &old_op.op_def).Finalize(node_def())); in TEST_F()
226 ExpectSuccess(old_op.op_def); in TEST_F()
240 TF_ASSERT_OK(NodeDefBuilder("less_strict", &old_op.op_def) in TEST_F()
243 ExpectSuccess(old_op.op_def); in TEST_F()
258 TF_ASSERT_OK(NodeDefBuilder("remove_restriction", &old_op.op_def) in TEST_F()
261 ExpectSuccess(old_op.op_def); in TEST_F()
277 TF_ASSERT_OK(NodeDefBuilder("attr_order", &old_op.op_def) in TEST_F()
[all …]
Dnode_def_util_test.cc37 return op_reg_data.op_def; in ToOpDef()
52 void ExpectSuccess(const NodeDef& good, const OpDef& op_def) { in ExpectSuccess() argument
53 EXPECT_EQ(Status::OK(), ValidateNodeDef(good, op_def)) in ExpectSuccess()
55 << "; OpDef: " << SummarizeOpDef(op_def); in ExpectSuccess()
58 void ExpectFailure(const NodeDef& bad, const OpDef& op_def, in ExpectFailure() argument
60 Status status = ValidateNodeDef(bad, op_def); in ExpectFailure()
63 << "; OpDef: " << SummarizeOpDef(op_def); in ExpectFailure()
68 << "; OpDef: " << SummarizeOpDef(op_def); in ExpectFailure()
73 << "; OpDef: " << SummarizeOpDef(op_def) << "\nActual error: " << status in ExpectFailure()
410 const OpDef op_def = ToOpDef(OpDefBuilder("Simple") in TEST() local
[all …]
Dop_def_util.cc151 const OpDef::AttrDef* FindAttr(StringPiece name, const OpDef& op_def) { in FindAttr() argument
152 for (int i = 0; i < op_def.attr_size(); ++i) { in FindAttr()
153 if (op_def.attr(i).name() == name) { in FindAttr()
154 return &op_def.attr(i); in FindAttr()
160 OpDef::AttrDef* FindAttrMutable(StringPiece name, OpDef* op_def) { in FindAttrMutable() argument
161 for (int i = 0; i < op_def->attr_size(); ++i) { in FindAttrMutable()
162 if (op_def->attr(i).name() == name) { in FindAttrMutable()
163 return op_def->mutable_attr(i); in FindAttrMutable()
169 const OpDef::ArgDef* FindInputArg(StringPiece name, const OpDef& op_def) { in FindInputArg() argument
170 for (int i = 0; i < op_def.input_arg_size(); ++i) { in FindInputArg()
[all …]
Dop_def_builder.cc110 strings::StrCat(__VA_ARGS__, AttrError(orig, op_def->name()))); \
148 void FinalizeAttr(StringPiece spec, bool allow_attr_type_any, OpDef* op_def, in FinalizeAttr() argument
150 OpDef::AttrDef* attr = op_def->add_attr(); in FinalizeAttr()
335 __VA_ARGS__, InOutError(is_output, orig, op_def->name()))); \
340 void FinalizeInputOrOutput(StringPiece spec, bool is_output, OpDef* op_def, in FinalizeInputOrOutput() argument
343 is_output ? op_def->add_output_arg() : op_def->add_input_arg(); in FinalizeInputOrOutput()
371 const OpDef::AttrDef* attr = FindAttr(type_or_attr, *op_def); in FinalizeInputOrOutput()
396 OpDef::AttrDef* attr = FindAttrMutable(arg->number_attr(), op_def); in FinalizeInputOrOutput()
404 OpDef::AttrDef* attr = FindAttrMutable(arg->type_list_attr(), op_def); in FinalizeInputOrOutput()
418 op_def->set_is_stateful(true); in FinalizeInputOrOutput()
[all …]

12345678910>>...12