/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/ |
D | simple_selectors.cc | 49 std::unique_ptr<GPUOperation> SelectLSTM(const OperationDef& op_def, in SelectLSTM() argument 51 return absl::make_unique<GPUOperation>(CreateLSTM(op_def, gpu_info)); in SelectLSTM() 55 const OperationDef& op_def) { in SelectReLU() argument 56 return absl::make_unique<GPUOperation>(CreateReLU(op_def, attr)); in SelectReLU() 61 const OperationDef& op_def) { in SelectPReLU() argument 62 return absl::make_unique<GPUOperation>(CreatePReLU(gpu_info, op_def, attr)); in SelectPReLU() 66 const OperationDef& op_def) { in SelectPooling() argument 67 return absl::make_unique<GPUOperation>(CreatePooling(op_def, attr)); in SelectPooling() 71 const MaxUnpooling2DAttributes& attr, const OperationDef& op_def) { in SelectMaxUnpooling() argument 72 return absl::make_unique<GPUOperation>(CreateMaxUnpooling(op_def, attr)); in SelectMaxUnpooling() [all …]
|
D | simple_selectors.h | 30 std::unique_ptr<GPUOperation> SelectLSTM(const OperationDef& op_def, 34 const OperationDef& op_def); 38 const OperationDef& op_def); 41 const OperationDef& op_def); 44 const MaxUnpooling2DAttributes& attr, const OperationDef& op_def); 46 void SelectAdd(const OperationDef& op_def, const std::vector<int>& channels, 50 const OperationDef& op_def, 55 const OperationDef& op_def, const GpuInfo& gpu_info, 60 const OperationDef& op_def); 63 const OperationDef& op_def, [all …]
|
D | operation_selector.cc | 82 const OperationDef& op_def, ModelHints hints, in WinogradFromNode() argument 99 gpu_info, shape_0, op_def.src_tensors[0].storage_type, in WinogradFromNode() 100 op_def.src_tensors[0].data_type, op_def.src_tensors[0].layout); in WinogradFromNode() 101 td_0.data_type = op_def.src_tensors[0].data_type; in WinogradFromNode() 102 td_0.layout = op_def.src_tensors[0].layout; in WinogradFromNode() 105 gpu_info, shape_1, op_def.src_tensors[0].storage_type, in WinogradFromNode() 106 op_def.src_tensors[0].data_type, op_def.src_tensors[0].layout); in WinogradFromNode() 107 td_1.data_type = op_def.src_tensors[0].data_type; in WinogradFromNode() 108 td_1.layout = op_def.src_tensors[0].layout; in WinogradFromNode() 114 winograd_up_def.precision = op_def.precision; in WinogradFromNode() [all …]
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/default/ |
D | convolution_selector.cc | 38 const GpuInfo& gpu_info, const OperationDef& op_def, in SelectConvolutionAdreno() argument 40 if (IsConvConstantsSupported(gpu_info, op_def, attr)) { in SelectConvolutionAdreno() 41 GPUOperation conv = CreateConvConstants(gpu_info, op_def, attr); in SelectConvolutionAdreno() 44 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionAdreno() 51 const GpuInfo& gpu_info, const OperationDef& op_def, in SelectConvolutionWinogradAdreno() argument 54 CreateConvPowerVRWino4x4To6x6(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionWinogradAdreno() 61 const OperationDef& op_def, ModelHints hints, in SelectConvolutionDynamicWeightsAdreno() argument 64 gpu_info, op_def, attr, weights_shape, &dst_shape); in SelectConvolutionDynamicWeightsAdreno() 71 const GpuInfo& gpu_info, const OperationDef& op_def) { in SelectConvolutionNVidia() argument 72 if (IsConvConstantsSupported(gpu_info, op_def, attr)) { in SelectConvolutionNVidia() [all …]
|
D | convolution_transposed_selector.cc | 30 const OperationDef& op_def) { in SelectConvolutionTransposedAdreno() argument 33 CreateConvolutionTransposedThin(gpu_info, op_def, attr); in SelectConvolutionTransposedAdreno() 37 CreateConvolutionTransposed3x3Thin(gpu_info, op_def, attr); in SelectConvolutionTransposedAdreno() 41 CreateConvolutionTransposed(gpu_info, op_def, attr); in SelectConvolutionTransposedAdreno() 48 const OperationDef& op_def) { in SelectConvolutionTransposedPowerVR() argument 51 CreateConvolutionTransposedThin(gpu_info, op_def, attr); in SelectConvolutionTransposedPowerVR() 55 CreateConvolutionTransposed3x3Thin(gpu_info, op_def, attr); in SelectConvolutionTransposedPowerVR() 57 } else if (IsConvolutionTransposed3x3Supported(op_def, attr)) { in SelectConvolutionTransposedPowerVR() 59 CreateConvolutionTransposed3x3(gpu_info, op_def, attr); in SelectConvolutionTransposedPowerVR() 61 } else if (IsConvolutionTransposed4x4Supported(op_def, attr)) { in SelectConvolutionTransposedPowerVR() [all …]
|
D | fully_connected_selector.cc | 29 const OperationDef& op_def, int batch_size) { in SelectFullyConnectedGeneric() argument 30 if (op_def.IsBatchSupported()) { in SelectFullyConnectedGeneric() 32 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedGeneric() 35 FullyConnected fc = CreateFullyConnected(gpu_info, op_def, attr); in SelectFullyConnectedGeneric() 42 const OperationDef& op_def, int batch_size) { in SelectFullyConnectedAdreno() argument 43 if (op_def.IsBatchSupported()) { in SelectFullyConnectedAdreno() 45 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedAdreno() 48 FullyConnected fc = CreateFullyConnected(gpu_info, op_def, attr); in SelectFullyConnectedAdreno() 55 const OperationDef& op_def, int batch_size) { in SelectFullyConnectedPowerVR() argument 56 if (op_def.IsBatchSupported()) { in SelectFullyConnectedPowerVR() [all …]
|
D | dw_convolution_selector.cc | 28 const OperationDef& op_def) { in SelectDWConvolutionAdreno() argument 31 CreateDepthwiseConv3x3(gpu_info, op_def, attr)); in SelectDWConvolutionAdreno() 34 CreateDepthwiseConvolution2D(gpu_info, op_def, attr)); in SelectDWConvolutionAdreno() 40 const OperationDef& op_def) { in SelectDWConvolutionPowerVR() argument 43 CreateDepthwiseConv3x3(gpu_info, op_def, attr)); in SelectDWConvolutionPowerVR() 46 CreateDepthwiseConvolution2D(gpu_info, op_def, attr)); in SelectDWConvolutionPowerVR() 52 const OperationDef& op_def) { in SelectDWConvolutionMali() argument 53 const auto storage_type = op_def.src_tensors[0].storage_type; in SelectDWConvolutionMali() 58 !buffer_type && op_def.precision != CalculationsPrecision::F32) { in SelectDWConvolutionMali() 60 CreateDepthwiseConv3x3(gpu_info, op_def, attr)); in SelectDWConvolutionMali() [all …]
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/ |
D | elementwise_test_util.cc | 35 OperationDef op_def; in AbsTest() local 36 op_def.precision = precision; in AbsTest() 38 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AbsTest() 39 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in AbsTest() 42 env->GetGpuInfo(), op_def, OperationType::ABS); in AbsTest() 62 OperationDef op_def; in CosTest() local 63 op_def.precision = precision; in CosTest() 65 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in CosTest() 66 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in CosTest() 69 env->GetGpuInfo(), op_def, OperationType::COS); in CosTest() [all …]
|
D | padding_test_util.cc | 40 OperationDef op_def; in PaddingAppendWidthTest() local 41 op_def.precision = precision; in PaddingAppendWidthTest() 43 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingAppendWidthTest() 44 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingAppendWidthTest() 46 GPUOperation operation = CreatePadding(op_def, attr); in PaddingAppendWidthTest() 70 OperationDef op_def; in PaddingPrependWidthTest() local 71 op_def.precision = precision; in PaddingPrependWidthTest() 73 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingPrependWidthTest() 74 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingPrependWidthTest() 76 GPUOperation operation = CreatePadding(op_def, attr); in PaddingPrependWidthTest() [all …]
|
D | resize_test_util.cc | 41 OperationDef op_def; in ResizeBilinearAlignedTest() local 42 op_def.precision = precision; in ResizeBilinearAlignedTest() 44 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearAlignedTest() 45 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearAlignedTest() 47 Resize operation = CreateResize(op_def, attr); in ResizeBilinearAlignedTest() 73 OperationDef op_def; in ResizeBilinearNonAlignedTest() local 74 op_def.precision = precision; in ResizeBilinearNonAlignedTest() 76 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearNonAlignedTest() 77 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearNonAlignedTest() 79 Resize operation = CreateResize(op_def, attr); in ResizeBilinearNonAlignedTest() [all …]
|
D | relu_test_util.cc | 40 OperationDef op_def; in ReLUNoClipNoAlphaTest() local 41 op_def.precision = precision; in ReLUNoClipNoAlphaTest() 43 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReLUNoClipNoAlphaTest() 44 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ReLUNoClipNoAlphaTest() 46 GPUOperation operation = CreateReLU(op_def, attr); in ReLUNoClipNoAlphaTest() 69 OperationDef op_def; in ReLUClipTest() local 70 op_def.precision = precision; in ReLUClipTest() 72 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReLUClipTest() 73 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ReLUClipTest() 75 GPUOperation operation = CreateReLU(op_def, attr); in ReLUClipTest() [all …]
|
D | reduce_test_util.cc | 37 OperationDef op_def; in MeanHWTest() local 38 op_def.precision = precision; in MeanHWTest() 40 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in MeanHWTest() 41 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in MeanHWTest() 44 CreateReduce(axis, src_tensor.shape, OperationType::MEAN, op_def, in MeanHWTest() 64 OperationDef op_def; in ReduceSumChannelsTest() local 65 op_def.precision = precision; in ReduceSumChannelsTest() 67 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReduceSumChannelsTest() 68 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ReduceSumChannelsTest() 72 op_def, env->GetGpuInfo()); in ReduceSumChannelsTest() [all …]
|
D | concat_test_util.cc | 42 OperationDef op_def; in ConcatWidthTest() local 43 op_def.precision = precision; in ConcatWidthTest() 45 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatWidthTest() 46 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatWidthTest() 47 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatWidthTest() 49 GPUOperation operation = CreateConcatXY(op_def, attr); in ConcatWidthTest() 75 OperationDef op_def; in ConcatHeightTest() local 76 op_def.precision = precision; in ConcatHeightTest() 78 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatHeightTest() 79 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatHeightTest() [all …]
|
D | quantize_and_dequantize_test_util.cc | 45 OperationDef op_def; in QuantAndDequant_Dim2Bits8Test() local 46 op_def.precision = precision; in QuantAndDequant_Dim2Bits8Test() 48 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in QuantAndDequant_Dim2Bits8Test() 49 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in QuantAndDequant_Dim2Bits8Test() 51 GPUOperation operation = CreateQuantizeAndDequantize(op_def, attr); in QuantAndDequant_Dim2Bits8Test() 82 OperationDef op_def; in QuantAndDequant_Dim3Bits8_NegativeRangeTest() local 83 op_def.precision = precision; in QuantAndDequant_Dim3Bits8_NegativeRangeTest() 85 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in QuantAndDequant_Dim3Bits8_NegativeRangeTest() 86 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in QuantAndDequant_Dim3Bits8_NegativeRangeTest() 88 GPUOperation operation = CreateQuantizeAndDequantize(op_def, attr); in QuantAndDequant_Dim3Bits8_NegativeRangeTest() [all …]
|
D | conv_weights_converter_test_util.cc | 32 const OperationDef& op_def) { in ConvolutionWeightsConverterTest() argument 93 auto converter = ConverterToConvWeights(op_def, weight_desc); in ConvolutionWeightsConverterTest() 126 OperationDef op_def; in ConverterToConvWeights1x1OutX4Test() local 127 op_def.precision = precision; in ConverterToConvWeights1x1OutX4Test() 129 op_def.src_tensors.push_back({data_type, storage, Layout::BHWC}); in ConverterToConvWeights1x1OutX4Test() 130 op_def.dst_tensors.push_back( in ConverterToConvWeights1x1OutX4Test() 133 weights, conv_weight_desc, env, op_def)); in ConverterToConvWeights1x1OutX4Test() 160 OperationDef op_def; in ConverterToConvWeights1x1OutX4UnalignedTest() local 161 op_def.precision = precision; in ConverterToConvWeights1x1OutX4UnalignedTest() 163 op_def.src_tensors.push_back({data_type, storage, Layout::BHWC}); in ConverterToConvWeights1x1OutX4UnalignedTest() [all …]
|
D | add_test_util.cc | 37 OperationDef op_def; in AddTwoEqualTensorsTest() local 38 op_def.precision = precision; in AddTwoEqualTensorsTest() 40 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddTwoEqualTensorsTest() 41 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddTwoEqualTensorsTest() 42 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in AddTwoEqualTensorsTest() 44 GPUOperation operation = CreateAdd(op_def, channels, channels[0]); in AddTwoEqualTensorsTest() 67 OperationDef op_def; in AddFirstTensorHasMoreChannelsThanSecondTest() local 68 op_def.precision = precision; in AddFirstTensorHasMoreChannelsThanSecondTest() 70 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddFirstTensorHasMoreChannelsThanSecondTest() 71 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddFirstTensorHasMoreChannelsThanSecondTest() [all …]
|
D | pooling.cc | 27 std::string GetAveragePoolingKernelCode(const OperationDef& op_def, in GetAveragePoolingKernelCode() argument 30 auto src_desc = op_def.src_tensors[0]; in GetAveragePoolingKernelCode() 32 if (op_def.IsBatchSupported()) { in GetAveragePoolingKernelCode() 36 auto dst_desc = op_def.dst_tensors[0]; in GetAveragePoolingKernelCode() 37 if (op_def.IsBatchSupported()) { in GetAveragePoolingKernelCode() 55 if (op_def.dst_tensors[0].HasAxis(axis)) { in GetAveragePoolingKernelCode() 58 if (op_def.src_tensors[0].HasAxis(axis)) { in GetAveragePoolingKernelCode() 72 op_def.src_tensors[0].storage_type == TensorStorageType::BUFFER || in GetAveragePoolingKernelCode() 73 op_def.src_tensors[0].storage_type == TensorStorageType::IMAGE_BUFFER; in GetAveragePoolingKernelCode() 78 if (op_def.dst_tensors[0].HasAxis(Axis::DEPTH)) { in GetAveragePoolingKernelCode() [all …]
|
D | space_to_depth_test_util.cc | 36 OperationDef op_def; in SpaceToDepthTensorShape1x2x2x1BlockSize2Test() local 37 op_def.precision = precision; in SpaceToDepthTensorShape1x2x2x1BlockSize2Test() 39 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in SpaceToDepthTensorShape1x2x2x1BlockSize2Test() 40 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in SpaceToDepthTensorShape1x2x2x1BlockSize2Test() 42 GPUOperation operation = CreateSpaceToDepth(op_def, attr); in SpaceToDepthTensorShape1x2x2x1BlockSize2Test() 63 OperationDef op_def; in SpaceToDepthTensorShape1x2x2x2BlockSize2Test() local 64 op_def.precision = precision; in SpaceToDepthTensorShape1x2x2x2BlockSize2Test() 66 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in SpaceToDepthTensorShape1x2x2x2BlockSize2Test() 67 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in SpaceToDepthTensorShape1x2x2x2BlockSize2Test() 69 GPUOperation operation = CreateSpaceToDepth(op_def, attr); in SpaceToDepthTensorShape1x2x2x2BlockSize2Test() [all …]
|
D | pooling_test_util.cc | 43 OperationDef op_def; in AveragePoolingTest() local 44 op_def.precision = precision; in AveragePoolingTest() 46 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AveragePoolingTest() 47 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in AveragePoolingTest() 49 GPUOperation operation = CreatePooling(op_def, attr); in AveragePoolingTest() 74 OperationDef op_def; in AveragePoolingNonEmptyPaddingTest() local 75 op_def.precision = precision; in AveragePoolingNonEmptyPaddingTest() 77 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AveragePoolingNonEmptyPaddingTest() 78 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in AveragePoolingNonEmptyPaddingTest() 80 GPUOperation operation = CreatePooling(op_def, attr); in AveragePoolingNonEmptyPaddingTest() [all …]
|
D | softmax_test_util.cc | 38 OperationDef op_def; in SoftmaxTest() local 39 op_def.precision = precision; in SoftmaxTest() 41 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in SoftmaxTest() 42 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in SoftmaxTest() 44 GPUOperation operation = CreateSoftmax(op_def); in SoftmaxTest() 78 OperationDef op_def; in SoftmaxBigNumberTest() local 79 op_def.precision = precision; in SoftmaxBigNumberTest() 81 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in SoftmaxBigNumberTest() 82 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in SoftmaxBigNumberTest() 84 GPUOperation operation = CreateSoftmax(op_def); in SoftmaxBigNumberTest() [all …]
|
D | conv_powervr_test_util.cc | 46 OperationDef op_def; in ConvPowerVR1x1SimpleWeightsTest() local 47 op_def.precision = precision; in ConvPowerVR1x1SimpleWeightsTest() 49 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConvPowerVR1x1SimpleWeightsTest() 50 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ConvPowerVR1x1SimpleWeightsTest() 53 CreateConvPowerVR(env->GetGpuInfo(), op_def, attr); in ConvPowerVR1x1SimpleWeightsTest() 83 OperationDef op_def; in ConvPowerVR1x1Test() local 84 op_def.precision = precision; in ConvPowerVR1x1Test() 86 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConvPowerVR1x1Test() 87 op_def.dst_tensors.push_back({data_type, storage, Layout::HWC}); in ConvPowerVR1x1Test() 90 CreateConvPowerVR(env->GetGpuInfo(), op_def, attr); in ConvPowerVR1x1Test() [all …]
|
/external/tensorflow/tensorflow/core/framework/ |
D | op_compatibility_test.cc | 46 const OpDef* op_def; in RegisteredOpDef() local 47 TF_CHECK_OK(OpRegistry::Global()->LookUpOpDef(node_def()->op(), &op_def)); in RegisteredOpDef() 48 return op_def; in RegisteredOpDef() 225 TF_ASSERT_OK(NodeDefBuilder("add_attr", &old_op.op_def).Finalize(node_def())); in TEST_F() 226 ExpectSuccess(old_op.op_def); in TEST_F() 240 TF_ASSERT_OK(NodeDefBuilder("less_strict", &old_op.op_def) in TEST_F() 243 ExpectSuccess(old_op.op_def); in TEST_F() 258 TF_ASSERT_OK(NodeDefBuilder("remove_restriction", &old_op.op_def) in TEST_F() 261 ExpectSuccess(old_op.op_def); in TEST_F() 277 TF_ASSERT_OK(NodeDefBuilder("attr_order", &old_op.op_def) in TEST_F() [all …]
|
D | node_def_util_test.cc | 37 return op_reg_data.op_def; in ToOpDef() 52 void ExpectSuccess(const NodeDef& good, const OpDef& op_def) { in ExpectSuccess() argument 53 EXPECT_EQ(Status::OK(), ValidateNodeDef(good, op_def)) in ExpectSuccess() 55 << "; OpDef: " << SummarizeOpDef(op_def); in ExpectSuccess() 58 void ExpectFailure(const NodeDef& bad, const OpDef& op_def, in ExpectFailure() argument 60 Status status = ValidateNodeDef(bad, op_def); in ExpectFailure() 63 << "; OpDef: " << SummarizeOpDef(op_def); in ExpectFailure() 68 << "; OpDef: " << SummarizeOpDef(op_def); in ExpectFailure() 73 << "; OpDef: " << SummarizeOpDef(op_def) << "\nActual error: " << status in ExpectFailure() 410 const OpDef op_def = ToOpDef(OpDefBuilder("Simple") in TEST() local [all …]
|
D | op_def_util.cc | 151 const OpDef::AttrDef* FindAttr(StringPiece name, const OpDef& op_def) { in FindAttr() argument 152 for (int i = 0; i < op_def.attr_size(); ++i) { in FindAttr() 153 if (op_def.attr(i).name() == name) { in FindAttr() 154 return &op_def.attr(i); in FindAttr() 160 OpDef::AttrDef* FindAttrMutable(StringPiece name, OpDef* op_def) { in FindAttrMutable() argument 161 for (int i = 0; i < op_def->attr_size(); ++i) { in FindAttrMutable() 162 if (op_def->attr(i).name() == name) { in FindAttrMutable() 163 return op_def->mutable_attr(i); in FindAttrMutable() 169 const OpDef::ArgDef* FindInputArg(StringPiece name, const OpDef& op_def) { in FindInputArg() argument 170 for (int i = 0; i < op_def.input_arg_size(); ++i) { in FindInputArg() [all …]
|
D | op_def_builder.cc | 110 strings::StrCat(__VA_ARGS__, AttrError(orig, op_def->name()))); \ 148 void FinalizeAttr(StringPiece spec, bool allow_attr_type_any, OpDef* op_def, in FinalizeAttr() argument 150 OpDef::AttrDef* attr = op_def->add_attr(); in FinalizeAttr() 335 __VA_ARGS__, InOutError(is_output, orig, op_def->name()))); \ 340 void FinalizeInputOrOutput(StringPiece spec, bool is_output, OpDef* op_def, in FinalizeInputOrOutput() argument 343 is_output ? op_def->add_output_arg() : op_def->add_input_arg(); in FinalizeInputOrOutput() 371 const OpDef::AttrDef* attr = FindAttr(type_or_attr, *op_def); in FinalizeInputOrOutput() 396 OpDef::AttrDef* attr = FindAttrMutable(arg->number_attr(), op_def); in FinalizeInputOrOutput() 404 OpDef::AttrDef* attr = FindAttrMutable(arg->type_list_attr(), op_def); in FinalizeInputOrOutput() 418 op_def->set_is_stateful(true); in FinalizeInputOrOutput() [all …]
|