Home
last modified time | relevance | path

Searched refs:op_idx (Results 1 – 21 of 21) sorted by relevance

/external/tensorflow/tensorflow/lite/tools/optimize/
Dquantize_model.cc46 int subgraph_index, int op_idx, const string& operator_name) { in GetOperatorProperty() argument
48 operator_property::GetOperatorProperty(model, subgraph_index, op_idx); in GetOperatorProperty()
50 model->subgraphs[subgraph_index]->operators[op_idx].get(); in GetOperatorProperty()
295 for (int op_idx = subgraph->operators.size() - 1; op_idx >= 0; op_idx--) { in ApplyConstraints() local
296 OperatorT* op = subgraph->operators[op_idx].get(); in ApplyConstraints()
298 GetOperatorProperty(operator_names, model, subgraph_idx, op_idx, in ApplyConstraints()
349 subgraph->operators.insert(subgraph->operators.begin() + op_idx, in ApplyConstraints()
401 ModelT* model, int32_t subgraph_idx, size_t* op_idx, in QuantizeOpInput() argument
408 OperatorT* op = subgraph->operators[*op_idx].get(); in QuantizeOpInput()
415 input_idx, op->inputs.size(), EnumNameBuiltinOperator(op_code), *op_idx, in QuantizeOpInput()
[all …]
Dquantization_wrapper_utils.cc45 for (size_t op_idx = 0; op_idx < subgraph->operators.size(); op_idx++) { in IntermediateTensorExists() local
46 OperatorT* op = subgraph->operators[op_idx].get(); in IntermediateTensorExists()
79 for (size_t op_idx = 0; op_idx < subgraph->operators.size(); op_idx++) { in AddIntemediateTensorsToFusedOp() local
81 OperatorT* op = subgraph->operators[op_idx].get(); in AddIntemediateTensorsToFusedOp()
83 operator_property::GetOperatorProperty(model, subgraph_idx, op_idx); in AddIntemediateTensorsToFusedOp()
92 auto name = CreateTensorName(op_idx, i); in AddIntemediateTensorsToFusedOp()
Dmodel_utils.cc127 for (int op_idx = subgraph->operators.size() - 1; op_idx >= 0; op_idx--) { in SetOperatorCodeVersion() local
128 OperatorT* op = subgraph->operators[op_idx].get(); in SetOperatorCodeVersion()
131 operator_property::GetOperatorProperty(model, subgraph_idx, op_idx); in SetOperatorCodeVersion()
Dquantize_weights.cc41 int32_t op_idx; member
57 for (size_t op_idx = 0; op_idx < subgraph->operators.size(); ++op_idx) { in GetTensorConsumers() local
58 OperatorT* op = subgraph->operators[op_idx].get(); in GetTensorConsumers()
65 {op, static_cast<int32_t>(op_idx), static_cast<int32_t>(i)}); in GetTensorConsumers()
444 min_op_idx = std::min(dequant_op_info.op_idx, min_op_idx); in QuantizeWeightsInt8()
532 min_op_idx = std::min(dequant_op_info.op_idx, min_op_idx); in QuantizeWeightsFloat16()
Dquantize_weights_test.cc118 for (size_t op_idx = 0; op_idx < subgraph->operators()->size(); ++op_idx) { in GetProducerOpCode() local
119 const auto op = subgraph->operators()->Get(op_idx); in GetProducerOpCode()
Dquantization_utils.h133 float GetEffectiveScale(ModelT* model, SubGraphT* subgraph, int op_idx,
Dquantization_utils.cc624 float GetEffectiveScale(ModelT* model, SubGraphT* subgraph, int op_idx, in GetEffectiveScale() argument
629 OperatorT* op = subgraph->operators[op_idx].get(); in GetEffectiveScale()
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/AMDGPU/
DR600AsmPrinter.cpp57 for (unsigned op_idx = 0; op_idx < numOperands; op_idx++) { in EmitProgramInfoR600() local
58 const MachineOperand &MO = MI.getOperand(op_idx); in EmitProgramInfoR600()
/external/tensorflow/tensorflow/lite/tools/
Dverifier.cc383 for (int op_idx = 0; op_idx < subgraph.operators()->Length(); ++op_idx) { in VerifySubGraphConsistency() local
384 const auto* op = subgraph.operators()->Get(op_idx); in VerifySubGraphConsistency()
403 input_idx, op_idx, in VerifySubGraphConsistency()
414 output_idx, op_idx, in VerifySubGraphConsistency()
421 output_idx, op_idx, in VerifySubGraphConsistency()
428 output_idx, op_idx, in VerifySubGraphConsistency()
435 output_idx, op_idx, in VerifySubGraphConsistency()
/external/angle/third_party/spirv-tools/src/source/opt/
Dfix_storage_class.cpp33 [&uses](Instruction* use, uint32_t op_idx) { in Process() argument
34 uses.push_back({use, op_idx}); in Process()
175 uint32_t op_idx, std::set<uint32_t>* seen) { in PropagateType() argument
187 if (op_idx == 2) { in PropagateType()
200 if (op_idx > 2) { in PropagateType()
Dfix_storage_class.h79 bool PropagateType(Instruction* inst, uint32_t type_id, uint32_t op_idx,
/external/deqp-deps/SPIRV-Tools/source/opt/
Dfix_storage_class.cpp33 [&uses](Instruction* use, uint32_t op_idx) { in Process() argument
34 uses.push_back({use, op_idx}); in Process()
175 uint32_t op_idx, std::set<uint32_t>* seen) { in PropagateType() argument
187 if (op_idx == 2) { in PropagateType()
200 if (op_idx > 2) { in PropagateType()
Dfix_storage_class.h79 bool PropagateType(Instruction* inst, uint32_t type_id, uint32_t op_idx,
/external/swiftshader/third_party/SPIRV-Tools/source/opt/
Dfix_storage_class.cpp33 [&uses](Instruction* use, uint32_t op_idx) { in Process() argument
34 uses.push_back({use, op_idx}); in Process()
175 uint32_t op_idx, std::set<uint32_t>* seen) { in PropagateType() argument
187 if (op_idx == 2) { in PropagateType()
200 if (op_idx > 2) { in PropagateType()
Dfix_storage_class.h79 bool PropagateType(Instruction* inst, uint32_t type_id, uint32_t op_idx,
/external/tensorflow/tensorflow/compiler/xla/service/cpu/
Dcpu_layout_assignment.cc123 } else if (optional<int64> op_idx = in AddBackendConstraints() local
125 const HloInstruction* op = instruction->operand(*op_idx); in AddBackendConstraints()
127 ColMajorShape(op->shape()), instruction, *op_idx)); in AddBackendConstraints()
/external/llvm/lib/Target/AMDGPU/
DAMDGPUAsmPrinter.cpp262 for (unsigned op_idx = 0; op_idx < numOperands; op_idx++) { in EmitProgramInfoR600() local
263 const MachineOperand &MO = MI.getOperand(op_idx); in EmitProgramInfoR600()
332 for (unsigned op_idx = 0; op_idx < numOperands; op_idx++) { in getSIProgramInfo() local
333 const MachineOperand &MO = MI.getOperand(op_idx); in getSIProgramInfo()
/external/tensorflow/tensorflow/core/grappler/optimizers/
Dscoped_allocator_optimizer.cc701 for (int op_idx = 0; op_idx < ops.size(); ++op_idx) { in RewireSubgraph() local
702 NodeDef* old_op = ops[op_idx]; in RewireSubgraph()
742 *n->mutable_input(i) = strings::StrCat(sas_name, ":", op_idx); in RewireSubgraph()
/external/tensorflow/tensorflow/stream_executor/rocm/
Drocm_dnn.cc1000 miopenStatus_t SetConvolutionArgs(const int op_idx, const float* alpha, in SetConvolutionArgs() argument
1003 auto status = wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &conv_op); in SetConvolutionArgs()
1018 miopenStatus_t SetBiasArgs(const int op_idx, const float* alpha, in SetBiasArgs() argument
1021 auto status = wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &bias_op); in SetBiasArgs()
1036 miopenStatus_t SetBatchNormInferenceArgs(const int op_idx, const float* alpha, in SetBatchNormInferenceArgs() argument
1043 wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &batchnorm_op); in SetBatchNormInferenceArgs()
1060 const int op_idx, const float* alpha, const float* beta, in SetBatchNormForwardArgs() argument
1066 wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &batchnorm_op); in SetBatchNormForwardArgs()
1083 miopenStatus_t SetBatchNormBackwardArgs(const int op_idx, const float* alpha, in SetBatchNormBackwardArgs() argument
1091 wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &batchnorm_op); in SetBatchNormBackwardArgs()
[all …]
/external/tensorflow/tensorflow/compiler/xla/service/
Dtuple_points_to_analysis.cc694 for (int64 op_idx : alias_user->OperandIndices(alias.instruction())) { in GetAllUsesOfInstructionAtIndex() local
695 uses.emplace_back(alias_user, op_idx); in GetAllUsesOfInstructionAtIndex()
/external/tensorflow/tensorflow/python/tpu/
Dtensor_tracer.py383 def reason(op_idx, details): argument
386 return '%d %s'%(op_idx, details)