/external/tensorflow/tensorflow/lite/tools/optimize/ |
D | quantize_model.cc | 46 int subgraph_index, int op_idx, const string& operator_name) { in GetOperatorProperty() argument 48 operator_property::GetOperatorProperty(model, subgraph_index, op_idx); in GetOperatorProperty() 50 model->subgraphs[subgraph_index]->operators[op_idx].get(); in GetOperatorProperty() 295 for (int op_idx = subgraph->operators.size() - 1; op_idx >= 0; op_idx--) { in ApplyConstraints() local 296 OperatorT* op = subgraph->operators[op_idx].get(); in ApplyConstraints() 298 GetOperatorProperty(operator_names, model, subgraph_idx, op_idx, in ApplyConstraints() 349 subgraph->operators.insert(subgraph->operators.begin() + op_idx, in ApplyConstraints() 401 ModelT* model, int32_t subgraph_idx, size_t* op_idx, in QuantizeOpInput() argument 408 OperatorT* op = subgraph->operators[*op_idx].get(); in QuantizeOpInput() 415 input_idx, op->inputs.size(), EnumNameBuiltinOperator(op_code), *op_idx, in QuantizeOpInput() [all …]
|
D | quantization_wrapper_utils.cc | 45 for (size_t op_idx = 0; op_idx < subgraph->operators.size(); op_idx++) { in IntermediateTensorExists() local 46 OperatorT* op = subgraph->operators[op_idx].get(); in IntermediateTensorExists() 79 for (size_t op_idx = 0; op_idx < subgraph->operators.size(); op_idx++) { in AddIntemediateTensorsToFusedOp() local 81 OperatorT* op = subgraph->operators[op_idx].get(); in AddIntemediateTensorsToFusedOp() 83 operator_property::GetOperatorProperty(model, subgraph_idx, op_idx); in AddIntemediateTensorsToFusedOp() 92 auto name = CreateTensorName(op_idx, i); in AddIntemediateTensorsToFusedOp()
|
D | model_utils.cc | 127 for (int op_idx = subgraph->operators.size() - 1; op_idx >= 0; op_idx--) { in SetOperatorCodeVersion() local 128 OperatorT* op = subgraph->operators[op_idx].get(); in SetOperatorCodeVersion() 131 operator_property::GetOperatorProperty(model, subgraph_idx, op_idx); in SetOperatorCodeVersion()
|
D | quantize_weights.cc | 41 int32_t op_idx; member 57 for (size_t op_idx = 0; op_idx < subgraph->operators.size(); ++op_idx) { in GetTensorConsumers() local 58 OperatorT* op = subgraph->operators[op_idx].get(); in GetTensorConsumers() 65 {op, static_cast<int32_t>(op_idx), static_cast<int32_t>(i)}); in GetTensorConsumers() 444 min_op_idx = std::min(dequant_op_info.op_idx, min_op_idx); in QuantizeWeightsInt8() 532 min_op_idx = std::min(dequant_op_info.op_idx, min_op_idx); in QuantizeWeightsFloat16()
|
D | quantize_weights_test.cc | 118 for (size_t op_idx = 0; op_idx < subgraph->operators()->size(); ++op_idx) { in GetProducerOpCode() local 119 const auto op = subgraph->operators()->Get(op_idx); in GetProducerOpCode()
|
D | quantization_utils.h | 133 float GetEffectiveScale(ModelT* model, SubGraphT* subgraph, int op_idx,
|
D | quantization_utils.cc | 624 float GetEffectiveScale(ModelT* model, SubGraphT* subgraph, int op_idx, in GetEffectiveScale() argument 629 OperatorT* op = subgraph->operators[op_idx].get(); in GetEffectiveScale()
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/AMDGPU/ |
D | R600AsmPrinter.cpp | 57 for (unsigned op_idx = 0; op_idx < numOperands; op_idx++) { in EmitProgramInfoR600() local 58 const MachineOperand &MO = MI.getOperand(op_idx); in EmitProgramInfoR600()
|
/external/tensorflow/tensorflow/lite/tools/ |
D | verifier.cc | 383 for (int op_idx = 0; op_idx < subgraph.operators()->Length(); ++op_idx) { in VerifySubGraphConsistency() local 384 const auto* op = subgraph.operators()->Get(op_idx); in VerifySubGraphConsistency() 403 input_idx, op_idx, in VerifySubGraphConsistency() 414 output_idx, op_idx, in VerifySubGraphConsistency() 421 output_idx, op_idx, in VerifySubGraphConsistency() 428 output_idx, op_idx, in VerifySubGraphConsistency() 435 output_idx, op_idx, in VerifySubGraphConsistency()
|
/external/angle/third_party/spirv-tools/src/source/opt/ |
D | fix_storage_class.cpp | 33 [&uses](Instruction* use, uint32_t op_idx) { in Process() argument 34 uses.push_back({use, op_idx}); in Process() 175 uint32_t op_idx, std::set<uint32_t>* seen) { in PropagateType() argument 187 if (op_idx == 2) { in PropagateType() 200 if (op_idx > 2) { in PropagateType()
|
D | fix_storage_class.h | 79 bool PropagateType(Instruction* inst, uint32_t type_id, uint32_t op_idx,
|
/external/deqp-deps/SPIRV-Tools/source/opt/ |
D | fix_storage_class.cpp | 33 [&uses](Instruction* use, uint32_t op_idx) { in Process() argument 34 uses.push_back({use, op_idx}); in Process() 175 uint32_t op_idx, std::set<uint32_t>* seen) { in PropagateType() argument 187 if (op_idx == 2) { in PropagateType() 200 if (op_idx > 2) { in PropagateType()
|
D | fix_storage_class.h | 79 bool PropagateType(Instruction* inst, uint32_t type_id, uint32_t op_idx,
|
/external/swiftshader/third_party/SPIRV-Tools/source/opt/ |
D | fix_storage_class.cpp | 33 [&uses](Instruction* use, uint32_t op_idx) { in Process() argument 34 uses.push_back({use, op_idx}); in Process() 175 uint32_t op_idx, std::set<uint32_t>* seen) { in PropagateType() argument 187 if (op_idx == 2) { in PropagateType() 200 if (op_idx > 2) { in PropagateType()
|
D | fix_storage_class.h | 79 bool PropagateType(Instruction* inst, uint32_t type_id, uint32_t op_idx,
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | cpu_layout_assignment.cc | 123 } else if (optional<int64> op_idx = in AddBackendConstraints() local 125 const HloInstruction* op = instruction->operand(*op_idx); in AddBackendConstraints() 127 ColMajorShape(op->shape()), instruction, *op_idx)); in AddBackendConstraints()
|
/external/llvm/lib/Target/AMDGPU/ |
D | AMDGPUAsmPrinter.cpp | 262 for (unsigned op_idx = 0; op_idx < numOperands; op_idx++) { in EmitProgramInfoR600() local 263 const MachineOperand &MO = MI.getOperand(op_idx); in EmitProgramInfoR600() 332 for (unsigned op_idx = 0; op_idx < numOperands; op_idx++) { in getSIProgramInfo() local 333 const MachineOperand &MO = MI.getOperand(op_idx); in getSIProgramInfo()
|
/external/tensorflow/tensorflow/core/grappler/optimizers/ |
D | scoped_allocator_optimizer.cc | 701 for (int op_idx = 0; op_idx < ops.size(); ++op_idx) { in RewireSubgraph() local 702 NodeDef* old_op = ops[op_idx]; in RewireSubgraph() 742 *n->mutable_input(i) = strings::StrCat(sas_name, ":", op_idx); in RewireSubgraph()
|
/external/tensorflow/tensorflow/stream_executor/rocm/ |
D | rocm_dnn.cc | 1000 miopenStatus_t SetConvolutionArgs(const int op_idx, const float* alpha, in SetConvolutionArgs() argument 1003 auto status = wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &conv_op); in SetConvolutionArgs() 1018 miopenStatus_t SetBiasArgs(const int op_idx, const float* alpha, in SetBiasArgs() argument 1021 auto status = wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &bias_op); in SetBiasArgs() 1036 miopenStatus_t SetBatchNormInferenceArgs(const int op_idx, const float* alpha, in SetBatchNormInferenceArgs() argument 1043 wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &batchnorm_op); in SetBatchNormInferenceArgs() 1060 const int op_idx, const float* alpha, const float* beta, in SetBatchNormForwardArgs() argument 1066 wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &batchnorm_op); in SetBatchNormForwardArgs() 1083 miopenStatus_t SetBatchNormBackwardArgs(const int op_idx, const float* alpha, in SetBatchNormBackwardArgs() argument 1091 wrap::miopenFusionPlanGetOp(fusion_plan_, op_idx, &batchnorm_op); in SetBatchNormBackwardArgs() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | tuple_points_to_analysis.cc | 694 for (int64 op_idx : alias_user->OperandIndices(alias.instruction())) { in GetAllUsesOfInstructionAtIndex() local 695 uses.emplace_back(alias_user, op_idx); in GetAllUsesOfInstructionAtIndex()
|
/external/tensorflow/tensorflow/python/tpu/ |
D | tensor_tracer.py | 383 def reason(op_idx, details): argument 386 return '%d %s'%(op_idx, details)
|