/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | xla_compilation_device.cc | 91 OpKernel* op_kernel, XlaContext& context) { in AttachLocationToMetadata() argument 93 context.StackTraceForNodeName(op_kernel->def().name())) { in AttachLocationToMetadata() 101 void XlaCompilationDevice::Compute(OpKernel* op_kernel, in Compute() argument 104 << FormatNodeDefForError(op_kernel->def()); in Compute() 108 metadata.set_op_type(op_kernel->type_string()); in Compute() 109 metadata.set_op_name(op_kernel->name()); in Compute() 110 AttachLocationToMetadata(metadata, op_kernel, xla_context); in Compute() 114 ParseShardingFromDevice(op_kernel->def(), std::numeric_limits<int>::max(), in Compute() 121 GetFrontendAttributesFromAttrSlice(AttrSlice(op_kernel->def())); in Compute() 138 op_kernel->Compute(context); in Compute()
|
D | const_analysis.cc | 88 Status GetCompileTimeConstInputs(const NodeDef& node, const OpKernel* op_kernel, in GetCompileTimeConstInputs() argument 92 DCHECK(op_def != nullptr || op_kernel != nullptr); in GetCompileTimeConstInputs() 167 return XlaOpRegistry::CompileTimeConstantInputs(*op_kernel, in GetCompileTimeConstInputs() 287 Status GetCompileTimeConstInputs(const OpKernel* op_kernel, in GetCompileTimeConstInputs() argument 290 return GetCompileTimeConstInputs(op_kernel->def(), op_kernel, in GetCompileTimeConstInputs()
|
D | xla_op_registry.h | 199 static Status CompileTimeConstantInputs(const OpKernel& op_kernel, in CompileTimeConstantInputs() argument 201 return CompileTimeConstantInputs(op_kernel.def(), /*op_kernel=*/&op_kernel, in CompileTimeConstantInputs() 296 const OpKernel* op_kernel,
|
D | graph_compiler.cc | 141 std::unique_ptr<OpKernel> op_kernel(op_kernel_raw); in Compile() local 151 params.op_kernel = op_kernel.get(); in Compile() 172 VLOG(3) << "Translating " << params.op_kernel->name(); in Compile() 176 device_->Compute(CHECK_NOTNULL(params.op_kernel), &op_context); in Compile()
|
D | xla_op_kernel.cc | 139 TF_RETURN_IF_ERROR(context->op_kernel().InputRange(name, &start, &stop)); in InputIndex() 165 context_->op_kernel().type_string(), in ConstantInputReshaped() 172 "Input ", index, " to node `", context_->op_kernel().name(), in ConstantInputReshaped() 173 "` with op ", context_->op_kernel().type_string(), in ConstantInputReshaped() 186 context_->op_kernel().name(), " input ", index, " has shape ", in ConstantInputReshaped() 282 context_->op_kernel().name(), " input ", index, " has shape ", in ResolveInputDynamismIntoPred() 312 context_->op_kernel().name(), " input ", index, " has shape ", in ResolveInputDynamismIntoPredVector() 453 TF_RETURN_IF_ERROR(op_kernel().InputRange(name, &start, &stop)); in ConstantInputList() 726 xla_context()->StackTraceForNodeName(op_kernel().name())) { in StackTrace()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | cwise_ops_common.cc | 39 const string& op = ctx->op_kernel().type_string(); in SetComputeError() 41 DataTypeIsInteger(ctx->op_kernel().input_type(0))) { in SetComputeError() 44 DataTypeIsInteger(ctx->op_kernel().input_type(0)) && in SetComputeError() 45 DataTypeIsSigned(ctx->op_kernel().input_type(1))) { in SetComputeError() 62 TryGetNodeAttr(ctx->op_kernel().def(), "incompatible_shape_error", in BinaryOpState() 65 const string& op = ctx->op_kernel().type_string(); in BinaryOpState()
|
D | scoped_allocator_ops.cc | 61 VLOG(1) << "_ScopedAllocatorOp " << context->op_kernel().name() in Compute() 141 context->op_kernel().requested_input(i), " to node ", in Compute() 142 context->op_kernel().name(), " input bounds = [", input_lb, ", ", in Compute() 149 context->op_kernel().requested_input(i), " to node ", in Compute() 150 context->op_kernel().name(), " input bounds = [", input_lb, ", ", in Compute() 212 " to node ", context->op_kernel().name())); in Compute() 218 " to node ", context->op_kernel().name())); in Compute()
|
D | segment_reduction_ops_impl_1.cc | 43 void UnsortedSegmentReductionValidation(OpKernel* op_kernel, in UnsortedSegmentReductionValidation() argument 59 bool UnsortedSegmentReductionDoValidation(OpKernel* op_kernel, in UnsortedSegmentReductionDoValidation() argument 64 UnsortedSegmentReductionValidation(op_kernel, context, data, segment_ids, in UnsortedSegmentReductionDoValidation()
|
D | ops_testutil.cc | 59 for (int index = 0; index < params->op_kernel->num_outputs(); index++) { in SetOutputAttrs() 62 (params->op_kernel->output_memory_types()[index] == HOST_MEMORY); in SetOutputAttrs() 170 params_->op_kernel = kernel_.get(); in RunOpKernel()
|
/external/tensorflow/tensorflow/core/grappler/optimizers/ |
D | evaluation_utils.cc | 71 std::unique_ptr<OpKernel> op_kernel( in EvaluateNode() local 79 params.op_kernel = op_kernel.get(); in EvaluateNode() 83 const int num_outputs = op_kernel->num_outputs(); in EvaluateNode() 92 op_kernel->Compute(&op_context); in EvaluateNode()
|
/external/tensorflow/tensorflow/core/framework/ |
D | op_kernel_test.cc | 368 params.op_kernel = op.get(); in TEST_F() 452 params.op_kernel = op.get(); in TEST_F() 741 params.op_kernel = op.get(); in TEST_F() 826 std::unique_ptr<OpKernel> op_kernel = in TEST_F() local 829 auto* get_attr_kernel = static_cast<GetAttrKernel*>(op_kernel.get()); in TEST_F() 833 op_kernel = ExpectSuccess("GetAttrStringList", DEVICE_CPU, in TEST_F() 835 get_attr_kernel = static_cast<GetAttrKernel*>(op_kernel.get()); in TEST_F() 847 std::unique_ptr<OpKernel> op_kernel = ExpectSuccess( in TEST_F() local 850 auto* get_attr_kernel = static_cast<GetAttrKernel*>(op_kernel.get()); in TEST_F() 855 op_kernel = ExpectSuccess( in TEST_F() [all …]
|
D | device.h | 87 virtual void Compute(OpKernel* op_kernel, OpKernelContext* context) { in Compute() argument 88 op_kernel->Compute(context); in Compute() 92 virtual void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() argument 94 op_kernel->ComputeAsync(context, std::move(done)); in ComputeAsync()
|
D | op_kernel.cc | 327 params, static_cast<int>(params->op_kernel->output_types().size())) {} in OpKernelContext() 421 CHECK_LT(index, num_inputs()) << " name: " << op_kernel().name(); in input() 628 TF_RETURN_IF_ERROR(params_->op_kernel->InputRange(name, &start, &stop)); in input_list() 636 TF_RETURN_IF_ERROR(params_->op_kernel->InputRange(name, &start, &stop)); in mutable_input_list() 643 TF_RETURN_IF_ERROR(params_->op_kernel->OutputRange(name, &start, &stop)); in output_list() 658 " kernel=", params_->op_kernel->name()); in allocate_output() 663 " kernel=", params_->op_kernel->name()); in allocate_output() 681 TF_RETURN_IF_ERROR(params_->op_kernel->OutputRange(name, &start, &stop)); in allocate_output() 696 TF_RETURN_IF_ERROR(params_->op_kernel->OutputRange(name, &start, &stop)); in allocate_output() 723 LogMemory::RecordTensorAllocation(params_->op_kernel->name(), in allocate_tensor() [all …]
|
/external/tensorflow/tensorflow/lite/tools/ |
D | list_flex_ops.cc | 39 Json::Value op_kernel(Json::arrayValue); in OpListToJSONString() local 40 op_kernel.append(Json::Value(op.op_name)); in OpListToJSONString() 41 op_kernel.append(Json::Value(op.kernel_name)); in OpListToJSONString() 42 result.append(op_kernel); in OpListToJSONString()
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | renamed_device.h | 122 void Compute(OpKernel* op_kernel, OpKernelContext* context) override { in Compute() argument 123 underlying_device_->Compute(op_kernel, context); in Compute() 126 void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() argument 128 underlying_device_->ComputeAsync(op_kernel, context, std::move(done)); in ComputeAsync()
|
D | executor.cc | 556 OpKernel* op_kernel = item.kernel; in ProcessSync() local 562 op_kernel->name_view()); in ProcessSync() 564 [op_kernel, &ctx] { in ProcessSync() 565 return op_kernel->TraceString( in ProcessSync() 569 device->Compute(op_kernel, &ctx); in ProcessSync() 572 device->Compute(op_kernel, &ctx); in ProcessSync() 582 device->Compute(op_kernel, &ctx); in ProcessSync() 800 params.op_kernel = item.kernel; in Process() 1032 LogMemory::RecordTensorOutput(ctx->op_kernel().name(), in ProcessOutputs() 1041 LogMemory::RecordTensorOutput(ctx->op_kernel().name(), in ProcessOutputs()
|
/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_tfconv_op.h | 64 static void ConvertMklToTf(OpKernel* op_kernel, OpKernelContext* context, in ConvertMklToTf() argument 83 DataType input_data_type = op_kernel->input_type(input_number); in ConvertMklToTf() 84 DataType output_data_type = op_kernel->output_type(input_number); in ConvertMklToTf()
|
/external/tensorflow/tensorflow/compiler/jit/ |
D | xla_device.cc | 408 void XlaDevice::Compute(OpKernel* op_kernel, OpKernelContext* context) { in Compute() argument 409 VLOG(2) << "XlaDevice::Compute " << op_kernel->name() << ":" in Compute() 410 << op_kernel->type_string(); in Compute() 412 op_kernel->Compute(context); in Compute() 415 void XlaDevice::ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() argument 418 VLOG(2) << "XlaDevice::ComputeAsync " << op_kernel->name() << ":" in ComputeAsync() 419 << op_kernel->type_string(); in ComputeAsync() 420 op_kernel->ComputeAsync(context, done); in ComputeAsync()
|
/external/tensorflow/tensorflow/core/tpu/kernels/ |
D | tpu_handle_to_key_op.cc | 35 VLOG(1) << "TpuHandleToProtoKeyOp::Compute " << ctx->op_kernel().name() in Compute() 36 << " on device " << ctx->op_kernel().requested_device(); in Compute()
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | conv_ops.cc | 56 ctx->op_kernel().type_string(), ctx->Input(0), ctx->Input(1), attrs_); in Compile() 117 ctx->op_kernel().type_string(), input_shape, ctx->Input(1), in Compile() 178 ctx->op_kernel().type_string(), ctx->Input(0), filter_shape, in Compile()
|
D | reduction_ops_common.cc | 50 VLOG(1) << "ReductionOp: " << ctx->op_kernel().name(); in Compile() 106 string desc = ctx->op_kernel().name(); in Compile()
|
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
D | batch_resource_base.cc | 179 context->op_kernel().name_view().data()); in RegisterInput() 182 context->op_kernel().name_view().data()); in RegisterInput() 185 context->op_kernel().name_view().data()); in RegisterInput() 188 context->op_kernel().name_view().data()); in RegisterInput() 191 context->op_kernel().name_view().data()); in RegisterInput() 332 context->op_kernel().name_view().data()); in ConcatInputTensors() 334 context->op_kernel().name_view().data()); in ConcatInputTensors() 610 last_task_context->op_kernel().name_view().data()); in ProcessFuncBatch()
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
D | gpu_device.cc | 126 operation_ = context->op_kernel().name() + "/EigenAllocator"; in Reinitialize() 531 string BaseGPUDevice::ComputeOpKernelDebugString(const OpKernel& op_kernel, in ComputeOpKernelDebugString() argument 533 return strings::StrCat(op_kernel.name(), " op ", op_kernel.type_string(), in ComputeOpKernelDebugString() 538 void BaseGPUDevice::Compute(OpKernel* op_kernel, OpKernelContext* context) { in Compute() argument 557 << ComputeOpKernelDebugString(*op_kernel, stream_id); in Compute() 567 ScopedMemoryDebugAnnotation op_annotation(op_kernel->name_view().data(), in Compute() 569 op_kernel->Compute(context); in Compute() 579 << ComputeOpKernelDebugString(*op_kernel, stream_id); in Compute() 583 << ComputeOpKernelDebugString(*op_kernel, stream_id); in Compute() 598 << ComputeOpKernelDebugString(*op_kernel, stream_id); in Compute() [all …]
|
D | gpu_device.h | 65 void Compute(OpKernel* op_kernel, OpKernelContext* context) override; 69 void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, 163 std::string ComputeOpKernelDebugString(const OpKernel& op_kernel,
|
/external/tensorflow/tensorflow/core/kernels/data/ |
D | map_defun_op_test.cc | 107 Status CreateMapDefunContext(OpKernel* const op_kernel, in CreateMapDefunContext() argument 110 TF_RETURN_IF_ERROR(CheckOpKernelInput(*op_kernel, *inputs)); in CreateMapDefunContext() 111 TF_RETURN_IF_ERROR(CreateOpKernelContext(op_kernel, inputs, context)); in CreateMapDefunContext()
|