Home
last modified time | relevance | path

Searched refs:op_kernel (Results 1 – 25 of 77) sorted by relevance

1234

/external/tensorflow/tensorflow/compiler/tf2xla/
Dxla_compilation_device.cc91 OpKernel* op_kernel, XlaContext& context) { in AttachLocationToMetadata() argument
93 context.StackTraceForNodeName(op_kernel->def().name())) { in AttachLocationToMetadata()
101 void XlaCompilationDevice::Compute(OpKernel* op_kernel, in Compute() argument
104 << FormatNodeDefForError(op_kernel->def()); in Compute()
108 metadata.set_op_type(op_kernel->type_string()); in Compute()
109 metadata.set_op_name(op_kernel->name()); in Compute()
110 AttachLocationToMetadata(metadata, op_kernel, xla_context); in Compute()
114 ParseShardingFromDevice(op_kernel->def(), std::numeric_limits<int>::max(), in Compute()
121 GetFrontendAttributesFromAttrSlice(AttrSlice(op_kernel->def())); in Compute()
138 op_kernel->Compute(context); in Compute()
Dconst_analysis.cc88 Status GetCompileTimeConstInputs(const NodeDef& node, const OpKernel* op_kernel, in GetCompileTimeConstInputs() argument
92 DCHECK(op_def != nullptr || op_kernel != nullptr); in GetCompileTimeConstInputs()
167 return XlaOpRegistry::CompileTimeConstantInputs(*op_kernel, in GetCompileTimeConstInputs()
287 Status GetCompileTimeConstInputs(const OpKernel* op_kernel, in GetCompileTimeConstInputs() argument
290 return GetCompileTimeConstInputs(op_kernel->def(), op_kernel, in GetCompileTimeConstInputs()
Dxla_op_registry.h199 static Status CompileTimeConstantInputs(const OpKernel& op_kernel, in CompileTimeConstantInputs() argument
201 return CompileTimeConstantInputs(op_kernel.def(), /*op_kernel=*/&op_kernel, in CompileTimeConstantInputs()
296 const OpKernel* op_kernel,
Dgraph_compiler.cc141 std::unique_ptr<OpKernel> op_kernel(op_kernel_raw); in Compile() local
151 params.op_kernel = op_kernel.get(); in Compile()
172 VLOG(3) << "Translating " << params.op_kernel->name(); in Compile()
176 device_->Compute(CHECK_NOTNULL(params.op_kernel), &op_context); in Compile()
Dxla_op_kernel.cc139 TF_RETURN_IF_ERROR(context->op_kernel().InputRange(name, &start, &stop)); in InputIndex()
165 context_->op_kernel().type_string(), in ConstantInputReshaped()
172 "Input ", index, " to node `", context_->op_kernel().name(), in ConstantInputReshaped()
173 "` with op ", context_->op_kernel().type_string(), in ConstantInputReshaped()
186 context_->op_kernel().name(), " input ", index, " has shape ", in ConstantInputReshaped()
282 context_->op_kernel().name(), " input ", index, " has shape ", in ResolveInputDynamismIntoPred()
312 context_->op_kernel().name(), " input ", index, " has shape ", in ResolveInputDynamismIntoPredVector()
453 TF_RETURN_IF_ERROR(op_kernel().InputRange(name, &start, &stop)); in ConstantInputList()
726 xla_context()->StackTraceForNodeName(op_kernel().name())) { in StackTrace()
/external/tensorflow/tensorflow/core/kernels/
Dcwise_ops_common.cc39 const string& op = ctx->op_kernel().type_string(); in SetComputeError()
41 DataTypeIsInteger(ctx->op_kernel().input_type(0))) { in SetComputeError()
44 DataTypeIsInteger(ctx->op_kernel().input_type(0)) && in SetComputeError()
45 DataTypeIsSigned(ctx->op_kernel().input_type(1))) { in SetComputeError()
62 TryGetNodeAttr(ctx->op_kernel().def(), "incompatible_shape_error", in BinaryOpState()
65 const string& op = ctx->op_kernel().type_string(); in BinaryOpState()
Dscoped_allocator_ops.cc61 VLOG(1) << "_ScopedAllocatorOp " << context->op_kernel().name() in Compute()
141 context->op_kernel().requested_input(i), " to node ", in Compute()
142 context->op_kernel().name(), " input bounds = [", input_lb, ", ", in Compute()
149 context->op_kernel().requested_input(i), " to node ", in Compute()
150 context->op_kernel().name(), " input bounds = [", input_lb, ", ", in Compute()
212 " to node ", context->op_kernel().name())); in Compute()
218 " to node ", context->op_kernel().name())); in Compute()
Dsegment_reduction_ops_impl_1.cc43 void UnsortedSegmentReductionValidation(OpKernel* op_kernel, in UnsortedSegmentReductionValidation() argument
59 bool UnsortedSegmentReductionDoValidation(OpKernel* op_kernel, in UnsortedSegmentReductionDoValidation() argument
64 UnsortedSegmentReductionValidation(op_kernel, context, data, segment_ids, in UnsortedSegmentReductionDoValidation()
Dops_testutil.cc59 for (int index = 0; index < params->op_kernel->num_outputs(); index++) { in SetOutputAttrs()
62 (params->op_kernel->output_memory_types()[index] == HOST_MEMORY); in SetOutputAttrs()
170 params_->op_kernel = kernel_.get(); in RunOpKernel()
/external/tensorflow/tensorflow/core/grappler/optimizers/
Devaluation_utils.cc71 std::unique_ptr<OpKernel> op_kernel( in EvaluateNode() local
79 params.op_kernel = op_kernel.get(); in EvaluateNode()
83 const int num_outputs = op_kernel->num_outputs(); in EvaluateNode()
92 op_kernel->Compute(&op_context); in EvaluateNode()
/external/tensorflow/tensorflow/core/framework/
Dop_kernel_test.cc368 params.op_kernel = op.get(); in TEST_F()
452 params.op_kernel = op.get(); in TEST_F()
741 params.op_kernel = op.get(); in TEST_F()
826 std::unique_ptr<OpKernel> op_kernel = in TEST_F() local
829 auto* get_attr_kernel = static_cast<GetAttrKernel*>(op_kernel.get()); in TEST_F()
833 op_kernel = ExpectSuccess("GetAttrStringList", DEVICE_CPU, in TEST_F()
835 get_attr_kernel = static_cast<GetAttrKernel*>(op_kernel.get()); in TEST_F()
847 std::unique_ptr<OpKernel> op_kernel = ExpectSuccess( in TEST_F() local
850 auto* get_attr_kernel = static_cast<GetAttrKernel*>(op_kernel.get()); in TEST_F()
855 op_kernel = ExpectSuccess( in TEST_F()
[all …]
Ddevice.h87 virtual void Compute(OpKernel* op_kernel, OpKernelContext* context) { in Compute() argument
88 op_kernel->Compute(context); in Compute()
92 virtual void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() argument
94 op_kernel->ComputeAsync(context, std::move(done)); in ComputeAsync()
Dop_kernel.cc327 params, static_cast<int>(params->op_kernel->output_types().size())) {} in OpKernelContext()
421 CHECK_LT(index, num_inputs()) << " name: " << op_kernel().name(); in input()
628 TF_RETURN_IF_ERROR(params_->op_kernel->InputRange(name, &start, &stop)); in input_list()
636 TF_RETURN_IF_ERROR(params_->op_kernel->InputRange(name, &start, &stop)); in mutable_input_list()
643 TF_RETURN_IF_ERROR(params_->op_kernel->OutputRange(name, &start, &stop)); in output_list()
658 " kernel=", params_->op_kernel->name()); in allocate_output()
663 " kernel=", params_->op_kernel->name()); in allocate_output()
681 TF_RETURN_IF_ERROR(params_->op_kernel->OutputRange(name, &start, &stop)); in allocate_output()
696 TF_RETURN_IF_ERROR(params_->op_kernel->OutputRange(name, &start, &stop)); in allocate_output()
723 LogMemory::RecordTensorAllocation(params_->op_kernel->name(), in allocate_tensor()
[all …]
/external/tensorflow/tensorflow/lite/tools/
Dlist_flex_ops.cc39 Json::Value op_kernel(Json::arrayValue); in OpListToJSONString() local
40 op_kernel.append(Json::Value(op.op_name)); in OpListToJSONString()
41 op_kernel.append(Json::Value(op.kernel_name)); in OpListToJSONString()
42 result.append(op_kernel); in OpListToJSONString()
/external/tensorflow/tensorflow/core/common_runtime/
Drenamed_device.h122 void Compute(OpKernel* op_kernel, OpKernelContext* context) override { in Compute() argument
123 underlying_device_->Compute(op_kernel, context); in Compute()
126 void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() argument
128 underlying_device_->ComputeAsync(op_kernel, context, std::move(done)); in ComputeAsync()
Dexecutor.cc556 OpKernel* op_kernel = item.kernel; in ProcessSync() local
562 op_kernel->name_view()); in ProcessSync()
564 [op_kernel, &ctx] { in ProcessSync()
565 return op_kernel->TraceString( in ProcessSync()
569 device->Compute(op_kernel, &ctx); in ProcessSync()
572 device->Compute(op_kernel, &ctx); in ProcessSync()
582 device->Compute(op_kernel, &ctx); in ProcessSync()
800 params.op_kernel = item.kernel; in Process()
1032 LogMemory::RecordTensorOutput(ctx->op_kernel().name(), in ProcessOutputs()
1041 LogMemory::RecordTensorOutput(ctx->op_kernel().name(), in ProcessOutputs()
/external/tensorflow/tensorflow/core/kernels/mkl/
Dmkl_tfconv_op.h64 static void ConvertMklToTf(OpKernel* op_kernel, OpKernelContext* context, in ConvertMklToTf() argument
83 DataType input_data_type = op_kernel->input_type(input_number); in ConvertMklToTf()
84 DataType output_data_type = op_kernel->output_type(input_number); in ConvertMklToTf()
/external/tensorflow/tensorflow/compiler/jit/
Dxla_device.cc408 void XlaDevice::Compute(OpKernel* op_kernel, OpKernelContext* context) { in Compute() argument
409 VLOG(2) << "XlaDevice::Compute " << op_kernel->name() << ":" in Compute()
410 << op_kernel->type_string(); in Compute()
412 op_kernel->Compute(context); in Compute()
415 void XlaDevice::ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() argument
418 VLOG(2) << "XlaDevice::ComputeAsync " << op_kernel->name() << ":" in ComputeAsync()
419 << op_kernel->type_string(); in ComputeAsync()
420 op_kernel->ComputeAsync(context, done); in ComputeAsync()
/external/tensorflow/tensorflow/core/tpu/kernels/
Dtpu_handle_to_key_op.cc35 VLOG(1) << "TpuHandleToProtoKeyOp::Compute " << ctx->op_kernel().name() in Compute()
36 << " on device " << ctx->op_kernel().requested_device(); in Compute()
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/
Dconv_ops.cc56 ctx->op_kernel().type_string(), ctx->Input(0), ctx->Input(1), attrs_); in Compile()
117 ctx->op_kernel().type_string(), input_shape, ctx->Input(1), in Compile()
178 ctx->op_kernel().type_string(), ctx->Input(0), filter_shape, in Compile()
Dreduction_ops_common.cc50 VLOG(1) << "ReductionOp: " << ctx->op_kernel().name(); in Compile()
106 string desc = ctx->op_kernel().name(); in Compile()
/external/tensorflow/tensorflow/core/kernels/batching_util/
Dbatch_resource_base.cc179 context->op_kernel().name_view().data()); in RegisterInput()
182 context->op_kernel().name_view().data()); in RegisterInput()
185 context->op_kernel().name_view().data()); in RegisterInput()
188 context->op_kernel().name_view().data()); in RegisterInput()
191 context->op_kernel().name_view().data()); in RegisterInput()
332 context->op_kernel().name_view().data()); in ConcatInputTensors()
334 context->op_kernel().name_view().data()); in ConcatInputTensors()
610 last_task_context->op_kernel().name_view().data()); in ProcessFuncBatch()
/external/tensorflow/tensorflow/core/common_runtime/gpu/
Dgpu_device.cc126 operation_ = context->op_kernel().name() + "/EigenAllocator"; in Reinitialize()
531 string BaseGPUDevice::ComputeOpKernelDebugString(const OpKernel& op_kernel, in ComputeOpKernelDebugString() argument
533 return strings::StrCat(op_kernel.name(), " op ", op_kernel.type_string(), in ComputeOpKernelDebugString()
538 void BaseGPUDevice::Compute(OpKernel* op_kernel, OpKernelContext* context) { in Compute() argument
557 << ComputeOpKernelDebugString(*op_kernel, stream_id); in Compute()
567 ScopedMemoryDebugAnnotation op_annotation(op_kernel->name_view().data(), in Compute()
569 op_kernel->Compute(context); in Compute()
579 << ComputeOpKernelDebugString(*op_kernel, stream_id); in Compute()
583 << ComputeOpKernelDebugString(*op_kernel, stream_id); in Compute()
598 << ComputeOpKernelDebugString(*op_kernel, stream_id); in Compute()
[all …]
Dgpu_device.h65 void Compute(OpKernel* op_kernel, OpKernelContext* context) override;
69 void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context,
163 std::string ComputeOpKernelDebugString(const OpKernel& op_kernel,
/external/tensorflow/tensorflow/core/kernels/data/
Dmap_defun_op_test.cc107 Status CreateMapDefunContext(OpKernel* const op_kernel, in CreateMapDefunContext() argument
110 TF_RETURN_IF_ERROR(CheckOpKernelInput(*op_kernel, *inputs)); in CreateMapDefunContext()
111 TF_RETURN_IF_ERROR(CreateOpKernelContext(op_kernel, inputs, context)); in CreateMapDefunContext()

1234