Searched refs:op_kernel_context (Results 1 – 16 of 16) sorted by relevance
/external/tensorflow/tensorflow/compiler/mlir/tools/kernel_gen/tests/ |
D | ops.mlir | 8 func @alloc(%ctx: !tf_framework.op_kernel_context, 16 func @forwarding_alloc(%ctx: !tf_framework.op_kernel_context, 26 func @dealloc(%ctx: !tf_framework.op_kernel_context, 33 func @assert(%ctx: !tf_framework.op_kernel_context) { 46 tf_framework.null_context : !tf_framework.op_kernel_context
|
D | embed_tf_framework.mlir | 6 // CHECK-SAME: [[CTX:%.*]]: !tf_framework.op_kernel_context, 44 // CHECK-SAME: [[CTX:%.*]]: !tf_framework.op_kernel_context 45 func @assert(%arg0: !tf_framework.op_kernel_context) 73 // CHECK-SAME: [[CTX:%.*]]: !tf_framework.op_kernel_context 74 func @double_assert(%arg0: !tf_framework.op_kernel_context)
|
D | tf_framework_legalize_to_llvm.mlir | 10 func @alloc(%ctx: !tf_framework.op_kernel_context, 67 func @dealloc(%ctx: !tf_framework.op_kernel_context, 87 func @report_error(%ctx: !tf_framework.op_kernel_context) {
|
D | invalid.mlir | 3 func @alloc_raw(%ctx: !tf_framework.op_kernel_context, %size : index) {
|
D | print_memrefs.mlir | 4 %ctx: !tf_framework.op_kernel_context, %input: memref<*xf16>)
|
D | tf_kernel_gpu_launch_to_llvm.mlir | 19 func @launch(%ctx: !tf_framework.op_kernel_context, %memref: memref<?x10xf32>) {
|
D | tf_abi_knowledge.mlir | 10 func @abs(%ctx: !tf_framework.op_kernel_context, %arg0: memref<*xf32>, %size: index) 79 …func @add_same_shape(%arg0: !tf_framework.op_kernel_context, %arg1: memref<*xf32>, %arg2: memref<*… 166 …func @add_same_shape(%arg0: !tf_framework.op_kernel_context, %arg1: memref<*xf32>, %arg2: memref<*… 274 …func @add_one_scalar(%arg0: !tf_framework.op_kernel_context, %arg1: memref<*xf32>, %arg2: memref<*…
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | identity_op.cc | 35 ctx->op_kernel_context()->set_output( in Compile() 36 i, ctx->op_kernel_context()->input(i)); in Compile()
|
D | retval_op.cc | 38 const Tensor& input = ctx->op_kernel_context()->input(0); in Compile()
|
D | arg_op.cc | 56 ctx->op_kernel_context()->set_output(0, *val); in Compile()
|
D | while_op.cc | 623 ctx->op_kernel_context()->set_output( in Compile() 625 ctx->op_kernel_context()->input(update.input_index)); in Compile()
|
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
D | batch_resource_base.cc | 393 op_kernel_context = input_task.context, status = shared_status]() { in SplitInputTask() 394 const int num_output = op_kernel_context->num_outputs(); in SplitInputTask() 408 Concat(op_kernel_context, to_concatenate, &output_tensor); in SplitInputTask() 413 op_kernel_context->set_output(i, std::move(output_tensor)); in SplitInputTask() 415 op_kernel_context->SetStatus(status->status()); in SplitInputTask()
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | xla_op_kernel.h | 265 OpKernelContext* op_kernel_context() const { return context_; } in op_kernel_context() function
|
D | xla_compiler_test.cc | 105 ResourceMgr* rm = ctx->op_kernel_context()->resource_manager(); in Compile()
|
/external/tensorflow/tensorflow/core/kernels/data/ |
D | dataset_test_base.cc | 656 CreateIteratorContext(dataset->op_kernel_context(), &iterator_ctx)); in CheckSplitProviderShardedIteration() 863 CreateIteratorContext(dataset.op_kernel_context(), &iterator_ctx)); in MakeIterator()
|
D | dataset_test_base.h | 492 OpKernelContext* op_kernel_context() const { return ctx_.get(); } in op_kernel_context() function
|