Home
last modified time | relevance | path

Searched refs:in_tensor (Results 1 – 25 of 37) sorted by relevance

12

/external/pytorch/aten/src/ATen/native/cuda/
DTensorTransformations.cu51 scalar_t* in_tensor, in C10_LAUNCH_BOUNDS_1()
79 out_tensor[linear_index] = in_tensor[dst_offset]; in C10_LAUNCH_BOUNDS_1()
85 const scalar_t* in_tensor, in C10_LAUNCH_BOUNDS_1()
106 out_tensor[linear_index] = in_tensor[source_idx]; in C10_LAUNCH_BOUNDS_1()
115 auto in_tensor = self; in roll_cuda() local
117 in_tensor = self.contiguous(); in roll_cuda()
119 auto out_tensor = at::empty_like(in_tensor, LEGACY_CONTIGUOUS_MEMORY_FORMAT); in roll_cuda()
123 const int64_t N = in_tensor.numel(); in roll_cuda()
125 const int64_t size = in_tensor.size(dim); in roll_cuda()
133 TORCH_CHECK(cuda::getApplyGrid(N, dim_grid, in_tensor.get_device()), "unable to get dim grid"); in roll_cuda()
[all …]
/external/pytorch/torch/distributed/tensor/
D_tp_conv.py40 def _ring_send_recv_construct(in_tensor, d1, d2, left, right, rank, size): argument
42 send_to_right = in_tensor[:, :, :, -d1:].contiguous()
43 send_to_left = in_tensor[:, :, :, :d2].contiguous()
59 in_tensor = torch.cat([in_tensor, recv_from_right], dim=-1)
61 in_tensor = torch.cat([recv_from_left, in_tensor], dim=-1)
63 in_tensor = torch.cat([recv_from_left, in_tensor, recv_from_right], dim=-1)
65 return in_tensor
116 in_tensor = cast(torch.Tensor, local_tensor_args[0])
120 assert _is_supported(in_tensor.shape, weight.shape, stride, padding, dilation)
136 in_tensor = _ring_send_recv_construct(
[all …]
/external/tensorflow/tensorflow/lite/python/
Dlite_test.py121 in_tensor = array_ops.placeholder(
123 out_tensor = math_ops.add(in_tensor, in_tensor, name='add')
134 converter = lite.TFLiteConverter(frozen_graph_def, [in_tensor],
151 in_tensor = array_ops.placeholder(
153 math_ops.add(in_tensor, in_tensor, name='add')
208 in_tensor = array_ops.placeholder(
210 math_ops.add(in_tensor, in_tensor, name='add')
229 in_tensor = array_ops.placeholder(
231 out_tensor = in_tensor + in_tensor
235 converter = lite.TFLiteConverter.from_session(sess, [in_tensor],
[all …]
Dlite_flex_test.py52 in_tensor = array_ops.placeholder(shape=[1, 4], dtype=dtypes.float32)
53 out_tensor = in_tensor + in_tensor
57 converter = lite.TFLiteConverter.from_session(sess, [in_tensor],
81 in_tensor = array_ops.placeholder(
83 out_tensor = nn_ops.l2_loss(in_tensor)
84 converter = lite.TFLiteConverter.from_session(sess, [in_tensor],
94 in_tensor = array_ops.placeholder(shape=[1, 4], dtype=dtypes.float32)
95 out_tensor = in_tensor + in_tensor
99 converter = lite.TFLiteConverter.from_session(sess, [in_tensor],
216 in_tensor = array_ops.placeholder(
[all …]
Dtflite_convert_test.py117 in_tensor = array_ops.placeholder(
119 _ = in_tensor + in_tensor
138 in_tensor = array_ops.placeholder(
140 _ = in_tensor + in_tensor
157 in_tensor = random_ops.random_normal(shape=[1, 16, 16, 3], name='random')
158 _ = in_tensor + in_tensor
230 in_tensor = array_ops.placeholder(
232 out_tensor = in_tensor + in_tensor
233 inputs = {'x': in_tensor}
250 in_tensor = array_ops.placeholder(
[all …]
Dconvert_saved_model_test.py42 in_tensor = array_ops.placeholder(shape=shape, dtype=dtypes.float32)
43 out_tensor = in_tensor + in_tensor
44 inputs = {"x": in_tensor}
215 in_tensor = array_ops.placeholder(shape=[1, 28, 28], dtype=dtypes.float32)
216 out_tensor = in_tensor + in_tensor
217 sig_input_tensor = saved_model.utils.build_tensor_info(in_tensor)
Dconvert_test.py37 in_tensor = array_ops.placeholder(
39 out_tensor = in_tensor + in_tensor
44 sess.graph_def, input_tensors=[in_tensor], output_tensors=[out_tensor])
49 in_tensor = array_ops.placeholder(
52 in_tensor + in_tensor, min=0., max=1.)
57 input_tensors=[in_tensor],
65 in_tensor = array_ops.placeholder(
67 _ = in_tensor + in_tensor
Dutil_test.py69 in_tensor = array_ops.placeholder(dtype=dtypes.float32, shape=[4])
71 value=in_tensor, num_or_size_splits=[1, 1, 1, 1], axis=0)
155 in_tensor = array_ops.placeholder(
157 _ = in_tensor + in_tensor
165 in_tensor = array_ops.placeholder(
167 _ = in_tensor + in_tensor
Dmetrics_nonportable_test.py131 in_tensor = array_ops.placeholder(
133 math_ops.add(in_tensor, in_tensor, name='add')
453 in_tensor = array_ops.placeholder(
455 out_tensor = in_tensor + in_tensor
456 inputs = {'x': in_tensor}
Dlite_v2_test.py1434 in_tensor = tf.compat.v1.placeholder(
1437 out_tensor = in_tensor + constant
1438 inputs = {'x': in_tensor}
1488 in_tensor = tf.compat.v1.placeholder(
1491 assert_op = tf.Assert(tf.less_equal(in_tensor, constant), [in_tensor])
1493 out_tensor = in_tensor + constant
1494 inputs = {'x': in_tensor}
2385 in_tensor = tf.compat.v1.placeholder(
2387 out_tensor = in_tensor + in_tensor
2388 inputs = {'input': in_tensor}
[all …]
/external/executorch/backends/vulkan/runtime/graph/ops/impl/
DStaging.cpp67 const ValueRef in_tensor, in add_tensor_to_staging_node() argument
72 *graph.get_tensor(in_tensor), graph.int8_buffers_enabled()); in add_tensor_to_staging_node()
74 utils::uvec3 global_wg_size = graph.create_global_wg_size(in_tensor); in add_tensor_to_staging_node()
77 if (graph.is_buffer_storage(in_tensor)) { in add_tensor_to_staging_node()
79 {graph.sizes_ubo(in_tensor), in add_tensor_to_staging_node()
80 graph.strides_ubo(in_tensor), in add_tensor_to_staging_node()
81 graph.numel_ubo(in_tensor)}); in add_tensor_to_staging_node()
83 ubos.append({graph.sizes_ubo(in_tensor)}); in add_tensor_to_staging_node()
97 ubos.append({graph.numel_ubo(in_tensor)}); in add_tensor_to_staging_node()
106 {{out_staging, vkapi::kWrite}, {in_tensor, vkapi::kRead}}, in add_tensor_to_staging_node()
[all …]
/external/tensorflow/tensorflow/python/kernel_tests/
Dcollective_ops_multi_worker_test.py177 in_tensor = constant_op.constant([1.])
183 in_tensor,
201 in_tensor,
211 in_tensor,
223 in_tensor,
238 in_tensor = constant_op.constant([1.])
253 collective_ops.all_reduce(in_tensor, group_size, group_key,
259 collective_ops.all_reduce(in_tensor, group_size, group_key,
267 collective_ops.all_reduce(in_tensor, group_size, group_key, instance_key)
278 in_tensor = constant_op.constant([1.])
[all …]
Dcollective_ops_test.py377 in_tensor = constant_op.constant(in_value)
386 in_tensor,
396 in_tensor,
411 in_tensor = constant_op.constant(in_value)
420 in_tensor,
433 in_tensor,
738 in_tensor = constant_op.constant([1.])
757 in_tensor,
768 in_tensor,
784 in_tensor,
[all …]
/external/tensorflow/tensorflow/compiler/xla/experimental/xla_sharding/
Dxla_sharding_test.py83 in_tensor = array_ops.ones([4, 5, 6], dtype=dtypes.float32)
85 self.assertAllEqual(in_tensor, result)
101 in_tensor = array_ops.ones([4, 5, 6], dtype=dtypes.float32)
103 self.assertAllEqual(in_tensor, result)
118 in_tensor = array_ops.ones([4, 5, 6], dtype=dtypes.float32)
121 self.assertAllEqual(in_tensor, result)
158 in_tensor = array_ops.ones([4, 5, 6], dtype=dtypes.float32)
161 self.assertAllEqual(in_tensor, result)
/external/executorch/backends/cadence/aot/
Dsimplify_ops.py73 in_tensor = args[0].to_tensor() if isinstance(args[0], ProxyValue) else args[0]
76 dim = dim if dim >= 0 else dim + in_tensor.dim()
77 length = in_tensor.size(dim)
90 empty_shape = [x for x in in_tensor.shape if x != 0]
95 {"dtype": in_tensor.dtype},
Dreplace_ops.py255 in_tensor = args[0].to_tensor() if isinstance(args[0], ProxyValue) else args[0]
256 in_shape = in_tensor.shape
282 in_tensor = args[0].to_tensor() if isinstance(args[0], ProxyValue) else args[0]
285 if in_tensor.dim() <= 1:
288 assert in_tensor.dim() == 2, "t_copy expects a tensor with <= 2 dimensions"
427 in_tensor = args[0].to_tensor() if isinstance(args[0], ProxyValue) else args[0]
428 old_dims = tuple(range(in_tensor.dim()))
495 in_tensor = args[0]
500 in_tensor.to_tensor().shape
501 if isinstance(in_tensor, ProxyValue)
[all …]
/external/tensorflow/tensorflow/core/kernels/
Dcast_op.h35 typename TTypes<IN_OUT>::ConstFlat in_tensor, \
39 in_tensor.unaryExpr(LSBZeroSetter<IN_OUT, OUT_TYPE>()) \
42 out_tensor.device(d) = in_tensor.template cast<OUT_TYPE>(); \
53 typename TTypes<IN_OUT>::ConstFlat in_tensor, \
57 in_tensor.unaryExpr(LSBZeroSetter<IN_OUT, OUT_TYPE>()) \
60 out_tensor.device(d) = in_tensor.template cast<OUT_TYPE>(); \
80 typename TTypes<IN_OUT>::ConstFlat in_tensor, \
82 out_tensor.device(d) = in_tensor.template cast<OUT_TYPE>(); \
100 typename TTypes<IN_OUT>::ConstFlat in_tensor, \
102 out_tensor.device(d) = in_tensor.template cast<OUT_TYPE>(); \
/external/tensorflow/tensorflow/core/kernels/mkl/
Dmkl_transpose_op.cc65 Status MKLTransposeND(OpKernelContext* context, const Tensor& in_tensor, in MKLTransposeND() argument
72 memory::dims in_dims = TFShapeToMklDnnDims(in_tensor.shape()); in MKLTransposeND()
80 in.SetUsrMem(in_dims, in_strides, &in_tensor); in MKLTransposeND()
89 in.SetUsrMemDataHandle(&in_tensor, transpose_stream); in MKLTransposeND()
/external/ComputeLibrary/src/dynamic_fusion/sketch/utils/
DDependencyGraph.h194 for(auto in_tensor : inputs) variable
198 link_input(op, in_tensor);
368 void link_input(OperatorId op, TensorId in_tensor) in link_input() argument
371 if(!tensor_exists(in_tensor)) in link_input()
373 insert_new_tensor(in_tensor); in link_input()
375 ARM_COMPUTE_ERROR_ON(are_connected(op, in_tensor)); // Prevent repetitive linking in link_input()
376 _adj_src_tensors[op].push_back(in_tensor); in link_input()
377 _adj_dst_ops[in_tensor].push_back(op); in link_input()
/external/tensorflow/tensorflow/lite/toco/python/
Dtoco_from_protos_test.py33 def _run(self, sess, in_tensor, out_tensor, should_succeed): argument
52 input_array.name = TensorName(in_tensor)
53 input_array.shape.dims.extend(map(int, in_tensor.shape))
/external/executorch/backends/vulkan/tools/gpuinfo/include/
Dtextures.h61 vTensor in_tensor = in tex_cacheline_concurr() local
79 in_tensor.image(), in tex_cacheline_concurr()
161 vTensor in_tensor = api::vTensor(api::context(), sizes_nchw, vkapi::kFloat); in tex_bandwidth() local
193 in_tensor.image(), in tex_bandwidth()
/external/tensorflow/tensorflow/python/ops/
Dcollective_ops_gpu_test.py290 in_tensor = constant_op.constant(1.)
297 in_tensor,
318 in_tensor,
329 in_tensor,
/external/tensorflow/tensorflow/lite/python/metrics/
Dmetrics_nonportable_test.py131 in_tensor = array_ops.placeholder(
133 math_ops.add(in_tensor, in_tensor, name='add')
455 in_tensor = array_ops.placeholder(
457 out_tensor = in_tensor + in_tensor
458 inputs = {'x': in_tensor}
/external/executorch/backends/vulkan/runtime/
DVulkanBackend.cpp394 vTensorPtr in_tensor = graph->get_tensor(in_tensor_ref); in maybe_resize_input() local
397 et_tensor.dim() == in_tensor->sizes().size(), in maybe_resize_input()
399 static_cast<size_t>(in_tensor->sizes().size()), in maybe_resize_input()
405 if (in_tensor->sizes()[i] != et_tensor.sizes()[i]) { in maybe_resize_input()
416 in_tensor->numel() == et_tensor.numel(), in maybe_resize_input()
418 static_cast<size_t>(in_tensor->numel()), in maybe_resize_input()
/external/pytorch/test/distributed/algorithms/quantization/
Dtest_quantization.py289 in_tensor = torch.ones([sum(in_splits), size], dtype=dtype) * rank
296 in_tensor = in_tensor.cuda(rank_to_GPU)
304 in_tensor,

12