/external/pytorch/aten/src/ATen/native/cuda/ |
D | TensorTransformations.cu | 51 scalar_t* in_tensor, in C10_LAUNCH_BOUNDS_1() 79 out_tensor[linear_index] = in_tensor[dst_offset]; in C10_LAUNCH_BOUNDS_1() 85 const scalar_t* in_tensor, in C10_LAUNCH_BOUNDS_1() 106 out_tensor[linear_index] = in_tensor[source_idx]; in C10_LAUNCH_BOUNDS_1() 115 auto in_tensor = self; in roll_cuda() local 117 in_tensor = self.contiguous(); in roll_cuda() 119 auto out_tensor = at::empty_like(in_tensor, LEGACY_CONTIGUOUS_MEMORY_FORMAT); in roll_cuda() 123 const int64_t N = in_tensor.numel(); in roll_cuda() 125 const int64_t size = in_tensor.size(dim); in roll_cuda() 133 TORCH_CHECK(cuda::getApplyGrid(N, dim_grid, in_tensor.get_device()), "unable to get dim grid"); in roll_cuda() [all …]
|
/external/pytorch/torch/distributed/tensor/ |
D | _tp_conv.py | 40 def _ring_send_recv_construct(in_tensor, d1, d2, left, right, rank, size): argument 42 send_to_right = in_tensor[:, :, :, -d1:].contiguous() 43 send_to_left = in_tensor[:, :, :, :d2].contiguous() 59 in_tensor = torch.cat([in_tensor, recv_from_right], dim=-1) 61 in_tensor = torch.cat([recv_from_left, in_tensor], dim=-1) 63 in_tensor = torch.cat([recv_from_left, in_tensor, recv_from_right], dim=-1) 65 return in_tensor 116 in_tensor = cast(torch.Tensor, local_tensor_args[0]) 120 assert _is_supported(in_tensor.shape, weight.shape, stride, padding, dilation) 136 in_tensor = _ring_send_recv_construct( [all …]
|
/external/tensorflow/tensorflow/lite/python/ |
D | lite_test.py | 121 in_tensor = array_ops.placeholder( 123 out_tensor = math_ops.add(in_tensor, in_tensor, name='add') 134 converter = lite.TFLiteConverter(frozen_graph_def, [in_tensor], 151 in_tensor = array_ops.placeholder( 153 math_ops.add(in_tensor, in_tensor, name='add') 208 in_tensor = array_ops.placeholder( 210 math_ops.add(in_tensor, in_tensor, name='add') 229 in_tensor = array_ops.placeholder( 231 out_tensor = in_tensor + in_tensor 235 converter = lite.TFLiteConverter.from_session(sess, [in_tensor], [all …]
|
D | lite_flex_test.py | 52 in_tensor = array_ops.placeholder(shape=[1, 4], dtype=dtypes.float32) 53 out_tensor = in_tensor + in_tensor 57 converter = lite.TFLiteConverter.from_session(sess, [in_tensor], 81 in_tensor = array_ops.placeholder( 83 out_tensor = nn_ops.l2_loss(in_tensor) 84 converter = lite.TFLiteConverter.from_session(sess, [in_tensor], 94 in_tensor = array_ops.placeholder(shape=[1, 4], dtype=dtypes.float32) 95 out_tensor = in_tensor + in_tensor 99 converter = lite.TFLiteConverter.from_session(sess, [in_tensor], 216 in_tensor = array_ops.placeholder( [all …]
|
D | tflite_convert_test.py | 117 in_tensor = array_ops.placeholder( 119 _ = in_tensor + in_tensor 138 in_tensor = array_ops.placeholder( 140 _ = in_tensor + in_tensor 157 in_tensor = random_ops.random_normal(shape=[1, 16, 16, 3], name='random') 158 _ = in_tensor + in_tensor 230 in_tensor = array_ops.placeholder( 232 out_tensor = in_tensor + in_tensor 233 inputs = {'x': in_tensor} 250 in_tensor = array_ops.placeholder( [all …]
|
D | convert_saved_model_test.py | 42 in_tensor = array_ops.placeholder(shape=shape, dtype=dtypes.float32) 43 out_tensor = in_tensor + in_tensor 44 inputs = {"x": in_tensor} 215 in_tensor = array_ops.placeholder(shape=[1, 28, 28], dtype=dtypes.float32) 216 out_tensor = in_tensor + in_tensor 217 sig_input_tensor = saved_model.utils.build_tensor_info(in_tensor)
|
D | convert_test.py | 37 in_tensor = array_ops.placeholder( 39 out_tensor = in_tensor + in_tensor 44 sess.graph_def, input_tensors=[in_tensor], output_tensors=[out_tensor]) 49 in_tensor = array_ops.placeholder( 52 in_tensor + in_tensor, min=0., max=1.) 57 input_tensors=[in_tensor], 65 in_tensor = array_ops.placeholder( 67 _ = in_tensor + in_tensor
|
D | util_test.py | 69 in_tensor = array_ops.placeholder(dtype=dtypes.float32, shape=[4]) 71 value=in_tensor, num_or_size_splits=[1, 1, 1, 1], axis=0) 155 in_tensor = array_ops.placeholder( 157 _ = in_tensor + in_tensor 165 in_tensor = array_ops.placeholder( 167 _ = in_tensor + in_tensor
|
D | metrics_nonportable_test.py | 131 in_tensor = array_ops.placeholder( 133 math_ops.add(in_tensor, in_tensor, name='add') 453 in_tensor = array_ops.placeholder( 455 out_tensor = in_tensor + in_tensor 456 inputs = {'x': in_tensor}
|
D | lite_v2_test.py | 1434 in_tensor = tf.compat.v1.placeholder( 1437 out_tensor = in_tensor + constant 1438 inputs = {'x': in_tensor} 1488 in_tensor = tf.compat.v1.placeholder( 1491 assert_op = tf.Assert(tf.less_equal(in_tensor, constant), [in_tensor]) 1493 out_tensor = in_tensor + constant 1494 inputs = {'x': in_tensor} 2385 in_tensor = tf.compat.v1.placeholder( 2387 out_tensor = in_tensor + in_tensor 2388 inputs = {'input': in_tensor} [all …]
|
/external/executorch/backends/vulkan/runtime/graph/ops/impl/ |
D | Staging.cpp | 67 const ValueRef in_tensor, in add_tensor_to_staging_node() argument 72 *graph.get_tensor(in_tensor), graph.int8_buffers_enabled()); in add_tensor_to_staging_node() 74 utils::uvec3 global_wg_size = graph.create_global_wg_size(in_tensor); in add_tensor_to_staging_node() 77 if (graph.is_buffer_storage(in_tensor)) { in add_tensor_to_staging_node() 79 {graph.sizes_ubo(in_tensor), in add_tensor_to_staging_node() 80 graph.strides_ubo(in_tensor), in add_tensor_to_staging_node() 81 graph.numel_ubo(in_tensor)}); in add_tensor_to_staging_node() 83 ubos.append({graph.sizes_ubo(in_tensor)}); in add_tensor_to_staging_node() 97 ubos.append({graph.numel_ubo(in_tensor)}); in add_tensor_to_staging_node() 106 {{out_staging, vkapi::kWrite}, {in_tensor, vkapi::kRead}}, in add_tensor_to_staging_node() [all …]
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | collective_ops_multi_worker_test.py | 177 in_tensor = constant_op.constant([1.]) 183 in_tensor, 201 in_tensor, 211 in_tensor, 223 in_tensor, 238 in_tensor = constant_op.constant([1.]) 253 collective_ops.all_reduce(in_tensor, group_size, group_key, 259 collective_ops.all_reduce(in_tensor, group_size, group_key, 267 collective_ops.all_reduce(in_tensor, group_size, group_key, instance_key) 278 in_tensor = constant_op.constant([1.]) [all …]
|
D | collective_ops_test.py | 377 in_tensor = constant_op.constant(in_value) 386 in_tensor, 396 in_tensor, 411 in_tensor = constant_op.constant(in_value) 420 in_tensor, 433 in_tensor, 738 in_tensor = constant_op.constant([1.]) 757 in_tensor, 768 in_tensor, 784 in_tensor, [all …]
|
/external/tensorflow/tensorflow/compiler/xla/experimental/xla_sharding/ |
D | xla_sharding_test.py | 83 in_tensor = array_ops.ones([4, 5, 6], dtype=dtypes.float32) 85 self.assertAllEqual(in_tensor, result) 101 in_tensor = array_ops.ones([4, 5, 6], dtype=dtypes.float32) 103 self.assertAllEqual(in_tensor, result) 118 in_tensor = array_ops.ones([4, 5, 6], dtype=dtypes.float32) 121 self.assertAllEqual(in_tensor, result) 158 in_tensor = array_ops.ones([4, 5, 6], dtype=dtypes.float32) 161 self.assertAllEqual(in_tensor, result)
|
/external/executorch/backends/cadence/aot/ |
D | simplify_ops.py | 73 in_tensor = args[0].to_tensor() if isinstance(args[0], ProxyValue) else args[0] 76 dim = dim if dim >= 0 else dim + in_tensor.dim() 77 length = in_tensor.size(dim) 90 empty_shape = [x for x in in_tensor.shape if x != 0] 95 {"dtype": in_tensor.dtype},
|
D | replace_ops.py | 255 in_tensor = args[0].to_tensor() if isinstance(args[0], ProxyValue) else args[0] 256 in_shape = in_tensor.shape 282 in_tensor = args[0].to_tensor() if isinstance(args[0], ProxyValue) else args[0] 285 if in_tensor.dim() <= 1: 288 assert in_tensor.dim() == 2, "t_copy expects a tensor with <= 2 dimensions" 427 in_tensor = args[0].to_tensor() if isinstance(args[0], ProxyValue) else args[0] 428 old_dims = tuple(range(in_tensor.dim())) 495 in_tensor = args[0] 500 in_tensor.to_tensor().shape 501 if isinstance(in_tensor, ProxyValue) [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | cast_op.h | 35 typename TTypes<IN_OUT>::ConstFlat in_tensor, \ 39 in_tensor.unaryExpr(LSBZeroSetter<IN_OUT, OUT_TYPE>()) \ 42 out_tensor.device(d) = in_tensor.template cast<OUT_TYPE>(); \ 53 typename TTypes<IN_OUT>::ConstFlat in_tensor, \ 57 in_tensor.unaryExpr(LSBZeroSetter<IN_OUT, OUT_TYPE>()) \ 60 out_tensor.device(d) = in_tensor.template cast<OUT_TYPE>(); \ 80 typename TTypes<IN_OUT>::ConstFlat in_tensor, \ 82 out_tensor.device(d) = in_tensor.template cast<OUT_TYPE>(); \ 100 typename TTypes<IN_OUT>::ConstFlat in_tensor, \ 102 out_tensor.device(d) = in_tensor.template cast<OUT_TYPE>(); \
|
/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_transpose_op.cc | 65 Status MKLTransposeND(OpKernelContext* context, const Tensor& in_tensor, in MKLTransposeND() argument 72 memory::dims in_dims = TFShapeToMklDnnDims(in_tensor.shape()); in MKLTransposeND() 80 in.SetUsrMem(in_dims, in_strides, &in_tensor); in MKLTransposeND() 89 in.SetUsrMemDataHandle(&in_tensor, transpose_stream); in MKLTransposeND()
|
/external/ComputeLibrary/src/dynamic_fusion/sketch/utils/ |
D | DependencyGraph.h | 194 for(auto in_tensor : inputs) variable 198 link_input(op, in_tensor); 368 void link_input(OperatorId op, TensorId in_tensor) in link_input() argument 371 if(!tensor_exists(in_tensor)) in link_input() 373 insert_new_tensor(in_tensor); in link_input() 375 ARM_COMPUTE_ERROR_ON(are_connected(op, in_tensor)); // Prevent repetitive linking in link_input() 376 _adj_src_tensors[op].push_back(in_tensor); in link_input() 377 _adj_dst_ops[in_tensor].push_back(op); in link_input()
|
/external/tensorflow/tensorflow/lite/toco/python/ |
D | toco_from_protos_test.py | 33 def _run(self, sess, in_tensor, out_tensor, should_succeed): argument 52 input_array.name = TensorName(in_tensor) 53 input_array.shape.dims.extend(map(int, in_tensor.shape))
|
/external/executorch/backends/vulkan/tools/gpuinfo/include/ |
D | textures.h | 61 vTensor in_tensor = in tex_cacheline_concurr() local 79 in_tensor.image(), in tex_cacheline_concurr() 161 vTensor in_tensor = api::vTensor(api::context(), sizes_nchw, vkapi::kFloat); in tex_bandwidth() local 193 in_tensor.image(), in tex_bandwidth()
|
/external/tensorflow/tensorflow/python/ops/ |
D | collective_ops_gpu_test.py | 290 in_tensor = constant_op.constant(1.) 297 in_tensor, 318 in_tensor, 329 in_tensor,
|
/external/tensorflow/tensorflow/lite/python/metrics/ |
D | metrics_nonportable_test.py | 131 in_tensor = array_ops.placeholder( 133 math_ops.add(in_tensor, in_tensor, name='add') 455 in_tensor = array_ops.placeholder( 457 out_tensor = in_tensor + in_tensor 458 inputs = {'x': in_tensor}
|
/external/executorch/backends/vulkan/runtime/ |
D | VulkanBackend.cpp | 394 vTensorPtr in_tensor = graph->get_tensor(in_tensor_ref); in maybe_resize_input() local 397 et_tensor.dim() == in_tensor->sizes().size(), in maybe_resize_input() 399 static_cast<size_t>(in_tensor->sizes().size()), in maybe_resize_input() 405 if (in_tensor->sizes()[i] != et_tensor.sizes()[i]) { in maybe_resize_input() 416 in_tensor->numel() == et_tensor.numel(), in maybe_resize_input() 418 static_cast<size_t>(in_tensor->numel()), in maybe_resize_input()
|
/external/pytorch/test/distributed/algorithms/quantization/ |
D | test_quantization.py | 289 in_tensor = torch.ones([sum(in_splits), size], dtype=dtype) * rank 296 in_tensor = in_tensor.cuda(rank_to_GPU) 304 in_tensor,
|