/external/tensorflow/tensorflow/core/distributed_runtime/ |
D | collective_rma_distributed.cc | 43 const AllocatorAttributes& to_alloc_attr, Tensor* to_tensor, in RecvBufCall() argument 52 req_.set_num_bytes(to_tensor->TotalBytes()); in RecvBufCall() 53 req_.set_buf_ptr(reinterpret_cast<int64>(DMAHelper::base(to_tensor))); in RecvBufCall() 110 const AllocatorAttributes& to_alloc_attr, Tensor* to_tensor, in RecvFromPeer() argument 116 to_alloc_attr, to_tensor, client_locality, dev_to_dev_stream_index, in RecvFromPeer() 157 step_id_, "dynamic", to_tensor->dtype(), &to_tensor->shape()); in RecvFromPeer() 161 to_tensor->dtype(), to_tensor->shape()); in RecvFromPeer() 164 dst_tensor = to_tensor; in RecvFromPeer() 169 [this, state, to_device, to_alloc_attr, to_device_ctx, to_tensor, cpu_dev, in RecvFromPeer() 198 to_tensor, dev_to_dev_stream_index, in RecvFromPeer()
|
D | collective_rma_distributed.h | 43 const AllocatorAttributes& to_alloc_attr, Tensor* to_tensor,
|
/external/tensorflow/tensorflow/python/ops/ragged/ |
D | ragged_to_tensor_op_test.py | 97 dt = rt.to_tensor() 341 dt = rt.to_tensor(default_value=default, shape=shape) 400 self.evaluate(rt.to_tensor(default_value=default, shape=shape)) 405 rt_placeholder.to_tensor(default_value=default, shape=shape)) 409 actual = input_data.to_tensor( 416 actual = input_data.to_tensor( 424 actual = input_data.to_tensor( 431 actual = input_data.to_tensor( 438 actual = input_data.to_tensor(shape=[4, 4]) 460 actual = input_data.to_tensor(default_value=default_value) [all …]
|
D | ragged_squeeze_op_test.py | 54 self.assertAllEqual(ragged_conversion_ops.to_tensor(rt), dt) 114 self.assertAllEqual(ragged_conversion_ops.to_tensor(rt), dt) 169 self.assertAllEqual(ragged_conversion_ops.to_tensor(rt_s), dt_s)
|
D | ragged_conversion_ops.py | 48 def to_tensor(rt_input, default_value=None, name=None): function 50 return rt_input.to_tensor(default_value, name) 57 return rt_input.to_tensor(default_value=default_value, shape=shape)
|
D | ragged_string_ops.py | 768 to_tensor = False 771 to_tensor = True 792 dense_shape) if to_tensor else output 823 dense_shape) if to_tensor else output
|
/external/tensorflow/tensorflow/python/keras/layers/preprocessing/benchmarks/ |
D | category_cross_hash_dense_benchmark.py | 63 data_a.to_tensor(default_value="", shape=(batch_size, max_length)), 65 data_b.to_tensor(default_value="", shape=(batch_size, max_length)), 71 data_a.to_tensor(default_value="", shape=(batch_size, max_length)), 73 data_b.to_tensor(default_value="", shape=(batch_size, max_length)),
|
D | bucketized_column_dense_benchmark.py | 59 keras_data = {"data": data.to_tensor(default_value=0.0)} 62 fc_data = {"data": data.to_tensor(default_value=0.0)}
|
D | category_hash_dense_benchmark.py | 58 "data": data.to_tensor(default_value="", shape=(batch_size, max_length)) 63 "data": data.to_tensor(default_value="", shape=(batch_size, max_length))
|
D | embedding_dense_benchmark.py | 59 keras_data = {"data": data.to_tensor(default_value=0)} 62 fc_data = {"data": data.to_tensor(default_value=0)}
|
D | category_vocab_list_dense_benchmark.py | 56 "data": data.to_tensor(default_value="", shape=(batch_size, max_length)) 61 "data": data.to_tensor(default_value="", shape=(batch_size, max_length))
|
D | category_vocab_list_indicator_dense_benchmark.py | 62 "data": data.to_tensor(default_value="", shape=(batch_size, max_length)) 67 "data": data.to_tensor(default_value="", shape=(batch_size, max_length))
|
D | category_vocab_file_dense_benchmark.py | 74 "data": data.to_tensor( 80 "data": data.to_tensor(
|
/external/tensorflow/tensorflow/compiler/xla/experimental/xla_sharding/ |
D | xla_sharding.py | 234 def copy_sharding(from_tensor, to_tensor, use_sharding_op=False): argument 247 return to_tensor 250 to_tensor = tf2xla.sharding(to_tensor, sharding=sharding) 253 to_tensor.op._set_attr('_XlaSharding', attr_value) 254 return to_tensor
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | collective_rma_local.cc | 29 const AllocatorAttributes& to_alloc_attr, Tensor* to_tensor, in RecvFromPeer() argument 48 auto consumer_callback = [to_tensor, to_device_ctx, to_device, to_alloc_attr, in RecvFromPeer() 65 int64_t recv_bytes = to_tensor->TotalBytes(); in RecvFromPeer() 74 to_tensor, // dst Tensor* in RecvFromPeer()
|
D | collective_rma_local.h | 43 const AllocatorAttributes& to_alloc_attr, Tensor* to_tensor,
|
D | collective_test_util.h | 52 const AllocatorAttributes& to_alloc_attr, Tensor* to_tensor,
|
D | graph_execution_state.cc | 580 tensor_connection.to_tensor(), "\"."); in PruneGraph() 585 tensor_connection.to_tensor(), in PruneGraph() 590 &tensor_connection.to_tensor(), {from_node, from_id.second})); in PruneGraph() 695 feeds.emplace_back(ParseTensorName(tensor_connection.to_tensor())); in OptimizeGraph()
|
D | collective_test_util.cc | 59 const AllocatorAttributes& to_alloc_attr, Tensor* to_tensor, in RecvFromPeer() argument 65 to_alloc_attr, to_tensor, client_locality, dev_to_dev_stream_index, in RecvFromPeer()
|
/external/tflite-support/tensorflow_lite_support/custom_ops/python/ |
D | ragged_tensor_to_tensor_test.py | 33 return ragged_tensor.to_tensor()
|
/external/tensorflow/tensorflow/python/distribute/ |
D | input_lib_type_spec_test.py | 426 "dense": ragged_tensor.to_tensor(), 600 "dense": ragged_tensor.to_tensor(), 661 "dense": ragged_tensor.to_tensor(), 722 "dense": ragged_tensor.to_tensor(),
|
/external/tflite-support/tensorflow_lite_support/custom_ops/kernel/ |
D | ngrams_test.py | 112 input_tensor = tf.ragged.constant(test_case).to_tensor() 132 input_tensor = tf.ragged.constant(test_case).to_tensor()
|
D | whitespace_tokenizer_test.py | 54 return tokenizer.tokenize(test_case).to_tensor()
|
/external/tensorflow/tensorflow/python/keras/layers/ |
D | pooling_test.py | 70 dense_data = ragged_data.to_tensor() 90 dense_data = ragged_data.to_tensor()
|
/external/tensorflow/tensorflow/python/keras/ |
D | losses.py | 1253 rt.to_tensor() if isinstance(rt, ragged_tensor.RaggedTensor) else rt 1266 r = r.to_tensor() 1280 return loss_fn(y_true, y_pred.to_tensor())
|