/external/perfetto/src/tracing/ |
D | traced_proto_unittest.cc | 320 EXPECT_EQ(result.nested_size(), 2); in TEST_F() 333 EXPECT_EQ(result.nested_size(), 2); in TEST_F() 349 EXPECT_EQ(result.nested_size(), 2); in TEST_F() 361 EXPECT_EQ(result.nested_size(), 1); in TEST_F() 372 EXPECT_EQ(result.nested_size(), 1); in TEST_F() 385 EXPECT_EQ(result.nested_size(), 2); in TEST_F() 397 EXPECT_EQ(result.nested_size(), 2); in TEST_F() 409 EXPECT_EQ(result.nested_size(), 2); in TEST_F() 425 EXPECT_EQ(result.nested_size(), 2); in TEST_F()
|
/external/pytorch/aten/src/ATen/native/nested/ |
D | NestedTensorFactories.cpp | 49 auto nested_size = self_nt->get_nested_sizes().clone(); in empty_like_nested() local 50 int64_t buffer_size = get_numel_from_nested_size_tensor(nested_size); in empty_like_nested() 52 auto tensor = wrap_buffer(new_buffer, nested_size); in empty_like_nested() 63 auto nested_size = self_nt->get_nested_sizes().clone(); in empty_like_nested() local 66 auto tensor = wrap_buffer(new_buffer, nested_size, nested_strides, offsets); in empty_like_nested()
|
D | NestedTensorMath.cpp | 903 std::tuple<Tensor, Tensor> _nested_compute_contiguous_strides_offsets(const Tensor& nested_size) { in _nested_compute_contiguous_strides_offsets() argument 905 construct_nested_strides(nested_size), in _nested_compute_contiguous_strides_offsets() 906 construct_offsets(nested_size)); in _nested_compute_contiguous_strides_offsets()
|
/external/pytorch/torch/csrc/autograd/ |
D | input_metadata.cpp | 12 auto nested_size = input._nested_tensor_size(); in compute_variant_shape() local 13 return MetadataShape{std::in_place_type<at::Tensor>, nested_size}; in compute_variant_shape()
|
/external/pytorch/aten/src/ATen/ |
D | NestedTensorImpl.h | 17 at::Tensor construct_nested_strides(const at::Tensor& nested_size); 18 at::Tensor construct_offsets(const at::Tensor& nested_size);
|
/external/linux-kselftest/tools/testing/selftests/kvm/lib/x86_64/ |
D | processor.c | 964 static int nested_size = -1; in vcpu_save_state() local 966 if (nested_size == -1) { in vcpu_save_state() 967 nested_size = kvm_check_cap(KVM_CAP_NESTED_STATE); in vcpu_save_state() 968 TEST_ASSERT(nested_size <= sizeof(state->nested_), in vcpu_save_state() 970 nested_size, sizeof(state->nested_)); in vcpu_save_state() 993 if (nested_size) { in vcpu_save_state() 997 TEST_ASSERT(state->nested.size <= nested_size, in vcpu_save_state() 999 state->nested.size, nested_size); in vcpu_save_state()
|
/external/pytorch/test/dynamo/ |
D | test_subclasses.py | 45 def get_jagged_tensor(nested_size, offsets, requires_grad=True): argument 48 D = nested_size[1] 50 for s in nested_size[0]: 1807 def _get_jagged_tensor(self, nested_size, offsets, requires_grad=True): argument 1808 return get_jagged_tensor(nested_size, offsets, requires_grad)
|
/external/rust/android-crates-io/crates/grpcio-sys/grpc/tools/codegen/core/ |
D | gen_huffman_decompressor.cc | 1104 size_t nested_size = nested->Size(); in Choose() local 1109 if (nested_size < best_size) { in Choose() 1111 best_size = nested_size; in Choose()
|
/external/grpc-grpc/tools/codegen/core/ |
D | gen_huffman_decompressor.cc | 1156 size_t nested_size = nested->Size(); in Choose() local 1161 if (nested_size < best_size) { in Choose() 1163 best_size = nested_size; in Choose()
|
/external/pytorch/test/ |
D | test_nestedtensor.py | 2929 nested_size = torch.tensor([[1, 2], [2, 2]]) 2934 def grad_test_func(tensor, nested_size): argument 2936 tensor, nested_size, fuse_transform_0213=False 2941 data = (padded_tensor, nested_size) 2945 nested_size = torch.tensor([[1, 8], [2, 8]]) 2950 def grad_test_func(tensor, nested_size): argument 2952 tensor, nested_size, fuse_transform_0213=True 2957 data = (padded_tensor, nested_size) 3523 def _get_list_for_jagged_tensor(self, nested_size, device, requires_grad=True): argument 3524 Ds = nested_size[1:] [all …]
|