/external/eigen/unsupported/test/ |
D | cxx11_tensor_argmax_cuda.cu | 37 double* d_in; in test_cuda_simple_argmax() local 40 cudaMalloc((void**)(&d_in), in_bytes); in test_cuda_simple_argmax() 44 cudaMemcpy(d_in, in.data(), in_bytes, cudaMemcpyHostToDevice); in test_cuda_simple_argmax() 49 …Eigen::TensorMap<Eigen::Tensor<double, 3, Layout>, Aligned > gpu_in(d_in, Eigen::array<DenseIndex,… in test_cuda_simple_argmax() 63 cudaFree(d_in); in test_cuda_simple_argmax() 101 float* d_in; in test_cuda_argmax_dim() local 103 cudaMalloc((void**)(&d_in), in_bytes); in test_cuda_argmax_dim() 106 cudaMemcpy(d_in, tensor.data(), in_bytes, cudaMemcpyHostToDevice); in test_cuda_argmax_dim() 111 …Eigen::TensorMap<Eigen::Tensor<float, 4, DataLayout>, Aligned > gpu_in(d_in, Eigen::array<DenseInd… in test_cuda_argmax_dim() 140 cudaMemcpy(d_in, tensor.data(), in_bytes, cudaMemcpyHostToDevice); in test_cuda_argmax_dim() [all …]
|
D | cxx11_tensor_cuda.cu | 632 Scalar* d_in; in test_cuda_lgamma() local 634 cudaMalloc((void**)(&d_in), bytes); in test_cuda_lgamma() 637 cudaMemcpy(d_in, in.data(), bytes, cudaMemcpyHostToDevice); in test_cuda_lgamma() 642 Eigen::TensorMap<Eigen::Tensor<Scalar, 2> > gpu_in(d_in, 72, 97); in test_cuda_lgamma() 656 cudaFree(d_in); in test_cuda_lgamma() 686 Scalar* d_in; in test_cuda_digamma() local 688 cudaMalloc((void**)(&d_in), bytes); in test_cuda_digamma() 691 cudaMemcpy(d_in, in.data(), bytes, cudaMemcpyHostToDevice); in test_cuda_digamma() 696 Eigen::TensorMap<Eigen::Tensor<Scalar, 1> > gpu_in(d_in, 7); in test_cuda_digamma() 711 cudaFree(d_in); in test_cuda_digamma() [all …]
|
/external/tensorflow/tensorflow/compiler/tests/ |
D | conv3d_test.py | 108 d_in = d > 0 and d < y_shape[1] - 1 111 if d_in + h_in + w_in == 3: 113 elif d_in + h_in + w_in == 2: 115 elif d_in or h_in or w_in: 146 d_in = d % strides[1] == 0 and 0 < d < y_shape[1] - 1 149 if d_in + h_in + w_in == 3: 151 elif d_in + h_in + w_in == 2: 153 elif d_in or h_in or w_in: 189 d_in = d % strides[1] == 0 and pad < d < y_shape[1] - 1 - pad 192 if d_in + h_in + w_in == 3: [all …]
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | conv3d_transpose_test.py | 72 d_in = d > 0 and d < y_shape[1] - 1 75 if d_in + h_in + w_in == 3: 77 elif d_in + h_in + w_in == 2: 79 elif d_in or h_in or w_in: 110 d_in = d % strides[1] == 0 and 0 < d < y_shape[1] - 1 113 if d_in + h_in + w_in == 3: 115 elif d_in + h_in + w_in == 2: 117 elif d_in or h_in or w_in: 183 d_in = d % strides[1] == 0 and pad < d < y_shape[1] - 1 - pad 186 if d_in + h_in + w_in == 3: [all …]
|
/external/autotest/client/common_lib/ |
D | smogcheck_ina219.py | 94 def _validateRangeDict(self, d_ref, d_in): argument 108 if k not in d_in: 109 raise InaError('Key %s not present in dict %r' % (k, d_in)) 110 if type(v) != type(d_in[k]): 113 % (k, type(v), type(d_in[k]))) 115 self._validateRangeDict(v, d_in[k])
|
/external/eigen/test/ |
D | cuda_common.h | 36 typename Input::Scalar* d_in; in run_on_cuda() local 41 cudaMalloc((void**)(&d_in), in_bytes); in run_on_cuda() 44 cudaMemcpy(d_in, in.data(), in_bytes, cudaMemcpyHostToDevice); in run_on_cuda() 53 run_on_cuda_meta_kernel<<<Grids,Blocks>>>(ker, n, d_in, d_out); in run_on_cuda() 57 …cudaMemcpy(const_cast<typename Input::Scalar*>(in.data()), d_in, in_bytes, cudaMemcpyDeviceToHo… in run_on_cuda() 60 cudaFree(d_in); in run_on_cuda()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | where_op_gpu.cu.h | 73 const T* d_in, TIndex* d_out, int num_items, in operator() 78 d_in, is_nonzero); in operator() 88 const bool* d_in, TIndex* d_out, int num_items, 91 return cub::DeviceReduce::Sum(d_temp_storage, temp_storage_bytes, d_in,
|
/external/tensorflow/tensorflow/contrib/distributions/python/ops/bijectors/ |
D | masked_autoregressive.py | 325 d_in = n_in // num_blocks 330 col_slice = slice(col, col + d_in) 332 col += d_in
|
/external/freetype/src/base/ |
D | ftcalc.c | 1049 FT_Pos d_in, d_out, d_hypot; in ft_corner_is_flat() local 1071 d_in = FT_HYPOT( in_x, in_y ); in ft_corner_is_flat() 1079 return ( d_in + d_out - d_hypot ) < ( d_hypot >> 4 ); in ft_corner_is_flat()
|