/external/eigen/unsupported/test/ |
D | cxx11_tensor_cuda.cu | 81 float* d_out; in test_cuda_elementwise_small() local 84 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_elementwise_small() 97 d_out, Eigen::array<Eigen::DenseIndex, 1>(2)); in test_cuda_elementwise_small() 101 assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, in test_cuda_elementwise_small() 113 cudaFree(d_out); in test_cuda_elementwise_small() 134 float* d_out; in test_cuda_elementwise() local 138 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_elementwise() 150 …Eigen::TensorMap<Eigen::Tensor<float, 3> > gpu_out(d_out, Eigen::array<Eigen::DenseIndex, 3>(72,53… in test_cuda_elementwise() 154 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) … in test_cuda_elementwise() 168 cudaFree(d_out); in test_cuda_elementwise() [all …]
|
D | cxx11_tensor_random_cuda.cu | 30 float* d_out; in test_cuda_random_uniform() local 31 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_random_uniform() 36 Eigen::TensorMap<Eigen::Tensor<float, 2> > gpu_out(d_out, 72,97); in test_cuda_random_uniform() 40 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) … in test_cuda_random_uniform() 55 float* d_out; in test_cuda_random_normal() local 56 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_random_normal() 61 Eigen::TensorMap<Eigen::Tensor<float, 2> > gpu_out(d_out, 72,97); in test_cuda_random_normal() 66 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) … in test_cuda_random_normal()
|
D | cxx11_tensor_argmax_cuda.cu | 102 DenseIndex* d_out; in test_cuda_argmax_dim() local 104 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_argmax_dim() 112 Eigen::TensorMap<Eigen::Tensor<DenseIndex, 3, DataLayout>, Aligned > gpu_out(d_out, out_shape); in test_cuda_argmax_dim() 116 …assert(cudaMemcpyAsync(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.str… in test_cuda_argmax_dim() 144 …assert(cudaMemcpyAsync(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.str… in test_cuda_argmax_dim() 153 cudaFree(d_out); in test_cuda_argmax_dim() 191 DenseIndex* d_out; in test_cuda_argmin_dim() local 193 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_argmin_dim() 201 Eigen::TensorMap<Eigen::Tensor<DenseIndex, 3, DataLayout>, Aligned > gpu_out(d_out, out_shape); in test_cuda_argmin_dim() 205 …assert(cudaMemcpyAsync(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.str… in test_cuda_argmin_dim() [all …]
|
D | cxx11_tensor_complex_cwise_ops_cuda.cu | 29 std::complex<T>* d_out; in test_cuda_complex_cwise_ops() local 32 cudaMalloc((void**)(&d_out), complex_bytes); in test_cuda_complex_cwise_ops() 42 d_out, kNumItems); in test_cuda_complex_cwise_ops() 78 assert(cudaMemcpyAsync(actual.data(), d_out, complex_bytes, cudaMemcpyDeviceToHost, in test_cuda_complex_cwise_ops() 89 cudaFree(d_out); in test_cuda_complex_cwise_ops()
|
D | cxx11_tensor_device.cu | 285 float* d_out; in test_gpu() local 288 cudaMalloc((void**)(&d_out), out_bytes); in test_gpu() 295 Eigen::TensorMap<Eigen::Tensor<float, 3> > gpu_out(d_out, 40,50,70); in test_gpu() 299 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() 309 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() 319 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() 329 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() 344 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, context.device().stre… in test_gpu() 355 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, context.device().stre… in test_gpu() 369 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, context.device().stre… in test_gpu()
|
/external/eigen/test/ |
D | cuda_common.h | 37 typename Output::Scalar* d_out; in run_on_cuda() local 42 cudaMalloc((void**)(&d_out), out_bytes); in run_on_cuda() 45 cudaMemcpy(d_out, out.data(), out_bytes, cudaMemcpyHostToDevice); in run_on_cuda() 53 run_on_cuda_meta_kernel<<<Grids,Blocks>>>(ker, n, d_in, d_out); in run_on_cuda() 58 cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost); in run_on_cuda() 61 cudaFree(d_out); in run_on_cuda()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | where_op_gpu.cu.h | 70 const T* d_in, TIndex* d_out, int num_items, in operator() 77 is_nonzero_iter, d_out, num_items, stream, in operator() 85 const bool* d_in, TIndex* d_out, int num_items, 89 d_out, num_items, stream, 102 const T* d_flags, OutputIterator d_out, 112 is_nonzero_iter /*d_flags*/, d_out, d_num_selected_out, num_items, 121 const T* d_flags, OutputIterator d_out, 128 d_out, d_num_selected_out, num_items, stream, debug_synchronous);
|
/external/OpenCL-CTS/test_conformance/printf/ |
D | test_printf.cpp | 449 cl_mem d_out = NULL; in doTest() local 487 d_out = clCreateBuffer(context, CL_MEM_READ_WRITE, in doTest() 489 if(err!= CL_SUCCESS || d_out == NULL) { in doTest() 493 err = clSetKernelArg(kernel, 1, sizeof(cl_mem), &d_out); in doTest() 543 clEnqueueReadBuffer(queue, d_out, CL_TRUE, 0, sizeof(cl_int),&out32, in doTest() 548 clEnqueueReadBuffer(queue, d_out, CL_TRUE, 0, sizeof(cl_ulong),&out64, in doTest() 572 if(d_out) in doTest() 573 clReleaseMemObject(d_out); in doTest()
|
/external/tensorflow/tensorflow/python/framework/ |
D | subscribe_test.py | 72 d_out = self.evaluate([d]) 76 self.assertEqual(d_out, [42]) 152 d_out = self.evaluate([d]) 155 self.assertEqual(d_out, [11])
|
/external/freetype/src/base/ |
D | ftcalc.c | 1054 FT_Pos d_in, d_out, d_hypot; in ft_corner_is_flat() local 1077 d_out = FT_HYPOT( out_x, out_y ); in ft_corner_is_flat() 1084 return ( d_in + d_out - d_hypot ) < ( d_hypot >> 4 ); in ft_corner_is_flat()
|