/external/eigen/unsupported/test/ |
D | cxx11_tensor_cuda.cu | 81 float* d_out; in test_cuda_elementwise_small() local 84 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_elementwise_small() 97 d_out, Eigen::array<Eigen::DenseIndex, 1>(2)); in test_cuda_elementwise_small() 101 assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, in test_cuda_elementwise_small() 113 cudaFree(d_out); in test_cuda_elementwise_small() 134 float* d_out; in test_cuda_elementwise() local 138 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_elementwise() 150 …Eigen::TensorMap<Eigen::Tensor<float, 3> > gpu_out(d_out, Eigen::array<Eigen::DenseIndex, 3>(72,53… in test_cuda_elementwise() 154 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) … in test_cuda_elementwise() 168 cudaFree(d_out); in test_cuda_elementwise() [all …]
|
D | cxx11_tensor_random_cuda.cu | 30 float* d_out; in test_cuda_random_uniform() local 31 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_random_uniform() 36 Eigen::TensorMap<Eigen::Tensor<float, 2> > gpu_out(d_out, 72,97); in test_cuda_random_uniform() 40 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) … in test_cuda_random_uniform() 55 float* d_out; in test_cuda_random_normal() local 56 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_random_normal() 61 Eigen::TensorMap<Eigen::Tensor<float, 2> > gpu_out(d_out, 72,97); in test_cuda_random_normal() 66 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) … in test_cuda_random_normal()
|
D | cxx11_tensor_argmax_cuda.cu | 102 DenseIndex* d_out; in test_cuda_argmax_dim() local 104 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_argmax_dim() 112 Eigen::TensorMap<Eigen::Tensor<DenseIndex, 3, DataLayout>, Aligned > gpu_out(d_out, out_shape); in test_cuda_argmax_dim() 116 …assert(cudaMemcpyAsync(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.str… in test_cuda_argmax_dim() 144 …assert(cudaMemcpyAsync(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.str… in test_cuda_argmax_dim() 153 cudaFree(d_out); in test_cuda_argmax_dim() 191 DenseIndex* d_out; in test_cuda_argmin_dim() local 193 cudaMalloc((void**)(&d_out), out_bytes); in test_cuda_argmin_dim() 201 Eigen::TensorMap<Eigen::Tensor<DenseIndex, 3, DataLayout>, Aligned > gpu_out(d_out, out_shape); in test_cuda_argmin_dim() 205 …assert(cudaMemcpyAsync(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.str… in test_cuda_argmin_dim() [all …]
|
D | cxx11_tensor_complex_cwise_ops_cuda.cu | 29 std::complex<T>* d_out; in test_cuda_complex_cwise_ops() local 32 cudaMalloc((void**)(&d_out), complex_bytes); in test_cuda_complex_cwise_ops() 42 d_out, kNumItems); in test_cuda_complex_cwise_ops() 78 assert(cudaMemcpyAsync(actual.data(), d_out, complex_bytes, cudaMemcpyDeviceToHost, in test_cuda_complex_cwise_ops() 89 cudaFree(d_out); in test_cuda_complex_cwise_ops()
|
D | cxx11_tensor_device.cu | 285 float* d_out; in test_gpu() local 288 cudaMalloc((void**)(&d_out), out_bytes); in test_gpu() 295 Eigen::TensorMap<Eigen::Tensor<float, 3> > gpu_out(d_out, 40,50,70); in test_gpu() 299 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() 309 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() 319 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() 329 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() 344 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, context.device().stre… in test_gpu() 355 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, context.device().stre… in test_gpu() 369 …assert(cudaMemcpyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, context.device().stre… in test_gpu()
|
/external/eigen/test/ |
D | cuda_common.h | 37 typename Output::Scalar* d_out; in run_on_cuda() local 42 cudaMalloc((void**)(&d_out), out_bytes); in run_on_cuda() 45 cudaMemcpy(d_out, out.data(), out_bytes, cudaMemcpyHostToDevice); in run_on_cuda() 53 run_on_cuda_meta_kernel<<<Grids,Blocks>>>(ker, n, d_in, d_out); in run_on_cuda() 58 cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost); in run_on_cuda() 61 cudaFree(d_out); in run_on_cuda()
|
/external/freetype/src/base/ |
D | ftcalc.c | 969 FT_Pos d_in, d_out, d_hypot; in ft_corner_is_flat() local 992 d_out = FT_HYPOT( out_x, out_y ); in ft_corner_is_flat() 999 return ( d_in + d_out - d_hypot ) < ( d_hypot >> 4 ); in ft_corner_is_flat()
|
/external/pdfium/third_party/freetype/src/base/ |
D | ftcalc.c | 969 FT_Pos d_in, d_out, d_hypot; in ft_corner_is_flat() local 992 d_out = FT_HYPOT( out_x, out_y ); in ft_corner_is_flat() 999 return ( d_in + d_out - d_hypot ) < ( d_hypot >> 4 ); in ft_corner_is_flat()
|