/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorDeviceCuda.h | 62 if (status != cudaSuccess) { in initializeDeviceProp() 66 assert(status == cudaSuccess); in initializeDeviceProp() 71 if (status != cudaSuccess) { in initializeDeviceProp() 77 assert(status == cudaSuccess); in initializeDeviceProp() 122 assert(err == cudaSuccess); in initializeDeviceProp() 142 assert(err == cudaSuccess); in initializeDeviceProp() 145 assert(err == cudaSuccess); in initializeDeviceProp() 152 assert(err == cudaSuccess); in initializeDeviceProp() 155 assert(err == cudaSuccess); in initializeDeviceProp() 171 assert(err == cudaSuccess); in initializeDeviceProp() [all …]
|
/external/eigen/unsupported/test/ |
D | cxx11_tensor_device.cu | 71 assert(cudaMalloc((void**)(&kernel_1d_), 2*sizeof(float)) == cudaSuccess); in GPUContext() 73 …ert(cudaMemcpy(kernel_1d_, kernel_1d_val, 2*sizeof(float), cudaMemcpyHostToDevice) == cudaSuccess); in GPUContext() 75 assert(cudaMalloc((void**)(&kernel_2d_), 4*sizeof(float)) == cudaSuccess); in GPUContext() 77 …ert(cudaMemcpy(kernel_2d_, kernel_2d_val, 4*sizeof(float), cudaMemcpyHostToDevice) == cudaSuccess); in GPUContext() 79 assert(cudaMalloc((void**)(&kernel_3d_), 8*sizeof(float)) == cudaSuccess); in GPUContext() 81 …ert(cudaMemcpy(kernel_3d_, kernel_3d_val, 8*sizeof(float), cudaMemcpyHostToDevice) == cudaSuccess); in GPUContext() 84 assert(cudaFree(kernel_1d_) == cudaSuccess); in ~GPUContext() 85 assert(cudaFree(kernel_2d_) == cudaSuccess); in ~GPUContext() 86 assert(cudaFree(kernel_3d_) == cudaSuccess); in ~GPUContext() 299 assert(cudaMemcpy(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost) == cudaSuccess); in test_gpu() [all …]
|
D | cxx11_tensor_argmax_cuda.cu | 56 …out_max.data(), d_out_max, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_simple_argmax() 57 …out_min.data(), d_out_min, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_simple_argmax() 58 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_simple_argmax() 116 …(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_argmax_dim() 117 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_argmax_dim() 144 …(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_argmax_dim() 145 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_argmax_dim() 205 …(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_argmin_dim() 206 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_argmin_dim() 233 …(tensor_arg.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_argmin_dim() [all …]
|
D | cxx11_tensor_cuda.cu | 53 gpu_device.stream()) == cudaSuccess); in test_cuda_nullary() 55 gpu_device.stream()) == cudaSuccess); in test_cuda_nullary() 57 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_nullary() 102 gpu_device.stream()) == cudaSuccess); in test_cuda_elementwise_small() 103 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_elementwise_small() 154 …pyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_elementwise() 155 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_elementwise() 197 gpu_device.stream()) == cudaSuccess); in test_cuda_props() 198 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_props() 236 …pyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_reduction() [all …]
|
D | cxx11_tensor_random_cuda.cu | 40 …pyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_random_uniform() 41 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_random_uniform() 66 …pyAsync(out.data(), d_out, out_bytes, cudaMemcpyDeviceToHost, gpu_device.stream()) == cudaSuccess); in test_cuda_random_normal() 67 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_random_normal()
|
D | cxx11_tensor_complex_cuda.cu | 57 gpu_device.stream()) == cudaSuccess); in test_cuda_nullary() 59 gpu_device.stream()) == cudaSuccess); in test_cuda_nullary() 61 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_nullary()
|
D | cxx11_tensor_complex_cwise_ops_cuda.cu | 79 gpu_device.stream()) == cudaSuccess); in test_cuda_complex_cwise_ops() 80 assert(cudaStreamSynchronize(gpu_device.stream()) == cudaSuccess); in test_cuda_complex_cwise_ops()
|
/external/tensorflow/tensorflow/core/util/ |
D | gpu_kernel_helper_test.cu.cc | 32 EXPECT_EQ(cudaSuccess, err) << cudaGetErrorString(err); \ 39 ASSERT_EQ(cudaSuccess, err) << cudaGetErrorString(err); \ 168 ASSERT_EQ(cudaSuccess, err) << cudaGetErrorString(err); in SetUp() 316 ASSERT_EQ(cudaMallocManaged(&failure_count, sizeof(unsigned)), cudaSuccess); in TEST() 318 ASSERT_EQ(hipHostMalloc(&failure_count, sizeof(unsigned), 0), cudaSuccess); in TEST() 323 ASSERT_EQ(gpuDeviceSynchronize(), cudaSuccess); in TEST()
|
D | gpu_launch_config.h | 169 CHECK_EQ(err, cudaSuccess); in GetGpuLaunchConfig() 202 CHECK_EQ(err, cudaSuccess); in GetGpuLaunchConfigFixedBlockSize() 302 CHECK_EQ(err, cudaSuccess); in GetGpu3DLaunchConfig()
|
D | gpu_kernel_helper.h | 47 #define gpuSuccess cudaSuccess 112 if (result != cudaSuccess) { in GpuLaunchKernel()
|
D | gpu_device_functions.h | 65 #define cudaSuccess 0 macro 81 if (!SE_PREDICT_TRUE(error == cudaSuccess)) { \ 90 if (!SE_PREDICT_TRUE(error == cudaSuccess)) { \
|
/external/skqp/src/compute/common/cuda/ |
D | assert_cuda.c | 34 if (code != cudaSuccess) in assert_cuda()
|
/external/tensorflow/tensorflow/core/profiler/internal/gpu/ |
D | device_tracer_test.cc | 308 ASSERT_EQ(cudaSuccess, cudaMalloc(&devptr, size_in_bytes)); in TEST_F() 310 ASSERT_EQ(cudaSuccess, cudaMemset(devptr, test_value, size_in_bytes)); in TEST_F() 312 ASSERT_EQ(cudaSuccess, in TEST_F() 315 ASSERT_EQ(cudaSuccess, cudaFree(devptr)); in TEST_F()
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
D | gpu_managed_allocator.cc | 48 CHECK_EQ(cudaFree(ptr), cudaSuccess); in DeallocateRaw()
|
D | gpu_device.cc | 173 CHECK_EQ(err, cudaSuccess); in deallocate() 1202 if (err != cudaSuccess) { in CreateDevices() 1219 if (err != cudaSuccess) { in CreateDevices() 1225 if (err != cudaSuccess) { in CreateDevices() 1232 if (err != cudaSuccess) { in CreateDevices() 1272 if (err != cudaSuccess) { in CreateDevices()
|
/external/tensorflow/tensorflow/stream_executor/cuda/ |
D | cuda_driver_test.cc | 38 if (result == cudaSuccess) { in CheckCuda()
|
/external/tensorflow/tensorflow/compiler/tf2tensorrt/utils/ |
D | trt_int8_calibrator.cc | 73 if (status != cudaSuccess) { in setBatch()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | topk_op_gpu.h | 506 if (err != cudaSuccess) { 531 if (err != cudaSuccess) { 574 if (err != cudaSuccess) {
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | nccl_utils.h | 44 #define cudaSuccess hipSuccess macro
|
D | custom_call_test.cc | 35 #define gpuSuccess cudaSuccess
|
D | nccl_utils.cc | 92 if (s == cudaSuccess) { in ToStatus()
|
/external/tensorflow/tensorflow/core/grappler/clusters/ |
D | utils.cc | 85 if (error != cudaSuccess) { in GetLocalGPUInfo()
|
/external/tensorflow/tensorflow/core/nccl/ |
D | nccl_manager.cc | 49 #define cudaSuccess hipSuccess 65 if (cuda_status != cudaSuccess) { \
|
/external/tensorflow/tensorflow/compiler/tf2tensorrt/convert/ |
D | convert_graph.cc | 815 if (cuda_error != cudaSuccess) { in ConvertAfterShapes()
|
/external/tensorflow/tensorflow/compiler/tf2tensorrt/kernels/ |
D | trt_engine_op.cc | 1063 if (err != cudaSuccess) { in AllocateCalibrationResources()
|