Searched refs:CUDA (Results 1 – 25 of 104) sorted by relevance
12345
/external/eigen/doc/ |
D | UsingNVCC.dox | 4 /** \page TopicCUDA Using Eigen in CUDA kernels 8 Staring from CUDA 5.0, the CUDA compiler, \c nvcc, is able to properly parse %Eigen's code (almost). 9 A few adaptations of the %Eigen's code already allows to use some parts of %Eigen in your own CUDA … 10 To this end you need the devel branch of %Eigen, CUDA 5.0 or greater with GCC. 27 …- On 64bits system Eigen uses \c long \c int as the default type for indexes and sizes. On CUDA de… 28 …CUDA code compatible, this cannot be done automatically by %Eigen, and the user is thus required t…
|
/external/llvm/docs/ |
D | CompileCudaWithLLVM.rst | 2 Compiling CUDA C/C++ with LLVM 11 This document contains the user guides and the internals of compiling CUDA 12 C/C++ with LLVM. It is aimed at both users who want to compile CUDA with LLVM 14 familiarity with CUDA. Information about CUDA programming can be found in the 15 `CUDA programming guide 18 How to Build LLVM with CUDA Support 21 CUDA support is still in development and works the best in the trunk version 52 How to Compile CUDA C/C++ with LLVM 55 We assume you have installed the CUDA driver and runtime. Consult the `NVIDIA 56 CUDA installation guide [all …]
|
D | NVPTXUsage.rst | 21 This document assumes a basic familiarity with CUDA and the PTX 22 assembly language. Information about the CUDA Driver API and the PTX assembly 23 language can be found in the `CUDA documentation 100 copy data to it by name with the CUDA Driver API. 117 generated PTX compatible with the CUDA Driver API. 119 Example: 32-bit PTX for CUDA Driver API: ``nvptx-nvidia-cuda`` 121 Example: 64-bit PTX for CUDA Driver API: ``nvptx64-nvidia-cuda`` 223 map in the following way to CUDA builtins: 226 CUDA Builtin PTX Special Register Intrinsic 252 instruction, equivalent to the ``__syncthreads()`` call in CUDA. [all …]
|
/external/tensorflow/tensorflow/contrib/cmake/ |
D | CMakeLists.txt | 38 # GPU, CUDA and cuDNN options 40 set(tensorflow_CUDA_VERSION "9.0" CACHE STRING "CUDA version to build against") 55 # Options for linking CUDA/CUDNN libraries 72 …option(tensorflow_CUDA_LIBRARY_PATH "Designate the default CUDA library paths" /usr/local/cuda/lib… 304 …# In some Linux distros, find_package(CUDA) seems to require CMAKE_LIBRARY_PATH to include cuda-li… 310 find_package(CUDA ${tensorflow_CUDA_VERSION} REQUIRED EXACT) 312 # Test compatibility of compiler on CUDA 322 message(FATAL_ERROR "Selected compiler (or version) is not supported for CUDA") 378 # Remove "." from CUDA version variable. 413 # NOTE(mrry): Update these flags when the version of CUDA or cuDNN used
|
D | README.md | 54 - [NVidia CUDA Toolkit 8.0](https://developer.nvidia.com/cuda-downloads) 120 …In case TensorFlow fails to find the CUDA dll's during initialization, check your PATH environment… 121 It should contain the directory of the CUDA dlls and the directory of the CUDNN dll. 125 D:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v8.0\bin 210 GPU support. If GPU is enabled you need to install the CUDA 8.0 Toolkit and CUDNN 5.1.
|
/external/eigen/Eigen/ |
D | Core | 17 // Handle NVCC/CUDA/SYCL 19 // Do not try asserts on CUDA and SYCL! 32 // All functions callable from CUDA code must be qualified with __device__ 34 // Do not try to vectorize on CUDA and SYCL! 52 // When compiling CUDA device code with NVCC, pull in math functions from the 390 #include "src/Core/arch/CUDA/Half.h" 391 #include "src/Core/arch/CUDA/PacketMathHalf.h" 392 #include "src/Core/arch/CUDA/TypeCasting.h" 395 #include "src/Core/arch/CUDA/PacketMath.h" 396 #include "src/Core/arch/CUDA/MathFunctions.h" [all …]
|
/external/llvm/lib/Target/NVPTX/ |
D | NVPTXLowerKernelArgs.cpp | 200 if (TM && TM->getDrvInterface() == NVPTX::CUDA) { in runOnFunction() 224 else if (TM && TM->getDrvInterface() == NVPTX::CUDA) in runOnFunction()
|
D | NVPTXSubtarget.cpp | 56 if (TM.getDrvInterface() == NVPTX::CUDA) in hasImageHandles()
|
D | NVPTX.h | 66 CUDA enumerator
|
/external/clang/test/Driver/Inputs/CUDA_80/usr/local/cuda/ |
D | version.txt | 1 CUDA Version 8.0.42
|
/external/clang/include/clang/Basic/ |
D | DiagnosticDriverKinds.td | 25 def err_drv_cuda_bad_gpu_arch : Error<"Unsupported CUDA gpu architecture: %0">; 27 "cannot find CUDA installation. Provide its path via --cuda-path, or pass " 28 "-nocudainc to build without CUDA includes.">; 30 "GPU arch %1 requires CUDA version at least %3, but installation at %0 is %2. " 31 "Use --cuda-path to specify a different CUDA install, or pass "
|
D | LangOptions.def | 184 LANGOPT(CUDA , 1, 0, "CUDA") 190 LANGOPT(CUDAIsDevice , 1, 0, "compiling for CUDA device") 191 LANGOPT(CUDAAllowVariadicFunctions, 1, 0, "allowing variadic functions in CUDA device code")
|
/external/tensorflow/tensorflow/docs_src/install/ |
D | install_windows.md | 33 * CUDA® Toolkit 9.0. For details, see 38 * The NVIDIA drivers associated with CUDA Toolkit 9.0. 42 other CUDA DLLs. Ensure that you add the directory where you installed 44 * GPU card with CUDA Compute Capability 3.0 or higher. See 182 <pre>[...\stream_executor\dso_loader.cc] Couldn't open CUDA library nvcuda.dll</pre>
|
D | install_sources.md | 130 * GPU card with CUDA Compute Capability 3.0 or higher. See 216 If you want to compile tensorflow and have XCode 7.3 and CUDA 7.5 installed, 217 note that Xcode 7.3 is not yet compatible with CUDA 7.5. To remedy this 220 * Upgrade to CUDA 8.0. 288 Do you wish to build TensorFlow with CUDA support? [y/N] <b>Y</b> 289 CUDA support will be enabled for TensorFlow 290 Do you want to use clang as CUDA compiler? [y/N] 291 nvcc will be used as CUDA compiler 292 Please specify the Cuda SDK version you want to use, e.g. 7.0. [Leave empty to default to CUDA 9.0]… 293 Please specify the location where CUDA 9.0 toolkit is installed. Refer to README.md for more detail… [all …]
|
/external/clang/test/Frontend/ |
D | stdlang.c | 16 #if defined(CUDA)
|
/external/eigen/test/ |
D | CMakeLists.txt | 352 # CUDA unit tests 353 option(EIGEN_TEST_CUDA "Enable CUDA support in unit tests" OFF) 354 option(EIGEN_TEST_CUDA_CLANG "Use clang instead of nvcc to compile the CUDA tests" OFF) 362 find_package(CUDA 5.0)
|
/external/clang/test/CodeGenCUDA/Inputs/ |
D | device-code-2.ll | 1 ; Simple bit of IR to mimic CUDA's libdevice.
|
D | device-code.ll | 1 ; Simple bit of IR to mimic CUDA's libdevice. We want to be
|
/external/tensorflow/tensorflow/contrib/makefile/ |
D | README.md | 133 ## CUDA support for Tegra devices running Android (Nvidia Shield TV, etc) 148 #### Building all CUDA-enabled native binaries: 149 …CUDA-enabled versions of libtensorflow_inference.so and the benchmark binary. (libtensorflow_demo.… 158 #### Testing the CUDA-enabled benchmark via adb: 171 #### Building the CUDA-enabled TensorFlow AAR with Bazel: 184 #### Building the CUDA-enabled TensorFlow Android demo with Bazel: 215 #### Building the CUDA-enabled Android demo with gradle/Android Studio:
|
/external/vulkan-validation-layers/libs/glm/gtx/ |
D | fast_square_root.inl | 30 # ifdef __CUDACC__ // Wordaround for a CUDA compiler bug up to CUDA6 41 # ifdef __CUDACC__ // Wordaround for a CUDA compiler bug up to CUDA6
|
/external/tensorflow/tensorflow/contrib/mpi/ |
D | README.md | 31 …fers this requires MPI libraries that are built with CUDA support (CUDA Aware). When using non-CUD…
|
/external/clang/include/clang/Frontend/ |
D | LangStandards.def | 161 // CUDA 163 "NVIDIA CUDA(tm)",
|
/external/tensorflow/tensorflow/tools/ci_build/ |
D | Dockerfile.gpu | 25 # Configure the build for our CUDA configuration.
|
/external/tensorflow/tensorflow/docs_src/performance/ |
D | benchmarks.md | 81 * **CUDA / cuDNN:** 8.0 / 5.1 171 * **CUDA / cuDNN:** 8.0 / 5.1 245 * **CUDA / cuDNN:** 8.0 / 5.1 330 * **CUDA / cuDNN:** 8.0 / 5.1
|
/external/llvm/test/CodeGen/NVPTX/ |
D | nounroll.ll | 6 ; Compiled from the following CUDA code:
|
12345