Home
last modified time | relevance | path

Searched refs:CUDA (Results 1 – 25 of 104) sorted by relevance

12345

/external/eigen/doc/
DUsingNVCC.dox4 /** \page TopicCUDA Using Eigen in CUDA kernels
8 Staring from CUDA 5.0, the CUDA compiler, \c nvcc, is able to properly parse %Eigen's code (almost).
9 A few adaptations of the %Eigen's code already allows to use some parts of %Eigen in your own CUDA
10 To this end you need the devel branch of %Eigen, CUDA 5.0 or greater with GCC.
27 …- On 64bits system Eigen uses \c long \c int as the default type for indexes and sizes. On CUDA de…
28CUDA code compatible, this cannot be done automatically by %Eigen, and the user is thus required t…
/external/llvm/docs/
DCompileCudaWithLLVM.rst2 Compiling CUDA C/C++ with LLVM
11 This document contains the user guides and the internals of compiling CUDA
12 C/C++ with LLVM. It is aimed at both users who want to compile CUDA with LLVM
14 familiarity with CUDA. Information about CUDA programming can be found in the
15 `CUDA programming guide
18 How to Build LLVM with CUDA Support
21 CUDA support is still in development and works the best in the trunk version
52 How to Compile CUDA C/C++ with LLVM
55 We assume you have installed the CUDA driver and runtime. Consult the `NVIDIA
56 CUDA installation guide
[all …]
DNVPTXUsage.rst21 This document assumes a basic familiarity with CUDA and the PTX
22 assembly language. Information about the CUDA Driver API and the PTX assembly
23 language can be found in the `CUDA documentation
100 copy data to it by name with the CUDA Driver API.
117 generated PTX compatible with the CUDA Driver API.
119 Example: 32-bit PTX for CUDA Driver API: ``nvptx-nvidia-cuda``
121 Example: 64-bit PTX for CUDA Driver API: ``nvptx64-nvidia-cuda``
223 map in the following way to CUDA builtins:
226 CUDA Builtin PTX Special Register Intrinsic
252 instruction, equivalent to the ``__syncthreads()`` call in CUDA.
[all …]
/external/tensorflow/tensorflow/contrib/cmake/
DCMakeLists.txt38 # GPU, CUDA and cuDNN options
40 set(tensorflow_CUDA_VERSION "9.0" CACHE STRING "CUDA version to build against")
55 # Options for linking CUDA/CUDNN libraries
72 …option(tensorflow_CUDA_LIBRARY_PATH "Designate the default CUDA library paths" /usr/local/cuda/lib…
304 …# In some Linux distros, find_package(CUDA) seems to require CMAKE_LIBRARY_PATH to include cuda-li…
310 find_package(CUDA ${tensorflow_CUDA_VERSION} REQUIRED EXACT)
312 # Test compatibility of compiler on CUDA
322 message(FATAL_ERROR "Selected compiler (or version) is not supported for CUDA")
378 # Remove "." from CUDA version variable.
413 # NOTE(mrry): Update these flags when the version of CUDA or cuDNN used
DREADME.md54 - [NVidia CUDA Toolkit 8.0](https://developer.nvidia.com/cuda-downloads)
120 …In case TensorFlow fails to find the CUDA dll's during initialization, check your PATH environment…
121 It should contain the directory of the CUDA dlls and the directory of the CUDNN dll.
125 D:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v8.0\bin
210 GPU support. If GPU is enabled you need to install the CUDA 8.0 Toolkit and CUDNN 5.1.
/external/eigen/Eigen/
DCore17 // Handle NVCC/CUDA/SYCL
19 // Do not try asserts on CUDA and SYCL!
32 // All functions callable from CUDA code must be qualified with __device__
34 // Do not try to vectorize on CUDA and SYCL!
52 // When compiling CUDA device code with NVCC, pull in math functions from the
390 #include "src/Core/arch/CUDA/Half.h"
391 #include "src/Core/arch/CUDA/PacketMathHalf.h"
392 #include "src/Core/arch/CUDA/TypeCasting.h"
395 #include "src/Core/arch/CUDA/PacketMath.h"
396 #include "src/Core/arch/CUDA/MathFunctions.h"
[all …]
/external/llvm/lib/Target/NVPTX/
DNVPTXLowerKernelArgs.cpp200 if (TM && TM->getDrvInterface() == NVPTX::CUDA) { in runOnFunction()
224 else if (TM && TM->getDrvInterface() == NVPTX::CUDA) in runOnFunction()
DNVPTXSubtarget.cpp56 if (TM.getDrvInterface() == NVPTX::CUDA) in hasImageHandles()
DNVPTX.h66 CUDA enumerator
/external/clang/test/Driver/Inputs/CUDA_80/usr/local/cuda/
Dversion.txt1 CUDA Version 8.0.42
/external/clang/include/clang/Basic/
DDiagnosticDriverKinds.td25 def err_drv_cuda_bad_gpu_arch : Error<"Unsupported CUDA gpu architecture: %0">;
27 "cannot find CUDA installation. Provide its path via --cuda-path, or pass "
28 "-nocudainc to build without CUDA includes.">;
30 "GPU arch %1 requires CUDA version at least %3, but installation at %0 is %2. "
31 "Use --cuda-path to specify a different CUDA install, or pass "
DLangOptions.def184 LANGOPT(CUDA , 1, 0, "CUDA")
190 LANGOPT(CUDAIsDevice , 1, 0, "compiling for CUDA device")
191 LANGOPT(CUDAAllowVariadicFunctions, 1, 0, "allowing variadic functions in CUDA device code")
/external/tensorflow/tensorflow/docs_src/install/
Dinstall_windows.md33 * CUDA® Toolkit 9.0. For details, see
38 * The NVIDIA drivers associated with CUDA Toolkit 9.0.
42 other CUDA DLLs. Ensure that you add the directory where you installed
44 * GPU card with CUDA Compute Capability 3.0 or higher. See
182 <pre>[...\stream_executor\dso_loader.cc] Couldn't open CUDA library nvcuda.dll</pre>
Dinstall_sources.md130 * GPU card with CUDA Compute Capability 3.0 or higher. See
216 If you want to compile tensorflow and have XCode 7.3 and CUDA 7.5 installed,
217 note that Xcode 7.3 is not yet compatible with CUDA 7.5. To remedy this
220 * Upgrade to CUDA 8.0.
288 Do you wish to build TensorFlow with CUDA support? [y/N] <b>Y</b>
289 CUDA support will be enabled for TensorFlow
290 Do you want to use clang as CUDA compiler? [y/N]
291 nvcc will be used as CUDA compiler
292 Please specify the Cuda SDK version you want to use, e.g. 7.0. [Leave empty to default to CUDA 9.0]…
293 Please specify the location where CUDA 9.0 toolkit is installed. Refer to README.md for more detail…
[all …]
/external/clang/test/Frontend/
Dstdlang.c16 #if defined(CUDA)
/external/eigen/test/
DCMakeLists.txt352 # CUDA unit tests
353 option(EIGEN_TEST_CUDA "Enable CUDA support in unit tests" OFF)
354 option(EIGEN_TEST_CUDA_CLANG "Use clang instead of nvcc to compile the CUDA tests" OFF)
362 find_package(CUDA 5.0)
/external/clang/test/CodeGenCUDA/Inputs/
Ddevice-code-2.ll1 ; Simple bit of IR to mimic CUDA's libdevice.
Ddevice-code.ll1 ; Simple bit of IR to mimic CUDA's libdevice. We want to be
/external/tensorflow/tensorflow/contrib/makefile/
DREADME.md133 ## CUDA support for Tegra devices running Android (Nvidia Shield TV, etc)
148 #### Building all CUDA-enabled native binaries:
149CUDA-enabled versions of libtensorflow_inference.so and the benchmark binary. (libtensorflow_demo.…
158 #### Testing the CUDA-enabled benchmark via adb:
171 #### Building the CUDA-enabled TensorFlow AAR with Bazel:
184 #### Building the CUDA-enabled TensorFlow Android demo with Bazel:
215 #### Building the CUDA-enabled Android demo with gradle/Android Studio:
/external/vulkan-validation-layers/libs/glm/gtx/
Dfast_square_root.inl30 # ifdef __CUDACC__ // Wordaround for a CUDA compiler bug up to CUDA6
41 # ifdef __CUDACC__ // Wordaround for a CUDA compiler bug up to CUDA6
/external/tensorflow/tensorflow/contrib/mpi/
DREADME.md31 …fers this requires MPI libraries that are built with CUDA support (CUDA Aware). When using non-CUD…
/external/clang/include/clang/Frontend/
DLangStandards.def161 // CUDA
163 "NVIDIA CUDA(tm)",
/external/tensorflow/tensorflow/tools/ci_build/
DDockerfile.gpu25 # Configure the build for our CUDA configuration.
/external/tensorflow/tensorflow/docs_src/performance/
Dbenchmarks.md81 * **CUDA / cuDNN:** 8.0 / 5.1
171 * **CUDA / cuDNN:** 8.0 / 5.1
245 * **CUDA / cuDNN:** 8.0 / 5.1
330 * **CUDA / cuDNN:** 8.0 / 5.1
/external/llvm/test/CodeGen/NVPTX/
Dnounroll.ll6 ; Compiled from the following CUDA code:

12345