Home
last modified time | relevance | path

Searched refs:plan_cache (Results 1 – 3 of 3) sorted by relevance

/external/pytorch/aten/src/ATen/native/cuda/
DSpectralOps.cpp219 CuFFTParamsLRUCache& plan_cache = cufft_get_plan_cache(input.device().index()); in _exec_fft() local
220 std::unique_lock<std::mutex> guard(plan_cache.mutex, std::defer_lock); in _exec_fft()
239 if (use_caching && plan_cache.max_size() > 0) { in _exec_fft()
241 if (plan_cache.max_size() > 0) { // check again after acquiring the lock in _exec_fft()
242 config = &plan_cache.lookup(Params); in _exec_fft()
/external/pytorch/torch/csrc/jit/runtime/
Dgraph_executor.cpp649 for (auto& entry : plan_cache) { in getDebugState()
675 auto it = plan_cache.find(spec); in getOrCompile()
676 if (it != plan_cache.end()) { in getOrCompile()
682 auto r = plan_cache.emplace(std::move(spec), std::move(plan)); in getOrCompile()
792 std::unordered_map<ArgumentSpec, ExecutionPlan> plan_cache; member
/external/pytorch/test/
Dtest_spectral_ops.py833 plan_cache = torch.backends.cuda.cufft_plan_cache
835 plan_cache = torch.backends.cuda.cufft_plan_cache[device]
836 original = plan_cache.max_size
837 plan_cache.max_size = n
841 plan_cache.max_size = original