Searched refs:plan_cache (Results 1 – 3 of 3) sorted by relevance
219 CuFFTParamsLRUCache& plan_cache = cufft_get_plan_cache(input.device().index()); in _exec_fft() local220 std::unique_lock<std::mutex> guard(plan_cache.mutex, std::defer_lock); in _exec_fft()239 if (use_caching && plan_cache.max_size() > 0) { in _exec_fft()241 if (plan_cache.max_size() > 0) { // check again after acquiring the lock in _exec_fft()242 config = &plan_cache.lookup(Params); in _exec_fft()
649 for (auto& entry : plan_cache) { in getDebugState()675 auto it = plan_cache.find(spec); in getOrCompile()676 if (it != plan_cache.end()) { in getOrCompile()682 auto r = plan_cache.emplace(std::move(spec), std::move(plan)); in getOrCompile()792 std::unordered_map<ArgumentSpec, ExecutionPlan> plan_cache; member
833 plan_cache = torch.backends.cuda.cufft_plan_cache835 plan_cache = torch.backends.cuda.cufft_plan_cache[device]836 original = plan_cache.max_size837 plan_cache.max_size = n841 plan_cache.max_size = original