Home
last modified time | relevance | path

Searched refs:cache_enabled (Results 1 – 25 of 29) sorted by relevance

12

/external/pytorch/torch/cpu/amp/
Dautocast_mode.py26 cache_enabled: bool = True,
34 "cpu", enabled=enabled, dtype=dtype, cache_enabled=cache_enabled
/external/pytorch/torch/cuda/amp/
Dautocast_mode.py27 cache_enabled: bool = True,
35 "cuda", enabled=enabled, dtype=dtype, cache_enabled=cache_enabled
/external/tensorflow/tensorflow/core/distributed_runtime/rpc/
Dgrpc_worker_service.cc449 bool cache_enabled = (response_cache_ != nullptr && request_id != 0); in GrpcRecvTensorAsync() local
451 auto do_response = [response, done, cache_enabled](const Tensor& tensor, in GrpcRecvTensorAsync()
455 grpc::EncodeTensorToByteBuffer(is_dead, tensor, cache_enabled, response); in GrpcRecvTensorAsync()
464 if (cache_enabled && in GrpcRecvTensorAsync()
469 auto rendezvous_done = [this, request_id, do_response, cache_enabled]( in GrpcRecvTensorAsync()
472 if (cache_enabled) { in GrpcRecvTensorAsync()
595 bool cache_enabled = (response_cache_ != nullptr && request_id != 0); in RecvBufAsync() local
597 auto do_response = [this, response, done, cache_enabled]( in RecvBufAsync()
604 response->set_require_ack(cache_enabled); in RecvBufAsync()
612 if (cache_enabled && in RecvBufAsync()
[all …]
/external/compiler-rt/lib/sanitizer_common/
Dsanitizer_procmaps_common.cc67 MemoryMappingLayout::MemoryMappingLayout(bool cache_enabled) { in MemoryMappingLayout() argument
69 if (cache_enabled) { in MemoryMappingLayout()
79 if (cache_enabled) in MemoryMappingLayout()
Dsanitizer_procmaps.h36 explicit MemoryMappingLayout(bool cache_enabled);
Dsanitizer_procmaps_mac.cc24 MemoryMappingLayout::MemoryMappingLayout(bool cache_enabled) { in MemoryMappingLayout() argument
/external/python/mako/mako/
Dtemplate.py237 cache_enabled=True, argument
334 cache_enabled,
351 cache_enabled, argument
358 self.cache_enabled = cache_enabled
514 cache_enabled=True, argument
546 cache_enabled,
Dlookup.py164 cache_enabled=True, argument
210 "cache_enabled": cache_enabled,
Dcache.py89 if not self.template.cache_enabled:
/external/pytorch/torch/_higher_order_ops/
Dwrap.py70 cache_enabled: Optional[bool],
82 with torch.autocast(device_type, dtype, enabled, cache_enabled):
/external/pytorch/torch/amp/
Dautocast_mode.py223 cache_enabled: Optional[bool] = None,
272 if cache_enabled is not None:
273 self._cache_enabled = cache_enabled
/external/pytorch/aten/src/ATen/
Dautocast_mode.cpp83 thread_local bool cache_enabled = true; variable
109 return cache_enabled; in is_autocast_cache_enabled()
113 cache_enabled = enabled; in set_autocast_cache_enabled()
126 arg.is_leaf() && !arg.is_view() && cache_enabled && in cached_cast()
/external/executorch/extension/llm/modules/
Dattention.py161 self.cache_enabled = False
189 self.cache_enabled = True
/external/pytorch/test/
Dtest_jit_autocast.py800 with torch.cpu.amp.autocast(cache_enabled=False), torch.no_grad():
809 with torch.cpu.amp.autocast(cache_enabled=False), torch.no_grad():
824 with torch.cpu.amp.autocast(cache_enabled=False), torch.no_grad():
848 … with torch.cpu.amp.autocast(cache_enabled=False, dtype=torch.bfloat16), torch.no_grad():
Dtest_cuda.py2628 self, with_amp, cache_enabled, allow_unused_input argument
2700 device_type="cuda", enabled=with_amp, cache_enabled=cache_enabled
2742 device_type="cuda", enabled=with_amp, cache_enabled=cache_enabled
2779 self, with_amp, cache_enabled, allow_unused_input argument
2813 device_type="cuda", enabled=with_amp, cache_enabled=cache_enabled
2831 device_type="cuda", enabled=with_amp, cache_enabled=cache_enabled
Dtest_ops.py2409 cache_enabled = unittest.mock.patch(
2412 cache_enabled.start()
2413 self.addCleanup(cache_enabled.stop)
Dtest_jit_llga_fuser.py71 with torch.cpu.amp.autocast(cache_enabled=False, dtype=torch.bfloat16):
/external/mesa3d/src/broadcom/vulkan/
Dv3dv_pipeline_cache.c205 bool cache_enabled) in v3dv_pipeline_cache_init() argument
210 if (cache_enabled) { in v3dv_pipeline_cache_init()
Dv3dv_private.h2483 bool cache_enabled);
/external/python/mako/test/
Dtest_cache.py132 cache_enabled=False,
/external/python/mako/doc/build/
Dcaching.rst87 * ``cache_enabled`` - Setting this
96 cache_enabled = False
Dchangelog.rst1742 added "cache_enabled=True" flag to Template,
1746 cache_enabled configuration setting.
/external/pytorch/torch/_subclasses/
Dfake_tensor.py1139 self.cache_enabled = (
1691 if self.cache_enabled:
/external/bcc/libbpf-tools/x86/
Dvmlinux.h114836 bool (*cache_enabled)(struct mmc_host *); member
Dvmlinux_518.h114836 bool (*cache_enabled)(struct mmc_host *); member

12