/external/pytorch/torch/testing/_internal/distributed/rpc/ |
D | dist_autograd_test.py | 124 return rpc.rpc_sync(dst, my_rref_add, args=(rref_t1, t2)) 134 return rpc.rpc_sync( 140 return rpc.rpc_sync(worker_name(next_dst), my_py_add, args=(t1, t2)) 168 ret = rpc.rpc_sync(ps, my_rref_add, args=(rref_t1, t2)) 175 rpc.rpc_sync(ps, _set_rpc_done, args=(context_id, rank_diff)) 176 rpc.rpc_sync(ps, _check_rpc_done, args=(0,)) 182 ret = rpc.rpc_sync(ps, my_script_ref_add, args=(rref_t1, t2)) 189 rpc.rpc_sync(ps, _set_rpc_done, args=(context_id, rank_diff)) 190 rpc.rpc_sync(ps, _check_rpc_done, args=(0,)) 224 return rpc.rpc_sync(worker_name(dst), method, args=(args)) [all …]
|
D | rpc_test.py | 297 return rpc.rpc_sync(dst, torch.add, args=(torch.ones(2, 2), 1)) 301 return rpc.rpc_sync( 738 return rpc.rpc_sync(to, fn, args=args, kwargs=kwargs) 752 ret = rpc.rpc_sync(dst, add_rref_to_value, args=(rref, x + y)) 776 result = rpc.rpc_sync( 817 ret = rpc.rpc_sync( 900 ret = rpc.rpc_sync( 1001 c = rpc.rpc_sync( 1069 ps_gradient = rref.rpc_sync().get_gradient(rref) 1154 ret = rpc.rpc_sync(self_worker_info, torch.add, args=(torch.ones(2, 2), 1)) [all …]
|
D | faulty_agent_rpc_test.py | 256 rpc.rpc_sync( 298 rpc.rpc_sync(dst_worker, my_script_func, args=(torch.tensor(1),), timeout=1)
|
/external/pytorch/benchmarks/distributed/rpc/rl/ |
D | coordinator.py | 45 ob_ref.rpc_sync().set_state(state_size, batch) 47 self.agent_rref.rpc_sync().set_world( 84 agent_latency, agent_throughput = self.agent_rref.rpc_sync().finish_episode( 88 self.agent_rref.rpc_sync().reset_metrics()
|
D | observer.py | 8 from torch.distributed.rpc import rpc_sync 65 action = rpc_sync(
|
/external/pytorch/torch/testing/_internal/distributed/nn/api/ |
D | remote_module_test.py | 273 …ret1 = rpc.rpc_sync(dst_worker_name, get_remote_training_arg, args=(remote_module.get_module_rref(… 277 …ret2 = rpc.rpc_sync(dst_worker_name, get_remote_training_arg, args=(remote_module.get_module_rref(… 439 attrs = rpc.rpc_sync( 497 attrs = rpc.rpc_sync( 510 ret1 = rpc.rpc_sync(dst_worker2_name, remote_forward, (remote_module, args)) 512 ret2 = rpc.rpc_sync( 538 attrs = rpc.rpc_sync( 553 remote_module2 = rpc.rpc_sync( 560 ret1 = rpc.rpc_sync( 563 ret2 = rpc.rpc_sync( [all …]
|
/external/pytorch/torch/testing/_internal/distributed/rpc/jit/ |
D | rpc_test.py | 87 ret = ret = rpc.rpc_sync(dst_worker_name, rref_local_value, (rref,)) 111 ret = rpc.rpc_sync( 130 rpc.rpc_sync(dst, rref_list_mutate, args=(list_rref,)) 262 ret = rpc.rpc_sync(worker_name(dst_rank), rref_to_here, args=(rref,)) 266 rref1 = rpc.rpc_sync(worker_name(dst_rank), return_rref, args=(rref,)) 432 rref = rpc.rpc_sync( 462 rref = rpc.rpc_sync( 531 res = rpc.rpc_sync(dst_worker_name, two_args_two_kwargs, args, kwargs) 882 ret = rpc.rpc_sync(dst_worker_name, one_arg, args=(torch.ones(2, 2),)) 896 ret = rpc.rpc_sync(dst_worker_name, one_arg, args=(10, 20)) [all …]
|
D | dist_autograd_test.py | 71 loss = rpc.rpc_sync(dst_worker_name, torch.add, args=(t1, t2)).sum()
|
D | rpc_test_faulty.py | 218 rpc.rpc_sync(dst_worker, rref_to_here, args=(rref, ))
|
/external/pytorch/torch/testing/_internal/ |
D | dist_utils.py | 106 rpc.rpc_sync(f"worker{rank}", noop, args=()) 159 num_owners_on_rank, num_forks_on_rank = rpc.rpc_sync(
|
/external/pytorch/torch/testing/_internal/distributed/rpc/examples/ |
D | parameter_server_test.py | 88 m = self.ps_rref.rpc_sync().get_model() 93 m = rpc.rpc_sync(
|
D | reinforcement_learning_rpc_test.py | 15 from torch.distributed.rpc import RRef, rpc_sync, rpc_async, remote 38 return rpc_sync(rref.owner(), _call_method, args=args, kwargs=kwargs)
|
/external/pytorch/torch/testing/_internal/distributed/ |
D | ddp_under_dist_autograd_test.py | 84 return rpc.rpc_sync(rref.owner(), _call_method, args=args_tup, kwargs=kwargs) 461 rpc.rpc_sync(trainer, set_shutdown_signal, args=()) 463 rpc.rpc_sync(self.remote_worker_name(), set_shutdown_signal, args=()) 653 rpc.rpc_sync( 718 rpc.rpc_sync( 727 rpc.rpc_sync(
|
/external/pytorch/torch/distributed/nn/api/ |
D | remote_module.py | 265 self.module_rref = rpc.rpc_sync( 302 return rpc.rpc_sync(self.on, _param_rrefs, args=(self.module_rref, recurse)) 441 return self.module_rref.rpc_sync().train() # type: ignore[operator, union-attr] 444 return self.module_rref.rpc_sync().eval() # type: ignore[operator, union-attr]
|
/external/pytorch/torch/distributed/rpc/ |
D | api.py | 239 rpc_sync( 386 rpc_sync( 758 def rpc_sync(to, func, args=None, kwargs=None, timeout: float = UNSET_RPC_TIMEOUT): function
|
D | backend_registry.py | 303 device_count, device_map, devices = api.rpc_sync( 333 api.rpc_sync(
|
/external/pytorch/torch/distributed/benchmarks/ |
D | benchmark_ddp_rpc.py | 67 emb_rref.rpc_sync().forward( 142 rpc.rpc_sync(ps_name, _retrieve_embedding_parameters, args=(emb_rref,))
|
/external/pytorch/benchmarks/distributed/rpc/parameter_server/ |
D | launcher.py | 210 server_rref.rpc_sync().reset_state(server_rref) 233 metrics = server_rref.rpc_sync().get_metrics(server_rref)
|
/external/pytorch/docs/source/ |
D | rpc.rst | 37 :meth:`~torch.distributed.rpc.rpc_sync` (synchronous), 49 :meth:`~torch.distributed.rpc.rpc_sync` and 111 .. autofunction:: rpc_sync 195 :meth:`~torch.distributed.rpc.rpc_sync`,
|
/external/pytorch/torch/csrc/jit/runtime/ |
D | operator.cpp | 217 prim::rpc_sync, prim::rpc_remote}; in printerHasSpecialCaseFor() 334 prim::rpc_sync, in aliasAnalysisHasSpecialCaseFor()
|
D | register_distributed_ops.cpp | 246 prim::rpc_sync,
|
/external/pytorch/docs/source/rpc/ |
D | distributed_autograd.rst | 32 t3 = rpc.rpc_sync("worker1", my_add, args=(t1, t2)) 171 d = rpc.rpc_sync("worker1", torch.add, args=(a, b)) 172 e = rpc.rpc_sync("worker1", torch.mul, args=(b, c)) 258 t3 = rpc.rpc_sync("worker1", my_add, args=(t1, t2))
|
D | rref.rst | 31 :meth:`~torch.distributed.rpc.rpc_sync`, 50 :meth:`~torch.distributed.rpc.rpc_sync`, 80 owner calls :meth:`~torch.distributed.rpc.rpc_sync`,
|
/external/pytorch/torch/_C/ |
D | _distributed_rpc.pyi | 59 def rpc_sync(self, timeout: float = ...) -> Any: ...
|
/external/pytorch/aten/src/ATen/core/ |
D | interned_strings.h | 184 _(prim, rpc_sync) \
|