Home
last modified time | relevance | path

Searched refs:ParallelStyle (Results 1 – 7 of 7) sorted by relevance

/external/pytorch/torch/distributed/tensor/parallel/
Dapi.py14 from torch.distributed.tensor.parallel.style import ParallelStyle
25 parallelize_plan: Union[ParallelStyle, Dict[str, ParallelStyle]], argument
84 if isinstance(parallelize_plan, ParallelStyle):
Dstyle.py22 class ParallelStyle(ABC): class
35 class ColwiseParallel(ParallelStyle):
145 class RowwiseParallel(ParallelStyle):
251 class SequenceParallel(ParallelStyle):
333 class PrepareModuleInput(ParallelStyle):
463 class PrepareModuleOutput(ParallelStyle):
D__init__.py6 ParallelStyle,
/external/pytorch/test/distributed/_tensor/experimental/
Dtest_tp_transform.py11 ParallelStyle,
104 parallel_strategies: Dict[str, ParallelStyle] = {
142 parallel_strategies: Dict[str, ParallelStyle] = {
/external/pytorch/docs/source/
Ddistributed.tensor.parallel.rst38 parameters to DTensors, the following ``ParallelStyle`` s can be used in
50 ``ParallelStyle`` s, we assume the input/output activation tensors are evenly sharded on
55 and use ``use_local_output=False`` to return DTensor after each ``ParallelStyle``, where
/external/pytorch/torch/distributed/tensor/experimental/
D_tp_transform.py17 from torch.distributed.tensor.parallel.style import ColwiseParallel, ParallelStyle
39 parallel_strategies: Dict[str, ParallelStyle], argument
82 parallel_strategies: Dict[str, ParallelStyle], argument
109 parallel_strategies: Dict[str, ParallelStyle], argument
D_attention.py29 from torch.distributed.tensor.parallel.style import ParallelStyle
626 class _AttentionContextParallel(ParallelStyle):