Lines Matching full:destination
70 def reduce_add(inputs, destination=None): argument
78 destination (int, optional): a device on which the output will be
83 :attr:`destination` device.
85 destination = _get_device_index(destination, optional=True)
90 if inp.get_device() == destination:
100 "reduce_add expects destination to be on the same GPU with one of the tensors"
110 destination_device = torch.device(inputs[root_index].device.type, destination)
121 def reduce_add_coalesced(inputs, destination=None, buffer_size=10485760): argument
130 destination (int, optional): a device on which the output will be
136 inputs, placed on the ``destination`` device.
138 # TODO: When `len(inputs) == 1` and all inputs are on `destination`, just
146 result = reduce_add(tensor_at_gpus, destination) # this will be sparse too
159 flat_result = reduce_add(flat_tensors, destination)
218 def gather(tensors, dim=0, destination=None, *, out=None): argument
226 destination (torch.device, str, or int, optional): the output device.
234 :attr:`destination` must not be specified when :attr:`out` is specified.
237 - If :attr:`destination` is specified,
238 a tensor located on :attr:`destination` device, that is a result of
246 if destination == -1:
253 destination = _get_device_index(destination, allow_cpu=True, optional=True)
254 return torch._C._gather(tensors, dim, destination)
256 if destination is not None:
258 … f"'destination' must not be specified when 'out' is specified, but got destination={destination}"