Home
last modified time | relevance | path

Searched refs:all_arguments (Results 1 – 3 of 3) sorted by relevance

/external/tensorflow/tensorflow/compiler/xla/service/
Dservice.cc601 std::vector<std::vector<std::vector<const ShapedBuffer*>>> all_arguments; in ExecuteGraphParallel() local
669 all_arguments.push_back(replicated_arguments); in ExecuteGraphParallel()
670 all_arguments.insert(all_arguments.end(), executors.size() - 1, {{}}); in ExecuteGraphParallel()
703 TF_RETURN_IF_ERROR(RecordArguments(all_arguments[i].front(), stream.get(), in ExecuteGraphParallel()
726 executable_ptrs[0], all_arguments[0], execute_backend_.get(), in ExecuteGraphParallel()
735 ExecuteParallelAndRegisterResult(executable_ptrs, all_arguments, in ExecuteGraphParallel()
/external/pytorch/torch/csrc/jit/python/
Dpybind_utils.h903 size_t all_arguments = args.size() + kwargs.size(); in matchSchemaAllowFakeScriptObject() local
904 if (all_arguments > schema.arguments().size()) { in matchSchemaAllowFakeScriptObject()
910 all_arguments, in matchSchemaAllowFakeScriptObject()
989 size_t all_arguments = (self ? 1 : 0) + args.size() + kwargs.size(); in createStackForSchema() local
990 if (all_arguments > schema.arguments().size()) { in createStackForSchema()
996 all_arguments, in createStackForSchema()
/external/pytorch/torchgen/
Dmodel.py2444 all_arguments: list[str] = []
2445 all_arguments.extend(map(str, self.flat_positional))
2447 all_arguments.append("*")
2448 all_arguments.extend(map(str, self.flat_kwarg_only))
2449 all_arguments.extend(map(str, self.out))
2450 return ", ".join(all_arguments)