Searched refs:tensor_inputs (Results 1 – 13 of 13) sorted by relevance
/external/pytorch/aten/src/ATen/functorch/ |
D | BatchRulesHelper.h | 142 std::vector<std::pair<Tensor, std::optional<int64_t>>> tensor_inputs; 148 tensor_inputs.emplace_back(tensor_value, tensor_bdim); 152 Func(tensor_inputs); 160 TORCH_INTERNAL_ASSERT(tensor_idx < tensor_inputs.size()); 161 torch::jit::push(stack, tensor_inputs[tensor_idx].first); 177 … void handle_pointwise_ops(std::vector<std::pair<Tensor, std::optional<int64_t>>> &tensor_inputs) { 179 for (auto& tensor_input : tensor_inputs) { 183 for (auto& tensor_input: tensor_inputs) { 195 …void handle_variadic_bdims(std::vector<std::pair<Tensor, std::optional<int64_t>>> &tensor_inputs) { 196 for (auto & tensor_input : tensor_inputs) { [all …]
|
/external/pytorch/benchmarks/framework_overhead_benchmark/ |
D | pt_wrapper_module.py | 27 self.tensor_inputs = [] 30 self.tensor_inputs.append(torch.randn(1)) 32 self.module = torch.jit.trace(self.module, self.tensor_inputs) 47 self.module.forward(*self.tensor_inputs)
|
D | utils.py | 27 bench.add_input(*module.tensor_inputs)
|
/external/pytorch/torch/csrc/jit/backends/xnnpack/ |
D | xnnpack_graph_builder.cpp | 67 std::vector<at::Tensor> tensor_inputs, in runGraphOnInputs() argument 85 tensor_inputs.size() == _inputs.size(), in runGraphOnInputs() 87 for (int i = 0; i < tensor_inputs.size(); i++) { in runGraphOnInputs() 89 {_val_to_ids[_inputs[i]], tensor_inputs[i].data_ptr<float>()}); in runGraphOnInputs()
|
D | xnnpack_graph_builder.h | 86 std::vector<at::Tensor> tensor_inputs,
|
/external/tensorflow/tensorflow/python/saved_model/ |
D | function_deserialization.py | 69 tensor_inputs = [] 72 tensor_inputs.append( 75 tensor_inputs.append(arg) 76 …result = function._call_flat(tensor_inputs, function.captured_inputs) # pylint: disable=protected…
|
/external/pytorch/torch/csrc/jit/passes/ |
D | graph_fuser.cpp | 846 auto tensor_inputs = filter( in tryToMoveChunk() local 849 auto tensor_sizes = fmap(tensor_inputs, [&](Value* v) { in tryToMoveChunk() 994 auto tensor_inputs = filter(n->inputs(), [](Value* v) { in buildShapeExpressions() local 998 fmap(tensor_inputs, [&](Value* v) { return shape_of.at(v); }); in buildShapeExpressions()
|
D | tensorexpr_fuser.cpp | 469 auto tensor_inputs = filter(n->inputs(), [](Value* v) { in buildShapeExpressions() local 474 auto shapes = fmap(tensor_inputs, [&](Value* v) { in buildShapeExpressions()
|
/external/pytorch/torch/utils/ |
D | checkpoint.py | 252 tensor_inputs = [] 255 tensor_inputs.append(arg) 261 ctx.save_for_backward(*tensor_inputs)
|
/external/pytorch/functorch/csrc/dim/ |
D | dim.cpp | 2404 Slice<TensorInfo> tensor_inputs; in getsetitem_flat() local 2408 tensor_inputs.append(A, TensorInfo()); in getsetitem_flat() 2413 tensor_inputs.append(A, ti); in getsetitem_flat() 2532 if(tensor_inputs[i]) { in getsetitem_flat() 2535 for (auto l : tensor_inputs[i].levels) { in getsetitem_flat() 2550 … tensor_inputs[i] = TensorInfo {d->range(), Slice<DimEntry>(A, DimEntry(d)), false, TensorRef()}; in getsetitem_flat() 2580 if (tensor_inputs[i]) { in getsetitem_flat() 2583 TensorRef t = tensor_inputs[i].tensor; in getsetitem_flat() 2584 if (!tensor_inputs[i].has_device && device_holding_tensor) { in getsetitem_flat() 2587 …flat_inputs[i] = handle_from_tensor(A, _match_levels(A, t, tensor_inputs[i].levels, index_levels)); in getsetitem_flat()
|
/external/pytorch/torch/autograd/ |
D | gradcheck.py | 515 tensor_inputs = tuple(i for i in inputs if is_tensor_like(i) and i.requires_grad) 517 if any(i.is_complex() for i in tensor_inputs): 524 _allocate_jacobians_with_outputs(outputs, 1) for i in tensor_inputs 528 _allocate_jacobians_with_outputs(outputs, i.numel()) for i in tensor_inputs 1232 tensor_inputs = tuple(i for i in inputs if is_tensor_like(i) and i.requires_grad)
|
/external/tensorflow/tensorflow/python/eager/ |
D | function.py | 1788 tensor_inputs = [] 1797 tensor_inputs.append(arg.handle) 1800 tensor_inputs.append(arg) 1806 for i, tensor_input in enumerate(tensor_inputs): 1828 args = tensor_inputs + captured_inputs
|
/external/tensorflow/tensorflow/compiler/tf2tensorrt/convert/ |
D | convert_nodes.cc | 5743 std::vector<ITensorProxyPtr> tensor_inputs; in ConvertAddN() local 5744 tensor_inputs.reserve(inputs.size()); in ConvertAddN() 5747 tensor_inputs.push_back(input.tensor()); in ConvertAddN() 5751 tensor_inputs.push_back(params->converter->CreateConstantLayer( in ConvertAddN() 5755 ITensorProxyPtr lhs = tensor_inputs[0]; in ConvertAddN() 5757 ITensorProxyPtr rhs = tensor_inputs[i]; in ConvertAddN()
|