/external/tensorflow/tensorflow/core/kernels/data/ |
D | map_defun_op.cc | 138 const std::vector<TensorShape> arg_shapes; member 152 std::vector<TensorShape> arg_shapes, int64 batch_size, in ComputeOptions() 156 arg_shapes(std::move(arg_shapes)), in ComputeOptions() 196 compute_opts_->arg_shapes.at(index)); in GetArg() 287 std::vector<TensorShape> arg_shapes; in SetupArgs() local 288 arg_shapes.reserve(arguments.size()); in SetupArgs() 291 arg_shapes.push_back(arguments[i].shape()); in SetupArgs() 292 arg_shapes.at(i).RemoveDim(0); in SetupArgs() 296 ctx, arguments, captured_inputs, std::move(arg_shapes), batch_size, in SetupArgs()
|
/external/tensorflow/tensorflow/compiler/jit/ |
D | shape_inference.cc | 43 const std::map<int, InferredShape>& arg_shapes, in PropagateShapes() argument 70 auto it = arg_shapes.find(index); in PropagateShapes() 71 if (it != arg_shapes.end()) { in PropagateShapes() 135 Status InferShapes(Graph* graph, const std::map<int, InferredShape>& arg_shapes, in InferShapes() argument 152 TF_RETURN_IF_ERROR(PropagateShapes(*graph, arg_shapes, &shape_refiner)); in InferShapes()
|
D | xla_compilation_cache.cc | 73 for (const auto& a : arg_shapes) { in HumanString() 86 if (arg_shapes != other.arg_shapes) return false; in operator ==() 102 for (const auto& arg : signature.arg_shapes) { in operator ()() 129 signature.arg_shapes.emplace_back(arg.type, arg.DimensionSizes()); in BuildSignature()
|
D | shape_inference.h | 43 Status InferShapes(Graph* graph, const std::map<int, InferredShape>& arg_shapes,
|
D | encapsulate_util.cc | 315 std::map<int, InferredShape> arg_shapes; in PerformStaticShapeInferenceBeforeEncapsulation() local 318 InferShapes(g, arg_shapes, /*fnlib_def=*/nullptr, &shape_info)); in PerformStaticShapeInferenceBeforeEncapsulation()
|
D | xla_compilation_cache.h | 100 std::vector<std::pair<DataType, std::vector<int64>>> arg_shapes; member
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | index_ops_cpu.cc | 108 std::vector<xla::Shape> arg_shapes; in Compile() local 115 arg_shapes.push_back(std::move(arg_shape)); in Compile() 122 xla_shape, arg_shapes); in Compile() 125 xla_shape, arg_shapes); in Compile()
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | xla_jit_compiled_cpu_function.cc | 104 std::vector<const xla::Shape*> arg_shapes; in Compile() local 105 arg_shapes.reserve(program_shape->parameters_size()); in Compile() 107 arg_shapes.push_back(&program_shape->parameters(i)); in Compile() 114 client->Compile(computation, arg_shapes, build_options)); in Compile()
|
D | xla_compiler.cc | 767 std::vector<xla::Shape> arg_shapes(input_to_args->size()); in BuildArguments() local 771 args[(*input_to_args)[i]], is_entry_computation, &arg_shapes[i])); in BuildArguments() 775 input_shapes->push_back(xla::ShapeUtil::MakeTupleShape(arg_shapes)); in BuildArguments() 777 *input_shapes = arg_shapes; in BuildArguments() 857 << " shape: " << xla::ShapeUtil::HumanString(arg_shapes[i]) in BuildArguments()
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | shape_inference.cc | 349 absl::Span<const Shape* const> arg_shapes, const int64 dimension) { in InferConcatOpShape() argument 350 if (arg_shapes.empty()) { in InferConcatOpShape() 353 if (dimension < 0 || dimension >= arg_shapes[0]->rank()) { in InferConcatOpShape() 359 for (const Shape* shape : arg_shapes) { in InferConcatOpShape() 400 for (size_t i = 1; i < arg_shapes.size(); ++i) { in InferConcatOpShape() 401 new_dimensions[dimension] += arg_shapes[i]->dimensions(dimension); in InferConcatOpShape() 1073 absl::Span<const Shape* const> arg_shapes, const ProgramShape& to_apply, in InferMapShape() argument 1075 if (arg_shapes.empty()) { in InferMapShape() 1080 const Shape* arg_shape = arg_shapes[0]; in InferMapShape() 1081 for (size_t i = 1; i < arg_shapes.size(); ++i) { in InferMapShape() [all …]
|
D | shape_inference.h | 83 absl::Span<const Shape* const> arg_shapes, const ProgramShape& to_apply, 152 absl::Span<const Shape* const> arg_shapes, 241 absl::Span<const Shape* const> arg_shapes, int64 dimension); 271 absl::Span<const Shape* const> arg_shapes, const ProgramShape& to_apply);
|
/external/tensorflow/tensorflow/compiler/xla/python/ |
D | numpy_bridge.cc | 157 Safe_PyObjectPtr arg_shapes = make_safe(PyTuple_New(shape.parameters_size())); in PyProgramShapeInfoFromXlaProgramShape() local 159 PyTuple_SET_ITEM(arg_shapes.get(), i, in PyProgramShapeInfoFromXlaProgramShape() 165 PyTuple_Pack(2, arg_shapes.release(), result_shape.release())); in PyProgramShapeInfoFromXlaProgramShape()
|
D | xla_client.py | 642 arg_shapes, result_shape = shape_info 643 return ProgramShape([_wrap_shape(arg) for arg in arg_shapes],
|
/external/tensorflow/tensorflow/python/eager/ |
D | function.py | 1565 arg_shapes = _flat_shape_list(args, kwargs) 1573 flat_to_check=arg_shapes)): 1577 relaxed_arg_shapes = arg_shapes 1579 if len(arg_shapes) != len(relaxed_arg_shapes): 1582 % (len(arg_shapes), len(relaxed_arg_shapes))) 1585 arg_shapes, relaxed_arg_shapes)]
|
/external/tensorflow/tensorflow/python/framework/ |
D | function.py | 844 arg_shapes=None, argument 890 if arg_shapes is None: 891 arg_shapes = [None] * len(arg_types) 894 for (argname, argtype, argshape) in zip(arg_names, arg_types, arg_shapes):
|
D | func_graph.py | 610 arg_shapes = override_flat_arg_shapes[:len(flat_args)] 613 arg_shapes = None 616 args, arg_names, flat_shapes=arg_shapes)
|