/external/tensorflow/tensorflow/core/util/ |
D | bcast_test.cc | 26 string BCast(const tensorflow::BCast::Vec& x, const tensorflow::BCast::Vec& y, in BCast() function 28 tensorflow::BCast b(x, y, fewer_dims_optimization); in BCast() 44 string BCastBatchIndices(const tensorflow::BCast::Vec& x, in BCastBatchIndices() 45 const tensorflow::BCast::Vec& y, in BCastBatchIndices() 47 tensorflow::BCast b(x, y, fewer_dims_optimization, in BCastBatchIndices() 55 string BCastList3(const tensorflow::BCast::Vec& x, in BCastList3() 56 const tensorflow::BCast::Vec& y, in BCastList3() 57 const tensorflow::BCast::Vec& z, in BCastList3() 80 EXPECT_EQ("invalid", BCast({5, 3, 2}, {3}, use_optimization)); in TEST() 81 EXPECT_EQ("invalid", BCast({5, 3, 2}, {2, 2}, use_optimization)); in TEST() [all …]
|
D | bcast.cc | 21 BCast::Vec BCast::FromShape(const TensorShape& shape) { in FromShape() 30 TensorShape BCast::ToShape(const BCastList::Vec& vec) { in ToShape()
|
D | matmul_bcast.h | 34 using Vec = BCast::Vec; 42 absl::make_unique<BCast>(std::move(x_resized), std::move(y_resized)); in MatMulBCast() 88 std::unique_ptr<BCast> batch_bcast_;
|
D | bcast.h | 352 class BCast : public BCastList<2> { 364 BCast(const Vec& x, const Vec& y, const bool fewer_dims_optimization = true, 369 ~BCast() {} in ~BCast() 402 const BCast::Vec& vec) { in ToIndexArrayType() 411 const BCast::Vec& vec) { in ToIndexArray() 420 TF_DISALLOW_COPY_AND_ASSIGN(BCast);
|
/external/tensorflow/tensorflow/core/kernels/ |
D | betainc_op.cc | 83 auto merged_shape_vec = BCast::FromShape(merged_shape); in Compute() 84 BCast a_shaper(BCast::FromShape(a_shape), merged_shape_vec); in Compute() 85 BCast b_shaper(BCast::FromShape(b_shape), merged_shape_vec); in Compute() 86 BCast x_shaper(BCast::FromShape(x_shape), merged_shape_vec); in Compute() 97 functor.BCast(ctx->eigen_device<Device>(), a_value, \ in Compute() 98 BCast::ToIndexArray<NDIM>(a_shaper.x_bcast()), b_value, \ in Compute() 99 BCast::ToIndexArray<NDIM>(b_shaper.x_bcast()), x_value, \ in Compute() 100 BCast::ToIndexArray<NDIM>(x_shaper.x_bcast()), \ in Compute() 137 void Betainc<GPUDevice, T, NDIM>::BCast( \
|
D | bcast_ops.cc | 34 gtl::InlinedVector<BCast::Vec, 4> shapes; in Compute() 40 BCast::Vec vec; in Compute() 46 BCast bcast(shapes[0], shapes[1]); in Compute() 57 void Output(OpKernelContext* ctx, int idx, const BCast::Vec& v) { in Output() 83 gtl::InlinedVector<BCast::Vec, 4> shapes; in Compute() 89 BCast::Vec vec; in Compute() 95 BCast bcast(shapes[0], shapes[1]); in Compute() 107 void Output(OpKernelContext* ctx, int idx, const BCast::Vec& v) { in Output()
|
D | xent_op.cc | 46 BCast bcast(BCast::FromShape(logits_in.shape()), in Compute() 47 BCast::FromShape(labels_in.shape())); in Compute() 54 shape_in = BCast::ToShape(bcast.output_shape()); in Compute() 87 BCast::ToIndexArray<2>(bcast.x_bcast()), in Compute() 88 BCast::ToIndexArray<2>(bcast.y_bcast()), in Compute()
|
D | broadcast_to_op.cc | 79 BCast bcast(BCast::FromShape(input_shape), BCast::FromShape(output_shape), in Compute() 85 OP_REQUIRES(ctx, BCast::ToShape(bcast.output_shape()) == output_shape, in Compute() 114 const TensorShape& input_shape, const BCast& bcast) const; \
|
D | broadcast_to_op.h | 51 const Tensor &input_tensor, const BCast &bcast) const { in ReshapeAndBCast() 59 BCast::ToIndexArrayType<int, NDIMS>(bcast.x_bcast())); in ReshapeAndBCast() 64 BCast::ToIndexArrayType<Eigen::DenseIndex, NDIMS>(bcast.x_bcast())); in ReshapeAndBCast() 74 const BCast &bcast) const { in operator()
|
D | substr_op.cc | 136 BCast bcast(BCast::FromShape(input_shape), BCast::FromShape(pos_shape)); in Compute() 141 TensorShape output_shape = BCast::ToShape(bcast.result_shape()); in Compute() 162 pos_shaped.broadcast(BCast::ToIndexArray<1>(bcast.y_bcast())); in Compute() 172 len_shaped.broadcast(BCast::ToIndexArray<1>(bcast.y_bcast())); in Compute() 215 pos_shaped.broadcast(BCast::ToIndexArray<2>(bcast.y_bcast())); in Compute() 225 len_shaped.broadcast(BCast::ToIndexArray<2>(bcast.y_bcast())); in Compute()
|
D | cwise_ops_common.cc | 58 bcast(BCast::FromShape(in0.shape()), BCast::FromShape(in1.shape())) { in BinaryOpState() 77 const TensorShape output_shape = BCast::ToShape(bcast.output_shape()); in BinaryOpState()
|
D | sparse_dense_binary_op_shared.cc | 84 const auto lhs_dims = BCast::FromShape(TensorShape(shape_vec)); in Compute() 85 const auto rhs_dims = BCast::FromShape(dense_t->shape()); in Compute() 86 BCast b(lhs_dims, rhs_dims, false); // false for keeping the same num dims. in Compute() 126 .broadcast(BCast::ToIndexArray<NDIM>(b.y_bcast())); \ in Compute()
|
D | cwise_ops_common.h | 60 BCast bcast; 185 functor::BinaryFunctor<Device, Functor, 2>().BCast( in Compute() 188 BCast::ToIndexArray<2>(bcast.x_bcast()), in Compute() 190 BCast::ToIndexArray<2>(bcast.y_bcast()), error_ptr); in Compute() 192 functor::BinaryFunctor<Device, Functor, 3>().BCast( in Compute() 195 BCast::ToIndexArray<3>(bcast.x_bcast()), in Compute() 197 BCast::ToIndexArray<3>(bcast.y_bcast()), error_ptr); in Compute() 199 functor::BinaryFunctor<Device, Functor, 4>().BCast( in Compute() 202 BCast::ToIndexArray<4>(bcast.x_bcast()), in Compute() 204 BCast::ToIndexArray<4>(bcast.y_bcast()), error_ptr); in Compute() [all …]
|
D | cwise_op_select.cc | 163 BCast cond_bcast(bcast.output_shape(), cond->shape().dim_sizes(), false); in Compute() 164 BCast then_bcast(bcast.output_shape(), then->shape().dim_sizes(), false); in Compute() 165 BCast else_bcast(bcast.output_shape(), else_->shape().dim_sizes(), false); in Compute() 186 const TensorShape output_shape = BCast::ToShape(bcast.output_shape()); in Compute() 202 BCast::ToIndexArray<NDIMS>(cond_bcast.y_bcast()), \ in Compute() 203 BCast::ToIndexArray<NDIMS>(then_bcast.y_bcast()), \ in Compute() 204 BCast::ToIndexArray<NDIMS>(else_bcast.y_bcast())); \ in Compute()
|
D | random_binomial_op.cc | 177 const BCast& bcast, typename TTypes<T>::ConstFlat counts, in operator ()() 342 tensorflow::BCast bcast(counts_tensor.shape().dim_sizes(), in Compute() 362 TensorShape bcast_shape = BCast::ToShape(bcast.output_shape()); in Compute() 459 tensorflow::BCast bcast(counts_tensor.shape().dim_sizes(), in Compute() 479 TensorShape bcast_shape = BCast::ToShape(bcast.output_shape()); in Compute()
|
D | quantized_mul_op.cc | 292 BCast bcast(BCast::FromShape(x.shape()), BCast::FromShape(y.shape())); in Compute() 301 0, BCast::ToShape(bcast.output_shape()), &z)); in Compute()
|
D | betainc_op.h | 36 void BCast(const Device& d, typename TTypes<T, NDIM>::ConstTensor a, in BCast() function
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | bcast_ops.cc | 42 absl::InlinedVector<BCast::Vec, 2> shapes; in Compile() 50 shapes.push_back(BCast::Vec(shape.begin(), shape.end())); in Compile() 52 BCast bcast(shapes[0], shapes[1]); in Compile() 91 absl::InlinedVector<BCast::Vec, 4> shapes; in Compile() 100 shapes.push_back(BCast::Vec(vec.begin(), vec.end())); in Compile() 102 BCast bcast(shapes[0], shapes[1]); in Compile() 112 void Output(XlaOpKernelContext* ctx, int idx, const BCast::Vec& v) { in Output()
|
D | select_op.cc | 95 BCast bcast_then_else(BCast::FromShape(then_shape), in Compile() 96 BCast::FromShape(else_shape), in Compile() 104 BCast bcast(bcast_then_else.output_shape(), BCast::FromShape(cond_shape), in Compile() 109 BCast::ToShape(bcast_then_else.output_shape()).DebugString(), " vs. ", in Compile()
|
D | binary_ops.cc | 47 const BCast& broadcast_helper, \ 75 xla::XlaOp y, const BCast& broadcast_helper) { in DivNoNanImpl() 93 xla::XlaOp y, const BCast& broadcast_helper) { in MulNoNanImpl() 116 xla::XlaOp y, const BCast& broadcast_helper) { in FloorDivImpl() 146 const BCast& broadcast_helper) { in XlogyImpl() 155 const BCast& broadcast_helper) { in Xlog1pyImpl() 165 const BCast& broadcast_helper) { in XdivyImpl() 178 xla::XlaOp y, const BCast& broadcast_helper) { in FloorModImpl() 261 const BCast& broadcast_helper) { in IgammaImpl() 269 const BCast& broadcast_helper) { in IgammaGradAImpl() [all …]
|
D | cwise_ops.cc | 43 BCast bcast(BCast::FromShape(lhs_shape), BCast::FromShape(rhs_shape), in Compile() 90 xla::XlaOp lhs, xla::XlaOp rhs, const BCast& broadcast_helper) { in Broadcast()
|
D | cwise_ops.h | 61 const absl::Span<const int64>& rhs_shape, const BCast& broadcast_helper, 70 xla::XlaOp lhs, xla::XlaOp rhs, const BCast& broadcast_helper);
|
D | matrix_triangular_solve_op.cc | 45 MatMulBCast bcast(BCast::FromShape(lhs_shape), BCast::FromShape(rhs_shape)); in Compile()
|
/external/tensorflow/tensorflow/compiler/tf2xla/lib/ |
D | broadcast.cc | 102 BCast bcast(BCast::FromShape(lhs_tf_shape), BCast::FromShape(rhs_tf_shape)); in BroadcastOpsToSame()
|
/external/tensorflow/tensorflow/core/grappler/utils/ |
D | symbolic_shapes.cc | 26 BCast::Vec ShapeDims(const TensorShapeProto& shape) { in ShapeDims() 27 BCast::Vec dims; in ShapeDims() 102 BCast bcast(ShapeDims(left), ShapeDims(right), in ShapesBroadcastable() 118 BCast bcast(ShapeDims(left), ShapeDims(right), in ShapeAfterBroadcast()
|