/external/tensorflow/tensorflow/core/kernels/linalg/ |
D | svd_op_gpu.cu.cc | 133 u_shape.AddDim(m); in RunSVD() 134 u_shape.AddDim(m); in RunSVD() 136 v_shape.AddDim(n); in RunSVD() 137 v_shape.AddDim(n); in RunSVD() 145 u_shape.AddDim(p); in RunSVD() 146 u_shape.AddDim(m); in RunSVD() 148 v_shape.AddDim(p); in RunSVD() 149 v_shape.AddDim(n); in RunSVD() 296 input_shape.AddDim(n); in PerformSVD_MgeqN() 297 input_shape.AddDim(m); in PerformSVD_MgeqN() [all …]
|
D | lu_op.cc | 82 batch_shape.AddDim(input.dim_size(dim)); in Compute() 94 permutation_shape.AddDim(num_rows); in Compute()
|
D | einsum_op_impl.h | 344 transposed_shape.AddDim(input.dim_size(permutation[i])); in TransposeOperand() 502 output_shape.AddDim(dim); in ReduceOperand() 509 output_shape.AddDim(reshape[kFree]); in ReduceOperand() 510 output_shape.AddDim(reshape[kContract]); in ReduceOperand() 568 output_shape.AddDim(inputs[i].dim_size(free_axis)); in ContractOperands() 663 result_shape.AddDim(label_to_dim_sizes[label]); in Compute()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | spacetobatch_op.cc | 134 external_output_shape.AddDim(orig_input_tensor.dim_size(0) * in SpaceToBatchOpCompute() 141 external_output_shape.AddDim(size); in SpaceToBatchOpCompute() 143 internal_input_shape.AddDim(input_batch_size); in SpaceToBatchOpCompute() 144 internal_output_shape.AddDim(input_batch_size * block_shape_product); in SpaceToBatchOpCompute() 162 internal_input_shape.AddDim(input_size); in SpaceToBatchOpCompute() 164 internal_output_shape.AddDim(output_size); in SpaceToBatchOpCompute() 165 external_output_shape.AddDim(output_size); in SpaceToBatchOpCompute() 172 external_output_shape.AddDim(size); in SpaceToBatchOpCompute() 175 internal_input_shape.AddDim(depth); in SpaceToBatchOpCompute() 176 internal_output_shape.AddDim(depth); in SpaceToBatchOpCompute()
|
D | batchtospace_op.cc | 137 external_output_shape.AddDim(orig_input_batch_size / block_shape_product); in BatchToSpaceOpCompute() 143 external_output_shape.AddDim(size); in BatchToSpaceOpCompute() 145 internal_input_shape.AddDim(input_batch_size); in BatchToSpaceOpCompute() 146 internal_output_shape.AddDim(input_batch_size / block_shape_product); in BatchToSpaceOpCompute() 161 internal_input_shape.AddDim(input_size); in BatchToSpaceOpCompute() 162 internal_output_shape.AddDim(cropped_size); in BatchToSpaceOpCompute() 163 external_output_shape.AddDim(cropped_size); in BatchToSpaceOpCompute() 170 external_output_shape.AddDim(size); in BatchToSpaceOpCompute() 173 internal_input_shape.AddDim(depth); in BatchToSpaceOpCompute() 174 internal_output_shape.AddDim(depth); in BatchToSpaceOpCompute()
|
D | reduction_ops_common.cc | 24 for (auto size : out_reshape_) shape.AddDim(size); in out_reshape() 31 for (auto size : out_shape_) shape.AddDim(size); in out_shape() 39 shape.AddDim(data_reshape_[i]); in shuffled_shape() 42 shape.AddDim(data_reshape_[i]); in shuffled_shape()
|
D | reshape_op.h | 127 shape->AddDim(1); in ValidateSizes() 135 shape->AddDim(size); in ValidateSizes() 138 shape->AddDim(size); in ValidateSizes()
|
D | gather_op.cc | 142 result_shape.AddDim(params.dim_size(i)); in Compute() 146 result_shape.AddDim(params.dim_size(i)); in Compute() 150 result_shape.AddDim(indices.dim_size(i)); in Compute() 153 result_shape.AddDim(params.dim_size(i)); in Compute()
|
D | diag_op.cc | 55 out_shape.AddDim(diagonal.dim_size(i)); in Compute() 58 out_shape.AddDim(diagonal.dim_size(i)); in Compute() 95 out_shape.AddDim(tensor.dim_size(i)); in Compute()
|
D | decode_raw_op.cc | 59 out_shape.AddDim(0); in Compute() 71 out_shape.AddDim(added_dim); in Compute()
|
D | unsorted_segment_join_op.cc | 64 output_shape.AddDim(num_segments); in GetOutputShape() 67 output_shape.AddDim(input_shape.dim_size(index)); in GetOutputShape()
|
D | reshape_util.cc | 119 output_shape.AddDim(1); in ReshapeSparseTensor() 125 output_shape.AddDim(size); in ReshapeSparseTensor()
|
D | reduce_join_op.cc | 104 if (keep_dims) output_shape.AddDim(1); in GetOutputShape() 106 output_shape.AddDim(input_shape.dim_size(index)); in GetOutputShape()
|
D | dynamic_partition_op.cc | 72 shape.AddDim(partition_count[p]); in ValidateAndAllocateOutputs() 74 shape.AddDim((*data)->dim_size(i)); in ValidateAndAllocateOutputs()
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | matrix_triangular_solve_op.cc | 90 lhs_broadcast_shape.AddDim(m); in Broadcast() 91 lhs_broadcast_shape.AddDim(m); in Broadcast() 99 rhs_broadcast_shape.AddDim(m); in Broadcast() 100 rhs_broadcast_shape.AddDim(n); in Broadcast()
|
D | reshape_op.cc | 65 shape.AddDim(1); in Compile() 70 shape.AddDim(size); in Compile() 76 shape.AddDim(size); in Compile()
|
D | matrix_diag_ops.cc | 313 output_shape.AddDim(num_rows); in Compile() 314 output_shape.AddDim(num_cols); in Compile() 391 if (num_diags > 1) output_shape.AddDim(num_diags); in Compile() 395 output_shape.AddDim(max_diag_len); in Compile() 506 if (num_diags > 1) expected_diag_shape.AddDim(num_diags); in Compile() 510 expected_diag_shape.AddDim(max_diag_len); in Compile()
|
/external/tensorflow/tensorflow/core/framework/ |
D | tensor_shape_test.cc | 68 s.AddDim(100); in TEST() 180 s.AddDim(1); in TEST() 182 s.AddDim(100000); in TEST() 185 s.AddDim(2); in TEST() 187 s.AddDim(4); in TEST() 189 s.AddDim(3); in TEST() 266 void AddDim(int64 size); 409 AddDim(d.size()); in TensorShapeOld() 417 AddDim(s); in TensorShapeOld() 428 void TensorShapeOld::AddDim(int64 size) { in AddDim() function in tensorflow::__anon127dd62f0111::TensorShapeOld [all …]
|
D | tensor_shape.cc | 153 AddDim(d.size()); in TensorShapeBase() 396 void TensorShapeBase<Shape>::AddDim(int64 size) { in AddDim() function in tensorflow::TensorShapeBase 489 for (auto d : shape) AddDim(d.size); in AppendShape() 516 AddDim(dval); in InsertDim() 585 AddDim(dval); in set_dim() 645 AddDim(dval); in RemoveDimRange() 851 out.AddDim(size); in Concatenate() 867 for (auto dim : shape) out.AddDim(dim.size); in Concatenate()
|
/external/tensorflow/tensorflow/core/util/ |
D | example_proto_fast_parsing.cc | 1163 out_shape.AddDim(serialized.size()); in FastParseExample() 1165 out_shape.AddDim(dim); in FastParseExample() 1259 indices_shape.AddDim(total_num_features); in FastParseExample() 1260 indices_shape.AddDim(2); in FastParseExample() 1265 values_shape.AddDim(total_num_features); in FastParseExample() 1312 row_splits_shape.AddDim(serialized.size() + 1); in FastParseExample() 1323 values_shape.AddDim(total_num_features); in FastParseExample() 1378 values_shape.AddDim(batch_size); in FastParseExample() 1379 values_shape.AddDim(max_num_elements); in FastParseExample() 1381 values_shape.AddDim(config.dense[d].shape.dim_size(i)); in FastParseExample() [all …]
|
D | strided_slice_op.cc | 283 processing_shape->AddDim(shrink_i ? 1 : -1); in ValidateStridedSliceOp() 370 processing_shape->AddDim(size_i); in ValidateStridedSliceOp() 372 processing_shape->AddDim(-1); in ValidateStridedSliceOp() 400 final_shape->AddDim(processing_shape->dim_size(gather_index)); in ValidateStridedSliceOp() 406 final_shape->AddDim(1); in ValidateStridedSliceOp()
|
D | example_proto_helper.cc | 313 output_shapes->indices_shape.AddDim(total_num_features); in GetSparseTensorShapes() 314 output_shapes->indices_shape.AddDim(2); in GetSparseTensorShapes() 315 output_shapes->values_shape.AddDim(total_num_features); in GetSparseTensorShapes() 349 out_shape.AddDim(batch_size); in BatchExampleProtoToTensors() 352 for (const int dim : shape.dim_sizes()) out_shape.AddDim(dim); in BatchExampleProtoToTensors() 659 dense_shape.AddDim(dense_shapes[i].dim_size(d)); in GetDenseShapes()
|
/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_reshape_op.cc | 217 shape->AddDim(1); in ValidateSizes() 225 shape->AddDim(size); in ValidateSizes() 228 shape->AddDim(size); in ValidateSizes()
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | xla_resource.cc | 138 ta_shape.AddDim(max_array_size_); in SetZeroValue() 146 ta_shape.AddDim(max_array_size_); in SetZeroValue() 171 ta_shape.AddDim(max_array_size_); in GetOrCreateTensorArrayGradient()
|
/external/tensorflow/tensorflow/core/kernels/sparse/ |
D | mat_mul_op.cc | 242 if (rank == 3) output_shape.AddDim(batch_size); in AllocateOutput() 525 if (rank == 3) c_shape.AddDim(batch_size); in Compute() 527 c_shape.AddDim(b_outer_dim); in Compute() 528 c_shape.AddDim(a_outer_dim); in Compute() 530 c_shape.AddDim(a_outer_dim); in Compute() 531 c_shape.AddDim(b_outer_dim); in Compute() 641 b_t_transposed_shape.AddDim(batch_size); in Compute() 643 b_t_transposed_shape.AddDim(b_t.dim_size(row_dim + 1)); in Compute() 644 b_t_transposed_shape.AddDim(b_t.dim_size(row_dim)); in Compute()
|