Searched refs:new_dimensions (Results 1 – 10 of 10) sorted by relevance
/external/libvpx/vp9/encoder/ |
D | vp9_lookahead.c | 108 int larger_dimensions, new_dimensions; in vp9_lookahead_push() local 118 new_dimensions = width != buf->img.y_crop_width || in vp9_lookahead_push() 125 assert(!larger_dimensions || new_dimensions); in vp9_lookahead_push() 135 if (!new_dimensions && ctx->max_sz == 1 && active_map && !flags) { in vp9_lookahead_push() 179 } else if (new_dimensions) { in vp9_lookahead_push()
|
/external/libaom/av1/encoder/ |
D | lookahead.c | 110 int larger_dimensions, new_dimensions; in av1_lookahead_push() local 123 new_dimensions = width != buf->img.y_crop_width || in av1_lookahead_push() 130 assert(!larger_dimensions || new_dimensions); in av1_lookahead_push() 141 } else if (new_dimensions) { in av1_lookahead_push()
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | layout_normalization.cc | 368 std::vector<int64_t> new_dimensions = in HandleReverse() local 372 HloInstruction::CreateReverse(a0->shape(), a0, new_dimensions)); in HandleReverse() 444 std::vector<int64_t> new_dimensions; in TransformDimensionsForLayoutPreservingHlo() local 456 new_dimensions.push_back(tr_dim - degen_delta); in TransformDimensionsForLayoutPreservingHlo() 458 absl::c_sort(new_dimensions); in TransformDimensionsForLayoutPreservingHlo() 459 return new_dimensions; in TransformDimensionsForLayoutPreservingHlo()
|
D | space_to_batch_converter.cc | 747 std::vector<int64_t> new_dimensions(activations->shape().dimensions().begin(), in SplitAndTransposeMergedBatch() local 756 new_dimensions[batch_dimension] = old_batch_size; in SplitAndTransposeMergedBatch() 761 new_dimensions.insert(new_dimensions.begin() + spatial_dimensions[0], in SplitAndTransposeMergedBatch() 767 MakeReshapeHlo(new_dimensions, activations)); in SplitAndTransposeMergedBatch() 771 std::vector<int64_t> transpose_dims(new_dimensions.size()); in SplitAndTransposeMergedBatch() 774 std::vector<int64_t> trans_dims(new_dimensions.size()); in SplitAndTransposeMergedBatch() 800 std::vector<int64_t> new_dimensions(activations->shape().dimensions().begin(), in ChangeSpatialSizeOnSpaceToBatchedShape() local 1814 std::vector<int64_t> new_dimensions( in Propagate() local 1817 new_dimensions[space_dim] *= (batch_size / old_batch_size); in Propagate() 1818 new_dimensions[batch_dim] = old_batch_size; in Propagate() [all …]
|
D | shape_inference.cc | 413 std::vector<int64_t> new_dimensions(arg_shape->dimensions().begin(), in InferConcatOpShape() local 416 new_dimensions[dimension] += arg_shapes[i]->dimensions(dimension); in InferConcatOpShape() 419 Shape result = ShapeUtil::MakeShape(element_type, new_dimensions); in InferConcatOpShape() 2157 std::vector<int64_t> new_dimensions(shape.dimensions().begin(), in InferAllToAllShape() local 2159 new_dimensions[split_dimension] /= split_count; in InferAllToAllShape() 2160 new_dimensions[concat_dimension] *= split_count; in InferAllToAllShape() 2161 return ShapeUtil::MakeShape(shape.element_type(), new_dimensions); in InferAllToAllShape() 2269 std::vector<int64_t> new_dimensions; in InferReduceShape() local 2273 new_dimensions.push_back(arg.dimensions(i)); in InferReduceShape() 2280 new_dimensions, new_is_dynamic); in InferReduceShape() [all …]
|
D | algebraic_simplifier.cc | 3546 std::vector<int64_t> new_dimensions; in HandleBroadcast() local 3547 new_dimensions.reserve(operand->dimensions().size()); in HandleBroadcast() 3549 new_dimensions.push_back(dims[dim]); in HandleBroadcast() 3554 broadcast->shape(), operand->mutable_operand(0), new_dimensions)); in HandleBroadcast() 5431 std::vector<int64_t> new_dimensions; in HandleReduce() local 5432 new_dimensions.reserve(arg->dimensions().size() + in HandleReduce() 5435 reduce_dims.end(), std::back_inserter(new_dimensions)); in HandleReduce() 5439 new_dimensions, function)); in HandleReduce()
|
/external/tensorflow/tensorflow/compiler/xla/ |
D | literal_util.cc | 389 absl::Span<const int64_t> new_dimensions, in ReshapeSlice() argument 392 for (int64_t i = 0, end = new_dimensions.size(); i < end; ++i) { in ReshapeSlice() 393 new_num_elements *= new_dimensions[i]; in ReshapeSlice() 396 CHECK_EQ(new_dimensions.size(), minor_to_major.size()); in ReshapeSlice() 399 ShapeUtil::MakeShape(literal.shape().element_type(), new_dimensions)); in ReshapeSlice()
|
D | array.h | 532 void Reshape(absl::Span<const int64_t> new_dimensions) { 534 sizes_ = std::vector<int64_t>(new_dimensions.begin(), new_dimensions.end());
|
D | literal_util.h | 242 static Literal ReshapeSlice(absl::Span<const int64_t> new_dimensions,
|
/external/tensorflow/tensorflow/python/ops/ragged/ |
D | dynamic_ragged_shape.py | 774 new_dimensions = new_inner_rank - self.inner_rank 776 [not x.is_uniform() for x in self.row_partitions[-new_dimensions:]]): 779 new_dimensions = new_inner_rank - self.inner_rank 781 x.uniform_row_length() for x in self.row_partitions[-new_dimensions:]
|