/external/tensorflow/tensorflow/compiler/xla/ |
D | window_util.cc | 65 /* static */ string ToString(const WindowDimension& dim) { in ToString() argument 68 string str = StrCat("(size=", dim.size()); in ToString() 69 if (dim.stride() != 1) { in ToString() 70 StrAppend(&str, ",stride=", dim.stride()); in ToString() 72 if (dim.padding_low() != 0) { in ToString() 73 StrAppend(&str, ",padding_low=", dim.padding_low()); in ToString() 75 if (dim.padding_high() != 0) { in ToString() 76 StrAppend(&str, ",padding_high=", dim.padding_high()); in ToString() 78 if (dim.base_dilation() != 1) { in ToString() 79 StrAppend(&str, ",base_dilation=", dim.base_dilation()); in ToString() [all …]
|
D | array_test.cc | 28 EXPECT_EQ(uninit.dim(0), 2); in TEST() 29 EXPECT_EQ(uninit.dim(1), 3); in TEST() 36 EXPECT_EQ(fullof7.dim(0), 1); in TEST() 37 EXPECT_EQ(fullof7.dim(1), 2); in TEST() 38 EXPECT_EQ(fullof7.dim(2), 3); in TEST() 40 for (int64 n0 = 0; n0 < fullof7.dim(0); ++n0) { in TEST() 41 for (int64 n1 = 0; n1 < fullof7.dim(1); ++n1) { in TEST() 42 for (int64 n2 = 0; n2 < fullof7.dim(2); ++n2) { in TEST() 52 EXPECT_EQ(arr.dim(0), 2); in TEST() 53 EXPECT_EQ(arr.dim(1), 3); in TEST() [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | conv_grad_shape_utils.cc | 41 int dim) const { in SpatialPadding() 45 0, static_cast<int>((output_size(dim) - 1) * stride(dim) + in SpatialPadding() 46 (filter_size(dim) - 1) * dilation(dim) + in SpatialPadding() 47 1 - input_size(dim))); in SpatialPadding() 57 int filter_spatial_dim, ConvBackpropSpatialDimension* dim) { in ConvBackpropExtractAndVerifyDimension() argument 58 dim->input_size = input_shape.dim_size(spatial_dim); in ConvBackpropExtractAndVerifyDimension() 59 dim->filter_size = filter_shape.dim_size(filter_spatial_dim); in ConvBackpropExtractAndVerifyDimension() 60 dim->output_size = output_shape.dim_size(spatial_dim); in ConvBackpropExtractAndVerifyDimension() 61 dim->stride = strides[spatial_dim]; in ConvBackpropExtractAndVerifyDimension() 62 dim->dilation = dilations[spatial_dim]; in ConvBackpropExtractAndVerifyDimension() [all …]
|
D | ragged_tensor_to_sparse_kernel.cc | 86 for (int dim = rt_nested_splits_len - 2; dim >= 0; --dim) { in Compute() local 87 while (IsCompleted(pos, dim, rt_nested_splits)) { in Compute() 88 pos[dim] += 1; in Compute() 93 for (int dim = 0; dim < index_prefix.size(); ++dim) { in Compute() local 94 int start = dim > 0 ? rt_nested_splits[dim - 1](pos[dim - 1]) : 0; in Compute() 95 index_prefix[dim] = pos[dim] - start; in Compute() 105 int dim = 0; in Compute() local 107 sparse_indices(next_index, dim++) = index; in Compute() 109 sparse_indices(next_index, dim++) = i; // index_middle in Compute() 111 sparse_indices(next_index, dim++) = index; in Compute() [all …]
|
D | mirror_pad_op.h | 131 for (int dim = 0; dim < Dims; ++dim) { 132 eigen_assert(padding_[dim].first + op.offset() <= dimensions_[dim]); 133 eigen_assert(padding_[dim].second + op.offset() <= dimensions_[dim]); 134 dimensions_[dim] += padding_[dim].first + padding_[dim].second; 175 for (int dim = 0; dim < Dims; ++dim) { 176 coords[dim] = ToInputCoord(coords[dim], dim); 195 int dim = -1; 199 dim = k; 206 dim = k; 215 if (dim < 0) { [all …]
|
D | unique_op_test.cc | 39 TensorProto GetRandomInt32TensorProto(int dim, int max_int) { in GetRandomInt32TensorProto() argument 42 tensor_proto.mutable_tensor_shape()->add_dim()->set_size(dim); in GetRandomInt32TensorProto() 44 for (int i = 0; i < dim; ++i) { in GetRandomInt32TensorProto() 51 TensorProto GetRandomInt32TensorProtoWithRepeat(int dim, int repeat, in GetRandomInt32TensorProtoWithRepeat() argument 55 tensor_proto.mutable_tensor_shape()->add_dim()->set_size(dim); in GetRandomInt32TensorProtoWithRepeat() 57 for (int i = 0; i < dim; ++i) { in GetRandomInt32TensorProtoWithRepeat() 66 static void BM_Unique_INT32(int iters, int dim, int max_int) { in BM_Unique_INT32() argument 70 Tensor input(DT_INT32, TensorShape({dim})); in BM_Unique_INT32() 71 CHECK(input.FromProto(GetRandomInt32TensorProto(dim, max_int))); in BM_Unique_INT32() 79 testing::BytesProcessed(static_cast<int64>(iters) * dim * sizeof(int32)); in BM_Unique_INT32() [all …]
|
D | scatter_nd_op_cpu_impl.h | 106 for (int dim = IXDIM - 1; dim >= 0; --dim) { 107 if (dim == IXDIM - 1) { 108 batch_strides[dim] = 1; 110 batch_strides[dim] = 111 batch_strides[dim + 1] * output_shape_prefix[dim + 1]; 118 for (int dim = 0; dim < IXDIM; ++dim) { 119 const Index ix_d = internal::SubtleMustCopy(Tindices(loc, dim)); 120 out_of_bounds |= !FastBoundsCheck(ix_d, output_shape_prefix[dim]); 121 i += ix_d * batch_strides[dim]; 190 for (int dim = IXDIM - 1; dim >= 0; --dim) { [all …]
|
D | sobol_op.cc | 61 Eigen::VectorXi GetFirstPoint(int i, int dim, in GetFirstPoint() argument 67 Eigen::VectorXi integer_sequence = Eigen::VectorXi::Zero(dim); in GetFirstPoint() 71 for (int j = 0; j < dim; ++j) { in GetFirstPoint() 83 void CalculateSobolSample(int32_t dim, int32_t num_results, int32_t skip, in CalculateSobolSample() argument 92 Eigen::MatrixXi direction_numbers(dim, num_digits); in CalculateSobolSample() 97 for (int j = 0; j < dim; ++j) { in CalculateSobolSample() 108 ? GetFirstPoint(skip + start_point + 1, dim, direction_numbers) in CalculateSobolSample() 111 for (int j = 0; j < dim; ++j) { in CalculateSobolSample() 112 output(start_point * dim + j) = integer_sequence(j) * normalizing_constant; in CalculateSobolSample() 121 for (int j = 0; j < dim; ++j) { in CalculateSobolSample() [all …]
|
D | conv_grad_shape_utils.h | 56 int64 input_size(int dim) const { return spatial_dims[dim].input_size; } in input_size() 57 int64 filter_size(int dim) const { return spatial_dims[dim].filter_size; } in filter_size() 58 int64 output_size(int dim) const { return spatial_dims[dim].output_size; } in output_size() 59 int64 stride(int dim) const { return spatial_dims[dim].stride; } in stride() 60 int64 dilation(int dim) const { return spatial_dims[dim].dilation; } in dilation() 63 int SpatialPadding(const Padding& padding, int dim) const;
|
/external/tensorflow/tensorflow/compiler/tests/ |
D | lstm_layer_inference.config.pbtxt | 2 feed{ id{node_name:"inputs/x_seq_0/read"} shape{dim{size:128}dim{size:1024}} } 3 feed{ id{node_name:"inputs/x_seq_1/read"} shape{dim{size:128}dim{size:1024}} } 4 feed{ id{node_name:"inputs/x_seq_2/read"} shape{dim{size:128}dim{size:1024}} } 5 feed{ id{node_name:"inputs/x_seq_3/read"} shape{dim{size:128}dim{size:1024}} } 6 feed{ id{node_name:"inputs/x_seq_4/read"} shape{dim{size:128}dim{size:1024}} } 7 feed{ id{node_name:"inputs/pad_seq_0/read"} shape{dim{size:128}dim{size:1}} } 8 feed{ id{node_name:"inputs/pad_seq_1/read"} shape{dim{size:128}dim{size:1}} } 9 feed{ id{node_name:"inputs/pad_seq_2/read"} shape{dim{size:128}dim{size:1}} } 10 feed{ id{node_name:"inputs/pad_seq_3/read"} shape{dim{size:128}dim{size:1}} } 11 feed{ id{node_name:"inputs/pad_seq_4/read"} shape{dim{size:128}dim{size:1}} } [all …]
|
/external/tensorflow/tensorflow/core/grappler/utils/ |
D | symbolic_shapes.cc | 30 dims.push_back(shape.dim(i).size()); in ShapeDims() 36 bool IsKnown(const TensorShapeProto::Dim& dim) { return dim.size() >= 0; } in IsKnown() argument 38 bool IsKnownSymbolically(const TensorShapeProto::Dim& dim) { in IsKnownSymbolically() argument 39 return dim.size() <= -2; in IsKnownSymbolically() 42 bool IsUnknown(const TensorShapeProto::Dim& dim) { return dim.size() == -1; } in IsUnknown() argument 47 shape.dim().begin(), shape.dim().end(), in ShapeIsSymbolicallyDefined() 48 [](const TensorShapeProto::Dim& dim) { return !IsUnknown(dim); }); in ShapeIsSymbolicallyDefined() argument 67 for (const auto& dim : shape.dim()) { in NumCoefficients() local 68 if (dim.size() < 0) { in NumCoefficients() 71 num_coefficients *= dim.size(); in NumCoefficients() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/client/lib/ |
D | conv_grad_size_util.cc | 37 SpatialDimensionOutputSizeAndPadding dim; in GetWindowedOutputSize() local 40 dim.output_size = (input_size - effective_filter_size + stride) / stride; in GetWindowedOutputSize() 41 dim.pad_before = dim.pad_after = 0; in GetWindowedOutputSize() 44 dim.output_size = (input_size + stride - 1) / stride; in GetWindowedOutputSize() 46 std::max(int64{0}, (dim.output_size - 1) * stride + in GetWindowedOutputSize() 50 dim.pad_before = padding_needed / 2; in GetWindowedOutputSize() 51 dim.pad_after = padding_needed - dim.pad_before; in GetWindowedOutputSize() 54 if (dim.output_size < 0) { in GetWindowedOutputSize() 56 "Computed output size would be negative: ", dim.output_size, in GetWindowedOutputSize() 61 return dim; in GetWindowedOutputSize() [all …]
|
/external/tensorflow/tensorflow/core/profiler/internal/ |
D | tfprof_tensor.h | 55 bool AddValue(const T& value, TFProfTensorProto* dim) { in AddValue() argument 61 dim->add_value_double(double_val); in AddValue() 63 dim->value_double(dim->value_double_size() - 1)); in AddValue() 67 dim->add_value_int64(int64_val); in AddValue() 69 dim->value_int64(dim->value_int64_size() - 1)); in AddValue() 71 dim->add_value_str(sstream.str()); in AddValue() 73 dim->value_str(dim->value_str_size() - 1), "' "); in AddValue() 84 TFProfTensorProto* dim) { in BuildOutput() argument 94 dim->add_value_double(double_val); in BuildOutput() 96 dim->value_double(dim->value_double_size() - 1)); in BuildOutput() [all …]
|
/external/tensorflow/tensorflow/core/profiler/internal/testdata/ |
D | graph.pbtxt | 9 dim { 12 dim { 15 dim { 18 dim { 37 dim { 40 dim { 43 dim { 46 dim { 71 dim { 89 dim { [all …]
|
/external/eigen/test/ |
D | umeyama.cpp | 91 void run_test(int dim, int num_elements) in run_test() argument 102 MatrixX R = randMatrixSpecialUnitary<Scalar>(dim); in run_test() 103 VectorX t = Scalar(50)*VectorX::Random(dim,1); in run_test() 105 MatrixX cR_t = MatrixX::Identity(dim+1,dim+1); in run_test() 106 cR_t.block(0,0,dim,dim) = c*R; in run_test() 107 cR_t.block(0,dim,dim,1) = t; in run_test() 109 MatrixX src = MatrixX::Random(dim+1, num_elements); in run_test() 110 src.row(dim) = Matrix<Scalar, 1, Dynamic>::Constant(num_elements, Scalar(1)); in run_test() 114 MatrixX cR_t_umeyama = umeyama(src.block(0,0,dim,num_elements), dst.block(0,0,dim,num_elements)); in run_test() 129 const int dim = Dimension; in run_fixed_size_test() local [all …]
|
D | geo_alignedbox.cpp | 32 const Index dim = _box.dim(); in alignedbox() local 34 VectorType p0 = VectorType::Random(dim); in alignedbox() 35 VectorType p1 = VectorType::Random(dim); in alignedbox() 37 p1 = VectorType::Random(dim); } in alignedbox() 40 BoxType b0(dim); in alignedbox() 41 BoxType b1(VectorType::Random(dim),VectorType::Random(dim)); in alignedbox() 60 BoxType box1(VectorType::Random(dim)); in alignedbox() 61 box1.extend(VectorType::Random(dim)); in alignedbox() 62 BoxType box2(VectorType::Random(dim)); in alignedbox() 63 box2.extend(VectorType::Random(dim)); in alignedbox() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoopVectorize/X86/ |
D | register-assumption.ll | 8 br label %loop_exit.dim.11.critedge 10 loop_exit.dim.11.critedge: ; preds = %loop_body.dim.0 14 br label %loop_header.dim.017.preheader 16 loop_header.dim.017.preheader: ; preds = %loop_exit.dim.016, %loop_exit.dim.11.c… 17 br label %loop_body.dim.018 19 loop_body.dim.018: ; preds = %loop_body.dim.018, %loop_header.dim.01… 20 …%invar_address.dim.019.0135 = phi i64 [ 0, %loop_header.dim.017.preheader ], [ %0, %loop_body.dim.… 24 %0 = add nuw nsw i64 %invar_address.dim.019.0135, 1 26 br i1 %1, label %loop_header.dim.017.preheader, label %loop_body.dim.018
|
/external/llvm/test/Transforms/LoopVectorize/X86/ |
D | register-assumption.ll | 8 br label %loop_exit.dim.11.critedge 10 loop_exit.dim.11.critedge: ; preds = %loop_body.dim.0 14 br label %loop_header.dim.017.preheader 16 loop_header.dim.017.preheader: ; preds = %loop_exit.dim.016, %loop_exit.dim.11.c… 17 br label %loop_body.dim.018 19 loop_body.dim.018: ; preds = %loop_body.dim.018, %loop_header.dim.01… 20 …%invar_address.dim.019.0135 = phi i64 [ 0, %loop_header.dim.017.preheader ], [ %0, %loop_body.dim.… 24 %0 = add nuw nsw i64 %invar_address.dim.019.0135, 1 26 br i1 %1, label %loop_header.dim.017.preheader, label %loop_body.dim.018
|
/external/tensorflow/tensorflow/core/util/sparse/ |
D | sparse_tensor.h | 214 static inline int GetSliceIndex(const int dim, const int split_size, in GetSliceIndex() argument 217 DCHECK_GE(dim, 0); in GetSliceIndex() 218 if (residual == 0) return dim / split_size; in GetSliceIndex() 220 if (dim < offset) { in GetSliceIndex() 221 return dim / (split_size + 1); in GetSliceIndex() 223 return residual + ((dim - offset) / split_size); in GetSliceIndex() 228 static inline int GetDimensionInSlice(const int dim, const int split_size, in GetDimensionInSlice() argument 231 DCHECK_GE(dim, 0); in GetDimensionInSlice() 232 if (residual == 0) return dim % split_size; in GetDimensionInSlice() 234 if (dim < offset) { in GetDimensionInSlice() [all …]
|
/external/OpenCL-CTS/test_conformance/basic/ |
D | test_work_item_functions.cpp | 90 for( size_t dim = 1; dim <= 3; dim++ ) in test_work_item_functions() local 95 for( size_t j = 0; j < dim; j++ ) in test_work_item_functions() 108 …error = clEnqueueNDRangeKernel( queue, kernel, (cl_uint)dim, NULL, threads, localThreads, 0, NULL,… in test_work_item_functions() 118 if( testData[ q ].workDim != (cl_uint)dim ) in test_work_item_functions() 120 … return proper value for %d dimensions (expected %d, got %d)\n", (int)dim, (int)dim, (int)testData… in test_work_item_functions() 124 for( size_t j = 0; j < dim; j++ ) in test_work_item_functions() 129 … (int)j, (int)dim, (int)threads[ j ], (int)testData[ q ].globalSize[ j ] ); in test_work_item_functions() 136 … (int)j, (int)dim, (int)threads[ j ], (int)testData[ q ].globalID[ j ] ); in test_work_item_functions() 143 … (int)j, (int)dim, (int)localThreads[ j ], (int)testData[ q ].localSize[ j ] ); in test_work_item_functions() 150 … (int)j, (int)dim, (int)localThreads[ j ], (int)testData[ q ].localID[ j ] ); in test_work_item_functions() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | gpu_conv_runner.cc | 358 for (const WindowDimension& dim : window.dimensions()) { in GetGpuConvParams() local 359 CHECK_EQ(dims_reversed, dim.window_reversal()) << conv->ToString(); in GetGpuConvParams() 360 CHECK_EQ(dim.padding_low(), dim.padding_high()) << conv->ToString(); in GetGpuConvParams() 361 CHECK_EQ(dim.base_dilation(), 1) in GetGpuConvParams() 384 for (int dim = 0; dim < num_dimensions; ++dim) { in GetGpuConvParams() local 387 static_cast<DimIndex>(effective_num_dimensions - dim - 1), in GetGpuConvParams() 388 input_shape->dimensions(dnums.input_spatial_dimensions(dim))); in GetGpuConvParams() 398 for (int dim = 0; dim < num_dimensions; ++dim) { in GetGpuConvParams() local 400 static_cast<DimIndex>(effective_num_dimensions - dim - 1), in GetGpuConvParams() 401 filter_shape->dimensions(dnums.kernel_spatial_dimensions(dim))); in GetGpuConvParams() [all …]
|
/external/skia/src/codec/ |
D | SkScalingCodec.h | 20 SkISize dim = this->dimensions(); in onGetScaledDimensions() local 23 dim.fWidth = std::max(1, SkScalarRoundToInt(desiredScale * dim.fWidth)); in onGetScaledDimensions() 24 dim.fHeight = std::max(1, SkScalarRoundToInt(desiredScale * dim.fHeight)); in onGetScaledDimensions() 25 return dim; in onGetScaledDimensions() 29 SkISize dim = this->dimensions(); in onDimensionsSupported() local 32 return 1 <= w && w <= dim.width() && 1 <= h && h <= dim.height(); in onDimensionsSupported()
|
/external/eigen/unsupported/test/ |
D | cxx11_tensor_argmax.cpp | 166 for (int dim = 0; dim < 4; ++dim) { in test_argmax_dim() local 177 if (ix[dim] != 0) continue; in test_argmax_dim() 185 tensor_argmax = tensor.argmax(dim); in test_argmax_dim() 188 ptrdiff_t(2*3*5*7 / tensor.dimension(dim))); in test_argmax_dim() 199 if (ix[dim] != tensor.dimension(dim) - 1) continue; in test_argmax_dim() 207 tensor_argmax = tensor.argmax(dim); in test_argmax_dim() 210 ptrdiff_t(2*3*5*7 / tensor.dimension(dim))); in test_argmax_dim() 213 VERIFY_IS_EQUAL(tensor_argmax.data()[n], tensor.dimension(dim) - 1); in test_argmax_dim() 224 for (int dim = 0; dim < 4; ++dim) { in test_argmin_dim() local 235 if (ix[dim] != 0) continue; in test_argmin_dim() [all …]
|
/external/mesa3d/src/gallium/frontends/nine/ |
D | nine_ff.h | 85 unsigned dim = context->ff.tex_stage[s][D3DTSS_TEXTURETRANSFORMFLAGS] & 0x7; in nine_ff_get_projected_key() local 89 if (dim > 4) in nine_ff_get_projected_key() 90 dim = input_texture_coord[s]; in nine_ff_get_projected_key() 92 if (!dim && gen == NINED3DTSS_TCI_PASSTHRU) in nine_ff_get_projected_key() 93 dim = input_texture_coord[s]; in nine_ff_get_projected_key() 94 else if (!dim) in nine_ff_get_projected_key() 95 dim = 4; in nine_ff_get_projected_key() 97 if (dim == 1) /* NV behaviour */ in nine_ff_get_projected_key() 99 if (dim > input_texture_coord[s] && gen == NINED3DTSS_TCI_PASSTHRU) in nine_ff_get_projected_key() 102 dim = 4; in nine_ff_get_projected_key() [all …]
|
/external/angle/src/compiler/translator/ |
D | BuiltInFunctionEmulatorGLSL.cpp | 89 for (int dim = 2; dim <= 4; ++dim) in InitBuiltInAtanFunctionEmulatorForGLSLWorkarounds() local 92 ss << "emu_precision vec" << dim << " atan_emu(emu_precision vec" << dim in InitBuiltInAtanFunctionEmulatorForGLSLWorkarounds() 93 << " y, emu_precision vec" << dim << " x)\n" in InitBuiltInAtanFunctionEmulatorForGLSLWorkarounds() 96 << dim << "("; in InitBuiltInAtanFunctionEmulatorForGLSLWorkarounds() 97 for (int i = 0; i < dim; ++i) in InitBuiltInAtanFunctionEmulatorForGLSLWorkarounds() 100 if (i < dim - 1) in InitBuiltInAtanFunctionEmulatorForGLSLWorkarounds() 107 emu->addEmulatedFunctionWithDependency(BuiltInId::atan_Float1_Float1, ids[dim - 1], in InitBuiltInAtanFunctionEmulatorForGLSLWorkarounds()
|