/external/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Support/ |
D | MachineValueType.h | 918 static MVT getVectorVT(MVT VT, unsigned NumElements) { in getVectorVT() argument 923 if (NumElements == 1) return MVT::v1i1; in getVectorVT() 924 if (NumElements == 2) return MVT::v2i1; in getVectorVT() 925 if (NumElements == 4) return MVT::v4i1; in getVectorVT() 926 if (NumElements == 8) return MVT::v8i1; in getVectorVT() 927 if (NumElements == 16) return MVT::v16i1; in getVectorVT() 928 if (NumElements == 32) return MVT::v32i1; in getVectorVT() 929 if (NumElements == 64) return MVT::v64i1; in getVectorVT() 930 if (NumElements == 128) return MVT::v128i1; in getVectorVT() 931 if (NumElements == 256) return MVT::v256i1; in getVectorVT() [all …]
|
D | LowLevelTypeImpl.h | 57 static LLT vector(uint16_t NumElements, unsigned ScalarSizeInBits) { in vector() argument 58 assert(NumElements > 1 && "invalid number of vector elements"); in vector() 60 return LLT{/*isPointer=*/false, /*isVector=*/true, NumElements, in vector() 65 static LLT vector(uint16_t NumElements, LLT ScalarTy) { in vector() argument 66 assert(NumElements > 1 && "invalid number of vector elements"); in vector() 68 return LLT{ScalarTy.isPointer(), /*isVector=*/true, NumElements, in vector() 73 static LLT scalarOrVector(uint16_t NumElements, LLT ScalarTy) { in scalarOrVector() argument 74 return NumElements == 1 ? ScalarTy : LLT::vector(NumElements, ScalarTy); in scalarOrVector() 77 static LLT scalarOrVector(uint16_t NumElements, unsigned ScalarSize) { in scalarOrVector() argument 78 return scalarOrVector(NumElements, LLT::scalar(ScalarSize)); in scalarOrVector() [all …]
|
/external/llvm/include/llvm/CodeGen/ |
D | MachineValueType.h | 588 static MVT getVectorVT(MVT VT, unsigned NumElements) { in getVectorVT() argument 593 if (NumElements == 2) return MVT::v2i1; in getVectorVT() 594 if (NumElements == 4) return MVT::v4i1; in getVectorVT() 595 if (NumElements == 8) return MVT::v8i1; in getVectorVT() 596 if (NumElements == 16) return MVT::v16i1; in getVectorVT() 597 if (NumElements == 32) return MVT::v32i1; in getVectorVT() 598 if (NumElements == 64) return MVT::v64i1; in getVectorVT() 599 if (NumElements == 512) return MVT::v512i1; in getVectorVT() 600 if (NumElements == 1024) return MVT::v1024i1; in getVectorVT() 603 if (NumElements == 1) return MVT::v1i8; in getVectorVT() [all …]
|
/external/llvm/lib/Target/X86/ |
D | X86ShuffleDecodeConstantPool.cpp | 130 unsigned NumElements = MaskTySize / ElSize; in DecodeVPERMILPMask() local 131 assert((NumElements == 2 || NumElements == 4 || NumElements == 8 || in DecodeVPERMILPMask() 132 NumElements == 16) && in DecodeVPERMILPMask() 134 ShuffleMask.reserve(NumElements); in DecodeVPERMILPMask() 138 for (unsigned i = 0; i < NumElements; ++i) { in DecodeVPERMILPMask() 183 unsigned NumElements = MaskTySize / ElSize; in DecodeVPERMIL2PMask() local 184 assert((NumElements == 2 || NumElements == 4 || NumElements == 8) && in DecodeVPERMIL2PMask() 186 ShuffleMask.reserve(NumElements); in DecodeVPERMIL2PMask() 190 for (unsigned i = 0; i < NumElements; ++i) { in DecodeVPERMIL2PMask() 225 Index += Src * NumElements; in DecodeVPERMIL2PMask() [all …]
|
/external/swiftshader/third_party/subzero/crosstest/ |
D | test_vector_ops.cpp | 663 static typename std::enable_if<Vectors<Ty>::NumElements == 4, Ty>::type 665 const uint8_t NumElements = 4; in shufflevector() local 667 V1, V2, Idx0 % (NumElements * 2), Idx1 % (NumElements * 2), in shufflevector() 668 Idx2 % (NumElements * 2), Idx3 % (NumElements * 2)); in shufflevector() 679 static typename std::enable_if<Vectors<Ty>::NumElements == 8, Ty>::type 681 const uint8_t NumElements = 8; in shufflevector() local 683 V1, V2, Idx0 % (NumElements * 2), Idx1 % (NumElements * 2), in shufflevector() 684 Idx2 % (NumElements * 2), Idx3 % (NumElements * 2), in shufflevector() 685 Idx4 % (NumElements * 2), Idx5 % (NumElements * 2), in shufflevector() 686 Idx6 % (NumElements * 2), Idx7 % (NumElements * 2)); in shufflevector() [all …]
|
D | test_select_main.cpp | 42 static const size_t NumElements = Vectors<T>::NumElements; in testSelect() local 47 for (size_t j = 0; j < NumElements; ++j) { in testSelect() 78 static const size_t NumElements = 4; in testSelect() local 83 for (size_t j = 0; j < NumElements; ++j) { in testSelect() 108 static const size_t NumElements = Vectors<T>::NumElements; in testSelectI1() local 113 for (size_t j = 0; j < NumElements; ++j) { in testSelectI1()
|
D | test_icmp_main.cpp | 212 const static size_t NumElementsInType = Vectors<TypeUnsigned>::NumElements; in testsVecInt() 247 const static size_t NumElements = Vectors<T>::NumElements; in incrementI1Vector() local 248 for (Pos = 0; Pos < NumElements; ++Pos) { in incrementI1Vector() 255 return (Pos == NumElements); in incrementI1Vector() 271 const static size_t NumElements = Vectors<T>::NumElements; in testsVecI1() local 275 if (NumElements <= MAX_NUMBER_OF_ELEMENTS_FOR_EXHAUSTIVE_TESTING) { in testsVecI1() 306 for (size_t j = 0; j < NumElements; ++j) { in testsVecI1()
|
/external/tensorflow/tensorflow/lite/toco/ |
D | tooling_util_test.cc | 110 status = NumElements(std::vector<int>{1024, 1024, 2047}, &count); in TEST() 114 status = NumElements(std::vector<int>{1024, 0, 2048}, &count); in TEST() 118 status = NumElements(std::vector<int>{1, 2, -3}, &count); in TEST() 121 status = NumElements(std::vector<int>{1024, 1024, 2048}, &count); in TEST() 129 status = NumElements(std::vector<int32_t>{1024, 1024, 2047}, &count); in TEST() 133 status = NumElements(std::vector<int32_t>{1, 2, -3}, &count); in TEST() 136 status = NumElements(std::vector<int32_t>{1024, 1024, 2048}, &count); in TEST() 144 status = NumElements(std::vector<int64_t>{16777216, 16777216, 32767}, &count); in TEST() 148 status = NumElements(std::vector<int64_t>{1, 2, -3}, &count); in TEST() 151 status = NumElements(std::vector<int64_t>{16777216, 16777216, 32768}, &count); in TEST() [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | save_restore_v2_ops.cc | 47 const int num_tensors = static_cast<int>(tensor_names.NumElements()); in ValidateInputs() 49 context, prefix.NumElements() == 1, in ValidateInputs() 51 prefix.NumElements(), " instead.")); in ValidateInputs() 61 tensor_names.NumElements() == shape_and_slices.NumElements(), in ValidateInputs() 64 tensor_names.NumElements(), " vs. ", in ValidateInputs() 65 shape_and_slices.NumElements())); in ValidateInputs() 67 FastBoundsCheck(tensor_names.NumElements() + kFixedInputs, in ValidateInputs() 71 context, shape_and_slices.NumElements() == num_tensors, in ValidateInputs() 74 context->input(2).NumElements())); in ValidateInputs() 106 const int num_tensors = static_cast<int>(tensor_names.NumElements()); in Compute() [all …]
|
D | unravel_index_op.cc | 93 int64_t size = indices_tensor.NumElements(); in Compute() 104 TensorShape({dims_tensor.NumElements()}), in Compute() 115 TensorShape({dims_tensor.NumElements()}), in Compute() 126 ctx, ctx->allocate_output(0, TensorShape({dims_tensor.NumElements()}), in Compute() 136 TensorShape({dims_tensor.NumElements(), in Compute() 137 indices_tensor.NumElements()}), in Compute() 143 {static_cast<Eigen::Index>(dims_tensor.NumElements()), 1}}; in Compute() 145 {1, static_cast<Eigen::Index>(indices_tensor.NumElements())}); in Compute() 147 {1, static_cast<Eigen::Index>(indices_tensor.NumElements())}}; in Compute() 149 {static_cast<Eigen::Index>(dims_tensor.NumElements()), 1}); in Compute()
|
D | deserialize_sparse_variant_op.cc | 69 output_shape->NumElements(), &output_indices, in Compute() 108 2, {input_dims_to_stack + element_shape->NumElements()}, in Compute() 116 for (int64_t j = 0; j < element_shape->NumElements(); ++j) { in Compute() 136 output_shape->NumElements() - input_dims_to_stack == in Compute() 137 element_shape->NumElements(), in Compute() 141 i, "] was: ", output_shape->NumElements() - input_dims_to_stack, in Compute() 143 "] is: ", element_shape->NumElements())); in Compute() 145 for (int j = 0; j < element_shape->NumElements(); ++j) { in Compute() 155 const int output_rank = output_shape->NumElements(); in Compute() 180 const size_t num_index_rows = element_values->NumElements(); in Compute() [all …]
|
D | list_kernels.h | 223 if (output->NumElements() == 0) { in Compute() 233 t.shaped<T, 2>({1, t.NumElements()}))); in Compute() 235 if (!zeros.NumElements()) { in Compute() 246 {1, zeros.NumElements()}))); in Compute() 249 auto output_flat = output->shaped<T, 2>({1, output->NumElements()}); in Compute() 501 if ((c->num_inputs() <= 2 || i >= c->input(2).NumElements()) && in Compute() 511 OP_REQUIRES(c, i < c->input(2).NumElements(), in Compute() 515 c->input(2).NumElements(), " elements.")); in Compute() 527 if (output->NumElements() == 0) { in Compute() 539 if (element_tensor.NumElements() > 0) { in Compute() [all …]
|
D | quantized_batch_norm_op.cc | 177 OP_REQUIRES(context, input_min_tensor.NumElements() == 1, in Compute() 181 OP_REQUIRES(context, input_max_tensor.NumElements() == 1, in Compute() 186 OP_REQUIRES(context, mean_min_tensor.NumElements() == 1, in Compute() 190 OP_REQUIRES(context, mean_max_tensor.NumElements() == 1, in Compute() 195 OP_REQUIRES(context, var_min_tensor.NumElements() == 1, in Compute() 199 OP_REQUIRES(context, var_max_tensor.NumElements() == 1, in Compute() 204 OP_REQUIRES(context, beta_min_tensor.NumElements() == 1, in Compute() 208 OP_REQUIRES(context, beta_max_tensor.NumElements() == 1, in Compute() 213 OP_REQUIRES(context, gamma_min_tensor.NumElements() == 1, in Compute() 217 OP_REQUIRES(context, gamma_max_tensor.NumElements() == 1, in Compute() [all …]
|
D | edit_distance_op.cc | 67 if (hypothesis_values.NumElements() != hypothesis_indices.dim_size(0)) in ValidateShapes() 73 if (hypothesis_shape.NumElements() != hypothesis_indices.dim_size(1)) in ValidateShapes() 79 if (truth_shape.NumElements() < 2) in ValidateShapes() 83 truth_shape.NumElements()); in ValidateShapes() 84 if (truth_values.NumElements() != truth_indices.dim_size(0)) in ValidateShapes() 90 if (truth_shape.NumElements() != truth_indices.dim_size(1)) in ValidateShapes() 96 if (truth_shape.NumElements() != hypothesis_shape.NumElements()) in ValidateShapes() 138 hypothesis_shape->NumElements(), &hypothesis_st_shape)); in Compute() 142 truth_shape->NumElements(), &truth_st_shape)); in Compute()
|
D | sparse_slice_op.cc | 40 const int input_dims = input_shape.NumElements(); in operator ()() 126 const int input_dims = input_shape.NumElements(); in SparseSliceOpImpl() 127 OP_REQUIRES_ASYNC(context, input_dims == input_start.NumElements(), in SparseSliceOpImpl() 130 " but got length ", input_start.NumElements()), in SparseSliceOpImpl() 133 OP_REQUIRES_ASYNC(context, input_dims == input_size.NumElements(), in SparseSliceOpImpl() 136 " but got length ", input_size.NumElements()), in SparseSliceOpImpl()
|
D | reshape_op.h | 49 context, sizes.NumElements() < TensorShape::MaxDimensions(), in Compute() 52 sizes.NumElements())); in Compute() 102 OP_REQUIRES(context, shape.num_elements() == input.NumElements(), in Compute() 104 input.NumElements(), in Compute() 124 const int64_t num_dims = sizes.NumElements(); in ValidateSizes()
|
D | save_restore_tensor.cc | 49 const int64_t size = filename_t.NumElements(); in SaveTensors() 61 FastBoundsCheck(tensor_names_t.NumElements() + kFixedInputs, in SaveTensors() 64 const int N = static_cast<int>(tensor_names_t.NumElements()); in SaveTensors() 70 tensor_shapes_and_slices_t.NumElements() == static_cast<int64_t>(N), in SaveTensors() 74 tensor_shapes_and_slices_t.NumElements())); in SaveTensors() 155 const int64_t size = file_pattern_t.NumElements(); in RestoreTensor() 166 const int64_t size = tensor_name_t.NumElements(); in RestoreTensor() 337 for (int i = 0; i < restored_tensor->NumElements(); ++i) { in run() 343 << avg / restored_tensor->NumElements() << " total elts " in run() 344 << restored_tensor->NumElements(); in run()
|
D | quantized_mul_op.cc | 331 if (x.NumElements() == 1) { in Compute() 332 ScalarMultiply<T, Toutput>(context, y_data, offset_y, y.NumElements(), in Compute() 334 } else if (y.NumElements() == 1) { in Compute() 335 ScalarMultiply<T, Toutput>(context, x_data, offset_x, x.NumElements(), in Compute() 339 x.NumElements(), z_data); in Compute() 348 if (x.NumElements() < y.NumElements()) { in Compute() 350 vector_num_elements = x.NumElements(); in Compute() 353 tensor_num_elements = y.NumElements(); in Compute() 357 vector_num_elements = y.NumElements(); in Compute() 360 tensor_num_elements = x.NumElements(); in Compute()
|
/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_quantized_conv_ops.h | 63 DCHECK(min_b_vector.NumElements() == (*min_c_vector)->NumElements()); in MklQuantizationRangeForMultiplication() 64 DCHECK(max_b_vector.NumElements() == (*max_c_vector)->NumElements()); in MklQuantizationRangeForMultiplication() 65 size_t n_channel = min_b_vector.NumElements(); in MklQuantizationRangeForMultiplication()
|
/external/tensorflow/tensorflow/dtensor/cc/ |
D | dtensor_sharded_prefix_kernel.cc | 52 for (int i = 0; i < shape_and_slices->NumElements(); ++i) { in Compute() 67 OP_REQUIRES(ctx, tensor_names->NumElements() == layouts->NumElements(), in Compute() 71 tensor_names->NumElements(), in Compute() 72 " and layouts size : ", layouts->NumElements()))); in Compute() 77 const int num_tensors = static_cast<int>(tensor_names->NumElements()); in Compute()
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | call_once_test.cc | 55 ASSERT_EQ(NumElements(output), 1); in TEST_F() 69 ASSERT_EQ(NumElements(output), 1); in TEST_F() 84 ASSERT_EQ(NumElements(output), 1); in TEST_F() 96 ASSERT_EQ(NumElements(output), 1); in TEST_F() 111 ASSERT_EQ(NumElements(output), 1); in TEST_F()
|
/external/mesa3d/src/gallium/drivers/swr/rasterizer/jitter/ |
D | jit_pch.hpp | 142 static inline llvm::VectorType* getVectorType(llvm::Type *ElementType, unsigned NumElements) in getVectorType() argument 144 return llvm::VectorType::get(ElementType, NumElements, false); in getVectorType() 147 static inline llvm::VectorType* getVectorType(llvm::Type *ElementType, unsigned NumElements) in getVectorType() argument 149 return llvm::VectorType::get(ElementType, NumElements); in getVectorType()
|
/external/llvm/include/llvm/DebugInfo/CodeView/ |
D | StreamReader.h | 57 Error readArray(ArrayRef<T> &Array, uint32_t NumElements) { in readArray() argument 59 if (NumElements == 0) { in readArray() 64 if (NumElements > UINT32_MAX/sizeof(T)) in readArray() 67 if (auto EC = readBytes(Bytes, NumElements * sizeof(T))) in readArray() 69 Array = ArrayRef<T>(reinterpret_cast<const T *>(Bytes.data()), NumElements); in readArray()
|
/external/tensorflow/tensorflow/core/tfrt/utils/ |
D | fallback_tensor_test.cc | 30 ASSERT_EQ(immutable_tensor.tensor().NumElements(), 1); in TEST() 44 ASSERT_EQ(immutable_tensor.tensor().NumElements(), 1); in TEST() 60 ASSERT_EQ(fallback_tensor.tensor().NumElements(), 1); in TEST() 72 ASSERT_EQ(fallback_tensor.tensor().NumElements(), 1); in TEST()
|
/external/tensorflow/tensorflow/c/experimental/saved_model/core/ |
D | constant_loading_test.cc | 61 testing::FillNumericTensorBuffer(expected.dtype(), expected.NumElements(), in TEST_P() 83 EXPECT_EQ(revived_tensor->NumElements(), expected.NumElements()); in TEST_P() 89 testing::CheckBufferDataIsEqual(expected.dtype(), expected.NumElements(), in TEST_P()
|