/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorDimensions.h | 92 template <typename std::ptrdiff_t... Indices> 93 struct Sizes : internal::numeric_list<std::ptrdiff_t, Indices...> { 94 typedef internal::numeric_list<std::ptrdiff_t, Indices...> Base; 95 static const std::ptrdiff_t total_size = internal::arg_prod(Indices...); 102 return internal::arg_prod(Indices...); 137 template <typename std::ptrdiff_t... Indices> 138 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE std::ptrdiff_t array_prod(const Sizes<Indices...>&) { 139 return Sizes<Indices...>::total_size; 371 template <typename std::ptrdiff_t... Indices> struct array_size<const Sizes<Indices...> > { 372 static const std::ptrdiff_t value = Sizes<Indices...>::count; [all …]
|
D | TensorContraction.h | 63 typedef Indices_ Indices; 74 template<typename Indices, typename LhsXprType, typename RhsXprType> 75 class TensorContractionOp : public TensorBase<TensorContractionOp<Indices, LhsXprType, RhsXprType>,… 86 const LhsXprType& lhs, const RhsXprType& rhs, const Indices& dims) 90 const Indices& indices() const { return m_indices; } 104 const Indices m_indices; 111 typedef typename internal::traits<Derived>::Indices Indices; 116 typedef TensorContractionOp<Indices, LeftArgType, RightArgType> XprType; 143 static const int ContractDims = internal::array_size<Indices>::value; 571 template<typename Indices, typename LeftArgType, typename RightArgType, typename Device> [all …]
|
D | TensorConvolution.h | 256 template<typename Indices, typename InputXprType, typename KernelXprType> 257 class TensorConvolutionOp : public TensorBase<TensorConvolutionOp<Indices, InputXprType, KernelXprT… 268 …NE TensorConvolutionOp(const InputXprType& input, const KernelXprType& kernel, const Indices& dims) 272 const Indices& indices() const { return m_indices; } 286 const Indices m_indices; 290 template<typename Indices, typename InputArgType, typename KernelArgType, typename Device> 291 struct TensorEvaluator<const TensorConvolutionOp<Indices, InputArgType, KernelArgType>, Device> 293 typedef TensorConvolutionOp<Indices, InputArgType, KernelArgType> XprType; 296 static const int NumKernelDims = internal::array_size<Indices>::value; 753 template<typename Indices, typename InputArgType, typename KernelArgType> [all …]
|
/external/swiftshader/third_party/LLVM/include/llvm/CodeGen/ |
D | Analysis.h | 37 const unsigned *Indices, 42 ArrayRef<unsigned> Indices, 44 return ComputeLinearIndex(Ty, Indices.begin(), Indices.end(), CurIndex);
|
/external/llvm/lib/CodeGen/ |
D | InterleavedAccessPass.cpp | 231 SmallVector<unsigned, 4> Indices; in lowerInterleavedLoad() local 232 Indices.push_back(Index); in lowerInterleavedLoad() 246 Indices.push_back(Index); in lowerInterleavedLoad() 257 if (!TLI->lowerInterleavedLoad(LI, Shuffles, Indices, Factor)) in lowerInterleavedLoad() 299 SmallVector<int, 4> Indices; in tryReplaceExtracts() local 300 Shuffle->getShuffleMask(Indices); in tryReplaceExtracts() 301 for (unsigned I = 0; I < Indices.size(); ++I) in tryReplaceExtracts() 302 if (Indices[I] == Index) { in tryReplaceExtracts()
|
D | Analysis.cpp | 38 const unsigned *Indices, in ComputeLinearIndex() argument 42 if (Indices && Indices == IndicesEnd) in ComputeLinearIndex() 51 if (Indices && *Indices == unsigned(EI - EB)) in ComputeLinearIndex() 52 return ComputeLinearIndex(*EI, Indices+1, IndicesEnd, CurIndex); in ComputeLinearIndex() 55 assert(!Indices && "Unexpected out of bound"); in ComputeLinearIndex() 64 if (Indices) { in ComputeLinearIndex() 65 assert(*Indices < NumElts && "Unexpected out of bound"); in ComputeLinearIndex() 68 CurIndex += EltLinearOffset* *Indices; in ComputeLinearIndex() 69 return ComputeLinearIndex(EltTy, Indices+1, IndicesEnd, CurIndex); in ComputeLinearIndex()
|
/external/llvm/include/llvm/CodeGen/ |
D | Analysis.h | 55 const unsigned *Indices, 60 ArrayRef<unsigned> Indices, 62 return ComputeLinearIndex(Ty, Indices.begin(), Indices.end(), CurIndex);
|
/external/swiftshader/third_party/LLVM/lib/CodeGen/ |
D | Analysis.cpp | 35 const unsigned *Indices, in ComputeLinearIndex() argument 39 if (Indices && Indices == IndicesEnd) in ComputeLinearIndex() 48 if (Indices && *Indices == unsigned(EI - EB)) in ComputeLinearIndex() 49 return ComputeLinearIndex(*EI, Indices+1, IndicesEnd, CurIndex); in ComputeLinearIndex() 58 if (Indices && *Indices == i) in ComputeLinearIndex() 59 return ComputeLinearIndex(EltTy, Indices+1, IndicesEnd, CurIndex); in ComputeLinearIndex()
|
D | ShadowStackGC.cpp | 352 Value *Indices[] = { ConstantInt::get(Type::getInt32Ty(Context), 0), in CreateGEP() local 355 Value* Val = B.CreateGEP(BasePtr, Indices, Name); in CreateGEP() 365 Value *Indices[] = { ConstantInt::get(Type::getInt32Ty(Context), 0), in CreateGEP() local 367 Value *Val = B.CreateGEP(BasePtr, Indices, Name); in CreateGEP()
|
/external/clang/test/CXX/temp/temp.decls/temp.variadic/ |
D | metafunctions.cpp | 149 namespace Indices { namespace 153 template<unsigned I, unsigned N, int ...Indices> 154 struct build_indices_impl<I, N, int_tuple<Indices...> > 155 : build_indices_impl<I+1, N, int_tuple<Indices..., I> > { 158 template<unsigned N, int ...Indices> 159 struct build_indices_impl<N, N, int_tuple<Indices...> > { 160 typedef int_tuple<Indices...> type;
|
/external/clang/lib/Format/ |
D | SortJavaScriptImports.cpp | 143 SmallVector<unsigned, 16> Indices; in analyze() local 145 Indices.push_back(i); in analyze() 146 std::stable_sort(Indices.begin(), Indices.end(), in analyze() 150 bool ReferencesInOrder = std::is_sorted(Indices.begin(), Indices.end()); in analyze() 154 for (unsigned i = 0, e = Indices.size(); i != e; ++i) { in analyze() 155 JsModuleReference Reference = References[Indices[i]]; in analyze() 164 (Reference.IsExport != References[Indices[i + 1]].IsExport || in analyze() 165 Reference.Category != References[Indices[i + 1]].Category)) in analyze()
|
/external/swiftshader/third_party/LLVM/lib/Transforms/IPO/ |
D | ArgumentPromotion.cpp | 256 static bool PrefixIn(const ArgPromotion::IndicesVector &Indices, in PrefixIn() argument 259 Low = Set.upper_bound(Indices); in PrefixIn() 267 return Low != Set.end() && IsPrefix(*Low, Indices); in PrefixIn() 347 IndicesVector Indices; in isSafeToPromoteArgument() local 356 Indices.reserve(GEP->getNumIndices()); in isSafeToPromoteArgument() 360 Indices.push_back(CI->getSExtValue()); in isSafeToPromoteArgument() 367 MarkIndicesSafe(Indices, SafeToUnconditionallyLoad); in isSafeToPromoteArgument() 368 Indices.clear(); in isSafeToPromoteArgument() 556 IndicesVector Indices; in DoPromotion() local 557 Indices.reserve(User->getNumOperands() - 1); in DoPromotion() [all …]
|
/external/llvm/lib/Transforms/IPO/ |
D | ArgumentPromotion.cpp | 403 static bool PrefixIn(const IndicesVector &Indices, in PrefixIn() argument 406 Low = Set.upper_bound(Indices); in PrefixIn() 414 return Low != Set.end() && IsPrefix(*Low, Indices); in PrefixIn() 498 IndicesVector Indices; in isSafeToPromoteArgument() local 506 Indices.reserve(GEP->getNumIndices()); in isSafeToPromoteArgument() 510 Indices.push_back(CI->getSExtValue()); in isSafeToPromoteArgument() 517 MarkIndicesSafe(Indices, SafeToUnconditionallyLoad); in isSafeToPromoteArgument() 518 Indices.clear(); in isSafeToPromoteArgument() 707 IndicesVector Indices; in DoPromotion() local 708 Indices.reserve(UI->getNumOperands() - 1); in DoPromotion() [all …]
|
/external/llvm/lib/Target/Mips/ |
D | MipsSEISelLowering.cpp | 2442 SmallVector<int, 16> Indices, in lowerVECTOR_SHUFFLE_SHF() argument 2446 if (Indices.size() < 4) in lowerVECTOR_SHUFFLE_SHF() 2450 for (unsigned j = i; j < Indices.size(); j += 4) { in lowerVECTOR_SHUFFLE_SHF() 2451 int Idx = Indices[j]; in lowerVECTOR_SHUFFLE_SHF() 2522 SmallVector<int, 16> Indices, in isVECTOR_SHUFFLE_SPLATI() argument 2524 assert((Indices.size() % 2) == 0); in isVECTOR_SHUFFLE_SPLATI() 2527 for (const auto &V : Indices) { in isVECTOR_SHUFFLE_SPLATI() 2534 return fitsRegularPattern<int>(Indices.begin(), 1, Indices.end(), SplatIndex, in isVECTOR_SHUFFLE_SPLATI() 2554 SmallVector<int, 16> Indices, in lowerVECTOR_SHUFFLE_ILVEV() argument 2556 assert((Indices.size() % 2) == 0); in lowerVECTOR_SHUFFLE_ILVEV() [all …]
|
/external/swiftshader/third_party/LLVM/lib/Transforms/Instrumentation/ |
D | ProfilingUtils.cpp | 118 std::vector<Constant*> Indices(2); in IncrementCounterInBlock() local 119 Indices[0] = Constant::getNullValue(Type::getInt32Ty(Context)); in IncrementCounterInBlock() 120 Indices[1] = ConstantInt::get(Type::getInt32Ty(Context), CounterNum); in IncrementCounterInBlock() 122 ConstantExpr::getGetElementPtr(CounterArray, Indices); in IncrementCounterInBlock()
|
/external/llvm/lib/Target/WebAssembly/ |
D | WebAssemblyFixIrreducibleControlFlow.cpp | 198 DenseMap<MachineBasicBlock *, unsigned> Indices; in VisitLoop() local 203 auto Pair = Indices.insert(std::make_pair(MBB, 0)); in VisitLoop() 230 if (!Indices.count(Succ)) in VisitLoop() 243 .addImm(Indices[Succ]); in VisitLoop() 252 if (Op.isMBB() && Indices.count(Op.getMBB())) in VisitLoop()
|
D | WebAssemblyStoreResults.cpp | 91 SmallVector<SlotIndex, 4> Indices; in ReplaceDominatedUses() local 121 Indices.push_back(WhereIdx.getRegSlot()); in ReplaceDominatedUses() 127 LIS.extendToIndices(*ToLI, Indices); in ReplaceDominatedUses()
|
/external/swiftshader/third_party/LLVM/lib/Target/ |
D | TargetData.cpp | 533 ArrayRef<Value *> Indices) const { in getIndexedOffset() 539 TI = gep_type_begin(ptrTy, Indices); in getIndexedOffset() 540 for (unsigned CurIDX = 0, EndIDX = Indices.size(); CurIDX != EndIDX; in getIndexedOffset() 543 assert(Indices[CurIDX]->getType() == in getIndexedOffset() 546 unsigned FieldNo = cast<ConstantInt>(Indices[CurIDX])->getZExtValue(); in getIndexedOffset() 561 if (int64_t arrayIdx = cast<ConstantInt>(Indices[CurIDX])->getSExtValue()) in getIndexedOffset()
|
/external/llvm/lib/Target/NVPTX/ |
D | NVPTXFavorNonGenericAddrSpaces.cpp | 159 SmallVector<Value *, 8> Indices(GEP->idx_begin(), GEP->idx_end()); in hoistAddrSpaceCastFromGEP() local 167 GEP->getSourceElementType(), Cast->getOperand(0), Indices, in hoistAddrSpaceCastFromGEP() 180 Indices, GEP->isInBounds()); in hoistAddrSpaceCastFromGEP()
|
/external/swiftshader/third_party/LLVM/utils/TableGen/ |
D | CodeGenRegisters.cpp | 56 std::vector<Record*> Indices = TheDef->getValueAsListOfDefs("SubRegIndices"); in getSubRegs() local 57 if (SubList.size() != Indices.size()) in getSubRegs() 64 if (!SubRegs.insert(std::make_pair(Indices[i], SR)).second) in getSubRegs() 65 throw TGError(TheDef->getLoc(), "SubRegIndex " + Indices[i]->getName() + in getSubRegs() 86 Orphans.push_back(Orphan(SI->second, Indices[i], SI->first)); in getSubRegs() 147 std::vector<Record*> Indices = TheDef->getValueAsListOfDefs("SubRegIndices"); in addSubRegsPreOrder() local 148 for (unsigned i = 0, e = Indices.size(); i != e; ++i) { in addSubRegsPreOrder() 149 CodeGenRegister *SR = SubRegs.find(Indices[i])->second; in addSubRegsPreOrder() 165 std::vector<Record*> Indices = Def->getValueAsListOfDefs("SubRegIndices"); in expand() local 166 unsigned Dim = Indices.size(); in expand()
|
/external/llvm/lib/Transforms/Scalar/ |
D | SROA.cpp | 1313 SmallVectorImpl<Value *> &Indices, Twine NamePrefix) { in buildGEP() argument 1314 if (Indices.empty()) in buildGEP() 1319 if (Indices.size() == 1 && cast<ConstantInt>(Indices.back())->isZero()) in buildGEP() 1322 return IRB.CreateInBoundsGEP(nullptr, BasePtr, Indices, in buildGEP() 1337 SmallVectorImpl<Value *> &Indices, in getNaturalGEPWithType() argument 1340 return buildGEP(IRB, BasePtr, Indices, NamePrefix); in getNaturalGEPWithType() 1355 Indices.push_back(IRB.getIntN(PtrSize, 0)); in getNaturalGEPWithType() 1358 Indices.push_back(IRB.getInt32(0)); in getNaturalGEPWithType() 1363 Indices.push_back(IRB.getInt32(0)); in getNaturalGEPWithType() 1370 Indices.erase(Indices.end() - NumLayers, Indices.end()); in getNaturalGEPWithType() [all …]
|
/external/swiftshader/third_party/LLVM/bindings/ocaml/llvm/ |
D | llvm_ocaml.c | 710 CAMLprim LLVMValueRef llvm_const_gep(LLVMValueRef ConstantVal, value Indices) { in llvm_const_gep() argument 711 return LLVMConstGEP(ConstantVal, (LLVMValueRef*) Op_val(Indices), in llvm_const_gep() 712 Wosize_val(Indices)); in llvm_const_gep() 717 value Indices) { in llvm_const_in_bounds_gep() argument 718 return LLVMConstInBoundsGEP(ConstantVal, (LLVMValueRef*) Op_val(Indices), in llvm_const_in_bounds_gep() 719 Wosize_val(Indices)); in llvm_const_in_bounds_gep() 724 value Indices) { in llvm_const_extractvalue() argument 725 CAMLparam1(Indices); in llvm_const_extractvalue() 726 int size = Wosize_val(Indices); in llvm_const_extractvalue() 732 idxs[i] = Int_val(Field(Indices, i)); in llvm_const_extractvalue() [all …]
|
/external/llvm/lib/Analysis/ |
D | TypeMetadataUtils.cpp | 54 SmallVector<Value *, 8> Indices(GEP->op_begin() + 1, GEP->op_end()); in findLoadCallsAtConstantOffset() local 56 GEP->getSourceElementType(), Indices); in findLoadCallsAtConstantOffset()
|
/external/llvm/bindings/ocaml/llvm/ |
D | llvm_ocaml.c | 919 CAMLprim LLVMValueRef llvm_const_gep(LLVMValueRef ConstantVal, value Indices) { in llvm_const_gep() argument 920 return LLVMConstGEP(ConstantVal, (LLVMValueRef*) Op_val(Indices), in llvm_const_gep() 921 Wosize_val(Indices)); in llvm_const_gep() 926 value Indices) { in llvm_const_in_bounds_gep() argument 927 return LLVMConstInBoundsGEP(ConstantVal, (LLVMValueRef*) Op_val(Indices), in llvm_const_in_bounds_gep() 928 Wosize_val(Indices)); in llvm_const_in_bounds_gep() 939 value Indices) { in llvm_const_extractvalue() argument 940 CAMLparam1(Indices); in llvm_const_extractvalue() 941 int size = Wosize_val(Indices); in llvm_const_extractvalue() 947 idxs[i] = Int_val(Field(Indices, i)); in llvm_const_extractvalue() [all …]
|
/external/llvm/lib/DebugInfo/CodeView/ |
D | TypeRecord.cpp | 79 ArrayRef<TypeIndex> Indices; in deserialize() local 80 CV_DESERIALIZE(Data, L, CV_ARRAY_FIELD_N(Indices, L->NumArgs)); in deserialize() 81 return ArgListRecord(Kind, Indices); in deserialize() 246 ArrayRef<TypeIndex> Indices; in deserialize() local 247 CV_DESERIALIZE(Data, L, CV_ARRAY_FIELD_N(Indices, L->NumArgs)); in deserialize() 248 return BuildInfoRecord(Indices); in deserialize()
|