/external/skia/include/private/ |
D | SkNx_neon.h | 19 AI static float32x4_t emulate_vrndmq_f32(float32x4_t v) { in emulate_vrndmq_f32() 24 AI static float32x2_t emulate_vrndm_f32(float32x2_t v) { in emulate_vrndm_f32() 33 AI SkNx(float32x2_t vec) : fVec(vec) {} in SkNx() 35 AI SkNx() {} in SkNx() 36 AI SkNx(float val) : fVec(vdup_n_f32(val)) {} in SkNx() 37 AI SkNx(float a, float b) { fVec = (float32x2_t) { a, b }; } in SkNx() 39 AI static SkNx Load(const void* ptr) { return vld1_f32((const float*)ptr); } in Load() 40 AI void store(void* ptr) const { vst1_f32((float*)ptr, fVec); } in store() 42 AI static void Load2(const void* ptr, SkNx* x, SkNx* y) { in Load2() 48 AI static void Store2(void* dst, const SkNx& a, const SkNx& b) { in Store2() [all …]
|
D | SkNx_sse.h | 31 AI static __m128 emulate_mm_floor_ps(__m128 v) { in emulate_mm_floor_ps() 40 AI SkNx(const __m128& vec) : fVec(vec) {} in SkNx() 42 AI SkNx() {} in SkNx() 43 AI SkNx(float val) : fVec(_mm_set1_ps(val)) {} in SkNx() 44 AI static SkNx Load(const void* ptr) { in Load() 47 AI SkNx(float a, float b) : fVec(_mm_setr_ps(a,b,0,0)) {} in SkNx() 49 AI void store(void* ptr) const { _mm_storel_pi((__m64*)ptr, fVec); } in store() 51 AI static void Load2(const void* ptr, SkNx* x, SkNx* y) { in Load2() 57 AI static void Store2(void* dst, const SkNx& a, const SkNx& b) { in Store2() 62 AI static void Store3(void* dst, const SkNx& a, const SkNx& b, const SkNx& c) { in Store3() [all …]
|
D | SkNx.h | 20 #define AI SK_ALWAYS_INLINE macro 31 AI SkNx() = default; 32 AI SkNx(const Half& lo, const Half& hi) : fLo(lo), fHi(hi) {} in SkNx() 34 AI SkNx(T v) : fLo(v), fHi(v) {} in SkNx() 36 AI SkNx(T a, T b) : fLo(a) , fHi(b) { static_assert(N==2, ""); } in SkNx() 37 AI SkNx(T a, T b, T c, T d) : fLo(a,b), fHi(c,d) { static_assert(N==4, ""); } in SkNx() 38 AI SkNx(T a, T b, T c, T d, T e, T f, T g, T h) : fLo(a,b,c,d), fHi(e,f,g,h) { in SkNx() 41 AI SkNx(T a, T b, T c, T d, T e, T f, T g, T h, in SkNx() 47 AI T operator[](int k) const { 52 AI static SkNx Load(const void* vptr) { in Load() [all …]
|
/external/skqp/include/private/ |
D | SkNx_neon.h | 19 AI static float32x4_t emulate_vrndmq_f32(float32x4_t v) { in emulate_vrndmq_f32() 24 AI static float32x2_t emulate_vrndm_f32(float32x2_t v) { in emulate_vrndm_f32() 33 AI SkNx(float32x2_t vec) : fVec(vec) {} in SkNx() 35 AI SkNx() {} in SkNx() 36 AI SkNx(float val) : fVec(vdup_n_f32(val)) {} in SkNx() 37 AI SkNx(float a, float b) { fVec = (float32x2_t) { a, b }; } in SkNx() 39 AI static SkNx Load(const void* ptr) { return vld1_f32((const float*)ptr); } in Load() 40 AI void store(void* ptr) const { vst1_f32((float*)ptr, fVec); } in store() 42 AI static void Load2(const void* ptr, SkNx* x, SkNx* y) { in Load2() 48 AI static void Store2(void* dst, const SkNx& a, const SkNx& b) { in Store2() [all …]
|
D | SkNx_sse.h | 31 AI static __m128 emulate_mm_floor_ps(__m128 v) { in emulate_mm_floor_ps() 40 AI SkNx(const __m128& vec) : fVec(vec) {} in SkNx() 42 AI SkNx() {} in SkNx() 43 AI SkNx(float val) : fVec(_mm_set1_ps(val)) {} in SkNx() 44 AI static SkNx Load(const void* ptr) { in Load() 47 AI SkNx(float a, float b) : fVec(_mm_setr_ps(a,b,0,0)) {} in SkNx() 49 AI void store(void* ptr) const { _mm_storel_pi((__m64*)ptr, fVec); } in store() 51 AI static void Load2(const void* ptr, SkNx* x, SkNx* y) { in Load2() 57 AI static void Store2(void* dst, const SkNx& a, const SkNx& b) { in Store2() 62 AI static void Store3(void* dst, const SkNx& a, const SkNx& b, const SkNx& c) { in Store3() [all …]
|
D | SkNx.h | 20 #define AI SK_ALWAYS_INLINE macro 31 AI SkNx() = default; 32 AI SkNx(const Half& lo, const Half& hi) : fLo(lo), fHi(hi) {} in SkNx() 34 AI SkNx(T v) : fLo(v), fHi(v) {} in SkNx() 36 AI SkNx(T a, T b) : fLo(a) , fHi(b) { static_assert(N==2, ""); } in SkNx() 37 AI SkNx(T a, T b, T c, T d) : fLo(a,b), fHi(c,d) { static_assert(N==4, ""); } in SkNx() 38 AI SkNx(T a, T b, T c, T d, T e, T f, T g, T h) : fLo(a,b,c,d), fHi(e,f,g,h) { in SkNx() 41 AI SkNx(T a, T b, T c, T d, T e, T f, T g, T h, in SkNx() 47 AI T operator[](int k) const { 52 AI static SkNx Load(const void* vptr) { in Load() [all …]
|
/external/harfbuzz_ng/src/ |
D | hb-atomic.hh | 59 #define hb_atomic_int_impl_add(AI, V) __atomic_fetch_add ((AI), (V), __ATOMIC_ACQ_REL) argument 60 #define hb_atomic_int_impl_set_relaxed(AI, V) __atomic_store_n ((AI), (V), __ATOMIC_RELAXED) argument 61 #define hb_atomic_int_impl_set(AI, V) __atomic_store_n ((AI), (V), __ATOMIC_RELEASE) argument 62 #define hb_atomic_int_impl_get_relaxed(AI) __atomic_load_n ((AI), __ATOMIC_RELAXED) argument 63 #define hb_atomic_int_impl_get(AI) __atomic_load_n ((AI), __ATOMIC_ACQUIRE) argument 86 #define hb_atomic_int_impl_add(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->fetch_add ((V), … argument 87 #define hb_atomic_int_impl_set_relaxed(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V… argument 88 #define hb_atomic_int_impl_set(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V), std:… argument 89 #define hb_atomic_int_impl_get_relaxed(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (… argument 90 #define hb_atomic_int_impl_get(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (std::me… argument [all …]
|
/external/llvm-project/clang/include/clang/CodeGen/ |
D | CGFunctionInfo.h | 138 auto AI = ABIArgInfo(Direct); variable 139 AI.setCoerceToType(T); 140 AI.setPaddingType(Padding); 141 AI.setDirectOffset(Offset); 142 AI.setCanBeFlattened(CanBeFlattened); 143 return AI; 146 auto AI = getDirect(T); variable 147 AI.setInReg(true); 148 return AI; 153 auto AI = ABIArgInfo(Extend); variable [all …]
|
/external/llvm-project/llvm/lib/Target/AArch64/ |
D | AArch64StackTagging.cpp | 285 AllocaInst *AI; member 306 bool isInterestingAlloca(const AllocaInst &AI); 309 void tagAlloca(AllocaInst *AI, Instruction *InsertBefore, Value *Ptr, 311 void untagAlloca(AllocaInst *AI, Instruction *InsertBefore, uint64_t Size); 414 bool AArch64StackTagging::isInterestingAlloca(const AllocaInst &AI) { in isInterestingAlloca() argument 417 AI.getAllocatedType()->isSized() && AI.isStaticAlloca() && in isInterestingAlloca() 419 AI.getAllocationSizeInBits(*DL).getValue() > 0 && in isInterestingAlloca() 422 !AI.isUsedWithInAlloca() && in isInterestingAlloca() 424 !AI.isSwiftError() && in isInterestingAlloca() 426 !(SSI && SSI->isSafe(AI)); in isInterestingAlloca() [all …]
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/ |
D | AArch64StackTagging.cpp | 271 AllocaInst *AI; member 290 bool isInterestingAlloca(const AllocaInst &AI); 293 void tagAlloca(AllocaInst *AI, Instruction *InsertBefore, Value *Ptr, 295 void untagAlloca(AllocaInst *AI, Instruction *InsertBefore, uint64_t Size); 393 bool AArch64StackTagging::isInterestingAlloca(const AllocaInst &AI) { in isInterestingAlloca() argument 396 AI.getAllocatedType()->isSized() && AI.isStaticAlloca() && in isInterestingAlloca() 398 AI.getAllocationSizeInBits(*DL).getValue() > 0 && in isInterestingAlloca() 401 !AI.isUsedWithInAlloca() && in isInterestingAlloca() 403 !AI.isSwiftError(); in isInterestingAlloca() 407 void AArch64StackTagging::tagAlloca(AllocaInst *AI, Instruction *InsertBefore, in tagAlloca() argument [all …]
|
/external/clang/include/clang/CodeGen/ |
D | CGFunctionInfo.h | 124 auto AI = ABIArgInfo(Direct); variable 125 AI.setCoerceToType(T); 126 AI.setPaddingType(Padding); 127 AI.setDirectOffset(Offset); 128 AI.setCanBeFlattened(CanBeFlattened); 129 return AI; 132 auto AI = getDirect(T); variable 133 AI.setInReg(true); 134 return AI; 137 auto AI = ABIArgInfo(Extend); variable [all …]
|
/external/llvm/lib/CodeGen/ |
D | StackProtector.cpp | 61 StackProtector::getSSPLayout(const AllocaInst *AI) const { in getSSPLayout() 62 return AI ? Layout.lookup(AI) : SSPLK_None; in getSSPLayout() 166 bool StackProtector::HasAddressTaken(const Instruction *AI) { in HasAddressTaken() argument 167 for (const User *U : AI->users()) { in HasAddressTaken() 169 if (AI == SI->getValueOperand()) in HasAddressTaken() 172 if (AI == SI->getOperand(0)) in HasAddressTaken() 237 if (const AllocaInst *AI = dyn_cast<AllocaInst>(&I)) { in RequiresStackProtector() local 238 if (AI->isArrayAllocation()) { in RequiresStackProtector() 244 if (const auto *CI = dyn_cast<ConstantInt>(AI->getArraySize())) { in RequiresStackProtector() 248 Layout.insert(std::make_pair(AI, SSPLK_LargeArray)); in RequiresStackProtector() [all …]
|
D | SafeStack.cpp | 148 uint64_t getStaticAllocaAllocationSize(const AllocaInst* AI); 213 uint64_t SafeStack::getStaticAllocaAllocationSize(const AllocaInst* AI) { in getStaticAllocaAllocationSize() argument 214 uint64_t Size = DL->getTypeAllocSize(AI->getAllocatedType()); in getStaticAllocaAllocationSize() 215 if (AI->isArrayAllocation()) { in getStaticAllocaAllocationSize() 216 auto C = dyn_cast<ConstantInt>(AI->getArraySize()); in getStaticAllocaAllocationSize() 412 if (auto AI = dyn_cast<AllocaInst>(&I)) { in findInsts() local 415 uint64_t Size = getStaticAllocaAllocationSize(AI); in findInsts() 416 if (IsSafeStackAlloca(AI, Size)) in findInsts() 419 if (AI->isStaticAlloca()) { in findInsts() 421 StaticAllocas.push_back(AI); in findInsts() [all …]
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/ |
D | StackProtector.cpp | 164 bool StackProtector::HasAddressTaken(const Instruction *AI) { in HasAddressTaken() argument 165 for (const User *U : AI->users()) { in HasAddressTaken() 169 if (AI == cast<StoreInst>(I)->getValueOperand()) in HasAddressTaken() 175 if (AI == cast<AtomicCmpXchgInst>(I)->getNewValOperand()) in HasAddressTaken() 179 if (AI == cast<PtrToIntInst>(I)->getOperand(0)) in HasAddressTaken() 282 if (const AllocaInst *AI = dyn_cast<AllocaInst>(&I)) { in RequiresStackProtector() local 283 if (AI->isArrayAllocation()) { in RequiresStackProtector() 292 if (const auto *CI = dyn_cast<ConstantInt>(AI->getArraySize())) { in RequiresStackProtector() 296 Layout.insert(std::make_pair(AI, in RequiresStackProtector() 302 Layout.insert(std::make_pair(AI, in RequiresStackProtector() [all …]
|
/external/llvm/lib/Transforms/Utils/ |
D | PromoteMemoryToRegister.cpp | 51 bool llvm::isAllocaPromotable(const AllocaInst *AI) { in isAllocaPromotable() argument 54 unsigned AS = AI->getType()->getAddressSpace(); in isAllocaPromotable() 57 for (const User *U : AI->users()) { in isAllocaPromotable() 64 if (SI->getOperand(0) == AI) in isAllocaPromotable() 119 void AnalyzeAlloca(AllocaInst *AI) { in AnalyzeAlloca() 125 for (auto UI = AI->user_begin(), E = AI->user_end(); UI != E;) { in AnalyzeAlloca() 149 DbgDeclare = FindAllocaDbgDeclare(AI); in AnalyzeAlloca() 293 void ComputeLiveInBlocks(AllocaInst *AI, AllocaInfo &Info, 304 static void removeLifetimeIntrinsicUsers(AllocaInst *AI) { in removeLifetimeIntrinsicUsers() argument 308 for (auto UI = AI->user_begin(), UE = AI->user_end(); UI != UE;) { in removeLifetimeIntrinsicUsers() [all …]
|
D | MetaRenamer.cpp | 84 for (auto AI = M.alias_begin(), AE = M.alias_end(); AI != AE; ++AI) { in runOnModule() local 85 StringRef Name = AI->getName(); in runOnModule() 89 AI->setName("alias"); in runOnModule() 125 for (auto AI = F.arg_begin(), AE = F.arg_end(); AI != AE; ++AI) in runOnFunction() local 126 if (!AI->getType()->isVoidTy()) in runOnFunction() 127 AI->setName("arg"); in runOnFunction()
|
/external/llvm-project/llvm/lib/CodeGen/ |
D | StackProtector.cpp | 165 bool StackProtector::HasAddressTaken(const Instruction *AI, in HasAddressTaken() argument 168 for (const User *U : AI->users()) { in HasAddressTaken() 178 if (AI == cast<StoreInst>(I)->getValueOperand()) in HasAddressTaken() 184 if (AI == cast<AtomicCmpXchgInst>(I)->getNewValOperand()) in HasAddressTaken() 188 if (AI == cast<PtrToIntInst>(I)->getOperand(0)) in HasAddressTaken() 302 if (const AllocaInst *AI = dyn_cast<AllocaInst>(&I)) { in RequiresStackProtector() local 303 if (AI->isArrayAllocation()) { in RequiresStackProtector() 312 if (const auto *CI = dyn_cast<ConstantInt>(AI->getArraySize())) { in RequiresStackProtector() 316 Layout.insert(std::make_pair(AI, in RequiresStackProtector() 322 Layout.insert(std::make_pair(AI, in RequiresStackProtector() [all …]
|
/external/llvm-project/llvm/unittests/Transforms/Utils/ |
D | IntegerDivisionTest.cpp | 36 Function::arg_iterator AI = F->arg_begin(); in TEST() local 37 Value *A = &*AI++; in TEST() 38 Value *B = &*AI++; in TEST() 66 Function::arg_iterator AI = F->arg_begin(); in TEST() local 67 Value *A = &*AI++; in TEST() 68 Value *B = &*AI++; in TEST() 96 Function::arg_iterator AI = F->arg_begin(); in TEST() local 97 Value *A = &*AI++; in TEST() 98 Value *B = &*AI++; in TEST() 126 Function::arg_iterator AI = F->arg_begin(); in TEST() local [all …]
|
/external/llvm/unittests/Transforms/Utils/ |
D | IntegerDivision.cpp | 37 Function::arg_iterator AI = F->arg_begin(); in TEST() local 38 Value *A = &*AI++; in TEST() 39 Value *B = &*AI++; in TEST() 67 Function::arg_iterator AI = F->arg_begin(); in TEST() local 68 Value *A = &*AI++; in TEST() 69 Value *B = &*AI++; in TEST() 97 Function::arg_iterator AI = F->arg_begin(); in TEST() local 98 Value *A = &*AI++; in TEST() 99 Value *B = &*AI++; in TEST() 127 Function::arg_iterator AI = F->arg_begin(); in TEST() local [all …]
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Utils/ |
D | PromoteMemoryToRegister.cpp | 64 bool llvm::isAllocaPromotable(const AllocaInst *AI) { in isAllocaPromotable() argument 67 unsigned AS = AI->getType()->getAddressSpace(); in isAllocaPromotable() 70 for (const User *U : AI->users()) { in isAllocaPromotable() 77 if (SI->getOperand(0) == AI) in isAllocaPromotable() 129 void AnalyzeAlloca(AllocaInst *AI) { in AnalyzeAlloca() 135 for (auto UI = AI->user_begin(), E = AI->user_end(); UI != E;) { in AnalyzeAlloca() 157 DbgDeclares = FindDbgAddrUses(AI); in AnalyzeAlloca() 291 void ComputeLiveInBlocks(AllocaInst *AI, AllocaInfo &Info, 315 static void removeLifetimeIntrinsicUsers(AllocaInst *AI) { in removeLifetimeIntrinsicUsers() argument 319 for (auto UI = AI->user_begin(), UE = AI->user_end(); UI != UE;) { in removeLifetimeIntrinsicUsers() [all …]
|
D | MetaRenamer.cpp | 96 for (auto AI = M.alias_begin(), AE = M.alias_end(); AI != AE; ++AI) { in runOnModule() local 97 StringRef Name = AI->getName(); in runOnModule() 101 AI->setName("alias"); in runOnModule() 146 for (auto AI = F.arg_begin(), AE = F.arg_end(); AI != AE; ++AI) in runOnFunction() local 147 if (!AI->getType()->isVoidTy()) in runOnFunction() 148 AI->setName("arg"); in runOnFunction()
|
/external/llvm-project/llvm/lib/Transforms/Utils/ |
D | PromoteMemoryToRegister.cpp | 64 bool llvm::isAllocaPromotable(const AllocaInst *AI) { in isAllocaPromotable() argument 66 for (const User *U : AI->users()) { in isAllocaPromotable() 73 if (SI->getOperand(0) == AI) in isAllocaPromotable() 126 void AnalyzeAlloca(AllocaInst *AI) { in AnalyzeAlloca() 132 for (auto UI = AI->user_begin(), E = AI->user_end(); UI != E;) { in AnalyzeAlloca() 154 findDbgUsers(DbgUsers, AI); in AnalyzeAlloca() 288 void ComputeLiveInBlocks(AllocaInst *AI, AllocaInfo &Info, 312 static void removeIntrinsicUsers(AllocaInst *AI) { in removeIntrinsicUsers() argument 316 for (auto UI = AI->use_begin(), UE = AI->use_end(); UI != UE;) { in removeIntrinsicUsers() 358 static bool rewriteSingleStoreAlloca(AllocaInst *AI, AllocaInfo &Info, in rewriteSingleStoreAlloca() argument [all …]
|
D | MetaRenamer.cpp | 70 for (auto AI = F.arg_begin(), AE = F.arg_end(); AI != AE; ++AI) in MetaRename() local 71 if (!AI->getType()->isVoidTy()) in MetaRename() 72 AI->setName("arg"); in MetaRename() 95 for (auto AI = M.alias_begin(), AE = M.alias_end(); AI != AE; ++AI) { in MetaRename() local 96 StringRef Name = AI->getName(); in MetaRename() 100 AI->setName("alias"); in MetaRename()
|
/external/llvm-project/llvm/lib/Target/AMDGPU/ |
D | AMDGPUArgumentUsageInfo.cpp | 148 AMDGPUFunctionArgInfo AI; in fixedABILayout() local 149 AI.PrivateSegmentBuffer in fixedABILayout() 151 AI.DispatchPtr = ArgDescriptor::createRegister(AMDGPU::SGPR4_SGPR5); in fixedABILayout() 152 AI.QueuePtr = ArgDescriptor::createRegister(AMDGPU::SGPR6_SGPR7); in fixedABILayout() 156 AI.ImplicitArgPtr = ArgDescriptor::createRegister(AMDGPU::SGPR8_SGPR9); in fixedABILayout() 157 AI.DispatchID = ArgDescriptor::createRegister(AMDGPU::SGPR10_SGPR11); in fixedABILayout() 160 AI.WorkGroupIDX = ArgDescriptor::createRegister(AMDGPU::SGPR12); in fixedABILayout() 161 AI.WorkGroupIDY = ArgDescriptor::createRegister(AMDGPU::SGPR13); in fixedABILayout() 162 AI.WorkGroupIDZ = ArgDescriptor::createRegister(AMDGPU::SGPR14); in fixedABILayout() 165 AI.WorkItemIDX = ArgDescriptor::createRegister(AMDGPU::VGPR31, Mask); in fixedABILayout() [all …]
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/InstCombine/ |
D | InstCombineLoadStoreAlloca.cpp | 162 isOnlyCopiedFromConstantGlobal(AllocaInst *AI, in isOnlyCopiedFromConstantGlobal() argument 165 if (isOnlyCopiedFromConstantGlobal(AI, TheCopy, ToDelete)) in isOnlyCopiedFromConstantGlobal() 171 static bool isDereferenceableForAllocaSize(const Value *V, const AllocaInst *AI, in isDereferenceableForAllocaSize() argument 173 if (AI->isArrayAllocation()) in isDereferenceableForAllocaSize() 175 uint64_t AllocaSize = DL.getTypeStoreSize(AI->getAllocatedType()); in isDereferenceableForAllocaSize() 178 return isDereferenceableAndAlignedPointer(V, Align(AI->getAlignment()), in isDereferenceableForAllocaSize() 182 static Instruction *simplifyAllocaArraySize(InstCombiner &IC, AllocaInst &AI) { in simplifyAllocaArraySize() argument 184 if (!AI.isArrayAllocation()) { in simplifyAllocaArraySize() 186 if (AI.getArraySize()->getType()->isIntegerTy(32)) in simplifyAllocaArraySize() 191 AI.setOperand(0, V); in simplifyAllocaArraySize() [all …]
|