/third_party/skia/third_party/externals/swiftshader/third_party/subzero/src/ |
D | IceIntrinsics.cpp | 29 static_assert(sizeof(Intrinsics::IntrinsicInfo) == 4, 32 bool Intrinsics::isMemoryOrderValid(IntrinsicID ID, uint64_t Order, in isMemoryOrderValid() 79 Intrinsics::ValidateIntrinsicValue 80 Intrinsics::FullIntrinsicInfo::validateIntrinsic(const InstIntrinsic *Intrinsic, in validateIntrinsic() 86 return Intrinsics::BadReturnType; in validateIntrinsic() 88 return Intrinsics::BadReturnType; in validateIntrinsic() 91 return Intrinsics::WrongNumOfArgs; in validateIntrinsic() 96 return Intrinsics::WrongArgType; in validateIntrinsic() 99 return Intrinsics::IsValidIntrinsic; in validateIntrinsic() 102 Type Intrinsics::FullIntrinsicInfo::getArgType(SizeT Index) const { in getArgType()
|
D | IceTargetLoweringMIPS32.cpp | 259 return Intrinsics::MemoryOrderInvalid; in getConstantMemoryOrder() 583 Intrinsics::IntrinsicID ID = Intrinsic->getIntrinsicID(); in genTargetHelperCallFor() 584 if (isVectorType(DestTy) && ID == Intrinsics::Fabs) { in genTargetHelperCallFor() 586 Intrinsics::IntrinsicInfo Info = Intrinsic->getIntrinsicInfo(); in genTargetHelperCallFor() 614 case Intrinsics::AtomicLoad: { in genTargetHelperCallFor() 617 if (!Intrinsics::isMemoryOrderValid( in genTargetHelperCallFor() 636 case Intrinsics::AtomicStore: { in genTargetHelperCallFor() 640 if (!Intrinsics::isMemoryOrderValid( in genTargetHelperCallFor() 659 case Intrinsics::AtomicCmpxchg: { in genTargetHelperCallFor() 662 if (!Intrinsics::isMemoryOrderValid( in genTargetHelperCallFor() [all …]
|
D | IceTargetLoweringX8664.cpp | 658 return Intrinsics::MemoryOrderInvalid; in getConstantMemoryOrder() 703 Intrinsics::IntrinsicID ID = Intrin->getIntrinsicID(); in doLoadOpt() 704 if (ID == Intrinsics::AtomicLoad && in doLoadOpt() 705 Intrinsics::isMemoryOrderValid( in doLoadOpt() 3325 switch (Intrinsics::IntrinsicID ID = Instr->getIntrinsicID()) { in lowerIntrinsic() 3326 case Intrinsics::AtomicCmpxchg: { in lowerIntrinsic() 3327 if (!Intrinsics::isMemoryOrderValid( in lowerIntrinsic() 3342 case Intrinsics::AtomicFence: in lowerIntrinsic() 3343 if (!Intrinsics::isMemoryOrderValid( in lowerIntrinsic() 3350 case Intrinsics::AtomicFenceAll: in lowerIntrinsic() [all …]
|
D | IceTargetLoweringARM32.cpp | 717 Intrinsics::IntrinsicID ID = Intrinsic->getIntrinsicID(); in genTargetHelperCallFor() 721 case Intrinsics::Ctpop: { in genTargetHelperCallFor() 737 case Intrinsics::Longjmp: { in genTargetHelperCallFor() 749 case Intrinsics::Memcpy: { in genTargetHelperCallFor() 764 case Intrinsics::Memmove: { in genTargetHelperCallFor() 777 case Intrinsics::Memset: { in genTargetHelperCallFor() 800 case Intrinsics::Setjmp: { in genTargetHelperCallFor() 4637 return Intrinsics::MemoryOrderInvalid; in getConstantMemoryOrder() 4681 case Intrinsics::AtomicExchange: in createArithInst() 4683 case Intrinsics::AtomicAdd: in createArithInst() [all …]
|
D | IceTargetLoweringX8632.cpp | 649 return Intrinsics::MemoryOrderInvalid; in getConstantMemoryOrder() 692 Intrinsics::IntrinsicID ID = Intrin->getIntrinsicID(); in doLoadOpt() 693 if (ID == Intrinsics::AtomicLoad && in doLoadOpt() 695 Intrinsics::isMemoryOrderValid( in doLoadOpt() 3769 switch (Intrinsics::IntrinsicID ID = Instr->getIntrinsicID()) { in lowerIntrinsic() 3770 case Intrinsics::AtomicCmpxchg: { in lowerIntrinsic() 3771 if (!Intrinsics::isMemoryOrderValid( in lowerIntrinsic() 3786 case Intrinsics::AtomicFence: in lowerIntrinsic() 3787 if (!Intrinsics::isMemoryOrderValid( in lowerIntrinsic() 3794 case Intrinsics::AtomicFenceAll: in lowerIntrinsic() [all …]
|
D | IceConverter.cpp | 561 const Ice::Intrinsics::FullIntrinsicInfo *Info = in convertCallInstruction() 614 const Ice::Intrinsics::FullIntrinsicInfo *I) { in validateIntrinsic() 617 case Ice::Intrinsics::IsValidCall: in validateIntrinsic() 619 case Ice::Intrinsics::BadReturnType: { in validateIntrinsic() 627 case Ice::Intrinsics::WrongNumOfArgs: { in validateIntrinsic() 635 case Ice::Intrinsics::WrongCallArgType: { in validateIntrinsic()
|
D | IceIntrinsics.h | 28 namespace Intrinsics {
|
D | IceInst.h | 624 const Intrinsics::IntrinsicInfo &Info) { in create() 635 Intrinsics::IntrinsicInfo getIntrinsicInfo() const { return Info; } in getIntrinsicInfo() 636 Intrinsics::IntrinsicID getIntrinsicID() const { return Info.ID; } in getIntrinsicID() 643 const Intrinsics::IntrinsicInfo &Info) in InstIntrinsic() 646 const Intrinsics::IntrinsicInfo Info;
|
D | IceTargetLowering.cpp | 338 if (Intrinsic->getIntrinsicID() == Intrinsics::LoadSubVector) in doAddressOpt() 340 else if (Intrinsic->getIntrinsicID() == Intrinsics::StoreSubVector) in doAddressOpt()
|
/third_party/skia/src/sksl/ir/ |
D | SkSLFunctionCall.cpp | 343 namespace Intrinsics { namespace 453 Intrinsics::evaluate_radians); in optimize_intrinsic_call() 456 Intrinsics::evaluate_degrees); in optimize_intrinsic_call() 459 Intrinsics::evaluate_sin); in optimize_intrinsic_call() 462 Intrinsics::evaluate_cos); in optimize_intrinsic_call() 465 Intrinsics::evaluate_tan); in optimize_intrinsic_call() 468 Intrinsics::evaluate_sinh); in optimize_intrinsic_call() 471 Intrinsics::evaluate_cosh); in optimize_intrinsic_call() 474 Intrinsics::evaluate_tanh); in optimize_intrinsic_call() 477 Intrinsics::evaluate_asin); in optimize_intrinsic_call() [all …]
|
/third_party/skia/third_party/externals/swiftshader/src/Reactor/ |
D | SubzeroReactor.cpp | 273 Ice::Intrinsics::MemoryOrder stdToIceMemoryOrder(std::memory_order memoryOrder) in stdToIceMemoryOrder() 277 case std::memory_order_relaxed: return Ice::Intrinsics::MemoryOrderRelaxed; in stdToIceMemoryOrder() 278 case std::memory_order_consume: return Ice::Intrinsics::MemoryOrderConsume; in stdToIceMemoryOrder() 279 case std::memory_order_acquire: return Ice::Intrinsics::MemoryOrderAcquire; in stdToIceMemoryOrder() 280 case std::memory_order_release: return Ice::Intrinsics::MemoryOrderRelease; in stdToIceMemoryOrder() 281 case std::memory_order_acq_rel: return Ice::Intrinsics::MemoryOrderAcquireRelease; in stdToIceMemoryOrder() 282 case std::memory_order_seq_cst: return Ice::Intrinsics::MemoryOrderSequentiallyConsistent; in stdToIceMemoryOrder() 284 return Ice::Intrinsics::MemoryOrderInvalid; in stdToIceMemoryOrder() 1439 …Intrinsics::IntrinsicInfo intrinsic = { Ice::Intrinsics::LoadSubVector, Ice::Intrinsics::SideEffec… in createLoad() 1510 …Intrinsics::IntrinsicInfo intrinsic = { Ice::Intrinsics::StoreSubVector, Ice::Intrinsics::SideEffe… in createStore() [all …]
|
D | Optimizer.cpp | 663 if(instrinsic->getIntrinsicID() == Ice::Intrinsics::LoadSubVector) in asLoadSubVector() 676 if(instrinsic->getIntrinsicID() == Ice::Intrinsics::StoreSubVector) in asStoreSubVector()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/IR/ |
D | Intrinsics.td | 1 //===- Intrinsics.td - Defines all LLVM intrinsics ---------*- tablegen -*-===// 332 //===--------------- Variable Argument Handling Intrinsics ----------------===// 340 //===------------------- Garbage Collection Intrinsics --------------------===// 351 //===------------------- ObjC ARC runtime Intrinsics --------------------===// 423 //===--------------------- Code Generator Intrinsics ----------------------===// 506 //===------------------- Standard C Library Intrinsics --------------------===// 592 //===--------------- Constrained Floating Point Intrinsics ----------------===// 766 //===------------------------- Expect Intrinsics --------------------------===// 771 //===-------------------- Bit Manipulation Intrinsics ---------------------===// 790 //===------------------------ Debugger Intrinsics -------------------------===// [all …]
|
D | IntrinsicsXCore.td | 94 // Intrinsics for events. 104 // Intrinsics for threads.
|
D | CMakeLists.txt | 4 set(LLVM_TARGET_DEFINITIONS Intrinsics.td)
|
D | IntrinsicsRISCV.td | 16 // Atomic Intrinsics have multiple versions for different access widths, which
|
D | IntrinsicsAMDGPU.td | 136 // ABI Special Intrinsics 197 // Instruction Intrinsics 1510 // CI+ Intrinsics 1522 // VI Intrinsics 1566 // GFX10 Intrinsics 1829 // Special Intrinsics for backend internal use only. No frontend
|
D | IntrinsicsPowerPC.td | 37 // Intrinsics used to generate ctr-based loops. These should only be 46 // Intrinsics for [double]word extended forms of divide instructions 705 // Power8 Intrinsics 935 // PowerPC QPX Intrinsics.
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/ |
D | module.install.modulemap | 16 textual header "IR/Intrinsics.gen" 17 textual header "IR/Intrinsics.inc"
|
D | module.modulemap | 224 // Intrinsics.h because they need to be generated by tablegen first. 245 // Intrinsics.h 259 header "IR/Intrinsics.h"
|
/third_party/mesa3d/src/compiler/nir/ |
D | README | 61 Intrinsics 65 or is not referentially transparent is represented by an intrinsic. Intrinsics 74 - Intrinsics can be marked as side-effect free, which permits them to be 79 Intrinsics are used for:
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/InstCombine/ |
D | InstCombineTables.td | 2 include "llvm/IR/Intrinsics.td"
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Mips/ |
D | MSA.txt | 4 Intrinsics are lowered to SelectionDAG nodes where possible in order to enable
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/ |
D | AMDGPUGISel.td | 191 // FIXME: Intrinsics aren't marked as commutable, so we need to add an explcit
|
/third_party/mesa3d/docs/drivers/ |
D | llvmpipe.rst | 243 - `Intel Intrinsics
|