/external/v8/src/regexp/mips64/ |
D | regexp-macro-assembler-mips64.cc | 187 __ Ld(a0, register_location(reg)); in AdvanceRegister() local 219 __ Ld(a1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckAtStart() local 227 __ Ld(a1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckNotAtStart() local 254 __ Ld(a0, register_location(start_reg)); // Index of start of capture. in CheckNotBackReferenceIgnoreCase() local 255 __ Ld(a1, register_location(start_reg + 1)); // Index of end of capture. in CheckNotBackReferenceIgnoreCase() local 264 __ Ld(t1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckNotBackReferenceIgnoreCase() local 324 __ Ld(t1, register_location(start_reg)); // Index of start of capture. in CheckNotBackReferenceIgnoreCase() local 325 __ Ld(a2, register_location(start_reg + 1)); // Index of end of capture. in CheckNotBackReferenceIgnoreCase() local 380 __ Ld(end_of_input_address(), MemOperand(frame_pointer(), kInputEnd)); in CheckNotBackReferenceIgnoreCase() local 403 __ Ld(a0, register_location(start_reg)); in CheckNotBackReference() local [all …]
|
/external/v8/src/builtins/mips64/ |
D | builtins-mips64.cc | 48 __ Ld(a2, FieldMemOperand(a1, JSFunction::kPrototypeOrInitialMapOffset)); in Generate_InternalArrayConstructor() local 131 __ Ld(t1, MemOperand(t0)); in Generate_JSBuiltinsConstructStubHelper() local 145 __ Ld(cp, MemOperand(fp, ConstructFrameConstants::kContextOffset)); in Generate_JSBuiltinsConstructStubHelper() local 147 __ Ld(a1, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSBuiltinsConstructStubHelper() local 190 __ Ld(t2, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); in Generate_JSConstructStubGeneric() local 237 __ Ld(a1, MemOperand(fp, ConstructFrameConstants::kConstructorOffset)); in Generate_JSConstructStubGeneric() local 238 __ Ld(a0, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSConstructStubGeneric() local 262 __ Ld(t1, MemOperand(t0)); in Generate_JSConstructStubGeneric() local 286 __ Ld(cp, MemOperand(fp, ConstructFrameConstants::kContextOffset)); in Generate_JSConstructStubGeneric() local 315 __ Ld(v0, MemOperand(sp, 0 * kPointerSize)); in Generate_JSConstructStubGeneric() local [all …]
|
/external/v8/src/mips64/ |
D | deoptimizer-mips64.cc | 67 __ Ld(a2, MemOperand(sp, kSavedRegistersAreaSize)); in Generate() local 83 __ Ld(a1, MemOperand(fp, CommonFrameConstants::kContextOrFrameTypeOffset)); in Generate() local 85 __ Ld(a0, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset)); in Generate() local 103 __ Ld(a1, MemOperand(v0, Deoptimizer::input_offset())); in Generate() local 110 __ Ld(a2, MemOperand(sp, i * kPointerSize)); in Generate() local 146 __ Ld(a2, MemOperand(a1, FrameDescription::frame_size_offset())); in Generate() local 173 __ Ld(sp, MemOperand(a0, Deoptimizer::caller_frame_top_offset())); in Generate() local 181 __ Ld(a4, MemOperand(a0, Deoptimizer::output_offset())); // a4 is output_. in Generate() local 186 __ Ld(a2, MemOperand(a4, 0)); // output_[ix] in Generate() local 187 __ Ld(a3, MemOperand(a2, FrameDescription::frame_size_offset())); in Generate() local [all …]
|
D | code-stubs-mips64.cc | 71 __ Ld(a4, MemOperand(a4)); in Generate() local 97 __ Ld(a6, MemOperand(a5)); in Generate() local 194 __ Ld(t9, MemOperand(sp, kCArgsSlotsSize)); in Generate() local 356 __ Ld(s0, MemOperand(s5, kNextOffset)); in CallApiFunctionAndReturn() local 357 __ Ld(s1, MemOperand(s5, kLimitOffset)); in CallApiFunctionAndReturn() local 392 __ Ld(v0, return_value_operand); in CallApiFunctionAndReturn() local 405 __ Ld(kScratchReg, MemOperand(s5, kLimitOffset)); in CallApiFunctionAndReturn() local 413 __ Ld(s0, MemOperand(sp, stack_space_offset)); in CallApiFunctionAndReturn() local 423 __ Ld(a5, MemOperand(kScratchReg)); in CallApiFunctionAndReturn() local 549 __ Ld(scratch, FieldMemOperand(callback, AccessorInfo::kDataOffset)); in Generate() local [all …]
|
D | macro-assembler-mips64.h | 305 void Ld(Register rd, const MemOperand& rs); 390 Ld(dst, MemOperand(sp, 0)); in pop() 398 Ld(src2, MemOperand(sp, 0 * kPointerSize)); in Pop() 399 Ld(src1, MemOperand(sp, 1 * kPointerSize)); in Pop() 405 Ld(src3, MemOperand(sp, 0 * kPointerSize)); in Pop() 406 Ld(src2, MemOperand(sp, 1 * kPointerSize)); in Pop() 407 Ld(src1, MemOperand(sp, 2 * kPointerSize)); in Pop() 1259 Ld(scratch, MemOperand(scratch)); in GenerateSwitchTable() 1269 Ld(scratch, MemOperand(scratch, 6 * v8::internal::kInstrSize)); in GenerateSwitchTable()
|
D | macro-assembler-mips64.cc | 131 Ld(destination, MemOperand(s6, RootRegisterOffset(index))); in LoadRoot() 138 Ld(destination, MemOperand(s6, RootRegisterOffset(index))); in LoadRoot() 311 Ld(scratch, MemOperand(address)); in RecordWrite() 1307 Ld(rd, rs); in Uld() 1464 void TurboAssembler::Ld(Register rd, const MemOperand& rs) { in Ld() function in v8::internal::TurboAssembler 1963 Ld(ToRegister(i), MemOperand(sp, stack_offset)); in MultiPop() 3239 Ld(result, MemOperand(sp, 0)); in TruncateDoubleToI() 4129 Ld(destination, in LoadFromConstantsTable() 4135 Ld(destination, MemOperand(kRootRegister, offset)); in LoadRootRelative() 4409 Ld(a1, MemOperand(a1)); in MaybeDropFrames() [all …]
|
/external/v8/src/debug/mips64/ |
D | debug-mips64.cc | 36 __ Ld(a1, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset)); in GenerateFrameDropperTrampoline() local 41 __ Ld(a0, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); in GenerateFrameDropperTrampoline() local
|
/external/llvm/lib/Transforms/Scalar/ |
D | LoopVersioningLICM.cpp | 342 LoadInst *Ld = dyn_cast<LoadInst>(I); in instructionSafeForVersioning() local 343 if (!Ld || !Ld->isSimple()) { in instructionSafeForVersioning() 348 Value *Ptr = Ld->getPointerOperand(); in instructionSafeForVersioning()
|
D | LoopInterchange.cpp | 93 LoadInst *Ld = dyn_cast<LoadInst>(I); in populateDependencyMatrix() local 95 if (!St && !Ld) in populateDependencyMatrix() 97 if (Ld && !Ld->isSimple()) in populateDependencyMatrix()
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Transforms/Scalar/ |
D | LoopVersioningLICM.cpp | 375 LoadInst *Ld = dyn_cast<LoadInst>(I); in instructionSafeForVersioning() local 376 if (!Ld || !Ld->isSimple()) { in instructionSafeForVersioning() 381 Value *Ptr = Ld->getPointerOperand(); in instructionSafeForVersioning()
|
D | LoopUnrollAndJamPass.cpp | 236 if (auto *Ld = dyn_cast<LoadInst>(&I)) { in computeUnrollAndJamCount() local 237 Value *V = Ld->getPointerOperand(); in computeUnrollAndJamCount()
|
/external/llvm/lib/CodeGen/ |
D | GCRootLowering.cpp | 225 Value *Ld = new LoadInst(CI->getArgOperand(1), "", CI); in PerformDefaultLowering() local 226 Ld->takeName(CI); in PerformDefaultLowering() 227 CI->replaceAllUsesWith(Ld); in PerformDefaultLowering()
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/CodeGen/ |
D | GCRootLowering.cpp | 224 Value *Ld = new LoadInst(CI->getArgOperand(1), "", CI); in PerformDefaultLowering() local 225 Ld->takeName(CI); in PerformDefaultLowering() 226 CI->replaceAllUsesWith(Ld); in PerformDefaultLowering()
|
/external/swiftshader/third_party/LLVM/lib/CodeGen/ |
D | GCStrategy.cpp | 299 Value *Ld = new LoadInst(CI->getArgOperand(1), "", CI); in PerformDefaultLowering() local 300 Ld->takeName(CI); in PerformDefaultLowering() 301 CI->replaceAllUsesWith(Ld); in PerformDefaultLowering()
|
/external/ImageMagick/PerlMagick/t/reference/filter/ |
D | SigmoidalContrast.miff | 15 …c+"[&m0'�NJyVHGH6-8)'-$++(&+#,6)3C0(6&%*'-,#9F'q�Sk�MMs7T{=Z�DZ�AQu@d^Ld^Ld^Ld^Ld^Lf^Li^Mj_Nj_Ni]…
|
D | GaussianBlur.miff | 15 …Ld[Ld\Me]Nf^Oh_Qi`TjaVjaWi`Ue^Q`ZLZUFVRBXP@eP@OB�LB�H?�C:�?5�=1�</�</�<0�=0�=0�>1�>4�>7�>8�=7�<5�…
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/AMDGPU/ |
D | SMInstructions.td | 383 auto Ld = cast<LoadSDNode>(N); 384 return Ld->getAlignment() >= 4 && 385 …((((Ld->getAddressSpace() == AMDGPUASI.CONSTANT_ADDRESS) || (Ld->getAddressSpace() == AMDGPUASI.CO… 386 (Subtarget->getScalarizeGlobalBehavior() && Ld->getAddressSpace() == AMDGPUASI.GLOBAL_ADDRESS && 387 !Ld->isVolatile() && !N->isDivergent() &&
|
/external/llvm/lib/Analysis/ |
D | LoopAccessAnalysis.cpp | 1547 auto *Ld = dyn_cast<LoadInst>(&I); in analyzeLoop() local 1548 if (!Ld || (!Ld->isSimple() && !IsAnnotatedParallel)) { in analyzeLoop() 1549 emitAnalysis(LoopAccessReport(Ld) in analyzeLoop() 1556 Loads.push_back(Ld); in analyzeLoop() 1557 DepChecker->addAccess(Ld); in analyzeLoop() 1559 collectStridedAccess(Ld); in analyzeLoop()
|
/external/llvm/test/tools/llvm-objdump/ |
D | macho-LLVM-bundle.test | 16 # CHECK: <subdoc subdoc_name="Ld">
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/tools/llvm-objdump/ |
D | macho-LLVM-bundle.test | 16 # CHECK: <subdoc subdoc_name="Ld">
|
/external/llvm/lib/Target/X86/ |
D | X86ISelLowering.cpp | 3540 } else if (LoadSDNode *Ld = dyn_cast<LoadSDNode>(Arg)) { in MatchingStackOffset() local 3548 SDValue Ptr = Ld->getBasePtr(); in MatchingStackOffset() 5758 SDValue Ld; in LowerVectorBroadcast() local 5776 Ld = Splat; in LowerVectorBroadcast() 5777 ConstSplatVal = (Ld.getOpcode() == ISD::Constant || in LowerVectorBroadcast() 5778 Ld.getOpcode() == ISD::ConstantFP); in LowerVectorBroadcast() 5782 if (!ConstSplatVal && !BVOp->isOnlyUserOf(Ld.getNode())) in LowerVectorBroadcast() 5808 Ld = Sc.getOperand(0); in LowerVectorBroadcast() 5809 ConstSplatVal = (Ld.getOpcode() == ISD::Constant || in LowerVectorBroadcast() 5810 Ld.getOpcode() == ISD::ConstantFP); in LowerVectorBroadcast() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/AArch64/ |
D | AArch64ISelDAGToDAG.cpp | 1202 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectLoad() local 1203 SDValue SuperReg = SDValue(Ld, 0); in SelectLoad() 1208 ReplaceUses(SDValue(N, NumVecs), SDValue(Ld, 1)); in SelectLoad() 1213 cast<MachineSDNode>(Ld)->setMemRefs(MemOp, MemOp + 1); in SelectLoad() 1231 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectPostLoad() local 1234 ReplaceUses(SDValue(N, NumVecs), SDValue(Ld, 0)); in SelectPostLoad() 1237 SDValue SuperReg = SDValue(Ld, 1); in SelectPostLoad() 1246 ReplaceUses(SDValue(N, NumVecs + 1), SDValue(Ld, 2)); in SelectPostLoad() 1349 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectLoadLane() local 1350 SDValue SuperReg = SDValue(Ld, 0); in SelectLoadLane() [all …]
|
/external/antlr/runtime/ObjC/Framework/ |
D | ACNumber.m | 164 return [NSString stringWithFormat:@"%Ld", u.i];
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Analysis/ |
D | LoopAccessAnalysis.cpp | 1802 auto *Ld = dyn_cast<LoadInst>(&I); in analyzeLoop() local 1803 if (!Ld || (!Ld->isSimple() && !IsAnnotatedParallel)) { in analyzeLoop() 1804 recordAnalysis("NonSimpleLoad", Ld) in analyzeLoop() 1811 Loads.push_back(Ld); in analyzeLoop() 1812 DepChecker->addAccess(Ld); in analyzeLoop() 1814 collectStridedAccess(Ld); in analyzeLoop()
|
/external/llvm/lib/Target/AArch64/ |
D | AArch64ISelDAGToDAG.cpp | 1145 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectLoad() local 1146 SDValue SuperReg = SDValue(Ld, 0); in SelectLoad() 1151 ReplaceUses(SDValue(N, NumVecs), SDValue(Ld, 1)); in SelectLoad() 1168 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectPostLoad() local 1171 ReplaceUses(SDValue(N, NumVecs), SDValue(Ld, 0)); in SelectPostLoad() 1174 SDValue SuperReg = SDValue(Ld, 1); in SelectPostLoad() 1183 ReplaceUses(SDValue(N, NumVecs + 1), SDValue(Ld, 2)); in SelectPostLoad() 1281 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectLoadLane() local 1282 SDValue SuperReg = SDValue(Ld, 0); in SelectLoadLane() 1294 ReplaceUses(SDValue(N, NumVecs), SDValue(Ld, 1)); in SelectLoadLane() [all …]
|