/external/vixl/src/aarch64/ |
D | simulator-constants-aarch64.h | 86 const unsigned kPrintfArgCountOffset = 1 * kInstructionSize; 87 const unsigned kPrintfArgPatternListOffset = 2 * kInstructionSize; 88 const unsigned kPrintfLength = 3 * kInstructionSize; 110 const unsigned kTraceParamsOffset = 1 * kInstructionSize; 111 const unsigned kTraceCommandOffset = 2 * kInstructionSize; 112 const unsigned kTraceLength = 3 * kInstructionSize; 140 const unsigned kLogParamsOffset = 1 * kInstructionSize; 141 const unsigned kLogLength = 2 * kInstructionSize; 146 const unsigned kRuntimeCallWrapperOffset = 1 * kInstructionSize;
|
D | macro-assembler-aarch64.cc | 122 if (option == kBranchRequired) emit_size += kInstructionSize; in Emit() 125 VIXL_ASSERT(emit_size % kInstructionSize == 0); in Emit() 136 ExactAssemblyScopeWithoutPoolsCheck guard(masm_, kInstructionSize); in Emit() 143 ExactAssemblyScopeWithoutPoolsCheck guard(masm_, kInstructionSize); in Emit() 271 ExactAssemblyScopeWithoutPoolsCheck guard(masm_, kInstructionSize); in Emit() 297 ExactAssemblyScopeWithoutPoolsCheck guard(masm_, kInstructionSize); in Emit() 595 EmissionCheckScope guard(this, 2 * kInstructionSize); in B() 621 EmissionCheckScope guard(this, 2 * kInstructionSize); in Cbnz() 647 EmissionCheckScope guard(this, 2 * kInstructionSize); in Cbz() 672 EmissionCheckScope guard(this, 2 * kInstructionSize); in Tbnz() [all …]
|
D | instructions-aarch64.cc | 229 VIXL_ASSERT((LSSize_offset + LSSize_width) == (kInstructionSize * 8)); in CalcLSDataSize() 279 return encoded_max * kInstructionSize; in GetImmBranchForwardRange() 305 offset = GetImmBranch() * static_cast<int>(kInstructionSize); in GetImmPCOffsetTarget()
|
D | instructions-aarch64.h | 40 const unsigned kInstructionSize = 4; variable 411 (1 << kLoadLiteralImmBitwidth) / 2 - kInstructionSize; 467 return this + kInstructionSize; in GetNextInstruction()
|
D | debugger-aarch64.cc | 561 from -= (count - 1) * kInstructionSize; in PrintInstructions() 563 const Instruction* to = from + count * kInstructionSize; in PrintInstructions() 1264 debugger->WritePc(debugger->ReadPc() + steps * kInstructionSize); in Run()
|
/external/vixl/test/ |
D | test-code-generation-scopes.cc | 72 CodeBufferCheckScope scope(&masm, aarch64::kInstructionSize); in TEST() 101 CodeBufferCheckScope scope(&masm, 2 * aarch64::kInstructionSize); in TEST() 134 scope.Open(&masm, aarch64::kInstructionSize); in TEST() 164 CodeBufferCheckScope scope(&masm, aarch64::kInstructionSize); in TEST() 200 scope.Open(&masm, aarch64::kInstructionSize); in TEST() 230 EmissionCheckScope scope(&masm, aarch64::kInstructionSize); in TEST() 262 scope.Open(&masm, aarch64::kInstructionSize); in TEST() 292 EmissionCheckScope scope(&masm, aarch64::kInstructionSize); in TEST() 328 scope.Open(&masm, aarch64::kInstructionSize); in TEST() 386 VIXL_CHECK((expected + aarch64::kInstructionSize) == \ [all …]
|
/external/v8/src/arm64/ |
D | instructions-arm64.h | 125 return InstructionAtOffset(count * static_cast<int>(kInstructionSize)); 129 return InstructionAtOffset(count * static_cast<int>(kInstructionSize)); 311 kInstructionSize; in ImmBranchRange() 392 DCHECK(check == NO_CHECK || IsAligned(offset, kInstructionSize)); 399 DCHECK(check == NO_CHECK || IsAligned(offset, kInstructionSize)); 465 const unsigned kPrintfArgCountOffset = 1 * kInstructionSize; 466 const unsigned kPrintfArgPatternListOffset = 2 * kInstructionSize; 467 const unsigned kPrintfLength = 3 * kInstructionSize; 491 const unsigned kDebugCodeOffset = 1 * kInstructionSize; 492 const unsigned kDebugParamsOffset = 2 * kInstructionSize; [all …]
|
D | codegen-arm64.cc | 50 young_sequence_.length() / kInstructionSize); in CodeAgingHelper() 56 const int length = kCodeAgeStubEntryOffset / kInstructionSize; in CodeAgingHelper() 86 kNoCodeAgeSequenceLength / kInstructionSize); in PatchPlatformCodeAge()
|
D | deoptimizer-arm64.cc | 20 return 4 * kInstructionSize; in patch_size() 53 patch_size() / kInstructionSize); in PatchCodeForDeoptimization() 54 patcher.ldr_pcrel(ip0, (2 * kInstructionSize) >> kLoadLiteralScaleLog2); in PatchCodeForDeoptimization() 298 const int Deoptimizer::table_entry_size_ = 2 * kInstructionSize;
|
D | instructions-arm64.cc | 254 DCHECK(IsAligned(DistanceTo(target), kInstructionSize)); in SetBranchImmTarget() 290 DCHECK(IsAligned(DistanceTo(target), kInstructionSize)); in SetUnresolvedInternalReferenceImmTarget() 305 DCHECK(IsAligned(DistanceTo(source), kInstructionSize)); in SetImmLLiteral()
|
D | assembler-arm64-inl.h | 592 Address candidate = pc - 2 * kInstructionSize; 617 STATIC_ASSERT(Assembler::kCallSizeWithoutRelocation == 4 * kInstructionSize); 618 STATIC_ASSERT(Assembler::kCallSizeWithRelocation == 2 * kInstructionSize); 779 static const int kNoCodeAgeSequenceLength = 5 * kInstructionSize; 780 static const int kCodeAgeStubEntryOffset = 3 * kInstructionSize; 945 DCHECK(IsAligned(offset, kInstructionSize)); 1166 DCHECK((SizeLS_offset + SizeLS_width) == (kInstructionSize * 8));
|
D | assembler-arm64.cc | 366 return 4 * kInstructionSize + EntryCount() * kPointerSize; in WorstCaseSize() 378 int prologue_size = require_jump ? kInstructionSize : 0; in SizeIfEmittedAtCurrentPc() 379 prologue_size += 2 * kInstructionSize; in SizeIfEmittedAtCurrentPc() 381 0 : kInstructionSize; in SizeIfEmittedAtCurrentPc() 940 return RoundUp(size, kInstructionSize) / kInstructionSize; in ConstantPoolSizeAt() 945 return kPrintfLength / kInstructionSize; in ConstantPoolSizeAt() 2313 DCHECK(RoundUp(len, kInstructionSize) <= static_cast<size_t>(kGap)); in EmitStringData() 2317 STATIC_ASSERT(sizeof(pad) == kInstructionSize); in EmitStringData() 2318 EmitData(pad, RoundUp(pc_offset(), kInstructionSize) - pc_offset()); in EmitStringData() 2993 int pc_limit = pc_offset() + instructions * kInstructionSize; in BlockConstPoolFor() [all …]
|
D | assembler-arm64.h | 854 static const int kCallSizeWithoutRelocation = 4 * kInstructionSize; 855 static const int kCallSizeWithRelocation = 2 * kInstructionSize; 884 return SizeOfCodeGeneratedSince(label) / kInstructionSize; in InstructionsGeneratedSince() 892 kDebugBreakSlotInstructions * kInstructionSize; 1850 static const int kMaxVeneerCodeSize = 1 * kInstructionSize; 2016 next_constant_pool_check_ = pc_offset() + instructions * kInstructionSize; in SetNextConstPoolCheckIn() 2022 STATIC_ASSERT(sizeof(instruction) == kInstructionSize); in Emit() 2200 count * kInstructionSize + kGap) { in PatchingAssembler() 2205 : Assembler(isolate, start, count * kInstructionSize + kGap) { in PatchingAssembler()
|
D | code-stubs-arm64.h | 137 Instruction* instr2 = patcher.InstructionAt(kInstructionSize); in Patch()
|
/external/vixl/test/aarch64/ |
D | test-fuzz-aarch64.cc | 50 Instruction buffer[kInstructionSize]; in TEST() 69 Instruction buffer[kInstructionSize]; in TEST() 91 Instruction buffer[kInstructionSize]; 107 Instruction buffer[kInstructionSize];
|
D | test-assembler-aarch64.cc | 265 instruction += kInstructionSize; \ 291 VIXL_CHECK((expected + kInstructionSize) == (masm.GetLiteralPoolSize())) 571 ExactAssemblyScope scope(&masm, 3 * kInstructionSize); in TEST() 2042 VIXL_ASSERT((offset_into_page % kInstructionSize) == 0); in AdrpPageBoundaryHelper() 2076 for (size_t i = 2; i < (kPageSize / kInstructionSize); i += 2) { in AdrpPageBoundaryHelper() 2087 for (size_t i = 0; i < (kPageSize / kInstructionSize);) { in AdrpPageBoundaryHelper() 2088 if (i++ == (offset_into_page / kInstructionSize)) __ bind(&test); in AdrpPageBoundaryHelper() 2090 if (i++ == (offset_into_page / kInstructionSize)) __ bind(&test); in AdrpPageBoundaryHelper() 2094 for (size_t i = 0; i < (kPageSize / kInstructionSize); i += 2) { in AdrpPageBoundaryHelper() 2121 AdrpPageBoundaryHelper(kInstructionSize * 0); in TEST() [all …]
|
/external/vixl/benchmarks/aarch64/ |
D | bench-branch-link.cc | 57 MacroAssembler masm(instructions * kInstructionSize); in main() 58 ExactAssemblyScope scope(&masm, instructions * kInstructionSize); in main()
|
D | bench-dataop.cc | 59 const unsigned buffer_instruction_count = buffer_size / kInstructionSize; in main() 79 ExactAssemblyScope scope(&masm, remaining * kInstructionSize); in main()
|
D | bench-branch.cc | 59 const int buffer_instruction_count = buffer_size / kInstructionSize; in main() 82 ExactAssemblyScope scope(&masm, remaining * kInstructionSize); in main()
|
D | bench-branch-masm.cc | 60 buffer_size / (instructions_per_iteration * kInstructionSize); in main()
|
D | bench-branch-link-masm.cc | 53 size_t base_buf_size = iterations * 4 * kInstructionSize; in main()
|
/external/vixl/examples/aarch64/ |
D | non-const-visitor.cc | 97 for (instr = instr_start; instr < instr_end; instr += kInstructionSize) { in main() 122 for (instr = start; instr < end; instr += kInstructionSize) { in ModifyNonConstVisitorTestGeneratedCode()
|
D | literal.cc | 46 kInstructionSize + sizeof(int64_t), in LiteralExample()
|
D | custom-disassembler.cc | 165 for (instr = instr_start; instr < instr_end; instr += kInstructionSize) { in TestCustomDisassembler()
|
/external/v8/src/debug/arm64/ |
D | debug-arm64.cc | 68 patcher.ldr_pcrel(ip0, (2 * kInstructionSize) >> kLoadLiteralScaleLog2); in PatchDebugBreakSlot()
|