| /art/compiler/trampolines/ |
| D | trampoline_compiler.cc | 64 ArmVIXLAssembler assembler(arena); in CreateTrampoline() local 102 Arm64Assembler assembler(arena); in CreateTrampoline() local 141 MipsAssembler assembler(arena); in CreateTrampoline() local 173 Mips64Assembler assembler(arena); in CreateTrampoline() local 205 X86Assembler assembler(arena); in CreateTrampoline() local 226 x86_64::X86_64Assembler assembler(arena); in CreateTrampoline() local
|
| /art/compiler/optimizing/ |
| D | intrinsics_arm_vixl.cc | 92 ArmVIXLAssembler* assembler = down_cast<ArmVIXLAssembler*>(codegen->GetAssembler()); in EmitNativeCode() local 128 static void GenSystemArrayCopyBaseAddress(ArmVIXLAssembler* assembler, in GenSystemArrayCopyBaseAddress() 151 static void GenSystemArrayCopyEndAddress(ArmVIXLAssembler* assembler, in GenSystemArrayCopyEndAddress() 182 ArmVIXLAssembler* assembler = arm_codegen->GetAssembler(); in EmitNativeCode() local 279 static void MoveFPToInt(LocationSummary* locations, bool is64bit, ArmVIXLAssembler* assembler) { in MoveFPToInt() 289 static void MoveIntToFP(LocationSummary* locations, bool is64bit, ArmVIXLAssembler* assembler) { in MoveIntToFP() 346 ArmVIXLAssembler* assembler = codegen->GetAssembler(); in GenNumberOfLeadingZeros() local 395 ArmVIXLAssembler* assembler = codegen->GetAssembler(); in GenNumberOfTrailingZeros() local 444 static void MathAbsFP(HInvoke* invoke, ArmVIXLAssembler* assembler) { in MathAbsFP() 476 ArmVIXLAssembler* assembler) { in GenAbsInteger() [all …]
|
| D | intrinsics_mips64.cc | 149 static void MoveFPToInt(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { in MoveFPToInt() 186 static void MoveIntToFP(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { in MoveIntToFP() 225 Mips64Assembler* assembler) { in GenReverseBytes() 277 Mips64Assembler* assembler) { in GenNumberOfLeadingZeroes() 308 Mips64Assembler* assembler) { in GenNumberOfTrailingZeroes() 345 Mips64Assembler* assembler) { in GenReverse() 390 Mips64Assembler* assembler) { in GenBitCount() 474 static void MathAbsFP(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { in MathAbsFP() 511 static void GenAbsInteger(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { in GenAbsInteger() 547 Mips64Assembler* assembler) { in GenMinMaxFP() [all …]
|
| D | intrinsics_arm.cc | 51 static void GenSystemArrayCopyBaseAddress(ArmAssembler* assembler, in GenSystemArrayCopyBaseAddress() 74 static void GenSystemArrayCopyEndAddress(ArmAssembler* assembler, in GenSystemArrayCopyEndAddress() 110 ArmAssembler* assembler = arm_codegen->GetAssembler(); in EmitNativeCode() local 209 static void MoveFPToInt(LocationSummary* locations, bool is64bit, ArmAssembler* assembler) { in MoveFPToInt() 221 static void MoveIntToFP(LocationSummary* locations, bool is64bit, ArmAssembler* assembler) { in MoveIntToFP() 280 ArmAssembler* assembler = codegen->GetAssembler(); in GenNumberOfLeadingZeros() local 329 ArmAssembler* assembler = codegen->GetAssembler(); in GenNumberOfTrailingZeros() local 378 static void MathAbsFP(LocationSummary* locations, bool is64bit, ArmAssembler* assembler) { in MathAbsFP() 418 ArmAssembler* assembler) { in GenAbsInteger() 466 ArmAssembler* assembler) { in GenMinMax() [all …]
|
| D | intrinsics_utils.h | 51 Assembler* assembler = codegen->GetAssembler(); in EmitNativeCode() local
|
| D | intrinsics_mips.cc | 160 static void MoveFPToInt(LocationSummary* locations, bool is64bit, MipsAssembler* assembler) { in MoveFPToInt() 202 static void MoveIntToFP(LocationSummary* locations, bool is64bit, MipsAssembler* assembler) { in MoveIntToFP() 251 MipsAssembler* assembler) { in GenReverse() 445 MipsAssembler* assembler) { in GenNumberOfLeadingZeroes() 493 MipsAssembler* assembler) { in GenNumberOfTrailingZeroes() 616 MipsAssembler* assembler) { in GenBitCount() 751 MipsAssembler* assembler) { in MathAbsFP() 817 static void GenAbsInteger(LocationSummary* locations, bool is64bit, MipsAssembler* assembler) { in GenAbsInteger() 867 MipsAssembler* assembler) { in GenMinMaxFP() 1100 MipsAssembler* assembler) { in GenMinMax() [all …]
|
| D | intrinsics_x86_64.cc | 147 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in MoveFPToInt() 153 static void MoveIntToFP(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in MoveIntToFP() 197 X86_64Assembler* assembler) { in GenReverseBytes() 258 X86_64Assembler* assembler, in MathAbsFP() 302 static void GenAbsInteger(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in GenAbsInteger() 343 X86_64Assembler* assembler, in GenMinMaxFP() 472 X86_64Assembler* assembler) { in GenMinMax() 601 X86_64Assembler* assembler, in GenSSE41FPToFPIntrinsic() 678 X86_64Assembler* assembler = GetAssembler(); in VisitMathRoundFloat() local 722 X86_64Assembler* assembler = GetAssembler(); in VisitMathRoundDouble() local [all …]
|
| D | intrinsics_x86.cc | 201 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MoveFPToInt() 216 static void MoveIntToFP(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MoveIntToFP() 286 X86Assembler* assembler) { in GenReverseBytes() 325 X86Assembler* assembler = GetAssembler(); in VisitLongReverseBytes() local 365 X86Assembler* assembler, in MathAbsFP() 428 static void GenAbsInteger(LocationSummary* locations, X86Assembler* assembler) { in GenAbsInteger() 456 static void GenAbsLong(LocationSummary* locations, X86Assembler* assembler) { in GenAbsLong() 501 X86Assembler* assembler, in GenMinMaxFP() 667 X86Assembler* assembler) { in GenMinMax() 830 X86Assembler* assembler, in GenSSE41FPToFPIntrinsic() [all …]
|
| D | code_generator_x86_64.cc | 7013 X86_64Assembler* assembler = codegen_->GetAssembler(); in CreateJumpTable() local 7042 X86_64Assembler* assembler = GetAssembler(); in Finalize() local
|
| D | code_generator_x86.cc | 7659 X86Assembler* assembler = codegen_->GetAssembler(); in CreateJumpTable() local 7689 X86Assembler* assembler = GetAssembler(); in Finalize() local
|
| D | code_generator_arm_vixl.cc | 1879 static bool CanGenerateTest(HCondition* condition, ArmVIXLAssembler* assembler) { in CanGenerateTest() 5357 ArmVIXLAssembler* assembler = codegen_->GetAssembler(); in CanEncodeConstantAsImmediate() local
|
| D | code_generator_arm.cc | 1788 static bool CanGenerateTest(HCondition* condition, ArmAssembler* assembler) { in CanGenerateTest() 5348 ArmAssembler* assembler = codegen_->GetAssembler(); in CanEncodeConstantAsImmediate() local
|
| /art/compiler/utils/x86_64/ |
| D | assembler_x86_64_test.cc | 365 std::string shll_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shll_fn() 388 std::string shlq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shlq_fn() 411 std::string shrl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shrl_fn() 434 std::string shrq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shrq_fn() 457 std::string sarl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in sarl_fn() 480 std::string sarq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in sarq_fn() 503 std::string rorl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rorl_fn() 526 std::string roll_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in roll_fn() 549 std::string rorq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rorq_fn() 572 std::string rolq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rolq_fn() [all …]
|
| D | jni_macro_assembler_x86_64.cc | 144 static void DecreaseFrameSizeImpl(size_t adjust, X86_64Assembler* assembler) { in DecreaseFrameSizeImpl()
|
| /art/compiler/linker/arm/ |
| D | relative_patcher_thumb2.cc | 105 arm::Thumb2Assembler assembler(&arena); in CompileThunk() local
|
| /art/compiler/linker/arm64/ |
| D | relative_patcher_arm64.cc | 375 static void EmitGrayCheckAndFastPath(arm64::Arm64Assembler& assembler, in EmitGrayCheckAndFastPath() 406 arm64::Arm64Assembler assembler(&arena); in CompileThunk() local
|
| /art/compiler/utils/arm/ |
| D | assembler_thumb2.h | 589 ArrayRef<const FixupId> Dependents(const Thumb2Assembler& assembler) const { in Dependents()
|
| D | jni_macro_assembler_arm.cc | 184 static void DecreaseFrameSizeImpl(ArmAssembler* assembler, size_t adjust) { in DecreaseFrameSizeImpl() 286 static void EmitLoad(ArmAssembler* assembler, in EmitLoad()
|
| D | assembler_thumb2.cc | 31 void Thumb2Assembler::Fixup::ForExpandableDependencies(Thumb2Assembler* assembler, Function fn) { in ForExpandableDependencies() 54 void Thumb2Assembler::Fixup::PrepareDependents(Thumb2Assembler* assembler) { in PrepareDependents()
|
| /art/compiler/utils/ |
| D | assembler_thumb_test.cc | 181 void EmitAndCheck(arm::Thumb2Assembler* assembler, const char* testname, in EmitAndCheck() 192 void EmitAndCheck(arm::Thumb2Assembler* assembler, const char* testname) { in EmitAndCheck() 208 arm::Thumb2Assembler assembler; member in art::arm::Thumb2AssemblerTest 1617 JniAssemblerType assembler; member in art::arm::ArmVIXLAssemblerTest 1623 void EmitAndCheck(JniAssemblerType* assembler, const char* testname, in EmitAndCheck() 1634 void EmitAndCheck(JniAssemblerType* assembler, const char* testname) { in EmitAndCheck()
|
| /art/compiler/utils/x86/ |
| D | assembler_x86_test.cc | 280 std::string rorl_fn(AssemblerX86Test::Base* assembler_test, x86::X86Assembler* assembler) { in rorl_fn() 303 std::string roll_fn(AssemblerX86Test::Base* assembler_test, x86::X86Assembler* assembler) { in roll_fn()
|
| D | jni_macro_assembler_x86.cc | 113 static void DecreaseFrameSizeImpl(X86Assembler* assembler, size_t adjust) { in DecreaseFrameSizeImpl()
|