| /art/compiler/trampolines/ |
| D | trampoline_compiler.cc | 61 ArmVIXLAssembler assembler(allocator); in CreateTrampoline() local 96 Arm64Assembler assembler(allocator); in CreateTrampoline() local 131 Riscv64Assembler assembler(allocator); in CreateTrampoline() local 162 X86Assembler assembler(allocator); in CreateTrampoline() local 183 x86_64::X86_64Assembler assembler(allocator); in CreateTrampoline() local
|
| /art/compiler/optimizing/ |
| D | intrinsics_riscv64.cc | 43 Riscv64Assembler* assembler = codegen->GetAssembler(); in EmitNativeCode() local 105 Riscv64Assembler* assembler = codegen->GetAssembler(); in EmitNativeCode() local 211 Riscv64Assembler* assembler = GetAssembler(); in VisitDoubleDoubleToRawLongBits() local 221 Riscv64Assembler* assembler = GetAssembler(); in VisitDoubleLongBitsToDouble() local 231 Riscv64Assembler* assembler = GetAssembler(); in VisitFloatFloatToRawIntBits() local 241 Riscv64Assembler* assembler = GetAssembler(); in VisitFloatIntBitsToFloat() local 251 Riscv64Assembler* assembler = GetAssembler(); in VisitDoubleIsInfinite() local 264 Riscv64Assembler* assembler = GetAssembler(); in VisitFloatIsInfinite() local 289 Riscv64Assembler* assembler = GetAssembler(); in VisitMemoryPeekByte() local 298 Riscv64Assembler* assembler = GetAssembler(); in VisitMemoryPeekIntNative() local [all …]
|
| D | intrinsics_x86_64.cc | 77 static void GenArrayAddress(X86_64Assembler* assembler, in GenArrayAddress() 103 X86_64Assembler* assembler = x86_64_codegen->GetAssembler(); in EmitNativeCode() local 163 X86_64Assembler* assembler = x86_64_codegen->GetAssembler(); in EmitNativeCode() local 198 CpuRegister dst, XmmRegister src, bool is64bit, X86_64Assembler* assembler) { in MoveFPToInt() 207 XmmRegister dst, CpuRegister src, bool is64bit, X86_64Assembler* assembler) { in MoveIntToFP() 215 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in MoveFPToInt() 221 static void MoveIntToFP(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in MoveIntToFP() 289 X86_64Assembler* assembler = codegen->GetAssembler(); in GenIsInfinite() local 377 static void GenSSE41FPToFPIntrinsic(HInvoke* invoke, X86_64Assembler* assembler, int round_mode) { in GenSSE41FPToFPIntrinsic() 438 X86_64Assembler* assembler = GetAssembler(); in VisitMathRoundFloat() local [all …]
|
| D | intrinsics_arm_vixl.cc | 88 ArmVIXLAssembler* assembler = arm_codegen->GetAssembler(); in EmitNativeCode() local 179 static void MoveFPToInt(LocationSummary* locations, bool is64bit, ArmVIXLAssembler* assembler) { in MoveFPToInt() 189 static void MoveIntToFP(LocationSummary* locations, bool is64bit, ArmVIXLAssembler* assembler) { in MoveIntToFP() 260 ArmVIXLAssembler* assembler = codegen->GetAssembler(); in GenNumberOfLeadingZeros() local 305 ArmVIXLAssembler* assembler = codegen->GetAssembler(); in GenNumberOfTrailingZeros() local 351 ArmVIXLAssembler* assembler = GetAssembler(); in VisitMathSqrt() local 363 ArmVIXLAssembler* assembler = GetAssembler(); in VisitMathRint() local 380 ArmVIXLAssembler* assembler = GetAssembler(); in VisitMathRoundFloat() local 423 ArmVIXLAssembler* assembler = GetAssembler(); in VisitMemoryPeekByte() local 433 ArmVIXLAssembler* assembler = GetAssembler(); in VisitMemoryPeekIntNative() local [all …]
|
| D | intrinsics_x86.cc | 75 static void GenArrayAddress(X86Assembler* assembler, in GenArrayAddress() 100 X86Assembler* assembler = x86_codegen->GetAssembler(); in EmitNativeCode() local 172 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MoveFPToInt() 187 static void MoveIntToFP(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MoveIntToFP() 254 X86Assembler* assembler) { in GenReverseBytes() 293 X86Assembler* assembler = GetAssembler(); in VisitLongReverseBytes() local 339 static void GenSSE41FPToFPIntrinsic(HInvoke* invoke, X86Assembler* assembler, int round_mode) { in GenSSE41FPToFPIntrinsic() 400 X86Assembler* assembler = GetAssembler(); in VisitMathRoundFloat() local 459 X86Assembler* assembler = codegen->GetAssembler(); in GenFPToFPCall() local 496 static void GenLowestOneBit(X86Assembler* assembler, in GenLowestOneBit() [all …]
|
| D | jit_patches_arm64.h | 52 JitPatchesARM64(Arm64Assembler* assembler, ArenaAllocator* allocator) : in JitPatchesARM64()
|
| D | intrinsics_utils.h | 58 TAssembler* assembler = down_cast<TAssembler*>(codegen->GetAssembler()); in EmitNativeCode() local
|
| D | code_generator_arm64.cc | 1549 void CodeGeneratorARM64::PopFrameAndReturn(Arm64Assembler* assembler, in PopFrameAndReturn() 5599 Arm64Assembler assembler(GetGraph()->GetAllocator()); in EmitThunkCode() local 7443 static void EmitGrayCheckAndFastPath(arm64::Arm64Assembler& assembler, in EmitGrayCheckAndFastPath() 7482 static void LoadReadBarrierMarkIntrospectionEntrypoint(arm64::Arm64Assembler& assembler, in LoadReadBarrierMarkIntrospectionEntrypoint() 7491 void CodeGeneratorARM64::CompileBakerReadBarrierThunk(Arm64Assembler& assembler, in CompileBakerReadBarrierThunk()
|
| D | code_generator_arm_vixl.cc | 122 EmitAdrCode(ArmVIXLMacroAssembler* assembler, vixl32::Register rd, vixl32::Label* label) in EmitAdrCode() 6222 ArmVIXLAssembler* assembler = codegen->GetAssembler(); in CanEncode32BitConstantAsImmediate() local 10014 arm::ArmVIXLAssembler assembler(GetGraph()->GetAllocator()); in EmitThunkCode() local 10290 static void EmitGrayCheckAndFastPath(ArmVIXLAssembler& assembler, in EmitGrayCheckAndFastPath() 10317 static vixl32::Register LoadReadBarrierMarkIntrospectionEntrypoint(ArmVIXLAssembler& assembler) { in LoadReadBarrierMarkIntrospectionEntrypoint() 10329 void CodeGeneratorARMVIXL::CompileBakerReadBarrierThunk(ArmVIXLAssembler& assembler, in CompileBakerReadBarrierThunk()
|
| D | intrinsics_arm64.cc | 1332 Arm64Assembler* assembler = codegen->GetAssembler(); in EmitLoadExclusive() local 1386 Arm64Assembler* assembler = codegen->GetAssembler(); in EmitStoreExclusive() local 1449 Arm64Assembler* assembler = codegen->GetAssembler(); in GenerateCompareAndSet() local 1546 Arm64Assembler* assembler = arm64_codegen->GetAssembler(); in EmitNativeCode() local
|
| D | code_generator_riscv64.cc | 789 Riscv64Assembler* assembler = down_cast<CodeGeneratorRISCV64*>(codegen_)->GetAssembler(); in FpBinOp() local 848 Riscv64Assembler* assembler = down_cast<CodeGeneratorRISCV64*>(codegen_)->GetAssembler(); in FpUnOp() local
|
| D | code_generator_x86_64.cc | 8465 X86_64Assembler* assembler = codegen_->GetAssembler(); in CreateJumpTable() local 8494 X86_64Assembler* assembler = GetAssembler(); in Finalize() local
|
| D | code_generator_x86.cc | 9140 X86Assembler* assembler = codegen_->GetAssembler(); in CreateJumpTable() local 9170 X86Assembler* assembler = GetAssembler(); in Finalize() local
|
| /art/compiler/utils/x86_64/ |
| D | assembler_x86_64_test.cc | 654 std::string shll_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shll_fn() 675 std::string shlq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shlq_fn() 696 std::string shrl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shrl_fn() 716 std::string shrq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shrq_fn() 736 std::string sarl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in sarl_fn() 756 std::string sarq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in sarq_fn() 776 std::string rorl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rorl_fn() 796 std::string roll_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in roll_fn() 816 std::string rorq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rorq_fn() 836 std::string rolq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rolq_fn() [all …]
|
| D | jni_macro_assembler_x86_64.cc | 144 static void DecreaseFrameSizeImpl(size_t adjust, X86_64Assembler* assembler) { in DecreaseFrameSizeImpl()
|
| /art/compiler/utils/riscv64/ |
| D | assembler_riscv64.h | 2750 ScopedExtensionsOverride(Riscv64Assembler* assembler, Riscv64ExtensionMask enabled_extensions) in ScopedExtensionsOverride() 2761 static Riscv64ExtensionMask GetEnabledExtensions(Riscv64Assembler* assembler) { in GetEnabledExtensions() 2773 explicit ScopedExtensionsRestriction(Riscv64Assembler* assembler) in ScopedExtensionsRestriction() 2780 explicit ScopedExtensionsInclusion(Riscv64Assembler* assembler) in ScopedExtensionsInclusion() 2792 explicit ScratchRegisterScope(Riscv64Assembler* assembler) in ScratchRegisterScope()
|
| D | assembler_riscv64.cc | 6428 void AddConstImpl(Riscv64Assembler* assembler, in AddConstImpl()
|
| /art/compiler/utils/ |
| D | assembler_thumb_test.cc | 93 ArmVIXLJNIMacroAssembler assembler; member in art::arm::ArmVIXLAssemblerTest
|
| /art/compiler/utils/x86/ |
| D | jni_macro_assembler_x86.cc | 115 static void DecreaseFrameSizeImpl(X86Assembler* assembler, size_t adjust) { in DecreaseFrameSizeImpl()
|
| D | assembler_x86_test.cc | 495 std::string rorl_fn(AssemblerX86Test::Base* assembler_test, x86::X86Assembler* assembler) { in rorl_fn() 515 std::string roll_fn(AssemblerX86Test::Base* assembler_test, x86::X86Assembler* assembler) { in roll_fn()
|