/art/compiler/utils/arm/ |
D | jni_macro_assembler_arm_vixl.h | 191 void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) override; 210 ArmException(ArmManagedRegister scratch, size_t stack_adjust) in ArmException() argument 211 : scratch_(scratch), stack_adjust_(stack_adjust) {} in ArmException()
|
D | jni_macro_assembler_arm_vixl.cc | 612 void ArmVIXLJNIMacroAssembler::ExceptionPoll(ManagedRegister mscratch, size_t stack_adjust) { in ExceptionPoll() argument 613 CHECK_ALIGNED(stack_adjust, kStackAlignment); in ExceptionPoll() 618 new ArmVIXLJNIMacroAssembler::ArmException(mscratch.AsArm(), stack_adjust)); in ExceptionPoll()
|
/art/compiler/utils/arm64/ |
D | jni_macro_assembler_arm64.h | 172 void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) override; 186 Arm64Exception(Arm64ManagedRegister scratch, size_t stack_adjust) in Arm64Exception() argument 187 : scratch_(scratch), stack_adjust_(stack_adjust) {} in Arm64Exception()
|
D | jni_macro_assembler_arm64.cc | 622 void Arm64JNIMacroAssembler::ExceptionPoll(ManagedRegister m_scratch, size_t stack_adjust) { in ExceptionPoll() argument 623 CHECK_ALIGNED(stack_adjust, kStackAlignment); in ExceptionPoll() 625 exception_blocks_.emplace_back(new Arm64Exception(scratch, stack_adjust)); in ExceptionPoll()
|
/art/compiler/utils/x86/ |
D | jni_macro_assembler_x86.cc | 30 explicit X86ExceptionSlowPath(size_t stack_adjust) : stack_adjust_(stack_adjust) {} in X86ExceptionSlowPath() argument 519 void X86JNIMacroAssembler::ExceptionPoll(ManagedRegister /*scratch*/, size_t stack_adjust) { in ExceptionPoll() argument 520 X86ExceptionSlowPath* slow = new (__ GetAllocator()) X86ExceptionSlowPath(stack_adjust); in ExceptionPoll()
|
D | jni_macro_assembler_x86.h | 156 void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) override;
|
/art/compiler/utils/x86_64/ |
D | jni_macro_assembler_x86_64.cc | 579 explicit X86_64ExceptionSlowPath(size_t stack_adjust) : stack_adjust_(stack_adjust) {} in X86_64ExceptionSlowPath() argument 585 void X86_64JNIMacroAssembler::ExceptionPoll(ManagedRegister /*scratch*/, size_t stack_adjust) { in ExceptionPoll() argument 586 X86_64ExceptionSlowPath* slow = new (__ GetAllocator()) X86_64ExceptionSlowPath(stack_adjust); in ExceptionPoll()
|
D | jni_macro_assembler_x86_64.h | 182 void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) override;
|
/art/compiler/utils/ |
D | jni_macro_assembler.h | 207 virtual void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) = 0;
|
/art/compiler/optimizing/ |
D | intrinsics_x86.cc | 1456 int stack_adjust = kX86WordSize; in VisitStringGetCharsNoCheck() local 1457 __ cfi().AdjustCFAOffset(stack_adjust); in VisitStringGetCharsNoCheck() 1473 __ cfi().AdjustCFAOffset(stack_adjust); in VisitStringGetCharsNoCheck() 1510 __ cfi().AdjustCFAOffset(-stack_adjust); in VisitStringGetCharsNoCheck() 1514 __ cfi().AdjustCFAOffset(-stack_adjust); in VisitStringGetCharsNoCheck()
|
/art/compiler/utils/mips64/ |
D | assembler_mips64.h | 399 explicit Mips64ExceptionSlowPath(Mips64ManagedRegister scratch, size_t stack_adjust) in Mips64ExceptionSlowPath() argument 400 : scratch_(scratch), stack_adjust_(stack_adjust) {} in Mips64ExceptionSlowPath() 1434 void ExceptionPoll(ManagedRegister mscratch, size_t stack_adjust) override;
|
D | assembler_mips64.cc | 4069 void Mips64Assembler::ExceptionPoll(ManagedRegister mscratch, size_t stack_adjust) { in ExceptionPoll() argument 4071 exception_blocks_.emplace_back(scratch, stack_adjust); in ExceptionPoll()
|
/art/compiler/utils/mips/ |
D | assembler_mips.h | 248 explicit MipsExceptionSlowPath(MipsManagedRegister scratch, size_t stack_adjust) in MipsExceptionSlowPath() argument 249 : scratch_(scratch), stack_adjust_(stack_adjust) {} in MipsExceptionSlowPath() 1369 void ExceptionPoll(ManagedRegister mscratch, size_t stack_adjust) override;
|
D | assembler_mips.cc | 5232 void MipsAssembler::ExceptionPoll(ManagedRegister mscratch, size_t stack_adjust) { in ExceptionPoll() argument 5234 exception_blocks_.emplace_back(scratch, stack_adjust); in ExceptionPoll()
|