// Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file // for details. All rights reserved. Use of this source code is governed by a // BSD-style license that can be found in the LICENSE file. // // This is forked from Dart revision df52deea9f25690eb8b66c5995da92b70f7ac1fe // Please update the (git) revision if we merge changes from Dart. // https://code.google.com/p/dart/wiki/GettingTheSource #ifndef VM_ASSEMBLER_ARM_H_ #define VM_ASSEMBLER_ARM_H_ #ifndef VM_ASSEMBLER_H_ #error Do not include assembler_arm.h directly; use assembler.h instead. #endif #include "platform/assert.h" #include "platform/utils.h" #include "vm/constants_arm.h" #include "vm/cpu.h" #include "vm/hash_map.h" #include "vm/object.h" #include "vm/simulator.h" namespace dart { // Forward declarations. class RuntimeEntry; class StubEntry; #if 0 // Moved to ARM32::AssemblerARM32 as needed // Instruction encoding bits. enum { H = 1 << 5, // halfword (or byte) L = 1 << 20, // load (or store) S = 1 << 20, // set condition code (or leave unchanged) W = 1 << 21, // writeback base register (or leave unchanged) A = 1 << 21, // accumulate in multiply instruction (or not) B = 1 << 22, // unsigned byte (or word) D = 1 << 22, // high/lo bit of start of s/d register range N = 1 << 22, // long (or short) U = 1 << 23, // positive (or negative) offset/index P = 1 << 24, // offset/pre-indexed addressing (or post-indexed addressing) I = 1 << 25, // immediate shifter operand (or not) B0 = 1, B1 = 1 << 1, B2 = 1 << 2, B3 = 1 << 3, B4 = 1 << 4, B5 = 1 << 5, B6 = 1 << 6, B7 = 1 << 7, B8 = 1 << 8, B9 = 1 << 9, B10 = 1 << 10, B11 = 1 << 11, B12 = 1 << 12, B16 = 1 << 16, B17 = 1 << 17, B18 = 1 << 18, B19 = 1 << 19, B20 = 1 << 20, B21 = 1 << 21, B22 = 1 << 22, B23 = 1 << 23, B24 = 1 << 24, B25 = 1 << 25, B26 = 1 << 26, B27 = 1 << 27, }; #endif class Label : public ValueObject { public: Label() : position_(0) { } ~Label() { // Assert if label is being destroyed with unresolved branches pending. ASSERT(!IsLinked()); } // Returns the position for bound and linked labels. Cannot be used // for unused labels. intptr_t Position() const { ASSERT(!IsUnused()); return IsBound() ? -position_ - kWordSize : position_ - kWordSize; } bool IsBound() const { return position_ < 0; } bool IsUnused() const { return position_ == 0; } bool IsLinked() const { return position_ > 0; } private: intptr_t position_; void Reinitialize() { position_ = 0; } void BindTo(intptr_t position) { ASSERT(!IsBound()); position_ = -position - kWordSize; ASSERT(IsBound()); } void LinkTo(intptr_t position) { ASSERT(!IsBound()); position_ = position + kWordSize; ASSERT(IsLinked()); } friend class Assembler; DISALLOW_COPY_AND_ASSIGN(Label); }; // Encodes Addressing Mode 1 - Data-processing operands. class Operand : public ValueObject { public: // Data-processing operands - Uninitialized. Operand() : type_(-1), encoding_(-1) { } // Data-processing operands - Copy constructor. Operand(const Operand& other) : ValueObject(), type_(other.type_), encoding_(other.encoding_) { } // Data-processing operands - Assignment operator. Operand& operator=(const Operand& other) { type_ = other.type_; encoding_ = other.encoding_; return *this; } #if 0 // Moved to encodeRotatedImm8() in IceAssemblerARM32.cpp // Data-processing operands - Immediate. explicit Operand(uint32_t immediate) { ASSERT(immediate < (1 << kImmed8Bits)); type_ = 1; encoding_ = immediate; } // Moved to decodeOperand() and encodeRotatedImm8() in IceAssemblerARM32.cpp // Data-processing operands - Rotated immediate. Operand(uint32_t rotate, uint32_t immed8) { ASSERT((rotate < (1 << kRotateBits)) && (immed8 < (1 << kImmed8Bits))); type_ = 1; encoding_ = (rotate << kRotateShift) | (immed8 << kImmed8Shift); } // Moved to decodeOperand() in IceAssemblerARM32.cpp // Data-processing operands - Register. explicit Operand(Register rm) { type_ = 0; encoding_ = static_cast(rm); } // Moved to encodeShiftRotateImm5() in IceAssemblerARM32.cpp // Data-processing operands - Logical shift/rotate by immediate. Operand(Register rm, Shift shift, uint32_t shift_imm) { ASSERT(shift_imm < (1 << kShiftImmBits)); type_ = 0; encoding_ = shift_imm << kShiftImmShift | static_cast(shift) << kShiftShift | static_cast(rm); } // Moved to encodeShiftRotateReg() in IceAssemblerARM32.cpp // Data-processing operands - Logical shift/rotate by register. Operand(Register rm, Shift shift, Register rs) { type_ = 0; encoding_ = static_cast(rs) << kShiftRegisterShift | static_cast(shift) << kShiftShift | (1 << 4) | static_cast(rm); } // Already defined as ARM32::OperandARM32FlexImm::canHoldImm(). static bool CanHold(uint32_t immediate, Operand* o) { // Avoid the more expensive test for frequent small immediate values. if (immediate < (1 << kImmed8Bits)) { o->type_ = 1; o->encoding_ = (0 << kRotateShift) | (immediate << kImmed8Shift); return true; } // Note that immediate must be unsigned for the test to work correctly. for (int rot = 0; rot < 16; rot++) { uint32_t imm8 = (immediate << 2*rot) | (immediate >> (32 - 2*rot)); if (imm8 < (1 << kImmed8Bits)) { o->type_ = 1; o->encoding_ = (rot << kRotateShift) | (imm8 << kImmed8Shift); return true; } } return false; } #endif private: bool is_valid() const { return (type_ == 0) || (type_ == 1); } uint32_t type() const { ASSERT(is_valid()); return type_; } uint32_t encoding() const { ASSERT(is_valid()); return encoding_; } uint32_t type_; // Encodes the type field (bits 27-25) in the instruction. uint32_t encoding_; friend class Assembler; friend class Address; }; enum OperandSize { kByte, kUnsignedByte, kHalfword, kUnsignedHalfword, kWord, kUnsignedWord, kWordPair, kSWord, kDWord, kRegList, }; // Load/store multiple addressing mode. enum BlockAddressMode { // bit encoding P U W DA = (0|0|0) << 21, // decrement after IA = (0|4|0) << 21, // increment after DB = (8|0|0) << 21, // decrement before IB = (8|4|0) << 21, // increment before DA_W = (0|0|1) << 21, // decrement after with writeback to base IA_W = (0|4|1) << 21, // increment after with writeback to base DB_W = (8|0|1) << 21, // decrement before with writeback to base IB_W = (8|4|1) << 21 // increment before with writeback to base }; class Address : public ValueObject { public: enum OffsetKind { Immediate, IndexRegister, ScaledIndexRegister, }; // Memory operand addressing mode enum Mode { kModeMask = (8|4|1) << 21, // bit encoding P U W Offset = (8|4|0) << 21, // offset (w/o writeback to base) PreIndex = (8|4|1) << 21, // pre-indexed addressing with writeback PostIndex = (0|4|0) << 21, // post-indexed addressing with writeback NegOffset = (8|0|0) << 21, // negative offset (w/o writeback to base) NegPreIndex = (8|0|1) << 21, // negative pre-indexed with writeback NegPostIndex = (0|0|0) << 21 // negative post-indexed with writeback }; Address(const Address& other) : ValueObject(), encoding_(other.encoding_), kind_(other.kind_) { } Address& operator=(const Address& other) { encoding_ = other.encoding_; kind_ = other.kind_; return *this; } bool Equals(const Address& other) const { return (encoding_ == other.encoding_) && (kind_ == other.kind_); } #if 0 // Moved to decodeImmRegOffset() in IceAssemblerARM32.cpp. // Used to model stack offsets. explicit Address(Register rn, int32_t offset = 0, Mode am = Offset) { ASSERT(Utils::IsAbsoluteUint(12, offset)); kind_ = Immediate; if (offset < 0) { encoding_ = (am ^ (1 << kUShift)) | -offset; // Flip U to adjust sign. } else { encoding_ = am | offset; } encoding_ |= static_cast(rn) << kRnShift; } #endif // There is no register offset mode unless Mode is Offset, in which case the // shifted register case below should be used. Address(Register rn, Register r, Mode am); Address(Register rn, Register rm, Shift shift = LSL, uint32_t shift_imm = 0, Mode am = Offset) { Operand o(rm, shift, shift_imm); if ((shift == LSL) && (shift_imm == 0)) { kind_ = IndexRegister; } else { kind_ = ScaledIndexRegister; } encoding_ = o.encoding() | am | (static_cast(rn) << kRnShift); } // There is no shifted register mode with a register shift. Address(Register rn, Register rm, Shift shift, Register r, Mode am = Offset); static OperandSize OperandSizeFor(intptr_t cid); static bool CanHoldLoadOffset(OperandSize size, int32_t offset, int32_t* offset_mask); static bool CanHoldStoreOffset(OperandSize size, int32_t offset, int32_t* offset_mask); static bool CanHoldImmediateOffset(bool is_load, intptr_t cid, int64_t offset); private: Register rn() const { return Instr::At(reinterpret_cast(&encoding_))->RnField(); } Register rm() const { return ((kind() == IndexRegister) || (kind() == ScaledIndexRegister)) ? Instr::At(reinterpret_cast(&encoding_))->RmField() : kNoRegister; } Mode mode() const { return static_cast(encoding() & kModeMask); } uint32_t encoding() const { return encoding_; } #if 0 // Moved to encodeImmRegOffsetEnc3 in IceAssemblerARM32.cpp // Encoding for addressing mode 3. uint32_t encoding3() const; #endif // Encoding for vfp load/store addressing. uint32_t vencoding() const; OffsetKind kind() const { return kind_; } uint32_t encoding_; OffsetKind kind_; friend class Assembler; }; class FieldAddress : public Address { public: FieldAddress(Register base, int32_t disp) : Address(base, disp - kHeapObjectTag) { } // This addressing mode does not exist. FieldAddress(Register base, Register r); FieldAddress(const FieldAddress& other) : Address(other) { } FieldAddress& operator=(const FieldAddress& other) { Address::operator=(other); return *this; } }; class Assembler : public ValueObject { public: explicit Assembler(bool use_far_branches = false) : buffer_(), prologue_offset_(-1), use_far_branches_(use_far_branches), comments_(), constant_pool_allowed_(false) { } ~Assembler() { } void PopRegister(Register r) { Pop(r); } void Bind(Label* label); void Jump(Label* label) { b(label); } // Misc. functionality intptr_t CodeSize() const { return buffer_.Size(); } intptr_t prologue_offset() const { return prologue_offset_; } // Count the fixups that produce a pointer offset, without processing // the fixups. On ARM there are no pointers in code. intptr_t CountPointerOffsets() const { return 0; } const ZoneGrowableArray& GetPointerOffsets() const { ASSERT(buffer_.pointer_offsets().length() == 0); // No pointers in code. return buffer_.pointer_offsets(); } ObjectPoolWrapper& object_pool_wrapper() { return object_pool_wrapper_; } RawObjectPool* MakeObjectPool() { return object_pool_wrapper_.MakeObjectPool(); } bool use_far_branches() const { return FLAG_use_far_branches || use_far_branches_; } #if defined(TESTING) || defined(DEBUG) // Used in unit tests and to ensure predictable verification code size in // FlowGraphCompiler::EmitEdgeCounter. void set_use_far_branches(bool b) { use_far_branches_ = b; } #endif // TESTING || DEBUG void FinalizeInstructions(const MemoryRegion& region) { buffer_.FinalizeInstructions(region); } // Debugging and bringup support. void Stop(const char* message); void Unimplemented(const char* message); void Untested(const char* message); void Unreachable(const char* message); static void InitializeMemoryWithBreakpoints(uword data, intptr_t length); void Comment(const char* format, ...) PRINTF_ATTRIBUTE(2, 3); static bool EmittingComments(); const Code::Comments& GetCodeComments() const; static const char* RegisterName(Register reg); static const char* FpuRegisterName(FpuRegister reg); #if 0 // Moved to ARM32::AssemblerARM32::and_() // Data-processing instructions. void and_(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemblerARM32::eor() void eor(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemberARM32::sub() void sub(Register rd, Register rn, Operand o, Condition cond = AL); void subs(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemberARM32::rsb() void rsb(Register rd, Register rn, Operand o, Condition cond = AL); void rsbs(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemblerARM32::add() void add(Register rd, Register rn, Operand o, Condition cond = AL); void adds(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemblerARM32::adc() void adc(Register rd, Register rn, Operand o, Condition cond = AL); void adcs(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemblerARM32::sbc() void sbc(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemblerARM32::sbc() void sbcs(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemblerARM32::rsc() void rsc(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemblerARM32::tst(); void tst(Register rn, Operand o, Condition cond = AL); #endif void teq(Register rn, Operand o, Condition cond = AL); #if 0 // Moved to ARM32::AssemblerARM32::cmp() void cmp(Register rn, Operand o, Condition cond = AL); // Moved to ARM32::AssemblerARM32::cmn() void cmn(Register rn, Operand o, Condition cond = AL); // Moved to ARM32::IceAssemblerARM32::orr(). void orr(Register rd, Register rn, Operand o, Condition cond = AL); void orrs(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::IceAssemblerARM32::mov() void mov(Register rd, Operand o, Condition cond = AL); void movs(Register rd, Operand o, Condition cond = AL); // Moved to ARM32::IceAssemblerARM32::bic() void bic(Register rd, Register rn, Operand o, Condition cond = AL); void bics(Register rd, Register rn, Operand o, Condition cond = AL); // Moved to ARM32::IceAssemblerARM32::mvn() void mvn(Register rd, Operand o, Condition cond = AL); void mvns(Register rd, Operand o, Condition cond = AL); // Miscellaneous data-processing instructions. // Moved to ARM32::AssemblerARM32::clz() void clz(Register rd, Register rm, Condition cond = AL); // Multiply instructions. // Moved to ARM32::AssemblerARM32::mul() void mul(Register rd, Register rn, Register rm, Condition cond = AL); void muls(Register rd, Register rn, Register rm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::mla() void mla(Register rd, Register rn, Register rm, Register ra, Condition cond = AL); // Moved to ARM32::AssemblerARM32::mls() void mls(Register rd, Register rn, Register rm, Register ra, Condition cond = AL); #endif void smull(Register rd_lo, Register rd_hi, Register rn, Register rm, Condition cond = AL); #if 0 // Moved to ARM32::AssemblerARM32::umull(); void umull(Register rd_lo, Register rd_hi, Register rn, Register rm, Condition cond = AL); #endif void smlal(Register rd_lo, Register rd_hi, Register rn, Register rm, Condition cond = AL); void umlal(Register rd_lo, Register rd_hi, Register rn, Register rm, Condition cond = AL); // Emulation of this instruction uses IP and the condition codes. Therefore, // none of the registers can be IP, and the instruction can only be used // unconditionally. void umaal(Register rd_lo, Register rd_hi, Register rn, Register rm); // Division instructions. #if 0 // Moved to ARM32::AssemblerARM32::sdiv() void sdiv(Register rd, Register rn, Register rm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::udiv() void udiv(Register rd, Register rn, Register rm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::ldr() // Load/store instructions. void ldr(Register rd, Address ad, Condition cond = AL); // Moved to ARM32::AssemblerARM32::str() void str(Register rd, Address ad, Condition cond = AL); // Moved to ARM32::AssemblerARM32::ldr() void ldrb(Register rd, Address ad, Condition cond = AL); // Moved to ARM32::AssemblerARM32::str() void strb(Register rd, Address ad, Condition cond = AL); // Moved to ARM32::AssemblerARM32::ldr() void ldrh(Register rd, Address ad, Condition cond = AL); // Moved to ARM32::AssemblerARM32::str() void strh(Register rd, Address ad, Condition cond = AL); #endif void ldrsb(Register rd, Address ad, Condition cond = AL); void ldrsh(Register rd, Address ad, Condition cond = AL); // ldrd and strd actually support the full range of addressing modes, but // we don't use them, and we need to split them up into two instructions for // ARMv5TE, so we only support the base + offset mode. void ldrd(Register rd, Register rn, int32_t offset, Condition cond = AL); void strd(Register rd, Register rn, int32_t offset, Condition cond = AL); #if 0 // Folded into ARM32::AssemblerARM32::popList(), since it is its only use (and // doesn't implement ARM LDM instructions). void ldm(BlockAddressMode am, Register base, RegList regs, Condition cond = AL); // Folded into ARM32::AssemblerARM32::pushList(), since it is its only use // (and doesn't implement ARM STM instruction). void stm(BlockAddressMode am, Register base, RegList regs, Condition cond = AL); // Moved to ARM::AssemblerARM32::ldrex(); void ldrex(Register rd, Register rn, Condition cond = AL); // Moved to ARM::AssemblerARM32::strex(); void strex(Register rd, Register rt, Register rn, Condition cond = AL); #endif // Miscellaneous instructions. void clrex(); #if 0 // Moved to ARM32::AssemblerARM32::nop(). void nop(Condition cond = AL); // Moved to ARM32::AssemblerARM32::bkpt() // Note that gdb sets breakpoints using the undefined instruction 0xe7f001f0. void bkpt(uint16_t imm16); static int32_t BkptEncoding(uint16_t imm16) { // bkpt requires that the cond field is AL. return (AL << kConditionShift) | B24 | B21 | ((imm16 >> 4) << 8) | B6 | B5 | B4 | (imm16 & 0xf); } // Not ported. PNaCl doesn't allow breakpoint instructions. static uword GetBreakInstructionFiller() { return BkptEncoding(0); } // Floating point instructions (VFPv3-D16 and VFPv3-D32 profiles). // Moved to ARM32::AssemblerARM32::vmovsr(). void vmovsr(SRegister sn, Register rt, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vmovrs(). void vmovrs(Register rt, SRegister sn, Condition cond = AL); #endif void vmovsrr(SRegister sm, Register rt, Register rt2, Condition cond = AL); void vmovrrs(Register rt, Register rt2, SRegister sm, Condition cond = AL); #if 0 // Moved to ARM32::AssemblerARM32::vmovdrr(). void vmovdrr(DRegister dm, Register rt, Register rt2, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vmovrrd(). void vmovrrd(Register rt, Register rt2, DRegister dm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vmovqir(). void vmovdr(DRegister dd, int i, Register rt, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vmovss(). void vmovs(SRegister sd, SRegister sm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vmovdd(). void vmovd(DRegister dd, DRegister dm, Condition cond = AL); #endif void vmovq(QRegister qd, QRegister qm); #if 0 // Returns false if the immediate cannot be encoded. // Moved to ARM32::AssemblerARM32::vmovs(); bool vmovs(SRegister sd, float s_imm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vmovs(); bool vmovd(DRegister dd, double d_imm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vldrs() void vldrs(SRegister sd, Address ad, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vstrs() void vstrs(SRegister sd, Address ad, Condition cond = AL); #endif // Moved to ARM32::AssemblerARM32::vldrd() void vldrd(DRegister dd, Address ad, Condition cond = AL); #if 0 // Moved to Arm32::AssemblerARM32::vstrd() void vstrd(DRegister dd, Address ad, Condition cond = AL); #endif void vldms(BlockAddressMode am, Register base, SRegister first, SRegister last, Condition cond = AL); void vstms(BlockAddressMode am, Register base, SRegister first, SRegister last, Condition cond = AL); void vldmd(BlockAddressMode am, Register base, DRegister first, intptr_t count, Condition cond = AL); void vstmd(BlockAddressMode am, Register base, DRegister first, intptr_t count, Condition cond = AL); #if 0 // Moved to Arm32::AssemblerARM32::vadds() void vadds(SRegister sd, SRegister sn, SRegister sm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vaddd() void vaddd(DRegister dd, DRegister dn, DRegister dm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vaddqi(). void vaddqi(OperandSize sz, QRegister qd, QRegister qn, QRegister qm); // Moved to ARM32::AssemblerARM32::vaddqf(). void vaddqs(QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vsubs() void vsubs(SRegister sd, SRegister sn, SRegister sm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vsubd() void vsubd(DRegister dd, DRegister dn, DRegister dm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vsubqi(). void vsubqi(OperandSize sz, QRegister qd, QRegister qn, QRegister qm); // Moved to ARM32::AssemblerARM32::vsubqf(). void vsubqs(QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vmuls() void vmuls(SRegister sd, SRegister sn, SRegister sm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vmuld() void vmuld(DRegister dd, DRegister dn, DRegister dm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vmulqi(). void vmulqi(OperandSize sz, QRegister qd, QRegister qn, QRegister qm); // Moved to ARM32::AssemblerARM32::vmulqf(). void vmulqs(QRegister qd, QRegister qn, QRegister qm); // Moved to ARM32::AssemblerARM32::vshlqi(). void vshlqi(OperandSize sz, QRegister qd, QRegister qm, QRegister qn); // Moved to ARM32::AssemblerARM32::vshlqu(). void vshlqu(OperandSize sz, QRegister qd, QRegister qm, QRegister qn); // Moved to Arm32::AssemblerARM32::vmlas() void vmlas(SRegister sd, SRegister sn, SRegister sm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vmlad() void vmlad(DRegister dd, DRegister dn, DRegister dm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vmlss() void vmlss(SRegister sd, SRegister sn, SRegister sm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vmlsd() void vmlsd(DRegister dd, DRegister dn, DRegister dm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vdivs() void vdivs(SRegister sd, SRegister sn, SRegister sm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vdivd() void vdivd(DRegister dd, DRegister dn, DRegister dm, Condition cond = AL); #endif void vminqs(QRegister qd, QRegister qn, QRegister qm); void vmaxqs(QRegister qd, QRegister qn, QRegister qm); void vrecpeqs(QRegister qd, QRegister qm); void vrecpsqs(QRegister qd, QRegister qn, QRegister qm); void vrsqrteqs(QRegister qd, QRegister qm); void vrsqrtsqs(QRegister qd, QRegister qn, QRegister qm); #if 0 // Moved to ARM32::AssemblerARM32::vorrq() void veorq(QRegister qd, QRegister qn, QRegister qm); // Moved to ARM32::AssemblerARM32::vorrq() void vorrq(QRegister qd, QRegister qn, QRegister qm); #endif void vornq(QRegister qd, QRegister qn, QRegister qm); #if 0 // Moved to Arm32::AssemblerARM32::vandq(). void vandq(QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vandq(). void vmvnq(QRegister qd, QRegister qm); // Moved to Arm32::AssemblerARM32::vceqqi(). void vceqqi(OperandSize sz, QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vceqqs(). void vceqqs(QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vcgeqi(). void vcgeqi(OperandSize sz, QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vcugeqi(). void vcugeqi(OperandSize sz, QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vcgeqs(). void vcgeqs(QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vcgtqi(). void vcgtqi(OperandSize sz, QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vcugtqi(). void vcugtqi(OperandSize sz, QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vcgtqs(). void vcgtqs(QRegister qd, QRegister qn, QRegister qm); // Moved to Arm32::AssemblerARM32::vabss(). void vabss(SRegister sd, SRegister sm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vabsd(). void vabsd(DRegister dd, DRegister dm, Condition cond = AL); // Moved to Arm32::AssemblerARM32::vabsq(). void vabsqs(QRegister qd, QRegister qm); #endif void vnegs(SRegister sd, SRegister sm, Condition cond = AL); void vnegd(DRegister dd, DRegister dm, Condition cond = AL); #if 0 // Moved to ARM32::AssemblerARM32::vnegqs(). void vnegqs(QRegister qd, QRegister qm); // Moved to ARM32::AssemblerARM32::vsqrts(). void vsqrts(SRegister sd, SRegister sm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vsqrts(). void vsqrtd(DRegister dd, DRegister dm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtsd(). void vcvtsd(SRegister sd, DRegister dm, Condition cond = AL); // Moved to ARM32::AssemblerARM32:vcvtds(). void vcvtds(DRegister dd, SRegister sm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtis() void vcvtis(SRegister sd, SRegister sm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtid() void vcvtid(SRegister sd, DRegister dm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtsi() void vcvtsi(SRegister sd, SRegister sm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtdi() void vcvtdi(DRegister dd, SRegister sm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtus(). void vcvtus(SRegister sd, SRegister sm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtud(). void vcvtud(SRegister sd, DRegister dm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtsu() void vcvtsu(SRegister sd, SRegister sm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::vcvtdu() void vcvtdu(DRegister dd, SRegister sm, Condition cond = AL); // Moved to ARM23::AssemblerARM32::vcmps(). void vcmps(SRegister sd, SRegister sm, Condition cond = AL); // Moved to ARM23::AssemblerARM32::vcmpd(). void vcmpd(DRegister dd, DRegister dm, Condition cond = AL); // Moved to ARM23::AssemblerARM32::vcmpsz(). void vcmpsz(SRegister sd, Condition cond = AL); // Moved to ARM23::AssemblerARM32::vcmpdz(). void vcmpdz(DRegister dd, Condition cond = AL); // APSR_nzcv version moved to ARM32::AssemblerARM32::vmrsAPSR_nzcv() void vmrs(Register rd, Condition cond = AL); #endif void vmstat(Condition cond = AL); // Duplicates the operand of size sz at index idx from dm to all elements of // qd. This is a special case of vtbl. void vdup(OperandSize sz, QRegister qd, DRegister dm, int idx); // Each byte of dm is an index into the table of bytes formed by concatenating // a list of 'length' registers starting with dn. The result is placed in dd. void vtbl(DRegister dd, DRegister dn, int length, DRegister dm); // The words of qd and qm are interleaved with the low words of the result // in qd and the high words in qm. void vzipqw(QRegister qd, QRegister qm); // Branch instructions. #if 0 // Moved to ARM32::AssemblerARM32::b(); void b(Label* label, Condition cond = AL); // Moved to ARM32::AssemblerARM32::bl() void bl(Label* label, Condition cond = AL); // Moved to ARM32::AssemblerARM32::bx() void bx(Register rm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::blx() void blx(Register rm, Condition cond = AL); #endif void Branch(const StubEntry& stub_entry, Patchability patchable = kNotPatchable, Register pp = PP, Condition cond = AL); void BranchLink(const StubEntry& stub_entry, Patchability patchable = kNotPatchable); void BranchLink(const Code& code, Patchability patchable); // Branch and link to an entry address. Call sequence can be patched. void BranchLinkPatchable(const StubEntry& stub_entry); void BranchLinkPatchable(const Code& code); // Branch and link to [base + offset]. Call sequence is never patched. void BranchLinkOffset(Register base, int32_t offset); // Add signed immediate value to rd. May clobber IP. void AddImmediate(Register rd, int32_t value, Condition cond = AL); void AddImmediate(Register rd, Register rn, int32_t value, Condition cond = AL); void AddImmediateSetFlags(Register rd, Register rn, int32_t value, Condition cond = AL); void SubImmediateSetFlags(Register rd, Register rn, int32_t value, Condition cond = AL); void AndImmediate(Register rd, Register rs, int32_t imm, Condition cond = AL); // Test rn and immediate. May clobber IP. void TestImmediate(Register rn, int32_t imm, Condition cond = AL); // Compare rn with signed immediate value. May clobber IP. void CompareImmediate(Register rn, int32_t value, Condition cond = AL); // Signed integer division of left by right. Checks to see if integer // division is supported. If not, uses the FPU for division with // temporary registers tmpl and tmpr. tmpl and tmpr must be different // registers. void IntegerDivide(Register result, Register left, Register right, DRegister tmpl, DRegister tmpr); // Load and Store. // These three do not clobber IP. void LoadPatchableImmediate(Register rd, int32_t value, Condition cond = AL); void LoadDecodableImmediate(Register rd, int32_t value, Condition cond = AL); void LoadImmediate(Register rd, int32_t value, Condition cond = AL); // These two may clobber IP. void LoadSImmediate(SRegister sd, float value, Condition cond = AL); void LoadDImmediate(DRegister dd, double value, Register scratch, Condition cond = AL); void MarkExceptionHandler(Label* label); void Drop(intptr_t stack_elements); void RestoreCodePointer(); void LoadPoolPointer(Register reg = PP); void LoadIsolate(Register rd); void LoadObject(Register rd, const Object& object, Condition cond = AL); void LoadUniqueObject(Register rd, const Object& object, Condition cond = AL); void LoadFunctionFromCalleePool(Register dst, const Function& function, Register new_pp); void LoadNativeEntry(Register dst, const ExternalLabel* label, Patchability patchable, Condition cond = AL); void PushObject(const Object& object); void CompareObject(Register rn, const Object& object); // When storing into a heap object field, knowledge of the previous content // is expressed through these constants. enum FieldContent { kEmptyOrSmiOrNull, // Empty = garbage/zapped in release/debug mode. kHeapObjectOrSmi, kOnlySmi, }; void StoreIntoObject(Register object, // Object we are storing into. const Address& dest, // Where we are storing into. Register value, // Value we are storing. bool can_value_be_smi = true); void StoreIntoObjectOffset(Register object, int32_t offset, Register value, bool can_value_be_smi = true); void StoreIntoObjectNoBarrier(Register object, const Address& dest, Register value, FieldContent old_content = kHeapObjectOrSmi); void InitializeFieldNoBarrier(Register object, const Address& dest, Register value) { StoreIntoObjectNoBarrier(object, dest, value, kEmptyOrSmiOrNull); } void StoreIntoObjectNoBarrierOffset( Register object, int32_t offset, Register value, FieldContent old_content = kHeapObjectOrSmi); void StoreIntoObjectNoBarrier(Register object, const Address& dest, const Object& value, FieldContent old_content = kHeapObjectOrSmi); void StoreIntoObjectNoBarrierOffset( Register object, int32_t offset, const Object& value, FieldContent old_content = kHeapObjectOrSmi); // Store value_even, value_odd, value_even, ... into the words in the address // range [begin, end), assumed to be uninitialized fields in object (tagged). // The stores must not need a generational store barrier (e.g., smi/null), // and (value_even, value_odd) must be a valid register pair. // Destroys register 'begin'. void InitializeFieldsNoBarrier(Register object, Register begin, Register end, Register value_even, Register value_odd); // Like above, for the range [base+begin_offset, base+end_offset), unrolled. void InitializeFieldsNoBarrierUnrolled(Register object, Register base, intptr_t begin_offset, intptr_t end_offset, Register value_even, Register value_odd); // Stores a Smi value into a heap object field that always contains a Smi. void StoreIntoSmiField(const Address& dest, Register value); void LoadClassId(Register result, Register object, Condition cond = AL); void LoadClassById(Register result, Register class_id); void LoadClass(Register result, Register object, Register scratch); void CompareClassId(Register object, intptr_t class_id, Register scratch); void LoadClassIdMayBeSmi(Register result, Register object); void LoadTaggedClassIdMayBeSmi(Register result, Register object); void ComputeRange(Register result, Register value, Register scratch, Label* miss); void UpdateRangeFeedback(Register value, intptr_t idx, Register ic_data, Register scratch1, Register scratch2, Label* miss); intptr_t FindImmediate(int32_t imm); bool CanLoadFromObjectPool(const Object& object) const; void LoadFromOffset(OperandSize type, Register reg, Register base, int32_t offset, Condition cond = AL); void LoadFieldFromOffset(OperandSize type, Register reg, Register base, int32_t offset, Condition cond = AL) { LoadFromOffset(type, reg, base, offset - kHeapObjectTag, cond); } void StoreToOffset(OperandSize type, Register reg, Register base, int32_t offset, Condition cond = AL); void LoadSFromOffset(SRegister reg, Register base, int32_t offset, Condition cond = AL); void StoreSToOffset(SRegister reg, Register base, int32_t offset, Condition cond = AL); void LoadDFromOffset(DRegister reg, Register base, int32_t offset, Condition cond = AL); void StoreDToOffset(DRegister reg, Register base, int32_t offset, Condition cond = AL); void LoadMultipleDFromOffset(DRegister first, intptr_t count, Register base, int32_t offset); void StoreMultipleDToOffset(DRegister first, intptr_t count, Register base, int32_t offset); void CopyDoubleField(Register dst, Register src, Register tmp1, Register tmp2, DRegister dtmp); void CopyFloat32x4Field(Register dst, Register src, Register tmp1, Register tmp2, DRegister dtmp); void CopyFloat64x2Field(Register dst, Register src, Register tmp1, Register tmp2, DRegister dtmp); #if 0 // Moved to ARM32::AssemblerARM32::push(). void Push(Register rd, Condition cond = AL); // Moved to ARM32::AssemblerARM32::pop(). void Pop(Register rd, Condition cond = AL); // Moved to ARM32::AssemblerARM32::pushList(). void PushList(RegList regs, Condition cond = AL); // Moved to ARM32::AssemblerARM32::popList(). void PopList(RegList regs, Condition cond = AL); #endif void MoveRegister(Register rd, Register rm, Condition cond = AL); // Convenience shift instructions. Use mov instruction with shifter operand // for variants setting the status flags. #if 0 // Moved to ARM32::AssemblerARM32::lsl() void Lsl(Register rd, Register rm, const Operand& shift_imm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::lsl() void Lsl(Register rd, Register rm, Register rs, Condition cond = AL); // Moved to ARM32::AssemblerARM32::lsr() void Lsr(Register rd, Register rm, const Operand& shift_imm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::lsr() void Lsr(Register rd, Register rm, Register rs, Condition cond = AL); // Moved to ARM32::AssemblerARM32::asr() void Asr(Register rd, Register rm, const Operand& shift_imm, Condition cond = AL); // Moved to ARM32::AssemblerARM32::asr() void Asr(Register rd, Register rm, Register rs, Condition cond = AL); #endif void Asrs(Register rd, Register rm, const Operand& shift_imm, Condition cond = AL); void Ror(Register rd, Register rm, const Operand& shift_imm, Condition cond = AL); void Ror(Register rd, Register rm, Register rs, Condition cond = AL); void Rrx(Register rd, Register rm, Condition cond = AL); // Fill rd with the sign of rm. void SignFill(Register rd, Register rm, Condition cond = AL); void Vreciprocalqs(QRegister qd, QRegister qm); void VreciprocalSqrtqs(QRegister qd, QRegister qm); // If qm must be preserved, then provide a (non-QTMP) temporary. void Vsqrtqs(QRegister qd, QRegister qm, QRegister temp); void Vdivqs(QRegister qd, QRegister qn, QRegister qm); void SmiTag(Register reg, Condition cond = AL) { Lsl(reg, reg, Operand(kSmiTagSize), cond); } void SmiTag(Register dst, Register src, Condition cond = AL) { Lsl(dst, src, Operand(kSmiTagSize), cond); } void SmiUntag(Register reg, Condition cond = AL) { Asr(reg, reg, Operand(kSmiTagSize), cond); } void SmiUntag(Register dst, Register src, Condition cond = AL) { Asr(dst, src, Operand(kSmiTagSize), cond); } // Untag the value in the register assuming it is a smi. // Untagging shifts tag bit into the carry flag - if carry is clear // assumption was correct. In this case jump to the is_smi label. // Otherwise fall-through. void SmiUntag(Register dst, Register src, Label* is_smi) { ASSERT(kSmiTagSize == 1); Asrs(dst, src, Operand(kSmiTagSize)); b(is_smi, CC); } void CheckCodePointer(); // Function frame setup and tear down. void EnterFrame(RegList regs, intptr_t frame_space); void LeaveFrame(RegList regs); void Ret(); void ReserveAlignedFrameSpace(intptr_t frame_space); // Create a frame for calling into runtime that preserves all volatile // registers. Frame's SP is guaranteed to be correctly aligned and // frame_space bytes are reserved under it. void EnterCallRuntimeFrame(intptr_t frame_space); void LeaveCallRuntimeFrame(); void CallRuntime(const RuntimeEntry& entry, intptr_t argument_count); // Set up a Dart frame on entry with a frame pointer and PC information to // enable easy access to the RawInstruction object of code corresponding // to this frame. void EnterDartFrame(intptr_t frame_size); void LeaveDartFrame(RestorePP restore_pp = kRestoreCallerPP); // Set up a Dart frame for a function compiled for on-stack replacement. // The frame layout is a normal Dart frame, but the frame is partially set // up on entry (it is the frame of the unoptimized code). void EnterOsrFrame(intptr_t extra_size); // Set up a stub frame so that the stack traversal code can easily identify // a stub frame. void EnterStubFrame(); void LeaveStubFrame(); // The register into which the allocation stats table is loaded with // LoadAllocationStatsAddress should be passed to // IncrementAllocationStats(WithSize) as stats_addr_reg to update the // allocation stats. These are separate assembler macros so we can // avoid a dependent load too nearby the load of the table address. void LoadAllocationStatsAddress(Register dest, intptr_t cid, bool inline_isolate = true); void IncrementAllocationStats(Register stats_addr, intptr_t cid, Heap::Space space); void IncrementAllocationStatsWithSize(Register stats_addr_reg, Register size_reg, Heap::Space space); Address ElementAddressForIntIndex(bool is_load, bool is_external, intptr_t cid, intptr_t index_scale, Register array, intptr_t index, Register temp); Address ElementAddressForRegIndex(bool is_load, bool is_external, intptr_t cid, intptr_t index_scale, Register array, Register index); // If allocation tracing for |cid| is enabled, will jump to |trace| label, // which will allocate in the runtime where tracing occurs. void MaybeTraceAllocation(intptr_t cid, Register temp_reg, Label* trace, bool inline_isolate = true); // Inlined allocation of an instance of class 'cls', code has no runtime // calls. Jump to 'failure' if the instance cannot be allocated here. // Allocated instance is returned in 'instance_reg'. // Only the tags field of the object is initialized. void TryAllocate(const Class& cls, Label* failure, Register instance_reg, Register temp_reg); void TryAllocateArray(intptr_t cid, intptr_t instance_size, Label* failure, Register instance, Register end_address, Register temp1, Register temp2); // Emit data (e.g encoded instruction or immediate) in instruction stream. void Emit(int32_t value); // On some other platforms, we draw a distinction between safe and unsafe // smis. static bool IsSafe(const Object& object) { return true; } static bool IsSafeSmi(const Object& object) { return object.IsSmi(); } bool constant_pool_allowed() const { return constant_pool_allowed_; } void set_constant_pool_allowed(bool b) { constant_pool_allowed_ = b; } private: AssemblerBuffer buffer_; // Contains position independent code. ObjectPoolWrapper object_pool_wrapper_; int32_t prologue_offset_; bool use_far_branches_; #if 0 // If you are thinking of using one or both of these instructions directly, // instead LoadImmediate should probably be used. // Moved to ARM::AssemblerARM32::movw void movw(Register rd, uint16_t imm16, Condition cond = AL); // Moved to ARM::AssemblerARM32::movt void movt(Register rd, uint16_t imm16, Condition cond = AL); #endif void BindARMv6(Label* label); void BindARMv7(Label* label); void LoadWordFromPoolOffset(Register rd, int32_t offset, Register pp, Condition cond); void BranchLink(const ExternalLabel* label); class CodeComment : public ZoneAllocated { public: CodeComment(intptr_t pc_offset, const String& comment) : pc_offset_(pc_offset), comment_(comment) { } intptr_t pc_offset() const { return pc_offset_; } const String& comment() const { return comment_; } private: intptr_t pc_offset_; const String& comment_; DISALLOW_COPY_AND_ASSIGN(CodeComment); }; GrowableArray comments_; bool constant_pool_allowed_; void LoadObjectHelper(Register rd, const Object& object, Condition cond, bool is_unique, Register pp); #if 0 // Moved to ARM32::AssemblerARM32::emitType01() void EmitType01(Condition cond, int type, Opcode opcode, int set_cc, Register rn, Register rd, Operand o); // Moved to ARM32::AssemblerARM32::emitType05() void EmitType5(Condition cond, int32_t offset, bool link); // Moved to ARM32::AssemberARM32::emitMemOp() void EmitMemOp(Condition cond, bool load, bool byte, Register rd, Address ad); // Moved to AssemblerARM32::emitMemOpEnc3(); void EmitMemOpAddressMode3(Condition cond, int32_t mode, Register rd, Address ad); // Moved to ARM32::AssemblerARM32::emitMultiMemOp() void EmitMultiMemOp(Condition cond, BlockAddressMode am, bool load, Register base, RegList regs); #endif void EmitShiftImmediate(Condition cond, Shift opcode, Register rd, Register rm, Operand o); void EmitShiftRegister(Condition cond, Shift opcode, Register rd, Register rm, Operand o); #if 0 // Moved to ARM32::AssemblerARM32::emitMulOp() void EmitMulOp(Condition cond, int32_t opcode, Register rd, Register rn, Register rm, Register rs); // Moved to ARM32::AssemblerARM32::emitDivOp(); void EmitDivOp(Condition cond, int32_t opcode, Register rd, Register rn, Register rm); #endif void EmitMultiVSMemOp(Condition cond, BlockAddressMode am, bool load, Register base, SRegister start, uint32_t count); void EmitMultiVDMemOp(Condition cond, BlockAddressMode am, bool load, Register base, DRegister start, int32_t count); #if 0 // Moved to ARM32::AssemblerARM32::emitVFPsss void EmitVFPsss(Condition cond, int32_t opcode, SRegister sd, SRegister sn, SRegister sm); // Moved to ARM32::AssemblerARM32::emitVFPddd void EmitVFPddd(Condition cond, int32_t opcode, DRegister dd, DRegister dn, DRegister dm); // Moved to ARM32::AssemblerARM32::emitVFPsd void EmitVFPsd(Condition cond, int32_t opcode, SRegister sd, DRegister dm); // Moved to ARM32::AssemblerARM32::emitVFPds void EmitVFPds(Condition cond, int32_t opcode, DRegister dd, SRegister sm); // Moved to ARM32::AssemblerARM32::emitSIMDqqq() void EmitSIMDqqq(int32_t opcode, OperandSize sz, QRegister qd, QRegister qn, QRegister qm); #endif void EmitSIMDddd(int32_t opcode, OperandSize sz, DRegister dd, DRegister dn, DRegister dm); void EmitFarBranch(Condition cond, int32_t offset, bool link); #if 0 // Moved to ARM32::AssemblerARM32::emitBranch() void EmitBranch(Condition cond, Label* label, bool link); // Moved to ARM32::AssemblerARM32::encodeBranchoffset(). int32_t EncodeBranchOffset(int32_t offset, int32_t inst); // Moved to ARM32::AssemberARM32::decodeBranchOffset(). static int32_t DecodeBranchOffset(int32_t inst); #endif int32_t EncodeTstOffset(int32_t offset, int32_t inst); int32_t DecodeTstOffset(int32_t inst); void StoreIntoObjectFilter(Register object, Register value, Label* no_update); // Shorter filtering sequence that assumes that value is not a smi. void StoreIntoObjectFilterNoSmi(Register object, Register value, Label* no_update); // Helpers for write-barrier verification. // Returns VerifiedMemory::offset() as an Operand. Operand GetVerifiedMemoryShadow(); // Writes value to [base + offset] and also its shadow location, if enabled. void WriteShadowedField(Register base, intptr_t offset, Register value, Condition cond = AL); void WriteShadowedFieldPair(Register base, intptr_t offset, Register value_even, Register value_odd, Condition cond = AL); // Writes new_value to address and its shadow location, if enabled, after // verifying that its old value matches its shadow. void VerifiedWrite(const Address& address, Register new_value, FieldContent old_content); #if 0 // Added the following missing operations: // // ARM32::AssemblerARM32::uxt() (uxtb and uxth) // ARM32::AssemblerARM32::vpop() // ARM32::AssemblerARM32::vpush() // ARM32::AssemblerARM32::rbit() // ARM32::AssemblerARM32::vbslq() // ARM32::AssemblerARM32::veord() // ARM32::AssemblerARM32::vld1qr() // ARM32::AssemblerARM32::vshlqc // ARM32::AssemblerARM32::vshrqic // ARM32::AssemblerARM32::vshrquc // ARM32::AssemblerARM32::vst1qr() // ARM32::AssemblerARM32::vmorqi() // ARM32::AssemblerARM32::vmovqc() #endif DISALLOW_ALLOCATION(); DISALLOW_COPY_AND_ASSIGN(Assembler); }; } // namespace dart #endif // VM_ASSEMBLER_ARM_H_