/art/compiler/optimizing/ |
D | code_generator_x86.cc | 87 __ movl(reg_, Immediate(0)); in EmitNativeCode() local 181 __ movl(calling_convention.GetRegisterAt(0), Immediate(instruction_->GetStringIndex())); in EmitNativeCode() local 213 __ movl(calling_convention.GetRegisterAt(0), Immediate(cls_->GetTypeIndex())); in EmitNativeCode() local 351 __ movl(Address(ESP, stack_index), static_cast<Register>(reg_id)); in SaveCoreRegister() local 356 __ movl(static_cast<Register>(reg_id), Address(ESP, stack_index)); in RestoreCoreRegister() local 501 __ movl(Address(ESP, kCurrentMethodStackOffset), EAX); in GenerateFrameEntry() local 531 __ movl(reg, Address(ESP, kCurrentMethodStackOffset)); in LoadCurrentMethod() local 621 __ movl(destination.AsRegister<Register>(), source.AsRegister<Register>()); in Move32() local 626 __ movl(destination.AsRegister<Register>(), Address(ESP, source.GetStackIndex())); in Move32() local 640 __ movl(Address(ESP, destination.GetStackIndex()), source.AsRegister<Register>()); in Move32() local [all …]
|
D | code_generator_x86_64.cc | 94 __ movl(cpu_reg_, Immediate(0)); in EmitNativeCode() local 203 __ movl(CpuRegister(calling_convention.GetRegisterAt(0)), Immediate(cls_->GetTypeIndex())); in EmitNativeCode() local 250 __ movl(CpuRegister(calling_convention.GetRegisterAt(0)), in EmitNativeCode() local 384 __ movl(temp, Address(temp, ArtMethod::DexCacheResolvedMethodsOffset().SizeValue())); in GenerateStaticOrDirectCall() local 626 __ movl(destination.AsRegister<CpuRegister>(), in Move() local 648 __ movl(Address(CpuRegister(RSP), destination.GetStackIndex()), in Move() local 656 __ movl(Address(CpuRegister(RSP), destination.GetStackIndex()), Immediate(value)); in Move() local 659 __ movl(CpuRegister(TMP), Address(CpuRegister(RSP), source.GetStackIndex())); in Move() local 660 __ movl(Address(CpuRegister(RSP), destination.GetStackIndex()), CpuRegister(TMP)); in Move() local 702 __ movl(location.AsRegister<CpuRegister>(), imm); in Move() local [all …]
|
D | intrinsics_x86.cc | 79 __ movl(target_reg, EAX); in MoveFromReturnRegister() local 87 __ movl(target_reg_lo, EAX); in MoveFromReturnRegister() local 90 __ movl(target_reg_hi, EDX); in MoveFromReturnRegister() local 319 __ movl(output_lo, input_hi); in VisitLongReverseBytes() local 320 __ movl(output_hi, input_lo); in VisitLongReverseBytes() local 443 __ movl(temp, input_hi); in GenAbsLong() local 447 __ movl(output_lo, temp); in GenAbsLong() local 448 __ movl(output_hi, temp); in GenAbsLong() local 632 __ movl(temp, output_lo); in GenMinMax() local 636 __ movl(temp, output_hi); in GenMinMax() local [all …]
|
D | intrinsics_x86_64.cc | 75 __ movl(trg_reg, CpuRegister(RAX)); in MoveFromReturnRegister() local 364 __ movl(mask, out); in GenAbsInteger() local 944 __ movl(string_length, Address(string_obj, count_offset)); in GenerateStringIndexOf() local 954 __ movl(counter, string_length); in GenerateStringIndexOf() local 995 __ movl(out, Immediate(-1)); in GenerateStringIndexOf() local 1105 __ movl(out, Address(address, 0)); in GenPeek() local 1180 __ movl(Address(address, 0), in GenPoke() local 1183 __ movl(Address(address, 0), value.AsRegister<CpuRegister>()); in GenPoke() local 1255 __ movl(trg, Address(base, offset, ScaleFactor::TIMES_1, 0)); in GenUnsafeGet() local 1375 __ movl(Address(base, offset, ScaleFactor::TIMES_1, 0), value); in GenUnsafePut() local [all …]
|
/art/compiler/utils/x86/ |
D | assembler_x86.cc | 106 void X86Assembler::movl(Register dst, const Immediate& imm) { in movl() function in art::x86::X86Assembler 113 void X86Assembler::movl(Register dst, Register src) { in movl() function in art::x86::X86Assembler 120 void X86Assembler::movl(Register dst, const Address& src) { in movl() function in art::x86::X86Assembler 127 void X86Assembler::movl(const Address& dst, Register src) { in movl() function in art::x86::X86Assembler 134 void X86Assembler::movl(const Address& dst, const Immediate& imm) { in movl() function in art::x86::X86Assembler 141 void X86Assembler::movl(const Address& dst, Label* lbl) { in movl() function in art::x86::X86Assembler
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64.cc | 122 void X86_64Assembler::movl(CpuRegister dst, const Immediate& imm) { in movl() function in art::x86_64::X86_64Assembler 150 void X86_64Assembler::movl(CpuRegister dst, CpuRegister src) { in movl() function in art::x86_64::X86_64Assembler 166 void X86_64Assembler::movl(CpuRegister dst, const Address& src) { in movl() function in art::x86_64::X86_64Assembler 182 void X86_64Assembler::movl(const Address& dst, CpuRegister src) { in movl() function in art::x86_64::X86_64Assembler 189 void X86_64Assembler::movl(const Address& dst, const Immediate& imm) { in movl() function in art::x86_64::X86_64Assembler
|