/external/vixl/test/aarch64/ |
D | test-assembler-aarch64.cc | 7445 __ Fmov(tgt2, fp_tgt); in LoadFPValueHelper() local 9853 __ Fmov(s11, 1.0); in TEST() local 9854 __ Fmov(d22, -13.0); in TEST() local 9855 __ Fmov(s1, 255.0); in TEST() local 9856 __ Fmov(d2, 12.34567); in TEST() local 9857 __ Fmov(s3, 0.0); in TEST() local 9858 __ Fmov(d4, 0.0); in TEST() local 9859 __ Fmov(s5, kFP32PositiveInfinity); in TEST() local 9860 __ Fmov(d6, kFP64NegativeInfinity); in TEST() local 9882 __ Fmov(s20, 1.0); in TEST() local [all …]
|
D | test-utils-aarch64.cc | 365 __ Fmov(dn, value); in ClobberFP() local 370 __ Fmov(dn, first); in ClobberFP() local
|
/external/v8/src/crankshaft/arm64/ |
D | delayed-masm-arm64-inl.h | 32 void DelayedMasm::Fmov(FPRegister fd, FPRegister fn) { in Fmov() function 38 void DelayedMasm::Fmov(FPRegister fd, double imm) { in Fmov() function
|
D | lithium-gap-resolver-arm64.cc | 163 __ Fmov(SavedFPValueRegister(), cgen_->ToDoubleRegister(source)); in BreakCycle() local 188 __ Fmov(cgen_->ToDoubleRegister(saved_destination_), in RestoreValue() local 239 __ Fmov(result, cgen_->ToDouble(constant_source)); in EmitMove() local 270 __ Fmov(cgen_->ToDoubleRegister(destination), src); in EmitMove() local
|
D | lithium-codegen-arm64.cc | 2293 __ Fmov(temp, object); in DoCmpHoleAndBranchD() local 2409 __ Fmov(result, fp_zero); in DoConstantD() local 2414 __ Fmov(result, instr->value()); in DoConstantD() local 3203 __ Fmov(scratch, result); in DoLoadKeyedFixedDouble() local 3660 __ Fmov(double_scratch(), kFP64NegativeInfinity); in DoMathPowHalf() local 3731 __ Fmov(scratch_d, 0.5); in DoMathRoundD() local 3758 __ Fmov(dot_five, 0.5); in DoMathRoundI() local 3780 __ Fmov(result, input); in DoMathRoundI() local 5244 __ Fmov(scratch1, dbl_scratch1); in DoDeferredTaggedToI() local
|
/external/v8/src/compiler/arm64/ |
D | code-generator-arm64.cc | 1484 __ Fmov(i.OutputRegister32(), i.InputFloat32Register(0)); in AssembleArchInstruction() local 1488 __ Fmov(i.OutputRegister(), i.InputFloat64Register(0)); in AssembleArchInstruction() local 1495 __ Fmov(tmp, i.InputFloat64Register(0)); in AssembleArchInstruction() local 1497 __ Fmov(i.OutputFloat64Register(), tmp); in AssembleArchInstruction() local 1504 __ Fmov(tmp.W(), i.InputFloat32Register(0)); in AssembleArchInstruction() local 1506 __ Fmov(i.OutputFloat64Register(), tmp); in AssembleArchInstruction() local 1510 __ Fmov(i.OutputFloat64Register(), i.InputRegister(0)); in AssembleArchInstruction() local 1516 __ Fmov(i.OutputRegister(), i.InputDoubleRegister(0)); in AssembleArchInstruction() local 2043 __ Fmov(dst, src.ToFloat32()); in AssembleMove() local 2051 __ Fmov(temp, src.ToFloat32()); in AssembleMove() local [all …]
|
/external/v8/src/arm64/ |
D | macro-assembler-arm64-inl.h | 720 void MacroAssembler::Fmov(FPRegister fd, FPRegister fn) { in Fmov() function 732 void MacroAssembler::Fmov(FPRegister fd, Register rn) { in Fmov() function 738 void MacroAssembler::Fmov(FPRegister fd, double imm) { in Fmov() function 756 void MacroAssembler::Fmov(FPRegister fd, float imm) { in Fmov() function 778 void MacroAssembler::Fmov(Register rd, FPRegister fn) { in Fmov() function
|
D | code-stubs-arm64.cc | 90 __ Fmov(result, double_scratch); in Generate() local 751 __ Fmov(scratch1_double, base_double); in Generate() local 752 __ Fmov(base_double_copy, base_double); in Generate() local 753 __ Fmov(result_double, 1.0); in Generate() local 772 __ Fmov(scratch0_double, 1.0); in Generate() local 785 __ Fmov(base_double, base_double_copy); in Generate() local 1133 __ Fmov(fp_zero, 0.0); in Generate() local
|
D | macro-assembler-arm64.h | 446 void Fmov(FPRegister fd, T imm) { in Fmov() function
|
/external/vixl/src/aarch64/ |
D | macro-assembler-aarch64.h | 1343 void Fmov(const VRegister& vd, const VRegister& vn) { in Fmov() function 1354 void Fmov(const VRegister& vd, const Register& rn) { in Fmov() function 1360 void Fmov(const VRegister& vd, const XRegister& xn) { in Fmov() function 1363 void Fmov(const VRegister& vd, const WRegister& wn) { in Fmov() function 1366 void Fmov(const VRegister& vd, int index, const Register& rn) { in Fmov() function 1371 void Fmov(const Register& rd, const VRegister& vn, int index) { in Fmov() function 1385 void Fmov(VRegister vd, T imm) { in Fmov() function 1389 void Fmov(Register rd, VRegister vn) { in Fmov() function
|
D | macro-assembler-aarch64.cc | 1482 void MacroAssembler::Fmov(VRegister vd, double imm) { in Fmov() function in vixl::aarch64::MacroAssembler 1514 void MacroAssembler::Fmov(VRegister vd, float imm) { in Fmov() function in vixl::aarch64::MacroAssembler
|
/external/v8/src/builtins/arm64/ |
D | builtins-arm64.cc | 193 __ Fmov(x11, d5); in Generate_MathMaxMin() local 200 __ Fmov(x10, d5); in Generate_MathMaxMin() local
|