/external/vixl/test/aarch64/ |
D | test-assembler-aarch64.cc | 8032 __ Fmov(tgt2, fp_tgt); in LoadFPValueHelper() local 10443 __ Fmov(s1, 255.0); in TEST() local 10444 __ Fmov(d2, 12.34567); in TEST() local 10445 __ Fmov(s3, 0.0); in TEST() local 10446 __ Fmov(d4, 0.0); in TEST() local 10447 __ Fmov(s5, kFP32PositiveInfinity); in TEST() local 10448 __ Fmov(d6, kFP64NegativeInfinity); in TEST() local 10449 __ Fmov(h7, RawbitsToFloat16(0x6400U)); in TEST() local 10450 __ Fmov(h8, kFP16PositiveInfinity); in TEST() local 10451 __ Fmov(s11, 1.0); in TEST() local [all …]
|
D | test-utils-aarch64.cc | 409 __ Fmov(dn, value); in ClobberFP() local 414 __ Fmov(dn, first); in ClobberFP() local
|
D | test-disasm-aarch64.cc | 6205 COMPARE_MACRO(Fmov(v0.V2S(), 1.0f), "fmov v0.2s, #0x70 (1.0000)"); in TEST() 6206 COMPARE_MACRO(Fmov(v31.V2S(), -13.0f), "fmov v31.2s, #0xaa (-13.0000)"); in TEST() 6207 COMPARE_MACRO(Fmov(v0.V4S(), 1.0f), "fmov v0.4s, #0x70 (1.0000)"); in TEST() 6208 COMPARE_MACRO(Fmov(v31.V4S(), -13.0f), "fmov v31.4s, #0xaa (-13.0000)"); in TEST() 6209 COMPARE_MACRO(Fmov(v1.V2D(), 1.0), "fmov v1.2d, #0x70 (1.0000)"); in TEST() 6210 COMPARE_MACRO(Fmov(v29.V2D(), -13.0), "fmov v29.2d, #0xaa (-13.0000)"); in TEST() 6211 COMPARE_MACRO(Fmov(v0.V4H(), Float16(-5.0f)), "fmov v0.4h, #0x94 (-5.0000)"); in TEST() 6212 COMPARE_MACRO(Fmov(v31.V8H(), Float16(29.0f)), in TEST() 6214 COMPARE_MACRO(Fmov(v0.V4H(), Float16(-5.0)), "fmov v0.4h, #0x94 (-5.0000)"); in TEST() 6215 COMPARE_MACRO(Fmov(v31.V8H(), Float16(29.0)), "fmov v31.8h, #0x3d (29.0000)"); in TEST()
|
/external/v8/src/wasm/baseline/arm64/ |
D | liftoff-assembler-arm64.h | 169 Fmov(reg.fp().S(), value.to_f32_boxed().get_scalar()); in LoadConstant() 172 Fmov(reg.fp().D(), value.to_f64_boxed().get_scalar()); in LoadConstant() 311 Fmov(dst.S(), src.S()); in Move() 314 Fmov(dst.D(), src.D()); in Move() 469 Fmov(scratch.S(), src.W()); in emit_i32_popcnt() 472 Fmov(dst.W(), scratch.S()); in emit_i32_popcnt() 668 Fmov(dst.gp().W(), src.fp().S()); in emit_type_conversion() 709 Fmov(dst.gp().X(), src.fp().D()); in emit_type_conversion() 727 Fmov(dst.fp().S(), src.gp().W()); in emit_type_conversion() 745 Fmov(dst.fp().D(), src.gp().X()); in emit_type_conversion()
|
/external/vixl/src/aarch64/ |
D | macro-assembler-aarch64.cc | 1469 Fmov(tmp, value); in Fcmp() 1482 void MacroAssembler::Fmov(VRegister vd, double imm) { in Fmov() function in vixl::aarch64::MacroAssembler 1488 Fmov(vd, Float16(imm)); in Fmov() 1493 Fmov(vd, static_cast<float>(imm)); in Fmov() 1519 void MacroAssembler::Fmov(VRegister vd, float imm) { in Fmov() function in vixl::aarch64::MacroAssembler 1525 Fmov(vd, Float16(imm)); in Fmov() 1530 Fmov(vd, static_cast<double>(imm)); in Fmov() 1556 void MacroAssembler::Fmov(VRegister vd, Float16 imm) { in Fmov() function in vixl::aarch64::MacroAssembler 1561 Fmov(vd, FPToFloat(imm, kIgnoreDefaultNaN)); in Fmov() 1566 Fmov(vd, FPToDouble(imm, kIgnoreDefaultNaN)); in Fmov() [all …]
|
D | macro-assembler-aarch64.h | 1491 void Fmov(const VRegister& vd, const VRegister& vn) { in Fmov() function 1502 void Fmov(const VRegister& vd, const Register& rn) { in Fmov() function 1508 void Fmov(const VRegister& vd, const XRegister& xn) { in Fmov() function 1509 Fmov(vd, Register(xn)); in Fmov() 1511 void Fmov(const VRegister& vd, const WRegister& wn) { in Fmov() function 1512 Fmov(vd, Register(wn)); in Fmov() 1514 void Fmov(const VRegister& vd, int index, const Register& rn) { in Fmov() function 1519 void Fmov(const Register& rd, const VRegister& vn, int index) { in Fmov() function 1529 void Fmov(VRegister vd, double imm); 1530 void Fmov(VRegister vd, float imm); [all …]
|
/external/v8/src/arm64/ |
D | macro-assembler-arm64-inl.h | 510 Fmov(tmp, value); in Fcmp() 612 void TurboAssembler::Fmov(VRegister fd, VRegister fn) { in Fmov() function 623 void TurboAssembler::Fmov(VRegister fd, Register rn) { in Fmov() function 628 void TurboAssembler::Fmov(VRegister vd, double imm) { in Fmov() function 632 Fmov(vd, static_cast<float>(imm)); in Fmov() 656 void TurboAssembler::Fmov(VRegister vd, float imm) { in Fmov() function 659 Fmov(vd, static_cast<double>(imm)); in Fmov() 675 Fmov(vd, tmp); in Fmov() 683 void TurboAssembler::Fmov(Register rd, VRegister fn) { in Fmov() function
|
D | macro-assembler-arm64.h | 830 inline void Fmov(VRegister fd, VRegister fn); 831 inline void Fmov(VRegister fd, Register rn); 836 inline void Fmov(VRegister fd, double imm); 837 inline void Fmov(VRegister fd, float imm); 840 void Fmov(VRegister fd, T imm) { in Fmov() function 842 Fmov(fd, static_cast<double>(imm)); in Fmov() 844 inline void Fmov(Register rd, VRegister fn);
|
D | code-stubs-arm64.cc | 59 __ Fmov(fp_zero, 0.0); in Generate() local
|
D | macro-assembler-arm64.cc | 3063 Fmov(new_arg, old_arg); in PrintfNoPreserve() 3078 Fmov(pcs[i].VReg(), args[i].VReg()); in PrintfNoPreserve()
|
/external/v8/src/compiler/arm64/ |
D | code-generator-arm64.cc | 1486 __ Fmov(i.OutputRegister32(), i.InputFloat32Register(0)); in AssembleArchInstruction() local 1500 __ Fmov(i.OutputFloat64Register(), i.InputRegister(0)); in AssembleArchInstruction() local 1506 __ Fmov(i.OutputRegister(), i.InputDoubleRegister(0)); in AssembleArchInstruction() local 2656 __ Fmov(dst.S(), src.ToFloat32()); in AssembleMove() local 2659 __ Fmov(dst, src.ToFloat64().value()); in AssembleMove() local 2678 __ Fmov(temp, src.ToFloat32()); in AssembleMove() local 2688 __ Fmov(temp, src.ToFloat64().value()); in AssembleMove() local
|
/external/v8/src/builtins/arm64/ |
D | builtins-arm64.cc | 2957 __ Fmov(result, double_scratch); in Generate_DoubleToI() local 3053 __ Fmov(scratch1_double, base_double); in Generate_MathPowInternal() local 3054 __ Fmov(base_double_copy, base_double); in Generate_MathPowInternal() local 3055 __ Fmov(result_double, 1.0); in Generate_MathPowInternal() local 3074 __ Fmov(scratch0_double, 1.0); in Generate_MathPowInternal() local 3087 __ Fmov(base_double, base_double_copy); in Generate_MathPowInternal() local
|