/external/vixl/test/aarch64/ |
D | test-assembler-fp-aarch64.cc | 300 __ Fmov(tgt2, fp_tgt); in LoadFPValueHelper() local 332 __ Fmov(s1, 255.0); in TEST() local 333 __ Fmov(d2, 12.34567); in TEST() local 334 __ Fmov(s3, 0.0); in TEST() local 335 __ Fmov(d4, 0.0); in TEST() local 336 __ Fmov(s5, kFP32PositiveInfinity); in TEST() local 337 __ Fmov(d6, kFP64NegativeInfinity); in TEST() local 338 __ Fmov(h7, RawbitsToFloat16(0x6400U)); in TEST() local 339 __ Fmov(h8, kFP16PositiveInfinity); in TEST() local 340 __ Fmov(s11, 1.0); in TEST() local [all …]
|
D | test-assembler-neon-aarch64.cc | 3681 __ Fmov(v0.V4H(), 24.0); in TEST() local 3682 __ Fmov(v1.V4H(), 1024.0); in TEST() local 3683 __ Fmov(v2.V8H(), 5.5); in TEST() local 3684 __ Fmov(v3.V8H(), 2048.0); in TEST() local 3685 __ Fmov(v4.V8H(), kFP16PositiveInfinity); in TEST() local 3686 __ Fmov(v5.V8H(), kFP16NegativeInfinity); in TEST() local 3687 __ Fmov(v6.V4H(), RawbitsToFloat16(0x7c2f)); in TEST() local 3688 __ Fmov(v7.V8H(), RawbitsToFloat16(0xfe0f)); in TEST() local 3723 __ Fmov(v0.V4H(), 24.0); in TEST() local 3724 __ Fmov(v1.V4H(), 1024.0); in TEST() local [all …]
|
D | test-utils-aarch64.cc | 444 __ Fmov(dn, value); in ClobberFP() local 449 __ Fmov(dn, first); in ClobberFP() local
|
D | test-assembler-aarch64.cc | 7728 __ Fmov(d1, x1); in TEST() local 7729 __ Fmov(d2, x2); in TEST() local 7730 __ Fmov(d3, x3); in TEST() local 7731 __ Fmov(d4, x4); in TEST() local 7796 __ Fmov(d1, x1); in TEST() local 7797 __ Fmov(d2, x2); in TEST() local 7798 __ Fmov(d3, x3); in TEST() local 7799 __ Fmov(d4, x4); in TEST() local 7800 __ Fmov(d5, x1); in TEST() local 7801 __ Fmov(d6, x2); in TEST() local [all …]
|
D | test-disasm-neon-aarch64.cc | 3102 COMPARE_MACRO(Fmov(v0.V2S(), 1.0f), "fmov v0.2s, #0x70 (1.0000)"); in TEST() 3103 COMPARE_MACRO(Fmov(v31.V2S(), -13.0f), "fmov v31.2s, #0xaa (-13.0000)"); in TEST() 3104 COMPARE_MACRO(Fmov(v0.V4S(), 1.0f), "fmov v0.4s, #0x70 (1.0000)"); in TEST() 3105 COMPARE_MACRO(Fmov(v31.V4S(), -13.0f), "fmov v31.4s, #0xaa (-13.0000)"); in TEST() 3106 COMPARE_MACRO(Fmov(v1.V2D(), 1.0), "fmov v1.2d, #0x70 (1.0000)"); in TEST() 3107 COMPARE_MACRO(Fmov(v29.V2D(), -13.0), "fmov v29.2d, #0xaa (-13.0000)"); in TEST() 3109 COMPARE_MACRO(Fmov(v0.V4H(), Float16(-5.0f)), "fmov v0.4h, #0x94 (-5.0000)"); in TEST() 3110 COMPARE_MACRO(Fmov(v31.V8H(), Float16(29.0f)), in TEST() 3112 COMPARE_MACRO(Fmov(v0.V4H(), Float16(-5.0)), "fmov v0.4h, #0x94 (-5.0000)"); in TEST() 3113 COMPARE_MACRO(Fmov(v31.V8H(), Float16(29.0)), "fmov v31.8h, #0x3d (29.0000)"); in TEST() [all …]
|
/external/vixl/src/aarch64/ |
D | macro-assembler-aarch64.cc | 1479 Fmov(tmp, value); in Fcmp() 1492 void MacroAssembler::Fmov(VRegister vd, double imm) { in Fmov() function in vixl::aarch64::MacroAssembler 1498 Fmov(vd, Float16(imm)); in Fmov() 1503 Fmov(vd, static_cast<float>(imm)); in Fmov() 1529 void MacroAssembler::Fmov(VRegister vd, float imm) { in Fmov() function in vixl::aarch64::MacroAssembler 1535 Fmov(vd, Float16(imm)); in Fmov() 1540 Fmov(vd, static_cast<double>(imm)); in Fmov() 1566 void MacroAssembler::Fmov(VRegister vd, Float16 imm) { in Fmov() function in vixl::aarch64::MacroAssembler 1571 Fmov(vd, FPToFloat(imm, kIgnoreDefaultNaN)); in Fmov() 1576 Fmov(vd, FPToDouble(imm, kIgnoreDefaultNaN)); in Fmov() [all …]
|
D | macro-assembler-aarch64.h | 1489 void Fmov(const VRegister& vd, const VRegister& vn) { in Fmov() function 1500 void Fmov(const VRegister& vd, const Register& rn) { in Fmov() function 1506 void Fmov(const VRegister& vd, const XRegister& xn) { in Fmov() function 1507 Fmov(vd, Register(xn)); in Fmov() 1509 void Fmov(const VRegister& vd, const WRegister& wn) { in Fmov() function 1510 Fmov(vd, Register(wn)); in Fmov() 1512 void Fmov(const VRegister& vd, int index, const Register& rn) { in Fmov() function 1521 void Fmov(const Register& rd, const VRegister& vn, int index) { in Fmov() function 1535 void Fmov(VRegister vd, double imm); 1536 void Fmov(VRegister vd, float imm); [all …]
|
/external/v8/src/codegen/arm64/ |
D | macro-assembler-arm64-inl.h | 544 Fmov(tmp, value); in Fcmp() 646 void TurboAssembler::Fmov(VRegister fd, VRegister fn) { in Fmov() function 657 void TurboAssembler::Fmov(VRegister fd, Register rn) { in Fmov() function 662 void TurboAssembler::Fmov(VRegister vd, double imm) { in Fmov() function 666 Fmov(vd, static_cast<float>(imm)); in Fmov() 690 void TurboAssembler::Fmov(VRegister vd, float imm) { in Fmov() function 693 Fmov(vd, static_cast<double>(imm)); in Fmov() 709 Fmov(vd, tmp); in Fmov() 717 void TurboAssembler::Fmov(Register rd, VRegister fn) { in Fmov() function
|
D | macro-assembler-arm64.h | 934 inline void Fmov(VRegister fd, VRegister fn); 935 inline void Fmov(VRegister fd, Register rn); 940 inline void Fmov(VRegister fd, double imm); 941 inline void Fmov(VRegister fd, float imm); 944 void Fmov(VRegister fd, T imm) { in Fmov() function 946 Fmov(fd, static_cast<double>(imm)); in Fmov() 948 inline void Fmov(Register rd, VRegister fn);
|
D | macro-assembler-arm64.cc | 3175 Fmov(new_arg, old_arg); in PrintfNoPreserve() 3192 Fmov(pcs[i].Reg(), temp0); in PrintfNoPreserve() 3194 Fmov(pcs[i].Reg(), args[i].VReg()); in PrintfNoPreserve() 3206 Fmov(pcs[i].VReg(), args[i].VReg()); in PrintfNoPreserve()
|
/external/v8/src/wasm/baseline/arm64/ |
D | liftoff-assembler-arm64.h | 306 Fmov(reg.fp().S(), value.to_f32_boxed().get_scalar()); in LoadConstant() 309 Fmov(reg.fp().D(), value.to_f64_boxed().get_scalar()); in LoadConstant() 778 Fmov(dst.S(), src.S()); in Move() 780 Fmov(dst.D(), src.D()); in Move() 1025 Fmov(scratch.S(), src.W()); in emit_i32_popcnt() 1028 Fmov(dst.W(), scratch.S()); in emit_i32_popcnt() 1045 Fmov(scratch.D(), src.gp().X()); in emit_i64_popcnt() 1048 Fmov(dst.gp().X(), scratch.D()); in emit_i64_popcnt() 1196 Fmov(dst.S(), lhs.S()); in emit_f32_copysign() 1207 Fmov(dst.D(), lhs.D()); in emit_f64_copysign() [all …]
|
/external/vixl/benchmarks/aarch64/ |
D | bench-utils.cc | 372 __ Fmov(PickV(size), 1.25 * GetRandomBits(2)); in GenerateFPSequence() local
|
/external/v8/src/compiler/backend/arm64/ |
D | code-generator-arm64.cc | 1696 __ Fmov(i.OutputRegister32(), i.InputFloat32Register(0)); in AssembleArchInstruction() local 1710 __ Fmov(i.OutputFloat64Register(), i.InputRegister(0)); in AssembleArchInstruction() local 1716 __ Fmov(i.OutputRegister(), i.InputDoubleRegister(0)); in AssembleArchInstruction() local 3319 __ Fmov(dst.S(), src.ToFloat32()); in AssembleMove() local 3322 __ Fmov(dst, src.ToFloat64().value()); in AssembleMove() local 3341 __ Fmov(temp, src.ToFloat32()); in AssembleMove() local 3351 __ Fmov(temp, src.ToFloat64().value()); in AssembleMove() local
|
/external/v8/src/builtins/arm64/ |
D | builtins-arm64.cc | 623 __ Fmov(fp_zero, 0.0); in Generate_JSEntryVariant() local 3217 __ Fmov(result, double_scratch); in Generate_DoubleToI() local
|