/external/vixl/test/aarch64/ |
D | test-assembler-aarch64.cc | 5647 __ Movi(v0.V2D(), 0x0001020304050607, 0x08090a0b0c0d0e0f); in TEST() local 5648 __ Movi(v1.V2D(), 0x1011121314151617, 0x18191a1b1c1d1e1f); in TEST() local 5740 __ Movi(v0.V2D(), 0x0001020304050607, 0x08090a0b0c0d0e0f); in TEST() local 5741 __ Movi(v1.V2D(), 0x1011121314151617, 0x18191a1b1c1d1e1f); in TEST() local 5742 __ Movi(v2.V2D(), 0x2021222324252627, 0x28292a2b2c2d2e2f); in TEST() local 5839 __ Movi(v0.V2D(), 0x0001020304050607, 0x08090a0b0c0d0e0f); in TEST() local 5840 __ Movi(v1.V2D(), 0x1011121314151617, 0x18191a1b1c1d1e1f); in TEST() local 5841 __ Movi(v2.V2D(), 0x2021222324252627, 0x28292a2b2c2d2e2f); in TEST() local 5842 __ Movi(v3.V2D(), 0x2021222324252627, 0x28292a2b2c2d2e2f); in TEST() local 6898 __ Movi(v0.V2D(), 0, 0x2222222233333333); in TEST() local [all …]
|
D | test-disasm-aarch64.cc | 6185 COMPARE_MACRO(Movi(v4.V8B(), 0xaa), "movi v4.8b, #0xaa"); in TEST() 6186 COMPARE_MACRO(Movi(v1.V16B(), 0xcc), "movi v1.16b, #0xcc"); in TEST() 6188 COMPARE_MACRO(Movi(v4.V4H(), 0xaa, LSL, 0), "movi v4.4h, #0xaa, lsl #0"); in TEST() 6189 COMPARE_MACRO(Movi(v1.V8H(), 0xcc, LSL, 8), "movi v1.8h, #0xcc, lsl #8"); in TEST() 6191 COMPARE_MACRO(Movi(v4.V2S(), 0xaa, LSL, 0), "movi v4.2s, #0xaa, lsl #0"); in TEST() 6192 COMPARE_MACRO(Movi(v1.V2S(), 0xcc, LSL, 8), "movi v1.2s, #0xcc, lsl #8"); in TEST() 6193 COMPARE_MACRO(Movi(v4.V4S(), 0xaa, LSL, 16), "movi v4.4s, #0xaa, lsl #16"); in TEST() 6194 COMPARE_MACRO(Movi(v1.V4S(), 0xcc, LSL, 24), "movi v1.4s, #0xcc, lsl #24"); in TEST() 6196 COMPARE_MACRO(Movi(v4.V2S(), 0xaa, MSL, 8), "movi v4.2s, #0xaa, msl #8"); in TEST() 6197 COMPARE_MACRO(Movi(v1.V2S(), 0xcc, MSL, 16), "movi v1.2s, #0xcc, msl #16"); in TEST() [all …]
|
D | test-simulator-aarch64.cc | 1515 __ Movi(vd.V16B(), 0); in Test1OpNEON_Helper() local 2472 __ Movi(vd.V16B(), 0); in Test2OpImmNEON_Helper() local
|
/external/vixl/examples/aarch64/ |
D | neon-matrix-multiply.cc | 78 __ Movi(v0.V16B(), 0); in GenerateNEONMatrixMultiply() local 79 __ Movi(v1.V16B(), 0); in GenerateNEONMatrixMultiply() local 80 __ Movi(v2.V16B(), 0); in GenerateNEONMatrixMultiply() local 81 __ Movi(v3.V16B(), 0); in GenerateNEONMatrixMultiply() local
|
/external/vixl/src/aarch64/ |
D | macro-assembler-aarch64.cc | 1085 void MacroAssembler::Movi(const VRegister& vd, in Movi() function in vixl::aarch64::MacroAssembler 1110 void MacroAssembler::Movi(const VRegister& vd, uint64_t hi, uint64_t lo) { in Movi() function in vixl::aarch64::MacroAssembler 1114 Movi(vd.V2D(), lo); in Movi() 1513 Movi(vd, rawbits); in Fmov() 1550 Movi(vd, rawbits); in Fmov() 1587 Movi(vd, static_cast<uint64_t>(rawbits)); in Fmov()
|
D | macro-assembler-aarch64.h | 3084 void Movi(const VRegister& vd, 3088 void Movi(const VRegister& vd, uint64_t hi, uint64_t lo);
|
/external/v8/src/arm64/ |
D | macro-assembler-arm64-inl.h | 651 Movi(vd, bits); in Fmov() 678 Movi(vd, bits); in Fmov()
|
D | macro-assembler-arm64.cc | 498 void TurboAssembler::Movi(const VRegister& vd, uint64_t imm, Shift shift, in Movi() function in v8::internal::TurboAssembler 519 void TurboAssembler::Movi(const VRegister& vd, uint64_t hi, uint64_t lo) { in Movi() function in v8::internal::TurboAssembler 523 Movi(vd.V2D(), lo); in Movi()
|
D | macro-assembler-arm64.h | 846 void Movi(const VRegister& vd, uint64_t imm, Shift shift = LSL, 848 void Movi(const VRegister& vd, uint64_t hi, uint64_t lo);
|
/external/v8/src/compiler/arm64/ |
D | code-generator-arm64.cc | 2009 __ Movi(i.OutputSimd128Register().V16B(), 0); in AssembleArchInstruction() local 2115 __ Movi(temp, imm2, imm1); in AssembleArchInstruction() local
|