/art/runtime/arch/x86/ |
D | memcmp16_x86.S | 139 movdqa (%esi), %xmm0 143 movdqa 16(%esi), %xmm2 149 movdqa %xmm0, %xmm1 150 movdqa 32(%esi), %xmm0 151 movdqa 48(%esi), %xmm2 169 movdqa %xmm0, %xmm1 189 movdqa 16(%esi), %xmm1 190 movdqa %xmm1, %xmm2 194 movdqa 32(%esi), %xmm3 215 movdqa 16(%esi), %xmm0 [all …]
|
/art/runtime/arch/x86_64/ |
D | memcmp16_x86_64.S | 319 movdqa %xmm2, %xmm1 353 movdqa %xmm2, %xmm1 389 movdqa (%rdi), %xmm2 394 movdqa 16(%rdi), %xmm2 399 movdqa 32(%rdi), %xmm2 404 movdqa 48(%rdi), %xmm2 411 movdqa 64(%rdi), %xmm2 416 movdqa 80(%rdi), %xmm2 439 movdqa (%rdi), %xmm2 444 movdqa 16(%rdi), %xmm2 [all …]
|
/art/compiler/utils/x86/ |
D | assembler_x86_test.cc | 617 DriverStr(RepeatFF(&x86::X86Assembler::movdqa, "movdqa %{reg2}, %{reg1}"), "movdqa"); in TEST_F() 625 DriverStr(RepeatFF(&x86::X86Assembler::movdqa, "vmovdqa %{reg2}, %{reg1}"), "avx_movdqa"); in TEST_F() 629 DriverStr(RepeatFA(&x86::X86Assembler::movdqa, "movdqa {mem}, %{reg}"), "movdqa_load"); in TEST_F() 637 DriverStr(RepeatFA(&x86::X86Assembler::movdqa, "vmovdqa {mem}, %{reg}"), "avx_movdqa_load"); in TEST_F() 641 DriverStr(RepeatAF(&x86::X86Assembler::movdqa, "movdqa %{reg}, {mem}"), "movdqa_store"); in TEST_F() 649 DriverStr(RepeatAF(&x86::X86Assembler::movdqa, "vmovdqa %{reg}, {mem}"), "avx_movdqa_store"); in TEST_F()
|
D | assembler_x86.h | 463 void movdqa(XmmRegister dst, XmmRegister src); // move 464 void movdqa(XmmRegister dst, const Address& src); // load aligned 466 void movdqa(const Address& dst, XmmRegister src); // store aligned
|
D | assembler_x86.cc | 1184 void X86Assembler::movdqa(XmmRegister dst, XmmRegister src) { in movdqa() function in art::x86::X86Assembler 1215 void X86Assembler::movdqa(XmmRegister dst, const Address& src) { in movdqa() function in art::x86::X86Assembler 1277 void X86Assembler::movdqa(const Address& dst, XmmRegister src) { in movdqa() function in art::x86::X86Assembler
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64_test.cc | 1247 DriverStr(RepeatFF(&x86_64::X86_64Assembler::movdqa, "movdqa %{reg2}, %{reg1}"), "movdqa"); in TEST_F() 1255 DriverStr(RepeatFF(&x86_64::X86_64Assembler::movdqa, "vmovdqa %{reg2}, %{reg1}"), "avx_movdqa"); in TEST_F() 1259 DriverStr(RepeatAF(&x86_64::X86_64Assembler::movdqa, "movdqa %{reg}, {mem}"), "movdqa_s"); in TEST_F() 1267 DriverStr(RepeatAF(&x86_64::X86_64Assembler::movdqa, "vmovdqa %{reg}, {mem}"), "avx_movdqa_s"); in TEST_F() 1271 DriverStr(RepeatFA(&x86_64::X86_64Assembler::movdqa, "movdqa {mem}, %{reg}"), "movdqa_l"); in TEST_F() 1279 DriverStr(RepeatFA(&x86_64::X86_64Assembler::movdqa, "vmovdqa {mem}, %{reg}"), "avx_movdqa_l"); in TEST_F()
|
D | assembler_x86_64.h | 495 void movdqa(XmmRegister dst, XmmRegister src); // move 496 void movdqa(XmmRegister dst, const Address& src); // load aligned 498 void movdqa(const Address& dst, XmmRegister src); // store aligned
|
D | assembler_x86_64.cc | 1550 void X86_64Assembler::movdqa(XmmRegister dst, XmmRegister src) { in movdqa() function in art::x86_64::X86_64Assembler 1611 void X86_64Assembler::movdqa(XmmRegister dst, const Address& src) { in movdqa() function in art::x86_64::X86_64Assembler 1720 void X86_64Assembler::movdqa(const Address& dst, XmmRegister src) { in movdqa() function in art::x86_64::X86_64Assembler
|
/art/compiler/optimizing/ |
D | code_generator_vector_x86_64.cc | 1293 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad() 1305 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad() 1341 is_aligned16 ? __ movdqa(address, reg) : __ movdqu(address, reg); in VisitVecStore()
|
D | code_generator_vector_x86.cc | 1320 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad() 1332 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad() 1368 is_aligned16 ? __ movdqa(address, reg) : __ movdqu(address, reg); in VisitVecStore()
|