/art/runtime/arch/x86_64/ |
D | jni_entrypoints_x86_64.S | 77 PUSH_ARG rax 89 POP_ARG rax 120 movq THREAD_TOP_QUICK_FRAME_OFFSET(%rdi), %rax // uintptr_t tagged_quick_frame 121 andq LITERAL(TAGGED_JNI_SP_MASK_TOGGLED64), %rax // ArtMethod** sp 122 movq (%rax), %rax // ArtMethod* method 124 ART_METHOD_ACCESS_FLAGS_OFFSET(%rax) 147 testq %rax, %rax // check if returned method code is null 149 jmp *%rax // otherwise, tail call to intended method 160 testq LITERAL(1), %rax 170 PUSH_ARG rax [all …]
|
D | quick_entrypoints_x86_64.S | 86 PUSH rax 185 POP rax 315 movq %rax, %rdi 316 movq %rdx, %rax 323 jmp *%rax 459 movq %rdi, %rax // rax := method to be called 466 movq %rax, %rdi // rdi := method to be called 488 movq %rax, (%r8) // Store the result assuming its a long, int or Object* 553 movq %rdi, %rax // rax := method to be called 560 movq %rax, %rdi // rdi := method to be called [all …]
|
D | memcmp16_x86_64.S | 779 mov -16(%rdi), %rax 781 cmp %rax, %rcx 784 mov -8(%rdi), %rax 786 cmp %rax, %rcx 793 mov -12(%rdi), %rax 795 cmp %rax, %rcx 829 mov -18(%rdi), %rax 831 cmp %rax, %rcx 834 mov -10(%rdi), %rax 836 cmp %rax, %rcx [all …]
|
D | asm_support_x86_64.S | 535 movq REG_VAR(saved_rax), %rax // Restore RAX. 554 movq REG_VAR(saved_rax), %rax // Restore RAX.
|
/art/runtime/interpreter/mterp/x86_64ng/ |
D | arithmetic.S | 6 movzbq 2(rPC), %rax # rax <- BB 9 GET_WIDE_VREG %rax, %rax # eax <- vBB 12 GET_VREG %eax, %rax # eax <- vBB 21 $ext # rdx:rax <- sign-extended of rax 70 GET_WIDE_VREG %rax, rINSTq # eax <- vA 82 $ext # rdx:rax <- sign-extended of rax 128 movl rINST, %eax # rax <- 000000BA 130 GET_VREG %eax, %rax # eax <- vB 137 cdq # rax <- sign-extended of eax 156 movzbq 2(rPC), %rax # eax <- BB [all …]
|
D | other.S | 74 movq 2(rPC), %rax # rax <- HHHHhhhhBBBBbbbb 75 SET_WIDE_VREG %rax, rINSTq 80 movswq 2(rPC), %rax # rax <- ssssssssssssBBBB 81 SET_WIDE_VREG %rax, rINSTq # store 86 movslq 2(rPC), %rax # eax <- ssssssssBBBBbbbb 87 SET_WIDE_VREG %rax, rINSTq # store 92 movzwq 2(rPC), %rax # eax <- 000000000000BBBB 93 salq $$48, %rax # eax <- 00000000BBBB0000 94 SET_WIDE_VREG %rax, rINSTq # v[AA+0] <- eax 129 SET_VREG_OBJECT %edx, %rax # fp[A] <- fp[B] [all …]
|
D | array.S | 9 movzbq 2(rPC), %rax # eax <- BB 11 GET_VREG %edi, %rax # eax <- vBB (array object) 18 movq $data_offset(%rdi,%rsi,8), %rax 19 SET_WIDE_VREG %rax, rINSTq 66 movzbq 2(rPC), %rax # rax <- BB 68 GET_VREG %edi, %rax # edi <- vBB (array object) 99 movzbq 2(rPC), %rax # rax <- BB 101 GET_VREG %edi, %rax # edi <- vBB (array object) 122 SET_VREG rINST, %rax
|
D | main.S | 426 movq REG_VAR(inst), %rax 427 andq MACRO_LITERAL(0xf), %rax 428 GET_VREG %eax, %rax 434 movq REG_VAR(inst), %rax 435 andq MACRO_LITERAL(0xf), %rax 441 andq MACRO_LITERAL(0xf), %rax 443 GET_VREG %eax, %rax 451 movq REG_VAR(inst), %rax 452 andq MACRO_LITERAL(0xf), %rax 458 andq MACRO_LITERAL(0xf), %rax [all …]
|
D | floating_point.S | 20 movzbq 2(rPC), %rax # eax<- BB 21 GET_VREG_XMM${suff} %xmm0, %rax 57 movzbq 3(rPC), %rax # eax <- CC 60 v${instr}${suff} VREG_ADDRESS(%rax), %xmm0, %xmm0 65 ${instr}${suff} VREG_ADDRESS(%rax), %xmm0 177 movzbq 2(rPC), %rax # eax <- CC 179 fldl VREG_ADDRESS(%rax) # %st0 <- fp[vCC] 210 movzbq 2(rPC), %rax # eax <- CC 212 flds VREG_ADDRESS(%rax) # vCC to fp stack
|
D | invoke.S | 33 movq %rax, %rdi 93 movzw ART_METHOD_IMT_INDEX_OFFSET(%rax), %ecx 106 andq $$-4, %rax 107 movzw ART_METHOD_METHOD_INDEX_OFFSET(%rax), %ecx 134 movq %rax, %rdi
|
D | control_flow.S | 157 GET_WIDE_VREG %rax, rINSTq # eax <- vAA 160 movq %rax, %xmm0
|
D | object.S | 21 movq %rax, %rsi 165 movq %rax, %rsi
|
/art/compiler/utils/x86_64/ |
D | jni_macro_assembler_x86_64.cc | 525 CpuRegister rax(RAX); // RAX can be freely clobbered. It does not hold any argument. in TryToTransitionFromRunnableToNative() local 530 __ xorl(rax, rax); in TryToTransitionFromRunnableToNative() 561 CpuRegister rax(RAX); in TryToTransitionFromNativeToRunnable() local 566 __ movq(saved_rax, rax); // Save RAX. in TryToTransitionFromNativeToRunnable() 568 __ movl(rax, Immediate(kNativeStateValue)); in TryToTransitionFromNativeToRunnable() 575 __ movq(rax, saved_rax); // Restore RAX; MOV does not change flags. in TryToTransitionFromNativeToRunnable()
|
/art/compiler/optimizing/ |
D | intrinsics_x86_64.cc | 2301 CpuRegister rax(RAX); in GenCompareAndSetOrExchangeInt() local 2309 __ movzxb(rax, rax); in GenCompareAndSetOrExchangeInt() 2312 __ movsxb(rax, rax); in GenCompareAndSetOrExchangeInt() 2315 __ movsxw(rax, rax); in GenCompareAndSetOrExchangeInt() 2318 __ movzxw(rax, rax); in GenCompareAndSetOrExchangeInt() 2324 GenZFlagToResult(assembler, rax); in GenCompareAndSetOrExchangeInt() 4542 CpuRegister rax = rax_loc.AsRegister<CpuRegister>(); in GenerateVarHandleGetAndOp() local 4543 DCHECK_EQ(rax.AsRegister(), RAX); in GenerateVarHandleGetAndOp()
|
D | code_generator_x86_64.cc | 4417 CpuRegister rax = eax; in GenerateDivRemWithAnyConstant() local 4423 __ movq(numerator, rax); in GenerateDivRemWithAnyConstant() 4426 codegen_->Load64BitValue(rax, magic); in GenerateDivRemWithAnyConstant() 4445 __ movq(rax, rdx); in GenerateDivRemWithAnyConstant() 4447 __ addq(rdx, rax); in GenerateDivRemWithAnyConstant() 4450 __ movq(rax, numerator); in GenerateDivRemWithAnyConstant() 4458 __ subq(rax, rdx); in GenerateDivRemWithAnyConstant() 4459 __ movq(rdx, rax); in GenerateDivRemWithAnyConstant() 4461 __ movq(rax, rdx); in GenerateDivRemWithAnyConstant()
|
/art/runtime/arch/ |
D | stub_test.cc | 358 POP(%%rax) in Invoke3WithReferrerAndHidden()
|