Searched refs:rax (Results 1 – 25 of 223) sorted by relevance
123456789
9 vfmadd132pd (%rax), %xmm10, %xmm1117 vfmadd132ps (%rax), %xmm10, %xmm1125 vfmadd213pd (%rax), %xmm10, %xmm1133 vfmadd213ps (%rax), %xmm10, %xmm1141 vfmadd231pd (%rax), %xmm10, %xmm1149 vfmadd231ps (%rax), %xmm10, %xmm1157 vfmadd132pd (%rax), %ymm10, %ymm1165 vfmadd132ps (%rax), %ymm10, %ymm1173 vfmadd213pd (%rax), %ymm10, %ymm1181 vfmadd213ps (%rax), %ymm10, %ymm11[all …]
10 vphsubwd (%rcx,%rax), %xmm118 vphsubdq (%rcx,%rax), %xmm126 vphsubbw (%rax), %xmm142 vphaddwd (%rdx,%rax), %xmm750 vphadduwq (%rcx,%rax), %xmm658 vphadduwd (%rax), %xmm566 vphaddudq 8(%rcx,%rax), %xmm490 vphaddubd (%rax), %xmm5106 vphaddbw (%rcx,%rax), %xmm1114 vphaddbq (%rcx,%rax), %xmm1[all …]
313 vcmpps $0, (%rax), %xmm12, %xmm15325 vcmppd $0, (%rax), %xmm12, %xmm151377 vucomiss (%rax), %xmm121385 vcomiss (%rax), %xmm121393 vucomisd (%rax), %xmm121401 vcomisd (%rax), %xmm121409 vcvtsi2ss (%rax), %xmm11, %xmm121413 vcvtsi2ss (%rax), %xmm11, %xmm121421 vcvtsi2sd (%rax), %xmm11, %xmm121425 vcvtsi2sd (%rax), %xmm11, %xmm12[all …]
13 blsmskl (%rax), %r10d17 blsmskq (%rax), %r1029 blsil (%rax), %r10d33 blsiq (%rax), %r1045 blsrl (%rax), %r10d49 blsrq (%rax), %r1053 andnl (%rax), %r11d, %r10d57 andnq (%rax), %r11, %r1061 bextrl %r12d, (%rax), %r10d69 bextrq %r12, (%rax), %r10[all …]
8 monitor %rax, %rcx, %rdx14 mwait %rax, %rcx25 or %rax, %rdx27 shl $3, %rax70 xorq $1, %rax74 xorq $256, %rax90 call *%rax239 cmovnzq %rbx, %rax360 shld %bx, (%rax)361 shld %cl, %bx, (%rax)[all …]
61 addq $0x0000000000000000, %rax65 addq $0x000000000000007F, %rax69 addq $0xFFFFFFFFFFFFFF80, %rax73 addq $0xFFFFFFFFFFFFFFFF, %rax77 addq $0x0000000000000000, %rax81 addq $0xFF80, %rax85 addq $0xFFFF, %rax89 movq $0xFFFFFF80, %rax93 movq $0xFFFFFFFF, %rax97 addq $0x000000007FFFFFFF, %rax[all …]
11 movq %gs:(%rdi), %rax67 crc32b %dil,%rax71 crc32b %r11b,%rax75 crc32b 4(%rbx), %rax79 crc32q %rbx, %rax83 crc32q 4(%rbx), %rax129 movq 48879(,%riz), %rax133 movq -4(,%riz,8), %rax137 movq (%rcx,%riz), %rax141 movq (%rcx,%riz,8), %rax[all …]
99 mov ($ap),%rax106 mov %rax,$lo0107 mov ($np),%rax113 add %rax,$lo0 # discarded114 mov 8($ap),%rax123 add %rax,$hi1124 mov ($ap,$j,8),%rax134 add %rax,$hi0135 mov ($np,$j,8),%rax144 add %rax,$hi1[all …]
89 mov (+8*0)($SRC2), %rax92 add %rax, $X[0]100 mov (+8*$i)($SRC2), %rax103 add %rax, $X[$i]122 mov (+8*0)($SRC2), %rax124 add %rax, $X[0]132 mov (+8*$i)($SRC2), %rax134 add %rax, $X[$i]435 xor %rax, %rax441 adc \$0, %rax[all …]
83 mov %rsp,%rax89 mov %rax,8(%rsp,$num,8) # tp[num+1]=%rsp101 lea .Lmagic_masks(%rip),%rax104 movq 0(%rax,%r10,8),%xmm4 # set of masks denoting which105 movq 8(%rax,%r10,8),%xmm5 # cache line contains element106 movq 16(%rax,%r10,8),%xmm6 # denoted by 7th argument107 movq 24(%rax,%r10,8),%xmm7125 mov ($ap),%rax138 mov %rax,$lo0139 mov ($np),%rax[all …]
10 0388B8C90000 addl 51640(%rax), %ecx11 038C05B8C90000 addl 51640(%rbp,%rax), %ecx13 03B0BCC90000 addl 51644(%rax), %esi15 03B8C0C90000 addl 51648(%rax), %edi19 4403AC85A0020000 addl 672(%rbp,%rax,4), %r13d26 037C8500 addl (%rbp,%rax,4), %edi31 48034320 addq 32(%rbx), %rax37 4801C0 addq %rax, %rax38 4801C7 addq %rax, %rdi39 4801C2 addq %rax, %rdx[all …]
38 addl 51640(%rax), %ecx42 addl 51640(%rbp,%rax), %ecx50 addl 51644(%rax), %esi58 addl 51648(%rax), %edi74 addl 672(%rbp,%rax,4), %r13d102 addl (%rbp,%rax,4), %edi122 addq 32(%rbx), %rax146 addq %rax, %rax150 addq %rax, %rdi154 addq %rax, %rdx[all …]
149 4803442410 addq 16(%rsp), %rax154 4803442418 addq 24(%rsp), %rax157 4803442420 addq 32(%rsp), %rax159 4803442428 addq 40(%rsp), %rax167 4883C008 addq $8, %rax172 4805C61E0000 addq $7878, %rax174 4C01F8 addq %r15, %rax177 4801442470 addq %rax, 112(%rsp)178 4801442420 addq %rax, 32(%rsp)179 4801442430 addq %rax, 48(%rsp)[all …]
11 ULong rax; variable24 rdx = 0x11111111; rax = 0x22222222; in main()28 rax&0xff,rbx&0xff,rcx&0xff); in main()37 "\tmov " VG_SYM(rax) ",%rax\n" in main()42 "\tmov " VG_SYM(rax) "(%rip),%rax\n" in main()68 rdx = 0x99999999; rax = 0x77777777; in main()72 rax&0xff,rbx&0xff,rcx&0xff); in main()81 "\tmov " VG_SYM(rax) ",%rax\n" in main()86 "\tmov " VG_SYM(rax) "(%rip),%rax\n" in main()112 rdx = 0x11111111; rax = 0x22222222; in main()[all …]
10 unsigned long rdi, rsi, rcx, rax; in main() local29 rax = 0x751234560000ULL + (' ' << 8) + '0'; in main()31 : "=D" (rdi), "=c" (rcx), "+a" (rax) in main()37 || rax != 0x751234560000ULL + (' ' << 8) + '0') in main()41 : "=S" (rsi), "=a" (rax) in main()44 || rax != 0x20302030ULL) in main()58 rax = 0x123450000ULL + ('d' << 8) + 'c'; in main()60 : "=D" (rdi), "=c" (rcx), "+a" (rax) in main()64 || rax != 0x123450000ULL + ('d' << 8) + 'c') in main()67 rax = 0x543210000ULL + ('b' << 8) + 'a'; in main()[all …]
26 movsxd rax, dword ptr arg(3) ;stride39 movd [rdi+rax], mm245 movd [rdi+2*rax], mm347 add rdi, rax52 movd [rdi+2*rax], mm480 movsxd rax, dword ptr arg(1) ;src_stride;83 movq mm1, [rsi+rax]84 movq mm2, [rsi+rax*2]87 lea rsi, [rsi+rax*2]90 add rsi, rax[all …]
53 lea rax, [GLOBAL(k0_k5)]54 add rax, rdx57 cmp esi, DWORD PTR [rax]60 movdqa xmm4, XMMWORD PTR [rax] ;k0_k561 movdqa xmm5, XMMWORD PTR [rax+256] ;k2_k462 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k365 movsxd rax, dword ptr arg(1) ;src_pixels_per_line91 lea rsi, [rsi + rax]115 movdqa xmm5, XMMWORD PTR [rax+256] ;k2_k4116 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3[all …]
114 movq $__NR_rt_sigprocmask, %rax // syscall #123 testq %rax, %rax130 movq %rsi, %rax /* rax --> VexGuestAMD64State * */132 movq OFFSET_amd64_RDI(%rax), %rdi133 movq OFFSET_amd64_RSI(%rax), %rsi134 movq OFFSET_amd64_RDX(%rax), %rdx135 movq OFFSET_amd64_R10(%rax), %r10136 movq OFFSET_amd64_R8(%rax), %r8137 movq OFFSET_amd64_R9(%rax), %r9138 popq %rax /* syscallno -> %rax */[all …]
56 using v8::internal::rax;112 __ movq(rax, arg2); in TEST()141 __ pop(rax); in TEST()142 __ pop(rax); in TEST()143 __ pop(rax); in TEST()166 __ movq(rax, arg2); in TEST()167 __ addq(rax, arg1); in TEST()189 __ movq(rax, arg2); in TEST()191 __ movq(rax, rdx); in TEST()224 __ movq(rax, Operand(rbp, -3 * kStackElementSize)); in TEST()[all …]
72 using v8::internal::rax;110 __ cmovq(not_equal, rax, rdx); in ExitCode()143 __ movl(rax, Immediate(id)); in TestMoveSmi()182 __ xor_(rax, rax); // Success. in TEST()202 __ movl(rax, Immediate(id + 1)); in TestSmiCompare()205 __ movl(rax, Immediate(id + 2)); in TestSmiCompare()209 __ movl(rax, Immediate(id + 3)); in TestSmiCompare()212 __ movl(rax, Immediate(id + 4)); in TestSmiCompare()215 __ incq(rax); in TestSmiCompare()222 __ movl(rax, Immediate(id + 9)); in TestSmiCompare()[all …]
... 103 173: 48 03 04 00 add (%rax,%rax,1),%rax 104 ...
199 __ lea(rax, Operand(rsi, rdi, times_1, 0)); in CheckAtStart()200 __ cmpq(rax, Operand(rbp, kInputStart)); in CheckAtStart()211 __ lea(rax, Operand(rsi, rdi, times_1, 0)); in CheckNotAtStart()212 __ cmpq(rax, Operand(rbp, kInputStart)); in CheckNotAtStart()258 __ movzxwl(rax, in CheckCharacters()260 __ cmpl(rax, Immediate(static_cast<int32_t>(str[0]))); in CheckCharacters()277 __ movq(rax, combined_chars, RelocInfo::NONE); in CheckCharacters()278 __ cmpq(rax, Operand(rbx, byte_offset + i)); in CheckCharacters()297 __ movq(rax, combined_chars, RelocInfo::NONE); in CheckCharacters()298 __ cmpq(rax, in CheckCharacters()[all …]
15 # CHECK: addb %al, (%rax)129 # CHECK: vcvtsd2si %xmm0, %rax132 # CHECK: vmaskmovpd %xmm0, %xmm1, (%rax)167 # CHECK: vblendps $129, (%rax), %ymm5, %ymm1173 # CHECK: vblendpd $129, (%rax), %ymm5, %ymm1185 # CHECK: vdpps $129, (%rax), %ymm5, %ymm1335 # CHECK: crc32q %rax, %rax338 # CHECK: invept (%rax), %rax341 # CHECK: invvpid (%rax), %rax344 # CHECK: invpcid (%rax), %rax[all …]
31 mov rax, arg(1) ;zbin_ptr32 movq mm1, [rax]63 mov rax, arg(3) ;dequant_ptr64 movq mm2, [rax]67 mov rax, arg(7) ;dqcoeff_ptr69 movq [rax], mm374 mov rax, arg(1) ;zbin_ptr75 movq mm5, [rax+8]103 mov rax, arg(3) ;dequant_ptr104 movq mm6, [rax+8][all …]
30 ; X64_WIN-NEXT: movq (%rcx,%rax,8), %rax31 ; X64_WIN-NEXT: movl i1@SECREL(%rax), %eax45 ; X64_LINUX: movq %fs:0, %rax46 ; X64_LINUX-NEXT: leaq i1@TPOFF(%rax), %rax57 ; X64_WIN-NEXT: movq (%rcx,%rax,8), %rax58 ; X64_WIN-NEXT: leaq i1@SECREL(%rax), %rax71 ; X64_LINUX: movq i2@GOTTPOFF(%rip), %rax72 ; X64_LINUX-NEXT: movl %fs:(%rax), %eax83 ; X64_WIN-NEXT: movq (%rcx,%rax,8), %rax84 ; X64_WIN-NEXT: movl i2@SECREL(%rax), %eax[all …]