Searched refs:rax (Results 1 – 25 of 257) sorted by relevance
1234567891011
6 movq 0(%rsi),%rax8 addq %rax,%r813 movq 8(%rsi),%rax15 addq %rax,%r921 movq 16(%rsi),%rax23 addq %rax,%r1029 movq 24(%rsi),%rax31 addq %rax,%r1137 movq 32(%rsi),%rax39 addq %rax,%r12[all …]
100 mov ($ap),%rax107 mov %rax,$lo0108 mov ($np),%rax114 add %rax,$lo0 # discarded115 mov 8($ap),%rax124 add %rax,$hi1125 mov ($ap,$j,8),%rax135 add %rax,$hi0136 mov ($np,$j,8),%rax145 add %rax,$hi1[all …]
90 mov (+8*0)($SRC2), %rax93 add %rax, $X[0]101 mov (+8*$i)($SRC2), %rax104 add %rax, $X[$i]123 mov (+8*0)($SRC2), %rax125 add %rax, $X[0]133 mov (+8*$i)($SRC2), %rax135 add %rax, $X[$i]436 xor %rax, %rax442 adc \$0, %rax[all …]
84 mov %rsp,%rax90 mov %rax,8(%rsp,$num,8) # tp[num+1]=%rsp102 lea .Lmagic_masks(%rip),%rax105 movq 0(%rax,%r10,8),%xmm4 # set of masks denoting which106 movq 8(%rax,%r10,8),%xmm5 # cache line contains element107 movq 16(%rax,%r10,8),%xmm6 # denoted by 7th argument108 movq 24(%rax,%r10,8),%xmm7126 mov ($ap),%rax139 mov %rax,$lo0140 mov ($np),%rax[all …]
36 movq (%rsi),%rax43 movq %rax,%r1044 movq (%rcx),%rax50 addq %rax,%r1051 movq 8(%rsi),%rax60 addq %rax,%r1361 movq (%rsi,%r15,8),%rax71 addq %rax,%r1172 movq (%rcx,%r15,8),%rax81 addq %rax,%r13[all …]
23 movq %rsp,%rax29 movq %rax,8(%rsp,%r9,8)36 leaq .Lmagic_masks(%rip),%rax39 movq 0(%rax,%r10,8),%xmm440 movq 8(%rax,%r10,8),%xmm541 movq 16(%rax,%r10,8),%xmm642 movq 24(%rax,%r10,8),%xmm760 movq (%rsi),%rax73 movq %rax,%r1074 movq (%rcx),%rax[all …]
9 vfmadd132pd (%rax), %xmm10, %xmm1117 vfmadd132ps (%rax), %xmm10, %xmm1125 vfmadd213pd (%rax), %xmm10, %xmm1133 vfmadd213ps (%rax), %xmm10, %xmm1141 vfmadd231pd (%rax), %xmm10, %xmm1149 vfmadd231ps (%rax), %xmm10, %xmm1157 vfmadd132pd (%rax), %ymm10, %ymm1165 vfmadd132ps (%rax), %ymm10, %ymm1173 vfmadd213pd (%rax), %ymm10, %ymm1181 vfmadd213ps (%rax), %ymm10, %ymm11[all …]
10 vphsubwd (%rcx,%rax), %xmm118 vphsubdq (%rcx,%rax), %xmm126 vphsubbw (%rax), %xmm142 vphaddwd (%rdx,%rax), %xmm750 vphadduwq (%rcx,%rax), %xmm658 vphadduwd (%rax), %xmm566 vphaddudq 8(%rcx,%rax), %xmm490 vphaddubd (%rax), %xmm5106 vphaddbw (%rcx,%rax), %xmm1114 vphaddbq (%rcx,%rax), %xmm1[all …]
313 vcmpps $0, (%rax), %xmm12, %xmm15325 vcmppd $0, (%rax), %xmm12, %xmm151377 vucomiss (%rax), %xmm121385 vcomiss (%rax), %xmm121393 vucomisd (%rax), %xmm121401 vcomisd (%rax), %xmm121409 vcvtsi2ssl (%rax), %xmm11, %xmm121413 vcvtsi2ssl (%rax), %xmm11, %xmm121421 vcvtsi2sdl (%rax), %xmm11, %xmm121425 vcvtsi2sdl (%rax), %xmm11, %xmm12[all …]
13 blsmskl (%rax), %r10d17 blsmskq (%rax), %r1029 blsil (%rax), %r10d33 blsiq (%rax), %r1045 blsrl (%rax), %r10d49 blsrq (%rax), %r1053 andnl (%rax), %r11d, %r10d57 andnq (%rax), %r11, %r1061 bextrl %r12d, (%rax), %r10d69 bextrq %r12, (%rax), %r10[all …]
8 monitor %rax, %rcx, %rdx14 mwait %rax, %rcx25 or %rax, %rdx27 shl $3, %rax70 xorq $1, %rax74 xorq $256, %rax90 call *%rax239 cmovnzq %rbx, %rax360 shld %bx, (%rax)361 shld %cl, %bx, (%rax)[all …]
61 addq $0x0000000000000000, %rax65 addq $0x000000000000007F, %rax69 addq $0xFFFFFFFFFFFFFF80, %rax73 addq $0xFFFFFFFFFFFFFFFF, %rax77 addq $0x0000000000000000, %rax81 addq $0xFF80, %rax85 addq $0xFFFF, %rax89 movq $0xFFFFFF80, %rax93 movq $0xFFFFFFFF, %rax97 addq $0x000000007FFFFFFF, %rax[all …]
5 palignr $8, (%rax), %xmm110 palignr $16, (%rax), %xmm115 palignr $0, (%rax), %xmm120 vpalignr $8, (%rax), %xmm1, %xmm225 vpalignr $16, (%rax), %xmm1, %xmm230 vpalignr $0, (%rax), %xmm1, %xmm235 vpalignr $8, (%rax), %ymm1, %ymm240 vpalignr $16, (%rax), %ymm1, %ymm245 vpalignr $0, (%rax), %ymm1, %ymm250 pshufd $27, (%rax), %xmm1[all …]
11 movq %gs:(%rdi), %rax67 crc32b %dil,%rax71 crc32b %r11b,%rax75 crc32b 4(%rbx), %rax79 crc32q %rbx, %rax83 crc32q 4(%rbx), %rax129 movq 48879(,%riz), %rax133 movq -4(,%riz,8), %rax137 movq (%rcx,%riz), %rax141 movq (%rcx,%riz,8), %rax[all …]
10 0388B8C90000 addl 51640(%rax), %ecx11 038C05B8C90000 addl 51640(%rbp,%rax), %ecx13 03B0BCC90000 addl 51644(%rax), %esi15 03B8C0C90000 addl 51648(%rax), %edi19 4403AC85A0020000 addl 672(%rbp,%rax,4), %r13d26 037C8500 addl (%rbp,%rax,4), %edi31 48034320 addq 32(%rbx), %rax37 4801C0 addq %rax, %rax38 4801C7 addq %rax, %rdi39 4801C2 addq %rax, %rdx[all …]
38 addl 51640(%rax), %ecx42 addl 51640(%rbp,%rax), %ecx50 addl 51644(%rax), %esi58 addl 51648(%rax), %edi74 addl 672(%rbp,%rax,4), %r13d102 addl (%rbp,%rax,4), %edi122 addq 32(%rbx), %rax146 addq %rax, %rax150 addq %rax, %rdi154 addq %rax, %rdx[all …]
11 ULong rax; variable24 rdx = 0x11111111; rax = 0x22222222; in main()28 rax&0xff,rbx&0xff,rcx&0xff); in main()37 "\tmov " VG_SYM(rax) ",%rax\n" in main()42 "\tmov " VG_SYM(rax) "(%rip),%rax\n" in main()68 rdx = 0x99999999; rax = 0x77777777; in main()72 rax&0xff,rbx&0xff,rcx&0xff); in main()81 "\tmov " VG_SYM(rax) ",%rax\n" in main()86 "\tmov " VG_SYM(rax) "(%rip),%rax\n" in main()112 rdx = 0x11111111; rax = 0x22222222; in main()[all …]
10 unsigned long rdi, rsi, rcx, rax; in main() local29 rax = 0x751234560000ULL + (' ' << 8) + '0'; in main()31 : "=D" (rdi), "=c" (rcx), "+a" (rax) in main()37 || rax != 0x751234560000ULL + (' ' << 8) + '0') in main()41 : "=S" (rsi), "=a" (rax) in main()44 || rax != 0x20302030ULL) in main()58 rax = 0x123450000ULL + ('d' << 8) + 'c'; in main()60 : "=D" (rdi), "=c" (rcx), "+a" (rax) in main()64 || rax != 0x123450000ULL + ('d' << 8) + 'c') in main()67 rax = 0x543210000ULL + ('b' << 8) + 'a'; in main()[all …]
33 movsxd rax, dword ptr arg(1) ;src_stride;36 movq mm1, [rsi+rax]37 movq mm2, [rsi+rax*2]40 lea rsi, [rsi+rax*2]43 add rsi, rax53 movq mm4, [rsi+rax]55 movq mm5, [rsi+rax*2]58 lea rsi, [rsi+rax*2]64 movq mm0, [rsi+rax]65 movq mm1, [rsi+rax*2][all …]
32 movsxd rax, dword ptr arg(1) ;src_stride;35 movdqu xmm1, [rsi+rax]36 movdqu xmm2, [rsi+rax*2]39 lea rsi, [rsi+rax*2]42 add rsi, rax51 movdqu xmm4, [rsi+rax]53 movdqu xmm5, [rsi+rax*2]54 lea rsi, [rsi+rax*2]57 add rsi, rax66 movdqu xmm1, [rsi+rax][all …]
54 lea rax, [GLOBAL(k0_k5)]55 add rax, rdx58 cmp esi, DWORD PTR [rax]61 movdqa xmm4, XMMWORD PTR [rax] ;k0_k562 movdqa xmm5, XMMWORD PTR [rax+256] ;k2_k463 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k366 movsxd rax, dword ptr arg(1) ;src_pixels_per_line92 lea rsi, [rsi + rax]117 movdqa xmm5, XMMWORD PTR [rax+256] ;k2_k4118 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3[all …]
114 movq $__NR_rt_sigprocmask, %rax // syscall #123 testq %rax, %rax130 movq %rsi, %rax /* rax --> VexGuestAMD64State * */132 movq OFFSET_amd64_RDI(%rax), %rdi133 movq OFFSET_amd64_RSI(%rax), %rsi134 movq OFFSET_amd64_RDX(%rax), %rdx135 movq OFFSET_amd64_R10(%rax), %r10136 movq OFFSET_amd64_R8(%rax), %r8137 movq OFFSET_amd64_R9(%rax), %r9138 popq %rax /* syscallno -> %rax */[all …]
56 using v8::internal::rax;112 __ movq(rax, arg2); in TEST()141 __ pop(rax); in TEST()142 __ pop(rax); in TEST()143 __ pop(rax); in TEST()166 __ movq(rax, arg2); in TEST()167 __ addq(rax, arg1); in TEST()189 __ movq(rax, arg2); in TEST()191 __ movq(rax, rdx); in TEST()224 __ movq(rax, Operand(rbp, -3 * kStackElementSize)); in TEST()[all …]
... 103 173: 48 03 04 00 add (%rax,%rax,1),%rax 104 ...
199 __ lea(rax, Operand(rsi, rdi, times_1, 0)); in CheckAtStart()200 __ cmpq(rax, Operand(rbp, kInputStart)); in CheckAtStart()211 __ lea(rax, Operand(rsi, rdi, times_1, 0)); in CheckNotAtStart()212 __ cmpq(rax, Operand(rbp, kInputStart)); in CheckNotAtStart()258 __ movzxwl(rax, in CheckCharacters()260 __ cmpl(rax, Immediate(static_cast<int32_t>(str[0]))); in CheckCharacters()277 __ movq(rax, combined_chars, RelocInfo::NONE); in CheckCharacters()278 __ cmpq(rax, Operand(rbx, byte_offset + i)); in CheckCharacters()297 __ movq(rax, combined_chars, RelocInfo::NONE); in CheckCharacters()298 __ cmpq(rax, in CheckCharacters()[all …]