/arch/x86/platform/efi/ |
D | efi_stub_32.S | 40 movl $1f, %edx 41 subl $__PAGE_OFFSET, %edx 42 jmp *%edx 51 popl %edx 52 movl %edx, saved_return_addr 56 movl $2f, %edx 57 subl $__PAGE_OFFSET, %edx 58 pushl %edx 63 movl %cr0, %edx 64 andl $0x7fffffff, %edx [all …]
|
/arch/x86/um/ |
D | setjmp_32.S | 24 movl %eax,%edx 26 movl 4(%esp),%edx 30 movl %ebx,(%edx) 31 movl %esp,4(%edx) # Post-return %esp! 33 movl %ebp,8(%edx) 34 movl %esi,12(%edx) 35 movl %edi,16(%edx) 36 movl %ecx,20(%edx) # Return address 47 xchgl %eax,%edx 49 movl 4(%esp),%edx # jmp_ptr address [all …]
|
/arch/x86/lib/ |
D | atomic64_386_32.S | 46 movl 4(v), %edx 60 movl 4(v), %edx 69 adcl %edx, 4(v) 76 adcl 4(v), %edx 78 movl %edx, 4(v) 85 sbbl %edx, 4(v) 91 negl %edx 93 sbbl $0, %edx 95 adcl 4(v), %edx 97 movl %edx, 4(v) [all …]
|
D | copy_user_64.S | 34 cmpl $8,%edx 37 movl %edx,%ecx 38 andl $63,%edx 62 movl %edx,%ecx 63 andl $7,%edx 72 20: andl %edx,%edx 74 movl %edx,%ecx 87 addl %ecx,%edx 89 40: leal (%rdx,%rcx,8),%edx 91 50: movl %ecx,%edx [all …]
|
D | hweight.S | 17 movl %eax, %edx # w -> t 18 shrl %edx # t >>= 1 19 andl $0x55555555, %edx # t &= 0x55555555 20 subl %edx, %eax # w -= t 22 movl %eax, %edx # w -> t 24 andl $0x33333333, %edx # t &= 0x33333333 26 addl %edx, %eax # w = w_tmp + t 28 movl %eax, %edx # w -> t 29 shrl $4, %edx # t >>= 4 30 addl %edx, %eax # w_tmp += t [all …]
|
D | atomic64_cx8_32.S | 17 movl %ecx, %edx 55 movl %edx, %edi 61 movl %edx, %ecx 70 movl %ecx, %edx 89 movl %edx, %ecx 98 movl %ecx, %edx 113 movl %edx, %ecx 123 movl %ecx, %edx 136 movl %edx, %edi 144 movl %edx, %ecx [all …]
|
D | checksum_32.S | 83 movl %ecx, %edx 107 2: movl %edx, %ecx 108 andl $0x1c, %edx 110 shrl $2, %edx # This clears CF 113 dec %edx 151 movl %ecx, %edx 233 movl %edx, %ecx 315 SRC( movl 4(%esi), %edx ) 318 adcl %edx, %eax 319 DST( movl %edx, 4(%edi) ) [all …]
|
D | memcpy_64.S | 37 andl $7, %edx 39 movl %edx, %ecx 89 addl $0x20, %edx 121 addl $0x20, %edx 125 cmpl $16, %edx 142 cmpl $8, %edx 154 cmpl $4, %edx 167 subl $1, %edx 192 cmpl $8, %edx 206 subl %ecx, %edx [all …]
|
/arch/x86/math-emu/ |
D | wm_shrx.S | 48 movl 4(%esi),%edx /* msl */ 51 shrd %cl,%edx,%ebx 52 shr %cl,%edx 54 movl %edx,4(%esi) 66 movl 4(%esi),%edx /* msl */ 67 shrd %cl,%edx,%eax 68 shr %cl,%edx 69 movl %edx,(%esi) 82 xorl %edx,%edx 83 movl %edx,(%esi) [all …]
|
D | round_Xsig.S | 35 movl 8(%esi),%edx 41 orl %edx,%edx /* ms bits */ 45 movl %ebx,%edx 52 bsrl %edx,%ecx /* get the required shift in %ecx */ 56 shld %cl,%ebx,%edx 65 adcl $0,%edx 68 movl $0x80000000,%edx 72 movl %edx,8(%esi) 95 movl 8(%esi),%edx 101 orl %edx,%edx /* ms bits */ [all …]
|
D | reg_u_div.S | 93 movswl EXP(%esi),%edx 95 subl %eax,%edx 96 addl EXP_BIAS,%edx 99 cmpl EXP_WAY_UNDER,%edx 103 movl EXP_WAY_UNDER,%edx 122 movl SIGH(%esi),%edx /* Dividend */ 125 cmpl %ecx,%edx 129 subl %ecx,%edx /* Prevent the overflow */ 191 movl SIGH(%esi),%edx 198 cmpl SIGH(%ebx),%edx /* Test for imminent overflow */ [all …]
|
D | reg_u_sub.S | 68 movl PARAM6,%edx 71 xorl %edx,%edx /* register extension */ 83 shrd %cl,%ebx,%edx 95 shrd %cl,%eax,%edx 100 orl $1,%edx /* record the fact in the extension */ 108 movl %ebx,%edx 118 movl %eax,%edx 122 orl $1,%edx 130 movl %eax,%edx 131 rcrl %edx [all …]
|
D | div_Xsig.S | 130 movl FPU_accum_3,%edx 143 mov %edx,%eax 156 sbbl %edx,FPU_accum_3 162 sbbl %edx,FPU_accum_2 174 movl XsigH(%ebx),%edx 176 sbbl %edx,FPU_accum_2 188 movl FPU_accum_2,%edx /* get the reduced num */ 192 cmpl XsigH(%ebx),%edx 203 sbbl XsigH(%ebx),%edx 204 movl %edx,FPU_accum_2 [all …]
|
D | mul_Xsig.S | 43 movl %edx,-12(%ebp) 48 adcl %edx,-8(%ebp) 54 adcl %edx,-4(%ebp) 84 movl %edx,-12(%ebp) 88 addl %edx,-12(%ebp) 95 adcl %edx,-8(%ebp) 101 adcl %edx,-8(%ebp) 107 adcl %edx,-4(%ebp) 138 movl %edx,-12(%ebp) 142 addl %edx,-12(%ebp) [all …]
|
D | shr_Xsig.S | 38 movl 8(%esi),%edx /* msl */ 40 shrd %cl,%edx,%ebx 41 shr %cl,%edx 44 movl %edx,8(%esi) 56 movl 8(%esi),%edx /* msl */ 57 shrd %cl,%edx,%eax 58 shr %cl,%edx 60 movl %edx,4(%esi) 73 xorl %edx,%edx 75 movl %edx,4(%esi) [all …]
|
D | reg_round.S | 122 movl PARAM2,%edx 205 orl %edx,%ecx 222 orl %edx,%edx 241 orl %edx,%ecx 284 orl %edx,%ecx 298 orl %edx,%edx 316 orl %edx,%ecx 349 orl %edx,%edx 357 orl %edx,%edx 362 cmpl $0x80000000,%edx [all …]
|
D | wm_sqrt.S | 92 xorl %edx,%edx 101 rcrl $1,%edx 109 movl %edx,FPU_fsqrt_arg_0 116 shll %edx /* max result was 7fff... */ 117 testl $0x80000000,%edx /* but min was 3fff... */ 120 movl $0x80000000,%edx /* round up */ 123 movl %edx,%esi /* Our first guess */ 140 movl %ecx,%edx /* msw of the arg / 2 */ 145 movl %ecx,%edx 150 movl %ecx,%edx [all …]
|
D | reg_norm.S | 32 movl SIGH(%ebx),%edx 35 orl %edx,%edx /* ms bits */ 42 movl %eax,%edx 48 bsrl %edx,%ecx /* get the required shift in %ecx */ 51 shld %cl,%eax,%edx 55 movl %edx,SIGH(%ebx) 110 movl SIGH(%ebx),%edx 113 orl %edx,%edx /* ms bits */ 120 movl %eax,%edx 126 bsrl %edx,%ecx /* get the required shift in %ecx */ [all …]
|
D | reg_u_add.S | 46 movl %ecx,%edx 55 movl PARAM7,%edx 68 xorl %edx,%edx /* clear the extension */ 83 shrd %cl,%ebx,%edx 95 shrd %cl,%eax,%edx 100 orl $1,%edx /* record the fact in the extension */ 108 movl %ebx,%edx 117 movl %eax,%edx 121 orl $1,%edx 125 movl $1,%edx /* The shifted nr always at least one '1' */ [all …]
|
/arch/x86/boot/compressed/ |
D | efi_stub_32.S | 41 1: popl %edx 42 subl $1b, %edx 51 movl %ecx, saved_return_addr(%edx) 54 movl %ecx, efi_rt_function_ptr(%edx) 68 1: popl %edx 69 subl $1b, %edx 71 movl efi_rt_function_ptr(%edx), %ecx 77 movl saved_return_addr(%edx), %ecx
|
/arch/x86/kernel/ |
D | ftrace_32.S | 59 pushl %edx 64 movl 4*4(%esp), %edx 67 lea 4*4(%esp), %edx 72 movl 4(%edx), %edx /* edx has ebp */ 82 popl %edx 130 pushl %edx 137 movl 15*4(%esp), %edx /* Load parent ip (2nd parameter) */ 139 movl 0x4(%ebp), %edx /* Load parent ip (2nd parameter) */ 159 popl %edx 196 pushl %edx [all …]
|
/arch/x86/realmode/rm/ |
D | reboot.S | 38 xorl %edx, %edx 97 movl %cr0, %edx 98 andl $0x00000011, %edx 99 orl $0x60000000, %edx 100 movl %edx, %cr0 102 movl %cr0, %edx 103 testl $0x60000000, %edx /* If no cache bits -> no wbinvd */ 108 movl %edx, %cr0
|
/arch/x86/boot/ |
D | cpucheck.c | 136 u32 eax, edx; in check_cpu() local 138 asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); in check_cpu() 140 asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); in check_cpu() 151 u32 eax, edx; in check_cpu() local 153 asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); in check_cpu() 155 asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); in check_cpu() 163 u32 eax, edx; in check_cpu() local 166 asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); in check_cpu() 167 asm("wrmsr" : : "a" (~0), "d" (edx), "c" (ecx)); in check_cpu() 171 asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); in check_cpu()
|
/arch/x86/power/ |
D | hibernate_asm_32.S | 42 movl restore_pblist, %edx 46 testl %edx, %edx 49 movl pbe_address(%edx), %esi 50 movl pbe_orig_address(%edx), %edi 56 movl pbe_next(%edx), %edx
|
/arch/x86/kernel/cpu/ |
D | topology.c | 33 unsigned int eax, ebx, ecx, edx; in detect_extended_topology_early() local 38 cpuid_count(0xb, SMT_LEVEL, &eax, &ebx, &ecx, &edx); in detect_extended_topology_early() 51 c->initial_apicid = edx; in detect_extended_topology_early() 65 unsigned int eax, ebx, ecx, edx, sub_index; in detect_extended_topology() local 75 cpuid_count(0xb, SMT_LEVEL, &eax, &ebx, &ecx, &edx); in detect_extended_topology() 81 cpuid_count(0xb, sub_index, &eax, &ebx, &ecx, &edx); in detect_extended_topology()
|