/arch/sh/kernel/ |
D | relocate_kernel.S | 26 mov.l r15, @-r0 27 mov r0, r15 28 mov.l r14, @-r15 29 mov.l r13, @-r15 30 mov.l r12, @-r15 31 mov.l r11, @-r15 32 mov.l r10, @-r15 33 mov.l r9, @-r15 34 mov.l r8, @-r15 37 sts.l macl, @-r15 [all …]
|
D | entry-common.S | 85 mov.l @(r0,r15), r0 ! get status register 200 mov.l @(OFF_R4,r15), r4 ! arg0 201 mov.l @(OFF_R5,r15), r5 202 mov.l @(OFF_R6,r15), r6 203 mov.l @(OFF_R7,r15), r7 ! arg3 204 mov.l @(OFF_R3,r15), r3 ! syscall_nr 211 mov.l r0, @(OFF_R0,r15) ! Return value 215 mov.l @(r0,r15), r0 ! get status register 239 mov.l r0, @(OFF_R0,r15) ! Return value 305 mov.l @(OFF_R5,r15), r5 ! fn [all …]
|
/arch/sh/include/asm/ |
D | entry-macros.S | 29 mov r15, \ti 37 mov.l r0, @-r15 38 mov.l r1, @-r15 39 mov.l r2, @-r15 40 mov.l r3, @-r15 41 mov.l r4, @-r15 42 mov.l r5, @-r15 43 mov.l r6, @-r15 44 mov.l r7, @-r15 50 mov.l @r15+, r7 [all …]
|
/arch/powerpc/mm/ |
D | tlb_low_64e.S | 69 std r15,EX_TLB_R15(r12) 94 ld r15,EX_TLB_R15(r12) 128 srdi r15,r16,60 /* get region */ 134 cmpwi r15,0 /* user vs kernel check */ 153 rldicl r15,r16,64-PGDIR_SHIFT+3,64-PGD_INDEX_SIZE-3 155 clrrdi r15,r15,3 163 ldx r14,r14,r15 /* grab pgd entry */ 166 ldx r14,r14,r15 /* grab pgd entry */ 170 rldicl r15,r16,64-PUD_SHIFT+3,64-PUD_INDEX_SIZE-3 171 clrrdi r15,r15,3 [all …]
|
/arch/sh/kernel/cpu/sh2/ |
D | entry.S | 51 mov.l @(5*4,r15),r3 ! previous SR 55 mov.l r3,@(5*4,r15) ! update SR 64 mov r15,r2 ! r2 = user stack top 65 mov r0,r15 ! switch kernel stack 66 mov.l r1,@-r15 ! TRA 67 sts.l macl, @-r15 68 sts.l mach, @-r15 69 stc.l gbr, @-r15 71 mov.l r0,@-r15 ! original SR 72 sts.l pr,@-r15 [all …]
|
/arch/s390/kernel/ |
D | mcount.S | 39 stg %r14,(__SF_GPRS+8*8)(%r15) # save traced function caller 40 lgr %r1,%r15 45 aghi %r15,-TRACED_FUNC_FRAME_SIZE 46 stg %r1,__SF_BACKCHAIN(%r15) 47 stg %r0,(__SF_GPRS+8*8)(%r15) 48 stg %r15,(__SF_GPRS+9*8)(%r15) 50 aghi %r15,-STACK_FRAME_SIZE 51 stg %r1,(STACK_PTREGS_GPRS+15*8)(%r15) 53 stg %r1,__SF_BACKCHAIN(%r15) 54 stg %r0,(STACK_PTREGS_PSW+8)(%r15) [all …]
|
D | swsusp.S | 31 stmg %r6,%r15,__SF_GPRS(%r15) 32 lgr %r1,%r15 33 aghi %r15,-STACK_FRAME_OVERHEAD 34 stg %r1,__SF_BACKCHAIN(%r15) 40 stnsm __SF_EMPTY(%r15),0xfb 43 stpx __SF_EMPTY(%r15) 46 llgf %r10,__SF_EMPTY(%r15) 55 mvc 0x318(4,%r1),__SF_EMPTY(%r15) /* move prefix to lowcore */ 58 stmg %r0,%r15,0x280(%r1) /* store general registers */ 61 stck __SF_EMPTY(%r15) /* store clock */ [all …]
|
D | entry.S | 82 tml %r15,\stacksize - CONFIG_STACK_GUARD 100 slgr %r14,%r15 104 aghi %r15,-(STACK_FRAME_OVERHEAD + __PT_SIZE) 107 UPDATE_VTIME %r14,%r15,\timer 109 2: lg %r15,__LC_ASYNC_STACK # load async stack 110 3: la %r11,STACK_FRAME_OVERHEAD(%r15) 248 stmg %r6,%r15,__SF_GPRS(%r15) # store gprs of prev task 253 stg %r15,__THREAD_ksp(%r1) # store kernel stack of prev 256 lgr %r15,%r5 257 aghi %r15,STACK_INIT # end of kernel stack of next [all …]
|
D | head64.S | 35 larl %r15,init_thread_union 36 stg %r15,__LC_THREAD_INFO # cache thread info in lowcore 37 lg %r14,__TI_task(%r15) # cache current in lowcore 39 aghi %r15,1<<(PAGE_SHIFT+THREAD_ORDER) # init_task_union + THREAD_SIZE 40 stg %r15,__LC_KERNEL_STACK # set end of kernel stack 41 aghi %r15,-160 95 stctg %c0,%c15,0(%r15) 96 oi 6(%r15),0x60 # enable sigp emergency & external call 97 oi 4(%r15),0x10 # switch on low address proctection 98 lctlg %c0,%c15,0(%r15)
|
/arch/x86/platform/efi/ |
D | efi_stub_64.S | 43 movq %r15, efi_scratch(%rip) 45 movq %cr4, %r15 46 movq %r15, %r14 49 movq %r15, %cr4 51 movq efi_scratch(%rip), %r15 57 movq %r15, efi_scratch(%rip) # r15 59 movq %cr3, %r15 60 movq %r15, efi_scratch+8(%rip) # prev_cr3 61 movq efi_scratch+16(%rip), %r15 # EFI pgt 62 movq %r15, %cr3 [all …]
|
/arch/sh/kernel/cpu/sh2a/ |
D | entry.S | 50 bst.b #6,@(4*4,r15) !set cpu mode to SR.MD 59 mov r15,r2 ! r2 = user stack top 60 mov r0,r15 ! switch kernel stack 61 mov.l r1,@-r15 ! TRA 62 sts.l macl, @-r15 63 sts.l mach, @-r15 64 stc.l gbr, @-r15 66 mov.l r0,@-r15 ! original SR 67 sts.l pr,@-r15 69 mov.l r0,@-r15 ! original PC [all …]
|
/arch/sh/kernel/cpu/shmobile/ |
D | sleep.S | 56 sts.l pr, @-r15 57 mov.l r14, @-r15 58 mov.l r13, @-r15 59 mov.l r12, @-r15 60 mov.l r11, @-r15 61 mov.l r10, @-r15 62 mov.l r9, @-r15 63 mov.l r8, @-r15 91 mov.l r15, @(SH_SLEEP_SP, r5) 92 mov r5, r15 [all …]
|
/arch/sh/lib/ |
D | mcount.S | 16 mov.l r4, @-r15; \ 17 mov.l r5, @-r15; \ 18 mov.l r6, @-r15; \ 19 mov.l r7, @-r15; \ 20 sts.l pr, @-r15; \ 22 mov.l @(20,r15),r4; \ 26 lds.l @r15+, pr; \ 27 mov.l @r15+, r7; \ 28 mov.l @r15+, r6; \ 29 mov.l @r15+, r5; \ [all …]
|
D | udivsi3_i4i.S | 75 mov.l r4,@-r15 77 mov.l r1,@-r15 88 mov.l r4,@-r15 91 mov.l r5,@-r15 102 mov.l @r15+,r5 106 mov.l @r15+,r4 114 mov.l @r15+,r5 116 mov.l @r15+,r4 129 mov.l r4,@-r15 131 mov.l r1,@-r15 [all …]
|
D | udivsi3_i4i-Os.S | 49 mov.l r4,@-r15 56 mov.l r5,@-r15 73 mov.l @r15+,r5 75 mov.l @r15+,r4 94 mov.l r5,@-r15 102 mov.l @r15+,r5 103 mov.l @r15+,r4 113 mov.l r4,@-r15 115 mov.l r5,@-r15
|
/arch/sh/kernel/cpu/sh3/ |
D | entry.S | 142 mov r15, r4 157 mov r15, r4 183 mov r15, r4 ! regs 198 mov k4, r15 212 ! - restore r8, r9, r10, r11, r12, r13, r14, r15 from the stack 222 mov.l @r15+, r0 223 mov.l @r15+, r1 224 mov.l @r15+, r2 225 mov.l @r15+, r3 226 mov.l @r15+, r4 [all …]
|
D | swsusp.S | 27 mov.l 1f, r15 72 mov.l @r15+, r0 73 mov.l @r15+, r1 74 mov.l @r15+, r2 75 mov.l @r15+, r3 76 mov.l @r15+, r4 77 mov.l @r15+, r5 78 mov.l @r15+, r6 79 mov.l @r15+, r7 98 mov r15, r2 ! save sp in r2 [all …]
|
/arch/x86/crypto/ |
D | salsa20-x86_64-asm_64.S | 33 # r15_stack = r15 34 movq %r15,32(%rsp) 148 movq 112(%rsp),%r15 150 mov %r15,%rbx 152 shr $32,%r15 154 movq %r15,176(%rsp) 156 mov $20,%r15 160 movq %r15,184(%rsp) 162 movq 160(%rsp),%r15 170 lea (%rdi,%r15),%rbp [all …]
|
/arch/powerpc/kernel/ |
D | idle_power7.S | 63 3: lwz r15,0(r14) 64 andi. r15,r15,PNV_CORE_IDLE_LOCK_BIT 67 lwarx r15,0,r14 168 lwarx r15,0,r14 170 andi. r9,r15,PNV_CORE_IDLE_LOCK_BIT 173 andc r15,r15,r7 /* Clear thread bit */ 175 andi. r15,r15,PNV_CORE_IDLE_THREAD_BITS 189 stwcx. r15,0,r14 198 ori r15,r15,PNV_CORE_IDLE_LOCK_BIT 199 stwcx. r15,0,r14 [all …]
|
/arch/microblaze/kernel/ |
D | mcount.S | 82 rtsd r15, 8; 89 rtsd r15, 8 93 swi r15, r1, 0; 113 addik r6, r15, 0; /* MS: load current function addr */ 114 bralid r15, prepare_ftrace_return; 136 addik r5, r15, -4; /* MS: load current function addr */ 138 brald r15, r20; /* MS: jump to ftrace handler */ 141 lwi r15, r1, 0; 144 rtsd r15, 8; /* MS: jump back */ 152 swi r15, r1, 0; [all …]
|
/arch/s390/kernel/vdso32/ |
D | gettimeofday.S | 22 ahi %r15,-16 33 stcke 0(%r15) /* Store TOD clock */ 34 lm %r0,%r1,1(%r15) 40 st %r0,0(%r15) 46 4: al %r0,0(%r15) 51 5: mvc 0(4,%r15),__VDSO_XTIME_SEC+4(%r5) 56 l %r4,0(%r15) /* get tv_sec from stack */ 75 ahi %r15,16
|
/arch/ia64/lib/ |
D | ip_fast_csum.S | 38 add r15=4,in0 // second source pointer 43 (p7) ld4 r21=[r15],8 47 ld4 r23=[r15],8 103 dep r15=in3,in2,32,16 107 mux1 r15=r15,@rev 114 shr.u r15=r15,16 122 add r15=r15,in4 124 add r8=r8,r15
|
/arch/powerpc/crypto/ |
D | sha1-powerpc-asm.S | 33 add r0,RE(t),r15; \ 46 add r0,RE(t),r15; \ 60 add r0,RE(t),r15; \ 70 add r0,RE(t),r15; \ 89 add r0,RE(t),r15; \ 132 lis r15,0x5a82 /* K0-19 */ 133 ori r15,r15,0x7999 140 lis r15,0x6ed9 /* K20-39 */ 141 ori r15,r15,0xeba1 144 lis r15,0x8f1b /* K40-59 */ [all …]
|
/arch/s390/net/ |
D | bpf_jit.S | 54 clg %r3,STK_OFF_HLEN(%r15); /* Offset + SIZE > hlen? */ \ 62 la %r4,STK_OFF_TMP(%r15); /* Arg3 = temp bufffer */ \ 65 LOAD %r14,STK_OFF_TMP(%r15); /* Load from temp bufffer */ \ 83 clg %r3,STK_OFF_HLEN(%r15) # Offset >= hlen? 91 la %r4,STK_OFF_TMP(%r15) # Arg3 = pointer to temp buffer 94 llgc %r14,STK_OFF_TMP(%r15) # Load result from temp buffer 118 ltgr %r15,%r15 # Set condition code
|
/arch/ia64/kernel/ |
D | gate.S | 124 ld8 r15=[base1] // get address of new RBS base (or NULL) 127 cmp.ne p1,p0=r15,r0 // do we need to switch rbs? (note: pr is saved by kernel) 166 ld8 r15=[base0] // fetch sc_ar_bsp 169 cmp.ne p1,p0=r14,r15 // do we need to restore the rbs? 190 mov r15=__NR_rt_sigreturn 204 mov ar.bspstore=r15 // switch over to new register backing store area 214 sub r15=r16,r15 219 shl r15=r15,16 221 st8 [r14]=r15 // save sc_loadrs 250 extr.u r18=r15,3,6 // r18 <- rse_slot_num(bsp0) [all …]
|