/arch/powerpc/mm/ |
D | tlb_low_64e.S | 67 std r14,EX_TLB_R14(r12) 68 ld r14,PACAPGD(r13) 78 ld r14,EX_TLB_CR(r12) 85 mtcr r14 86 ld r14,EX_TLB_R14(r12) 147 cmpldi cr0,r14,0 156 ldx r14,r14,r15 /* grab pgd entry */ 159 ldx r14,r14,r15 /* grab pgd entry */ 165 cmpdi cr0,r14,0 167 ldx r14,r14,r15 /* grab pud entry */ [all …]
|
/arch/m32r/lib/ |
D | ashxdi3.S | 24 jc r14 || cmpu r2, r3 30 jmp r14 36 or r1, r3 || jmp r14 44 jc r14 || cmpu r2, r3 49 jmp r14 55 or r0, r3 || jmp r14 61 jc r14 || cmpu r2, r3 66 jmp r14 72 or r1, r3 || jmp r14 81 jc r14 || cmpu r2, r3 [all …]
|
D | memcpy.S | 25 jc r14 || cmpeq r0, r1 ; return if r2=0 26 jc r14 ; return if r0=r1 38 addi r4, #4 || jc r14 ; return if r2=0 54 jmp r14 88 jmp r14
|
D | memset.S | 24 jc r14 51 jmp r14 69 jc r14 79 jc r14 95 jmp r14 134 jmp r14 156 jmp r14 174 jmp r14
|
/arch/powerpc/crypto/ |
D | aes-spe-keys.S | 26 stw r14,8(r1); /* save registers */ \ 31 lwz r14,8(r1); /* restore registers */ \ 93 mr r14,r8 /* apply LS_BOX to 4th temp */ 94 rotlwi r14,r14,8 95 LS_BOX(r14, r15, r4) 96 xor r14,r14,r0 97 xor r5,r5,r14 /* xor next 4 keys */ 105 GF8_MUL(r0, r0, r4, r14) /* multiply RCO by 2 in GF */ 139 mr r14,r10 /* apply LS_BOX to 6th temp */ 140 rotlwi r14,r14,8 [all …]
|
/arch/arm/crypto/ |
D | sha1-armv4-large.S | 65 mov r14,sp 94 str r9,[r14,#-4]! 119 str r9,[r14,#-4]! 144 str r9,[r14,#-4]! 169 str r9,[r14,#-4]! 194 str r9,[r14,#-4]! 196 cmp r14,sp 222 str r9,[r14,#-4]! 224 ldr r9,[r14,#15*4] 225 ldr r10,[r14,#13*4] [all …]
|
/arch/ia64/kernel/ |
D | gate.S | 167 mov r14=ar.bsp 169 cmp.ne p1,p0=r14,r15 // do we need to restore the rbs? 201 adds r14=(RNAT_OFF+SIGCONTEXT_OFF),sp 208 st8 [r14]=r19 // save sc_ar_rnat 211 adds r14=(LOADRS_OFF+SIGCONTEXT_OFF),sp 221 st8 [r14]=r15 // save sc_loadrs 255 sub r14=r14,r17 // r14 (bspstore1) <- bsp1 - (sc_loadrs >> 16) 259 extr.u r14=r14,3,6 // r14 <- rse_slot_num(bspstore1) 261 add r14=r14,r17 // r14 <- rse_slot_num(bspstore1) + (sc_loadrs >> 19) 263 shr.u r14=r14,6 // r14 <- (rse_slot_num(bspstore1) + (sc_loadrs >> 19))/0x40 [all …]
|
D | head.S | 211 SAVE_REGION_REGS(r2, r8,r9,r10,r11,r12,r13,r14,r15); 277 STORE_REGION_REGS(r16, r8,r9,r10,r11,r12,r13,r14,r15); 640 adds r14=256,in0 647 ldf.fill.nta f48=[r14],loc0 652 ldf.fill.nta f80=[r14],loc0 657 ldf.fill.nta f112=[r14],loc1 662 ldf.fill.nta f49=[r14],loc0 667 ldf.fill.nta f81=[r14],loc0 672 ldf.fill.nta f113=[r14],loc1 677 ldf.fill.nta f50=[r14],loc0 [all …]
|
/arch/s390/kernel/ |
D | mcount.S | 17 br %r14 25 br %r14 40 stmg %r2,%r14,(STACK_PTREGS_GPRS+2*8)(%r15) 53 lgr %r3,%r14 55 basr %r14,%r1 63 brasl %r14,prepare_ftrace_return 79 brasl %r14,ftrace_return_to_handler 81 lgr %r14,%r2 83 br %r14
|
D | entry.S | 61 brasl %r14,trace_hardirqs_on_caller 68 brasl %r14,trace_hardirqs_off_caller 76 brasl %r14,lockdep_sys_exit 83 lghi %r14,\savearea 91 lgr %r14,%r9 92 slg %r14,BASED(.Lcritical_start) 93 clg %r14,BASED(.Lcritical_length) 96 brasl %r14,cleanup_critical 99 0: lg %r14,__LC_ASYNC_STACK # are we already on the async stack? 100 slgr %r14,%r15 [all …]
|
D | head_kdump.S | 33 brasl %r14,startup_kdump_relocated 47 lghi %r14,DATAMOVER_ADDR # Jump to copied data mover 48 basr %r14,%r14 76 lg %r14,.Lstartup_kdump-0b(%r13) 77 basr %r14,%r14 # Start relocated kernel
|
D | reipl.S | 88 brasl %r14,store_status 101 bas %r14,.Ldisab-.Lpg0(%r13) 111 bas %r14,.Ldisab-.Lpg0(%r13) 114 bas %r14,.Ldisab-.Lpg0(%r13) 120 .Ldisab: sll %r14,1 121 srl %r14,1 # need to kill hi bit to avoid specification exceptions. 122 st %r14,.Ldispsw+12-.Lpg0(%r13)
|
D | swsusp.S | 34 brasl %r14,save_fpu_regs 98 brasl %r14,swsusp_save 105 br %r14 120 brasl %r14,arch_set_page_states 203 basr %r14,%r3 261 brasl %r14,arch_set_page_states 264 brasl %r14,s390_early_resume 269 br %r14
|
/arch/arm/mach-s3c24xx/ |
D | sleep.S | 67 mov r14, #S3C24XX_PA_GPIO 68 ldr r12, [ r14, #0x54 ] 71 str r12, [ r14, #0x54 ] 78 ldrb r14, [ r3, #S3C2410_UTRSTAT ] 79 tst r14, #S3C2410_UTRSTAT_TXE
|
/arch/s390/net/ |
D | bpf_jit.S | 55 LOAD %r14,-SIZE(%r3,%r12); /* Get data from skb */ \ 63 brasl %r14,skb_copy_bits; /* Get data from skb */ \ 64 LOAD %r14,STK_OFF_TMP(%r15); /* Load from temp bufffer */ \ 81 llgc %r14,0(%r3,%r12) # Get byte from skb 89 brasl %r14,skb_copy_bits # Get data from skb 90 llgc %r14,STK_OFF_TMP(%r15) # Load result from temp buffer 101 brasl %r14,bpf_internal_load_pointer_neg_helper; \ 104 LOAD %r14,0(%r2); /* Get data from pointer */ \
|
/arch/s390/lib/ |
D | mem.S | 27 bzr %r14 42 br %r14 47 ber %r14 59 br %r14 73 bzr %r14 82 br %r14
|
/arch/powerpc/include/asm/ |
D | exception-64e.h | 107 std r14,EX_TLB_R14(r12); \ 108 addi r14,r12,EX_TLB_SIZE; \ 115 mtspr SPRN_SPRG_TLB_EXFRAME,r14; \ 138 ld r14,EX_TLB_CR(r12); \ 144 mtcr r14; \ 146 ld r14,EX_TLB_R14(r12); \ 191 std r14,EX_TLB_ESR(r12); /* save ESR */ 193 std r14,PACA_EXTLB+EX_TLB_ESR(r13); /* save ESR */
|
/arch/powerpc/lib/ |
D | checksum_64.S | 67 std r14,STK_REG(R14)(r1) 87 ld r14,40(r3) 100 adde r0,r0,r14 114 ld r14,40(r3) 124 adde r0,r0,r14 128 ld r14,STK_REG(R14)(r1) 269 std r14,STK_REG(R14)(r1) 289 source; ld r14,40(r3) 306 dest; std r14,40(r4) 308 adde r0,r0,r14 [all …]
|
/arch/x86/entry/ |
D | entry_64_compat.S | 102 xorq %r14, %r14 /* nospec r14 */ 238 xorq %r14, %r14 /* nospec r14 */ 358 pushq %r14 /* pt_regs->r14 */ 359 xorq %r14, %r14 /* nospec r14 */
|
/arch/nios2/kernel/ |
D | insnemu.S | 46 ldw r14, PT_R14(sp) 141 stw r14, 56(sp) 195 movi r14, 32 336 subi r14, r14, 1 337 bne r14, zero, divide_loop 489 subi r14, r14, 1 490 bne r14, zero, multiply_loop 572 ldw r14, 56(sp)
|
/arch/powerpc/kernel/ |
D | idle_book3s.S | 99 3: lwz r15,0(r14) 103 lwarx r15,0,r14 214 ld r14,PACA_CORE_IDLE_STATE_PTR(r13) 216 lwarx r15,0,r14 237 stwcx. r15,0,r14 247 stwcx. r15,0,r14 259 stw r0,0(r14) 292 ld r14,PACA_CORE_IDLE_STATE_PTR(r13) 295 lwarx r15,0,r14 300 stwcx. r15,0,r14 [all …]
|
D | exceptions-64e.S | 359 std r14,PACA_EXGEN+EX_R14(r13); \ 363 std r14,PACA_EXGEN+EX_R14(r13); 366 std r14,PACA_EXCRIT+EX_R14(r13); \ 370 std r14,PACA_EXDBG+EX_R14(r13); \ 374 std r14,PACA_EXMC+EX_R14(r13); \ 576 mfspr r14,SPRN_DEAR 587 mr r14,r10 600 mfspr r14,SPRN_DEAR 609 mfspr r14,SPRN_ESR 612 std r14,_DSISR(r1) [all …]
|
/arch/m32r/include/asm/ |
D | assembler.h | 86 push lr ; r14 171 pop r14 172 mvtc r14, psw 173 pop r14 174 mvtc r14, bpc 176 pop r14 177 mvtc r14, cr3 ; spu 179 pop lr ; r14
|
/arch/x86/um/ |
D | setjmp_64.S | 14 # %r14 32 movq %r14,40(%rdi) 50 movq 40(%rdi),%r14
|
/arch/powerpc/kvm/ |
D | book3s_interrupts.S | 42 PPC_LL r14, VCPU_GPR(R14)(vcpu); \ 88 mfcr r14 89 stw r14, _CCR(r1) 186 PPC_STL r14, VCPU_GPR(R14)(r7) 224 lwz r14, _CCR(r1) 225 mtcr r14
|