/arch/powerpc/mm/ |
D | tlb_low_64e.S | 67 std r14,EX_TLB_R14(r12) 68 ld r14,PACAPGD(r13) 85 ld r14,EX_TLB_CR(r12) 92 mtcr r14 93 ld r14,EX_TLB_R14(r12) 154 cmpldi cr0,r14,0 163 ldx r14,r14,r15 /* grab pgd entry */ 166 ldx r14,r14,r15 /* grab pgd entry */ 172 cmpdi cr0,r14,0 174 ldx r14,r14,r15 /* grab pud entry */ [all …]
|
/arch/m32r/lib/ |
D | ashxdi3.S | 25 jc r14 || cmpu r2, r3 31 jmp r14 37 or r1, r3 || jmp r14 45 jc r14 || cmpu r2, r3 50 jmp r14 56 or r0, r3 || jmp r14 62 jc r14 || cmpu r2, r3 67 jmp r14 73 or r1, r3 || jmp r14 82 jc r14 || cmpu r2, r3 [all …]
|
D | memcpy.S | 26 jc r14 || cmpeq r0, r1 ; return if r2=0 27 jc r14 ; return if r0=r1 39 addi r4, #4 || jc r14 ; return if r2=0 55 jmp r14 89 jmp r14
|
D | memset.S | 25 jc r14 52 jmp r14 70 jc r14 80 jc r14 96 jmp r14 135 jmp r14 157 jmp r14 175 jmp r14
|
/arch/powerpc/crypto/ |
D | aes-spe-keys.S | 26 stw r14,8(r1); /* save registers */ \ 31 lwz r14,8(r1); /* restore registers */ \ 93 mr r14,r8 /* apply LS_BOX to 4th temp */ 94 rotlwi r14,r14,8 95 LS_BOX(r14, r15, r4) 96 xor r14,r14,r0 97 xor r5,r5,r14 /* xor next 4 keys */ 105 GF8_MUL(r0, r0, r4, r14) /* multiply RCO by 2 in GF */ 139 mr r14,r10 /* apply LS_BOX to 6th temp */ 140 rotlwi r14,r14,8 [all …]
|
/arch/arm/crypto/ |
D | sha1-armv4-large.S | 65 mov r14,sp 94 str r9,[r14,#-4]! 119 str r9,[r14,#-4]! 144 str r9,[r14,#-4]! 169 str r9,[r14,#-4]! 194 str r9,[r14,#-4]! 196 cmp r14,sp 222 str r9,[r14,#-4]! 224 ldr r9,[r14,#15*4] 225 ldr r10,[r14,#13*4] [all …]
|
/arch/ia64/kernel/ |
D | gate.S | 168 mov r14=ar.bsp 170 cmp.ne p1,p0=r14,r15 // do we need to restore the rbs? 202 adds r14=(RNAT_OFF+SIGCONTEXT_OFF),sp 209 st8 [r14]=r19 // save sc_ar_rnat 212 adds r14=(LOADRS_OFF+SIGCONTEXT_OFF),sp 222 st8 [r14]=r15 // save sc_loadrs 256 sub r14=r14,r17 // r14 (bspstore1) <- bsp1 - (sc_loadrs >> 16) 260 extr.u r14=r14,3,6 // r14 <- rse_slot_num(bspstore1) 262 add r14=r14,r17 // r14 <- rse_slot_num(bspstore1) + (sc_loadrs >> 19) 264 shr.u r14=r14,6 // r14 <- (rse_slot_num(bspstore1) + (sc_loadrs >> 19))/0x40 [all …]
|
D | head.S | 212 SAVE_REGION_REGS(r2, r8,r9,r10,r11,r12,r13,r14,r15); 278 STORE_REGION_REGS(r16, r8,r9,r10,r11,r12,r13,r14,r15); 641 adds r14=256,in0 648 ldf.fill.nta f48=[r14],loc0 653 ldf.fill.nta f80=[r14],loc0 658 ldf.fill.nta f112=[r14],loc1 663 ldf.fill.nta f49=[r14],loc0 668 ldf.fill.nta f81=[r14],loc0 673 ldf.fill.nta f113=[r14],loc1 678 ldf.fill.nta f50=[r14],loc0 [all …]
|
/arch/s390/kernel/ |
D | mcount.S | 17 GEN_BR_THUNK %r14 22 BR_EX %r14 36 BR_EX %r14 58 stmg %r2,%r14,(STACK_PTREGS_GPRS+2*8)(%r15) 71 lgr %r3,%r14 73 BASR_EX %r14,%r1 81 brasl %r14,prepare_ftrace_return 97 brasl %r14,ftrace_return_to_handler 99 lgr %r14,%r2 101 BR_EX %r14
|
D | entry.S | 64 brasl %r14,trace_hardirqs_on_caller 71 brasl %r14,trace_hardirqs_off_caller 79 brasl %r14,lockdep_sys_exit 86 lghi %r14,\savearea 94 lgr %r14,%r9 95 slg %r14,BASED(.Lcritical_start) 96 clg %r14,BASED(.Lcritical_length) 99 brasl %r14,cleanup_critical 102 0: lg %r14,__LC_ASYNC_STACK # are we already on the async stack? 103 slgr %r14,%r15 [all …]
|
D | head_kdump.S | 34 brasl %r14,startup_kdump_relocated 48 lghi %r14,DATAMOVER_ADDR # Jump to copied data mover 49 basr %r14,%r14 77 lg %r14,.Lstartup_kdump-0b(%r13) 78 basr %r14,%r14 # Start relocated kernel
|
D | swsusp.S | 28 GEN_BR_THUNK %r14 38 brasl %r14,save_fpu_regs 102 brasl %r14,swsusp_save 109 BR_EX %r14 124 brasl %r14,arch_set_page_states 206 brasl %r14,sclp_early_printk 264 brasl %r14,arch_set_page_states 267 brasl %r14,s390_early_resume 272 BR_EX %r14
|
D | reipl.S | 92 brasl %r14,store_status 105 bas %r14,.Ldisab-.Lpg0(%r13) 115 bas %r14,.Ldisab-.Lpg0(%r13) 118 bas %r14,.Ldisab-.Lpg0(%r13) 124 .Ldisab: sll %r14,1 125 srl %r14,1 # need to kill hi bit to avoid specification exceptions. 126 st %r14,.Ldispsw+12-.Lpg0(%r13)
|
D | base.S | 17 GEN_BR_THUNK %r14 27 BASR_EX %r14,%r9 47 BASR_EX %r14,%r9 67 BASR_EX %r14,%r9 124 BR_EX %r14
|
/arch/arm/mach-s3c24xx/ |
D | sleep.S | 68 mov r14, #S3C24XX_PA_GPIO 69 ldr r12, [ r14, #0x54 ] 72 str r12, [ r14, #0x54 ] 79 ldrb r14, [ r3, #S3C2410_UTRSTAT ] 80 tst r14, #S3C2410_UTRSTAT_TXE
|
/arch/s390/net/ |
D | bpf_jit.S | 57 LOAD %r14,-SIZE(%r3,%r12); /* Get data from skb */ \ 65 brasl %r14,skb_copy_bits; /* Get data from skb */ \ 66 LOAD %r14,STK_OFF_TMP(%r15); /* Load from temp bufffer */ \ 86 llgc %r14,0(%r3,%r12) # Get byte from skb 94 brasl %r14,skb_copy_bits # Get data from skb 95 llgc %r14,STK_OFF_TMP(%r15) # Load result from temp buffer 106 brasl %r14,bpf_internal_load_pointer_neg_helper; \ 109 LOAD %r14,0(%r2); /* Get data from pointer */ \
|
/arch/powerpc/include/asm/ |
D | exception-64e.h | 107 std r14,EX_TLB_R14(r12); \ 108 addi r14,r12,EX_TLB_SIZE; \ 115 mtspr SPRN_SPRG_TLB_EXFRAME,r14; \ 138 ld r14,EX_TLB_CR(r12); \ 144 mtcr r14; \ 146 ld r14,EX_TLB_R14(r12); \ 191 std r14,EX_TLB_ESR(r12); /* save ESR */ 193 std r14,PACA_EXTLB+EX_TLB_ESR(r13); /* save ESR */
|
/arch/powerpc/lib/ |
D | checksum_64.S | 67 std r14,STK_REG(R14)(r1) 87 ld r14,40(r3) 100 adde r0,r0,r14 114 ld r14,40(r3) 124 adde r0,r0,r14 128 ld r14,STK_REG(R14)(r1) 261 std r14,STK_REG(R14)(r1) 281 source; ld r14,40(r3) 298 dest; std r14,40(r4) 300 adde r0,r0,r14 [all …]
|
/arch/s390/lib/ |
D | mem.S | 12 GEN_BR_THUNK %r14 40 BR_EX %r14 47 BR_EX %r14 85 BR_EX %r14 103 BR_EX %r14 127 BR_EX %r14
|
/arch/nios2/kernel/ |
D | insnemu.S | 46 ldw r14, PT_R14(sp) 141 stw r14, 56(sp) 195 movi r14, 32 336 subi r14, r14, 1 337 bne r14, zero, divide_loop 489 subi r14, r14, 1 490 bne r14, zero, multiply_loop 572 ldw r14, 56(sp)
|
/arch/powerpc/kernel/ |
D | exceptions-64e.S | 384 std r14,PACA_EXGEN+EX_R14(r13); \ 388 std r14,PACA_EXGEN+EX_R14(r13); 391 std r14,PACA_EXCRIT+EX_R14(r13); \ 395 std r14,PACA_EXDBG+EX_R14(r13); \ 399 std r14,PACA_EXMC+EX_R14(r13); \ 601 mfspr r14,SPRN_DEAR 612 mr r14,r10 625 mfspr r14,SPRN_DEAR 634 mfspr r14,SPRN_ESR 637 std r14,_DSISR(r1) [all …]
|
/arch/m32r/include/asm/ |
D | assembler.h | 87 push lr ; r14 172 pop r14 173 mvtc r14, psw 174 pop r14 175 mvtc r14, bpc 177 pop r14 178 mvtc r14, cr3 ; spu 180 pop lr ; r14
|
/arch/x86/um/ |
D | setjmp_64.S | 15 # %r14 33 movq %r14,40(%rdi) 51 movq 40(%rdi),%r14
|
/arch/powerpc/kvm/ |
D | book3s_interrupts.S | 42 PPC_LL r14, VCPU_GPR(R14)(vcpu); \ 88 mfcr r14 89 stw r14, _CCR(r1) 186 PPC_STL r14, VCPU_GPR(R14)(r7) 224 lwz r14, _CCR(r1) 225 mtcr r14
|
/arch/x86/lib/ |
D | csum-copy_64.S | 56 movq %r14, 4*8(%rsp) 94 movq 48(%rdi), %r14 107 adcq %r14, %rax 126 movq %r14, 48(%rsi) 204 movq 4*8(%rsp), %r14
|