/arch/powerpc/platforms/52xx/ |
D | mpc52xx_sleep.S | 21 lwz r8, 0x14(r6) /* intr->main_mask */ 22 ori r8, r8, 0x1 23 xori r8, r8, 0x1 24 stw r8, 0x14(r6) 28 li r8, 0x1 29 stw r8, 0x40(r6) /* intr->main_emulate */ 69 lwz r8, 0x4(r4) /* sdram->ctrl */ 71 oris r8, r8, 0x8000 /* mode_en */ 72 stw r8, 0x4(r4) 75 ori r8, r8, 0x0002 /* soft_pre */ [all …]
|
/arch/powerpc/lib/ |
D | strlen_32.S | 55 2: subf r8, r7, r9 56 and. r8, r8, r6 58 andc. r8, r8, r9 60 andc r8, r9, r6 62 subfe r8, r6, r8 63 nor r8, r8, r9 64 cntlzw r8, r8 66 srwi r8, r8, 3 67 add r3, r3, r8 72 orc r8, r8, r8 [all …]
|
D | copypage_64.S | 31 srd r8,r5,r11 33 mtctr r8 41 srdi r8,r5,7 /* page is copied in 128 byte strides */ 42 addi r8,r8,-1 /* one stride copied outside loop */ 44 mtctr r8 49 ldu r8,24(r4) 55 std r8,32(r3) 65 ld r8,64(r4) 71 std r8,96(r3) 81 ldu r8,128(r4) [all …]
|
/arch/powerpc/mm/book3s32/ |
D | hash_low.S | 50 lis r8, (mmu_hash_lock - PAGE_OFFSET)@h 51 ori r8, r8, (mmu_hash_lock - PAGE_OFFSET)@l 54 11: lwz r6,0(r8) 57 10: lwarx r6,0,r8 60 stwcx. r0,0,r8 76 lwz r8,0(r5) /* get pmd entry */ 77 rlwinm. r8,r8,0,0,19 /* extract address of pte page */ 79 rlwinm r8,r4,13,19,29 /* Compute pgdir/pmd offset */ 80 lwzx r8,r8,r5 /* Get L1 entry */ 81 rlwinm. r8,r8,0,0,20 /* extract pt base address */ [all …]
|
/arch/arm/lib/ |
D | memset.S | 36 stmfd sp!, {r8, lr} 39 UNWIND( .save {r8, lr} ) 40 mov r8, r1 44 stmiage ip!, {r1, r3, r8, lr} @ 64 bytes at a time. 45 stmiage ip!, {r1, r3, r8, lr} 46 stmiage ip!, {r1, r3, r8, lr} 47 stmiage ip!, {r1, r3, r8, lr} 49 ldmfdeq sp!, {r8, pc} @ Now <64 bytes to go. 54 stmiane ip!, {r1, r3, r8, lr} 55 stmiane ip!, {r1, r3, r8, lr} [all …]
|
/arch/powerpc/kernel/vdso32/ |
D | cacheflush.S | 35 subf r8,r6,r4 /* compute length */ 36 add r8,r8,r5 /* ensure we get enough */ 38 srw. r8,r8,r9 /* compute line count */ 41 mtctr r8 52 subf r8,r6,r4 /* compute length */ 53 add r8,r8,r5 55 srw. r8,r8,r9 /* compute line count */ 58 mtctr r8
|
/arch/powerpc/kernel/vdso64/ |
D | cacheflush.S | 35 subf r8,r6,r4 /* compute length */ 36 add r8,r8,r5 /* ensure we get enough */ 38 srd. r8,r8,r9 /* compute line count */ 41 mtctr r8 52 subf r8,r6,r4 /* compute length */ 53 add r8,r8,r5 55 srd. r8,r8,r9 /* compute line count */ 58 mtctr r8
|
/arch/sh/kernel/cpu/sh2a/ |
D | entry.S | 71 mov r2,r8 ! r8 = previus stack top 74 mov.l @r8+,r2 75 mov.l @r8+,r0 76 mov.l @r8+,r1 84 mov r2,r8 ! r8 = previous stack top 87 mov.l @r8+,r2 ! old R2 88 mov.l @r8+,r0 ! old R0 89 mov.l @r8+,r1 ! old R1 90 mov.l @r8+,r10 ! old PC 91 mov.l @r8+,r11 ! old SR [all …]
|
/arch/ia64/lib/ |
D | ip_fast_csum.S | 33 #define ret0 r8 119 add r8=r16,r17 122 add r8=r8,r18 124 add r8=r8,r19 127 add r8=r8,r15 129 shr.u r10=r8,32 // now fold sum into short 130 zxt4 r11=r8 132 add r8=r10,r11 134 shr.u r10=r8,16 // yeah, keep it rolling 135 zxt2 r11=r8 [all …]
|
D | strncpy_from_user.S | 25 mov r8=0 34 EX(.Lexit, ld1 r8=[in1],1) 36 EX(.Lexit, st1 [in0]=r8,1) 37 cmp.ne p6,p7=r8,r0 42 (p6) mov r8=in2 // buffer filled up---return buffer length 43 (p7) sub r8=in1,r9,1 // return string length (excluding NUL character)
|
/arch/microblaze/lib/ |
D | fastcopy.S | 103 andi r8, r6, 0xfffffffc /* as = s & ~3 */ 105 lwi r11, r8, 0 /* h = *(as + 0) */ 115 lwi r12, r8, 4 /* v = *(as + 4) */ 120 lwi r12, r8, 8 /* v = *(as + 8) */ 125 lwi r12, r8, 12 /* v = *(as + 12) */ 130 lwi r12, r8, 16 /* v = *(as + 16) */ 135 lwi r12, r8, 20 /* v = *(as + 20) */ 140 lwi r12, r8, 24 /* v = *(as + 24) */ 145 lwi r12, r8, 28 /* v = *(as + 28) */ 150 lwi r12, r8, 32 /* v = *(as + 32) */ [all …]
|
/arch/arm/mm/ |
D | abort-lv4t.S | 29 ldr r8, [r4] @ read arm instruction 31 tst r8, #1 << 20 @ L = 1 -> write? 33 and r7, r8, #15 << 24 58 mov r1, r8 62 tst r8, #1 << 21 @ check writeback bit 67 and r6, r8, r7 68 and r9, r8, r7, lsl #1 70 and r9, r8, r7, lsl #2 72 and r9, r8, r7, lsl #3 77 and r9, r8, #15 << 16 @ Extract 'n' from instruction [all …]
|
/arch/sh/lib64/ |
D | udivdi3.S | 35 mulu.l r5,r3,r8 38 shlld r8,r0,r8 40 sub r2,r8,r2 45 shlld r5,r0,r8 49 add r8,r21,r8 62 add r8,r7,r8 65 add r8,r5,r2 72 shlri r25,32,r8 77 mulu.l r5,r8,r5 82 shlri r5,14-1,r8 [all …]
|
/arch/sh/kernel/ |
D | relocate_kernel.S | 32 mov.l r8, @-r15 45 stc sr, r8 46 or r9, r8 47 ldc r8, sr 60 stc sr, r8 61 and r9, r8 62 ldc r8, sr 92 stc sr, r8 93 and r9, r8 94 ldc r8, sr [all …]
|
D | entry-common.S | 84 get_current_thread_info r8, r0 232 get_current_thread_info r8, r0 247 mov r8, r0 249 mov.l 1f, r8 250 add r0, r8 251 mov.l @r8, r8 252 jsr @r8 298 mov.l 1f, r8 299 jsr @r8 307 mov.l 1f, r8 [all …]
|
/arch/ia64/include/asm/ |
D | futex.h | 11 register unsigned long r8 __asm ("r8") = 0; \ 17 : "+r" (r8), "=r" (oldval) \ 20 ret = r8; \ 25 register unsigned long r8 __asm ("r8") = 0; \ 38 : "+r" (r8), "=r" (val), "=&r" (oldval), \ 42 if (unlikely (r8)) \ 45 ret = r8; \ 93 register unsigned long r8 __asm ("r8") = 0; in futex_atomic_cmpxchg_inatomic() 101 : "+r" (r8), "=&r" (prev) in futex_atomic_cmpxchg_inatomic() 106 return r8; in futex_atomic_cmpxchg_inatomic()
|
/arch/powerpc/boot/ |
D | crt0.S | 74 9: lwz r8,0(r12) /* get tag */ 75 cmpwi r8,0 77 cmpwi r8,RELA 81 11: addis r8,r8,(-RELACOUNT)@ha 82 cmpwi r8,RELACOUNT@l 113 lwz r8,p_etext-p_base(r10) 117 cmplw cr0,r9,r8 124 lwz r8,p_end-p_base(r10) 128 cmplw cr0,r9,r8 132 lwz r8,p_pstack-p_base(r10) [all …]
|
/arch/powerpc/crypto/ |
D | aes-spe-keys.S | 32 xor r8,r8,r8; \ 79 LOAD_KEY(r8,r4,12) 83 stw r8,12(r3) 88 mr r14,r8 /* apply LS_BOX to 4th temp */ 95 xor r8,r8,r7 99 stw r8,12(r3) 121 LOAD_KEY(r8,r4,12) 127 stw r8,12(r3) 141 xor r8,r8,r7 142 xor r9,r9,r8 [all …]
|
/arch/x86/kernel/ |
D | relocate_kernel_64.S | 77 movq %r8, %r12 83 movq PTR(PA_CONTROL_PAGE)(%rsi), %r8 100 lea PAGE_SIZE(%r8), %rsp 103 addq $(identity_mapped - relocate_kernel), %r8 104 pushq %r8 201 popq %r8 202 subq $(1b - relocate_kernel), %r8 203 movq CP_PA_SWAP_PAGE(%r8), %r10 204 movq CP_PA_BACKUP_PAGES_MAP(%r8), %rdi 205 movq CP_PA_TABLE_PAGE(%r8), %rax [all …]
|
/arch/csky/abiv1/ |
D | memcpy.S | 39 stw r8, (sp, 0) 43 ldw r8, (r3, 8) 47 stw r8, (r7, 8) 54 ldw r8, (sp, 0) 158 stw r8, (sp, 0) 165 ldw r8, (r3, 8) 179 mov r10, r8 180 GET_AFTER_BITS r8 24 181 or r8, r1 190 stw r8, (r7, 8) [all …]
|
/arch/powerpc/kvm/ |
D | book3s_64_slb.S | 43 li r8, 0 44 stb r8, 3(r11) 120 li r8, SLB_NUM_BOLTED 121 stb r8, 3(r11) 127 li r8, SLBSHADOW_SAVEAREA 131 LDX_BE r10, r11, r8 137 addi r8, r8, SHADOW_SLB_ENTRY_LEN
|
D | book3s_hv_interrupts.S | 68 mfspr r8,SPRN_DEC 74 extsw r8,r8 75 32: mtspr SPRN_HDEC,r8 76 add r8,r8,r7 77 std r8,HSTATE_DECEXP(r13) 114 mfspr r8, SPRN_MMCR2 140 std r8, HSTATE_MMCR2(r13) 147 mfspr r8, SPRN_PMC5 153 stw r8, HSTATE_PMC5(r13)
|
/arch/arm/mach-omap1/ |
D | ams-delta-fiq-handler.S | 102 mov r8, #2 @ reset FIQ agreement 103 str r8, [r12, #IRQ_CONTROL_REG_OFFSET] 108 mov r8, #1 109 orr r8, r11, r8, lsl r10 @ mask spurious interrupt 110 str r8, [r12, #IRQ_MIR_REG_OFFSET] 142 ldr r8, [r12, #OMAP1510_GPIO_DATA_INPUT] @ fetch GPIO input 148 ands r8, r8, #KEYBRD_DATA_MASK @ check start bit - detected? 151 @ r8 contains KEYBRD_DATA_MASK, use it 152 str r8, [r9, #BUF_STATE] @ enter data processing state 167 @ r8 still contains GPIO input bits [all …]
|
/arch/powerpc/kernel/ |
D | misc_32.S | 93 mulhwu r8,r3,r6 95 adde r4,r4,r8 109 lis r8,__got2_end@ha 110 addi r8,r8,__got2_end@l 111 subf r8,r7,r8 112 srwi. r8,r8,2 114 mtctr r8 448 lwz r8,12(r4); \ 452 stw r8,12(r3); \ 525 rlwinm r8,r7,0,32 # t3 = (count < 32) ? 32 : 0 [all …]
|
/arch/x86/crypto/ |
D | aegis128-aesni-asm.S | 78 mov LEN, %r8 79 and $0x1, %r8 82 mov LEN, %r8 83 and $0x1E, %r8 84 add SRC, %r8 85 mov (%r8), %r9b 88 mov LEN, %r8 89 and $0x2, %r8 92 mov LEN, %r8 93 and $0x1C, %r8 [all …]
|