/arch/microblaze/lib/ |
D | fastcopy.S | 79 andi r9, r6, 3 /* t1 = s & 3 */ 81 bnei r9, a_block_unaligned 84 lwi r9, r6, 0 /* t1 = *(s + 0) */ 88 swi r9, r5, 0 /* *(d + 0) = t1 */ 92 lwi r9, r6, 16 /* t1 = *(s + 16) */ 96 swi r9, r5, 16 /* *(d + 16) = t1 */ 111 addi r9, r9, -1 112 beqi r9, a_block_u1 /* t1 was 1 => 1 byte offset */ 113 addi r9, r9, -1 114 beqi r9, a_block_u2 /* t1 was 2 => 2 byte offset */ [all …]
|
/arch/arm/crypto/ |
D | sha1-armv4-large.S | 73 ldrb r9,[r1,#3] 77 orr r9,r9,r10,lsl#8 79 orr r9,r9,r11,lsl#16 81 orr r9,r9,r12,lsl#24 83 ldr r9,[r1],#4 @ handles unaligned 88 rev r9,r9 @ byte swap 92 add r7,r7,r9 @ E+=X[i] 94 str r9,[r14,#-4]! 98 ldrb r9,[r1,#3] 102 orr r9,r9,r10,lsl#8 [all …]
|
/arch/powerpc/mm/ |
D | slb_low.S | 36 srdi r9,r3,60 /* get region */ 38 cmpldi cr7,r9,0xc /* cmp PAGE_OFFSET for later use */ 67 cmpldi cr0,r9,0xf 96 srdi. r9,r10,USER_ESID_BITS 114 ld r9,PACALOWSLICESPSIZE(r13) 117 ld r9,PACAHIGHSLICEPSIZE(r13) 122 srd r9,r9,r11 123 andi. r9,r9,0xf 124 mulli r9,r9,MMUPSIZEDEFSIZE 130 add r11,r11,r9 [all …]
|
/arch/powerpc/boot/ |
D | crt0.S | 56 li r9,0 63 lwz r9,4(r12) /* get RELA pointer in r9 */ 80 cmpwi r9,0 83 add r9,r9,r11 /* Relocate RELA pointer */ 85 2: lbz r0,4+3(r9) /* ELF32_R_INFO(reloc->r_info) */ 88 lwz r12,0(r9) /* reloc->r_offset */ 89 lwz r0,8(r9) /* reloc->r_addend */ 92 addi r9,r9,12 96 3: lwz r9,p_start-p_base(r10) /* note: these are relocated now */ 98 4: dcbf r0,r9 [all …]
|
D | virtex405-head.S | 21 mfdccr r9 22 cmplwi r9,0 24 lis r9,0 27 1: dccci r0,r9 28 addi r9,r9,0x20
|
/arch/powerpc/kvm/ |
D | book3s_32_sr.S | 41 #define XCHG_SR(n) lwz r9, (SVCPU_SR+(n*4))(r3); \ 42 mtsr n, r9 69 li r9, 0 70 KVM_KILL_BAT(0, r9) 71 KVM_KILL_BAT(1, r9) 72 KVM_KILL_BAT(2, r9) 73 KVM_KILL_BAT(3, r9) 113 lis r9, BATS@ha 114 addi r9, r9, BATS@l 115 tophys(r9, r9) [all …]
|
D | book3s_hv_rmhandlers.S | 162 lwz r9, VCPU_PMC + 20(r4) 172 mtspr SPRN_PMC6, r9 261 addi r9,r5,VCORE_ENTRY_EXIT 262 21: lwarx r3,0,r9 266 stwcx. r3,0,r9 270 ld r9,VCPU_KVM(r4) /* pointer to struct kvm */ 274 ld r6,KVM_SDR1(r9) 275 lwz r7,KVM_LPID(r9) 292 10: ld r8,KVM_LPCR(r9) 294 ld r8,KVM_RMOR(r9) [all …]
|
/arch/powerpc/lib/ |
D | memcpy_64.S | 39 ld r9,0(r4) 46 mr r8,r9 48 1: ld r9,8(r4) 51 stdu r9,16(r3) 58 lwz r9,8(r4) 60 stw r9,0(r3) 63 lhz r9,8(r4) 65 sth r9,0(r3) 68 lbz r9,8(r4) 69 stb r9,0(r3) [all …]
|
D | copypage_64.S | 26 li r9,0 31 dcbt r9,r4 32 dcbz r9,r3 33 add r9,r9,r12 48 ld r9,8(r4) 54 std r9,40(r3) 64 ld r9,72(r4) 70 std r9,104(r3) 82 ld r9,8(r4) 88 std r9,40(r3) [all …]
|
D | copyuser_64.S | 64 mr r9,r7 70 70: std r9,0(r3) 72 22: ld r9,0(r4) 78 72: std r9,0(r3) 86 244: ld r9,0(r4) 88 245: std r9,0(r3) 91 23: lwz r9,0(r4) 93 73: stw r9,0(r3) 96 44: lhz r9,0(r4) 98 74: sth r9,0(r3) [all …]
|
/arch/avr32/lib/ |
D | findbit.S | 18 mov r9, r11 23 sub r9, 32 34 sub r9, r11, r10 50 add r9, r10 51 sub r9, 32 59 sub r9, 32 67 rsub r9, r11 68 add r10, r9 82 mov r9, r11 87 sub r9, 32 [all …]
|
D | memcpy.S | 22 mov r9, r11 23 andl r9, 3, COH 27 2: mov r9, r12 37 reteq r9 48 retal r9 52 movlt r9, r12 54 add r10, r9 55 lsl r9, 2 56 add pc, pc, r9 64 add pc, pc, r9 [all …]
|
D | io-writesb.S | 15 tst r11, r9 24 mov r9, 3 27 tst r11, r9 34 bfextu r9, r8, 24, 8 35 st.b r12[0], r9 36 bfextu r9, r8, 16, 8 37 st.b r12[0], r9 38 bfextu r9, r8, 8, 8 39 st.b r12[0], r9
|
D | csum_partial_copy_generic.S | 45 add r9, r5 46 acr r9 54 6: mov r12, r9 69 add r9, r5 70 acr r9 77 mov r9, -EFAULT 80 st.w r8[0], r9 87 mov r9, 0 91 mov r9, -EFAULT 95 st.w r8[0], r9 [all …]
|
D | csum_partial.S | 21 4: ld.w r9, r12++ 22 add r10, r9 32 mov r9, 0 36 ld.uh r9, r12++ 39 lsl r9, 16 42 7: or r9, r8 43 add r10, r9
|
/arch/arm/mm/ |
D | abort-lv4t.S | 62 and r9, r8, r7, lsl #1 63 add r6, r6, r9, lsr #1 64 and r9, r8, r7, lsl #2 65 add r6, r6, r9, lsr #2 66 and r9, r8, r7, lsl #3 67 add r6, r6, r9, lsr #3 71 and r9, r8, #15 << 16 @ Extract 'n' from instruction 72 ldr r7, [r2, r9, lsr #14] @ Get register 'Rn' 76 str r7, [r2, r9, lsr #14] @ Put register 'Rn' 83 and r9, r8, #0x00f @ get Rm / low nibble of immediate value [all …]
|
/arch/powerpc/kernel/vdso32/ |
D | gettimeofday.S | 43 mr r9, r3 /* datapage ptr in r9 */ 54 lwz r4,CFG_TZ_MINUTEWEST(r9)/* fill tz */ 55 lwz r5,CFG_TZ_DSTTIME(r9) 84 mr r9,r3 /* datapage ptr in r9 */ 101 lwz r5,WTOM_CLOCK_SEC(r9) 102 lwz r6,WTOM_CLOCK_NSEC(r9) 109 add r9,r9,r0 110 lwz r0,(CFG_TB_UPDATE_COUNT+LOPART)(r9) 198 1: lwz r8,(CFG_TB_UPDATE_COUNT+LOPART)(r9) 202 add r9,r9,r0 [all …]
|
/arch/score/lib/ |
D | checksum.S | 36 lw r9, [src, offset + 0x04]; \ 40 ADDC(sum, r9); \ 44 lw r9, [src, offset + 0x14]; \ 48 ADDC(sum, r9); \ 60 ldi r9, 0x0 69 lbu r9, [src] 70 slli r9,r9, 0x8 /*Little endian*/ 71 ADDC(sum, r9) 83 lhu r9, [src] 85 ADDC(sum, r9) [all …]
|
/arch/powerpc/kernel/ |
D | entry_32.S | 136 stw r9,_MSR(r11) 137 andi. r2,r9,MSR_PR 161 rlwinm r9,r1,0,0,(31-THREAD_SHIFT) 162 lwz r9,TI_CPU(r9) 163 slwi r9,r9,3 164 add r11,r11,r9 177 lwz r9,KSP_LIMIT(r12) 178 cmplw r1,r9 /* if r1 <= ksp_limit */ 182 rlwinm r9,r1,0,0,31-THREAD_SHIFT 183 tophys(r9,r9) /* check local flags */ [all …]
|
D | idle_6xx.S | 138 rlwinm r9,r1,0,0,31-THREAD_SHIFT /* current thread_info */ 139 lwz r8,TI_LOCAL_FLAGS(r9) /* set napping bit */ 141 stw r8,TI_LOCAL_FLAGS(r9) /* it will return to our caller */ 157 lwz r9,_LINK(r11) /* interrupted in ppc6xx_idle: */ 158 stw r9,_NIP(r11) /* make it do a blr */ 171 mfspr r9,SPRN_HID0 172 andis. r9,r9,HID0_NAP@h 174 addis r9,r11,(nap_save_msscr0-KERNELBASE)@ha 175 lwz r9,nap_save_msscr0@l(r9) 176 mtspr SPRN_MSSCR0, r9 [all …]
|
D | head_40x.S | 122 stw r9,GPR9(r11); \ 132 mfspr r9,SPRN_SRR1; \ 134 rlwinm r9,r9,0,14,12; /* clear MSR_WE (necessary?) */\ 164 stw r9,GPR9(r11); \ 169 mfspr r9,SPRN_ESR; /* in them at the point where the */\ 170 stw r9,_ESR(r11); /* exception was taken */\ 173 mfspr r9,SPRN_SRR3; \ 176 rlwinm r9,r9,0,14,12; /* clear MSR_WE (necessary?) */\ 220 copyee(r10, r9); \ 267 stw r9, 4(r0) [all …]
|
D | reloc_32.S | 43 lwz r9, (p_rela - 0b)(r12) 44 add r9, r9, r12 /* runtime address of .rela.dyn section */ 93 subf r7, r7, r9 /* cur_offset */ 109 lwz r4, 4(r9) /* r4 = rela.r_info */ 136 lwz r4, 0(r9) /* r_offset */ 137 lwz r0, 8(r9) /* r_addend */ 146 lwz r4, 0(r9) /* r_offset */ 147 lwz r0, 8(r9) /* r_addend */ 157 lwz r4, 0(r9) /* r_offset */ 158 lwz r0, 8(r9) /* r_addend */ [all …]
|
D | fsl_booke_entry_mapping.S | 45 mfspr r9,SPRN_TLB1CFG 46 andi. r9,r9,0xfff 60 cmpw r6,r9 /* Are we done? */ 95 subfic r9,r6,-4096 96 and r9,r9,r7 98 or r25,r8,r9 120 1: mflr r9 121 rlwimi r7,r9,0,20,31 151 li r9,0x0c 152 tlbivax 0,r9 [all …]
|
/arch/arm/mach-s3c24xx/ |
D | sleep-s3c2412.S | 60 ldrne r9, [r1] 61 strne r9, [r1] 62 ldrne r9, [r2] 63 strne r9, [r2] 64 ldrne r9, [r3] 65 strne r9, [r3]
|
/arch/cris/arch-v32/lib/ |
D | checksumcopy.S | 31 _mloop: movem [$r10+],$r9 ; read 10 longwords 33 movem $r9,[$r11+] ; write 10 longwords 45 addc $r9,$r13 61 ;; r9 can be used as temporary. 62 move.d $r13,$r9 63 lsrq 16,$r9 ; r0 = checksum >> 16 68 add.d $r9,$r13 ; checksum += r0 72 _wloop: move.w [$r10+],$r9 73 addu.w $r9,$r13 76 move.w $r9,[$r11+] [all …]
|