/arch/arm/include/asm/ |
D | assembler.h | 27 #define push lsl 28 #define get_byte_0 lsl #0 32 #define put_byte_0 lsl #0 33 #define put_byte_1 lsl #8 34 #define put_byte_2 lsl #16 35 #define put_byte_3 lsl #24 37 #define pull lsl 42 #define get_byte_3 lsl #0 43 #define put_byte_0 lsl #24 44 #define put_byte_1 lsl #16 [all …]
|
/arch/arm/lib/ |
D | io-writesw-armv3.S | 29 orr r3, r3, r3, lsl #16 47 mov ip, r3, lsl #16 52 orr ip, ip, ip, lsl #16 55 mov ip, r4, lsl #16 60 orr ip, ip, ip, lsl #16 63 mov ip, r5, lsl #16 68 orr ip, ip, ip, lsl #16 71 mov ip, r6, lsl #16 76 orr ip, ip, ip, lsl #16 90 mov ip, r3, lsl #16 [all …]
|
D | io-readsw-armv3.S | 40 orr ip, ip, ip, lsl #8 49 orr r3, r3, r4, lsl #16 54 orr r4, r4, r5, lsl #16 59 orr r5, r5, r6, lsl #16 64 orr r6, r6, lr, lsl #16 80 orr r3, r3, r4, lsl #16 85 orr r4, r4, r5, lsl #16 95 orr r3, r3, r4, lsl #16
|
D | io-readsw-armv4.S | 15 orr \rd, \hw1, \hw2, lsl #16 17 orr \rd, \hw2, \hw1, lsl #16 21 .Linsw_align: movs ip, r1, lsl #31 71 .Lno_insw_4: movs r2, r2, lsl #31 88 #define pull_hbyte1 lsl #24 92 #define push_hbyte0 lsl #24 109 _BE_ONLY_( mov ip, ip, lsl #24 ) 114 orr ip, ip, r3, lsl #8
|
D | muldi3.S | 34 bic xl, xl, ip, lsl #16 35 bic yl, yl, yh, lsl #16 40 adds xl, xl, yh, lsl #16 42 adds xl, xl, ip, lsl #16
|
D | ashldi3.S | 44 movmi ah, ah, lsl r2 45 movpl ah, al, lsl r3 47 mov al, al, lsl r2
|
D | bitops.h | 7 mov r3, r2, lsl r3 20 mov r3, r2, lsl r3 @ create mask 35 mov r3, r3, lsl r2 58 tst r2, r0, lsl r3 59 \instr r2, r2, r0, lsl r3
|
D | div64.S | 69 mov ip, ip, lsl yl 70 mov yl, r4, lsl yl 78 movcc yl, yl, lsl #1 79 movcc ip, ip, lsl #1 104 4: movs xl, xl, lsl #1 127 mov xl, xl, lsl xh 132 7: movs xl, xl, lsl #1 180 orr yl, yl, xh, lsl ip 181 mov xh, xl, lsl ip
|
D | io-writesw-armv4.S | 25 .Loutsw_align: movs ip, r1, lsl #31 58 .Lno_outsw_4: movs r2, r2, lsl #31 70 #define pull_hbyte0 lsl #8 74 #define push_hbyte1 lsl #8
|
D | lib1funcs.S | 48 mov \divisor, \divisor, lsl \result 49 mov \curbit, \curbit, lsl \result 59 moveq \divisor, \divisor, lsl #3 69 movlo \divisor, \divisor, lsl #4 70 movlo \curbit, \curbit, lsl #4 77 movlo \divisor, \divisor, lsl #1 78 movlo \curbit, \curbit, lsl #1 144 mov \divisor, \divisor, lsl \order 156 movlo \divisor, \divisor, lsl #4 164 movlo \divisor, \divisor, lsl #1
|
D | getuser.S | 42 orr r2, r2, r3, lsl #8 44 orr r2, r3, r2, lsl #8
|
D | csumpartial.S | 61 orr td0, td0, td3, lsl #8 63 orr td0, td3, td0, lsl #8 94 orrne td0, td0, ip, lsl #8 96 orrne td0, ip, td0, lsl #8
|
/arch/arm/mach-ep93xx/include/mach/ |
D | entry-macro.S | 38 movs \tmp, \irqstat, lsl #16 42 movs \tmp, \irqstat, lsl #8 46 movs \tmp, \irqstat, lsl #4 50 movs \tmp, \irqstat, lsl #2 54 movs \tmp, \irqstat, lsl #1
|
/arch/avr32/lib/ |
D | __avr32_lsl64.S | 23 lsl r10, r10, r12 24 lsl r11, r11, r12 29 lsl r11, r10, r9
|
D | csum_partial.S | 39 lsl r9, 16 41 lsl r8, 8
|
D | findbit.S | 37 lsl r8, 2 101 lsl r8, 2 131 lsl r8, 2 161 lsl r8, 2
|
D | memcpy.S | 40 lsl r10, 2 54 lsl r9, 2
|
/arch/arm/boot/compressed/ |
D | ll_char_wr.S | 43 mov r1, r1, lsl #3 59 add r0, r0, r5, lsl #3 @ Move to bottom of character 68 ldr r7, [lr, r7, lsl #2] 73 ldr r7, [lr, r7, lsl #2] 87 ldr ip, [lr, ip, lsl #2] 90 ldr ip, [lr, ip, lsl #2] @ avoid r4 97 ldr ip, [lr, ip, lsl #2] 100 ldr ip, [lr, ip, lsl #2] @ avoid r4
|
/arch/arm/mm/ |
D | abort-lv4t.S | 67 and r2, r8, r7, lsl #1 69 and r2, r8, r7, lsl #2 71 and r2, r8, r7, lsl #3 79 subne r7, r7, r6, lsl #2 @ Undo increment 80 addeq r7, r7, r6, lsl #2 @ Undo decrement 92 ldreq r6, [sp, r5, lsl #2] @ { load Rm value } 106 movs r2, r8, lsl #20 @ Get offset 121 ldr r6, [sp, r7, lsl #2] @ Get register 'Rm' 129 mov r6, r6, lsl r5 @ 0: LSL #!0 207 addeq r7, r7, r6, lsl #2 @ increment SP if PUSH [all …]
|
D | tlb-v7.S | 40 orr r0, r3, r0, lsl #PAGE_SHIFT @ Create initial MVA 41 mov r1, r1, lsl #PAGE_SHIFT 68 mov r0, r0, lsl #PAGE_SHIFT 69 mov r1, r1, lsl #PAGE_SHIFT
|
D | tlb-v6.S | 42 orr r0, r3, r0, lsl #PAGE_SHIFT @ Create initial MVA 43 mov r1, r1, lsl #PAGE_SHIFT 73 mov r0, r0, lsl #PAGE_SHIFT 74 mov r1, r1, lsl #PAGE_SHIFT
|
/arch/m68k/math-emu/ |
D | fp_movem.S | 157 3: lsl.b #1,%d1 164 lsl.w #1,%d2 165 lsl.l #7,%d2 166 lsl.l #8,%d2 176 4: lsl.b #1,%d1 307 lsl.l #5,%d1 312 3: lsl.b #1,%d1 320 lsl.l #5,%d1 331 3: lsl.b #1,%d1 341 4: lsl.b #1,%d1
|
D | fp_util.S | 131 lsl.l #8,%d0 | shift mantissa 168 lsl.l #8,%d0 | shift high mantissa 169 lsl.l #3,%d0 182 lsl.l #8,%d0 183 lsl.l #3,%d0
|
/arch/blackfin/mm/ |
D | sram-alloc.c | 718 struct sram_list_struct *lsl, **tmp; in sram_free_with_lsl() local 726 lsl = *tmp; in sram_free_with_lsl() 728 *tmp = lsl->next; in sram_free_with_lsl() 729 kfree(lsl); in sram_free_with_lsl() 738 struct sram_list_struct *lsl = NULL; in sram_alloc_with_lsl() local 741 lsl = kzalloc(sizeof(struct sram_list_struct), GFP_KERNEL); in sram_alloc_with_lsl() 742 if (!lsl) in sram_alloc_with_lsl() 758 kfree(lsl); in sram_alloc_with_lsl() 761 lsl->addr = addr; in sram_alloc_with_lsl() 762 lsl->length = size; in sram_alloc_with_lsl() [all …]
|
/arch/arm/mach-clps711x/include/mach/ |
D | entry-macro.S | 31 mov \mask, \mask, lsl #16 40 mov \mask, \mask, lsl #16
|