/arch/xtensa/include/asm/ |
D | initialize_mmu.h | 103 slli a4, a0, 5 104 srli a4, a4, 5 106 add a4, a4, a5 107 jx a4 112 2: movi a4, 0x20000000 113 add a5, a2, a4 116 add a5, a5, a4 127 movi a4, XCHAL_KSEG_PADDR + CA_WRITEBACK 128 wdtlb a4, a5 129 witlb a4, a5 [all …]
|
/arch/xtensa/lib/ |
D | strnlen_user.S | 37 # a4/ tmp 50 addi a4, a2, -4 # because we overincrement at the end; 69 add a10, a10, a4 # a10 = end of last 4B chunk 72 EX(10f) l32i a9, a4, 4 # get next word of string 73 addi a4, a4, 4 # advance string pointer 79 blt a4, a10, .Loop 83 EX(10f) l32i a9, a4, 4 # load 4 bytes for remaining checks 87 addi a4, a4, 2 # advance string pointer 96 addi a4, a4, 1 # advance string pointer 98 sub a2, a4, a2 # compute length [all …]
|
D | memcopy.S | 68 loopnez a4, .Lbytecopydone 70 beqz a4, .Lbytecopydone 71 add a7, a3, a4 # a7 = end address for source 90 _bltui a4, 7, .Lbytecopy # do short copies byte by byte 95 addi a4, a4, -1 102 _bltui a4, 6, .Lbytecopy # do short copies byte by byte 106 addi a4, a4, -2 116 # a2/ dst, a3/ src, a4/ len 122 srli a7, a4, 4 # number of loop iterations with 16B 152 bbci.l a4, 3, .L2 [all …]
|
D | memset.S | 38 # a2/ dst, a3/ c, a4/ length 48 srli a7, a4, 4 # number of loop iterations with 16B 50 bnez a4, .Laligned 77 bbci.l a4, 3, .L2 83 bbci.l a4, 2, .L3 88 bbci.l a4, 1, .L4 93 bbci.l a4, 0, .L5 105 bltui a4, 8, .Lbyteset # do short copies byte by byte 111 addi a4, a4, -1 119 addi a4, a4, -2 [all …]
|
D | checksum.S | 62 ONES_ADD(a4, a6) 63 ONES_ADD(a4, a7) 66 ONES_ADD(a4, a6) 67 ONES_ADD(a4, a7) 70 ONES_ADD(a4, a6) 71 ONES_ADD(a4, a7) 74 ONES_ADD(a4, a6) 75 ONES_ADD(a4, a7) 91 ONES_ADD(a4, a6) 99 ONES_ADD(a4, a6) [all …]
|
D | strncpy_user.S | 40 # a4/ len 54 # a2/ dst, a3/ src, a4/ len 56 beqz a4, .Lret # if len is zero 64 srli a12, a4, 2 # number of loop iterations with 4B per loop 75 addi a4, a4, -1 # decrement len 76 beqz a4, .Lret # if len is zero 85 addi a4, a4, -1 # decrement len 86 beqz a4, .Lret # if len is zero 92 addi a4, a4, -1 # decrement len 93 bnez a4, .Lsrcaligned # if len is nonzero [all …]
|
D | usercopy.S | 64 # a2/ dst, a3/ src, a4/ len 66 mov a11, a4 # preserve original len for error case 71 srli a7, a4, 4 # number of loop iterations with 16B 76 bnez a4, .Lsrcunaligned 85 bltui a4, 7, .Lbytecopy # do short copies byte by byte 92 addi a4, a4, -1 97 bltui a4, 6, .Lbytecopy # do short copies byte by byte 104 addi a4, a4, -2 115 loopnez a4, .Lbytecopydone 117 beqz a4, .Lbytecopydone [all …]
|
/arch/xtensa/boot/boot-redboot/ |
D | bootstrap.S | 55 movi a4, 1 56 wsr a4, ps 61 sll a4, a4 62 wsr a4, windowstart 65 movi a4, 0x00040000 66 wsr a4, ps 80 # addi a4, a0, __start - __start_a0 81 # mov a0, a4 83 movi a4, __start 85 add a4, a0, a4 [all …]
|
/arch/xtensa/kernel/ |
D | align.S | 165 s32i a4, a2, PT_AREG4 176 movi a4, fast_unaligned_fixup 177 s32i a4, a3, EXC_TABLE_FIXUP 205 l32i a4, a3, 0 # load 2 words 209 __src_b a4, a4, a5 # a4 has the instruction 213 extui a5, a4, INSN_OP0, 4 # get insn.op0 nibble 221 _bbci.l a4, OP1_SI_BIT + INSN_OP1, .Lload 226 extui a6, a4, INSN_T, 4 # get source register 249 extui a5, a4, INSN_OP0, 4 257 extui a5, a4, INSN_OP1, 4 [all …]
|
D | coprocessor.S | 32 xchal_cp##x##_store a2 a3 a4 a5 a6; \ 49 xchal_cp##x##_load a2 a3 a4 a5 a6; \ 132 s32i a4, a1, PT_AREG4 155 l32i a4, a0, 0 157 beqz a4, 1f # skip 'save' if no previous owner 161 l32i a5, a4, THREAD_CPENABLE 163 s32i a5, a4, THREAD_CPENABLE 175 add a2, a2, a4 187 1: GET_THREAD_INFO (a4, a1) 188 s32i a4, a0, 0 [all …]
|
D | mcount.S | 27 movi a4, ftrace_trace_function 28 l32i a4, a4, 0 30 bne a3, a4, 1f 42 callx4 a4
|
D | entry.S | 174 s32i a4, a1, PT_AREG4 227 addi a0, a4, -1 502 _bbsi.l a4, TIF_NEED_RESCHED, 3f 504 bnone a4, a2, 5f 506 2: l32i a4, a1, PT_DEPC 507 bgeui a4, VALID_DOUBLE_EXCEPTION_ADDRESS, 4f 530 _bbci.l a4, TIF_NEED_RESCHED, 4f 534 l32i a4, a2, TI_PRE_COUNT 535 bnez a4, 4f 548 _bbci.l a4, TIF_DB_DISABLED, 7f [all …]
|
D | head.S | 224 l32i a4, a2, 0 # start destination (in RAM) 228 beq a4, a5, 1b # skip, empty entry 229 beq a4, a6, 1b # skip, source and dest. are the same 233 s32i a7, a4, 0 # store word 234 addi a4, a4, 4 235 bltu a4, a5, 2b 246 __loopt a2, a3, a4, 2
|
/arch/riscv/lib/ |
D | tishift.S | 13 sext.w a4,a5 17 sll a4,a1,a4 19 or a0,a0,a4 24 negw a0,a4 36 sext.w a4,a5 40 sll a4,a1,a4 42 or a0,a0,a4 47 negw a0,a4 59 sext.w a4,a5 63 srl a4,a0,a4 [all …]
|
D | memcpy.S | 19 andi a4, a1, SZREG-1 20 bne a3, a4, 4f 30 sub a4, a3, a1 37 sub a2, a2, a4 /* Update count */ 40 andi a4, a2, ~((16*SZREG)-1) 41 beqz a4, 4f 42 add a3, a1, a4 44 REG_L a4, 0(a1) 54 REG_S a4, 0(t6) 64 REG_L a4, 10*SZREG(a1) [all …]
|
D | memset.S | 27 sub a4, a3, t0 32 sub a2, a2, a4 /* Update count */ 47 andi a4, a2, ~(SZREG-1) 48 add a3, t0, a4 50 andi a4, a4, 31*SZREG /* Calculate remainder */ 51 beqz a4, 3f /* Shortcut if no remainder */ 52 neg a4, a4 53 addi a4, a4, 32*SZREG /* Calculate initial offset */ 56 sub t0, t0, a4 62 srli a4, a4, 1 [all …]
|
D | memmove.S | 55 add a4, a1, a2 161 andi a5, a4, (SZREG - 1) /* Find the alignment offset of src (a4) */ 163 sub a5, a4, t4 /* Find the difference between src and dest */ 164 andi a4, a4, -SZREG /* Align the src pointer */ 195 REG_L t1, ( 0 * SZREG)(a4) 197 REG_L t0, (-1 * SZREG)(a4) 206 REG_L t1, (-2 * SZREG)(a4) 207 addi a4, a4, (-2 * SZREG) 217 add a4, t4, a5 /* Restore the src pointer */ 244 REG_L t1, (-1 * SZREG)(a4) [all …]
|
/arch/xtensa/mm/ |
D | misc.S | 62 __loopi a2, a4, PAGE_SIZE, 32 87 __endl a2, a4 123 mov a4, a2 141 idtlb a4 161 addi a6, a4, (PAGE_KERNEL | _PAGE_HW_WRITE) 173 1: __loopi a2, a4, PAGE_SIZE, 32 198 __endl a2, a4 229 mov a4, a2 235 idtlb a4 253 mov a4, a2 [all …]
|
/arch/mips/kernel/ |
D | linux32.c | 77 unsigned long, unused, unsigned long, a4, unsigned long, a5) 79 return ksys_pread64(fd, buf, count, merge_64(a4, a5)); 83 size_t, count, u32, unused, u64, a4, u64, a5) 85 return ksys_pwrite64(fd, buf, count, merge_64(a4, a5)); 110 unsigned long a4, unsigned long a5, in sys32_sync_file_range() argument 114 merge_64(a2, a3), merge_64(a4, a5), in sys32_sync_file_range() 120 unsigned long a4, unsigned long a5, in sys32_fadvise64_64() argument 124 merge_64(a2, a3), merge_64(a4, a5), in sys32_fadvise64_64()
|
/arch/m68k/kernel/ |
D | relocate_kernel.S | 26 lea %pc@(.Lcopy),%a4 27 2: addl #0x00000000,%a4 /* virt_to_phys() */ 37 jmp %a4@ /* Jump to physical .Lcopy */ 49 lea %pc@(.Lcont040),%a4 50 5: addl #0x00000000,%a4 /* virt_to_phys() */ 56 movel %a4,%d0 63 jmp %a4@ /* Jump to physical .Lcont040 */
|
/arch/riscv/kernel/ |
D | entry.S | 186 REG_L a4, PT_A4(sp) 377 REG_L a4, PT_A4(sp) 528 li a4, TASK_THREAD_RA 529 add a3, a0, a4 530 add a4, a1, a4 546 REG_L ra, TASK_THREAD_RA_RA(a4) 547 REG_L sp, TASK_THREAD_SP_RA(a4) 548 REG_L s0, TASK_THREAD_S0_RA(a4) 549 REG_L s1, TASK_THREAD_S1_RA(a4) 550 REG_L s2, TASK_THREAD_S2_RA(a4) [all …]
|
/arch/powerpc/platforms/powernv/ |
D | opal-call.c | 17 s64 a4, s64 a5, s64 a6, s64 a7, in __trace_opal_entry() argument 32 args[4] = a4; in __trace_opal_entry() 70 s64 a4, s64 a5, s64 a6, s64 a7, in __opal_call_trace() argument 75 __trace_opal_entry(a0, a1, a2, a3, a4, a5, a6, a7, opcode); in __opal_call_trace() 76 ret = __opal_call(a0, a1, a2, a3, a4, a5, a6, a7, opcode, msr); in __opal_call_trace() 87 s64 a4, s64 a5, s64 a6, s64 a7, in __opal_call_trace() argument 97 int64_t a4, int64_t a5, int64_t a6, int64_t a7, int64_t opcode) in opal_call() argument 110 return __opal_call(a0, a1, a2, a3, a4, a5, a6, a7, opcode, msr); in opal_call() 116 ret = __opal_call_trace(a0, a1, a2, a3, a4, a5, a6, a7, opcode, msr); in opal_call() 118 ret = __opal_call(a0, a1, a2, a3, a4, a5, a6, a7, opcode, msr); in opal_call() [all …]
|
/arch/x86/include/asm/xen/ |
D | hypercall.h | 133 #define __HYPERCALL_4ARG(a1,a2,a3,a4) \ argument 134 __HYPERCALL_3ARG(a1,a2,a3) __arg4 = (unsigned long)(a4); 135 #define __HYPERCALL_5ARG(a1,a2,a3,a4,a5) \ argument 136 __HYPERCALL_4ARG(a1,a2,a3,a4) __arg5 = (unsigned long)(a5); 189 #define _hypercall4(type, name, a1, a2, a3, a4) \ argument 192 __HYPERCALL_4ARG(a1, a2, a3, a4); \ 203 unsigned long a3, unsigned long a4, in xen_single_call() argument 207 __HYPERCALL_5ARG(a1, a2, a3, a4, a5); in xen_single_call() 239 unsigned long a3, unsigned long a4, in privcmd_call() argument 245 res = xen_single_call(call, a1, a2, a3, a4, a5); in privcmd_call()
|
/arch/m68k/ifpsp060/src/ |
D | itest.S | 2909 lea EAMEM+0x4(%a6),%a4 2915 mulu.l -(%a4),%d2:%d3 3206 mov.l &-0x2,%a4 3212 mulu.l (0x10.b,%a0,%a4.l*8),%d2:%d3 3230 mov.l &0x2,%a4 3236 mulu.l (-0x10.b,%a0,%a4.l*8),%d2:%d3 3325 lea EAMEM(%a6),%a4 3332 mulu.l (0x10.b,%a4,%d4.w*1),%d2:%d3 3496 lea EAMEM(%a6),%a4 3502 mulu.l (%a4),%d2:%d3 [all …]
|
/arch/arm/include/asm/ |
D | xor.h | 21 : "=r" (dst), "=r" (a1), "=r" (a2), "=r" (a3), "=r" (a4) \ 34 __XOR(a1, b1); __XOR(a2, b2); __XOR(a3, b3); __XOR(a4, b4) 44 : "0" (dst), "r" (a1), "r" (a2), "r" (a3), "r" (a4)) 53 register unsigned int a4 __asm__("r7"); in xor_arm4regs_2() 74 register unsigned int a4 __asm__("r7"); in xor_arm4regs_3()
|