/arch/hexagon/kernel/ |
D | vm_vectors.S | 15 jump 1f; 16 jump 1f; 17 jump 1f; 18 jump 1f; 19 jump 1f; 21 jump 1f; 22 jump 1f; 28 jump 1b; /* Reset */ 29 jump _K_enter_machcheck; 30 jump _K_enter_genex; [all …]
|
D | vm_entry.S | 221 jump event_dispatch; \ 248 jump event_dispatch; \ 285 if (!P0.new) jump:nt restore_all; 307 P0 = cmp.eq(R0, #0); if (!P0.new) jump:nt check_work_pending; 372 if (P0) jump check_work_pending 378 jump check_work_pending
|
D | head.S | 214 jump 1b
|
/arch/hexagon/mm/ |
D | copy_user_template.S | 20 if (!p0.new) jump:nt .Ldone 27 if (!p0.new) jump:nt .Loop_not_aligned_8 33 if (!p1) jump .Lsmall 47 jump .Lsmall 53 if (p0.new) jump:nt .Lalign 57 if (!p0.new) jump:nt .Loop_not_aligned_4 62 if (!p1) jump .Lsmall 77 jump .Lsmall 83 if (!p0.new) jump:nt .Loop_not_aligned 88 if (!p1) jump .Lsmall [all …]
|
/arch/hexagon/lib/ |
D | memset.S | 42 if p1 jump 2f /* skip byte loop */ 60 if !p0 jump 3f /* skip initial byte store */ 73 if !p0 jump 4f /* skip initial half store */ 86 if !p0 jump 5f /* skip initial word store */ 99 if !p0 jump 7f /* skip double loop */ 126 if !p0 jump 8f /* skip final word store */ 137 if !p0 jump 9f /* skip final half store */ 159 if (r2==#0) jump:nt .L1 164 if (p0.new) jump:nt .L3 179 if (!p0.new) jump:nt .L8 [all …]
|
D | memcpy_likely_aligned.S | 16 if (!p0) jump:nt .Lmemcpy_call 22 if (p0) jump:nt .Lmemcpy_call 54 jump memcpy
|
D | memcpy.S | 204 if (p2.new) jump:NT .Ldwordaligned; /* %8 < 97 */ 207 if(!p0) jump .Lbytes23orless; /* %1, <24 */ 275 if(nokernel) jump .Lskip64; 291 if(noprolog) jump .Lnoprolog32; 347 if(nokernel.new) jump:NT .Lepilog; /* likely not taken */ 352 if(p3) jump .Lodd_alignment; 395 jump .Lepilog; 449 if(p3) jump .Lepilog60;
|
/arch/powerpc/kernel/ |
D | module_64.c | 110 u32 jump[7]; member 506 if ((unsigned long)entry->jump % 8 != 0) { in create_ftrace_stub() 511 BUILD_BUG_ON(sizeof(stub_insns) > sizeof(entry->jump)); in create_ftrace_stub() 512 memcpy(entry->jump, stub_insns, sizeof(stub_insns)); in create_ftrace_stub() 523 entry->jump[2] |= IMM_H18(reladdr); in create_ftrace_stub() 524 entry->jump[3] |= IMM_L(reladdr); in create_ftrace_stub() 534 entry->jump[1] |= PPC_HA(reladdr); in create_ftrace_stub() 535 entry->jump[2] |= PPC_LO(reladdr); in create_ftrace_stub() 603 if ((unsigned long)entry->jump % 8 != 0) { in create_stub() 608 BUILD_BUG_ON(sizeof(ppc64_stub_insns) > sizeof(entry->jump)); in create_stub() [all …]
|
D | module_32.c | 149 if (entry->jump[0] != PPC_RAW_LIS(_R12, PPC_HA(val))) in entry_matches() 151 if (entry->jump[1] != PPC_RAW_ADDI(_R12, _R12, PPC_LO(val))) in entry_matches() 172 while (entry->jump[0]) { in do_plt_call() 177 if (patch_instruction(&entry->jump[0], ppc_inst(PPC_RAW_LIS(_R12, PPC_HA(val))))) in do_plt_call() 179 if (patch_instruction(&entry->jump[1], ppc_inst(PPC_RAW_ADDI(_R12, _R12, PPC_LO(val))))) in do_plt_call() 181 if (patch_instruction(&entry->jump[2], ppc_inst(PPC_RAW_MTCTR(_R12)))) in do_plt_call() 183 if (patch_instruction(&entry->jump[3], ppc_inst(PPC_RAW_BCTR()))) in do_plt_call()
|
/arch/xtensa/kernel/ |
D | align.S | 174 _beqi a0, OP0_L32I_N, 1f # L32I.N, jump 197 _beqi a0, OP0_L32I_N, .Lload_w# l32i.n: jump 236 _beqi a5, OP0_L32I_N, .Lload # L32I.N, jump 250 jx a5 # jump into table 274 _beqi a5, OP0_L32I_N, .Lload_w# l32i.n: jump 282 _beqi a5, OP1_L32I, .Lload_w # l32i: jump 305 jx a4 # jump to entry for target register 367 bbsi.l a0, PS_UM_BIT, 2f # jump if user mode 386 _beqz a5, 1f # s32i.n: jump 394 _beqi a5, OP1_S32I, 1f # jump if 32 bit store [all …]
|
/arch/m68k/ifpsp060/ |
D | fskeleton.S | 83 bral trap | jump to trap handler 102 bral trap | jump to trap handler 121 bral trap | jump to trap handler 140 bral trap | jump to trap handler 159 bral trap | jump to trap handler 178 bral trap | jump to trap handler 200 bral trap | jump to trap handler 214 bral trap | jump to trap handler 253 bral trap | jump to trap handler
|
/arch/loongarch/kernel/ |
D | alternative.c | 56 static inline bool in_alt_jump(unsigned long jump, void *start, void *end) in in_alt_jump() argument 58 return jump >= (unsigned long)start && jump < (unsigned long)end; in in_alt_jump()
|
/arch/powerpc/include/asm/ |
D | module.h | 22 unsigned int jump[4]; member
|
/arch/sh/boards/mach-hp6xx/ |
D | pm_wakeup.S | 28 ! jump to handler
|
/arch/alpha/ |
D | Makefile | 16 cflags-y += $(call cc-option, -fno-jump-tables)
|
/arch/x86/realmode/rm/ |
D | trampoline_64.S | 93 # flush prefetch and jump to startup_32 208 # Now jump into the kernel using virtual addresses
|
/arch/arm/mach-mvebu/ |
D | pmsu_ll.S | 63 mov pc, r0 @ jump to this value
|
/arch/x86/kernel/acpi/ |
D | wakeup_32.S | 38 # jump to place where we left off
|
/arch/x86/boot/ |
D | pmjump.S | 31 jmp 1f # Short jump to serialize on 386/486
|
/arch/loongarch/vdso/ |
D | Makefile | 26 -fno-stack-protector -fno-jump-tables -DDISABLE_BRANCH_PROFILING \
|
/arch/x86/ |
D | Makefile | 83 KBUILD_CFLAGS += $(call cc-option,-fcf-protection=branch -fno-jump-tables) 203 KBUILD_CFLAGS += -fno-jump-tables
|
/arch/m68k/fpsp040/ |
D | do_func.S | 7 | The opcode and tag bits form an index into a jump table in 83 leal tblpre,%a1 |start of monster jump table 279 | Alias extended denorms to norms for the jump table. 448 leal pscalet,%a1 |load start of jump table
|
/arch/x86/include/uapi/asm/ |
D | bootparam.h | 76 __u16 jump; member
|
/arch/arm/kernel/ |
D | sleep.S | 118 ret r3 @ jump to virtual address
|
/arch/arm/lib/ |
D | memset.S | 146 b 7b @ jump into the middle of memset
|