| /kernel/linux/linux-5.10/arch/powerpc/platforms/ps3/ |
| D | hvcall.S | 19 std r0, 16(r1); \ 24 ld r0, 16(r1); \ 41 std r0, 16(r1); \ 43 stdu r3, -8(r1); \ 48 addi r1, r1, 8; \ 49 ld r11, -8(r1); \ 52 ld r0, 16(r1); \ 60 std r0, 16(r1); \ 62 std r3, -8(r1); \ 63 stdu r4, -16(r1); \ [all …]
|
| /kernel/linux/linux-6.6/arch/powerpc/platforms/ps3/ |
| D | hvcall.S | 19 std r0, 16(r1); \ 24 ld r0, 16(r1); \ 41 std r0, 16(r1); \ 43 stdu r3, -8(r1); \ 48 addi r1, r1, 8; \ 49 ld r11, -8(r1); \ 52 ld r0, 16(r1); \ 60 std r0, 16(r1); \ 62 std r3, -8(r1); \ 63 stdu r4, -16(r1); \ [all …]
|
| /kernel/linux/linux-5.10/arch/microblaze/kernel/ |
| D | entry-nommu.S | 54 swi r1, r0, PER_CPU(ENTRY_SP) /* save the current sp */ 60 addik r1, r1, (-PT_SIZE) /* room for pt_regs (delay slot) */ 62 lwi r1, r0, PER_CPU(CURRENT_SAVE) /* get the saved current */ 63 lwi r1, r1, TS_THREAD_INFO /* get the thread info */ 65 addik r1, r1, THREAD_SIZE - PT_SIZE 67 swi r11, r1, PT_MODE /* store the mode */ 69 swi r2, r1, PT_R2 70 swi r3, r1, PT_R3 71 swi r4, r1, PT_R4 72 swi r5, r1, PT_R5 [all …]
|
| D | mcount.S | 18 addik r1, r1, -120; \ 19 swi r2, r1, 4; \ 20 swi r3, r1, 8; \ 21 swi r4, r1, 12; \ 22 swi r5, r1, 116; \ 23 swi r6, r1, 16; \ 24 swi r7, r1, 20; \ 25 swi r8, r1, 24; \ 26 swi r9, r1, 28; \ 27 swi r10, r1, 32; \ [all …]
|
| D | entry.S | 178 swi r2, r1, PT_R2; /* Save SDA */ \ 179 swi r3, r1, PT_R3; \ 180 swi r4, r1, PT_R4; \ 181 swi r5, r1, PT_R5; \ 182 swi r6, r1, PT_R6; \ 183 swi r7, r1, PT_R7; \ 184 swi r8, r1, PT_R8; \ 185 swi r9, r1, PT_R9; \ 186 swi r10, r1, PT_R10; \ 187 swi r11, r1, PT_R11; /* save clobbered regs after rval */\ [all …]
|
| /kernel/linux/linux-6.6/tools/testing/selftests/bpf/progs/ |
| D | verifier_xdp_direct_packet_access.c | 14 r2 = *(u32*)(r1 + %[xdp_md_data]); \ in end_mangling_bad_access_1() 15 r3 = *(u32*)(r1 + %[xdp_md_data_end]); \ in end_mangling_bad_access_1() 16 r1 = r2; \ in end_mangling_bad_access_1() 17 r1 += 8; \ in end_mangling_bad_access_1() 19 if r1 > r3 goto l0_%=; \ in end_mangling_bad_access_1() 20 r0 = *(u64*)(r1 - 8); \ in end_mangling_bad_access_1() 35 r2 = *(u32*)(r1 + %[xdp_md_data]); \ in end_mangling_bad_access_2() 36 r3 = *(u32*)(r1 + %[xdp_md_data_end]); \ in end_mangling_bad_access_2() 37 r1 = r2; \ in end_mangling_bad_access_2() 38 r1 += 8; \ in end_mangling_bad_access_2() [all …]
|
| D | verifier_value_ptr_arith.c | 45 __msg_unpriv("R1 tried to add from different maps, paths or scalars") 50 r0 = *(u32*)(r1 + %[__sk_buff_len]); \ in value_ptr_unknown_vs_const() 51 r1 = 0; \ in value_ptr_unknown_vs_const() 52 *(u64*)(r10 - 8) = r1; \ in value_ptr_unknown_vs_const() 56 r1 = %[map_hash_16b] ll; \ in value_ptr_unknown_vs_const() 58 l0_%=: r1 = %[map_array_48b] ll; \ in value_ptr_unknown_vs_const() 63 r1 = 6; \ in value_ptr_unknown_vs_const() 64 r1 = -r1; \ in value_ptr_unknown_vs_const() 65 r1 &= 0x7; \ in value_ptr_unknown_vs_const() 67 l3_%=: r1 = 3; \ in value_ptr_unknown_vs_const() [all …]
|
| D | verifier_stack_ptr.c | 29 r1 = r10; \ in ptr_to_stack_store_load() 30 r1 += -10; \ in ptr_to_stack_store_load() 32 *(u64*)(r1 + 2) = r0; \ in ptr_to_stack_store_load() 33 r0 = *(u64*)(r1 + 2); \ in ptr_to_stack_store_load() 45 r1 = r10; \ in load_bad_alignment_on_off() 46 r1 += -8; \ in load_bad_alignment_on_off() 48 *(u64*)(r1 + 2) = r0; \ in load_bad_alignment_on_off() 49 r0 = *(u64*)(r1 + 2); \ in load_bad_alignment_on_off() 61 r1 = r10; \ in load_bad_alignment_on_reg() 62 r1 += -10; \ in load_bad_alignment_on_reg() [all …]
|
| D | verifier_bounds_mix_sign_unsign.c | 24 r1 = 0; \ in signed_and_unsigned_positive_bounds() 25 *(u64*)(r10 - 8) = r1; \ in signed_and_unsigned_positive_bounds() 28 r1 = %[map_hash_8b] ll; \ in signed_and_unsigned_positive_bounds() 31 r1 = *(u64*)(r10 - 16); \ in signed_and_unsigned_positive_bounds() 33 if r2 >= r1 goto l0_%=; \ in signed_and_unsigned_positive_bounds() 34 if r1 s> 4 goto l0_%=; \ in signed_and_unsigned_positive_bounds() 35 r0 += r1; \ in signed_and_unsigned_positive_bounds() 36 r1 = 0; \ in signed_and_unsigned_positive_bounds() 37 *(u8*)(r0 + 0) = r1; \ in signed_and_unsigned_positive_bounds() 56 r1 = 0; \ in checks_mixing_signed_and_unsigned() [all …]
|
| D | verifier_bounds.c | 22 r1 = 0; \ in bounds_map_value_variant_1() 23 *(u64*)(r10 - 8) = r1; \ in bounds_map_value_variant_1() 26 r1 = %[map_hash_8b] ll; \ in bounds_map_value_variant_1() 29 r1 = *(u8*)(r0 + 0); \ in bounds_map_value_variant_1() 30 if r1 > 0xff goto l0_%=; \ in bounds_map_value_variant_1() 33 r1 -= r3; \ in bounds_map_value_variant_1() 34 r1 >>= 56; \ in bounds_map_value_variant_1() 35 r0 += r1; \ in bounds_map_value_variant_1() 50 __msg_unpriv("R1 has unknown scalar with mixed signed bounds") 54 r1 = 0; \ in bounds_map_value_variant_2() [all …]
|
| D | verifier_masking.c | 16 r2 -= r1; \ in test_out_of_bounds_1() 17 r2 |= r1; \ in test_out_of_bounds_1() 20 r1 &= r2; \ in test_out_of_bounds_1() 21 r0 = r1; \ in test_out_of_bounds_1() 36 r2 -= r1; \ in test_out_of_bounds_2() 37 r2 |= r1; \ in test_out_of_bounds_2() 40 r1 &= r2; \ in test_out_of_bounds_2() 41 r0 = r1; \ in test_out_of_bounds_2() 56 r2 -= r1; \ in test_out_of_bounds_3() 57 r2 |= r1; \ in test_out_of_bounds_3() [all …]
|
| /kernel/linux/linux-5.10/drivers/memory/ |
| D | ti-emif-sram-pm.S | 57 ldr r1, [r0, #EMIF_SDRAM_CONFIG] 58 str r1, [r2, #EMIF_SDCFG_VAL_OFFSET] 60 ldr r1, [r0, #EMIF_SDRAM_REFRESH_CONTROL] 61 str r1, [r2, #EMIF_REF_CTRL_VAL_OFFSET] 63 ldr r1, [r0, #EMIF_SDRAM_TIMING_1] 64 str r1, [r2, #EMIF_TIMING1_VAL_OFFSET] 66 ldr r1, [r0, #EMIF_SDRAM_TIMING_2] 67 str r1, [r2, #EMIF_TIMING2_VAL_OFFSET] 69 ldr r1, [r0, #EMIF_SDRAM_TIMING_3] 70 str r1, [r2, #EMIF_TIMING3_VAL_OFFSET] [all …]
|
| /kernel/linux/linux-6.6/drivers/memory/ |
| D | ti-emif-sram-pm.S | 50 ldr r1, [r0, #EMIF_SDRAM_CONFIG] 51 str r1, [r2, #EMIF_SDCFG_VAL_OFFSET] 53 ldr r1, [r0, #EMIF_SDRAM_REFRESH_CONTROL] 54 str r1, [r2, #EMIF_REF_CTRL_VAL_OFFSET] 56 ldr r1, [r0, #EMIF_SDRAM_TIMING_1] 57 str r1, [r2, #EMIF_TIMING1_VAL_OFFSET] 59 ldr r1, [r0, #EMIF_SDRAM_TIMING_2] 60 str r1, [r2, #EMIF_TIMING2_VAL_OFFSET] 62 ldr r1, [r0, #EMIF_SDRAM_TIMING_3] 63 str r1, [r2, #EMIF_TIMING3_VAL_OFFSET] [all …]
|
| /kernel/linux/linux-5.10/tools/testing/selftests/powerpc/pmu/ebb/ |
| D | busy_loop.S | 11 stdu %r1, -168(%r1) 12 std r14, 160(%r1) 13 std r15, 152(%r1) 14 std r16, 144(%r1) 15 std r17, 136(%r1) 16 std r18, 128(%r1) 17 std r19, 120(%r1) 18 std r20, 112(%r1) 19 std r21, 104(%r1) 20 std r22, 96(%r1) [all …]
|
| /kernel/linux/linux-6.6/tools/testing/selftests/powerpc/pmu/ebb/ |
| D | busy_loop.S | 11 stdu %r1, -168(%r1) 12 std r14, 160(%r1) 13 std r15, 152(%r1) 14 std r16, 144(%r1) 15 std r17, 136(%r1) 16 std r18, 128(%r1) 17 std r19, 120(%r1) 18 std r20, 112(%r1) 19 std r21, 104(%r1) 20 std r22, 96(%r1) [all …]
|
| /kernel/linux/linux-6.6/arch/microblaze/kernel/ |
| D | mcount.S | 18 addik r1, r1, -120; \ 19 swi r2, r1, 4; \ 20 swi r3, r1, 8; \ 21 swi r4, r1, 12; \ 22 swi r5, r1, 116; \ 23 swi r6, r1, 16; \ 24 swi r7, r1, 20; \ 25 swi r8, r1, 24; \ 26 swi r9, r1, 28; \ 27 swi r10, r1, 32; \ [all …]
|
| D | entry.S | 180 swi r2, r1, PT_R2; /* Save SDA */ \ 181 swi r3, r1, PT_R3; \ 182 swi r4, r1, PT_R4; \ 183 swi r5, r1, PT_R5; \ 184 swi r6, r1, PT_R6; \ 185 swi r7, r1, PT_R7; \ 186 swi r8, r1, PT_R8; \ 187 swi r9, r1, PT_R9; \ 188 swi r10, r1, PT_R10; \ 189 swi r11, r1, PT_R11; /* save clobbered regs after rval */\ [all …]
|
| /kernel/linux/linux-5.10/arch/powerpc/kernel/ |
| D | entry_64.S | 79 mr r10,r1 80 ld r1,PACAKSAVE(r13) 81 std r10,0(r1) 82 std r11,_NIP(r1) 83 std r12,_MSR(r1) 84 std r0,GPR0(r1) 85 std r10,GPR1(r1) 86 std r2,GPR2(r1) 91 std r3,GPR3(r1) 92 std r4,GPR4(r1) [all …]
|
| /kernel/linux/linux-5.10/arch/arm/mach-tegra/ |
| D | sleep-tegra30.S | 160 cpu_to_csr_reg r1, r3 161 add r1, r1, r12 @ virtual CSR address for this CPU 178 str r12, [r1] 186 ldr r3, [r1] @ read CSR 187 str r3, [r1] @ clear CSR 259 mov32 r1, tegra30_iram_start 260 sub r0, r0, r1 261 mov32 r1, TEGRA_IRAM_LPx_RESUME_AREA 262 add r0, r0, r1 323 mov r1, #(1 << 28) [all …]
|
| /kernel/linux/linux-6.6/arch/arm/mach-tegra/ |
| D | sleep-tegra30.S | 209 cpu_to_csr_reg r1, r3 210 add r1, r1, r12 @ virtual CSR address for this CPU 227 str r12, [r1] 235 ldr r3, [r1] @ read CSR 236 str r3, [r1] @ clear CSR 308 mov32 r1, tegra30_iram_start 309 sub r0, r0, r1 310 mov32 r1, TEGRA_IRAM_LPx_RESUME_AREA 311 add r0, r0, r1 372 mov r1, #(1 << 28) [all …]
|
| /kernel/linux/linux-5.10/crypto/ |
| D | serpent_generic.c | 233 static noinline void __serpent_setkey_sbox(u32 r0, u32 r1, u32 r2, in __serpent_setkey_sbox() argument 237 S3(r3, r4, r0, r1, r2); store_and_load_keys(r1, r2, r4, r3, 28, 24); in __serpent_setkey_sbox() 238 S4(r1, r2, r4, r3, r0); store_and_load_keys(r2, r4, r3, r0, 24, 20); in __serpent_setkey_sbox() 239 S5(r2, r4, r3, r0, r1); store_and_load_keys(r1, r2, r4, r0, 20, 16); in __serpent_setkey_sbox() 240 S6(r1, r2, r4, r0, r3); store_and_load_keys(r4, r3, r2, r0, 16, 12); in __serpent_setkey_sbox() 241 S7(r4, r3, r2, r0, r1); store_and_load_keys(r1, r2, r0, r4, 12, 8); in __serpent_setkey_sbox() 242 S0(r1, r2, r0, r4, r3); store_and_load_keys(r0, r2, r4, r1, 8, 4); in __serpent_setkey_sbox() 243 S1(r0, r2, r4, r1, r3); store_and_load_keys(r3, r4, r1, r0, 4, 0); in __serpent_setkey_sbox() 244 S2(r3, r4, r1, r0, r2); store_and_load_keys(r2, r4, r3, r0, 0, -4); in __serpent_setkey_sbox() 245 S3(r2, r4, r3, r0, r1); store_and_load_keys(r0, r1, r4, r2, -4, -8); in __serpent_setkey_sbox() [all …]
|
| /kernel/linux/linux-6.6/crypto/ |
| D | serpent_generic.c | 228 static noinline void __serpent_setkey_sbox(u32 r0, u32 r1, u32 r2, in __serpent_setkey_sbox() argument 232 S3(r3, r4, r0, r1, r2); store_and_load_keys(r1, r2, r4, r3, 28, 24); in __serpent_setkey_sbox() 233 S4(r1, r2, r4, r3, r0); store_and_load_keys(r2, r4, r3, r0, 24, 20); in __serpent_setkey_sbox() 234 S5(r2, r4, r3, r0, r1); store_and_load_keys(r1, r2, r4, r0, 20, 16); in __serpent_setkey_sbox() 235 S6(r1, r2, r4, r0, r3); store_and_load_keys(r4, r3, r2, r0, 16, 12); in __serpent_setkey_sbox() 236 S7(r4, r3, r2, r0, r1); store_and_load_keys(r1, r2, r0, r4, 12, 8); in __serpent_setkey_sbox() 237 S0(r1, r2, r0, r4, r3); store_and_load_keys(r0, r2, r4, r1, 8, 4); in __serpent_setkey_sbox() 238 S1(r0, r2, r4, r1, r3); store_and_load_keys(r3, r4, r1, r0, 4, 0); in __serpent_setkey_sbox() 239 S2(r3, r4, r1, r0, r2); store_and_load_keys(r2, r4, r3, r0, 0, -4); in __serpent_setkey_sbox() 240 S3(r2, r4, r3, r0, r1); store_and_load_keys(r0, r1, r4, r2, -4, -8); in __serpent_setkey_sbox() [all …]
|
| /kernel/linux/linux-6.6/arch/powerpc/kernel/ |
| D | interrupt_64.S | 22 ld r12,_NIP(r1) 28 ld r12,_MSR(r1) 33 ld r12,_NIP(r1) 39 ld r12,_MSR(r1) 52 mr r10,r1 53 ld r1,PACAKSAVE(r13) 54 std r10,0(r1) 55 std r11,_LINK(r1) 56 std r11,_NIP(r1) /* Saved LR is also the next instruction */ 57 std r12,_MSR(r1) [all …]
|
| /kernel/linux/linux-5.10/arch/arc/lib/ |
| D | strlen.S | 15 asl_s r1,r0,3 17 asl r7,r4,r1 19 sub r1,r2,r7 20 bic_s r1,r1,r2 24 or.eq r12,r12,r1 30 mov_s r1,31 31 sub3 r7,r1,r0 32 sub r1,r2,r4 33 bic_s r1,r1,r2 34 bmsk r1,r1,r7 [all …]
|
| /kernel/linux/linux-6.6/arch/arc/lib/ |
| D | strlen.S | 15 asl_s r1,r0,3 17 asl r7,r4,r1 19 sub r1,r2,r7 20 bic_s r1,r1,r2 24 or.eq r12,r12,r1 30 mov_s r1,31 31 sub3 r7,r1,r0 32 sub r1,r2,r4 33 bic_s r1,r1,r2 34 bmsk r1,r1,r7 [all …]
|