| /kernel/linux/linux-5.10/arch/arm64/crypto/ |
| D | chacha-neon-core.S | 42 ld1 {v12.4s}, [x10] 46 add v0.4s, v0.4s, v1.4s 47 eor v3.16b, v3.16b, v0.16b 51 add v2.4s, v2.4s, v3.4s 52 eor v4.16b, v1.16b, v2.16b 53 shl v1.4s, v4.4s, #12 54 sri v1.4s, v4.4s, #20 57 add v0.4s, v0.4s, v1.4s 58 eor v3.16b, v3.16b, v0.16b 59 tbl v3.16b, {v3.16b}, v12.16b [all …]
|
| D | sm3-ce-core.S | 11 .irp b, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12 12 .set .Lv\b\().4s, \b 44 sm3ss1 v5.4s, v8.4s, \t0\().4s, v9.4s 45 shl \t1\().4s, \t0\().4s, #1 46 sri \t1\().4s, \t0\().4s, #31 47 sm3tt1\ab v8.4s, v5.4s, v10.4s, \i 48 sm3tt2\ab v9.4s, v5.4s, \s0\().4s, \i 53 ext \s4\().16b, \s1\().16b, \s2\().16b, #12 54 ext v6.16b, \s0\().16b, \s1\().16b, #12 55 ext v7.16b, \s2\().16b, \s3\().16b, #8 [all …]
|
| D | sha256-core.S_shipped | 42 // on Cortex-A53 (or by 4 cycles per round). 83 b.ne .Lv8_entry 85 b.ne .Lneon_entry 95 sub sp,sp,#4*4 98 ldp w22,w23,[x0,#2*4] 99 ldp w24,w25,[x0,#4*4] 101 ldp w26,w27,[x0,#6*4] 106 ldp w3,w4,[x1],#2*4 107 ldr w19,[x30],#4 // *K++ 120 eor w19,w20,w21 // a^b, b^c in next round [all …]
|
| D | aes-ce-core.S | 13 ld1 {v0.16b}, [x2] 14 ld1 {v1.4s}, [x0], #16 18 mov v3.16b, v1.16b 19 b 2f 20 0: mov v2.16b, v1.16b 21 ld1 {v3.4s}, [x0], #16 22 1: aese v0.16b, v2.16b 23 aesmc v0.16b, v0.16b 24 2: ld1 {v1.4s}, [x0], #16 25 aese v0.16b, v3.16b [all …]
|
| D | aes-ce-ccm-core.S | 20 ld1 {v0.16b}, [x0] /* load mac */ 23 eor v1.16b, v1.16b, v1.16b 27 ins v1.b[0], w7 28 ext v1.16b, v1.16b, v1.16b, #1 /* rotate in the input bytes */ 30 cbnz w8, 0b 31 eor v0.16b, v0.16b, v1.16b 32 1: ld1 {v3.4s}, [x4] /* load first round key */ 39 mov v5.16b, v3.16b 40 b 4f 41 2: mov v4.16b, v3.16b [all …]
|
| D | sha512-ce-core.S | 15 .irp b,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19 16 .set .Lq\b, \b 17 .set .Lv\b\().2d, \b 40 .align 4 88 ext v6.16b, v\i2\().16b, v\i3\().16b, #8 89 ext v5.16b, v5.16b, v5.16b, #8 90 ext v7.16b, v\i1\().16b, v\i2\().16b, #8 93 ext v5.16b, v\in3\().16b, v\in4\().16b, #8 113 /* load first 4 round constants */ 122 CPU_LE( rev64 v12.16b, v12.16b ) [all …]
|
| D | ghash-ce-core.S | 71 ext t3.8b, \ad\().8b, \ad\().8b, #1 // A1 72 ext t5.8b, \ad\().8b, \ad\().8b, #2 // A2 73 ext t7.8b, \ad\().8b, \ad\().8b, #3 // A3 79 tbl t3.16b, {\ad\().16b}, perm1.16b // A1 80 tbl t5.16b, {\ad\().16b}, perm2.16b // A2 81 tbl t7.16b, {\ad\().16b}, perm3.16b // A3 87 __pmull_p8_tail \rq, \ad\().8b, SHASH.8b, 8b,, sh1, sh2, sh3, sh4 91 __pmull_p8_tail \rq, \ad\().8b, SHASH2.8b, 8b,, ss1, ss2, ss3, ss4 95 __pmull_p8_tail \rq, \ad\().16b, SHASH.16b, 16b, 2, sh1, sh2, sh3, sh4 99 pmull\t t3.8h, t3.\nb, \bd // F = A1*B [all …]
|
| D | aes-ce.S | 32 ld1 {v17.4s-v18.4s}, [\rk], #32 33 1111: ld1 {v19.4s-v20.4s}, [\rk], #32 34 2222: ld1 {v21.4s-v24.4s}, [\rk], #64 35 ld1 {v25.4s-v28.4s}, [\rk], #64 36 ld1 {v29.4s-v31.4s}, [\rk] 58 aes\de \i0\().16b, \k\().16b 59 aes\mc \i0\().16b, \i0\().16b 61 aes\de \i1\().16b, \k\().16b 62 aes\mc \i1\().16b, \i1\().16b 64 aes\de \i2\().16b, \k\().16b [all …]
|
| /kernel/linux/linux-5.10/arch/x86/lib/ |
| D | usercopy_32.c | 47 "3: lea 0(%2,%0,4),%0\n" \ 48 " jmp 2b\n" \ 50 _ASM_EXTABLE_UA(0b, 3b) \ 51 _ASM_EXTABLE_UA(1b, 2b) \ 53 : "r"(size & 3), "0"(size / 4), "1"(addr), "a"(0)); \ 102 "1: movl 32(%4), %%eax\n" in __copy_user_intel() 105 "2: movl 64(%4), %%eax\n" in __copy_user_intel() 107 "3: movl 0(%4), %%eax\n" in __copy_user_intel() 108 "4: movl 4(%4), %%edx\n" in __copy_user_intel() 110 "6: movl %%edx, 4(%3)\n" in __copy_user_intel() [all …]
|
| D | copy_user_64.S | 33 jnz 100b 40 _ASM_EXTABLE_CPY(100b, 103b) 41 _ASM_EXTABLE_CPY(101b, 103b) 69 4: movq 3*8(%rsi),%r11 74 9: movq 4*8(%rsi),%r8 78 13: movq %r8,4*8(%rdi) 85 jnz 1b 96 jnz 18b 105 jnz 21b 120 _ASM_EXTABLE_CPY(1b, 30b) [all …]
|
| /kernel/linux/linux-5.10/arch/arc/include/asm/ |
| D | uaccess.h | 63 case 4: __arc_get_user_one(*(k), u, "ld", __ret); break; \ 80 " .align 4\n" \ 85 " j 2b\n" \ 88 " .align 4\n" \ 89 " .word 1b,3b\n" \ 98 "4: ld %R1,[%2, 4]\n" \ 101 " .align 4\n" \ 107 " j 2b\n" \ 110 " .align 4\n" \ 111 " .word 1b,3b\n" \ [all …]
|
| /kernel/linux/linux-5.10/Documentation/userspace-api/media/v4l/ |
| D | pixfmt-rgb.rst | 41 - 4 50 - 4 59 - 4 68 - 4 84 - b\ :sub:`1` 85 - b\ :sub:`0` 96 - b\ :sub:`3` 97 - b\ :sub:`2` 98 - b\ :sub:`1` 99 - b\ :sub:`0` [all …]
|
| /kernel/linux/linux-5.10/drivers/staging/comedi/drivers/ni_routing/ni_route_values/ |
| D | ni_eseries.c | 54 [B(NI_PFI(0))] = { 55 [B(NI_AI_StartTrigger)] = I(NI_PFI_OUTPUT_AI_START1), 57 [B(NI_PFI(1))] = { 58 [B(NI_AI_ReferenceTrigger)] = I(NI_PFI_OUTPUT_AI_START2), 60 [B(NI_PFI(2))] = { 61 [B(NI_AI_ConvertClock)] = I(NI_PFI_OUTPUT_AI_CONVERT), 63 [B(NI_PFI(3))] = { 64 [B(NI_CtrSource(1))] = I(NI_PFI_OUTPUT_G_SRC1), 66 [B(NI_PFI(4))] = { 67 [B(NI_CtrGate(1))] = I(NI_PFI_OUTPUT_G_GATE1), [all …]
|
| D | ni_660x.c | 49 [B(NI_PFI(8))] = { 50 [B(NI_CtrInternalOutput(7))] = I(1), 52 [B(NI_PFI(10))] = { 53 [B(NI_CtrGate(7))] = I(1), 55 [B(NI_PFI(11))] = { 56 [B(NI_CtrSource(7))] = I(1), 58 [B(NI_PFI(12))] = { 59 [B(NI_CtrInternalOutput(6))] = I(1), 61 [B(NI_PFI(14))] = { 62 [B(NI_CtrGate(6))] = I(1), [all …]
|
| D | ni_mseries.c | 57 [B(NI_PFI(0))] = { 58 [B(TRIGGER_LINE(0))] = I(18), 59 [B(TRIGGER_LINE(1))] = I(19), 60 [B(TRIGGER_LINE(2))] = I(20), 61 [B(TRIGGER_LINE(3))] = I(21), 62 [B(TRIGGER_LINE(4))] = I(22), 63 [B(TRIGGER_LINE(5))] = I(23), 64 [B(TRIGGER_LINE(6))] = I(24), 65 [B(TRIGGER_LINE(7))] = I(25), 66 [B(NI_CtrSource(0))] = I(9), [all …]
|
| /kernel/linux/linux-5.10/arch/powerpc/lib/ |
| D | copy_32.S | 16 lwz r7,4(r4); \ 20 stw r7,4(r6); \ 27 lwz r7,4(r4); \ 34 8 ## n ## 4: \ 35 stw r7,4(r6); \ 46 b 104f; \ 49 b 105f; \ 50 EX_TABLE(8 ## n ## 0b,9 ## n ## 0b); \ 51 EX_TABLE(8 ## n ## 1b,9 ## n ## 0b); \ 52 EX_TABLE(8 ## n ## 2b,9 ## n ## 0b); \ [all …]
|
| /kernel/linux/linux-5.10/arch/m68k/lib/ |
| D | uaccess.c | 21 " jne 1b\n" in __generic_copy_from_user() 23 " jeq 4f\n" in __generic_copy_from_user() 26 "4: btst #0,%5\n" in __generic_copy_from_user() 28 "5: "MOVES".b (%1)+,%3\n" in __generic_copy_from_user() 29 " move.b %3,(%2)+\n" in __generic_copy_from_user() 38 " jeq 6b\n" in __generic_copy_from_user() 40 " jra 6b\n" in __generic_copy_from_user() 44 " .align 4\n" in __generic_copy_from_user() 45 " .long 1b,10b\n" in __generic_copy_from_user() 46 " .long 3b,30b\n" in __generic_copy_from_user() [all …]
|
| /kernel/linux/linux-5.10/arch/s390/lib/ |
| D | uaccess.c | 113 "6: jz 4f\n" in copy_from_user_mvcos() 117 " j 0b\n" in copy_from_user_mvcos() 118 "2: la %4,4095(%1)\n"/* %4 = ptr + 4095 */ in copy_from_user_mvcos() 119 " nr %4,%3\n" /* %4 = (ptr + 4095) & -4096 */ in copy_from_user_mvcos() 120 " slgr %4,%1\n" in copy_from_user_mvcos() 121 " clgr %0,%4\n" /* copy crosses next page boundary? */ in copy_from_user_mvcos() 123 "3: .insn ss,0xc80000000000,0(%4,%2),0(%1),0\n" in copy_from_user_mvcos() 124 "7: slgr %0,%4\n" in copy_from_user_mvcos() 126 "4: slgr %0,%0\n" in copy_from_user_mvcos() 128 EX_TABLE(0b,2b) EX_TABLE(3b,5b) EX_TABLE(6b,2b) EX_TABLE(7b,5b) in copy_from_user_mvcos() [all …]
|
| /kernel/linux/linux-5.10/arch/x86/crypto/ |
| D | twofish-x86_64-asm_64.S | 14 #define b_offset 4 61 * b input register containing b 64 * operations on a and b are interleaved to increase performance 66 #define encrypt_round(a,b,c,d,round)\ argument 67 movzx b ## B, %edi;\ 68 mov s1(%r11,%rdi,4),%r8d;\ 69 movzx a ## B, %edi;\ 70 mov s2(%r11,%rdi,4),%r9d;\ 71 movzx b ## H, %edi;\ 72 ror $16, b ## D;\ [all …]
|
| D | twofish-i586-asm_32.S | 17 #define ctx 4 /* Twofish context structure */ 20 #define b_offset 4 62 * b input register containing b 65 * operations on a and b are interleaved to increase performance 67 #define encrypt_round(a,b,c,d,round)\ argument 69 movzx b ## B, %edi;\ 70 mov s1(%ebp,%edi,4),d ## D;\ 71 movzx a ## B, %edi;\ 72 mov s2(%ebp,%edi,4),%esi;\ 73 movzx b ## H, %edi;\ [all …]
|
| /kernel/linux/linux-5.10/arch/parisc/lib/ |
| D | lusercopy.S | 63 2: b $lclu_done 66 ASM_EXCEPTIONTABLE_ENTRY(1b,2b) 92 b $lslen_done 95 3: b $lslen_done 98 ASM_EXCEPTIONTABLE_ENTRY(1b,3b) 99 ASM_EXCEPTIONTABLE_ENTRY(2b,3b) 183 b .Lalign_loop64 186 ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done) 187 ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done) 205 ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done) [all …]
|
| /kernel/linux/linux-5.10/arch/powerpc/boot/ |
| D | string.S | 18 bne 1b 31 bdnzf 2,1b /* dec ctr, branch if ctr != 0 && !cr0.eq */ 40 bne 1b 45 bne 1b 55 bne 1b 68 beq 1b 81 bdnzt eq,1b 89 bne 1b 97 addi r6,r3,-4 98 cmplwi 0,r5,4 [all …]
|
| /kernel/linux/linux-5.10/arch/arm/mm/ |
| D | abort-lv4t.S | 37 /* 0 */ b .data_arm_lateldrhpost @ ldrh rd, [rn], #m/rm 38 /* 1 */ b .data_arm_lateldrhpre @ ldrh rd, [rn, #m/rm] 39 /* 2 */ b .data_unknown 40 /* 3 */ b .data_unknown 41 /* 4 */ b .data_arm_lateldrpostconst @ ldr rd, [rn], #m 42 /* 5 */ b .data_arm_lateldrpreconst @ ldr rd, [rn, #m] 43 /* 6 */ b .data_arm_lateldrpostreg @ ldr rd, [rn], rm 44 /* 7 */ b .data_arm_lateldrprereg @ ldr rd, [rn, rm] 45 /* 8 */ b .data_arm_ldmstm @ ldm*a rn, <rlist> 46 /* 9 */ b .data_arm_ldmstm @ ldm*b rn, <rlist> [all …]
|
| /kernel/linux/linux-5.10/arch/arm/crypto/ |
| D | sha256-core.S_shipped | 113 sub sp,sp,#16*4 @ alloca(X[16]) 116 ldr r2,[r1],#4 123 @ ldr r2,[r1],#4 @ 0 125 str r1,[sp,#17*4] @ make room for r1 128 add r4,r4,r12 @ h+=Maj(a,b,c) from the past 135 add r4,r4,r12 @ h+=Maj(a,b,c) from the past 139 ldrb r12,[r1],#4 142 str r1,[sp,#17*4] @ make room for r1 148 ldr r12,[r14],#4 @ *K256++ 150 str r2,[sp,#0*4] [all …]
|
| /kernel/linux/linux-5.10/arch/mips/include/asm/ |
| D | unaligned-emul.h | 19 "4:\tli\t%1, %3\n\t" \ 23 STR(PTR)"\t1b, 4b\n\t" \ 24 STR(PTR)"\t2b, 4b\n\t" \ 40 "4:\tli\t%1, %3\n\t" \ 44 STR(PTR)"\t1b, 4b\n\t" \ 45 STR(PTR)"\t2b, 4b\n\t" \ 65 "4:"type##_lbu("$1", "3(%2)")"\n\t" \ 77 STR(PTR)"\t1b, 11b\n\t" \ 78 STR(PTR)"\t2b, 11b\n\t" \ 79 STR(PTR)"\t3b, 11b\n\t" \ [all …]
|