/kernel/linux/linux-5.10/lib/crypto/ |
D | sha256.c | 48 u32 a, b, c, d, e, f, g, h, t1, t2; in sha256_transform() local 66 t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; in sha256_transform() 68 t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; in sha256_transform() 70 t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; in sha256_transform() 72 t2 = e0(f) + Maj(f, g, h); a += t1; e = t1 + t2; in sha256_transform() 74 t2 = e0(e) + Maj(e, f, g); h += t1; d = t1 + t2; in sha256_transform() 76 t2 = e0(d) + Maj(d, e, f); g += t1; c = t1 + t2; in sha256_transform() 78 t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; in sha256_transform() 80 t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; in sha256_transform() 83 t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; in sha256_transform() [all …]
|
/kernel/linux/linux-5.10/drivers/block/paride/ |
D | bpck.c | 35 #define t2(pat) {PC ^= pat; out_p(2,PC);} macro 56 case 0: w0(r & 0xf); w0(r); t2(2); t2(4); in bpck_read_regr() 58 t2(4); in bpck_read_regr() 62 case 1: w0(r & 0xf); w0(r); t2(2); in bpck_read_regr() 63 e2(); t2(0x20); in bpck_read_regr() 64 t2(4); h = r0(); in bpck_read_regr() 65 t2(1); t2(0x20); in bpck_read_regr() 89 t2(2); in bpck_write_regr() 91 o2(); t2(4); t2(1); in bpck_write_regr() 115 w0(0x40); t2(2); t2(1); in bpck_write_block() [all …]
|
/kernel/linux/linux-5.10/arch/mips/kernel/ |
D | octeon_switch.S | 43 LONG_ADDI t2, a0, THREAD_CVMSEG /* Where to store CVMSEG to */ 51 LONG_S t8, 0(t2) /* Store CVMSEG to thread storage */ 52 LONG_ADDU t2, LONGSIZE*2 /* Increment loc in thread storage */ 54 LONG_S t9, -LONGSIZE(t2)/* Store CVMSEG to thread storage */ 78 set_saved_sp t0, t1, t2 105 dmfc2 t2, 0x0200 110 sd t2, OCTEON_CP2_CRC_POLY(a0) 124 dmfc2 t2, 0x0081 130 sd t2, OCTEON_CP2_3DES_KEY+8(a0) 131 dmfc2 t2, 0x0102 [all …]
|
D | scall64-n64.S | 56 dsubu t2, v0, __NR_64_Linux 57 sltiu t0, t2, __NR_64_Linux_syscalls 60 dsll t0, t2, 3 # offset into table 61 dla t2, sys_call_table 62 daddu t0, t2, t0 63 ld t2, (t0) # syscall routine 64 beqz t2, illegal_syscall 66 jalr t2 # Do The Real Thing (TM)
|
/kernel/linux/linux-5.10/arch/arm/crypto/ |
D | sha512-armv4.pl | 75 $t2="r11"; 102 ldr $t2,[sp,#$Hoff+0] @ h.lo 117 adds $Tlo,$Tlo,$t2 118 ldr $t2,[sp,#$Goff+0] @ g.lo 122 eor $t0,$t0,$t2 130 eor $t0,$t0,$t2 131 ldr $t2,[$Ktbl,#$lo] @ K[i].lo 139 adds $Tlo,$Tlo,$t2 140 and $t0,$t2,#0xff 143 ldr $t2,[sp,#$Boff+0] @ b.lo [all …]
|
D | sha256-armv4.pl | 63 $t2="r12"; 81 add $a,$a,$t2 @ h+=Maj(a,b,c) from the past 88 add $a,$a,$t2 @ h+=Maj(a,b,c) from the past 89 ldrb $t2,[$inp,#2] 91 orr $t1,$t1,$t2,lsl#8 92 ldrb $t2,[$inp],#4 98 orr $t1,$t1,$t2,lsl#24 103 ldr $t2,[$Ktbl],#4 @ *K256++ 109 add $h,$h,$t2 @ h+=K256[i] 114 and $t2,$t2,#0xff [all …]
|
/kernel/linux/linux-5.10/arch/mips/netlogic/common/ |
D | smpboot.S | 63 PTR_LA t2, nlm_boot_siblings 64 dsubu t2, t1 65 daddu t2, t0 67 jalr t2 107 andi t2, t0, 0x3 /* thread num */ 116 bnez t2, 1f /* skip thread programming */ 124 li t2, 6 /* XLR thread mode mask */ 125 nor t3, t2, zero 126 and t2, t1, t2 /* t2 - current thread mode */ 130 beq v1, t2, 1f /* same as request value */ [all …]
|
D | reset.S | 62 lui t2, 0x4080 /* Enable Unaligned Access, L2HPE */ 63 or t1, t1, t2 103 li t2, 0 /* index */ 106 sll v0, t2, 5 123 addi t2, 1 124 bne t3, t2, 11b 188 li t2, 0x40000 189 mul t3, t2, t1 /* t3 = node * 0x40000 */ 195 li t2, SYS_CPU_COHERENT_BASE 196 add t2, t2, t3 /* t2 <- SYS offset for node */ [all …]
|
/kernel/linux/linux-5.10/lib/mpi/ |
D | mpi-inv.c | 33 MPI u, v, u1, u2 = NULL, u3, v1, v2 = NULL, v3, t1, t2 = NULL, t3; in mpi_invm() local 65 t2 = mpi_alloc_set_ui(1); in mpi_invm() 66 t2->sign = 1; in mpi_invm() 74 t2 = mpi_alloc_set_ui(0); in mpi_invm() 81 if (mpi_test_bit(t1, 0) || mpi_test_bit(t2, 0)) { in mpi_invm() 84 mpi_sub(t2, t2, u); in mpi_invm() 87 mpi_rshift(t2, t2, 1); in mpi_invm() 102 mpi_set(u2, t2); in mpi_invm() 108 mpi_sub(v2, u, t2); in mpi_invm() 116 mpi_sub(t2, u2, v2); in mpi_invm() [all …]
|
/kernel/linux/linux-5.10/include/net/netfilter/ |
D | nf_conntrack_tuple.h | 125 const struct nf_conntrack_tuple *t2) in __nf_ct_tuple_src_equal() argument 127 return (nf_inet_addr_cmp(&t1->src.u3, &t2->src.u3) && in __nf_ct_tuple_src_equal() 128 t1->src.u.all == t2->src.u.all && in __nf_ct_tuple_src_equal() 129 t1->src.l3num == t2->src.l3num); in __nf_ct_tuple_src_equal() 133 const struct nf_conntrack_tuple *t2) in __nf_ct_tuple_dst_equal() argument 135 return (nf_inet_addr_cmp(&t1->dst.u3, &t2->dst.u3) && in __nf_ct_tuple_dst_equal() 136 t1->dst.u.all == t2->dst.u.all && in __nf_ct_tuple_dst_equal() 137 t1->dst.protonum == t2->dst.protonum); in __nf_ct_tuple_dst_equal() 141 const struct nf_conntrack_tuple *t2) in nf_ct_tuple_equal() argument 143 return __nf_ct_tuple_src_equal(t1, t2) && in nf_ct_tuple_equal() [all …]
|
/kernel/linux/linux-5.10/arch/x86/crypto/ |
D | glue_helper-asm-avx2.S | 56 t1x, t2, t2x, t3, t3x, t4, t5) \ argument 66 vinserti128 $1, t2x, t3, t2; /* ab: le0 ; cd: le1 */ \ 67 vpshufb t1, t2, x0; \ 70 add2_le128(t2, t0, t4, t3, t5); /* ab: le2 ; cd: le3 */ \ 71 vpshufb t1, t2, x1; \ 72 add2_le128(t2, t0, t4, t3, t5); \ 73 vpshufb t1, t2, x2; \ 74 add2_le128(t2, t0, t4, t3, t5); \ 75 vpshufb t1, t2, x3; \ 76 add2_le128(t2, t0, t4, t3, t5); \ [all …]
|
/kernel/linux/linux-5.10/arch/alpha/lib/ |
D | stxcpy.S | 47 lda t2, -1 # e1 : build a mask against false zero 48 mskqh t2, a1, t2 # e0 : detection in the src word 50 ornot t1, t2, t2 # .. e1 : 52 cmpbge zero, t2, t8 # .. e1 : bits set iff null found 132 ldq_u t2, 8(a1) # e0 : 136 extqh t2, a1, t4 # e0 : 148 or t6, t2, t2 # e1 : already extracted before 149 cmpbge zero, t2, t8 # e0 : testing eos 157 extql t2, a1, t0 # e0 : position ho-bits of lo word 158 ldq_u t2, 8(a1) # .. e1 : read next high-order source word [all …]
|
D | ev6-stxcpy.S | 58 lda t2, -1 # E : build a mask against false zero 59 mskqh t2, a1, t2 # U : detection in the src word (stall) 61 ornot t1, t2, t2 # E : (stall) 64 cmpbge zero, t2, t8 # E : bits set iff null found 155 ldq_u t2, 8(a1) # L : 158 extqh t2, a1, t4 # U : (stall on a1) 171 or t6, t2, t2 # E : already extracted before (stall) 172 cmpbge zero, t2, t8 # E : testing eos (stall) 180 extql t2, a1, t0 # U : position ho-bits of lo word 181 ldq_u t2, 8(a1) # U : read next high-order source word [all …]
|
D | stxncpy.S | 55 lda t2, -1 # e1 : build a mask against false zero 56 mskqh t2, a1, t2 # e0 : detection in the src word 58 ornot t1, t2, t2 # .. e1 : 60 cmpbge zero, t2, t8 # .. e1 : bits set iff null found 127 and a2, 7, t2 # e1 : 130 sll t10, t2, t10 # e0 : t10 = bitmask of last count byte 158 ldq_u t2, 8(a1) # e0 : load second src word 161 extqh t2, a1, t4 # e0 : 174 or t6, t2, t2 # .. e1 : 175 cmpbge zero, t2, t8 # e0 : find nulls in second partial [all …]
|
D | ev6-stxncpy.S | 66 lda t2, -1 # E : build a mask against false zero 67 mskqh t2, a1, t2 # U : detection in the src word (stall) 69 ornot t1, t2, t2 # E : (stall) 72 cmpbge zero, t2, t8 # E : bits set iff null found 160 and a2, 7, t2 # E : (stall) 164 sll t10, t2, t10 # U : t10 = bitmask of last count byte 197 ldq_u t2, 8(a1) # L : Latency=3 load second src word 200 extqh t2, a1, t4 # U : (3 cycle stall on t2) 215 or t6, t2, t2 # E : (stall) 217 cmpbge zero, t2, t8 # E : find nulls in second partial [all …]
|
D | ev67-strrchr.S | 35 and a1, 0xff, t2 # E : 00000000000000ch 41 or t2, t4, a1 # E : 000000000000chch 47 sll a1, 32, t2 # U : 0000chch00000000 51 or t2, t3, t2 # E : 0000chchchch0000 52 or a1, t2, a1 # E : chchchchchchchch 57 xor t0, a1, t2 # E : make bytes == c zero 60 cmpbge zero, t2, t3 # E : bits set iff byte == c 75 xor t0, a1, t2 # E : 78 cmpbge zero, t2, t3 # E : bits set iff byte == c 97 ctlz t8, t2 # U0 : Latency=3 (0x40 for t8=0) [all …]
|
D | strchr.S | 29 cmpbge zero, t0, t2 # .. e1 : bits set iff byte == zero 37 or t2, t3, t0 # e1 : bits set iff char match or zero match 45 cmpbge zero, t0, t2 # e0 : bits set iff byte == 0 47 or t2, t3, t0 # e0 : 56 and t0, 0xf0, t2 # e0 : binary search for that set bit 59 cmovne t2, 4, t2 # .. e1 : 62 addq t2, t3, t2 # e0 : 64 addq v0, t2, v0 # e0 :
|
D | strrchr.S | 36 xor t0, a1, t2 # e0 : make bytes == c zero 38 cmpbge zero, t2, t3 # e0 : bits set iff byte == c 49 xor t0, a1, t2 # e0 : 51 cmpbge zero, t2, t3 # e0 : bits set iff byte == c 71 and t8, 0xf0, t2 # e0 : binary search for the high bit set 72 cmovne t2, t2, t8 # .. e1 (zdb) 73 cmovne t2, 4, t2 # e0 : 79 addq t2, t1, t1 # e0 :
|
/kernel/linux/linux-5.10/arch/mips/dec/ |
D | int-handler.S | 135 lw t2,cpu_fpu_mask 143 and t2,t0 144 bnez t2,fpu # handle FPU immediately 214 2: lw t2,(t1) 216 and t2,t0 217 beq zero,t2,2b 241 li t2,4 # nr of bits / 2 247 li t2,4 # nr of bits / 2 253 li t2,8 # nr of bits / 2 262 li t2,8 # nr of bits / 2 [all …]
|
/kernel/linux/linux-5.10/crypto/ |
D | sha512_generic.c | 101 u64 a, b, c, d, e, f, g, h, t1, t2; in sha512_transform() local 127 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; in sha512_transform() 129 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; in sha512_transform() 131 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; in sha512_transform() 133 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; in sha512_transform() 135 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; in sha512_transform() 137 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; in sha512_transform() 139 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; in sha512_transform() 141 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; in sha512_transform() 148 a = b = c = d = e = f = g = h = t1 = t2 = 0; in sha512_transform()
|
/kernel/linux/linux-5.10/arch/powerpc/crypto/ |
D | aes-spe-keys.S | 39 #define LS_BOX(r, t1, t2) \ argument 40 lis t2,PPC_AES_4K_ENCTAB@h; \ 41 ori t2,t2,PPC_AES_4K_ENCTAB@l; \ 42 rlwimi t2,r,4,20,27; \ 43 lbz t1,8(t2); \ 45 rlwimi t2,r,28,20,27; \ 46 lbz t1,8(t2); \ 48 rlwimi t2,r,20,20,27; \ 49 lbz t1,8(t2); \ 51 rlwimi t2,r,12,20,27; \ [all …]
|
/kernel/linux/linux-5.10/arch/mips/lib/ |
D | csum_partial.S | 27 #undef t2 31 #define t2 $10 macro 113 move t2, a1 190 andi t2, a1, 0x40 204 beqz t2, 1f 205 andi t2, a1, 0x20 213 beqz t2, .Ldo_end_words 223 andi t2, a1, 0x3 237 move a1, t2 262 lbu t2, (src) [all …]
|
/kernel/linux/linux-5.10/arch/mips/include/asm/mach-ip27/ |
D | kernel-entry-init.h | 38 dsll t2, NASID_SHFT # Same for data nasid 40 or t2, t2, t0 # Physical load address of kernel data 42 dsrl t2, 12 # 4K pfn 44 dsll t2, 6 # Get pfn into place 49 or t0, t0, t2 69 move t2, t1 # text and data are here 84 lh t2, KV_RW_NASID_OFFSET(t0)
|
/kernel/linux/linux-5.10/arch/arm64/crypto/ |
D | sha512-armv8.pl | 104 ($t0,$t1,$t2,$t3)=map("$reg_t$_",(16,17,19,28)); 141 add $h,$h,$t2 // h+=K[i] 144 bic $t2,$g,$e 146 orr $t1,$t1,$t2 // Ch(e,f,g) 147 eor $t2,$a,$b // a^b, b^c in next round 153 and $t3,$t3,$t2 // (b^c)&=(a^b) 158 ldr $t3,[$Ktbl],#$SZ // *K++, $t2 in next round 163 add $h,$h,$t2 // h+=K[i] 167 bic $t2,$g,$e 172 orr $t1,$t1,$t2 // Ch(e,f,g) [all …]
|
/kernel/linux/linux-5.10/arch/alpha/include/asm/ |
D | word-at-a-time.h | 42 unsigned long t1, t2, t3; in find_zero() 47 t2 = bits & 0xcc; in find_zero() 50 if (t2) t2 = 2; in find_zero() 52 return t1 + t2 + t3; in find_zero()
|