Home
last modified time | relevance | path

Searched refs:t4 (Results 1 – 25 of 73) sorted by relevance

123

/kernel/linux/linux-5.10/arch/alpha/lib/
Dev67-strrchr.S36 insbl a1, 1, t4 # U : 000000000000ch00
41 or t2, t4, a1 # E : 000000000000chch
48 sll a1, 48, t4 # U : chch000000000000
50 or t4, a1, a1 # E : chch00000000chch
56 mskqh t5, a0, t4 # E : Complete garbage mask
58 cmpbge zero, t4, t4 # E : bits set iff byte is garbage
61 andnot t1, t4, t1 # E : clear garbage from null test
62 andnot t3, t4, t3 # E : clear garbage from char test
84 negq t1, t4 # E : isolate first null byte match
85 and t1, t4, t4 # E :
[all …]
Dstrrchr.S31 lda t4, -1 # .. e1 : build garbage mask
34 mskqh t4, a0, t4 # e0 :
37 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage
39 andnot t1, t4, t1 # .. e1 : clear garbage from null test
40 andnot t3, t4, t3 # e0 : clear garbage from char test
56 negq t1, t4 # e0 : isolate first null byte match
57 and t1, t4, t4 # e1 :
58 subq t4, 1, t5 # e0 : build a mask of the bytes up to...
59 or t4, t5, t4 # e1 : ... and including the null
61 and t3, t4, t3 # e0 : mask out char matches after null
Dstrchr.S27 lda t4, -1 # .. e1 : build garbage mask
30 mskqh t4, a0, t4 # e0 :
33 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage
38 andnot t0, t4, t0 # e0 : clear garbage bits
58 and t0, 0xaa, t4 # e0 :
61 cmovne t4, 1, t4 # .. e1 :
63 addq v0, t4, v0 # .. e1 :
Dev67-strchr.S40 lda t4, -1 # E : build garbage mask
42 mskqh t4, a0, t4 # U : only want relevant part of first quad
50 cmpbge zero, t4, t4 # E : bits set iff byte is garbage
58 andnot t0, t4, t0 # E : clear garbage bits
Dstxcpy.S136 extqh t2, a1, t4 # e0 :
138 or t1, t4, t1 # .. e1 :
238 and a0, 7, t4 # .. e1 : find dest misalignment
246 beq t4, 1f # .. e1 :
251 subq a1, t4, a1 # .. e1 : sub dest misalignment from src addr
256 cmplt t4, t5, t12 # e0 :
Dstxncpy.S161 extqh t2, a1, t4 # e0 :
162 or t1, t4, t1 # e1 : first aligned src word complete
295 and a0, 7, t4 # .. e1 : find dest misalignment
303 beq t4, 1f # .. e1 :
307 subq a1, t4, a1 # .. e1 : sub dest misalignment from src addr
312 1: cmplt t4, t5, t12 # e1 :
333 mskqh t1, t4, t1 # e0 :
Dev6-stxcpy.S158 extqh t2, a1, t4 # U : (stall on a1)
161 or t1, t4, t1 # E :
268 and a0, 7, t4 # E : find dest misalignment
275 beq t4, 1f # U :
284 subq a1, t4, a1 # E : sub dest misalignment from src addr
287 cmplt t4, t5, t12 # E :
Dev6-stxncpy.S200 extqh t2, a1, t4 # U : (3 cycle stall on t2)
202 or t1, t4, t1 # E : first aligned src word complete (stall)
340 and a0, 7, t4 # E : find dest misalignment
347 beq t4, 1f # U :
354 subq a1, t4, a1 # E : sub dest misalignment from src addr
359 1: cmplt t4, t5, t12 # E :
381 mskqh t1, t4, t1 # U :
/kernel/linux/linux-5.10/arch/arm64/crypto/
Dcrct10dif-ce-core.S84 t4 .req v18
136 ext t4.8b, ad.8b, ad.8b, #1 // A1
140 pmull t4.8h, t4.8b, fold_consts.8b // F = A1*B
150 tbl t4.16b, {ad.16b}, perm1.16b // A1
154 pmull2 t4.8h, t4.16b, fold_consts.16b // F = A1*B
162 0: eor t4.16b, t4.16b, t8.16b // L = E + F
166 uzp1 t8.2d, t4.2d, t5.2d
167 uzp2 t4.2d, t4.2d, t5.2d
173 eor t8.16b, t8.16b, t4.16b
174 and t4.16b, t4.16b, k32_48.16b
[all …]
/kernel/linux/linux-5.10/arch/ia64/lib/
Dmemcpy.S36 # define t4 r22 macro
190 sub t4=r0,dst // t4 = -dst
194 shl t4=t4,3 // t4 = 8*(dst & 7)
202 mov pr=t4,0x38 // (p5,p4,p3)=(dst & 7)
225 mov t4=ip
228 adds t4=.memcpy_loops-1b,t4
243 add t4=t0,t4
255 mov b6=t4
Dcopy_page_mck.S80 #define t4 r20 macro
85 #define t10 t4 // alias!
144 (p[D]) ld8 t4 = [src1], 3*8 // M1
156 (p[D]) st8 [dst1] = t4, 3*8
Dcopy_page.S44 .rotr t1[PIPE_DEPTH], t2[PIPE_DEPTH], t3[PIPE_DEPTH], t4[PIPE_DEPTH], \
79 (p[0]) ld8 t4[0]=[src2],16
80 (EPI) st8 [tgt2]=t4[PIPE_DEPTH-1],16
/kernel/linux/linux-5.10/arch/arm/crypto/
Dsha256-armv4.pl51 $inp="r1"; $t4="r1";
78 str $inp,[sp,#17*4] @ make room for $t4
95 str $inp,[sp,#17*4] @ make room for $t4
127 ldr $t4,[sp,#`($i+15)%16`*4] @ from future BODY_16_xx
144 @ ldr $t4,[sp,#`($i+14)%16`*4]
147 mov $t2,$t4,ror#$sigma1[0]
149 eor $t2,$t2,$t4,ror#$sigma1[1]
152 eor $t2,$t2,$t4,lsr#$sigma1[2] @ sigma1(X[i+14])
153 ldr $t4,[sp,#`($i+9)%16`*4]
159 add $t1,$t1,$t4 @ X[i]
[all …]
Daes-cipher-core.S42 .macro __hround, out0, out1, in0, in1, in2, in3, t3, t4, enc, sz, op, oldcpsr
65 __select \t4, \in0, 3
68 __select \t4, \in2, 3
72 __load \t4, \t4, 3, \sz, \op
87 eor \out1, \out1, \t4, ror #8
/kernel/linux/linux-5.10/arch/x86/crypto/
Dcamellia-aesni-avx2-asm_64.S63 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument
68 vbroadcasti128 .Linv_shift_row, t4; \
76 vpshufb t4, x0, x0; \
77 vpshufb t4, x7, x7; \
78 vpshufb t4, x3, x3; \
79 vpshufb t4, x6, x6; \
80 vpshufb t4, x2, x2; \
81 vpshufb t4, x5, x5; \
82 vpshufb t4, x1, x1; \
83 vpshufb t4, x4, x4; \
[all …]
Dglue_helper-asm-avx2.S56 t1x, t2, t2x, t3, t3x, t4, t5) \ argument
59 vpaddq t0, t0, t4; /* ab: -2:0 ; cd: -2:0 */\
70 add2_le128(t2, t0, t4, t3, t5); /* ab: le2 ; cd: le3 */ \
72 add2_le128(t2, t0, t4, t3, t5); \
74 add2_le128(t2, t0, t4, t3, t5); \
76 add2_le128(t2, t0, t4, t3, t5); \
78 add2_le128(t2, t0, t4, t3, t5); \
80 add2_le128(t2, t0, t4, t3, t5); \
82 add2_le128(t2, t0, t4, t3, t5); \
Dcamellia-aesni-avx-asm_64.S51 #define roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument
56 vmovdqa .Linv_shift_row, t4; \
62 vpshufb t4, x0, x0; \
63 vpshufb t4, x7, x7; \
64 vpshufb t4, x1, x1; \
65 vpshufb t4, x4, x4; \
66 vpshufb t4, x2, x2; \
67 vpshufb t4, x5, x5; \
68 vpshufb t4, x3, x3; \
69 vpshufb t4, x6, x6; \
[all …]
/kernel/linux/linux-5.10/lib/zlib_dfltcc/
Ddfltcc_util.h38 const Byte *t4 = op2 ? *op2 : NULL; in dfltcc() local
44 register const Byte *r4 __asm__("r4") = t4; in dfltcc()
60 t2 = r2; t3 = r3; t4 = r4; t5 = r5; in dfltcc()
67 *op2 = t4; in dfltcc()
/kernel/linux/linux-5.10/arch/mips/lib/
Dcsum_partial.S33 #define t4 $12 macro
182 CSUM_BIGCHUNK1(src, 0x00, sum, t0, t1, t3, t4)
193 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4)
194 CSUM_BIGCHUNK(src, 0x20, sum, t0, t1, t3, t4)
195 CSUM_BIGCHUNK(src, 0x40, sum, t0, t1, t3, t4)
196 CSUM_BIGCHUNK(src, 0x60, sum, t0, t1, t3, t4)
208 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4)
209 CSUM_BIGCHUNK(src, 0x20, sum, t0, t1, t3, t4)
217 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4)
476 LOAD(t4, UNIT(4)(src))
[all …]
/kernel/linux/linux-5.10/scripts/
Dmakelst27 t4=`field 1 $t3`
29 t6=`printf "%lu" $((0x$t4 - 0x$t5))`
/kernel/linux/linux-5.10/arch/riscv/kernel/
Dmcount.S82 la t4, ftrace_stub
86 bne t1, t4, do_ftrace_graph_caller
95 bne t5, t4, do_trace
/kernel/linux/linux-5.10/arch/sparc/lib/
Dblockops.S28 #define MIRROR_BLOCK(dst, src, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument
31 ldd [src + offset + 0x08], t4; \
35 std t4, [dst + offset + 0x08]; \
/kernel/linux/linux-5.10/arch/s390/include/asm/
Dstacktrace.h141 #define CALL_LARGS_4(t1, a1, t2, a2, t3, a3, t4, a4) \ argument
143 long arg4 = (long)(t4)(a4)
144 #define CALL_LARGS_5(t1, a1, t2, a2, t3, a3, t4, a4, t5, a5) \ argument
145 CALL_LARGS_4(t1, a1, t2, a2, t3, a3, t4, a4); \
/kernel/linux/linux-5.10/arch/mips/kernel/
Dscall32-o32.S52 addu t4, t0, 32
53 and t5, t4
195 lw t4, 20(sp)
198 sw t4, 16(sp)
/kernel/linux/linux-5.10/drivers/media/pci/cx88/
Dcx88-dsp.c72 u32 t2, t4, t6, t8; in int_cos() local
86 t4 = t2 * x / 32768 * x / 32768 / 3 / 4; in int_cos()
87 t6 = t4 * x / 32768 * x / 32768 / 5 / 6; in int_cos()
89 ret = 32768 - t2 + t4 - t6 + t8; in int_cos()

123