/arch/x86/crypto/ |
D | nh-avx2-x86_64.S | 24 #define T0 %ymm8 macro 42 vpaddd \k0, T3, T0 48 vpshufd $0x10, T0, T4 49 vpshufd $0x32, T0, T0 56 vpmuludq T4, T0, T0 60 vpaddq T0, PASS0_SUMS, PASS0_SUMS 142 vpunpcklqdq PASS1_SUMS, PASS0_SUMS, T0 // T0 = (0A 1A 0C 1C) 147 vinserti128 $0x1, T2_XMM, T0, T4 // T4 = (0A 1A 2A 3A) 149 vperm2i128 $0x31, T2, T0, T0 // T0 = (0C 1C 2C 3C) 153 vpaddq T1, T0, T0 [all …]
|
D | aegis128-aesni-asm.S | 19 #define T0 %xmm6 macro 54 movdqa STATE4, T0 59 aesenc T0, STATE3 121 movq (%r8), T0 122 pxor T0, MSG 144 movq T0, %r10 150 psrldq $8, T0 151 movq T0, %r10 385 movdqa MSG, T0 386 pxor \s1, T0 [all …]
|
D | nh-sse2-x86_64.S | 20 #define T0 %xmm8 macro 112 movdqa PASS0_SUMS, T0 114 punpcklqdq PASS1_SUMS, T0 // => (PASS0_SUM_A PASS1_SUM_A) 118 paddq PASS0_SUMS, T0 120 movdqu T0, 0x00(HASH)
|
D | twofish-x86_64-asm_64-3way.S | 77 #define do16bit_ror(rot, op1, op2, T0, T1, tmp1, tmp2, ab, dst) \ argument 81 op1##l T0(CTX, tmp2, 4), dst ## d; \
|
D | camellia-x86_64-asm_64.S | 79 #define xor2ror16(T0, T1, tmp1, tmp2, ab, dst) \ argument 83 xorq T0(, tmp2, 8), dst; \
|
/arch/mips/kvm/ |
D | entry.c | 31 #define T0 8 macro 38 #define T0 12 macro 305 UASM_i_LW(&p, T0, offsetof(struct kvm_vcpu_arch, pc), K1); in kvm_mips_build_enter_guest() 306 UASM_i_MTC0(&p, T0, C0_EPC); in kvm_mips_build_enter_guest() 345 uasm_i_mfc0(&p, T0, C0_GUESTCTL1); in kvm_mips_build_enter_guest() 347 uasm_i_ext(&p, T1, T0, MIPS_GCTL1_ID_SHIFT, in kvm_mips_build_enter_guest() 349 uasm_i_ins(&p, T0, T1, MIPS_GCTL1_RID_SHIFT, in kvm_mips_build_enter_guest() 351 uasm_i_mtc0(&p, T0, C0_GUESTCTL1); in kvm_mips_build_enter_guest() 369 UASM_i_LW(&p, T0, offsetof(struct kvm_vcpu_arch, cop0), K1); in kvm_mips_build_enter_guest() 370 UASM_i_LW(&p, T0, offsetof(struct mips_coproc, reg[MIPS_CP0_STATUS][0]), in kvm_mips_build_enter_guest() [all …]
|
/arch/arm64/crypto/ |
D | sha512-armv8.pl | 109 my ($T0,$T1,$T2)=(@X[($i-8)&15],@X[($i-9)&15],@X[($i-10)&15]); 110 $T0=@X[$i+3] if ($i<11); 142 eor $T0,$e,$e,ror#`$Sigma1[2]-$Sigma1[1]` 148 eor $t0,$t0,$T0,ror#$Sigma1[1] // Sigma1(e) 149 ror $T0,$a,#$Sigma0[0] 156 eor $t1,$T0,$t1,ror#$Sigma0[1] // Sigma0(a) 168 ror $T0,$a,#$Sigma0[0] 175 eor $T0,$T0,$a,ror#$Sigma0[1] 182 eor $t1,$T0,$a,ror#$Sigma0[2] // Sigma0(a) 463 my ($T0,$T1,$T2,$T3,$T4,$T5,$T6,$T7) = map("q$_",(4..7,16..19)); [all …]
|
D | nh-neon-core.S | 25 T0 .req v8 43 add T0.4s, T3.4s, \k0\().4s 49 mov T4.d[0], T0.d[1] 53 umlal PASS0_SUMS.2d, T0.2s, T4.2s 99 addp T0.2d, PASS0_SUMS.2d, PASS1_SUMS.2d 101 st1 {T0.16b,T1.16b}, [HASH]
|
/arch/sparc/crypto/ |
D | aes_asm.S | 7 #define ENCRYPT_TWO_ROUNDS(KEY_BASE, I0, I1, T0, T1) \ argument 8 AES_EROUND01(KEY_BASE + 0, I0, I1, T0) \ 10 AES_EROUND01(KEY_BASE + 4, T0, T1, I0) \ 11 AES_EROUND23(KEY_BASE + 6, T0, T1, I1) 13 #define ENCRYPT_TWO_ROUNDS_2(KEY_BASE, I0, I1, I2, I3, T0, T1, T2, T3) \ argument 14 AES_EROUND01(KEY_BASE + 0, I0, I1, T0) \ 18 AES_EROUND01(KEY_BASE + 4, T0, T1, I0) \ 19 AES_EROUND23(KEY_BASE + 6, T0, T1, I1) \ 23 #define ENCRYPT_TWO_ROUNDS_LAST(KEY_BASE, I0, I1, T0, T1) \ argument 24 AES_EROUND01(KEY_BASE + 0, I0, I1, T0) \ [all …]
|
/arch/arm/crypto/ |
D | sha256-armv4.pl | 292 my ($T0,$T1,$T2,$T3,$T4,$T5)=("q8","q9","q10","q11","d24","d25"); 312 &vext_8 ($T0,@X[0],@X[1],4); # X[1..4] 320 &vshr_u32 ($T2,$T0,$sigma0[0]); 326 &vshr_u32 ($T1,$T0,$sigma0[2]); 329 &vsli_32 ($T2,$T0,32-$sigma0[0]); 332 &vshr_u32 ($T3,$T0,$sigma0[1]); 338 &vsli_32 ($T3,$T0,32-$sigma0[1]); 386 &vld1_32 ("{$T0}","[$Ktbl,:128]!"); 398 &vadd_i32 ($T0,$T0,@X[0]); 400 &vst1_32 ("{$T0}","[$Xfer,:128]!"); [all …]
|
D | nh-neon-core.S | 36 T0 .req q8 58 vadd.u32 T0, T3, \k0 114 vst1.8 {T0-T1}, [HASH]
|
/arch/mips/mm/ |
D | page.c | 45 #define T0 8 macro 475 build_copy_load(&buf, T0, off); in build_copy_page() 483 build_copy_store(&buf, T0, off); in build_copy_page() 497 build_copy_load(&buf, T0, off); in build_copy_page() 505 build_copy_store(&buf, T0, off); in build_copy_page() 523 build_copy_load(&buf, T0, off); in build_copy_page() 528 build_copy_store(&buf, T0, off); in build_copy_page() 541 build_copy_load(&buf, T0, off); in build_copy_page() 546 build_copy_store(&buf, T0, off); in build_copy_page() 565 build_copy_load(&buf, T0, off); in build_copy_page() [all …]
|
/arch/x86/lib/ |
D | x86-opcode-map.txt | 1031 1: prefetch T0
|