/external/chromium_org/third_party/openssl/openssl/crypto/modes/asm/ |
D | ghash-x86_64.pl | 64 $Xi="%rdi"; 173 movzb 15($Xi),$Zlo 176 &loop ($Xi); 178 mov $Zlo,8($Xi) 179 mov $Zhi,($Xi) 241 &mov ($Zlo,"8($Xi)"); 242 &mov ($Zhi,"0($Xi)"); 252 &mov ("($Xi)",$Zhi); 253 &mov ("8($Xi)","%rdx"); 288 &mov ($dat,"$j($Xi)") if (--$j%4==0); [all …]
|
D | ghash-x86.pl | 818 ($Xi,$Xhi)=("xmm0","xmm1"); $Hkey="xmm2"; 825 my ($Xhi,$Xi,$Hkey)=@_; 827 &movdqa ($Xhi,$Xi); # 828 &pshufd ($T1,$Xi,0b01001110); 830 &pxor ($T1,$Xi); # 833 &pclmulqdq ($Xi,$Hkey,0x00); ####### 836 &xorps ($T1,$Xi); # 843 &pxor ($Xi,$T2); # 852 my ($Xhi,$Xi,$Hkey)=@_; 854 &movdqa ($T1,$Xi); # [all …]
|
D | ghash-armv4.pl | 63 $Xi="r0"; # argument block 91 str $_,[$Xi,#$i] 93 str $_,[$Xi,#$i] 96 strb $_,[$Xi,#$i+3] 98 strb $Tlh,[$Xi,#$i+2] 100 strb $Thl,[$Xi,#$i+1] 101 strb $Thh,[$Xi,#$i] 144 ldrb $nhi,[$Xi,#15] 162 ldrb $nhi,[$Xi,#14] 195 ldrplb $Tll,[$Xi,$cnt] [all …]
|
D | ghash-s390x.pl | 58 $Xi="%r2"; # argument block 96 la %r1,0($Xi) # H lies right after Xi in gcm128_context 108 aghi $Xi,-1 113 lg $Zlo,8+1($Xi) # Xi 134 la %r1,0($Xi) # H lies right after Xi in gcm128_context 147 aghi $Xi,-1 152 lg $Zlo,8+1($Xi) # Xi 153 lg $Zhi,0+1($Xi) 159 stg $Zlo,8+1($Xi) 160 stg $Zhi,0+1($Xi) [all …]
|
/external/openssl/crypto/modes/asm/ |
D | ghash-x86_64.pl | 64 $Xi="%rdi"; 173 movzb 15($Xi),$Zlo 176 &loop ($Xi); 178 mov $Zlo,8($Xi) 179 mov $Zhi,($Xi) 241 &mov ($Zlo,"8($Xi)"); 242 &mov ($Zhi,"0($Xi)"); 252 &mov ("($Xi)",$Zhi); 253 &mov ("8($Xi)","%rdx"); 288 &mov ($dat,"$j($Xi)") if (--$j%4==0); [all …]
|
D | ghash-x86.pl | 818 ($Xi,$Xhi)=("xmm0","xmm1"); $Hkey="xmm2"; 825 my ($Xhi,$Xi,$Hkey)=@_; 827 &movdqa ($Xhi,$Xi); # 828 &pshufd ($T1,$Xi,0b01001110); 830 &pxor ($T1,$Xi); # 833 &pclmulqdq ($Xi,$Hkey,0x00); ####### 836 &xorps ($T1,$Xi); # 843 &pxor ($Xi,$T2); # 852 my ($Xhi,$Xi,$Hkey)=@_; 854 &movdqa ($T1,$Xi); # [all …]
|
D | ghash-s390x.pl | 58 $Xi="%r2"; # argument block 96 la %r1,0($Xi) # H lies right after Xi in gcm128_context 108 aghi $Xi,-1 113 lg $Zlo,8+1($Xi) # Xi 134 la %r1,0($Xi) # H lies right after Xi in gcm128_context 147 aghi $Xi,-1 152 lg $Zlo,8+1($Xi) # Xi 153 lg $Zhi,0+1($Xi) 159 stg $Zlo,8+1($Xi) 160 stg $Zhi,0+1($Xi) [all …]
|
D | ghash-armv4.pl | 77 $Xi="r0"; # argument block 105 str $_,[$Xi,#$i] 107 str $_,[$Xi,#$i] 110 strb $_,[$Xi,#$i+3] 112 strb $Tlh,[$Xi,#$i+2] 114 strb $Thl,[$Xi,#$i+1] 115 strb $Thh,[$Xi,#$i] 158 ldrb $nhi,[$Xi,#15] 176 ldrb $nhi,[$Xi,#14] 209 ldrplb $Tll,[$Xi,$cnt] [all …]
|
D | ghash-sparcv9.pl | 64 $Xi="%i0"; # input argument block 86 ldub [$Xi+15],$xi0 87 ldub [$Xi+14],$xi1 136 ldub [$Xi+$cnt],$xi1 188 stx $Zlo,[$Xi+8] 190 stx $Zhi,[$Xi] 208 stx $Zlo,[$Xi+8] 210 stx $Zhi,[$Xi] 226 ldub [$Xi+15],$nlo 238 ldub [$Xi+14],$nlo [all …]
|
/external/chromium_org/third_party/openssl/openssl/crypto/modes/ |
D | gcm128.c | 145 static void gcm_gmult_8bit(u64 Xi[2], const u128 Htable[256]) in gcm_gmult_8bit() 148 const u8 *xi = (const u8 *)Xi+15; in gcm_gmult_8bit() 221 if ((u8 *)Xi==xi) break; in gcm_gmult_8bit() 236 Xi[0] = BSWAP8(Z.hi); in gcm_gmult_8bit() 237 Xi[1] = BSWAP8(Z.lo); in gcm_gmult_8bit() 239 u8 *p = (u8 *)Xi; in gcm_gmult_8bit() 248 Xi[0] = Z.hi; in gcm_gmult_8bit() 249 Xi[1] = Z.lo; in gcm_gmult_8bit() 252 #define GCM_MUL(ctx,Xi) gcm_gmult_8bit(ctx->Xi.u,ctx->Htable) argument 335 static void gcm_gmult_4bit(u64 Xi[2], const u128 Htable[16]) in gcm_gmult_4bit() [all …]
|
/external/openssl/crypto/modes/ |
D | gcm128.c | 145 static void gcm_gmult_8bit(u64 Xi[2], const u128 Htable[256]) in gcm_gmult_8bit() 148 const u8 *xi = (const u8 *)Xi+15; in gcm_gmult_8bit() 221 if ((u8 *)Xi==xi) break; in gcm_gmult_8bit() 236 Xi[0] = BSWAP8(Z.hi); in gcm_gmult_8bit() 237 Xi[1] = BSWAP8(Z.lo); in gcm_gmult_8bit() 239 u8 *p = (u8 *)Xi; in gcm_gmult_8bit() 248 Xi[0] = Z.hi; in gcm_gmult_8bit() 249 Xi[1] = Z.lo; in gcm_gmult_8bit() 252 #define GCM_MUL(ctx,Xi) gcm_gmult_8bit(ctx->Xi.u,ctx->Htable) argument 335 static void gcm_gmult_4bit(u64 Xi[2], const u128 Htable[16]) in gcm_gmult_4bit() [all …]
|
/external/openssl/crypto/sha/asm/ |
D | sha1-armv4-large.pl | 85 $Xi="r14"; 91 ldr $t0,[$Xi,#15*4] 92 ldr $t1,[$Xi,#13*4] 93 ldr $t2,[$Xi,#7*4] 95 ldr $t3,[$Xi,#2*4] 102 str $t0,[$Xi,#-4]! 135 str $t0,[$Xi,#-4]! 191 mov $Xi,sp 202 teq $Xi,sp 221 teq $Xi,sp @ preserve carry [all …]
|
D | sha1-sparcv9.pl | 32 $Xi="%g4"; 54 my $xi=($i&1)?@X[($i/2)%8]:$Xi; 72 " srlx @X[(($i+1)/2)%8],32,$Xi\n"; 91 sllx @X[($j+6)%8],32,$Xi ! Xupdate($i) 96 or $tmp1,$Xi,$Xi 98 xor $Xi,@X[$j%8],@X[$j%8] 99 srlx @X[$j%8],31,$Xi 101 and $Xi,$rot1m,$Xi 104 or $Xi,@X[$j%8],@X[$j%8] 116 $xi=$Xi; [all …]
|
D | sha1-thumb.pl | 40 $Xi="r12"; 133 mov $Xi,sp 170 mov $t0,$Xi 175 mov $Xi,$t1 181 cmp $Xi,$t0 187 mov $Xi,$t1 193 cmp $Xi,$t0 201 mov $Xi,$t1 207 cmp $Xi,$t0 211 mov $Xi,sp
|
D | sha1-586.pl | 415 my $Xi=4; # 4xSIMD Xupdate round, start pre-seeded 539 &movdqa (&QWP(64+16*(($Xi-4)%3),"esp"),@X[-4&7]);# save X[] to backtrace buffer 558 &movdqa (&QWP(0+16*(($Xi-1)&3),"esp"),@X[3]); # X[]+K xfer to IALU 587 &movdqa (@X[2],&QWP(64+16*(($Xi-6)%3),"esp")) if ($Xi>5); # restore X[] from backtrace buffer 595 &movdqa (@X[4],&QWP(112-16+16*(($Xi)/5),"esp")); # K_XX_XX 600 &movdqa (@X[1],@X[-2&7]) if ($Xi<7); 606 $Xi++; push(@X,shift(@X)); # "rotate" X[] 615 &movdqa (@X[2],@X[-1&7]) if ($Xi==8); 624 &movdqa (&QWP(64+16*(($Xi-4)%3),"esp"),@X[-4&7]); # save X[] to backtrace buffer 627 if ($Xi%5) { [all …]
|
D | sha1-x86_64.pl | 295 my $Xi=4; 403 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 439 &movdqa (@Tx[2],eval(16*(($Xi)/5))."($K_XX_XX)"); # K_XX_XX 447 $Xi++; push(@X,shift(@X)); # "rotate" X[] 457 &movdqa (@Tx[0],@X[-1&7]) if ($Xi==8); 468 if ($Xi%5) { 471 &movdqa (@Tx[2],eval(16*($Xi/5))."($K_XX_XX)"); 484 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 504 &movdqa (@Tx[1],@X[0]) if ($Xi<19); 514 $Xi++; push(@X,shift(@X)); # "rotate" X[] [all …]
|
D | sha1-sparcv9a.pl | 58 $Xi="%o7"; 152 ld [$Xfer+`4*($i%16)`],$Xi 160 add $Xi,$e,$e 171 ld [$Xfer+`4*($i%16)`],$Xi 180 add $Xi,$e,$e 204 ld [$Xfer+`4*($i%16)`],$Xi 217 add $Xi,$e,$e 222 ld [$Xfer+`4*($i%16)`],$Xi 237 add $Xi,$e,$e 242 ld [$Xfer+`4*($i%16)`],$Xi [all …]
|
/external/chromium_org/third_party/openssl/openssl/crypto/sha/asm/ |
D | sha1-sparcv9.pl | 32 $Xi="%g4"; 54 my $xi=($i&1)?@X[($i/2)%8]:$Xi; 72 " srlx @X[(($i+1)/2)%8],32,$Xi\n"; 91 sllx @X[($j+6)%8],32,$Xi ! Xupdate($i) 96 or $tmp1,$Xi,$Xi 98 xor $Xi,@X[$j%8],@X[$j%8] 99 srlx @X[$j%8],31,$Xi 101 and $Xi,$rot1m,$Xi 104 or $Xi,@X[$j%8],@X[$j%8] 116 $xi=$Xi; [all …]
|
D | sha1-thumb.pl | 40 $Xi="r12"; 133 mov $Xi,sp 170 mov $t0,$Xi 175 mov $Xi,$t1 181 cmp $Xi,$t0 187 mov $Xi,$t1 193 cmp $Xi,$t0 201 mov $Xi,$t1 207 cmp $Xi,$t0 211 mov $Xi,sp
|
D | sha1-armv4-large.pl | 71 $Xi="r14"; 77 ldr $t0,[$Xi,#15*4] 78 ldr $t1,[$Xi,#13*4] 79 ldr $t2,[$Xi,#7*4] 81 ldr $t3,[$Xi,#2*4] 88 str $t0,[$Xi,#-4]! 121 str $t0,[$Xi,#-4]! 167 mov $Xi,sp 178 teq $Xi,sp 197 teq $Xi,sp @ preserve carry [all …]
|
D | sha1-586.pl | 415 my $Xi=4; # 4xSIMD Xupdate round, start pre-seeded 539 &movdqa (&QWP(64+16*(($Xi-4)%3),"esp"),@X[-4&7]);# save X[] to backtrace buffer 558 &movdqa (&QWP(0+16*(($Xi-1)&3),"esp"),@X[3]); # X[]+K xfer to IALU 587 &movdqa (@X[2],&QWP(64+16*(($Xi-6)%3),"esp")) if ($Xi>5); # restore X[] from backtrace buffer 595 &movdqa (@X[4],&QWP(112-16+16*(($Xi)/5),"esp")); # K_XX_XX 600 &movdqa (@X[1],@X[-2&7]) if ($Xi<7); 606 $Xi++; push(@X,shift(@X)); # "rotate" X[] 615 &movdqa (@X[2],@X[-1&7]) if ($Xi==8); 624 &movdqa (&QWP(64+16*(($Xi-4)%3),"esp"),@X[-4&7]); # save X[] to backtrace buffer 627 if ($Xi%5) { [all …]
|
D | sha1-x86_64.pl | 295 my $Xi=4; 403 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 439 &movdqa (@Tx[2],eval(16*(($Xi)/5))."($K_XX_XX)"); # K_XX_XX 447 $Xi++; push(@X,shift(@X)); # "rotate" X[] 457 &movdqa (@Tx[0],@X[-1&7]) if ($Xi==8); 468 if ($Xi%5) { 471 &movdqa (@Tx[2],eval(16*($Xi/5))."($K_XX_XX)"); 484 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 504 &movdqa (@Tx[1],@X[0]) if ($Xi<19); 514 $Xi++; push(@X,shift(@X)); # "rotate" X[] [all …]
|
D | sha1-sparcv9a.pl | 58 $Xi="%o7"; 152 ld [$Xfer+`4*($i%16)`],$Xi 160 add $Xi,$e,$e 171 ld [$Xfer+`4*($i%16)`],$Xi 180 add $Xi,$e,$e 204 ld [$Xfer+`4*($i%16)`],$Xi 217 add $Xi,$e,$e 222 ld [$Xfer+`4*($i%16)`],$Xi 237 add $Xi,$e,$e 242 ld [$Xfer+`4*($i%16)`],$Xi [all …]
|
/external/openssl/crypto/aes/asm/ |
D | aesni-sha1-x86_64.pl | 110 my $Xi=4; 285 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 321 &movdqa (@Tx[2],eval(16*(($Xi)/5))."($K_XX_XX)"); # K_XX_XX 329 $Xi++; push(@X,shift(@X)); # "rotate" X[] 339 &movdqa (@Tx[0],@X[-1&7]) if ($Xi==8); 350 if ($Xi%5) { 353 &movdqa (@Tx[2],eval(16*($Xi/5))."($K_XX_XX)"); 366 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 386 &movdqa (@Tx[1],@X[0]) if ($Xi<19); 396 $Xi++; push(@X,shift(@X)); # "rotate" X[] [all …]
|
/external/chromium_org/third_party/openssl/openssl/crypto/aes/asm/ |
D | aesni-sha1-x86_64.pl | 110 my $Xi=4; 285 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 321 &movdqa (@Tx[2],eval(16*(($Xi)/5))."($K_XX_XX)"); # K_XX_XX 329 $Xi++; push(@X,shift(@X)); # "rotate" X[] 339 &movdqa (@Tx[0],@X[-1&7]) if ($Xi==8); 350 if ($Xi%5) { 353 &movdqa (@Tx[2],eval(16*($Xi/5))."($K_XX_XX)"); 366 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 386 &movdqa (@Tx[1],@X[0]) if ($Xi<19); 396 $Xi++; push(@X,shift(@X)); # "rotate" X[] [all …]
|