/third_party/openssl/crypto/modes/asm/ |
D | ghash-parisc.pl | 76 $nlo="%r3"; 131 ldb 15($Xi),$nlo 134 and $mask0xf0,$nlo,$nhi 135 depd,z $nlo,59,4,$nlo 137 ldd $nlo($Hll),$Zll 138 ldd $nlo($Hhh),$Zhh 143 ldb 14($Xi),$nlo 147 and $mask0xf0,$nlo,$nhi 148 depd,z $nlo,59,4,$nlo 163 ldd $nlo($Hll),$Tll [all …]
|
D | ghash-alpha.pl | 45 $nlo="a4"; # $20 60 extbl $Xlo,7,$nlo 61 and $nlo,0xf0,$nhi 62 sll $nlo,4,$nlo 63 and $nlo,0xf0,$nlo 65 addq $nlo,$Htbl,$nlo 66 ldq $Zlo,8($nlo) 68 ldq $Zhi,0($nlo) 73 extbl $Xlo,6,$nlo 83 and $nlo,0xf0,$nhi [all …]
|
D | ghash-sparcv9.pl | 70 $nlo="%l1"; 106 ldub [$inp+15],$nlo 116 xor $xi0,$nlo,$nlo 117 and $nlo,0xf0,$nhi 118 and $nlo,0x0f,$nlo 119 sll $nlo,4,$nlo 120 ldx [$Htblo+$nlo],$Zlo 121 ldx [$Htbl+$nlo],$Zhi 123 ldub [$inp+14],$nlo 137 xor $xi1,$nlo,$nlo [all …]
|
D | ghash-armv4.pl | 109 $nlo="r12"; 193 ldrb $nlo,[$inp,#15] 196 eor $nlo,$nlo,$nhi 197 and $nhi,$nlo,#0xf0 198 and $nlo,$nlo,#0x0f 201 add $Zhh,$Htbl,$nlo,lsl#4 204 ldrb $nlo,[$inp,#14] 218 eor $nlo,$nlo,$nhi 219 and $nhi,$nlo,#0xf0 220 and $nlo,$nlo,#0x0f [all …]
|
D | ghash-s390x.pl | 72 $nlo="%r8"; 164 sllg $nlo,$Zlo,4 166 ngr $nlo,$tmp 171 lg $Zlo,8($nlo,$Htbl) 172 lg $Zhi,0($nlo,$Htbl) 174 sllg $nlo,$xi,4 176 ngr $nlo,$tmp 195 xg $Zlo,8($nlo,$Htbl) 197 xg $Zhi,0($nlo,$Htbl) 198 sllg $nlo,$xi,4 [all …]
|
D | ghash-x86.pl | 376 my $nlo = $Zlh; 381 &xor ($nlo,$nlo); # avoid partial register stalls on PIII 383 &mov (&LB($nlo),&LB($nhi)); 384 &shl (&LB($nlo),4); 386 &movq ($Zlo,&QWP(8,$Htbl,$nlo)); 387 &movq ($Zhi,&QWP(0,$Htbl,$nlo)); 392 my $nix = $odd ? $nlo : $nhi; 394 &shl (&LB($nlo),4) if ($odd); 399 &mov (&LB($nlo),&BP($cnt/2,$inp)) if (!$odd && $cnt>=0); 405 &mov ($nhi,$nlo) if (!$odd && $cnt>=0); [all …]
|
D | ghash-x86_64.pl | 129 $nlo="%rax"; 161 xor $nlo,$nlo 163 mov `&LB("$Zlo")`,`&LB("$nlo")` 165 shl \$4,`&LB("$nlo")` 167 mov 8($Htbl,$nlo),$Zlo 168 mov ($Htbl,$nlo),$Zhi 178 mov ($inp,$cnt),`&LB("$nlo")` 183 mov `&LB("$nlo")`,`&LB("$nhi")` 186 shl \$4,`&LB("$nlo")` 195 xor 8($Htbl,$nlo),$Zlo [all …]
|
/third_party/openssl/crypto/bn/asm/ |
D | s390x-mont.pl | 86 $nlo="%r11"; 141 lg $nlo,0($np) # 142 _dswap $nlo 144 algr $nlo,$alo # +="tp[0]" 160 lg $nlo,0($j,$np) 161 _dswap $nlo 163 algr $nlo,$NHI 166 algr $nlo,$alo 169 stg $nlo,$stdframe-8($j,$sp) # tp[j-1]= 193 lg $nlo,0($np) # np[0] [all …]
|
D | sparcv9a-mont.pl | 115 $nlo="%f20"; $nlo_="%f21"; $nhi="%f22"; $nhi_="%f23"; 203 fzeros $nlo 213 fxtod $nlo,$nlo 231 std $nlo,[$np_l+$j] ! save smashed np[j] in double format 237 fmuld $nlo,$na,$nloa 239 fmuld $nlo,$nb,$nlob 242 fmuld $nlo,$nc,$nloc 245 fmuld $nlo,$nd,$nlod 283 fzeros $nlo 289 fxtod $nlo,$nlo [all …]
|
D | armv4-mont.pl | 84 $nlo="r12"; # ip 167 mov $nlo,#0 168 umlal $alo,$nlo,$nj,$n0 @ np[0]*n0+"t[0]" 178 umlal $nlo,$nhi,$nj,$n0 @ np[j]*n0 179 adds $nlo,$nlo,$alo 180 str $nlo,[$tp],#4 @ tp[j-1]=,tp++ 181 adc $nlo,$nhi,#0 185 adds $nlo,$nlo,$ahi 190 str $nlo,[$num] @ tp[num-1]= 208 mov $nlo,#0 [all …]
|
D | alpha-mont.pl | 46 $nlo="t10"; 116 mulq $nj,$m1,$nlo 131 addq $nlo,$hi1,$lo1 133 mulq $nj,$m1,$nlo 155 addq $nlo,$hi1,$lo1 204 mulq $nj,$m1,$nlo 222 addq $nlo,$hi1,$lo1 #L1 225 mulq $nj,$m1,$nlo #U1 248 addq $nlo,$hi1,$lo1
|
D | vis3-mont.pl | 45 ($n0,$m0,$m1,$lo0,$hi0, $lo1,$hi1,$aj,$alo,$nj,$nlo,$tj)= 148 mulx $nj, $m1, $nlo ! np[1]*m1 167 addcc $nlo, $hi1, $lo1 179 mulx $nj, $m1, $nlo ! np[j]*m1 192 addcc $nlo, $hi1, $lo1 236 mulx $nj, $m1, $nlo ! np[1]*m1 248 addcc $nlo, $hi1, $lo1 255 mulx $nj, $m1, $nlo ! np[j]*m1 271 addcc $nlo, $hi1, $lo1
|
D | mips-mont.pl | 114 $nlo=$s6; 207 mflo ($nlo,$nhi,$m1) 222 $ADDU $lo1,$nlo,$hi1 237 mflo ($nlo,$nj,$m1) 248 $ADDU $lo1,$nlo,$hi1 294 mflo ($nlo,$nhi,$m1) 310 $ADDU $lo1,$nlo,$hi1 328 mflo ($nlo,$nj,$m1) 343 $ADDU $lo1,$nlo,$hi1
|
D | sparct4-mont.pl | 669 ($n0,$m0,$m1,$lo0,$hi0, $lo1,$hi1,$aj,$alo,$nj,$nlo,$tj)= 738 mulx $nj, $m1, $nlo ! np[1]*m1 750 addcc $nlo, $hi1, $lo1 759 mulx $nj, $m1, $nlo ! np[j]*m1 772 addcc $nlo, $hi1, $lo1 814 mulx $nj, $m1, $nlo ! np[1]*m1 827 addcc $nlo, $hi1, $lo1 834 mulx $nj, $m1, $nlo ! np[j]*m1 850 addcc $nlo, $hi1, $lo1 983 mulx $nj, $m1, $nlo ! np[1]*m1 [all …]
|
D | armv8-mont.pl | 55 $lo1,$hi1,$nj,$m1,$nlo,$nhi, 103 mul $nlo,$nj,$m1 // np[1]*m1 126 adds $lo1,$nlo,$hi1 132 mul $nlo,$nj,$m1 // np[j]*m1 143 adds $lo1,$nlo,$hi1 174 mul $nlo,$nj,$m1 // np[1]*m1 188 adds $lo1,$nlo,$hi1 197 mul $nlo,$nj,$m1 // np[j]*m1 210 adds $lo1,$nlo,$hi1
|
D | ppc-mont.pl | 126 my $nlo="r31"; 191 $UMULL $nlo,$nj,$m1 ; np[1]*m1 203 addc $lo1,$nlo,$hi1 206 $UMULL $nlo,$nj,$m1 ; np[j]*m1 219 addc $lo1,$nlo,$hi1 250 $UMULL $nlo,$nj,$m1 ; np[1]*m1 263 addc $lo1,$nlo,$hi1 268 $UMULL $nlo,$nj,$m1 ; np[j]*m1 284 addc $lo1,$nlo,$hi1
|
/third_party/openssl/crypto/modes/ |
D | gcm128.c | 308 size_t rem, nlo, nhi; in gcm_gmult_4bit() local 314 nlo = ((const u8 *)Xi)[15]; in gcm_gmult_4bit() 315 nhi = nlo >> 4; in gcm_gmult_4bit() 316 nlo &= 0xf; in gcm_gmult_4bit() 318 Z.hi = Htable[nlo].hi; in gcm_gmult_4bit() 319 Z.lo = Htable[nlo].lo; in gcm_gmult_4bit() 336 nlo = ((const u8 *)Xi)[cnt]; in gcm_gmult_4bit() 337 nhi = nlo >> 4; in gcm_gmult_4bit() 338 nlo &= 0xf; in gcm_gmult_4bit() 348 Z.hi ^= Htable[nlo].hi; in gcm_gmult_4bit() [all …]
|
/third_party/openssl/crypto/modes/asm/arm32/ |
D | ghash-armv4.S | 60 ldmia r7,{r4-r7} @ load Htbl[nlo] 86 ldmia r11,{r8-r11} @ load Htbl[nlo] 221 ldmia r7,{r4-r7} @ load Htbl[nlo] 245 ldmia r11,{r8-r11} @ load Htbl[nlo]
|
/third_party/icu/icu4c/source/data/misc/ |
D | supplementalData.txt | 6288 "nlo",
|
/third_party/skia/third_party/externals/icu/source/data/misc/ |
D | supplementalData.txt | 6288 "nlo",
|