/external/boringssl/src/crypto/aes/asm/ |
D | bsaes-x86_64.pl | 105 my ($inp,$out,$len,$key,$ivp)=("%rdi","%rsi","%rdx","%rcx"); 935 my ($out,$inp,$rounds,$const)=("%rax","%rcx","%r10d","%r11"); 966 movdqu ($inp), %xmm7 # load round 0 key 967 lea 0x10($inp), $inp 975 movdqu ($inp), %xmm6 # load round 1 key 1018 lea 0x10($inp), $inp 1028 movdqu ($inp), %xmm6 # load next round key 1054 mov 240($inp),%r10d # pass rounds 1055 mov $inp,%rcx # pass key 1068 movdqu 0x00($inp), @XMM[0] # load input [all …]
|
D | aesni-x86_64.pl | 202 $inp="%rdi"; 254 { my ($inp,$out,$key) = @_4args; 261 movups ($inp),$inout0 # load input 277 movups ($inp),$inout0 # load input 608 movdqu ($inp),$inout0 # load 8 input blocks 609 movdqu 0x10($inp),$inout1 610 movdqu 0x20($inp),$inout2 611 movdqu 0x30($inp),$inout3 612 movdqu 0x40($inp),$inout4 613 movdqu 0x50($inp),$inout5 [all …]
|
D | aesni-x86.pl | 77 $inp="esi"; 415 &mov ($inp,&wparam(0)); 431 &movdqu ($inout0,&QWP(0,$inp)); 432 &movdqu ($inout1,&QWP(0x10,$inp)); 433 &movdqu ($inout2,&QWP(0x20,$inp)); 434 &movdqu ($inout3,&QWP(0x30,$inp)); 435 &movdqu ($inout4,&QWP(0x40,$inp)); 436 &movdqu ($inout5,&QWP(0x50,$inp)); 437 &lea ($inp,&DWP(0x60,$inp)); 443 &movdqu ($inout0,&QWP(0,$inp)); [all …]
|
D | bsaes-armv7.pl | 65 my ($inp,$out,$len,$key)=("r0","r1","r2","r3"); 917 my ($out,$inp,$rounds,$const)=("r12","r4","r5","r6"); 948 vld1.8 {@XMM[7]}, [$inp]! @ load round 0 key 954 vld1.8 {@XMM[15]}, [$inp]! @ load round 1 key 987 vld1.8 {@XMM[15]}, [$inp]! @ load next round key 1017 ldr r5,[$inp,#240] @ pass rounds 1018 mov r4,$inp @ pass key 1036 vld1.8 {@XMM[0]-@XMM[1]}, [$inp]! @ load input 1037 vld1.8 {@XMM[2]-@XMM[3]}, [$inp]! 1039 vld1.8 {@XMM[4]-@XMM[5]}, [$inp]! [all …]
|
D | aesv8-armx.pl | 68 my ($inp,$bits,$out,$ptr,$rounds)=("x0","w1","x2","x3","w12"); 92 cmp $inp,#0 108 vld1.8 {$in0},[$inp],#16 170 vld1.8 {$in1},[$inp],#8 205 vld1.8 {$in1},[$inp] 271 add $inp,$out,x12,lsl#4 // end of key schedule 274 vld1.32 {v1.16b},[$inp] 275 vst1.32 {v0.16b},[$inp],x4 280 vld1.32 {v1.16b},[$inp] 283 vst1.32 {v0.16b},[$inp],x4 [all …]
|
D | vpaes-x86.pl | 58 my ($round, $base, $magic, $key, $const, $inp, $out)= 372 &movdqu ("xmm0",&QWP(0,$inp)); # load key (unaligned) 436 &movdqu ("xmm0",&QWP(8,$inp)); # load key part 2 (very unaligned) 467 &movdqu ("xmm0",&QWP(16,$inp)); # load key part 2 (unaligned) 699 &lea ($inp,&DWP($k_dksd,$const)); 705 &movdqa ("xmm2",&QWP(0,$inp)); 707 &movdqa ("xmm3",&QWP(0x10,$inp)); 712 &movdqa ("xmm2",&QWP(0x20,$inp)); 715 &movdqa ("xmm3",&QWP(0x30,$inp)); 720 &movdqa ("xmm2",&QWP(0x40,$inp)); [all …]
|
D | aes-x86_64.pl | 56 $inp="%r8"; 550 lea 128($sbox),$inp # size optimization 551 mov 0-128($inp),$acc1 # prefetch Te4 552 mov 32-128($inp),$acc2 553 mov 64-128($inp),$t0 554 mov 96-128($inp),$t1 555 mov 128-128($inp),$acc1 556 mov 160-128($inp),$acc2 557 mov 192-128($inp),$t0 558 mov 224-128($inp),$t1 [all …]
|
/external/boringssl/src/crypto/rc4/asm/ |
D | rc4-x86_64.pl | 120 $inp="%rdx"; # arg3 141 mov $inp,%r12 145 my $inp="%r12"; 166 sub $inp,$out 185 xorb ($inp),$TY#b 186 movb $TY#b,($out,$inp) 187 lea 1($inp),$inp 217 xor ($inp),%r8 218 mov %r8,($out,$inp) 219 lea 8($inp),$inp [all …]
|
D | rc4-586.pl | 69 $inp="esi"; 119 &movq ("mm2",&QWP(0,$inp)) if ($i==1); 120 &movq (&QWP(-8,(@XX[0],$inp)),"mm1") if ($i==0); 141 &movq ("mm0",&QWP(0,$inp)) if ($i<=0); 142 &movq (&QWP(-8,($out,$inp)),"mm2") if ($i==0); 159 &mov ($inp,&wparam(2)); # load inp 172 &lea ($tx,&DWP(0,$inp,$ty)); 173 &sub ($out,$inp); # re-bias out 201 &lea ($ty,&DWP(-8,$inp,$ty)); 214 &cmp ($inp,&DWP(-4,$dat)); [all …]
|
D | rc4-md5-x86_64.pl | 57 my ($dat,$in0,$out,$ctx,$inp,$len, $func,$nargs); 63 ($ctx,$inp,$len) = ("%rdi","%rsi","%rdx"); 66 ($dat,$in0,$out,$ctx,$inp,$len) = ("%rdi","%rsi","%rdx","%rcx","%r8","%r9"); 131 $D#md5# mov $inp,%r15 137 $inp="%r15" if ($md5); 138 $inp=$in0 if (!$md5); 185 $inp="%r15"; 188 mov $in0,$inp 196 add $inp,$len # pointer to the end of input 229 #md5# add 4*`$j`($inp),$a [all …]
|
/external/boringssl/src/crypto/bn/asm/ |
D | rsaz-x86_64.pl | 121 ($out, $inp, $mod) = ("%rdi", "%rsi", "%rbp"); # common internal API 123 my ($out,$inp,$mod,$n0,$times) = ("%rdi","%rsi","%rdx","%rcx","%r8d"); 144 movq ($inp), %rdx 145 movq 8($inp), %rax 164 movq 16($inp), %rax 169 movq 24($inp), %rax 175 movq 32($inp), %rax 181 movq 40($inp), %rax 187 movq 48($inp), %rax 193 movq 56($inp), %rax [all …]
|
D | x86-mont.pl | 276 $inp="esi"; # integer path uses these registers differently 280 &mov ($inp,$_ap); 284 &mov ("edx",$inp); 292 &mov ("eax",&DWP(0,$inp)); 301 &mov ("eax",&DWP(0,$inp,$j,4)); # ap[j+1] 310 &mov ($inp,$_np); 319 &mov ("eax",&DWP(0,$inp)); # np[0] 322 &mov ("eax",&DWP(4,$inp)); # np[1] 335 &mov ("eax",&DWP(0,$inp,$j,4)); # ap[j+1] 346 &mov ($inp,$_np); [all …]
|
/external/ppp/pppd/ |
D | eap.c | 115 static void eap_input __P((int unit, u_char *inp, int inlen)); 119 static int eap_printpkt __P((u_char *inp, int inlen, 345 b64enc(bs, inp, inlen, outp) in b64enc() argument 347 u_char *inp; 354 bs->bs_bits = (bs->bs_bits << 8) | *inp++; 393 b64dec(bs, inp, inlen, outp) in b64dec() argument 395 u_char *inp; 403 if ((cp = strchr(base64, *inp++)) == NULL) 1248 write_pseudonym(esp, inp, len, id) in write_pseudonym() argument 1250 u_char *inp; [all …]
|
D | fsm.c | 339 u_char *inp; local 347 inp = inpacket; 352 GETCHAR(code, inp); 353 GETCHAR(id, inp); 354 GETSHORT(len, inp); 376 fsm_rconfreq(f, id, inp, len); 380 fsm_rconfack(f, id, inp, len); 385 fsm_rconfnakrej(f, code, id, inp, len); 389 fsm_rtermreq(f, id, inp, len); 397 fsm_rcoderej(f, inp, len); [all …]
|
D | upap.c | 319 u_char *inp; local 327 inp = inpacket; 332 GETCHAR(code, inp); 333 GETCHAR(id, inp); 334 GETSHORT(len, inp); 350 upap_rauthreq(u, inp, id, len); 354 upap_rauthack(u, inp, id, len); 358 upap_rauthnak(u, inp, id, len); 371 upap_rauthreq(u, inp, id, len) in upap_rauthreq() argument 373 u_char *inp; [all …]
|
/external/boringssl/src/crypto/modes/asm/ |
D | ghash-x86.pl | 138 $inp = "edi"; 161 my $cnt = $inp; 263 &mov ($inp,&wparam(0)); # load Xi 266 &mov ($Zhh,&DWP(0,$inp)); # load Xi[16] 267 &mov ($Zhl,&DWP(4,$inp)); 268 &mov ($Zlh,&DWP(8,$inp)); 269 &mov ($Zll,&DWP(12,$inp)); 284 &mov ($inp,&wparam(0)); 287 &mov (&DWP(12,$inp),$Zll); 288 &mov (&DWP(8,$inp),$Zlh); [all …]
|
D | ghash-x86_64.pl | 146 my $inp = shift; 167 mov ($inp,$cnt),`&LB("$nlo")` 251 $inp="%rdx"; 268 mov $inp,%r14 # reassign couple of args 271 { my $inp="%r14"; 305 &add ($len,$inp); # pointer to the end of data 310 &xor ($Zhi,"($inp)"); 311 &mov ("%rdx","8($inp)"); 312 &lea ($inp,"16($inp)"); 387 &cmp ($inp,$len); [all …]
|
/external/boringssl/src/crypto/sha/asm/ |
D | sha256-armv4.pl | 56 $inp="r1"; $t4="r1"; 81 @ ldr $t1,[$inp],#4 @ $i 83 str $inp,[sp,#17*4] @ make room for $t4 92 @ ldrb $t1,[$inp,#3] @ $i 94 ldrb $t2,[$inp,#2] 95 ldrb $t0,[$inp,#1] 97 ldrb $t2,[$inp],#4 100 str $inp,[sp,#17*4] @ make room for $t4 124 ldr $t1,[$inp],#4 @ prefetch 126 ldrb $t1,[$inp,#3] [all …]
|
D | sha512-x86_64.pl | 164 $inp="%rsi"; # 2nd arg 304 lea ($inp,%rdx,$SZ),%rdx # inp+num*16*$SZ 307 mov $inp,$_inp # save inp, 2nd arh 329 $code.=" mov $SZ*$i($inp),$T1\n"; 352 lea 16*$SZ($inp),$inp 363 cmp $_end,$inp 533 my ($ctx,$inp,$num,$Tbl)=("%rdi","%rsi","%rdx","%rcx"); 569 movdqu ($inp),@MSG[0] 570 movdqu 0x10($inp),@MSG[1] 571 movdqu 0x20($inp),@MSG[2] [all …]
|
D | sha1-x86_64.pl | 121 $inp="%rsi"; # 2nd arg 126 $inp="%r9"; 145 mov `4*$i`($inp),$xi[0] 149 mov `4*$j`($inp),$xi[1] 285 mov %rsi,$inp # reassigned argument 318 lea `16*4`($inp),$inp 336 my ($ctx,$inp,$num)=("%rdi","%rsi","%rdx"); 359 movdqu ($inp),@MSG[0] 361 movdqu 0x10($inp),@MSG[1] 363 movdqu 0x20($inp),@MSG[2] [all …]
|
D | sha1-armv4-large.pl | 87 $inp="r1"; 127 ldrb $t1,[$inp,#2] 128 ldrb $t0,[$inp,#3] 129 ldrb $t2,[$inp,#1] 131 ldrb $t3,[$inp],#4 138 ldr $t0,[$inp],#4 @ handles unaligned 204 add $len,$inp,$len,lsl#6 @ $len to point at the end of $inp 266 teq $inp,$len 459 &teq ($inp,$len); 461 &subeq ($inp,$inp,64); # reload last block to avoid SEGV [all …]
|
D | sha1-586.pl | 422 my ($ctx,$inp,$num)=("edi","esi","ecx"); 448 &mov ($inp,&wparam(1)); 457 &movdqu (@MSG[0],&QWP(0,$inp)); 459 &movdqu (@MSG[1],&QWP(0x10,$inp)); 461 &movdqu (@MSG[2],&QWP(0x20,$inp)); 463 &movdqu (@MSG[3],&QWP(0x30,$inp)); 471 &lea ("eax",&DWP(0x40,$inp)); 474 &cmovne ($inp,"eax"); 494 &movdqu (@MSG[0],&QWP(0,$inp)); 498 &movdqu (@MSG[1],&QWP(0x10,$inp)); [all …]
|
D | sha512-armv4.pl | 69 $inp="r1"; 297 add $len,$inp,$len,lsl#7 @ len to point at the end of inp 334 ldrb $Tlo,[$inp,#7] 335 ldrb $t0, [$inp,#6] 336 ldrb $t1, [$inp,#5] 337 ldrb $t2, [$inp,#4] 338 ldrb $Thi,[$inp,#3] 339 ldrb $t3, [$inp,#2] 341 ldrb $t0, [$inp,#1] 343 ldrb $t1, [$inp],#8 [all …]
|
/external/libhevc/common/ |
D | ihevc_trans_macros.h | 37 #define QUANT(out, inp, quant_coeff, qp_div, log2_trans_size, q_add) \ argument 51 …sign = (inp)<0 ? -1:1; … 53 …tmp = (LWORD64)(abs(inp)); … 62 #define QUANT_HBD(out, inp, quant_coeff, qp_div, log2_trans_size, q_add, bit_depth) \ argument 76 …sign = (inp)<0 ? -1:1; … 78 …tmp = (LWORD64)(abs(inp)); … 87 #define QUANT_NO_WEIGHTMAT(out, inp, quant_coeff, qp_div, log2_trans_size, q_add) \ argument 101 …sign = (inp)<0 ? -1:1; … 103 …tmp = (WORD32)(abs(inp)); … 112 #define QUANT_NO_WEIGHTMAT_HBD(out, inp, quant_coeff, qp_div, log2_trans_size, q_add, bit_depth) \ argument [all …]
|
/external/boringssl/src/crypto/evp/ |
D | asn1.c | 67 EVP_PKEY *d2i_PrivateKey(int type, EVP_PKEY **out, const uint8_t **inp, in d2i_PrivateKey() argument 87 !ret->ameth->old_priv_decode(ret, inp, len)) { in d2i_PrivateKey() 89 PKCS8_PRIV_KEY_INFO *p8 = d2i_PKCS8_PRIV_KEY_INFO(NULL, inp, len); in d2i_PrivateKey() 114 EVP_PKEY *d2i_AutoPrivateKey(EVP_PKEY **out, const uint8_t **inp, long len) { in d2i_AutoPrivateKey() argument 118 p = *inp; in d2i_AutoPrivateKey() 132 PKCS8_PRIV_KEY_INFO *p8 = d2i_PKCS8_PRIV_KEY_INFO(NULL, inp, len); in d2i_AutoPrivateKey() 152 return d2i_PrivateKey(keytype, out, inp, len); in d2i_AutoPrivateKey()
|