1.text 2 3.globl gcm_gmult_4bit 4.type gcm_gmult_4bit,@function 5.align 16 6gcm_gmult_4bit: 7 pushq %rbx 8 pushq %rbp 9 pushq %r12 10.Lgmult_prologue: 11 12 movzbq 15(%rdi),%r8 13 leaq .Lrem_4bit(%rip),%r11 14 xorq %rax,%rax 15 xorq %rbx,%rbx 16 movb %r8b,%al 17 movb %r8b,%bl 18 shlb $4,%al 19 movq $14,%rcx 20 movq 8(%rsi,%rax,1),%r8 21 movq (%rsi,%rax,1),%r9 22 andb $240,%bl 23 movq %r8,%rdx 24 jmp .Loop1 25 26.align 16 27.Loop1: 28 shrq $4,%r8 29 andq $15,%rdx 30 movq %r9,%r10 31 movb (%rdi,%rcx,1),%al 32 shrq $4,%r9 33 xorq 8(%rsi,%rbx,1),%r8 34 shlq $60,%r10 35 xorq (%rsi,%rbx,1),%r9 36 movb %al,%bl 37 xorq (%r11,%rdx,8),%r9 38 movq %r8,%rdx 39 shlb $4,%al 40 xorq %r10,%r8 41 decq %rcx 42 js .Lbreak1 43 44 shrq $4,%r8 45 andq $15,%rdx 46 movq %r9,%r10 47 shrq $4,%r9 48 xorq 8(%rsi,%rax,1),%r8 49 shlq $60,%r10 50 xorq (%rsi,%rax,1),%r9 51 andb $240,%bl 52 xorq (%r11,%rdx,8),%r9 53 movq %r8,%rdx 54 xorq %r10,%r8 55 jmp .Loop1 56 57.align 16 58.Lbreak1: 59 shrq $4,%r8 60 andq $15,%rdx 61 movq %r9,%r10 62 shrq $4,%r9 63 xorq 8(%rsi,%rax,1),%r8 64 shlq $60,%r10 65 xorq (%rsi,%rax,1),%r9 66 andb $240,%bl 67 xorq (%r11,%rdx,8),%r9 68 movq %r8,%rdx 69 xorq %r10,%r8 70 71 shrq $4,%r8 72 andq $15,%rdx 73 movq %r9,%r10 74 shrq $4,%r9 75 xorq 8(%rsi,%rbx,1),%r8 76 shlq $60,%r10 77 xorq (%rsi,%rbx,1),%r9 78 xorq %r10,%r8 79 xorq (%r11,%rdx,8),%r9 80 81 bswapq %r8 82 bswapq %r9 83 movq %r8,8(%rdi) 84 movq %r9,(%rdi) 85 86 movq 16(%rsp),%rbx 87 leaq 24(%rsp),%rsp 88.Lgmult_epilogue: 89 .byte 0xf3,0xc3 90.size gcm_gmult_4bit,.-gcm_gmult_4bit 91.globl gcm_ghash_4bit 92.type gcm_ghash_4bit,@function 93.align 16 94gcm_ghash_4bit: 95 pushq %rbx 96 pushq %rbp 97 pushq %r12 98 pushq %r13 99 pushq %r14 100 pushq %r15 101 subq $280,%rsp 102.Lghash_prologue: 103 movq %rdx,%r14 104 movq %rcx,%r15 105 subq $-128,%rsi 106 leaq 16+128(%rsp),%rbp 107 xorl %edx,%edx 108 movq 0+0-128(%rsi),%r8 109 movq 0+8-128(%rsi),%rax 110 movb %al,%dl 111 shrq $4,%rax 112 movq %r8,%r10 113 shrq $4,%r8 114 movq 16+0-128(%rsi),%r9 115 shlb $4,%dl 116 movq 16+8-128(%rsi),%rbx 117 shlq $60,%r10 118 movb %dl,0(%rsp) 119 orq %r10,%rax 120 movb %bl,%dl 121 shrq $4,%rbx 122 movq %r9,%r10 123 shrq $4,%r9 124 movq %r8,0(%rbp) 125 movq 32+0-128(%rsi),%r8 126 shlb $4,%dl 127 movq %rax,0-128(%rbp) 128 movq 32+8-128(%rsi),%rax 129 shlq $60,%r10 130 movb %dl,1(%rsp) 131 orq %r10,%rbx 132 movb %al,%dl 133 shrq $4,%rax 134 movq %r8,%r10 135 shrq $4,%r8 136 movq %r9,8(%rbp) 137 movq 48+0-128(%rsi),%r9 138 shlb $4,%dl 139 movq %rbx,8-128(%rbp) 140 movq 48+8-128(%rsi),%rbx 141 shlq $60,%r10 142 movb %dl,2(%rsp) 143 orq %r10,%rax 144 movb %bl,%dl 145 shrq $4,%rbx 146 movq %r9,%r10 147 shrq $4,%r9 148 movq %r8,16(%rbp) 149 movq 64+0-128(%rsi),%r8 150 shlb $4,%dl 151 movq %rax,16-128(%rbp) 152 movq 64+8-128(%rsi),%rax 153 shlq $60,%r10 154 movb %dl,3(%rsp) 155 orq %r10,%rbx 156 movb %al,%dl 157 shrq $4,%rax 158 movq %r8,%r10 159 shrq $4,%r8 160 movq %r9,24(%rbp) 161 movq 80+0-128(%rsi),%r9 162 shlb $4,%dl 163 movq %rbx,24-128(%rbp) 164 movq 80+8-128(%rsi),%rbx 165 shlq $60,%r10 166 movb %dl,4(%rsp) 167 orq %r10,%rax 168 movb %bl,%dl 169 shrq $4,%rbx 170 movq %r9,%r10 171 shrq $4,%r9 172 movq %r8,32(%rbp) 173 movq 96+0-128(%rsi),%r8 174 shlb $4,%dl 175 movq %rax,32-128(%rbp) 176 movq 96+8-128(%rsi),%rax 177 shlq $60,%r10 178 movb %dl,5(%rsp) 179 orq %r10,%rbx 180 movb %al,%dl 181 shrq $4,%rax 182 movq %r8,%r10 183 shrq $4,%r8 184 movq %r9,40(%rbp) 185 movq 112+0-128(%rsi),%r9 186 shlb $4,%dl 187 movq %rbx,40-128(%rbp) 188 movq 112+8-128(%rsi),%rbx 189 shlq $60,%r10 190 movb %dl,6(%rsp) 191 orq %r10,%rax 192 movb %bl,%dl 193 shrq $4,%rbx 194 movq %r9,%r10 195 shrq $4,%r9 196 movq %r8,48(%rbp) 197 movq 128+0-128(%rsi),%r8 198 shlb $4,%dl 199 movq %rax,48-128(%rbp) 200 movq 128+8-128(%rsi),%rax 201 shlq $60,%r10 202 movb %dl,7(%rsp) 203 orq %r10,%rbx 204 movb %al,%dl 205 shrq $4,%rax 206 movq %r8,%r10 207 shrq $4,%r8 208 movq %r9,56(%rbp) 209 movq 144+0-128(%rsi),%r9 210 shlb $4,%dl 211 movq %rbx,56-128(%rbp) 212 movq 144+8-128(%rsi),%rbx 213 shlq $60,%r10 214 movb %dl,8(%rsp) 215 orq %r10,%rax 216 movb %bl,%dl 217 shrq $4,%rbx 218 movq %r9,%r10 219 shrq $4,%r9 220 movq %r8,64(%rbp) 221 movq 160+0-128(%rsi),%r8 222 shlb $4,%dl 223 movq %rax,64-128(%rbp) 224 movq 160+8-128(%rsi),%rax 225 shlq $60,%r10 226 movb %dl,9(%rsp) 227 orq %r10,%rbx 228 movb %al,%dl 229 shrq $4,%rax 230 movq %r8,%r10 231 shrq $4,%r8 232 movq %r9,72(%rbp) 233 movq 176+0-128(%rsi),%r9 234 shlb $4,%dl 235 movq %rbx,72-128(%rbp) 236 movq 176+8-128(%rsi),%rbx 237 shlq $60,%r10 238 movb %dl,10(%rsp) 239 orq %r10,%rax 240 movb %bl,%dl 241 shrq $4,%rbx 242 movq %r9,%r10 243 shrq $4,%r9 244 movq %r8,80(%rbp) 245 movq 192+0-128(%rsi),%r8 246 shlb $4,%dl 247 movq %rax,80-128(%rbp) 248 movq 192+8-128(%rsi),%rax 249 shlq $60,%r10 250 movb %dl,11(%rsp) 251 orq %r10,%rbx 252 movb %al,%dl 253 shrq $4,%rax 254 movq %r8,%r10 255 shrq $4,%r8 256 movq %r9,88(%rbp) 257 movq 208+0-128(%rsi),%r9 258 shlb $4,%dl 259 movq %rbx,88-128(%rbp) 260 movq 208+8-128(%rsi),%rbx 261 shlq $60,%r10 262 movb %dl,12(%rsp) 263 orq %r10,%rax 264 movb %bl,%dl 265 shrq $4,%rbx 266 movq %r9,%r10 267 shrq $4,%r9 268 movq %r8,96(%rbp) 269 movq 224+0-128(%rsi),%r8 270 shlb $4,%dl 271 movq %rax,96-128(%rbp) 272 movq 224+8-128(%rsi),%rax 273 shlq $60,%r10 274 movb %dl,13(%rsp) 275 orq %r10,%rbx 276 movb %al,%dl 277 shrq $4,%rax 278 movq %r8,%r10 279 shrq $4,%r8 280 movq %r9,104(%rbp) 281 movq 240+0-128(%rsi),%r9 282 shlb $4,%dl 283 movq %rbx,104-128(%rbp) 284 movq 240+8-128(%rsi),%rbx 285 shlq $60,%r10 286 movb %dl,14(%rsp) 287 orq %r10,%rax 288 movb %bl,%dl 289 shrq $4,%rbx 290 movq %r9,%r10 291 shrq $4,%r9 292 movq %r8,112(%rbp) 293 shlb $4,%dl 294 movq %rax,112-128(%rbp) 295 shlq $60,%r10 296 movb %dl,15(%rsp) 297 orq %r10,%rbx 298 movq %r9,120(%rbp) 299 movq %rbx,120-128(%rbp) 300 addq $-128,%rsi 301 movq 8(%rdi),%r8 302 movq 0(%rdi),%r9 303 addq %r14,%r15 304 leaq .Lrem_8bit(%rip),%r11 305 jmp .Louter_loop 306.align 16 307.Louter_loop: 308 xorq (%r14),%r9 309 movq 8(%r14),%rdx 310 leaq 16(%r14),%r14 311 xorq %r8,%rdx 312 movq %r9,(%rdi) 313 movq %rdx,8(%rdi) 314 shrq $32,%rdx 315 xorq %rax,%rax 316 roll $8,%edx 317 movb %dl,%al 318 movzbl %dl,%ebx 319 shlb $4,%al 320 shrl $4,%ebx 321 roll $8,%edx 322 movq 8(%rsi,%rax,1),%r8 323 movq (%rsi,%rax,1),%r9 324 movb %dl,%al 325 movzbl %dl,%ecx 326 shlb $4,%al 327 movzbq (%rsp,%rbx,1),%r12 328 shrl $4,%ecx 329 xorq %r8,%r12 330 movq %r9,%r10 331 shrq $8,%r8 332 movzbq %r12b,%r12 333 shrq $8,%r9 334 xorq -128(%rbp,%rbx,8),%r8 335 shlq $56,%r10 336 xorq (%rbp,%rbx,8),%r9 337 roll $8,%edx 338 xorq 8(%rsi,%rax,1),%r8 339 xorq (%rsi,%rax,1),%r9 340 movb %dl,%al 341 xorq %r10,%r8 342 movzwq (%r11,%r12,2),%r12 343 movzbl %dl,%ebx 344 shlb $4,%al 345 movzbq (%rsp,%rcx,1),%r13 346 shrl $4,%ebx 347 shlq $48,%r12 348 xorq %r8,%r13 349 movq %r9,%r10 350 xorq %r12,%r9 351 shrq $8,%r8 352 movzbq %r13b,%r13 353 shrq $8,%r9 354 xorq -128(%rbp,%rcx,8),%r8 355 shlq $56,%r10 356 xorq (%rbp,%rcx,8),%r9 357 roll $8,%edx 358 xorq 8(%rsi,%rax,1),%r8 359 xorq (%rsi,%rax,1),%r9 360 movb %dl,%al 361 xorq %r10,%r8 362 movzwq (%r11,%r13,2),%r13 363 movzbl %dl,%ecx 364 shlb $4,%al 365 movzbq (%rsp,%rbx,1),%r12 366 shrl $4,%ecx 367 shlq $48,%r13 368 xorq %r8,%r12 369 movq %r9,%r10 370 xorq %r13,%r9 371 shrq $8,%r8 372 movzbq %r12b,%r12 373 movl 8(%rdi),%edx 374 shrq $8,%r9 375 xorq -128(%rbp,%rbx,8),%r8 376 shlq $56,%r10 377 xorq (%rbp,%rbx,8),%r9 378 roll $8,%edx 379 xorq 8(%rsi,%rax,1),%r8 380 xorq (%rsi,%rax,1),%r9 381 movb %dl,%al 382 xorq %r10,%r8 383 movzwq (%r11,%r12,2),%r12 384 movzbl %dl,%ebx 385 shlb $4,%al 386 movzbq (%rsp,%rcx,1),%r13 387 shrl $4,%ebx 388 shlq $48,%r12 389 xorq %r8,%r13 390 movq %r9,%r10 391 xorq %r12,%r9 392 shrq $8,%r8 393 movzbq %r13b,%r13 394 shrq $8,%r9 395 xorq -128(%rbp,%rcx,8),%r8 396 shlq $56,%r10 397 xorq (%rbp,%rcx,8),%r9 398 roll $8,%edx 399 xorq 8(%rsi,%rax,1),%r8 400 xorq (%rsi,%rax,1),%r9 401 movb %dl,%al 402 xorq %r10,%r8 403 movzwq (%r11,%r13,2),%r13 404 movzbl %dl,%ecx 405 shlb $4,%al 406 movzbq (%rsp,%rbx,1),%r12 407 shrl $4,%ecx 408 shlq $48,%r13 409 xorq %r8,%r12 410 movq %r9,%r10 411 xorq %r13,%r9 412 shrq $8,%r8 413 movzbq %r12b,%r12 414 shrq $8,%r9 415 xorq -128(%rbp,%rbx,8),%r8 416 shlq $56,%r10 417 xorq (%rbp,%rbx,8),%r9 418 roll $8,%edx 419 xorq 8(%rsi,%rax,1),%r8 420 xorq (%rsi,%rax,1),%r9 421 movb %dl,%al 422 xorq %r10,%r8 423 movzwq (%r11,%r12,2),%r12 424 movzbl %dl,%ebx 425 shlb $4,%al 426 movzbq (%rsp,%rcx,1),%r13 427 shrl $4,%ebx 428 shlq $48,%r12 429 xorq %r8,%r13 430 movq %r9,%r10 431 xorq %r12,%r9 432 shrq $8,%r8 433 movzbq %r13b,%r13 434 shrq $8,%r9 435 xorq -128(%rbp,%rcx,8),%r8 436 shlq $56,%r10 437 xorq (%rbp,%rcx,8),%r9 438 roll $8,%edx 439 xorq 8(%rsi,%rax,1),%r8 440 xorq (%rsi,%rax,1),%r9 441 movb %dl,%al 442 xorq %r10,%r8 443 movzwq (%r11,%r13,2),%r13 444 movzbl %dl,%ecx 445 shlb $4,%al 446 movzbq (%rsp,%rbx,1),%r12 447 shrl $4,%ecx 448 shlq $48,%r13 449 xorq %r8,%r12 450 movq %r9,%r10 451 xorq %r13,%r9 452 shrq $8,%r8 453 movzbq %r12b,%r12 454 movl 4(%rdi),%edx 455 shrq $8,%r9 456 xorq -128(%rbp,%rbx,8),%r8 457 shlq $56,%r10 458 xorq (%rbp,%rbx,8),%r9 459 roll $8,%edx 460 xorq 8(%rsi,%rax,1),%r8 461 xorq (%rsi,%rax,1),%r9 462 movb %dl,%al 463 xorq %r10,%r8 464 movzwq (%r11,%r12,2),%r12 465 movzbl %dl,%ebx 466 shlb $4,%al 467 movzbq (%rsp,%rcx,1),%r13 468 shrl $4,%ebx 469 shlq $48,%r12 470 xorq %r8,%r13 471 movq %r9,%r10 472 xorq %r12,%r9 473 shrq $8,%r8 474 movzbq %r13b,%r13 475 shrq $8,%r9 476 xorq -128(%rbp,%rcx,8),%r8 477 shlq $56,%r10 478 xorq (%rbp,%rcx,8),%r9 479 roll $8,%edx 480 xorq 8(%rsi,%rax,1),%r8 481 xorq (%rsi,%rax,1),%r9 482 movb %dl,%al 483 xorq %r10,%r8 484 movzwq (%r11,%r13,2),%r13 485 movzbl %dl,%ecx 486 shlb $4,%al 487 movzbq (%rsp,%rbx,1),%r12 488 shrl $4,%ecx 489 shlq $48,%r13 490 xorq %r8,%r12 491 movq %r9,%r10 492 xorq %r13,%r9 493 shrq $8,%r8 494 movzbq %r12b,%r12 495 shrq $8,%r9 496 xorq -128(%rbp,%rbx,8),%r8 497 shlq $56,%r10 498 xorq (%rbp,%rbx,8),%r9 499 roll $8,%edx 500 xorq 8(%rsi,%rax,1),%r8 501 xorq (%rsi,%rax,1),%r9 502 movb %dl,%al 503 xorq %r10,%r8 504 movzwq (%r11,%r12,2),%r12 505 movzbl %dl,%ebx 506 shlb $4,%al 507 movzbq (%rsp,%rcx,1),%r13 508 shrl $4,%ebx 509 shlq $48,%r12 510 xorq %r8,%r13 511 movq %r9,%r10 512 xorq %r12,%r9 513 shrq $8,%r8 514 movzbq %r13b,%r13 515 shrq $8,%r9 516 xorq -128(%rbp,%rcx,8),%r8 517 shlq $56,%r10 518 xorq (%rbp,%rcx,8),%r9 519 roll $8,%edx 520 xorq 8(%rsi,%rax,1),%r8 521 xorq (%rsi,%rax,1),%r9 522 movb %dl,%al 523 xorq %r10,%r8 524 movzwq (%r11,%r13,2),%r13 525 movzbl %dl,%ecx 526 shlb $4,%al 527 movzbq (%rsp,%rbx,1),%r12 528 shrl $4,%ecx 529 shlq $48,%r13 530 xorq %r8,%r12 531 movq %r9,%r10 532 xorq %r13,%r9 533 shrq $8,%r8 534 movzbq %r12b,%r12 535 movl 0(%rdi),%edx 536 shrq $8,%r9 537 xorq -128(%rbp,%rbx,8),%r8 538 shlq $56,%r10 539 xorq (%rbp,%rbx,8),%r9 540 roll $8,%edx 541 xorq 8(%rsi,%rax,1),%r8 542 xorq (%rsi,%rax,1),%r9 543 movb %dl,%al 544 xorq %r10,%r8 545 movzwq (%r11,%r12,2),%r12 546 movzbl %dl,%ebx 547 shlb $4,%al 548 movzbq (%rsp,%rcx,1),%r13 549 shrl $4,%ebx 550 shlq $48,%r12 551 xorq %r8,%r13 552 movq %r9,%r10 553 xorq %r12,%r9 554 shrq $8,%r8 555 movzbq %r13b,%r13 556 shrq $8,%r9 557 xorq -128(%rbp,%rcx,8),%r8 558 shlq $56,%r10 559 xorq (%rbp,%rcx,8),%r9 560 roll $8,%edx 561 xorq 8(%rsi,%rax,1),%r8 562 xorq (%rsi,%rax,1),%r9 563 movb %dl,%al 564 xorq %r10,%r8 565 movzwq (%r11,%r13,2),%r13 566 movzbl %dl,%ecx 567 shlb $4,%al 568 movzbq (%rsp,%rbx,1),%r12 569 shrl $4,%ecx 570 shlq $48,%r13 571 xorq %r8,%r12 572 movq %r9,%r10 573 xorq %r13,%r9 574 shrq $8,%r8 575 movzbq %r12b,%r12 576 shrq $8,%r9 577 xorq -128(%rbp,%rbx,8),%r8 578 shlq $56,%r10 579 xorq (%rbp,%rbx,8),%r9 580 roll $8,%edx 581 xorq 8(%rsi,%rax,1),%r8 582 xorq (%rsi,%rax,1),%r9 583 movb %dl,%al 584 xorq %r10,%r8 585 movzwq (%r11,%r12,2),%r12 586 movzbl %dl,%ebx 587 shlb $4,%al 588 movzbq (%rsp,%rcx,1),%r13 589 shrl $4,%ebx 590 shlq $48,%r12 591 xorq %r8,%r13 592 movq %r9,%r10 593 xorq %r12,%r9 594 shrq $8,%r8 595 movzbq %r13b,%r13 596 shrq $8,%r9 597 xorq -128(%rbp,%rcx,8),%r8 598 shlq $56,%r10 599 xorq (%rbp,%rcx,8),%r9 600 roll $8,%edx 601 xorq 8(%rsi,%rax,1),%r8 602 xorq (%rsi,%rax,1),%r9 603 movb %dl,%al 604 xorq %r10,%r8 605 movzwq (%r11,%r13,2),%r13 606 movzbl %dl,%ecx 607 shlb $4,%al 608 movzbq (%rsp,%rbx,1),%r12 609 andl $240,%ecx 610 shlq $48,%r13 611 xorq %r8,%r12 612 movq %r9,%r10 613 xorq %r13,%r9 614 shrq $8,%r8 615 movzbq %r12b,%r12 616 movl -4(%rdi),%edx 617 shrq $8,%r9 618 xorq -128(%rbp,%rbx,8),%r8 619 shlq $56,%r10 620 xorq (%rbp,%rbx,8),%r9 621 movzwq (%r11,%r12,2),%r12 622 xorq 8(%rsi,%rax,1),%r8 623 xorq (%rsi,%rax,1),%r9 624 shlq $48,%r12 625 xorq %r10,%r8 626 xorq %r12,%r9 627 movzbq %r8b,%r13 628 shrq $4,%r8 629 movq %r9,%r10 630 shlb $4,%r13b 631 shrq $4,%r9 632 xorq 8(%rsi,%rcx,1),%r8 633 movzwq (%r11,%r13,2),%r13 634 shlq $60,%r10 635 xorq (%rsi,%rcx,1),%r9 636 xorq %r10,%r8 637 shlq $48,%r13 638 bswapq %r8 639 xorq %r13,%r9 640 bswapq %r9 641 cmpq %r15,%r14 642 jb .Louter_loop 643 movq %r8,8(%rdi) 644 movq %r9,(%rdi) 645 646 leaq 280(%rsp),%rsi 647 movq 0(%rsi),%r15 648 movq 8(%rsi),%r14 649 movq 16(%rsi),%r13 650 movq 24(%rsi),%r12 651 movq 32(%rsi),%rbp 652 movq 40(%rsi),%rbx 653 leaq 48(%rsi),%rsp 654.Lghash_epilogue: 655 .byte 0xf3,0xc3 656.size gcm_ghash_4bit,.-gcm_ghash_4bit 657.globl gcm_init_clmul 658.type gcm_init_clmul,@function 659.align 16 660gcm_init_clmul: 661 movdqu (%rsi),%xmm2 662 pshufd $78,%xmm2,%xmm2 663 664 665 pshufd $255,%xmm2,%xmm4 666 movdqa %xmm2,%xmm3 667 psllq $1,%xmm2 668 pxor %xmm5,%xmm5 669 psrlq $63,%xmm3 670 pcmpgtd %xmm4,%xmm5 671 pslldq $8,%xmm3 672 por %xmm3,%xmm2 673 674 675 pand .L0x1c2_polynomial(%rip),%xmm5 676 pxor %xmm5,%xmm2 677 678 679 movdqa %xmm2,%xmm0 680 movdqa %xmm0,%xmm1 681 pshufd $78,%xmm0,%xmm3 682 pshufd $78,%xmm2,%xmm4 683 pxor %xmm0,%xmm3 684 pxor %xmm2,%xmm4 685.byte 102,15,58,68,194,0 686.byte 102,15,58,68,202,17 687.byte 102,15,58,68,220,0 688 pxor %xmm0,%xmm3 689 pxor %xmm1,%xmm3 690 691 movdqa %xmm3,%xmm4 692 psrldq $8,%xmm3 693 pslldq $8,%xmm4 694 pxor %xmm3,%xmm1 695 pxor %xmm4,%xmm0 696 697 movdqa %xmm0,%xmm3 698 psllq $1,%xmm0 699 pxor %xmm3,%xmm0 700 psllq $5,%xmm0 701 pxor %xmm3,%xmm0 702 psllq $57,%xmm0 703 movdqa %xmm0,%xmm4 704 pslldq $8,%xmm0 705 psrldq $8,%xmm4 706 pxor %xmm3,%xmm0 707 pxor %xmm4,%xmm1 708 709 710 movdqa %xmm0,%xmm4 711 psrlq $5,%xmm0 712 pxor %xmm4,%xmm0 713 psrlq $1,%xmm0 714 pxor %xmm4,%xmm0 715 pxor %xmm1,%xmm4 716 psrlq $1,%xmm0 717 pxor %xmm4,%xmm0 718 movdqu %xmm2,(%rdi) 719 movdqu %xmm0,16(%rdi) 720 .byte 0xf3,0xc3 721.size gcm_init_clmul,.-gcm_init_clmul 722.globl gcm_gmult_clmul 723.type gcm_gmult_clmul,@function 724.align 16 725gcm_gmult_clmul: 726 movdqu (%rdi),%xmm0 727 movdqa .Lbswap_mask(%rip),%xmm5 728 movdqu (%rsi),%xmm2 729.byte 102,15,56,0,197 730 movdqa %xmm0,%xmm1 731 pshufd $78,%xmm0,%xmm3 732 pshufd $78,%xmm2,%xmm4 733 pxor %xmm0,%xmm3 734 pxor %xmm2,%xmm4 735.byte 102,15,58,68,194,0 736.byte 102,15,58,68,202,17 737.byte 102,15,58,68,220,0 738 pxor %xmm0,%xmm3 739 pxor %xmm1,%xmm3 740 741 movdqa %xmm3,%xmm4 742 psrldq $8,%xmm3 743 pslldq $8,%xmm4 744 pxor %xmm3,%xmm1 745 pxor %xmm4,%xmm0 746 747 movdqa %xmm0,%xmm3 748 psllq $1,%xmm0 749 pxor %xmm3,%xmm0 750 psllq $5,%xmm0 751 pxor %xmm3,%xmm0 752 psllq $57,%xmm0 753 movdqa %xmm0,%xmm4 754 pslldq $8,%xmm0 755 psrldq $8,%xmm4 756 pxor %xmm3,%xmm0 757 pxor %xmm4,%xmm1 758 759 760 movdqa %xmm0,%xmm4 761 psrlq $5,%xmm0 762 pxor %xmm4,%xmm0 763 psrlq $1,%xmm0 764 pxor %xmm4,%xmm0 765 pxor %xmm1,%xmm4 766 psrlq $1,%xmm0 767 pxor %xmm4,%xmm0 768.byte 102,15,56,0,197 769 movdqu %xmm0,(%rdi) 770 .byte 0xf3,0xc3 771.size gcm_gmult_clmul,.-gcm_gmult_clmul 772.globl gcm_ghash_clmul 773.type gcm_ghash_clmul,@function 774.align 16 775gcm_ghash_clmul: 776 movdqa .Lbswap_mask(%rip),%xmm5 777 778 movdqu (%rdi),%xmm0 779 movdqu (%rsi),%xmm2 780.byte 102,15,56,0,197 781 782 subq $16,%rcx 783 jz .Lodd_tail 784 785 movdqu 16(%rsi),%xmm8 786 787 788 789 790 791 movdqu (%rdx),%xmm3 792 movdqu 16(%rdx),%xmm6 793.byte 102,15,56,0,221 794.byte 102,15,56,0,245 795 pxor %xmm3,%xmm0 796 movdqa %xmm6,%xmm7 797 pshufd $78,%xmm6,%xmm3 798 pshufd $78,%xmm2,%xmm4 799 pxor %xmm6,%xmm3 800 pxor %xmm2,%xmm4 801.byte 102,15,58,68,242,0 802.byte 102,15,58,68,250,17 803.byte 102,15,58,68,220,0 804 pxor %xmm6,%xmm3 805 pxor %xmm7,%xmm3 806 807 movdqa %xmm3,%xmm4 808 psrldq $8,%xmm3 809 pslldq $8,%xmm4 810 pxor %xmm3,%xmm7 811 pxor %xmm4,%xmm6 812 movdqa %xmm0,%xmm1 813 pshufd $78,%xmm0,%xmm3 814 pshufd $78,%xmm8,%xmm4 815 pxor %xmm0,%xmm3 816 pxor %xmm8,%xmm4 817 818 leaq 32(%rdx),%rdx 819 subq $32,%rcx 820 jbe .Leven_tail 821 822.Lmod_loop: 823.byte 102,65,15,58,68,192,0 824.byte 102,65,15,58,68,200,17 825.byte 102,15,58,68,220,0 826 pxor %xmm0,%xmm3 827 pxor %xmm1,%xmm3 828 829 movdqa %xmm3,%xmm4 830 psrldq $8,%xmm3 831 pslldq $8,%xmm4 832 pxor %xmm3,%xmm1 833 pxor %xmm4,%xmm0 834 movdqu (%rdx),%xmm3 835 pxor %xmm6,%xmm0 836 pxor %xmm7,%xmm1 837 838 movdqu 16(%rdx),%xmm6 839.byte 102,15,56,0,221 840.byte 102,15,56,0,245 841 842 movdqa %xmm6,%xmm7 843 pshufd $78,%xmm6,%xmm9 844 pshufd $78,%xmm2,%xmm10 845 pxor %xmm6,%xmm9 846 pxor %xmm2,%xmm10 847 pxor %xmm3,%xmm1 848 849 movdqa %xmm0,%xmm3 850 psllq $1,%xmm0 851 pxor %xmm3,%xmm0 852 psllq $5,%xmm0 853 pxor %xmm3,%xmm0 854.byte 102,15,58,68,242,0 855 psllq $57,%xmm0 856 movdqa %xmm0,%xmm4 857 pslldq $8,%xmm0 858 psrldq $8,%xmm4 859 pxor %xmm3,%xmm0 860 pxor %xmm4,%xmm1 861 862.byte 102,15,58,68,250,17 863 movdqa %xmm0,%xmm4 864 psrlq $5,%xmm0 865 pxor %xmm4,%xmm0 866 psrlq $1,%xmm0 867 pxor %xmm4,%xmm0 868 pxor %xmm1,%xmm4 869 psrlq $1,%xmm0 870 pxor %xmm4,%xmm0 871 872.byte 102,69,15,58,68,202,0 873 movdqa %xmm0,%xmm1 874 pshufd $78,%xmm0,%xmm3 875 pshufd $78,%xmm8,%xmm4 876 pxor %xmm0,%xmm3 877 pxor %xmm8,%xmm4 878 879 pxor %xmm6,%xmm9 880 pxor %xmm7,%xmm9 881 movdqa %xmm9,%xmm10 882 psrldq $8,%xmm9 883 pslldq $8,%xmm10 884 pxor %xmm9,%xmm7 885 pxor %xmm10,%xmm6 886 887 leaq 32(%rdx),%rdx 888 subq $32,%rcx 889 ja .Lmod_loop 890 891.Leven_tail: 892.byte 102,65,15,58,68,192,0 893.byte 102,65,15,58,68,200,17 894.byte 102,15,58,68,220,0 895 pxor %xmm0,%xmm3 896 pxor %xmm1,%xmm3 897 898 movdqa %xmm3,%xmm4 899 psrldq $8,%xmm3 900 pslldq $8,%xmm4 901 pxor %xmm3,%xmm1 902 pxor %xmm4,%xmm0 903 pxor %xmm6,%xmm0 904 pxor %xmm7,%xmm1 905 906 movdqa %xmm0,%xmm3 907 psllq $1,%xmm0 908 pxor %xmm3,%xmm0 909 psllq $5,%xmm0 910 pxor %xmm3,%xmm0 911 psllq $57,%xmm0 912 movdqa %xmm0,%xmm4 913 pslldq $8,%xmm0 914 psrldq $8,%xmm4 915 pxor %xmm3,%xmm0 916 pxor %xmm4,%xmm1 917 918 919 movdqa %xmm0,%xmm4 920 psrlq $5,%xmm0 921 pxor %xmm4,%xmm0 922 psrlq $1,%xmm0 923 pxor %xmm4,%xmm0 924 pxor %xmm1,%xmm4 925 psrlq $1,%xmm0 926 pxor %xmm4,%xmm0 927 testq %rcx,%rcx 928 jnz .Ldone 929 930.Lodd_tail: 931 movdqu (%rdx),%xmm3 932.byte 102,15,56,0,221 933 pxor %xmm3,%xmm0 934 movdqa %xmm0,%xmm1 935 pshufd $78,%xmm0,%xmm3 936 pshufd $78,%xmm2,%xmm4 937 pxor %xmm0,%xmm3 938 pxor %xmm2,%xmm4 939.byte 102,15,58,68,194,0 940.byte 102,15,58,68,202,17 941.byte 102,15,58,68,220,0 942 pxor %xmm0,%xmm3 943 pxor %xmm1,%xmm3 944 945 movdqa %xmm3,%xmm4 946 psrldq $8,%xmm3 947 pslldq $8,%xmm4 948 pxor %xmm3,%xmm1 949 pxor %xmm4,%xmm0 950 951 movdqa %xmm0,%xmm3 952 psllq $1,%xmm0 953 pxor %xmm3,%xmm0 954 psllq $5,%xmm0 955 pxor %xmm3,%xmm0 956 psllq $57,%xmm0 957 movdqa %xmm0,%xmm4 958 pslldq $8,%xmm0 959 psrldq $8,%xmm4 960 pxor %xmm3,%xmm0 961 pxor %xmm4,%xmm1 962 963 964 movdqa %xmm0,%xmm4 965 psrlq $5,%xmm0 966 pxor %xmm4,%xmm0 967 psrlq $1,%xmm0 968 pxor %xmm4,%xmm0 969 pxor %xmm1,%xmm4 970 psrlq $1,%xmm0 971 pxor %xmm4,%xmm0 972.Ldone: 973.byte 102,15,56,0,197 974 movdqu %xmm0,(%rdi) 975 .byte 0xf3,0xc3 976.LSEH_end_gcm_ghash_clmul: 977.size gcm_ghash_clmul,.-gcm_ghash_clmul 978.align 64 979.Lbswap_mask: 980.byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0 981.L0x1c2_polynomial: 982.byte 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xc2 983.align 64 984.type .Lrem_4bit,@object 985.Lrem_4bit: 986.long 0,0,0,471859200,0,943718400,0,610271232 987.long 0,1887436800,0,1822425088,0,1220542464,0,1423966208 988.long 0,3774873600,0,4246732800,0,3644850176,0,3311403008 989.long 0,2441084928,0,2376073216,0,2847932416,0,3051356160 990.type .Lrem_8bit,@object 991.Lrem_8bit: 992.value 0x0000,0x01C2,0x0384,0x0246,0x0708,0x06CA,0x048C,0x054E 993.value 0x0E10,0x0FD2,0x0D94,0x0C56,0x0918,0x08DA,0x0A9C,0x0B5E 994.value 0x1C20,0x1DE2,0x1FA4,0x1E66,0x1B28,0x1AEA,0x18AC,0x196E 995.value 0x1230,0x13F2,0x11B4,0x1076,0x1538,0x14FA,0x16BC,0x177E 996.value 0x3840,0x3982,0x3BC4,0x3A06,0x3F48,0x3E8A,0x3CCC,0x3D0E 997.value 0x3650,0x3792,0x35D4,0x3416,0x3158,0x309A,0x32DC,0x331E 998.value 0x2460,0x25A2,0x27E4,0x2626,0x2368,0x22AA,0x20EC,0x212E 999.value 0x2A70,0x2BB2,0x29F4,0x2836,0x2D78,0x2CBA,0x2EFC,0x2F3E 1000.value 0x7080,0x7142,0x7304,0x72C6,0x7788,0x764A,0x740C,0x75CE 1001.value 0x7E90,0x7F52,0x7D14,0x7CD6,0x7998,0x785A,0x7A1C,0x7BDE 1002.value 0x6CA0,0x6D62,0x6F24,0x6EE6,0x6BA8,0x6A6A,0x682C,0x69EE 1003.value 0x62B0,0x6372,0x6134,0x60F6,0x65B8,0x647A,0x663C,0x67FE 1004.value 0x48C0,0x4902,0x4B44,0x4A86,0x4FC8,0x4E0A,0x4C4C,0x4D8E 1005.value 0x46D0,0x4712,0x4554,0x4496,0x41D8,0x401A,0x425C,0x439E 1006.value 0x54E0,0x5522,0x5764,0x56A6,0x53E8,0x522A,0x506C,0x51AE 1007.value 0x5AF0,0x5B32,0x5974,0x58B6,0x5DF8,0x5C3A,0x5E7C,0x5FBE 1008.value 0xE100,0xE0C2,0xE284,0xE346,0xE608,0xE7CA,0xE58C,0xE44E 1009.value 0xEF10,0xEED2,0xEC94,0xED56,0xE818,0xE9DA,0xEB9C,0xEA5E 1010.value 0xFD20,0xFCE2,0xFEA4,0xFF66,0xFA28,0xFBEA,0xF9AC,0xF86E 1011.value 0xF330,0xF2F2,0xF0B4,0xF176,0xF438,0xF5FA,0xF7BC,0xF67E 1012.value 0xD940,0xD882,0xDAC4,0xDB06,0xDE48,0xDF8A,0xDDCC,0xDC0E 1013.value 0xD750,0xD692,0xD4D4,0xD516,0xD058,0xD19A,0xD3DC,0xD21E 1014.value 0xC560,0xC4A2,0xC6E4,0xC726,0xC268,0xC3AA,0xC1EC,0xC02E 1015.value 0xCB70,0xCAB2,0xC8F4,0xC936,0xCC78,0xCDBA,0xCFFC,0xCE3E 1016.value 0x9180,0x9042,0x9204,0x93C6,0x9688,0x974A,0x950C,0x94CE 1017.value 0x9F90,0x9E52,0x9C14,0x9DD6,0x9898,0x995A,0x9B1C,0x9ADE 1018.value 0x8DA0,0x8C62,0x8E24,0x8FE6,0x8AA8,0x8B6A,0x892C,0x88EE 1019.value 0x83B0,0x8272,0x8034,0x81F6,0x84B8,0x857A,0x873C,0x86FE 1020.value 0xA9C0,0xA802,0xAA44,0xAB86,0xAEC8,0xAF0A,0xAD4C,0xAC8E 1021.value 0xA7D0,0xA612,0xA454,0xA596,0xA0D8,0xA11A,0xA35C,0xA29E 1022.value 0xB5E0,0xB422,0xB664,0xB7A6,0xB2E8,0xB32A,0xB16C,0xB0AE 1023.value 0xBBF0,0xBA32,0xB874,0xB9B6,0xBCF8,0xBD3A,0xBF7C,0xBEBE 1024 1025.byte 71,72,65,83,72,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 1026.align 64 1027