• Home
  • Raw
  • Download

Lines Matching refs:rip

383         vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
398 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
402 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
405 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
410 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
433 vpaddd ONE(%rip), %xmm9, %xmm9 # INCR CNT to get Yn
435 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
448 lea SHIFT_MASK+16(%rip), %r12
465 lea SHIFT_MASK+16(%rip), %r12
482 vpshufb SHUF_MASK(%rip), %xmm2, %xmm2
491 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
495 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9 # shuffle xmm9 back to output as ciphertext
550 vpshufb SHUF_MASK(%rip), %xmm14, %xmm14 # perform a 16Byte swap
619 vpshufb SHUF_MASK(%rip), \T7, \T7
665 vpshufb SHUF_MASK(%rip), \T7, \T7
685 vpshufb SHUF_MASK(%rip), %xmm0, %xmm0
690 vpshufb SHUF_MASK(%rip), %xmm6, %xmm6
701 vpcmpeqd TWOONE(%rip), %xmm2, %xmm2
702 vpand POLY(%rip), %xmm2, %xmm2
772 lea SHIFT_MASK(%rip), %r12
799 vmovdqa SHUF_MASK(%rip), %xmm10
834 vmovdqa SHUF_MASK(%rip), %xmm1
853 vmovdqa SHUF_MASK(%rip), %xmm10
1012 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1014 vpshufb SHUF_MASK(%rip), reg_i, reg_i # perform a 16Byte swap
1063 vpshufb SHUF_MASK(%rip), reg_i, reg_i # prepare ciphertext for GHASH computations
1090 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1092 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
1094 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1096 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1098 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1100 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
1102 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1104 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
1106 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1108 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
1110 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1112 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
1114 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1116 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
1118 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1120 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1216 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
1218 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1219 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
1220 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
1221 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
1222 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
1223 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
1224 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1248 vpaddd ONE(%rip), \CTR, \XMM1 # INCR CNT
1249 vpaddd ONE(%rip), \XMM1, \XMM2
1250 vpaddd ONE(%rip), \XMM2, \XMM3
1251 vpaddd ONE(%rip), \XMM3, \XMM4
1252 vpaddd ONE(%rip), \XMM4, \XMM5
1253 vpaddd ONE(%rip), \XMM5, \XMM6
1254 vpaddd ONE(%rip), \XMM6, \XMM7
1255 vpaddd ONE(%rip), \XMM7, \XMM8
1258 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
1259 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1260 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
1261 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
1262 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
1263 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
1264 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
1265 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1267 vpaddd ONEf(%rip), \CTR, \XMM1 # INCR CNT
1268 vpaddd ONEf(%rip), \XMM1, \XMM2
1269 vpaddd ONEf(%rip), \XMM2, \XMM3
1270 vpaddd ONEf(%rip), \XMM3, \XMM4
1271 vpaddd ONEf(%rip), \XMM4, \XMM5
1272 vpaddd ONEf(%rip), \XMM5, \XMM6
1273 vpaddd ONEf(%rip), \XMM6, \XMM7
1274 vpaddd ONEf(%rip), \XMM7, \XMM8
1585 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
1586 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1587 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
1588 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
1589 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
1590 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
1591 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
1592 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1894 vmovdqa POLY2(%rip), \T3
1960 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1962 vpshufb SHUF_MASK(%rip), reg_i, reg_i # perform a 16Byte swap
2013 vpshufb SHUF_MASK(%rip), reg_i, reg_i # prepare ciphertext for GHASH computations
2040 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2042 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
2044 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2046 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2048 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2050 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
2052 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2054 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
2056 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2058 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
2060 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2062 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
2064 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2066 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
2068 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2070 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2167 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
2170 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2171 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
2172 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
2173 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
2174 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
2175 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
2176 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2203 vpaddd ONE(%rip), \CTR, \XMM1 # INCR CNT
2204 vpaddd ONE(%rip), \XMM1, \XMM2
2205 vpaddd ONE(%rip), \XMM2, \XMM3
2206 vpaddd ONE(%rip), \XMM3, \XMM4
2207 vpaddd ONE(%rip), \XMM4, \XMM5
2208 vpaddd ONE(%rip), \XMM5, \XMM6
2209 vpaddd ONE(%rip), \XMM6, \XMM7
2210 vpaddd ONE(%rip), \XMM7, \XMM8
2213 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
2214 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2215 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
2216 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
2217 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
2218 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
2219 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
2220 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2222 vpaddd ONEf(%rip), \CTR, \XMM1 # INCR CNT
2223 vpaddd ONEf(%rip), \XMM1, \XMM2
2224 vpaddd ONEf(%rip), \XMM2, \XMM3
2225 vpaddd ONEf(%rip), \XMM3, \XMM4
2226 vpaddd ONEf(%rip), \XMM4, \XMM5
2227 vpaddd ONEf(%rip), \XMM5, \XMM6
2228 vpaddd ONEf(%rip), \XMM6, \XMM7
2229 vpaddd ONEf(%rip), \XMM7, \XMM8
2505 vmovdqa POLY2(%rip), \T3
2535 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
2536 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2537 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
2538 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
2539 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
2540 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
2541 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
2542 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2709 vmovdqa POLY2(%rip), \T3