Lines Matching refs:reg
37 #define BLOCK(i, reg) \ in xor_avx_2() argument
39 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p1[i / sizeof(*p1)])); \ in xor_avx_2()
40 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_2()
42 asm volatile("vmovdqa %%ymm" #reg ", %0" : \ in xor_avx_2()
64 #define BLOCK(i, reg) \ in xor_avx_3() argument
66 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p2[i / sizeof(*p2)])); \ in xor_avx_3()
67 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_3()
69 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_3()
71 asm volatile("vmovdqa %%ymm" #reg ", %0" : \ in xor_avx_3()
94 #define BLOCK(i, reg) \ in xor_avx_4() argument
96 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p3[i / sizeof(*p3)])); \ in xor_avx_4()
97 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_4()
99 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_4()
101 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_4()
103 asm volatile("vmovdqa %%ymm" #reg ", %0" : \ in xor_avx_4()
127 #define BLOCK(i, reg) \ in xor_avx_5() argument
129 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p4[i / sizeof(*p4)])); \ in xor_avx_5()
130 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_5()
132 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_5()
134 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_5()
136 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_5()
138 asm volatile("vmovdqa %%ymm" #reg ", %0" : \ in xor_avx_5()