Lines Matching +full:0 +full:x0f
28 static const u8 __aligned(16) x0f[16] = { in raid6_2data_recov_ssse3()
29 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, in raid6_2data_recov_ssse3()
30 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f}; in raid6_2data_recov_ssse3()
60 asm volatile("movdqa %0,%%xmm7" : : "m" (x0f[0])); in raid6_2data_recov_ssse3()
63 asm volatile("movdqa %0,%%xmm6" : : "m" (qmul[0])); in raid6_2data_recov_ssse3()
64 asm volatile("movdqa %0,%%xmm14" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3()
65 asm volatile("movdqa %0,%%xmm15" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3()
73 asm volatile("movdqa %0,%%xmm1" : : "m" (q[0])); in raid6_2data_recov_ssse3()
74 asm volatile("movdqa %0,%%xmm9" : : "m" (q[16])); in raid6_2data_recov_ssse3()
75 asm volatile("movdqa %0,%%xmm0" : : "m" (p[0])); in raid6_2data_recov_ssse3()
76 asm volatile("movdqa %0,%%xmm8" : : "m" (p[16])); in raid6_2data_recov_ssse3()
77 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3()
78 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3()
79 asm volatile("pxor %0,%%xmm0" : : "m" (dp[0])); in raid6_2data_recov_ssse3()
80 asm volatile("pxor %0,%%xmm8" : : "m" (dp[16])); in raid6_2data_recov_ssse3()
85 asm volatile("movdqa %0,%%xmm5" : : "m" (qmul[16])); in raid6_2data_recov_ssse3()
130 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3()
131 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3()
135 asm volatile("movdqa %%xmm0,%0" : "=m" (dp[0])); in raid6_2data_recov_ssse3()
136 asm volatile("movdqa %%xmm8,%0" : "=m" (dp[16])); in raid6_2data_recov_ssse3()
144 asm volatile("movdqa %0,%%xmm1" : : "m" (*q)); in raid6_2data_recov_ssse3()
145 asm volatile("movdqa %0,%%xmm0" : : "m" (*p)); in raid6_2data_recov_ssse3()
146 asm volatile("pxor %0,%%xmm1" : : "m" (*dq)); in raid6_2data_recov_ssse3()
147 asm volatile("pxor %0,%%xmm0" : : "m" (*dp)); in raid6_2data_recov_ssse3()
150 * 0 = dp ^ p in raid6_2data_recov_ssse3()
152 asm volatile("movdqa %0,%%xmm4" : : "m" (qmul[0])); in raid6_2data_recov_ssse3()
153 asm volatile("movdqa %0,%%xmm5" : : "m" (qmul[16])); in raid6_2data_recov_ssse3()
167 asm volatile("movdqa %0,%%xmm4" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3()
168 asm volatile("movdqa %0,%%xmm1" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3()
180 asm volatile("movdqa %%xmm1,%0" : "=m" (*dq)); in raid6_2data_recov_ssse3()
183 asm volatile("movdqa %%xmm0,%0" : "=m" (*dp)); in raid6_2data_recov_ssse3()
202 static const u8 __aligned(16) x0f[16] = { in raid6_datap_recov_ssse3()
203 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, in raid6_datap_recov_ssse3()
204 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f}; in raid6_datap_recov_ssse3()
226 asm volatile("movdqa %0, %%xmm7" : : "m" (x0f[0])); in raid6_datap_recov_ssse3()
230 asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0])); in raid6_datap_recov_ssse3()
231 asm volatile("movdqa %0, %%xmm4" : : "m" (dq[16])); in raid6_datap_recov_ssse3()
232 asm volatile("pxor %0, %%xmm3" : : "m" (q[0])); in raid6_datap_recov_ssse3()
233 asm volatile("movdqa %0, %%xmm0" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
235 /* xmm3 = q[0] ^ dq[0] */ in raid6_datap_recov_ssse3()
237 asm volatile("pxor %0, %%xmm4" : : "m" (q[16])); in raid6_datap_recov_ssse3()
238 asm volatile("movdqa %0, %%xmm1" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
252 asm volatile("movdqa %0, %%xmm10" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
254 asm volatile("movdqa %0, %%xmm11" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
256 /* xmm1 = qmul[q[0] ^ dq[0]] */ in raid6_datap_recov_ssse3()
263 asm volatile("movdqa %0, %%xmm2" : : "m" (p[0])); in raid6_datap_recov_ssse3()
265 asm volatile("movdqa %0, %%xmm12" : : "m" (p[16])); in raid6_datap_recov_ssse3()
271 /* xmm2 = p[0] ^ qmul[q[0] ^ dq[0]] */ in raid6_datap_recov_ssse3()
277 asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0])); in raid6_datap_recov_ssse3()
278 asm volatile("movdqa %%xmm11, %0" : "=m" (dq[16])); in raid6_datap_recov_ssse3()
280 asm volatile("movdqa %%xmm2, %0" : "=m" (p[0])); in raid6_datap_recov_ssse3()
281 asm volatile("movdqa %%xmm12, %0" : "=m" (p[16])); in raid6_datap_recov_ssse3()
289 asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0])); in raid6_datap_recov_ssse3()
290 asm volatile("movdqa %0, %%xmm0" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
291 asm volatile("pxor %0, %%xmm3" : : "m" (q[0])); in raid6_datap_recov_ssse3()
292 asm volatile("movdqa %0, %%xmm1" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
297 asm volatile("movdqa %0, %%xmm2" : : "m" (p[0])); in raid6_datap_recov_ssse3()
311 asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0])); in raid6_datap_recov_ssse3()
312 asm volatile("movdqa %%xmm2, %0" : "=m" (p[0])); in raid6_datap_recov_ssse3()