Lines Matching full:volatile
56 asm volatile("movdqa %0,%%xmm7" : : "m" (x0f[0])); in raid6_2data_recov_ssse3()
59 asm volatile("movdqa %0,%%xmm6" : : "m" (qmul[0])); in raid6_2data_recov_ssse3()
60 asm volatile("movdqa %0,%%xmm14" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3()
61 asm volatile("movdqa %0,%%xmm15" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3()
69 asm volatile("movdqa %0,%%xmm1" : : "m" (q[0])); in raid6_2data_recov_ssse3()
70 asm volatile("movdqa %0,%%xmm9" : : "m" (q[16])); in raid6_2data_recov_ssse3()
71 asm volatile("movdqa %0,%%xmm0" : : "m" (p[0])); in raid6_2data_recov_ssse3()
72 asm volatile("movdqa %0,%%xmm8" : : "m" (p[16])); in raid6_2data_recov_ssse3()
73 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3()
74 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3()
75 asm volatile("pxor %0,%%xmm0" : : "m" (dp[0])); in raid6_2data_recov_ssse3()
76 asm volatile("pxor %0,%%xmm8" : : "m" (dp[16])); in raid6_2data_recov_ssse3()
80 asm volatile("movdqa %xmm6,%xmm4"); in raid6_2data_recov_ssse3()
81 asm volatile("movdqa %0,%%xmm5" : : "m" (qmul[16])); in raid6_2data_recov_ssse3()
82 asm volatile("movdqa %xmm6,%xmm12"); in raid6_2data_recov_ssse3()
83 asm volatile("movdqa %xmm5,%xmm13"); in raid6_2data_recov_ssse3()
84 asm volatile("movdqa %xmm1,%xmm3"); in raid6_2data_recov_ssse3()
85 asm volatile("movdqa %xmm9,%xmm11"); in raid6_2data_recov_ssse3()
86 asm volatile("movdqa %xmm0,%xmm2"); /* xmm2/10 = px */ in raid6_2data_recov_ssse3()
87 asm volatile("movdqa %xmm8,%xmm10"); in raid6_2data_recov_ssse3()
88 asm volatile("psraw $4,%xmm1"); in raid6_2data_recov_ssse3()
89 asm volatile("psraw $4,%xmm9"); in raid6_2data_recov_ssse3()
90 asm volatile("pand %xmm7,%xmm3"); in raid6_2data_recov_ssse3()
91 asm volatile("pand %xmm7,%xmm11"); in raid6_2data_recov_ssse3()
92 asm volatile("pand %xmm7,%xmm1"); in raid6_2data_recov_ssse3()
93 asm volatile("pand %xmm7,%xmm9"); in raid6_2data_recov_ssse3()
94 asm volatile("pshufb %xmm3,%xmm4"); in raid6_2data_recov_ssse3()
95 asm volatile("pshufb %xmm11,%xmm12"); in raid6_2data_recov_ssse3()
96 asm volatile("pshufb %xmm1,%xmm5"); in raid6_2data_recov_ssse3()
97 asm volatile("pshufb %xmm9,%xmm13"); in raid6_2data_recov_ssse3()
98 asm volatile("pxor %xmm4,%xmm5"); in raid6_2data_recov_ssse3()
99 asm volatile("pxor %xmm12,%xmm13"); in raid6_2data_recov_ssse3()
103 asm volatile("movdqa %xmm14,%xmm4"); in raid6_2data_recov_ssse3()
104 asm volatile("movdqa %xmm15,%xmm1"); in raid6_2data_recov_ssse3()
105 asm volatile("movdqa %xmm14,%xmm12"); in raid6_2data_recov_ssse3()
106 asm volatile("movdqa %xmm15,%xmm9"); in raid6_2data_recov_ssse3()
107 asm volatile("movdqa %xmm2,%xmm3"); in raid6_2data_recov_ssse3()
108 asm volatile("movdqa %xmm10,%xmm11"); in raid6_2data_recov_ssse3()
109 asm volatile("psraw $4,%xmm2"); in raid6_2data_recov_ssse3()
110 asm volatile("psraw $4,%xmm10"); in raid6_2data_recov_ssse3()
111 asm volatile("pand %xmm7,%xmm3"); in raid6_2data_recov_ssse3()
112 asm volatile("pand %xmm7,%xmm11"); in raid6_2data_recov_ssse3()
113 asm volatile("pand %xmm7,%xmm2"); in raid6_2data_recov_ssse3()
114 asm volatile("pand %xmm7,%xmm10"); in raid6_2data_recov_ssse3()
115 asm volatile("pshufb %xmm3,%xmm4"); in raid6_2data_recov_ssse3()
116 asm volatile("pshufb %xmm11,%xmm12"); in raid6_2data_recov_ssse3()
117 asm volatile("pshufb %xmm2,%xmm1"); in raid6_2data_recov_ssse3()
118 asm volatile("pshufb %xmm10,%xmm9"); in raid6_2data_recov_ssse3()
119 asm volatile("pxor %xmm4,%xmm1"); in raid6_2data_recov_ssse3()
120 asm volatile("pxor %xmm12,%xmm9"); in raid6_2data_recov_ssse3()
123 asm volatile("pxor %xmm5,%xmm1"); in raid6_2data_recov_ssse3()
124 asm volatile("pxor %xmm13,%xmm9"); in raid6_2data_recov_ssse3()
126 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3()
127 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3()
129 asm volatile("pxor %xmm1,%xmm0"); in raid6_2data_recov_ssse3()
130 asm volatile("pxor %xmm9,%xmm8"); in raid6_2data_recov_ssse3()
131 asm volatile("movdqa %%xmm0,%0" : "=m" (dp[0])); in raid6_2data_recov_ssse3()
132 asm volatile("movdqa %%xmm8,%0" : "=m" (dp[16])); in raid6_2data_recov_ssse3()
140 asm volatile("movdqa %0,%%xmm1" : : "m" (*q)); in raid6_2data_recov_ssse3()
141 asm volatile("movdqa %0,%%xmm0" : : "m" (*p)); in raid6_2data_recov_ssse3()
142 asm volatile("pxor %0,%%xmm1" : : "m" (*dq)); in raid6_2data_recov_ssse3()
143 asm volatile("pxor %0,%%xmm0" : : "m" (*dp)); in raid6_2data_recov_ssse3()
148 asm volatile("movdqa %0,%%xmm4" : : "m" (qmul[0])); in raid6_2data_recov_ssse3()
149 asm volatile("movdqa %0,%%xmm5" : : "m" (qmul[16])); in raid6_2data_recov_ssse3()
151 asm volatile("movdqa %xmm1,%xmm3"); in raid6_2data_recov_ssse3()
152 asm volatile("psraw $4,%xmm1"); in raid6_2data_recov_ssse3()
153 asm volatile("pand %xmm7,%xmm3"); in raid6_2data_recov_ssse3()
154 asm volatile("pand %xmm7,%xmm1"); in raid6_2data_recov_ssse3()
155 asm volatile("pshufb %xmm3,%xmm4"); in raid6_2data_recov_ssse3()
156 asm volatile("pshufb %xmm1,%xmm5"); in raid6_2data_recov_ssse3()
157 asm volatile("pxor %xmm4,%xmm5"); in raid6_2data_recov_ssse3()
159 asm volatile("movdqa %xmm0,%xmm2"); /* xmm2 = px */ in raid6_2data_recov_ssse3()
163 asm volatile("movdqa %0,%%xmm4" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3()
164 asm volatile("movdqa %0,%%xmm1" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3()
165 asm volatile("movdqa %xmm2,%xmm3"); in raid6_2data_recov_ssse3()
166 asm volatile("psraw $4,%xmm2"); in raid6_2data_recov_ssse3()
167 asm volatile("pand %xmm7,%xmm3"); in raid6_2data_recov_ssse3()
168 asm volatile("pand %xmm7,%xmm2"); in raid6_2data_recov_ssse3()
169 asm volatile("pshufb %xmm3,%xmm4"); in raid6_2data_recov_ssse3()
170 asm volatile("pshufb %xmm2,%xmm1"); in raid6_2data_recov_ssse3()
171 asm volatile("pxor %xmm4,%xmm1"); in raid6_2data_recov_ssse3()
174 asm volatile("pxor %xmm5,%xmm1"); in raid6_2data_recov_ssse3()
176 asm volatile("movdqa %%xmm1,%0" : "=m" (*dq)); in raid6_2data_recov_ssse3()
178 asm volatile("pxor %xmm1,%xmm0"); in raid6_2data_recov_ssse3()
179 asm volatile("movdqa %%xmm0,%0" : "=m" (*dp)); in raid6_2data_recov_ssse3()
222 asm volatile("movdqa %0, %%xmm7" : : "m" (x0f[0])); in raid6_datap_recov_ssse3()
226 asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0])); in raid6_datap_recov_ssse3()
227 asm volatile("movdqa %0, %%xmm4" : : "m" (dq[16])); in raid6_datap_recov_ssse3()
228 asm volatile("pxor %0, %%xmm3" : : "m" (q[0])); in raid6_datap_recov_ssse3()
229 asm volatile("movdqa %0, %%xmm0" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
233 asm volatile("pxor %0, %%xmm4" : : "m" (q[16])); in raid6_datap_recov_ssse3()
234 asm volatile("movdqa %0, %%xmm1" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
238 asm volatile("movdqa %xmm3, %xmm6"); in raid6_datap_recov_ssse3()
239 asm volatile("movdqa %xmm4, %xmm8"); in raid6_datap_recov_ssse3()
243 asm volatile("psraw $4, %xmm3"); in raid6_datap_recov_ssse3()
244 asm volatile("pand %xmm7, %xmm6"); in raid6_datap_recov_ssse3()
245 asm volatile("pand %xmm7, %xmm3"); in raid6_datap_recov_ssse3()
246 asm volatile("pshufb %xmm6, %xmm0"); in raid6_datap_recov_ssse3()
247 asm volatile("pshufb %xmm3, %xmm1"); in raid6_datap_recov_ssse3()
248 asm volatile("movdqa %0, %%xmm10" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
249 asm volatile("pxor %xmm0, %xmm1"); in raid6_datap_recov_ssse3()
250 asm volatile("movdqa %0, %%xmm11" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
254 asm volatile("psraw $4, %xmm4"); in raid6_datap_recov_ssse3()
255 asm volatile("pand %xmm7, %xmm8"); in raid6_datap_recov_ssse3()
256 asm volatile("pand %xmm7, %xmm4"); in raid6_datap_recov_ssse3()
257 asm volatile("pshufb %xmm8, %xmm10"); in raid6_datap_recov_ssse3()
258 asm volatile("pshufb %xmm4, %xmm11"); in raid6_datap_recov_ssse3()
259 asm volatile("movdqa %0, %%xmm2" : : "m" (p[0])); in raid6_datap_recov_ssse3()
260 asm volatile("pxor %xmm10, %xmm11"); in raid6_datap_recov_ssse3()
261 asm volatile("movdqa %0, %%xmm12" : : "m" (p[16])); in raid6_datap_recov_ssse3()
265 asm volatile("pxor %xmm1, %xmm2"); in raid6_datap_recov_ssse3()
269 asm volatile("pxor %xmm11, %xmm12"); in raid6_datap_recov_ssse3()
273 asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0])); in raid6_datap_recov_ssse3()
274 asm volatile("movdqa %%xmm11, %0" : "=m" (dq[16])); in raid6_datap_recov_ssse3()
276 asm volatile("movdqa %%xmm2, %0" : "=m" (p[0])); in raid6_datap_recov_ssse3()
277 asm volatile("movdqa %%xmm12, %0" : "=m" (p[16])); in raid6_datap_recov_ssse3()
285 asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0])); in raid6_datap_recov_ssse3()
286 asm volatile("movdqa %0, %%xmm0" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
287 asm volatile("pxor %0, %%xmm3" : : "m" (q[0])); in raid6_datap_recov_ssse3()
288 asm volatile("movdqa %0, %%xmm1" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
292 asm volatile("movdqa %xmm3, %xmm6"); in raid6_datap_recov_ssse3()
293 asm volatile("movdqa %0, %%xmm2" : : "m" (p[0])); in raid6_datap_recov_ssse3()
294 asm volatile("psraw $4, %xmm3"); in raid6_datap_recov_ssse3()
295 asm volatile("pand %xmm7, %xmm6"); in raid6_datap_recov_ssse3()
296 asm volatile("pand %xmm7, %xmm3"); in raid6_datap_recov_ssse3()
297 asm volatile("pshufb %xmm6, %xmm0"); in raid6_datap_recov_ssse3()
298 asm volatile("pshufb %xmm3, %xmm1"); in raid6_datap_recov_ssse3()
299 asm volatile("pxor %xmm0, %xmm1"); in raid6_datap_recov_ssse3()
303 asm volatile("pxor %xmm1, %xmm2"); in raid6_datap_recov_ssse3()
307 asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0])); in raid6_datap_recov_ssse3()
308 asm volatile("movdqa %%xmm2, %0" : "=m" (p[0])); in raid6_datap_recov_ssse3()