Lines Matching refs:d

40 	int d, z, z0;  in raid6_avx21_gen_syndrome()  local
51 for (d = 0; d < bytes; d += 32) { in raid6_avx21_gen_syndrome()
52 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_avx21_gen_syndrome()
53 asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */ in raid6_avx21_gen_syndrome()
54 asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d])); in raid6_avx21_gen_syndrome()
56 asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z0-1][d])); in raid6_avx21_gen_syndrome()
58 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome()
65 asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome()
74 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx21_gen_syndrome()
76 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx21_gen_syndrome()
89 int d, z, z0; in raid6_avx21_xor_syndrome() local
99 for (d = 0 ; d < bytes ; d += 32) { in raid6_avx21_xor_syndrome()
100 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx21_xor_syndrome()
101 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx21_xor_syndrome()
110 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx21_xor_syndrome()
122 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx21_xor_syndrome()
124 asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d])); in raid6_avx21_xor_syndrome()
125 asm volatile("vmovdqa %%ymm2,%0" : "=m" (p[d])); in raid6_avx21_xor_syndrome()
147 int d, z, z0; in raid6_avx22_gen_syndrome() local
159 for (d = 0; d < bytes; d += 64) { in raid6_avx22_gen_syndrome()
160 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_avx22_gen_syndrome()
161 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d+32])); in raid6_avx22_gen_syndrome()
162 asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */ in raid6_avx22_gen_syndrome()
163 asm volatile("vmovdqa %0,%%ymm3" : : "m" (dptr[z0][d+32]));/* P[1] */ in raid6_avx22_gen_syndrome()
167 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx22_gen_syndrome()
168 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+32])); in raid6_avx22_gen_syndrome()
177 asm volatile("vmovdqa %0,%%ymm5" : : "m" (dptr[z][d])); in raid6_avx22_gen_syndrome()
178 asm volatile("vmovdqa %0,%%ymm7" : : "m" (dptr[z][d+32])); in raid6_avx22_gen_syndrome()
184 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx22_gen_syndrome()
185 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx22_gen_syndrome()
186 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx22_gen_syndrome()
187 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx22_gen_syndrome()
199 int d, z, z0; in raid6_avx22_xor_syndrome() local
209 for (d = 0 ; d < bytes ; d += 64) { in raid6_avx22_xor_syndrome()
210 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx22_xor_syndrome()
211 asm volatile("vmovdqa %0,%%ymm6" :: "m" (dptr[z0][d+32])); in raid6_avx22_xor_syndrome()
212 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx22_xor_syndrome()
213 asm volatile("vmovdqa %0,%%ymm3" : : "m" (p[d+32])); in raid6_avx22_xor_syndrome()
228 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx22_xor_syndrome()
230 :: "m" (dptr[z][d+32])); in raid6_avx22_xor_syndrome()
249 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx22_xor_syndrome()
250 asm volatile("vpxor %0,%%ymm6,%%ymm6" : : "m" (q[d+32])); in raid6_avx22_xor_syndrome()
252 asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d])); in raid6_avx22_xor_syndrome()
253 asm volatile("vmovdqa %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx22_xor_syndrome()
254 asm volatile("vmovdqa %%ymm2,%0" : "=m" (p[d])); in raid6_avx22_xor_syndrome()
255 asm volatile("vmovdqa %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx22_xor_syndrome()
279 int d, z, z0; in raid6_avx24_gen_syndrome() local
298 for (d = 0; d < bytes; d += 128) { in raid6_avx24_gen_syndrome()
300 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx24_gen_syndrome()
301 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+32])); in raid6_avx24_gen_syndrome()
302 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+64])); in raid6_avx24_gen_syndrome()
303 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+96])); in raid6_avx24_gen_syndrome()
320 asm volatile("vmovdqa %0,%%ymm5" : : "m" (dptr[z][d])); in raid6_avx24_gen_syndrome()
321 asm volatile("vmovdqa %0,%%ymm7" : : "m" (dptr[z][d+32])); in raid6_avx24_gen_syndrome()
322 asm volatile("vmovdqa %0,%%ymm13" : : "m" (dptr[z][d+64])); in raid6_avx24_gen_syndrome()
323 asm volatile("vmovdqa %0,%%ymm15" : : "m" (dptr[z][d+96])); in raid6_avx24_gen_syndrome()
333 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx24_gen_syndrome()
335 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx24_gen_syndrome()
337 asm volatile("vmovntdq %%ymm10,%0" : "=m" (p[d+64])); in raid6_avx24_gen_syndrome()
339 asm volatile("vmovntdq %%ymm11,%0" : "=m" (p[d+96])); in raid6_avx24_gen_syndrome()
341 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx24_gen_syndrome()
343 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx24_gen_syndrome()
345 asm volatile("vmovntdq %%ymm12,%0" : "=m" (q[d+64])); in raid6_avx24_gen_syndrome()
347 asm volatile("vmovntdq %%ymm14,%0" : "=m" (q[d+96])); in raid6_avx24_gen_syndrome()
360 int d, z, z0; in raid6_avx24_xor_syndrome() local
370 for (d = 0 ; d < bytes ; d += 128) { in raid6_avx24_xor_syndrome()
371 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx24_xor_syndrome()
372 asm volatile("vmovdqa %0,%%ymm6" :: "m" (dptr[z0][d+32])); in raid6_avx24_xor_syndrome()
373 asm volatile("vmovdqa %0,%%ymm12" :: "m" (dptr[z0][d+64])); in raid6_avx24_xor_syndrome()
374 asm volatile("vmovdqa %0,%%ymm14" :: "m" (dptr[z0][d+96])); in raid6_avx24_xor_syndrome()
375 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx24_xor_syndrome()
376 asm volatile("vmovdqa %0,%%ymm3" : : "m" (p[d+32])); in raid6_avx24_xor_syndrome()
377 asm volatile("vmovdqa %0,%%ymm10" : : "m" (p[d+64])); in raid6_avx24_xor_syndrome()
378 asm volatile("vmovdqa %0,%%ymm11" : : "m" (p[d+96])); in raid6_avx24_xor_syndrome()
385 asm volatile("prefetchnta %0" :: "m" (dptr[z][d])); in raid6_avx24_xor_syndrome()
386 asm volatile("prefetchnta %0" :: "m" (dptr[z][d+64])); in raid6_avx24_xor_syndrome()
407 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx24_xor_syndrome()
409 :: "m" (dptr[z][d+32])); in raid6_avx24_xor_syndrome()
411 :: "m" (dptr[z][d+64])); in raid6_avx24_xor_syndrome()
413 :: "m" (dptr[z][d+96])); in raid6_avx24_xor_syndrome()
423 asm volatile("prefetchnta %0" :: "m" (q[d])); in raid6_avx24_xor_syndrome()
424 asm volatile("prefetchnta %0" :: "m" (q[d+64])); in raid6_avx24_xor_syndrome()
448 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx24_xor_syndrome()
449 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx24_xor_syndrome()
450 asm volatile("vmovntdq %%ymm10,%0" : "=m" (p[d+64])); in raid6_avx24_xor_syndrome()
451 asm volatile("vmovntdq %%ymm11,%0" : "=m" (p[d+96])); in raid6_avx24_xor_syndrome()
452 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx24_xor_syndrome()
453 asm volatile("vpxor %0,%%ymm6,%%ymm6" : : "m" (q[d+32])); in raid6_avx24_xor_syndrome()
454 asm volatile("vpxor %0,%%ymm12,%%ymm12" : : "m" (q[d+64])); in raid6_avx24_xor_syndrome()
455 asm volatile("vpxor %0,%%ymm14,%%ymm14" : : "m" (q[d+96])); in raid6_avx24_xor_syndrome()
456 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx24_xor_syndrome()
457 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx24_xor_syndrome()
458 asm volatile("vmovntdq %%ymm12,%0" : "=m" (q[d+64])); in raid6_avx24_xor_syndrome()
459 asm volatile("vmovntdq %%ymm14,%0" : "=m" (q[d+96])); in raid6_avx24_xor_syndrome()