Lines Matching refs:d

46 	int d, z, z0;  in raid6_avx21_gen_syndrome()  local
57 for (d = 0; d < bytes; d += 32) { in raid6_avx21_gen_syndrome()
58 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_avx21_gen_syndrome()
59 asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */ in raid6_avx21_gen_syndrome()
60 asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d])); in raid6_avx21_gen_syndrome()
62 asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z0-1][d])); in raid6_avx21_gen_syndrome()
64 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome()
71 asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome()
80 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx21_gen_syndrome()
82 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx21_gen_syndrome()
95 int d, z, z0; in raid6_avx21_xor_syndrome() local
105 for (d = 0 ; d < bytes ; d += 32) { in raid6_avx21_xor_syndrome()
106 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx21_xor_syndrome()
107 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx21_xor_syndrome()
116 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx21_xor_syndrome()
128 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx21_xor_syndrome()
130 asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d])); in raid6_avx21_xor_syndrome()
131 asm volatile("vmovdqa %%ymm2,%0" : "=m" (p[d])); in raid6_avx21_xor_syndrome()
153 int d, z, z0; in raid6_avx22_gen_syndrome() local
165 for (d = 0; d < bytes; d += 64) { in raid6_avx22_gen_syndrome()
166 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_avx22_gen_syndrome()
167 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d+32])); in raid6_avx22_gen_syndrome()
168 asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */ in raid6_avx22_gen_syndrome()
169 asm volatile("vmovdqa %0,%%ymm3" : : "m" (dptr[z0][d+32]));/* P[1] */ in raid6_avx22_gen_syndrome()
173 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx22_gen_syndrome()
174 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+32])); in raid6_avx22_gen_syndrome()
183 asm volatile("vmovdqa %0,%%ymm5" : : "m" (dptr[z][d])); in raid6_avx22_gen_syndrome()
184 asm volatile("vmovdqa %0,%%ymm7" : : "m" (dptr[z][d+32])); in raid6_avx22_gen_syndrome()
190 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx22_gen_syndrome()
191 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx22_gen_syndrome()
192 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx22_gen_syndrome()
193 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx22_gen_syndrome()
205 int d, z, z0; in raid6_avx22_xor_syndrome() local
215 for (d = 0 ; d < bytes ; d += 64) { in raid6_avx22_xor_syndrome()
216 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx22_xor_syndrome()
217 asm volatile("vmovdqa %0,%%ymm6" :: "m" (dptr[z0][d+32])); in raid6_avx22_xor_syndrome()
218 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx22_xor_syndrome()
219 asm volatile("vmovdqa %0,%%ymm3" : : "m" (p[d+32])); in raid6_avx22_xor_syndrome()
234 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx22_xor_syndrome()
236 :: "m" (dptr[z][d+32])); in raid6_avx22_xor_syndrome()
255 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx22_xor_syndrome()
256 asm volatile("vpxor %0,%%ymm6,%%ymm6" : : "m" (q[d+32])); in raid6_avx22_xor_syndrome()
258 asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d])); in raid6_avx22_xor_syndrome()
259 asm volatile("vmovdqa %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx22_xor_syndrome()
260 asm volatile("vmovdqa %%ymm2,%0" : "=m" (p[d])); in raid6_avx22_xor_syndrome()
261 asm volatile("vmovdqa %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx22_xor_syndrome()
285 int d, z, z0; in raid6_avx24_gen_syndrome() local
304 for (d = 0; d < bytes; d += 128) { in raid6_avx24_gen_syndrome()
306 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx24_gen_syndrome()
307 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+32])); in raid6_avx24_gen_syndrome()
308 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+64])); in raid6_avx24_gen_syndrome()
309 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+96])); in raid6_avx24_gen_syndrome()
326 asm volatile("vmovdqa %0,%%ymm5" : : "m" (dptr[z][d])); in raid6_avx24_gen_syndrome()
327 asm volatile("vmovdqa %0,%%ymm7" : : "m" (dptr[z][d+32])); in raid6_avx24_gen_syndrome()
328 asm volatile("vmovdqa %0,%%ymm13" : : "m" (dptr[z][d+64])); in raid6_avx24_gen_syndrome()
329 asm volatile("vmovdqa %0,%%ymm15" : : "m" (dptr[z][d+96])); in raid6_avx24_gen_syndrome()
339 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx24_gen_syndrome()
341 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx24_gen_syndrome()
343 asm volatile("vmovntdq %%ymm10,%0" : "=m" (p[d+64])); in raid6_avx24_gen_syndrome()
345 asm volatile("vmovntdq %%ymm11,%0" : "=m" (p[d+96])); in raid6_avx24_gen_syndrome()
347 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx24_gen_syndrome()
349 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx24_gen_syndrome()
351 asm volatile("vmovntdq %%ymm12,%0" : "=m" (q[d+64])); in raid6_avx24_gen_syndrome()
353 asm volatile("vmovntdq %%ymm14,%0" : "=m" (q[d+96])); in raid6_avx24_gen_syndrome()
366 int d, z, z0; in raid6_avx24_xor_syndrome() local
376 for (d = 0 ; d < bytes ; d += 128) { in raid6_avx24_xor_syndrome()
377 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx24_xor_syndrome()
378 asm volatile("vmovdqa %0,%%ymm6" :: "m" (dptr[z0][d+32])); in raid6_avx24_xor_syndrome()
379 asm volatile("vmovdqa %0,%%ymm12" :: "m" (dptr[z0][d+64])); in raid6_avx24_xor_syndrome()
380 asm volatile("vmovdqa %0,%%ymm14" :: "m" (dptr[z0][d+96])); in raid6_avx24_xor_syndrome()
381 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx24_xor_syndrome()
382 asm volatile("vmovdqa %0,%%ymm3" : : "m" (p[d+32])); in raid6_avx24_xor_syndrome()
383 asm volatile("vmovdqa %0,%%ymm10" : : "m" (p[d+64])); in raid6_avx24_xor_syndrome()
384 asm volatile("vmovdqa %0,%%ymm11" : : "m" (p[d+96])); in raid6_avx24_xor_syndrome()
391 asm volatile("prefetchnta %0" :: "m" (dptr[z][d])); in raid6_avx24_xor_syndrome()
392 asm volatile("prefetchnta %0" :: "m" (dptr[z][d+64])); in raid6_avx24_xor_syndrome()
413 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx24_xor_syndrome()
415 :: "m" (dptr[z][d+32])); in raid6_avx24_xor_syndrome()
417 :: "m" (dptr[z][d+64])); in raid6_avx24_xor_syndrome()
419 :: "m" (dptr[z][d+96])); in raid6_avx24_xor_syndrome()
429 asm volatile("prefetchnta %0" :: "m" (q[d])); in raid6_avx24_xor_syndrome()
430 asm volatile("prefetchnta %0" :: "m" (q[d+64])); in raid6_avx24_xor_syndrome()
454 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx24_xor_syndrome()
455 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx24_xor_syndrome()
456 asm volatile("vmovntdq %%ymm10,%0" : "=m" (p[d+64])); in raid6_avx24_xor_syndrome()
457 asm volatile("vmovntdq %%ymm11,%0" : "=m" (p[d+96])); in raid6_avx24_xor_syndrome()
458 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx24_xor_syndrome()
459 asm volatile("vpxor %0,%%ymm6,%%ymm6" : : "m" (q[d+32])); in raid6_avx24_xor_syndrome()
460 asm volatile("vpxor %0,%%ymm12,%%ymm12" : : "m" (q[d+64])); in raid6_avx24_xor_syndrome()
461 asm volatile("vpxor %0,%%ymm14,%%ymm14" : : "m" (q[d+96])); in raid6_avx24_xor_syndrome()
462 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx24_xor_syndrome()
463 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx24_xor_syndrome()
464 asm volatile("vmovntdq %%ymm12,%0" : "=m" (q[d+64])); in raid6_avx24_xor_syndrome()
465 asm volatile("vmovntdq %%ymm14,%0" : "=m" (q[d+96])); in raid6_avx24_xor_syndrome()