Lines Matching refs:vecA

79     q31x4_t          vecB, vecA;  in arm_mat_cmplx_mult_q31_2x2_mve()  local
91 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_2x2_mve()
92 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
93 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
95 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_2x2_mve()
96 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
97 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
109 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_2x2_mve()
110 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
111 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
113 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_2x2_mve()
114 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
115 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
142 q31x4_t vecB, vecB1, vecA; in arm_mat_cmplx_mult_q31_3x3_mve() local
165 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_3x3_mve()
166 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
167 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
169 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_3x3_mve()
170 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
171 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
173 vecA = vldrwq_z_s32(&pInA0[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
174 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
175 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
177 vecA = vldrwq_z_s32(&pInA1[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
178 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
179 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
186 vecA = vldrwq_s32(pInA2); in arm_mat_cmplx_mult_q31_3x3_mve()
187 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
188 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
190 vecA = vldrwq_z_s32(&pInA2[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
191 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
192 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
206 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_3x3_mve()
207 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
208 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
210 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_3x3_mve()
211 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
212 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
214 vecA = vldrwq_z_s32(&pInA0[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
215 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
216 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
218 vecA = vldrwq_z_s32(&pInA1[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
219 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
220 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
227 vecA = vldrwq_s32(pInA2); in arm_mat_cmplx_mult_q31_3x3_mve()
228 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
229 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
231 vecA = vldrwq_z_s32(&pInA2[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
232 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
233 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
247 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_3x3_mve()
248 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
249 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
251 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_3x3_mve()
252 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
253 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
255 vecA = vldrwq_z_s32(&pInA0[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
256 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
257 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
259 vecA = vldrwq_z_s32(&pInA1[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
260 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
261 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
268 vecA = vldrwq_s32(pInA2); in arm_mat_cmplx_mult_q31_3x3_mve()
269 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
270 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
272 vecA = vldrwq_z_s32(&pInA2[4], p0); in arm_mat_cmplx_mult_q31_3x3_mve()
273 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
274 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
298 q31x4_t vecB, vecB1, vecA; in arm_mat_cmplx_mult_q31_4x4_mve() local
317 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_4x4_mve()
318 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
319 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
321 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_4x4_mve()
322 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
323 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
325 vecA = vldrwq_s32(&pInA0[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
326 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
327 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
329 vecA = vldrwq_s32(&pInA1[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
330 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
331 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
338 vecA = vldrwq_s32(pInA2); in arm_mat_cmplx_mult_q31_4x4_mve()
339 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
340 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
342 vecA = vldrwq_s32(pInA3); in arm_mat_cmplx_mult_q31_4x4_mve()
343 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
344 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
346 vecA = vldrwq_s32(&pInA2[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
347 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
348 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
350 vecA = vldrwq_s32(&pInA3[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
351 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
352 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
368 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_4x4_mve()
369 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
370 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
372 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_4x4_mve()
373 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
374 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
376 vecA = vldrwq_s32(&pInA0[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
377 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
378 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
380 vecA = vldrwq_s32(&pInA1[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
381 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
382 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
389 vecA = vldrwq_s32(pInA2); in arm_mat_cmplx_mult_q31_4x4_mve()
390 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
391 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
393 vecA = vldrwq_s32(pInA3); in arm_mat_cmplx_mult_q31_4x4_mve()
394 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
395 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
397 vecA = vldrwq_s32(&pInA2[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
398 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
399 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
401 vecA = vldrwq_s32(&pInA3[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
402 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
403 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
418 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_4x4_mve()
419 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
420 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
422 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_4x4_mve()
423 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
424 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
426 vecA = vldrwq_s32(&pInA0[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
427 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
428 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
430 vecA = vldrwq_s32(&pInA1[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
431 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
432 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
439 vecA = vldrwq_s32(pInA2); in arm_mat_cmplx_mult_q31_4x4_mve()
440 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
441 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
443 vecA = vldrwq_s32(pInA3); in arm_mat_cmplx_mult_q31_4x4_mve()
444 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
445 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
447 vecA = vldrwq_s32(&pInA2[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
448 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
449 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
451 vecA = vldrwq_s32(&pInA3[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
452 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
453 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
469 vecA = vldrwq_s32(pInA0); in arm_mat_cmplx_mult_q31_4x4_mve()
470 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
471 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
473 vecA = vldrwq_s32(pInA1); in arm_mat_cmplx_mult_q31_4x4_mve()
474 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
475 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
477 vecA = vldrwq_s32(&pInA0[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
478 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
479 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
481 vecA = vldrwq_s32(&pInA1[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
482 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
483 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
490 vecA = vldrwq_s32(pInA2); in arm_mat_cmplx_mult_q31_4x4_mve()
491 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
492 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
494 vecA = vldrwq_s32(pInA3); in arm_mat_cmplx_mult_q31_4x4_mve()
495 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
496 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
498 vecA = vldrwq_s32(&pInA2[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
499 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
500 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
502 vecA = vldrwq_s32(&pInA3[4]); in arm_mat_cmplx_mult_q31_4x4_mve()
503 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
504 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
637 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31() local
646 vecA = vld1q(pSrcA0Vec); in arm_mat_cmplx_mult_q31()
648 acc0 = vmlsldavaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_q31()
649 acc1 = vmlaldavaxq(acc1, vecA, vecB); in arm_mat_cmplx_mult_q31()
652 vecA = vld1q(pSrcA1Vec); in arm_mat_cmplx_mult_q31()
655 acc2 = vmlsldavaq(acc2, vecA, vecB); in arm_mat_cmplx_mult_q31()
656 acc3 = vmlaldavaxq(acc3, vecA, vecB); in arm_mat_cmplx_mult_q31()
670 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31() local
680 vecA = vld1q(pSrcA0Vec); in arm_mat_cmplx_mult_q31()
681 acc0 = vmlsldavaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_q31()
682 acc1 = vmlaldavaxq(acc1, vecA, vecB); in arm_mat_cmplx_mult_q31()
683 vecA = vld1q(pSrcA1Vec); in arm_mat_cmplx_mult_q31()
684 acc2 = vmlsldavaq(acc2, vecA, vecB); in arm_mat_cmplx_mult_q31()
685 acc3 = vmlaldavaxq(acc3, vecA, vecB); in arm_mat_cmplx_mult_q31()
762 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31() local
770 vecA = vld1q(pSrcA0Vec); in arm_mat_cmplx_mult_q31()
772 acc0 = vmlsldavaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_q31()
773 acc1 = vmlaldavaxq(acc1, vecA, vecB); in arm_mat_cmplx_mult_q31()
787 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31() local
796 vecA = vld1q(pSrcA0Vec); in arm_mat_cmplx_mult_q31()
799 acc0 = vmlsldavaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_q31()
800 acc1 = vmlaldavaxq(acc1, vecA, vecB); in arm_mat_cmplx_mult_q31()