Lines Matching refs:vecB
79 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31_2x2_mve() local
90 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_2x2_mve()
92 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
93 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
96 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
97 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
108 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_2x2_mve()
110 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
111 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
114 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
115 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve()
142 q31x4_t vecB, vecB1, vecA; in arm_mat_cmplx_mult_q31_3x3_mve() local
162 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_3x3_mve()
166 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
167 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
170 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
171 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
187 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
188 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
203 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_3x3_mve()
207 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
208 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
211 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
212 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
228 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
229 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
244 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_3x3_mve()
248 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
249 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
252 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
253 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
269 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
270 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve()
298 q31x4_t vecB, vecB1, vecA; in arm_mat_cmplx_mult_q31_4x4_mve() local
314 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_4x4_mve()
318 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
319 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
322 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
323 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
339 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
340 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
343 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
344 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
365 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_4x4_mve()
369 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
370 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
373 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
374 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
390 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
391 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
394 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
395 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
415 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_4x4_mve()
419 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
420 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
423 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
424 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
440 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
441 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
444 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
445 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
466 vecB = vldrwq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_q31_4x4_mve()
470 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
471 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
474 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
475 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
491 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
492 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
495 acc2 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
496 acc3 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_4x4_mve()
637 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31() local
639 vecB = vldrwq_gather_shifted_offset(pInB, vecOffs); in arm_mat_cmplx_mult_q31()
648 acc0 = vmlsldavaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_q31()
649 acc1 = vmlaldavaxq(acc1, vecA, vecB); in arm_mat_cmplx_mult_q31()
655 acc2 = vmlsldavaq(acc2, vecA, vecB); in arm_mat_cmplx_mult_q31()
656 acc3 = vmlaldavaxq(acc3, vecA, vecB); in arm_mat_cmplx_mult_q31()
670 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31() local
672 vecB = vldrwq_gather_shifted_offset_z(pInB, vecOffs, p0); in arm_mat_cmplx_mult_q31()
681 acc0 = vmlsldavaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_q31()
682 acc1 = vmlaldavaxq(acc1, vecA, vecB); in arm_mat_cmplx_mult_q31()
684 acc2 = vmlsldavaq(acc2, vecA, vecB); in arm_mat_cmplx_mult_q31()
685 acc3 = vmlaldavaxq(acc3, vecA, vecB); in arm_mat_cmplx_mult_q31()
762 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31() local
764 vecB = vldrwq_gather_shifted_offset(pInB, vecOffs); in arm_mat_cmplx_mult_q31()
772 acc0 = vmlsldavaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_q31()
773 acc1 = vmlaldavaxq(acc1, vecA, vecB); in arm_mat_cmplx_mult_q31()
787 q31x4_t vecB, vecA; in arm_mat_cmplx_mult_q31() local
789 vecB = vldrwq_gather_shifted_offset_z(pInB, vecOffs, p0); in arm_mat_cmplx_mult_q31()
799 acc0 = vmlsldavaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_q31()
800 acc1 = vmlaldavaxq(acc1, vecA, vecB); in arm_mat_cmplx_mult_q31()