Lines Matching refs:vecB1

142     q31x4_t          vecB, vecB1, vecA;  in arm_mat_cmplx_mult_q31_3x3_mve()  local
163 vecB1 = vldrwq_gather_shifted_offset(pInB, vecColBOffs1); in arm_mat_cmplx_mult_q31_3x3_mve()
174 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
175 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
178 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
179 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
191 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
192 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
204 vecB1 = vldrwq_gather_shifted_offset(pInB, vecColBOffs1); in arm_mat_cmplx_mult_q31_3x3_mve()
215 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
216 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
219 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
220 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
232 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
233 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
245 vecB1 = vldrwq_gather_shifted_offset(pInB, vecColBOffs1); in arm_mat_cmplx_mult_q31_3x3_mve()
256 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
257 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
260 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
261 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
273 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
274 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve()
298 q31x4_t vecB, vecB1, vecA; in arm_mat_cmplx_mult_q31_4x4_mve() local
315 vecB1 = vldrwq_gather_shifted_offset(pInB, vecColBOffs1); in arm_mat_cmplx_mult_q31_4x4_mve()
326 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
327 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
330 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
331 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
347 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
348 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
351 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
352 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
366 vecB1 = vldrwq_gather_shifted_offset(pInB, vecColBOffs1); in arm_mat_cmplx_mult_q31_4x4_mve()
377 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
378 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
381 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
382 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
398 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
399 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
402 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
403 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
416 vecB1 = vldrwq_gather_shifted_offset(pInB, vecColBOffs1); in arm_mat_cmplx_mult_q31_4x4_mve()
427 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
428 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
431 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
432 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
448 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
449 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
452 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
453 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
467 vecB1 = vldrwq_gather_shifted_offset(pInB, vecColBOffs1); in arm_mat_cmplx_mult_q31_4x4_mve()
478 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
479 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
482 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
483 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
499 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
500 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
503 acc2 = vmlsldavaq_s32(acc2, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()
504 acc3 = vmlaldavaxq_s32(acc3, vecA, vecB1); in arm_mat_cmplx_mult_q31_4x4_mve()