Lines Matching refs:vecB

77     f16x8_t        vecB, vecA0, vecA1;  in arm_mat_cmplx_mult_f16_2x2_mve()  local
102 vecB = vldrhq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_f16_2x2_mve()
104 acc0 = vcmulq(vecA0, vecB); in arm_mat_cmplx_mult_f16_2x2_mve()
105 acc0 = vcmlaq_rot90(acc0, vecA0, vecB); in arm_mat_cmplx_mult_f16_2x2_mve()
107 acc1 = vcmulq(vecA1, vecB); in arm_mat_cmplx_mult_f16_2x2_mve()
108 acc1 = vcmlaq_rot90(acc1, vecA1, vecB); in arm_mat_cmplx_mult_f16_2x2_mve()
153 f16x8_t vecB, vecA0, vecA1, vecA2; in arm_mat_cmplx_mult_f16_3x3_mve() local
172 vecB = vldrhq_gather_shifted_offset_z(pInB, vecColBOffs0, p0); in arm_mat_cmplx_mult_f16_3x3_mve()
174 acc0 = vcmulq(vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
175 acc0 = vcmlaq_rot90(acc0, vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
177 acc1 = vcmulq(vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
178 acc1 = vcmlaq_rot90(acc1, vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
180 acc2 = vcmulq(vecA2, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
181 acc2 = vcmlaq_rot90(acc2, vecA2, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
192 vecB = vldrhq_gather_shifted_offset_z(pInB, vecColBOffs0, p0); in arm_mat_cmplx_mult_f16_3x3_mve()
194 acc0 = vcmulq(vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
195 acc0 = vcmlaq_rot90(acc0, vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
197 acc1 = vcmulq(vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
198 acc1 = vcmlaq_rot90(acc1, vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
200 acc2 = vcmulq(vecA2, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
201 acc2 = vcmlaq_rot90(acc2, vecA2, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
212 vecB = vldrhq_gather_shifted_offset_z(pInB, vecColBOffs0, p0); in arm_mat_cmplx_mult_f16_3x3_mve()
214 acc0 = vcmulq(vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
215 acc0 = vcmlaq_rot90(acc0, vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
217 acc1 = vcmulq(vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
218 acc1 = vcmlaq_rot90(acc1, vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
220 acc2 = vcmulq(vecA2, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
221 acc2 = vcmlaq_rot90(acc2, vecA2, vecB); in arm_mat_cmplx_mult_f16_3x3_mve()
251 f16x8_t vecB, vecA; in arm_mat_cmplx_mult_f16_4x4_mve() local
262 vecB = vldrhq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_f16_4x4_mve()
265 acc0 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
266 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
269 acc1 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
270 acc1 = vcmlaq_rot90(acc1, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
273 acc2 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
274 acc2 = vcmlaq_rot90(acc2, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
277 acc3 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
278 acc3 = vcmlaq_rot90(acc3, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
291 vecB = vldrhq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_f16_4x4_mve()
294 acc0 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
295 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
298 acc1 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
299 acc1 = vcmlaq_rot90(acc1, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
302 acc2 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
303 acc2 = vcmlaq_rot90(acc2, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
306 acc3 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
307 acc3 = vcmlaq_rot90(acc3, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
320 vecB = vldrhq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_f16_4x4_mve()
323 acc0 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
324 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
327 acc1 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
328 acc1 = vcmlaq_rot90(acc1, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
331 acc2 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
332 acc2 = vcmlaq_rot90(acc2, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
335 acc3 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
336 acc3 = vcmlaq_rot90(acc3, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
349 vecB = vldrhq_gather_shifted_offset(pInB, vecColBOffs0); in arm_mat_cmplx_mult_f16_4x4_mve()
352 acc0 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
353 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
356 acc1 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
357 acc1 = vcmlaq_rot90(acc1, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
360 acc2 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
361 acc2 = vcmlaq_rot90(acc2, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
364 acc3 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
365 acc3 = vcmlaq_rot90(acc3, vecA, vecB); in arm_mat_cmplx_mult_f16_4x4_mve()
503 f16x8_t vecB, vecA; in arm_mat_cmplx_mult_f16() local
505 vecB = vldrhq_gather_shifted_offset_f16(pInB, vecOffs); in arm_mat_cmplx_mult_f16()
512 acc0 = vcmlaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16()
513 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16()
516 acc1 = vcmlaq(acc1, vecA, vecB); in arm_mat_cmplx_mult_f16()
517 acc1 = vcmlaq_rot90(acc1, vecA, vecB); in arm_mat_cmplx_mult_f16()
520 acc2 = vcmlaq(acc2, vecA, vecB); in arm_mat_cmplx_mult_f16()
521 acc2 = vcmlaq_rot90(acc2, vecA, vecB); in arm_mat_cmplx_mult_f16()
524 acc3 = vcmlaq(acc3, vecA, vecB); in arm_mat_cmplx_mult_f16()
525 acc3 = vcmlaq_rot90(acc3, vecA, vecB); in arm_mat_cmplx_mult_f16()
540 f16x8_t vecB, vecA; in arm_mat_cmplx_mult_f16() local
542 vecB = vldrhq_gather_shifted_offset_z_f16(pInB, vecOffs, p0); in arm_mat_cmplx_mult_f16()
549 acc0 = vcmlaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16()
550 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16()
553 acc1 = vcmlaq(acc1, vecA, vecB); in arm_mat_cmplx_mult_f16()
554 acc1 = vcmlaq_rot90(acc1, vecA, vecB); in arm_mat_cmplx_mult_f16()
557 acc2 = vcmlaq(acc2, vecA, vecB); in arm_mat_cmplx_mult_f16()
558 acc2 = vcmlaq_rot90(acc2, vecA, vecB); in arm_mat_cmplx_mult_f16()
561 acc3 = vcmlaq(acc3, vecA, vecB); in arm_mat_cmplx_mult_f16()
562 acc3 = vcmlaq_rot90(acc3, vecA, vecB); in arm_mat_cmplx_mult_f16()
639 f16x8_t vecB, vecA; in arm_mat_cmplx_mult_f16() local
641 vecB = vldrhq_gather_shifted_offset(pInB, vecOffs); in arm_mat_cmplx_mult_f16()
649 acc0 = vcmlaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16()
650 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16()
664 f16x8_t vecB, vecA; in arm_mat_cmplx_mult_f16() local
666 vecB = vldrhq_gather_shifted_offset_z(pInB, vecOffs, p0); in arm_mat_cmplx_mult_f16()
669 acc0 = vcmlaq(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16()
670 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f16()