| /hal_nxp-latest/mcux/mcux-sdk/CMSIS/DSP/PrivateInclude/ |
| D | arm_vec_filtering.h | 38 #define MVE_INTR_CORR_QUAD_INC_X_FIXED_SIZE_F32(acc0, acc1, acc2, acc3, pX, pY, count)\ argument 86 acc0 = vecAddAcrossF32Mve(acc0Vec); \ 127 #define MVE_INTR_CORR_DUAL_INC_X_DEC_SIZE_F32(acc0, acc1, pX, pY, count)\ argument 167 acc0 = vecAddAcrossF32Mve(acc0Vec); \ 171 #define MVE_INTR_CORR_DUAL_INC_X_FIXED_SIZE_F32(acc0, acc1, pX, pY, count)\ argument 209 acc0 = vecAddAcrossF32Mve(acc0Vec); \ 213 #define MVE_INTR_CORR_DUAL_DEC_Y_INC_SIZE_F32(acc0, acc1, pX, pY, count)\ argument 252 acc0 = vecAddAcrossF32Mve(acc0Vec); \ 256 #define MVE_INTR_CONV_DUAL_INC_X_DEC_SIZE_F32(acc0, acc1, pX, pY, count) … argument 288 … acc0 = vecAddAcrossF32Mve(acc0Vec); \ [all …]
|
| /hal_nxp-latest/mcux/mcux-sdk/CMSIS/DSP/Source/MatrixFunctions/ |
| D | arm_mat_cmplx_mult_f32.c | 82 f32x4_t acc0, acc1; in arm_mat_cmplx_mult_f32_2x2_mve() local 96 acc0 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f32_2x2_mve() 97 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f32_2x2_mve() 103 pOut[0 * CMPLX_DIM * MATRIX_DIM2 + 0] = acc0[0] + acc0[2]; in arm_mat_cmplx_mult_f32_2x2_mve() 104 pOut[0 * CMPLX_DIM * MATRIX_DIM2 + 1] = acc0[1] + acc0[3]; in arm_mat_cmplx_mult_f32_2x2_mve() 117 acc0 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f32_2x2_mve() 118 acc0 = vcmlaq_rot90(acc0, vecA, vecB); in arm_mat_cmplx_mult_f32_2x2_mve() 124 pOut[0 * CMPLX_DIM * MATRIX_DIM2 + 0] = acc0[0] + acc0[2]; in arm_mat_cmplx_mult_f32_2x2_mve() 125 pOut[0 * CMPLX_DIM * MATRIX_DIM2 + 1] = acc0[1] + acc0[3]; in arm_mat_cmplx_mult_f32_2x2_mve() 147 f32x4_t acc0, acc1, acc2; in arm_mat_cmplx_mult_f32_3x3_mve() local [all …]
|
| D | arm_mat_mult_q31.c | 78 q63_t acc0, acc1; in arm_mat_mult_q31_2x2_mve() local 95 acc0 = vrmlaldavhq(vecA0, vecB); in arm_mat_mult_q31_2x2_mve() 98 acc0 = asrl(acc0, 23); in arm_mat_mult_q31_2x2_mve() 101 pOut[0 * MATRIX_DIM2] = (q31_t) acc0; in arm_mat_mult_q31_2x2_mve() 110 acc0 = vrmlaldavhq(vecA0, vecB); in arm_mat_mult_q31_2x2_mve() 113 acc0 = asrl(acc0, 23); in arm_mat_mult_q31_2x2_mve() 116 pOut[0 * MATRIX_DIM2] = (q31_t) acc0; in arm_mat_mult_q31_2x2_mve() 138 q63_t acc0, acc1, acc2; in arm_mat_mult_q31_3x3_mve() local 151 acc0 = vrmlaldavhq(vecA, vecB); in arm_mat_mult_q31_3x3_mve() 157 acc0 = asrl(acc0, 23); in arm_mat_mult_q31_3x3_mve() [all …]
|
| D | arm_mat_cmplx_mult_q31.c | 78 q63_t acc0, acc1, acc2, acc3; in arm_mat_cmplx_mult_q31_2x2_mve() local 92 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve() 99 pOut[0 * CMPLX_DIM * MATRIX_DIM2 + 0] = (q31_t) asrl(acc0, 31); in arm_mat_cmplx_mult_q31_2x2_mve() 110 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve() 119 pOut[0 * CMPLX_DIM * MATRIX_DIM2 + 0] = (q31_t) asrl(acc0, 31); in arm_mat_cmplx_mult_q31_2x2_mve() 141 q63_t acc0, acc1, acc2, acc3; in arm_mat_cmplx_mult_q31_3x3_mve() local 166 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve() 174 acc0 = vmlsldavaq_s32(acc0, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve() 181 pOut[0 * CMPLX_DIM * MATRIX_DIM3 + 0] = (q31_t) asrl(acc0, 31); in arm_mat_cmplx_mult_q31_3x3_mve() 187 acc0 = vmlsldavq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve() [all …]
|
| D | arm_mat_cmplx_mult_f16.c | 76 f16x8_t acc0, acc1; in arm_mat_cmplx_mult_f16_2x2_mve() local 104 acc0 = vcmulq(vecA0, vecB); in arm_mat_cmplx_mult_f16_2x2_mve() 105 acc0 = vcmlaq_rot90(acc0, vecA0, vecB); in arm_mat_cmplx_mult_f16_2x2_mve() 117 vecTmp = (f16x8_t) vrev64q_s32((int32x4_t) acc0); in arm_mat_cmplx_mult_f16_2x2_mve() 118 vecTmp = vaddq(vecTmp, acc0); in arm_mat_cmplx_mult_f16_2x2_mve() 151 f16x8_t acc0, acc1, acc2; in arm_mat_cmplx_mult_f16_3x3_mve() local 173 acc0 = vcmulq(vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve() 174 acc0 = vcmlaq_rot90(acc0, vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve() 182 mve_cmplx_sum_intra_vec_f16(acc0, &pOut[0 * CMPLX_DIM * MATRIX_DIM]); in arm_mat_cmplx_mult_f16_3x3_mve() 193 acc0 = vcmulq(vecA0, vecB); in arm_mat_cmplx_mult_f16_3x3_mve() [all …]
|
| D | arm_mat_mult_q15.c | 79 q63_t acc0, acc1; in arm_mat_mult_q15_2x2_mve() local 92 acc0 = vmlaldavq(vecA0, vecB); in arm_mat_mult_q15_2x2_mve() 95 acc0 = asrl(acc0, 15); in arm_mat_mult_q15_2x2_mve() 98 pOut[0 * MATRIX_DIM2] = (q15_t) __SSAT(acc0, 16); in arm_mat_mult_q15_2x2_mve() 107 acc0 = vmlaldavq(vecA0, vecB); in arm_mat_mult_q15_2x2_mve() 110 acc0 = asrl(acc0, 15); in arm_mat_mult_q15_2x2_mve() 113 pOut[0 * MATRIX_DIM2] = (q15_t) __SSAT(acc0, 16); in arm_mat_mult_q15_2x2_mve() 136 q63_t acc0, acc1, acc2; in arm_mat_mult_q15_3x3_mve() local 151 acc0 = vmlaldavq(vecA0, vecB); in arm_mat_mult_q15_3x3_mve() 155 acc0 = asrl(acc0, 15); in arm_mat_mult_q15_3x3_mve() [all …]
|
| D | arm_mat_mult_q7.c | 71 q31_t acc0, acc1; in arm_mat_mult_q7_2x2_mve() local 84 acc0 = vmladavq_s8(vecA0, vecB); in arm_mat_mult_q7_2x2_mve() 87 pOut[0 * MATRIX_DIM] = (q7_t) __SSAT(acc0 >> 7, 8); in arm_mat_mult_q7_2x2_mve() 96 acc0 = vmladavq_s8(vecA0, vecB); in arm_mat_mult_q7_2x2_mve() 99 pOut[0 * MATRIX_DIM] = (q7_t) __SSAT(acc0 >> 7, 8); in arm_mat_mult_q7_2x2_mve() 121 q31_t acc0, acc1, acc2; in arm_mat_mult_q7_3x3_mve() local 136 acc0 = vmladavq_s8(vecA0, vecB); in arm_mat_mult_q7_3x3_mve() 140 pOut[0 * MATRIX_DIM] = (q7_t) __SSAT(acc0 >> 7, 8); in arm_mat_mult_q7_3x3_mve() 150 acc0 = vmladavq_s8(vecA0, vecB); in arm_mat_mult_q7_3x3_mve() 154 pOut[0 * MATRIX_DIM] = (q7_t) __SSAT(acc0 >> 7, 8); in arm_mat_mult_q7_3x3_mve() [all …]
|
| D | arm_mat_vec_mult_f16.c | 79 f16x8_t vecIn, acc0, acc1, acc2, acc3; in arm_mat_vec_mult_f16() local 96 acc0 = vdupq_n_f16(0.0f); in arm_mat_vec_mult_f16() 115 acc0 = vfmaq(acc0, vecIn, vecA); in arm_mat_vec_mult_f16() 140 acc0 = vfmaq(acc0, vecIn, vecA); in arm_mat_vec_mult_f16() 151 *px++ = vecAddAcrossF16Mve(acc0); in arm_mat_vec_mult_f16() 169 f16x8_t vecIn, acc0, acc1; in arm_mat_vec_mult_f16() local 184 acc0 = vdupq_n_f16(0.0f); in arm_mat_vec_mult_f16() 198 acc0 = vfmaq(acc0, vecIn, vecA); in arm_mat_vec_mult_f16() 217 acc0 = vfmaq(acc0, vecIn, vecA); in arm_mat_vec_mult_f16() 224 *px++ = vecAddAcrossF16Mve(acc0); in arm_mat_vec_mult_f16() [all …]
|
| D | arm_mat_vec_mult_f32.c | 83 f32x4_t vecIn, acc0, acc1, acc2, acc3; in arm_mat_vec_mult_f32() local 100 acc0 = vdupq_n_f32(0.0f); in arm_mat_vec_mult_f32() 119 acc0 = vfmaq(acc0, vecIn, vecA); in arm_mat_vec_mult_f32() 144 acc0 = vfmaq(acc0, vecIn, vecA); in arm_mat_vec_mult_f32() 155 *px++ = vecAddAcrossF32Mve(acc0); in arm_mat_vec_mult_f32() 173 f32x4_t vecIn, acc0, acc1; in arm_mat_vec_mult_f32() local 188 acc0 = vdupq_n_f32(0.0f); in arm_mat_vec_mult_f32() 202 acc0 = vfmaq(acc0, vecIn, vecA); in arm_mat_vec_mult_f32() 221 acc0 = vfmaq(acc0, vecIn, vecA); in arm_mat_vec_mult_f32() 228 *px++ = vecAddAcrossF32Mve(acc0); in arm_mat_vec_mult_f32() [all …]
|
| D | arm_mat_vec_mult_q31.c | 74 q63_t acc0, acc1, acc2; in arm_mat_vec_mult_q31() local 86 acc0 = 0LL; in arm_mat_vec_mult_q31() 107 acc0 = vmlaldavaq(acc0, vecIn, vecMatA0); in arm_mat_vec_mult_q31() 127 acc0 = vmlaldavaq(acc0, vecIn, vecMatA0); in arm_mat_vec_mult_q31() 132 *px++ = asrl(acc0, 31); in arm_mat_vec_mult_q31() 150 q63_t acc0, acc1; in arm_mat_vec_mult_q31() local 165 acc0 = 0LL; in arm_mat_vec_mult_q31() 182 acc0 = vmlaldavaq(acc0, vecIn, vecMatA0); in arm_mat_vec_mult_q31() 201 acc0 = vmlaldavaq(acc0, vecIn, vecMatA0); in arm_mat_vec_mult_q31() 205 *px++ = asrl(acc0, 31); in arm_mat_vec_mult_q31() [all …]
|
| D | arm_mat_vec_mult_q15.c | 77 q63_t acc0, acc1, acc2; in arm_mat_vec_mult_q15() local 89 acc0 = 0LL; in arm_mat_vec_mult_q15() 110 acc0 = vmlaldavaq(acc0, vecIn, vecMatA0); in arm_mat_vec_mult_q15() 130 acc0 = vmlaldavaq(acc0, vecIn, vecMatA0); in arm_mat_vec_mult_q15() 135 *px++ = MVE_ASRL_SAT16(acc0, 15); in arm_mat_vec_mult_q15() 153 q63_t acc0, acc1; in arm_mat_vec_mult_q15() local 168 acc0 = 0LL; in arm_mat_vec_mult_q15() 185 acc0 = vmlaldavaq(acc0, vecIn, vecMatA0); in arm_mat_vec_mult_q15() 204 acc0 = vmlaldavaq(acc0, vecIn, vecMatA0); in arm_mat_vec_mult_q15() 208 *px++ = MVE_ASRL_SAT16(acc0, 15); in arm_mat_vec_mult_q15() [all …]
|
| /hal_nxp-latest/mcux/mcux-sdk/CMSIS/DSP/Source/FilteringFunctions/ |
| D | arm_fir_q31.c | 235 q63_t acc0=0, acc1=0, acc2=0, acc3=0; in arm_fir_q31_1_4_mve() local 254 acc0 = vrmlaldavhq(vecIn0, vecCoeffs); in arm_fir_q31_1_4_mve() 265 acc0 = asrl(acc0, 23); in arm_fir_q31_1_4_mve() 270 *pOutput++ = (q31_t) acc0; in arm_fir_q31_1_4_mve() 295 acc0 = vrmlaldavhq(vecIn0, vecCoeffs); in arm_fir_q31_1_4_mve() 303 acc0 = asrl(acc0, 23); in arm_fir_q31_1_4_mve() 307 *pOutput++ = (q31_t) acc0; in arm_fir_q31_1_4_mve() 323 acc0 = vrmlaldavhq(vecIn0, vecCoeffs); in arm_fir_q31_1_4_mve() 328 acc0 = asrl(acc0, 23); in arm_fir_q31_1_4_mve() 331 *pOutput++ = (q31_t) acc0; in arm_fir_q31_1_4_mve() [all …]
|
| D | arm_fir_interpolate_q15.c | 118 q63_t acc0 = 0LL; in arm_fir_interpolate_q15() local 129 acc0 = vmlaldavaq(acc0, vecState, vecCoef); in arm_fir_interpolate_q15() 152 acc0 = vmlaldavaq(acc0, vecState, vecCoef); in arm_fir_interpolate_q15() 164 acc0 = asrl(acc0, 15); in arm_fir_interpolate_q15() 169 *pDst++ = (q15_t) __SSAT(acc0, 16); in arm_fir_interpolate_q15() 182 q63_t acc0 = 0LL; in arm_fir_interpolate_q15() local 192 acc0 = vmlaldavaq(acc0, vecState, vecCoef); in arm_fir_interpolate_q15() 212 acc0 = vmlaldavaq(acc0, vecState, vecCoef); in arm_fir_interpolate_q15() 221 acc0 = asrl(acc0, 15); in arm_fir_interpolate_q15() 225 *pDst++ = (q15_t) __SSAT(acc0, 16);; in arm_fir_interpolate_q15() [all …]
|
| D | arm_fir_interpolate_q31.c | 115 q63_t acc0 = 0LL; in arm_fir_interpolate_q31() local 126 acc0 = vrmlaldavhaq(acc0, vecState, vecCoef); in arm_fir_interpolate_q31() 149 acc0 = vrmlaldavhaq(acc0, vecState, vecCoef); in arm_fir_interpolate_q31() 161 acc0 = asrl(acc0, 31 - 8); in arm_fir_interpolate_q31() 166 *pDst++ = (q31_t) acc0; in arm_fir_interpolate_q31() 179 q63_t acc0 = 0LL; in arm_fir_interpolate_q31() local 189 acc0 = vrmlaldavhaq(acc0, vecState, vecCoef); in arm_fir_interpolate_q31() 209 acc0 = vrmlaldavhaq(acc0, vecState, vecCoef); in arm_fir_interpolate_q31() 218 acc0 = asrl(acc0, 31 - 8); in arm_fir_interpolate_q31() 222 *pDst++ = (q31_t) acc0; in arm_fir_interpolate_q31() [all …]
|
| D | arm_fir_decimate_fast_q31.c | 74 q63_t acc0; /* Accumulator */ in arm_fir_decimate_fast_q31() local 106 acc0 = 0; in arm_fir_decimate_fast_q31() 138 acc0 = (q31_t) ((((q63_t) acc0 << 32) + ((q63_t) x0 * c0)) >> 32); in arm_fir_decimate_fast_q31() 153 acc0 = (q31_t) ((((q63_t) acc0 << 32) + ((q63_t) x0 * c0)) >> 32); in arm_fir_decimate_fast_q31() 168 acc0 = (q31_t) ((((q63_t) acc0 << 32) + ((q63_t) x0 * c0)) >> 32); in arm_fir_decimate_fast_q31() 183 acc0 = (q31_t) ((((q63_t) acc0 << 32) + ((q63_t) x0 * c0)) >> 32); in arm_fir_decimate_fast_q31() 207 acc0 = (q31_t) ((((q63_t) acc0 << 32) + ((q63_t) x0 * c0)) >> 32); in arm_fir_decimate_fast_q31() 221 *pDst++ = (q31_t) (acc0 << 1); in arm_fir_decimate_fast_q31() 252 acc0 = 0; in arm_fir_decimate_fast_q31() 274 acc0 = (q31_t) ((((q63_t) acc0 << 32) + ((q63_t) x0 * c0)) >> 32); in arm_fir_decimate_fast_q31() [all …]
|
| D | arm_fir_q15.c | 216 q63_t acc0, acc1, acc2, acc3; in arm_fir_q15() local 251 acc0 = 0LL; in arm_fir_q15() 272 acc0 = vmlaldavaq(acc0, vecIn0, vecCoeffs); in arm_fir_q15() 291 *pOutput++ = (q15_t) MVE_ASRL_SAT16(acc0, 15); in arm_fir_q15() 311 acc0 = 0LL; in arm_fir_q15() 331 acc0 = vmlaldavaq(acc0, vecIn0, vecCoeffs); in arm_fir_q15() 347 acc0 = asrl(acc0, 15); in arm_fir_q15() 351 *pOutput++ = (q15_t) MVE_ASRL_SAT16(acc0, 15); in arm_fir_q15() 362 acc0 = 0LL; in arm_fir_q15() 380 acc0 = vmlaldavaq(acc0, vecIn0, vecCoeffs); in arm_fir_q15() [all …]
|
| D | arm_fir_q7.c | 207 q31_t acc0, acc1, acc2, acc3; in arm_fir_q7() local 261 acc0 = 0; in arm_fir_q7() 284 acc0 = vmladavaq(acc0, vecIn0, vecCoeffs); in arm_fir_q7() 305 *pOutput++ = (q7_t) __SSAT((acc0 >> 7U), 8); in arm_fir_q7() 325 acc0 = 0; in arm_fir_q7() 341 acc0 = vmladavaq(acc0, vecIn0, vecCoeffs); in arm_fir_q7() 354 *pOutput++ = (q7_t) __SSAT((acc0 >> 7U), 8); in arm_fir_q7() 365 acc0 = 0; in arm_fir_q7() 380 acc0 = vmladavaq(acc0, vecIn0, vecCoeffs); in arm_fir_q7() 390 *pOutput++ = (q7_t) __SSAT((acc0 >> 7U), 8); in arm_fir_q7() [all …]
|
| D | arm_conv_q15.c | 116 int64_t acc0 = 0LL; in arm_conv_q15() local 122 MVE_INTR_CONV_DUAL_INC_Y_INC_SIZE_Q15(acc0, acc1, pX, pY, count); in arm_conv_q15() 123 *pDst++ = (q15_t) acc0; in arm_conv_q15() 143 int64_t acc0 = 0LL; in arm_conv_q15() local 155 MVE_INTR_CONV_QUAD_INC_X_FIXED_SIZE_Q15(acc0, acc1, acc2, acc3, pX, pY, count); in arm_conv_q15() 156 *pDst++ = (q15_t) acc0; in arm_conv_q15() 166 int64_t acc0 = 0LL; in arm_conv_q15() local 176 MVE_INTR_CONV_DUAL_INC_X_FIXED_SIZE_Q15(acc0, acc1, pX, pY, count); in arm_conv_q15() 177 *pDst++ = (q15_t) acc0; in arm_conv_q15() 198 int64_t acc0 = 0LL; in arm_conv_q15() local [all …]
|
| D | arm_correlate_q15.c | 136 int64_t acc0 = 0LL; in arm_correlate_q15() local 146 MVE_INTR_CORR_DUAL_DEC_Y_INC_SIZE_Q15(acc0, acc1, pX, pY, count); in arm_correlate_q15() 148 *pDst = (q15_t) acc0; in arm_correlate_q15() 170 int64_t acc0 = 0LL; in arm_correlate_q15() local 182 MVE_INTR_CORR_QUAD_INC_X_FIXED_SIZE_Q15(acc0, acc1, acc2, acc3, pX, pY, srcBLen); in arm_correlate_q15() 184 *pDst = (q15_t) acc0; in arm_correlate_q15() 197 int64_t acc0 = 0LL; in arm_correlate_q15() local 207 MVE_INTR_CORR_DUAL_INC_X_FIXED_SIZE_Q15(acc0, acc1, pX, pY, srcBLen); in arm_correlate_q15() 209 *pDst = (q15_t) acc0; in arm_correlate_q15() 233 int64_t acc0 = 0LL; in arm_correlate_q15() local [all …]
|
| /hal_nxp-latest/mcux/mcux-sdk/CMSIS/DSP/Source/SVMFunctions/ |
| D | arm_svm_linear_predict_f16.c | 80 f16x8_t vecIn, acc0, acc1, acc2, acc3; in arm_svm_linear_predict_f16() local 97 acc0 = vdupq_n_f16(0.0f); in arm_svm_linear_predict_f16() 115 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_linear_predict_f16() 139 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_linear_predict_f16() 150 acc0 = vmulq_n_f16(acc0,*pDualCoef++); in arm_svm_linear_predict_f16() 151 acc0 = vfmaq_n_f16(acc0,acc1,*pDualCoef++); in arm_svm_linear_predict_f16() 152 acc0 = vfmaq_n_f16(acc0,acc2,*pDualCoef++); in arm_svm_linear_predict_f16() 153 acc0 = vfmaq_n_f16(acc0,acc3,*pDualCoef++); in arm_svm_linear_predict_f16() 155 sum += (_Float16)vecAddAcrossF16Mve(acc0); in arm_svm_linear_predict_f16() 169 f16x8_t vecIn, acc0, acc1; in arm_svm_linear_predict_f16() local [all …]
|
| D | arm_svm_linear_predict_f32.c | 77 f32x4_t vecIn, acc0, acc1, acc2, acc3; in arm_svm_linear_predict_f32() local 94 acc0 = vdupq_n_f32(0.0f); in arm_svm_linear_predict_f32() 112 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_linear_predict_f32() 136 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_linear_predict_f32() 148 acc0 = vmulq_n_f32(acc0,*pDualCoef++); in arm_svm_linear_predict_f32() 149 acc0 = vfmaq_n_f32(acc0,acc1,*pDualCoef++); in arm_svm_linear_predict_f32() 150 acc0 = vfmaq_n_f32(acc0,acc2,*pDualCoef++); in arm_svm_linear_predict_f32() 151 acc0 = vfmaq_n_f32(acc0,acc3,*pDualCoef++); in arm_svm_linear_predict_f32() 153 sum += vecAddAcrossF32Mve(acc0); in arm_svm_linear_predict_f32() 167 f32x4_t vecIn, acc0, acc1; in arm_svm_linear_predict_f32() local [all …]
|
| D | arm_svm_sigmoid_predict_f16.c | 83 f16x8_t vecIn, acc0, acc1, acc2, acc3; in arm_svm_sigmoid_predict_f16() local 100 acc0 = vdupq_n_f16(0.0f); in arm_svm_sigmoid_predict_f16() 118 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_sigmoid_predict_f16() 142 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_sigmoid_predict_f16() 154 vtmp = vsetq_lane(vecAddAcrossF16Mve(acc0), vtmp, 0); in arm_svm_sigmoid_predict_f16() 177 f16x8_t vecIn, acc0, acc1; in arm_svm_sigmoid_predict_f16() local 192 acc0 = vdupq_n_f16(0.0f); in arm_svm_sigmoid_predict_f16() 205 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_sigmoid_predict_f16() 223 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_sigmoid_predict_f16() 231 vtmp = vsetq_lane(vecAddAcrossF16Mve(acc0), vtmp, 0); in arm_svm_sigmoid_predict_f16() [all …]
|
| D | arm_svm_polynomial_predict_f16.c | 83 f16x8_t vecIn, acc0, acc1, acc2, acc3; in arm_svm_polynomial_predict_f16() local 100 acc0 = vdupq_n_f16(0.0f); in arm_svm_polynomial_predict_f16() 118 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_polynomial_predict_f16() 142 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_polynomial_predict_f16() 154 vtmp = vsetq_lane(vecAddAcrossF16Mve(acc0), vtmp, 0); in arm_svm_polynomial_predict_f16() 178 f16x8_t vecIn, acc0, acc1; in arm_svm_polynomial_predict_f16() local 193 acc0 = vdupq_n_f16(0.0f); in arm_svm_polynomial_predict_f16() 206 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_polynomial_predict_f16() 224 acc0 = vfmaq(acc0, vecIn, vecA); in arm_svm_polynomial_predict_f16() 232 vtmp = vsetq_lane(vecAddAcrossF16Mve(acc0), vtmp, 0); in arm_svm_polynomial_predict_f16() [all …]
|
| /hal_nxp-latest/mcux/mcux-sdk/CMSIS/DSP/Source/ComplexMathFunctions/ |
| D | arm_cmplx_mag_q15.c | 65 q31_t acc0; in arm_cmplx_mag_q15() local 97 acc0 = __SMUAD(in, in); in arm_cmplx_mag_q15() 100 arm_sqrt_q15((q15_t) (acc0 >> 17), pDst++); in arm_cmplx_mag_q15() 118 q31_t acc0; /* Accumulators */ in arm_cmplx_mag_q15() local 121 q31_t acc0, acc1; /* Accumulators */ in arm_cmplx_mag_q15() local 135 acc0 = __SMUAD(in, in); in arm_cmplx_mag_q15() 137 arm_sqrt_q15((q15_t) (acc0 >> 17), pDst++); in arm_cmplx_mag_q15() 140 acc0 = __SMUAD(in, in); in arm_cmplx_mag_q15() 141 arm_sqrt_q15((q15_t) (acc0 >> 17), pDst++); in arm_cmplx_mag_q15() 144 acc0 = __SMUAD(in, in); in arm_cmplx_mag_q15() [all …]
|
| D | arm_cmplx_mag_squared_q15.c | 61 q31_t acc0; /* Accumulators */ in arm_cmplx_mag_squared_q15() local 96 acc0 = __SMUAD(in, in); in arm_cmplx_mag_squared_q15() 99 *pDst++ = (q15_t) (acc0 >> 17); in arm_cmplx_mag_squared_q15() 118 q31_t acc0; /* Accumulators */ in arm_cmplx_mag_squared_q15() local 121 q31_t acc0, acc1; /* Accumulators */ in arm_cmplx_mag_squared_q15() local 135 acc0 = __SMUAD(in, in); in arm_cmplx_mag_squared_q15() 137 *pDst++ = (q15_t) (acc0 >> 17); in arm_cmplx_mag_squared_q15() 140 acc0 = __SMUAD(in, in); in arm_cmplx_mag_squared_q15() 141 *pDst++ = (q15_t) (acc0 >> 17); in arm_cmplx_mag_squared_q15() 144 acc0 = __SMUAD(in, in); in arm_cmplx_mag_squared_q15() [all …]
|