Searched refs:vecDiff0 (Results 1 – 4 of 4) sorted by relevance
/cmsis-dsp-latest/Source/TransformFunctions/ |
D | arm_cfft_f16.c | 94 f16x8_t vecSum0, vecDiff0, vecSum1, vecDiff1; in _arm_radix4_butterfly_f16_mve() local 144 vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */ in _arm_radix4_butterfly_f16_mve() 171 vecTmp0 = MVE_CMPLX_SUB_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_f16_mve() 184 vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_f16_mve() 221 vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */ in _arm_radix4_butterfly_f16_mve() 239 vecTmp0 = MVE_CMPLX_SUB_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_f16_mve() 242 vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_f16_mve() 300 f16x8_t vecSum0, vecDiff0, vecSum1, vecDiff1; in _arm_radix4_butterfly_inverse_f16_mve() local 352 vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */ in _arm_radix4_butterfly_inverse_f16_mve() 378 vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_inverse_f16_mve() [all …]
|
D | arm_cfft_q31.c | 44 q31x4_t vecSum0, vecDiff0, vecSum1, vecDiff1; in _arm_radix4_butterfly_q31_mve() local 100 vecDiff0 = vhsubq(vecA, vecC); in _arm_radix4_butterfly_q31_mve() 126 vecTmp0 = MVE_CMPLX_SUB_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_q31_mve() 138 vecTmp0 = MVE_CMPLX_ADD_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_q31_mve() 185 vecDiff0 = vhsubq(vecA, vecC); in _arm_radix4_butterfly_q31_mve() 204 vecTmp0 = MVE_CMPLX_SUB_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_q31_mve() 207 vecTmp0 = MVE_CMPLX_ADD_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_q31_mve() 296 q31x4_t vecSum0, vecDiff0, vecSum1, vecDiff1; in _arm_radix4_butterfly_inverse_q31_mve() local 350 vecDiff0 = vhsubq(vecA, vecC); in _arm_radix4_butterfly_inverse_q31_mve() 376 vecTmp0 = MVE_CMPLX_ADD_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_inverse_q31_mve() [all …]
|
D | arm_cfft_q15.c | 42 q15x8_t vecSum0, vecDiff0, vecSum1, vecDiff1; in _arm_radix4_butterfly_q15_mve() local 96 vecDiff0 = vhsubq(vecA, vecC); in _arm_radix4_butterfly_q15_mve() 122 vecTmp0 = MVE_CMPLX_SUB_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_q15_mve() 135 vecTmp0 = MVE_CMPLX_ADD_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_q15_mve() 174 vecDiff0 = vhsubq(vecA, vecC); in _arm_radix4_butterfly_q15_mve() 193 vecTmp0 = MVE_CMPLX_SUB_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_q15_mve() 196 vecTmp0 = MVE_CMPLX_ADD_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_q15_mve() 278 q15x8_t vecSum0, vecDiff0, vecSum1, vecDiff1; in _arm_radix4_butterfly_inverse_q15_mve() local 334 vecDiff0 = vhsubq(vecA, vecC); in _arm_radix4_butterfly_inverse_q15_mve() 360 vecTmp0 = MVE_CMPLX_ADD_FX_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_inverse_q15_mve() [all …]
|
D | arm_cfft_f32.c | 95 f32x4_t vecSum0, vecDiff0, vecSum1, vecDiff1; in _arm_radix4_butterfly_f32_mve() local 147 vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */ in _arm_radix4_butterfly_f32_mve() 174 vecTmp0 = MVE_CMPLX_SUB_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_f32_mve() 187 vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_f32_mve() 224 vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */ in _arm_radix4_butterfly_f32_mve() 242 vecTmp0 = MVE_CMPLX_SUB_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_f32_mve() 245 vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_f32_mve() 303 f32x4_t vecSum0, vecDiff0, vecSum1, vecDiff1; in _arm_radix4_butterfly_inverse_f32_mve() local 355 vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */ in _arm_radix4_butterfly_inverse_f32_mve() 381 vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1); in _arm_radix4_butterfly_inverse_f32_mve() [all …]
|