Lines Matching refs:MMAU__X3
39 #define MMAU__X3 0xF000400CUL /*!< Accumulator register X3 */ macro
428 *((uint32_t volatile *)(MMAU__UMUL | MMAU__X3)) = lval2; in MMAU_d_umul_ll()
446 *((uint32_t volatile *)(MMAU__UMULD | MMAU__X3)) = lval; in MMAU_d_umul_dl()
465 *((uint32_t volatile *)(MMAU__UMULD | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_umuls_dl()
482 *((uint32_t volatile *)(MMAU__UMULDA | MMAU__X3)) = lval; in MMAU_d_umula_l()
501 *((uint32_t volatile *)(MMAU__UMULDA | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_umulas_l()
521 *((uint32_t volatile *)(MMAU__UMAC | MMAU__X3)) = lval2; in MMAU_d_umac_ll()
542 *((uint32_t volatile *)(MMAU__UMAC | MMAU__X3 | MMAU__SAT)) = lval2; in MMAU_d_umacs_ll()
563 *((uint32_t volatile *)(MMAU__UMACD | MMAU__X3)) = lval; in MMAU_d_umac_dl()
585 *((uint32_t volatile *)(MMAU__UMACD | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_umacs_dl()
606 *((uint32_t volatile *)(MMAU__UMACDA | MMAU__X3)) = lval; in MMAU_d_umaca_dl()
628 *((uint32_t volatile *)(MMAU__UMACDA | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_umacas_dl()
646 *((uint32_t volatile *)(MMAU__UDIV | MMAU__X3)) = lden; in MMAU_l_udiv_ll()
664 *((uint32_t volatile *)(MMAU__UDIVD | MMAU__X3)) = lden; in MMAU_d_udiv_dl()
699 *((uint32_t volatile *)(MMAU__UDIVDA | MMAU__X3)) = lden1; in MMAU_d_udiva_l()
732 *((uint32_t volatile *)(MMAU__USQR | MMAU__X3)) = lrad; in MMAU_l_usqr_l()
765 *((uint32_t volatile *)(MMAU__USQR | MMAU__X3)) = lrad; in MMAU_s_usqr_l()
814 *((int32_t volatile *)(MMAU__SMUL | MMAU__X3)) = lval2; in MMAU_d_smul_ll()
832 *((int32_t volatile *)(MMAU__SMULD | MMAU__X3)) = lval; in MMAU_d_smul_dl()
851 *((int32_t volatile *)(MMAU__SMULD | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_smuls_dl()
868 *((int32_t volatile *)(MMAU__SMULDA | MMAU__X3)) = lval; in MMAU_d_smula_l()
887 *((int32_t volatile *)(MMAU__SMULDA | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_smulas_l()
907 *((int32_t volatile *)(MMAU__SMAC | MMAU__X3)) = lval2; in MMAU_d_smac_ll()
928 *((int32_t volatile *)(MMAU__SMAC | MMAU__X3 | MMAU__SAT)) = lval2; in MMAU_d_smacs_ll()
949 *((int32_t volatile *)(MMAU__SMACD | MMAU__X3)) = lval; in MMAU_d_smac_dl()
973 *((int32_t volatile *)(MMAU__SMACD | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_smacs_dl()
993 *((int32_t volatile *)(MMAU__SMACDA | MMAU__X3)) = lval; in MMAU_d_smaca_dl()
1017 *((int32_t volatile *)(MMAU__SMACDA | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_smacas_dl()
1035 *((int32_t volatile *)(MMAU__SDIV | MMAU__X3)) = lden; in MMAU_l_sdiv_ll()
1054 *((int32_t volatile *)(MMAU__SDIV | MMAU__X3 | MMAU__SAT)) = lden; in MMAU_l_sdivs_ll()
1072 *((int32_t volatile *)(MMAU__SDIVD | MMAU__X3)) = lden; in MMAU_d_sdiv_dl()
1091 *((int32_t volatile *)(MMAU__SDIVD | MMAU__X3 | MMAU__SAT)) = lden; in MMAU_d_sdivs_dl()
1145 *((int32_t volatile *)(MMAU__SDIVDA | MMAU__X3)) = lden1; in MMAU_d_sdiva_l()
1163 *((int32_t volatile *)(MMAU__SDIVDA | MMAU__X3 | MMAU__SAT)) = lden1; in MMAU_d_sdivas_l()
1235 *((frac32_t volatile *)(MMAU__QMUL | MMAU__X3)) = lval2; in MMAU_l_mul_ll()
1254 *((frac32_t volatile *)(MMAU__QMUL | MMAU__X3 | MMAU__SAT)) = lval2; in MMAU_l_muls_ll()
1272 *((frac32_t volatile *)(MMAU__QMUL | MMAU__X3)) = lval2; in MMAU_d_mul_ll()
1293 *((frac32_t volatile *)(MMAU__QMUL | MMAU__X3 | MMAU__SAT)) = lval2; in MMAU_d_muls_ll()
1311 *((frac32_t volatile *)(MMAU__QMULD | MMAU__X3)) = lval; in MMAU_d_mul_dl()
1332 *((frac32_t volatile *)(MMAU__QMULD | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_muls_dl()
1349 *((frac32_t volatile *)(MMAU__QMULDA | MMAU__X3)) = lval; in MMAU_d_mula_l()
1370 *((frac32_t volatile *)(MMAU__QMULDA | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_mulas_l()
1388 *((frac32_t volatile *)(MMAU__QMULD | MMAU__X3)) = lval; in MMAU_l_mul_dl()
1408 *((frac32_t volatile *)(MMAU__QMULD | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_l_muls_dl()
1426 *((frac32_t volatile *)(MMAU__QMULDA | MMAU__X3)) = lval; in MMAU_l_mula_l()
1445 *((frac32_t volatile *)(MMAU__QMULDA | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_l_mulas_l()
1465 *((frac32_t volatile *)(MMAU__QMAC | MMAU__X3)) = lval2; in MMAU_d_mac_ll()
1488 *((frac32_t volatile *)(MMAU__QMAC | MMAU__X3 | MMAU__SAT)) = lval2; in MMAU_d_macs_ll()
1509 *((frac32_t volatile *)(MMAU__QMACD | MMAU__X3)) = lval; in MMAU_d_mac_dl()
1533 *((frac32_t volatile *)(MMAU__QMACD | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_macs_dl()
1554 *((frac32_t volatile *)(MMAU__QMACDA | MMAU__X3)) = lval; in MMAU_d_maca_dl()
1578 *((frac32_t volatile *)(MMAU__QMACDA | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_d_macas_dl()
1598 *((frac32_t volatile *)(MMAU__QMAC | MMAU__X3)) = lval2; in MMAU_l_mac_ll()
1619 *((frac32_t volatile *)(MMAU__QMAC | MMAU__X3 | MMAU__SAT)) = lval2; in MMAU_l_macs_ll()
1640 *((frac32_t volatile *)(MMAU__QMACD | MMAU__X3)) = lval; in MMAU_l_mac_dl()
1662 *((frac32_t volatile *)(MMAU__QMACD | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_l_macs_dl()
1683 *((frac32_t volatile *)(MMAU__QMACDA | MMAU__X3)) = lval; in MMAU_l_maca_dl()
1705 *((frac32_t volatile *)(MMAU__QMACDA | MMAU__X3 | MMAU__SAT)) = lval; in MMAU_l_macas_dl()
1723 *((frac32_t volatile *)(MMAU__QDIV | MMAU__X3)) = lden; in MMAU_l_div_ll()
1744 *((frac32_t volatile *)(MMAU__QDIV | MMAU__X3 | MMAU__SAT)) = lden; in MMAU_l_divs_ll()
1763 *((frac32_t volatile *)(MMAU__QDIVDA | MMAU__X3 | MMAU__SAT)) = lden; in MMAU_l_divas_l()
1781 *((frac32_t volatile *)(MMAU__QDIVD | MMAU__X3)) = lden; in MMAU_d_div_dl()
1802 *((frac32_t volatile *)(MMAU__QDIVD | MMAU__X3 | MMAU__SAT)) = lden; in MMAU_d_divs_dl()
1819 *((frac32_t volatile *)(MMAU__QDIVDA | MMAU__X3)) = lden1; in MMAU_d_diva_l()
1839 *((frac32_t volatile *)(MMAU__QDIVDA | MMAU__X3 | MMAU__SAT)) = lden1; in MMAU_d_divas_l()
1857 *((frac32_t volatile *)(MMAU__QDIVDA | MMAU__X3)) = lden; in MMAU_l_diva_l()
1875 *((frac32_t volatile *)(MMAU__QSQRD | MMAU__X3)) = lrad; in MMAU_l_sqr_l()