Lines Matching refs:op2
258 __STATIC_FORCEINLINE uint32_t __ROR(uint32_t op1, uint32_t op2) in __ROR() argument
260 op2 %= 32U; in __ROR()
261 if (op2 == 0U) in __ROR()
265 return (op1 >> op2) | (op1 << (32U - op2)); in __ROR()
1398 __STATIC_FORCEINLINE uint32_t __SADD8(uint32_t op1, uint32_t op2) in __SADD8() argument
1402 __ASM volatile ("sadd8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SADD8()
1406 __STATIC_FORCEINLINE uint32_t __QADD8(uint32_t op1, uint32_t op2) in __QADD8() argument
1410 __ASM volatile ("qadd8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __QADD8()
1414 __STATIC_FORCEINLINE uint32_t __SHADD8(uint32_t op1, uint32_t op2) in __SHADD8() argument
1418 __ASM volatile ("shadd8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SHADD8()
1422 __STATIC_FORCEINLINE uint32_t __UADD8(uint32_t op1, uint32_t op2) in __UADD8() argument
1426 __ASM volatile ("uadd8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UADD8()
1430 __STATIC_FORCEINLINE uint32_t __UQADD8(uint32_t op1, uint32_t op2) in __UQADD8() argument
1434 __ASM volatile ("uqadd8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UQADD8()
1438 __STATIC_FORCEINLINE uint32_t __UHADD8(uint32_t op1, uint32_t op2) in __UHADD8() argument
1442 __ASM volatile ("uhadd8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UHADD8()
1447 __STATIC_FORCEINLINE uint32_t __SSUB8(uint32_t op1, uint32_t op2) in __SSUB8() argument
1451 __ASM volatile ("ssub8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SSUB8()
1455 __STATIC_FORCEINLINE uint32_t __QSUB8(uint32_t op1, uint32_t op2) in __QSUB8() argument
1459 __ASM volatile ("qsub8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __QSUB8()
1463 __STATIC_FORCEINLINE uint32_t __SHSUB8(uint32_t op1, uint32_t op2) in __SHSUB8() argument
1467 __ASM volatile ("shsub8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SHSUB8()
1471 __STATIC_FORCEINLINE uint32_t __USUB8(uint32_t op1, uint32_t op2) in __USUB8() argument
1475 __ASM volatile ("usub8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __USUB8()
1479 __STATIC_FORCEINLINE uint32_t __UQSUB8(uint32_t op1, uint32_t op2) in __UQSUB8() argument
1483 __ASM volatile ("uqsub8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UQSUB8()
1487 __STATIC_FORCEINLINE uint32_t __UHSUB8(uint32_t op1, uint32_t op2) in __UHSUB8() argument
1491 __ASM volatile ("uhsub8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UHSUB8()
1496 __STATIC_FORCEINLINE uint32_t __SADD16(uint32_t op1, uint32_t op2) in __SADD16() argument
1500 __ASM volatile ("sadd16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SADD16()
1504 __STATIC_FORCEINLINE uint32_t __QADD16(uint32_t op1, uint32_t op2) in __QADD16() argument
1508 __ASM volatile ("qadd16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __QADD16()
1512 __STATIC_FORCEINLINE uint32_t __SHADD16(uint32_t op1, uint32_t op2) in __SHADD16() argument
1516 __ASM volatile ("shadd16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SHADD16()
1520 __STATIC_FORCEINLINE uint32_t __UADD16(uint32_t op1, uint32_t op2) in __UADD16() argument
1524 __ASM volatile ("uadd16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UADD16()
1528 __STATIC_FORCEINLINE uint32_t __UQADD16(uint32_t op1, uint32_t op2) in __UQADD16() argument
1532 __ASM volatile ("uqadd16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UQADD16()
1536 __STATIC_FORCEINLINE uint32_t __UHADD16(uint32_t op1, uint32_t op2) in __UHADD16() argument
1540 __ASM volatile ("uhadd16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UHADD16()
1544 __STATIC_FORCEINLINE uint32_t __SSUB16(uint32_t op1, uint32_t op2) in __SSUB16() argument
1548 __ASM volatile ("ssub16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SSUB16()
1552 __STATIC_FORCEINLINE uint32_t __QSUB16(uint32_t op1, uint32_t op2) in __QSUB16() argument
1556 __ASM volatile ("qsub16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __QSUB16()
1560 __STATIC_FORCEINLINE uint32_t __SHSUB16(uint32_t op1, uint32_t op2) in __SHSUB16() argument
1564 __ASM volatile ("shsub16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SHSUB16()
1568 __STATIC_FORCEINLINE uint32_t __USUB16(uint32_t op1, uint32_t op2) in __USUB16() argument
1572 __ASM volatile ("usub16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __USUB16()
1576 __STATIC_FORCEINLINE uint32_t __UQSUB16(uint32_t op1, uint32_t op2) in __UQSUB16() argument
1580 __ASM volatile ("uqsub16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UQSUB16()
1584 __STATIC_FORCEINLINE uint32_t __UHSUB16(uint32_t op1, uint32_t op2) in __UHSUB16() argument
1588 __ASM volatile ("uhsub16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UHSUB16()
1592 __STATIC_FORCEINLINE uint32_t __SASX(uint32_t op1, uint32_t op2) in __SASX() argument
1596 __ASM volatile ("sasx %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SASX()
1600 __STATIC_FORCEINLINE uint32_t __QASX(uint32_t op1, uint32_t op2) in __QASX() argument
1604 __ASM volatile ("qasx %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __QASX()
1608 __STATIC_FORCEINLINE uint32_t __SHASX(uint32_t op1, uint32_t op2) in __SHASX() argument
1612 __ASM volatile ("shasx %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SHASX()
1616 __STATIC_FORCEINLINE uint32_t __UASX(uint32_t op1, uint32_t op2) in __UASX() argument
1620 __ASM volatile ("uasx %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UASX()
1624 __STATIC_FORCEINLINE uint32_t __UQASX(uint32_t op1, uint32_t op2) in __UQASX() argument
1628 __ASM volatile ("uqasx %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UQASX()
1632 __STATIC_FORCEINLINE uint32_t __UHASX(uint32_t op1, uint32_t op2) in __UHASX() argument
1636 __ASM volatile ("uhasx %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UHASX()
1640 __STATIC_FORCEINLINE uint32_t __SSAX(uint32_t op1, uint32_t op2) in __SSAX() argument
1644 __ASM volatile ("ssax %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SSAX()
1648 __STATIC_FORCEINLINE uint32_t __QSAX(uint32_t op1, uint32_t op2) in __QSAX() argument
1652 __ASM volatile ("qsax %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __QSAX()
1656 __STATIC_FORCEINLINE uint32_t __SHSAX(uint32_t op1, uint32_t op2) in __SHSAX() argument
1660 __ASM volatile ("shsax %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SHSAX()
1664 __STATIC_FORCEINLINE uint32_t __USAX(uint32_t op1, uint32_t op2) in __USAX() argument
1668 __ASM volatile ("usax %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __USAX()
1672 __STATIC_FORCEINLINE uint32_t __UQSAX(uint32_t op1, uint32_t op2) in __UQSAX() argument
1676 __ASM volatile ("uqsax %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UQSAX()
1680 __STATIC_FORCEINLINE uint32_t __UHSAX(uint32_t op1, uint32_t op2) in __UHSAX() argument
1684 __ASM volatile ("uhsax %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UHSAX()
1688 __STATIC_FORCEINLINE uint32_t __USAD8(uint32_t op1, uint32_t op2) in __USAD8() argument
1692 __ASM volatile ("usad8 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __USAD8()
1696 __STATIC_FORCEINLINE uint32_t __USADA8(uint32_t op1, uint32_t op2, uint32_t op3) in __USADA8() argument
1700 __ASM volatile ("usada8 %0, %1, %2, %3" : "=r" (result) : "r" (op1), "r" (op2), "r" (op3) ); in __USADA8()
1726 __STATIC_FORCEINLINE uint32_t __UXTAB16(uint32_t op1, uint32_t op2) in __UXTAB16() argument
1730 __ASM volatile ("uxtab16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __UXTAB16()
1742 __STATIC_FORCEINLINE uint32_t __SXTAB16(uint32_t op1, uint32_t op2) in __SXTAB16() argument
1746 __ASM volatile ("sxtab16 %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SXTAB16()
1750 __STATIC_FORCEINLINE uint32_t __SMUAD (uint32_t op1, uint32_t op2) in __SMUAD() argument
1754 __ASM volatile ("smuad %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SMUAD()
1758 __STATIC_FORCEINLINE uint32_t __SMUADX (uint32_t op1, uint32_t op2) in __SMUADX() argument
1762 __ASM volatile ("smuadx %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SMUADX()
1766 __STATIC_FORCEINLINE uint32_t __SMLAD (uint32_t op1, uint32_t op2, uint32_t op3) in __SMLAD() argument
1770 __ASM volatile ("smlad %0, %1, %2, %3" : "=r" (result) : "r" (op1), "r" (op2), "r" (op3) ); in __SMLAD()
1774 __STATIC_FORCEINLINE uint32_t __SMLADX (uint32_t op1, uint32_t op2, uint32_t op3) in __SMLADX() argument
1778 __ASM volatile ("smladx %0, %1, %2, %3" : "=r" (result) : "r" (op1), "r" (op2), "r" (op3) ); in __SMLADX()
1782 __STATIC_FORCEINLINE uint64_t __SMLALD (uint32_t op1, uint32_t op2, uint64_t acc) in __SMLALD() argument
1791 …ld %0, %1, %2, %3" : "=r" (llr.w32[0]), "=r" (llr.w32[1]): "r" (op1), "r" (op2) , "0" (llr.w32[0])… in __SMLALD()
1793 …ld %0, %1, %2, %3" : "=r" (llr.w32[1]), "=r" (llr.w32[0]): "r" (op1), "r" (op2) , "0" (llr.w32[1])… in __SMLALD()
1799 __STATIC_FORCEINLINE uint64_t __SMLALDX (uint32_t op1, uint32_t op2, uint64_t acc) in __SMLALDX() argument
1808 …dx %0, %1, %2, %3" : "=r" (llr.w32[0]), "=r" (llr.w32[1]): "r" (op1), "r" (op2) , "0" (llr.w32[0])… in __SMLALDX()
1810 …dx %0, %1, %2, %3" : "=r" (llr.w32[1]), "=r" (llr.w32[0]): "r" (op1), "r" (op2) , "0" (llr.w32[1])… in __SMLALDX()
1816 __STATIC_FORCEINLINE uint32_t __SMUSD (uint32_t op1, uint32_t op2) in __SMUSD() argument
1820 __ASM volatile ("smusd %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SMUSD()
1824 __STATIC_FORCEINLINE uint32_t __SMUSDX (uint32_t op1, uint32_t op2) in __SMUSDX() argument
1828 __ASM volatile ("smusdx %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SMUSDX()
1832 __STATIC_FORCEINLINE uint32_t __SMLSD (uint32_t op1, uint32_t op2, uint32_t op3) in __SMLSD() argument
1836 __ASM volatile ("smlsd %0, %1, %2, %3" : "=r" (result) : "r" (op1), "r" (op2), "r" (op3) ); in __SMLSD()
1840 __STATIC_FORCEINLINE uint32_t __SMLSDX (uint32_t op1, uint32_t op2, uint32_t op3) in __SMLSDX() argument
1844 __ASM volatile ("smlsdx %0, %1, %2, %3" : "=r" (result) : "r" (op1), "r" (op2), "r" (op3) ); in __SMLSDX()
1848 __STATIC_FORCEINLINE uint64_t __SMLSLD (uint32_t op1, uint32_t op2, uint64_t acc) in __SMLSLD() argument
1857 …ld %0, %1, %2, %3" : "=r" (llr.w32[0]), "=r" (llr.w32[1]): "r" (op1), "r" (op2) , "0" (llr.w32[0])… in __SMLSLD()
1859 …ld %0, %1, %2, %3" : "=r" (llr.w32[1]), "=r" (llr.w32[0]): "r" (op1), "r" (op2) , "0" (llr.w32[1])… in __SMLSLD()
1865 __STATIC_FORCEINLINE uint64_t __SMLSLDX (uint32_t op1, uint32_t op2, uint64_t acc) in __SMLSLDX() argument
1874 …dx %0, %1, %2, %3" : "=r" (llr.w32[0]), "=r" (llr.w32[1]): "r" (op1), "r" (op2) , "0" (llr.w32[0])… in __SMLSLDX()
1876 …dx %0, %1, %2, %3" : "=r" (llr.w32[1]), "=r" (llr.w32[0]): "r" (op1), "r" (op2) , "0" (llr.w32[1])… in __SMLSLDX()
1882 __STATIC_FORCEINLINE uint32_t __SEL (uint32_t op1, uint32_t op2) in __SEL() argument
1886 __ASM volatile ("sel %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __SEL()
1890 __STATIC_FORCEINLINE int32_t __QADD( int32_t op1, int32_t op2) in __QADD() argument
1894 __ASM volatile ("qadd %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __QADD()
1898 __STATIC_FORCEINLINE int32_t __QSUB( int32_t op1, int32_t op2) in __QSUB() argument
1902 __ASM volatile ("qsub %0, %1, %2" : "=r" (result) : "r" (op1), "r" (op2) ); in __QSUB()
1916 __STATIC_FORCEINLINE int32_t __SMMLA (int32_t op1, int32_t op2, int32_t op3) in __SMMLA() argument
1920 __ASM volatile ("smmla %0, %1, %2, %3" : "=r" (result): "r" (op1), "r" (op2), "r" (op3) ); in __SMMLA()