Lines Matching refs:DST

58 #define DST	regs[insn->dst_reg]  macro
1706 DST = DST OP (SRC & 63); \ in ___bpf_prog_run()
1709 DST = (u32) DST OP ((u32) SRC & 31); \ in ___bpf_prog_run()
1712 DST = DST OP IMM; \ in ___bpf_prog_run()
1715 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run()
1720 DST = DST OP SRC; \ in ___bpf_prog_run()
1723 DST = (u32) DST OP (u32) SRC; \ in ___bpf_prog_run()
1726 DST = DST OP IMM; \ in ___bpf_prog_run()
1729 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run()
1742 DST = (u32) -DST; in ___bpf_prog_run()
1745 DST = -DST; in ___bpf_prog_run()
1750 DST = (u32) SRC; in ___bpf_prog_run()
1753 DST = (u32)(s8) SRC; in ___bpf_prog_run()
1756 DST = (u32)(s16) SRC; in ___bpf_prog_run()
1761 DST = (u32) IMM; in ___bpf_prog_run()
1766 DST = SRC; in ___bpf_prog_run()
1769 DST = (s8) SRC; in ___bpf_prog_run()
1772 DST = (s16) SRC; in ___bpf_prog_run()
1775 DST = (s32) SRC; in ___bpf_prog_run()
1780 DST = IMM; in ___bpf_prog_run()
1783 DST = (u64) (u32) insn[0].imm | ((u64) (u32) insn[1].imm) << 32; in ___bpf_prog_run()
1787 DST = (u64) (u32) (((s32) DST) >> (SRC & 31)); in ___bpf_prog_run()
1790 DST = (u64) (u32) (((s32) DST) >> IMM); in ___bpf_prog_run()
1793 (*(s64 *) &DST) >>= (SRC & 63); in ___bpf_prog_run()
1796 (*(s64 *) &DST) >>= IMM; in ___bpf_prog_run()
1801 div64_u64_rem(DST, SRC, &AX); in ___bpf_prog_run()
1802 DST = AX; in ___bpf_prog_run()
1805 AX = div64_s64(DST, SRC); in ___bpf_prog_run()
1806 DST = DST - AX * SRC; in ___bpf_prog_run()
1813 AX = (u32) DST; in ___bpf_prog_run()
1814 DST = do_div(AX, (u32) SRC); in ___bpf_prog_run()
1817 AX = abs((s32)DST); in ___bpf_prog_run()
1819 if ((s32)DST < 0) in ___bpf_prog_run()
1820 DST = (u32)-AX; in ___bpf_prog_run()
1822 DST = (u32)AX; in ___bpf_prog_run()
1829 div64_u64_rem(DST, IMM, &AX); in ___bpf_prog_run()
1830 DST = AX; in ___bpf_prog_run()
1833 AX = div64_s64(DST, IMM); in ___bpf_prog_run()
1834 DST = DST - AX * IMM; in ___bpf_prog_run()
1841 AX = (u32) DST; in ___bpf_prog_run()
1842 DST = do_div(AX, (u32) IMM); in ___bpf_prog_run()
1845 AX = abs((s32)DST); in ___bpf_prog_run()
1847 if ((s32)DST < 0) in ___bpf_prog_run()
1848 DST = (u32)-AX; in ___bpf_prog_run()
1850 DST = (u32)AX; in ___bpf_prog_run()
1857 DST = div64_u64(DST, SRC); in ___bpf_prog_run()
1860 DST = div64_s64(DST, SRC); in ___bpf_prog_run()
1867 AX = (u32) DST; in ___bpf_prog_run()
1869 DST = (u32) AX; in ___bpf_prog_run()
1872 AX = abs((s32)DST); in ___bpf_prog_run()
1874 if (((s32)DST < 0) == ((s32)SRC < 0)) in ___bpf_prog_run()
1875 DST = (u32)AX; in ___bpf_prog_run()
1877 DST = (u32)-AX; in ___bpf_prog_run()
1884 DST = div64_u64(DST, IMM); in ___bpf_prog_run()
1887 DST = div64_s64(DST, IMM); in ___bpf_prog_run()
1894 AX = (u32) DST; in ___bpf_prog_run()
1896 DST = (u32) AX; in ___bpf_prog_run()
1899 AX = abs((s32)DST); in ___bpf_prog_run()
1901 if (((s32)DST < 0) == ((s32)IMM < 0)) in ___bpf_prog_run()
1902 DST = (u32)AX; in ___bpf_prog_run()
1904 DST = (u32)-AX; in ___bpf_prog_run()
1911 DST = (__force u16) cpu_to_be16(DST); in ___bpf_prog_run()
1914 DST = (__force u32) cpu_to_be32(DST); in ___bpf_prog_run()
1917 DST = (__force u64) cpu_to_be64(DST); in ___bpf_prog_run()
1924 DST = (__force u16) cpu_to_le16(DST); in ___bpf_prog_run()
1927 DST = (__force u32) cpu_to_le32(DST); in ___bpf_prog_run()
1930 DST = (__force u64) cpu_to_le64(DST); in ___bpf_prog_run()
1937 DST = (__force u16) __swab16(DST); in ___bpf_prog_run()
1940 DST = (__force u32) __swab32(DST); in ___bpf_prog_run()
1943 DST = (__force u64) __swab64(DST); in ___bpf_prog_run()
2004 if ((SIGN##64) DST CMP_OP (SIGN##64) SRC) { \ in ___bpf_prog_run()
2010 if ((SIGN##32) DST CMP_OP (SIGN##32) SRC) { \ in ___bpf_prog_run()
2016 if ((SIGN##64) DST CMP_OP (SIGN##64) IMM) { \ in ___bpf_prog_run()
2022 if ((SIGN##32) DST CMP_OP (SIGN##32) IMM) { \ in ___bpf_prog_run()
2054 *(SIZE *)(unsigned long) (DST + insn->off) = SRC; \ in ___bpf_prog_run()
2057 *(SIZE *)(unsigned long) (DST + insn->off) = IMM; \ in ___bpf_prog_run()
2060 DST = *(SIZE *)(unsigned long) (SRC + insn->off); \ in ___bpf_prog_run()
2063 bpf_probe_read_kernel_common(&DST, sizeof(SIZE), \ in ___bpf_prog_run()
2065 DST = *((SIZE *)&DST); \ in ___bpf_prog_run()
2076 DST = *(SIZE *)(unsigned long) (SRC + insn->off); \ in ___bpf_prog_run()
2079 bpf_probe_read_kernel_common(&DST, sizeof(SIZE), \ in ___bpf_prog_run()
2081 DST = *((SIZE *)&DST); \ in ___bpf_prog_run()
2093 (DST + insn->off)); \ in ___bpf_prog_run()
2096 (DST + insn->off)); \ in ___bpf_prog_run()
2102 (atomic_t *)(unsigned long) (DST + insn->off)); \ in ___bpf_prog_run()
2106 (atomic64_t *)(unsigned long) (DST + insn->off)); \ in ___bpf_prog_run()
2121 (atomic_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
2125 (atomic64_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
2131 (atomic_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
2135 (atomic64_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()