Lines Matching refs:DST
62 #define DST regs[insn->dst_reg] macro
1819 DST = DST OP (SRC & 63); \ in ___bpf_prog_run()
1822 DST = (u32) DST OP ((u32) SRC & 31); \ in ___bpf_prog_run()
1825 DST = DST OP IMM; \ in ___bpf_prog_run()
1828 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run()
1833 DST = DST OP SRC; \ in ___bpf_prog_run()
1836 DST = (u32) DST OP (u32) SRC; \ in ___bpf_prog_run()
1839 DST = DST OP IMM; \ in ___bpf_prog_run()
1842 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run()
1855 DST = (u32) -DST; in ___bpf_prog_run()
1858 DST = -DST; in ___bpf_prog_run()
1863 DST = (u32) SRC; in ___bpf_prog_run()
1866 DST = (u32)(s8) SRC; in ___bpf_prog_run()
1869 DST = (u32)(s16) SRC; in ___bpf_prog_run()
1874 DST = (u32) IMM; in ___bpf_prog_run()
1879 DST = SRC; in ___bpf_prog_run()
1882 DST = (s8) SRC; in ___bpf_prog_run()
1885 DST = (s16) SRC; in ___bpf_prog_run()
1888 DST = (s32) SRC; in ___bpf_prog_run()
1893 DST = IMM; in ___bpf_prog_run()
1896 DST = (u64) (u32) insn[0].imm | ((u64) (u32) insn[1].imm) << 32; in ___bpf_prog_run()
1900 DST = (u64) (u32) (((s32) DST) >> (SRC & 31)); in ___bpf_prog_run()
1903 DST = (u64) (u32) (((s32) DST) >> IMM); in ___bpf_prog_run()
1906 (*(s64 *) &DST) >>= (SRC & 63); in ___bpf_prog_run()
1909 (*(s64 *) &DST) >>= IMM; in ___bpf_prog_run()
1914 div64_u64_rem(DST, SRC, &AX); in ___bpf_prog_run()
1915 DST = AX; in ___bpf_prog_run()
1918 AX = div64_s64(DST, SRC); in ___bpf_prog_run()
1919 DST = DST - AX * SRC; in ___bpf_prog_run()
1926 AX = (u32) DST; in ___bpf_prog_run()
1927 DST = do_div(AX, (u32) SRC); in ___bpf_prog_run()
1930 AX = abs_s32((s32)DST); in ___bpf_prog_run()
1932 if ((s32)DST < 0) in ___bpf_prog_run()
1933 DST = (u32)-AX; in ___bpf_prog_run()
1935 DST = (u32)AX; in ___bpf_prog_run()
1942 div64_u64_rem(DST, IMM, &AX); in ___bpf_prog_run()
1943 DST = AX; in ___bpf_prog_run()
1946 AX = div64_s64(DST, IMM); in ___bpf_prog_run()
1947 DST = DST - AX * IMM; in ___bpf_prog_run()
1954 AX = (u32) DST; in ___bpf_prog_run()
1955 DST = do_div(AX, (u32) IMM); in ___bpf_prog_run()
1958 AX = abs_s32((s32)DST); in ___bpf_prog_run()
1960 if ((s32)DST < 0) in ___bpf_prog_run()
1961 DST = (u32)-AX; in ___bpf_prog_run()
1963 DST = (u32)AX; in ___bpf_prog_run()
1970 DST = div64_u64(DST, SRC); in ___bpf_prog_run()
1973 DST = div64_s64(DST, SRC); in ___bpf_prog_run()
1980 AX = (u32) DST; in ___bpf_prog_run()
1982 DST = (u32) AX; in ___bpf_prog_run()
1985 AX = abs_s32((s32)DST); in ___bpf_prog_run()
1987 if (((s32)DST < 0) == ((s32)SRC < 0)) in ___bpf_prog_run()
1988 DST = (u32)AX; in ___bpf_prog_run()
1990 DST = (u32)-AX; in ___bpf_prog_run()
1997 DST = div64_u64(DST, IMM); in ___bpf_prog_run()
2000 DST = div64_s64(DST, IMM); in ___bpf_prog_run()
2007 AX = (u32) DST; in ___bpf_prog_run()
2009 DST = (u32) AX; in ___bpf_prog_run()
2012 AX = abs_s32((s32)DST); in ___bpf_prog_run()
2014 if (((s32)DST < 0) == ((s32)IMM < 0)) in ___bpf_prog_run()
2015 DST = (u32)AX; in ___bpf_prog_run()
2017 DST = (u32)-AX; in ___bpf_prog_run()
2024 DST = (__force u16) cpu_to_be16(DST); in ___bpf_prog_run()
2027 DST = (__force u32) cpu_to_be32(DST); in ___bpf_prog_run()
2030 DST = (__force u64) cpu_to_be64(DST); in ___bpf_prog_run()
2037 DST = (__force u16) cpu_to_le16(DST); in ___bpf_prog_run()
2040 DST = (__force u32) cpu_to_le32(DST); in ___bpf_prog_run()
2043 DST = (__force u64) cpu_to_le64(DST); in ___bpf_prog_run()
2050 DST = (__force u16) __swab16(DST); in ___bpf_prog_run()
2053 DST = (__force u32) __swab32(DST); in ___bpf_prog_run()
2056 DST = (__force u64) __swab64(DST); in ___bpf_prog_run()
2117 if ((SIGN##64) DST CMP_OP (SIGN##64) SRC) { \ in ___bpf_prog_run()
2123 if ((SIGN##32) DST CMP_OP (SIGN##32) SRC) { \ in ___bpf_prog_run()
2129 if ((SIGN##64) DST CMP_OP (SIGN##64) IMM) { \ in ___bpf_prog_run()
2135 if ((SIGN##32) DST CMP_OP (SIGN##32) IMM) { \ in ___bpf_prog_run()
2168 *(SIZE *)(unsigned long) (DST + insn->off) = SRC; \ in ___bpf_prog_run()
2171 *(SIZE *)(unsigned long) (DST + insn->off) = IMM; \ in ___bpf_prog_run()
2174 DST = *(SIZE *)(unsigned long) (SRC + insn->off); \ in ___bpf_prog_run()
2177 bpf_probe_read_kernel_common(&DST, sizeof(SIZE), \ in ___bpf_prog_run()
2179 DST = *((SIZE *)&DST); \ in ___bpf_prog_run()
2190 DST = *(SIZE *)(unsigned long) (SRC + insn->off); \ in ___bpf_prog_run()
2193 bpf_probe_read_kernel_common(&DST, sizeof(SIZE), \ in ___bpf_prog_run()
2195 DST = *((SIZE *)&DST); \ in ___bpf_prog_run()
2207 (DST + insn->off)); \ in ___bpf_prog_run()
2210 (DST + insn->off)); \ in ___bpf_prog_run()
2218 (atomic_t *)(unsigned long) (DST + insn->off)); \ in ___bpf_prog_run()
2222 (atomic64_t *)(unsigned long) (DST + insn->off)); \ in ___bpf_prog_run()
2244 (atomic_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
2248 (atomic64_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
2256 (atomic_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
2260 (atomic64_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
2272 DST = (SIZE)smp_load_acquire( \ in ___bpf_prog_run()
2291 (SIZE *)(unsigned long)(DST + insn->off), (SIZE)SRC); \ in ___bpf_prog_run()