Lines Matching refs:insns

68 		struct sock_filter insns[MAX_INSNS];  member
71 void *insns; member
104 self->u.ptr.insns = insn; in bpf_fill_maxinsns1()
123 self->u.ptr.insns = insn; in bpf_fill_maxinsns2()
150 self->u.ptr.insns = insn; in bpf_fill_maxinsns3()
169 self->u.ptr.insns = insn; in bpf_fill_maxinsns4()
192 self->u.ptr.insns = insn; in bpf_fill_maxinsns5()
214 self->u.ptr.insns = insn; in bpf_fill_maxinsns6()
240 self->u.ptr.insns = insn; in bpf_fill_maxinsns7()
263 self->u.ptr.insns = insn; in bpf_fill_maxinsns8()
289 self->u.ptr.insns = insn; in bpf_fill_maxinsns9()
314 self->u.ptr.insns = insn; in bpf_fill_maxinsns10()
343 self->u.ptr.insns = insn; in __bpf_fill_ja()
372 self->u.ptr.insns = insn; in bpf_fill_maxinsns12()
395 self->u.ptr.insns = insn; in bpf_fill_maxinsns13()
425 self->u.ptr.insns = insn; in bpf_fill_ld_abs_get_processor_id()
450 self->u.ptr.insns = insn; in __bpf_fill_stxdw()
467 static int __bpf_ld_imm64(struct bpf_insn insns[2], u8 reg, s64 imm64) in __bpf_ld_imm64()
471 memcpy(insns, tmp, sizeof(tmp)); in __bpf_ld_imm64()
483 struct bpf_insn *insns; in __bpf_fill_max_jmp() local
487 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); in __bpf_fill_max_jmp()
488 if (!insns) in __bpf_fill_max_jmp()
491 i = __bpf_ld_imm64(insns, R1, 0x0123456789abcdefULL); in __bpf_fill_max_jmp()
492 insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1); in __bpf_fill_max_jmp()
493 insns[i++] = BPF_JMP_IMM(jmp, R0, imm, S16_MAX); in __bpf_fill_max_jmp()
494 insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 2); in __bpf_fill_max_jmp()
495 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_max_jmp()
505 insns[i++] = BPF_ALU32_REG(op, R0, R1); in __bpf_fill_max_jmp()
507 insns[i++] = BPF_ALU64_REG(op, R0, R1); in __bpf_fill_max_jmp()
510 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_max_jmp()
511 self->u.ptr.insns = insns; in __bpf_fill_max_jmp()
662 self->u.ptr.insns = insn; in __bpf_fill_alu_shift()
772 self->u.ptr.insns = insn; in __bpf_fill_alu_shift_same_reg()
827 struct bpf_insn *insns; in __bpf_fill_pattern() local
839 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); in __bpf_fill_pattern()
840 if (!insns) in __bpf_fill_pattern()
844 insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0); in __bpf_fill_pattern()
862 &insns[i], in __bpf_fill_pattern()
879 i += (*emit)(self, arg, &insns[i], in __bpf_fill_pattern()
884 insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1); in __bpf_fill_pattern()
885 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_pattern()
888 self->u.ptr.insns = insns; in __bpf_fill_pattern()
917 struct bpf_insn *insns, s64 dst, s64 imm) in __bpf_emit_alu64_imm() argument
923 if (!insns) in __bpf_emit_alu64_imm()
927 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_alu64_imm()
928 i += __bpf_ld_imm64(&insns[i], R3, res); in __bpf_emit_alu64_imm()
929 insns[i++] = BPF_ALU64_IMM(op, R1, imm); in __bpf_emit_alu64_imm()
930 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); in __bpf_emit_alu64_imm()
931 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_alu64_imm()
938 struct bpf_insn *insns, s64 dst, s64 imm) in __bpf_emit_alu32_imm() argument
944 if (!insns) in __bpf_emit_alu32_imm()
948 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_alu32_imm()
949 i += __bpf_ld_imm64(&insns[i], R3, (u32)res); in __bpf_emit_alu32_imm()
950 insns[i++] = BPF_ALU32_IMM(op, R1, imm); in __bpf_emit_alu32_imm()
951 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); in __bpf_emit_alu32_imm()
952 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_alu32_imm()
959 struct bpf_insn *insns, s64 dst, s64 src) in __bpf_emit_alu64_reg() argument
965 if (!insns) in __bpf_emit_alu64_reg()
969 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_alu64_reg()
970 i += __bpf_ld_imm64(&insns[i], R2, src); in __bpf_emit_alu64_reg()
971 i += __bpf_ld_imm64(&insns[i], R3, res); in __bpf_emit_alu64_reg()
972 insns[i++] = BPF_ALU64_REG(op, R1, R2); in __bpf_emit_alu64_reg()
973 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); in __bpf_emit_alu64_reg()
974 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_alu64_reg()
981 struct bpf_insn *insns, s64 dst, s64 src) in __bpf_emit_alu32_reg() argument
987 if (!insns) in __bpf_emit_alu32_reg()
991 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_alu32_reg()
992 i += __bpf_ld_imm64(&insns[i], R2, src); in __bpf_emit_alu32_reg()
993 i += __bpf_ld_imm64(&insns[i], R3, (u32)res); in __bpf_emit_alu32_reg()
994 insns[i++] = BPF_ALU32_REG(op, R1, R2); in __bpf_emit_alu32_reg()
995 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); in __bpf_emit_alu32_reg()
996 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_alu32_reg()
1221 struct bpf_insn *insns; in __bpf_fill_alu_imm_regs() local
1227 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); in __bpf_fill_alu_imm_regs()
1228 if (!insns) in __bpf_fill_alu_imm_regs()
1248 i += __bpf_ld_imm64(&insns[i], rd, dst); in __bpf_fill_alu_imm_regs()
1251 insns[i++] = BPF_ALU32_IMM(op, rd, imm); in __bpf_fill_alu_imm_regs()
1253 insns[i++] = BPF_ALU64_IMM(op, rd, imm); in __bpf_fill_alu_imm_regs()
1255 insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, res, 2); in __bpf_fill_alu_imm_regs()
1256 insns[i++] = BPF_MOV64_IMM(R0, __LINE__); in __bpf_fill_alu_imm_regs()
1257 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_alu_imm_regs()
1259 insns[i++] = BPF_ALU64_IMM(BPF_RSH, rd, 32); in __bpf_fill_alu_imm_regs()
1260 insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, res >> 32, 2); in __bpf_fill_alu_imm_regs()
1261 insns[i++] = BPF_MOV64_IMM(R0, __LINE__); in __bpf_fill_alu_imm_regs()
1262 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_alu_imm_regs()
1265 insns[i++] = BPF_MOV64_IMM(R0, 1); in __bpf_fill_alu_imm_regs()
1266 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_alu_imm_regs()
1268 self->u.ptr.insns = insns; in __bpf_fill_alu_imm_regs()
1405 struct bpf_insn *insns; in __bpf_fill_alu_reg_pairs() local
1409 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); in __bpf_fill_alu_reg_pairs()
1410 if (!insns) in __bpf_fill_alu_reg_pairs()
1438 i += __bpf_ld_imm64(&insns[i], rd, dst); in __bpf_fill_alu_reg_pairs()
1439 i += __bpf_ld_imm64(&insns[i], rs, src); in __bpf_fill_alu_reg_pairs()
1442 insns[i++] = BPF_ALU32_REG(op, rd, rs); in __bpf_fill_alu_reg_pairs()
1444 insns[i++] = BPF_ALU64_REG(op, rd, rs); in __bpf_fill_alu_reg_pairs()
1446 insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, val, 2); in __bpf_fill_alu_reg_pairs()
1447 insns[i++] = BPF_MOV64_IMM(R0, __LINE__); in __bpf_fill_alu_reg_pairs()
1448 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_alu_reg_pairs()
1450 insns[i++] = BPF_ALU64_IMM(BPF_RSH, rd, 32); in __bpf_fill_alu_reg_pairs()
1451 insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, val >> 32, 2); in __bpf_fill_alu_reg_pairs()
1452 insns[i++] = BPF_MOV64_IMM(R0, __LINE__); in __bpf_fill_alu_reg_pairs()
1453 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_alu_reg_pairs()
1457 insns[i++] = BPF_MOV64_IMM(R0, 1); in __bpf_fill_alu_reg_pairs()
1458 insns[i++] = BPF_EXIT_INSN(); in __bpf_fill_alu_reg_pairs()
1460 self->u.ptr.insns = insns; in __bpf_fill_alu_reg_pairs()
1595 struct bpf_insn *insns, s64 dst, s64 src) in __bpf_emit_atomic64() argument
1601 if (!insns) in __bpf_emit_atomic64()
1618 i += __bpf_ld_imm64(&insns[i], R0, keep); in __bpf_emit_atomic64()
1619 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_atomic64()
1620 i += __bpf_ld_imm64(&insns[i], R2, src); in __bpf_emit_atomic64()
1621 i += __bpf_ld_imm64(&insns[i], R3, res); in __bpf_emit_atomic64()
1622 i += __bpf_ld_imm64(&insns[i], R4, fetch); in __bpf_emit_atomic64()
1623 i += __bpf_ld_imm64(&insns[i], R5, keep); in __bpf_emit_atomic64()
1625 insns[i++] = BPF_STX_MEM(BPF_DW, R10, R1, -8); in __bpf_emit_atomic64()
1626 insns[i++] = BPF_ATOMIC_OP(BPF_DW, op, R10, R2, -8); in __bpf_emit_atomic64()
1627 insns[i++] = BPF_LDX_MEM(BPF_DW, R1, R10, -8); in __bpf_emit_atomic64()
1629 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); in __bpf_emit_atomic64()
1630 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_atomic64()
1632 insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R4, 1); in __bpf_emit_atomic64()
1633 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_atomic64()
1635 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R5, 1); in __bpf_emit_atomic64()
1636 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_atomic64()
1642 struct bpf_insn *insns, s64 dst, s64 src) in __bpf_emit_atomic32() argument
1648 if (!insns) in __bpf_emit_atomic32()
1665 i += __bpf_ld_imm64(&insns[i], R0, keep); in __bpf_emit_atomic32()
1666 i += __bpf_ld_imm64(&insns[i], R1, (u32)dst); in __bpf_emit_atomic32()
1667 i += __bpf_ld_imm64(&insns[i], R2, src); in __bpf_emit_atomic32()
1668 i += __bpf_ld_imm64(&insns[i], R3, (u32)res); in __bpf_emit_atomic32()
1669 i += __bpf_ld_imm64(&insns[i], R4, fetch); in __bpf_emit_atomic32()
1670 i += __bpf_ld_imm64(&insns[i], R5, keep); in __bpf_emit_atomic32()
1672 insns[i++] = BPF_STX_MEM(BPF_W, R10, R1, -4); in __bpf_emit_atomic32()
1673 insns[i++] = BPF_ATOMIC_OP(BPF_W, op, R10, R2, -4); in __bpf_emit_atomic32()
1674 insns[i++] = BPF_LDX_MEM(BPF_W, R1, R10, -4); in __bpf_emit_atomic32()
1676 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); in __bpf_emit_atomic32()
1677 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_atomic32()
1679 insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R4, 1); in __bpf_emit_atomic32()
1680 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_atomic32()
1682 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R5, 1); in __bpf_emit_atomic32()
1683 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_atomic32()
1689 struct bpf_insn *insns, s64 dst, s64 src) in __bpf_emit_cmpxchg64() argument
1693 if (!insns) in __bpf_emit_cmpxchg64()
1696 i += __bpf_ld_imm64(&insns[i], R0, ~dst); in __bpf_emit_cmpxchg64()
1697 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_cmpxchg64()
1698 i += __bpf_ld_imm64(&insns[i], R2, src); in __bpf_emit_cmpxchg64()
1701 insns[i++] = BPF_STX_MEM(BPF_DW, R10, R1, -8); in __bpf_emit_cmpxchg64()
1702 insns[i++] = BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -8); in __bpf_emit_cmpxchg64()
1703 insns[i++] = BPF_LDX_MEM(BPF_DW, R3, R10, -8); in __bpf_emit_cmpxchg64()
1705 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 2); in __bpf_emit_cmpxchg64()
1706 insns[i++] = BPF_MOV64_IMM(R0, __LINE__); in __bpf_emit_cmpxchg64()
1707 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_cmpxchg64()
1709 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R3, 2); in __bpf_emit_cmpxchg64()
1710 insns[i++] = BPF_MOV64_IMM(R0, __LINE__); in __bpf_emit_cmpxchg64()
1711 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_cmpxchg64()
1714 insns[i++] = BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -8); in __bpf_emit_cmpxchg64()
1715 insns[i++] = BPF_LDX_MEM(BPF_DW, R3, R10, -8); in __bpf_emit_cmpxchg64()
1717 insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R3, 2); in __bpf_emit_cmpxchg64()
1718 insns[i++] = BPF_MOV64_IMM(R0, __LINE__); in __bpf_emit_cmpxchg64()
1719 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_cmpxchg64()
1721 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R1, 2); in __bpf_emit_cmpxchg64()
1722 insns[i++] = BPF_MOV64_IMM(R0, __LINE__); in __bpf_emit_cmpxchg64()
1723 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_cmpxchg64()
1729 struct bpf_insn *insns, s64 dst, s64 src) in __bpf_emit_cmpxchg32() argument
1733 if (!insns) in __bpf_emit_cmpxchg32()
1736 i += __bpf_ld_imm64(&insns[i], R0, ~dst); in __bpf_emit_cmpxchg32()
1737 i += __bpf_ld_imm64(&insns[i], R1, (u32)dst); in __bpf_emit_cmpxchg32()
1738 i += __bpf_ld_imm64(&insns[i], R2, src); in __bpf_emit_cmpxchg32()
1741 insns[i++] = BPF_STX_MEM(BPF_W, R10, R1, -4); in __bpf_emit_cmpxchg32()
1742 insns[i++] = BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R2, -4); in __bpf_emit_cmpxchg32()
1743 insns[i++] = BPF_ZEXT_REG(R0), /* Zext always inserted by verifier */ in __bpf_emit_cmpxchg32()
1744 insns[i++] = BPF_LDX_MEM(BPF_W, R3, R10, -4); in __bpf_emit_cmpxchg32()
1746 insns[i++] = BPF_JMP32_REG(BPF_JEQ, R1, R3, 2); in __bpf_emit_cmpxchg32()
1747 insns[i++] = BPF_MOV32_IMM(R0, __LINE__); in __bpf_emit_cmpxchg32()
1748 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_cmpxchg32()
1750 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R3, 2); in __bpf_emit_cmpxchg32()
1751 insns[i++] = BPF_MOV32_IMM(R0, __LINE__); in __bpf_emit_cmpxchg32()
1752 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_cmpxchg32()
1755 i += __bpf_ld_imm64(&insns[i], R0, dst); in __bpf_emit_cmpxchg32()
1756 insns[i++] = BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R2, -4); in __bpf_emit_cmpxchg32()
1757 insns[i++] = BPF_ZEXT_REG(R0), /* Zext always inserted by verifier */ in __bpf_emit_cmpxchg32()
1758 insns[i++] = BPF_LDX_MEM(BPF_W, R3, R10, -4); in __bpf_emit_cmpxchg32()
1760 insns[i++] = BPF_JMP32_REG(BPF_JEQ, R2, R3, 2); in __bpf_emit_cmpxchg32()
1761 insns[i++] = BPF_MOV32_IMM(R0, __LINE__); in __bpf_emit_cmpxchg32()
1762 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_cmpxchg32()
1764 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R1, 2); in __bpf_emit_cmpxchg32()
1765 insns[i++] = BPF_MOV32_IMM(R0, __LINE__); in __bpf_emit_cmpxchg32()
1766 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_cmpxchg32()
2020 self->u.ptr.insns = insn; in __bpf_fill_atomic_reg_pairs()
2177 self->u.ptr.insns = insn; in bpf_fill_ld_imm64_magn()
2239 self->u.ptr.insns = insn; in __bpf_fill_ld_imm64_bytes()
2303 struct bpf_insn *insns, s64 dst, s64 imm) in __bpf_emit_jmp_imm() argument
2307 if (insns) { in __bpf_emit_jmp_imm()
2311 insns[i++] = BPF_ALU32_IMM(BPF_MOV, R0, match); in __bpf_emit_jmp_imm()
2313 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_jmp_imm()
2314 insns[i++] = BPF_JMP_IMM(op, R1, imm, 1); in __bpf_emit_jmp_imm()
2316 insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1); in __bpf_emit_jmp_imm()
2317 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_jmp_imm()
2326 struct bpf_insn *insns, s64 dst, s64 imm) in __bpf_emit_jmp32_imm() argument
2330 if (insns) { in __bpf_emit_jmp32_imm()
2334 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_jmp32_imm()
2335 insns[i++] = BPF_JMP32_IMM(op, R1, imm, 1); in __bpf_emit_jmp32_imm()
2337 insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1); in __bpf_emit_jmp32_imm()
2338 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_jmp32_imm()
2347 struct bpf_insn *insns, s64 dst, s64 src) in __bpf_emit_jmp_reg() argument
2351 if (insns) { in __bpf_emit_jmp_reg()
2355 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_jmp_reg()
2356 i += __bpf_ld_imm64(&insns[i], R2, src); in __bpf_emit_jmp_reg()
2357 insns[i++] = BPF_JMP_REG(op, R1, R2, 1); in __bpf_emit_jmp_reg()
2359 insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1); in __bpf_emit_jmp_reg()
2360 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_jmp_reg()
2369 struct bpf_insn *insns, s64 dst, s64 src) in __bpf_emit_jmp32_reg() argument
2373 if (insns) { in __bpf_emit_jmp32_reg()
2377 i += __bpf_ld_imm64(&insns[i], R1, dst); in __bpf_emit_jmp32_reg()
2378 i += __bpf_ld_imm64(&insns[i], R2, src); in __bpf_emit_jmp32_reg()
2379 insns[i++] = BPF_JMP32_REG(op, R1, R2, 1); in __bpf_emit_jmp32_reg()
2381 insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1); in __bpf_emit_jmp32_reg()
2382 insns[i++] = BPF_EXIT_INSN(); in __bpf_emit_jmp32_reg()
2704 struct bpf_insn *insns; in __bpf_fill_staggered_jumps() local
2707 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); in __bpf_fill_staggered_jumps()
2708 if (!insns) in __bpf_fill_staggered_jumps()
2712 insns[0] = BPF_ALU64_IMM(BPF_MOV, R0, 0); in __bpf_fill_staggered_jumps()
2713 insns[1] = BPF_ALU64_IMM(BPF_MOV, R1, r1); in __bpf_fill_staggered_jumps()
2714 insns[2] = BPF_ALU64_IMM(BPF_MOV, R2, r2); in __bpf_fill_staggered_jumps()
2715 insns[3] = BPF_JMP_IMM(BPF_JA, 0, 0, 3 * size / 2); in __bpf_fill_staggered_jumps()
2719 struct bpf_insn *ins = &insns[4 + 3 * ind]; in __bpf_fill_staggered_jumps()
2734 insns[len - 1] = BPF_EXIT_INSN(); in __bpf_fill_staggered_jumps()
2736 self->u.ptr.insns = insns; in __bpf_fill_staggered_jumps()
3066 .u.insns = {
3085 .u.insns = {
3097 .u.insns = {
3112 .u.insns = {
3136 .u.insns = {
3152 .u.insns = {
3164 .u.insns = {
3175 .u.insns = {
3185 .u.insns = {
3198 .u.insns = {
3212 .u.insns = {
3225 .u.insns = {
3239 .u.insns = {
3260 .u.insns = {
3271 .u.insns = {
3282 .u.insns = {
3293 .u.insns = {
3312 .u.insns = {
3326 .u.insns = {
3340 .u.insns = {
3351 .u.insns = {
3362 .u.insns = {
3377 .u.insns = {
3395 .u.insns = {
3433 .u.insns = {
3460 .u.insns = {
3473 .u.insns = {
3495 .u.insns = {
3508 .u.insns = {
3521 .u.insns = {
3534 .u.insns = {
3547 .u.insns = {
3566 .u.insns = {
3598 .u.insns = {
3643 .u.insns = {
3696 .u.insns = {
4536 .u.insns = {
4547 .u.insns = {
4559 .u.insns = {
4572 .u.insns = {
4584 .u.insns = {
4666 .u.insns = {
4677 .u.insns = {
4689 .u.insns = {
4779 .u.insns = {
4852 .u.insns = {
4865 .u.insns = {
10406 .u.insns = {
10422 .u.insns = {
10438 .u.insns = {
10454 .u.insns = {
10466 .u.insns = {
10478 .u.insns = {
10493 .u.insns = {
10508 .u.insns = {
10523 .u.insns = {
10534 .u.insns = {
10552 .u.insns = {
10562 .u.insns = {
10573 .u.insns = {
10584 .u.insns = {
10595 .u.insns = {
10606 .u.insns = {
10617 .u.insns = {
10631 .u.insns = {
10647 .u.insns = {
10663 .u.insns = {
10679 .u.insns = {
10690 .u.insns = {
10701 .u.insns = {
10712 .u.insns = {
10730 .u.insns = {
10748 .u.insns = {
10766 .u.insns = {
10784 .u.insns = {
10802 .u.insns = {
10813 .u.insns = {
10824 .u.insns = {
10835 .u.insns = {
10852 .u.insns = {
10862 .u.insns = {
10872 .u.insns = {
10881 .u.insns = {
10891 .u.insns = {
10901 .u.insns = {
10914 .u.insns = {
10931 .u.insns = {
10948 .u.insns = {
10958 .u.insns = {
10968 .u.insns = {
10977 .u.insns = {
10987 .u.insns = {
10997 .u.insns = {
11014 .u.insns = {
11031 .u.insns = {
11048 .u.insns = {
11065 .u.insns = {
11075 .u.insns = {
11085 .u.insns = {
11094 .u.insns = {
11104 .u.insns = {
11114 .u.insns = {
11125 .u.insns = {
11139 .u.insns = {
11151 .u.insns = {
11163 .u.insns = {
11175 .u.insns = {
11187 .u.insns = {
11203 .u.insns = {
11219 .u.insns = {
11233 .u.insns = {
11249 .u.insns = {
11263 .u.insns = {
11279 .u.insns = {
11293 .u.insns = {
11309 .u.insns = {
11323 .u.insns = {
11339 .u.insns = {
11353 .u.insns = {
11369 .u.insns = {
11476 .u.insns = {
14417 fp = tests[which].u.insns; in filter_length()
14428 return tests[which].u.ptr.insns; in filter_pointer()
14430 return tests[which].u.insns; in filter_pointer()
14805 kfree(tests[i].u.ptr.insns); in test_bpf()
14806 tests[i].u.ptr.insns = NULL; in test_bpf()
14844 struct bpf_insn insns[MAX_INSNS]; member
14907 .insns = {
14916 .insns = {
14926 .insns = {
14936 .insns = {
14946 .insns = {
14964 .insns = {
14976 .insns = {
14988 .insns = {
15010 .insns = {
15022 .insns = {
15063 struct bpf_insn *insn = &test->insns[len]; in prepare_tail_call_tests()
15080 memcpy(fp->insnsi, test->insns, len * sizeof(struct bpf_insn)); in prepare_tail_call_tests()