/Linux-v6.6/tools/testing/selftests/bpf/prog_tests/ |
D | align.c | 89 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 4), 90 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 2), 92 BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 4), 93 BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 2), 140 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), \ 218 BPF_ALU64_IMM(BPF_ADD, BPF_REG_5, 14), 220 BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 4), 257 BPF_ALU64_IMM(BPF_ADD, BPF_REG_5, 14), 258 BPF_ALU64_REG(BPF_ADD, BPF_REG_5, BPF_REG_6), 260 BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 4), [all …]
|
D | cgroup_attach_multi.c | 43 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), /* r2 = fp - 4 */ in prog_load_cnt() 48 BPF_ATOMIC_OP(BPF_DW, BPF_ADD, BPF_REG_0, BPF_REG_1, 0), in prog_load_cnt() 54 BPF_ATOMIC_OP(BPF_W, BPF_ADD, BPF_REG_0, BPF_REG_1, 0), in prog_load_cnt() 60 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 0x1), in prog_load_cnt()
|
D | sockopt.c | 291 BPF_ALU64_IMM(BPF_ADD, BPF_REG_6, 1), 380 BPF_ALU64_IMM(BPF_ADD, BPF_REG_6, 1), 690 BPF_ALU64_IMM(BPF_ADD, BPF_REG_6, 1), 727 BPF_ALU64_IMM(BPF_ADD, BPF_REG_6, 1), 831 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, 1), 877 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, 1),
|
/Linux-v6.6/tools/testing/selftests/bpf/verifier/ |
D | map_kptr.c | 8 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 28 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 48 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 63 BPF_ALU64_REG(BPF_ADD, BPF_REG_3, BPF_REG_2), 79 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 94 BPF_ALU64_REG(BPF_ADD, BPF_REG_3, BPF_REG_2), 111 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 117 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 7), 133 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 147 BPF_ALU64_REG(BPF_ADD, BPF_REG_1, BPF_REG_2), [all …]
|
D | calls.c | 28 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 43 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 58 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 73 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 88 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 103 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 122 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 127 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), 165 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 172 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -4), [all …]
|
D | atomic_fetch_add.c | 9 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_1, -8), 31 BPF_ATOMIC_OP(BPF_W, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_1, -4), 50 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_10, -8), 62 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_2, -8), 75 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_2, BPF_REG_0, -8), 96 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_2, BPF_REG_3, 0),
|
D | atomic_fetch.c | 8 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 32 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 56 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 79 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 121 __ATOMIC_FETCH_OP_TEST(BPF_REG_1, BPF_REG_2, 1, BPF_ADD | BPF_FETCH, 2, 3), 122 __ATOMIC_FETCH_OP_TEST(BPF_REG_0, BPF_REG_1, 1, BPF_ADD | BPF_FETCH, 2, 3), 123 __ATOMIC_FETCH_OP_TEST(BPF_REG_1, BPF_REG_0, 1, BPF_ADD | BPF_FETCH, 2, 3), 124 __ATOMIC_FETCH_OP_TEST(BPF_REG_2, BPF_REG_3, 1, BPF_ADD | BPF_FETCH, 2, 3), 125 __ATOMIC_FETCH_OP_TEST(BPF_REG_4, BPF_REG_5, 1, BPF_ADD | BPF_FETCH, 2, 3), 126 __ATOMIC_FETCH_OP_TEST(BPF_REG_9, BPF_REG_8, 1, BPF_ADD | BPF_FETCH, 2, 3),
|
D | bpf_st_mem.c | 6 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, -42), 24 BPF_ALU64_REG(BPF_ADD, BPF_REG_0, BPF_REG_1), 26 BPF_ALU64_REG(BPF_ADD, BPF_REG_0, BPF_REG_1), 28 BPF_ALU64_REG(BPF_ADD, BPF_REG_0, BPF_REG_1), 53 BPF_ALU64_REG(BPF_ADD, BPF_REG_0, BPF_REG_10),
|
D | precise.c | 8 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 18 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 30 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, 1), /* R2=scalar(umin=1, umax=8) */ 32 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 67 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 77 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 89 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, 1), /* R2=scalar(umin=1, umax=8) */ 91 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 242 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 249 BPF_ALU64_REG(BPF_ADD, BPF_REG_0, BPF_REG_6),
|
D | atomic_invalid.c | 14 __INVALID_ATOMIC_ACCESS_TEST(BPF_ADD), 15 __INVALID_ATOMIC_ACCESS_TEST(BPF_ADD | BPF_FETCH), 16 __INVALID_ATOMIC_ACCESS_TEST(BPF_ADD), 17 __INVALID_ATOMIC_ACCESS_TEST(BPF_ADD | BPF_FETCH),
|
D | basic_instr.c | 5 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, 2), 8 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -1), 200 BPF_ALU64_REG(BPF_ADD, BPF_REG_1, BPF_REG_2), 213 BPF_ALU64_REG(BPF_ADD, BPF_REG_1, BPF_REG_2),
|
D | jump.c | 59 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 63 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -16), 67 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -32), 71 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -40), 75 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -48), 79 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -56), 388 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, -32767),
|
D | ctx_skb.c | 47 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 71 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 94 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 323 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), 340 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), 357 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), 360 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, 6), 1160 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 42), // 3. r3 += 42 1163 BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 14), // 6. r4 += 14 1182 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 42), // 3. r3 += 42 [all …]
|
D | jmp32.c | 798 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 809 BPF_ALU64_REG(BPF_ADD, BPF_REG_8, BPF_REG_6), 825 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 837 BPF_ALU64_REG(BPF_ADD, BPF_REG_8, BPF_REG_6), 853 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 865 BPF_ALU64_REG(BPF_ADD, BPF_REG_8, BPF_REG_6),
|
D | atomic_bounds.c | 18 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_1, -8),
|
/Linux-v6.6/lib/ |
D | test_bpf.c | 145 insn[i] = __BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, k); in bpf_fill_maxinsns3() 499 BPF_LSH, BPF_RSH, BPF_ARSH, BPF_ADD, in __bpf_fill_max_jmp() 570 case BPF_ADD: in __bpf_alu_result() 1053 return __bpf_fill_alu64_imm(self, BPF_ADD); in bpf_fill_alu64_add_imm() 1099 return __bpf_fill_alu32_imm(self, BPF_ADD); in bpf_fill_alu32_add_imm() 1145 return __bpf_fill_alu64_reg(self, BPF_ADD); in bpf_fill_alu64_add_reg() 1191 return __bpf_fill_alu32_reg(self, BPF_ADD); in bpf_fill_alu32_add_reg() 1313 return __bpf_fill_alu_imm_regs(self, BPF_ADD, false); in bpf_fill_alu64_add_imm_regs() 1374 return __bpf_fill_alu_imm_regs(self, BPF_ADD, true); in bpf_fill_alu32_add_imm_regs() 1505 return __bpf_fill_alu_reg_pairs(self, BPF_ADD, false); in bpf_fill_alu64_add_reg_pairs() [all …]
|
/Linux-v6.6/tools/testing/selftests/bpf/ |
D | test_sysctl.c | 231 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 272 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 314 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -24), 369 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -16), 418 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 460 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 496 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 535 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 574 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 610 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), [all …]
|
D | test_cgroup_storage.c | 25 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 0x1), in main() 33 BPF_ATOMIC_OP(BPF_DW, BPF_ADD, BPF_REG_0, BPF_REG_1, 0), in main()
|
D | disasm.c | 74 [BPF_ADD >> 4] = "+=", 102 [BPF_ADD >> 4] = "add", 212 (insn->imm == BPF_ADD || insn->imm == BPF_AND || in print_bpf_insn() 221 (insn->imm == (BPF_ADD | BPF_FETCH) || in print_bpf_insn()
|
/Linux-v6.6/samples/bpf/ |
D | test_cgrp2_attach.c | 52 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), /* r2 = fp - 4 */ in prog_load() 57 BPF_ATOMIC_OP(BPF_DW, BPF_ADD, BPF_REG_0, BPF_REG_1, 0), in prog_load() 63 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), /* r2 = fp - 4 */ in prog_load() 69 BPF_ATOMIC_OP(BPF_DW, BPF_ADD, BPF_REG_0, BPF_REG_1, 0), in prog_load()
|
D | cookie_uid_helper_example.c | 95 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), in prog_load() 138 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, -32), in prog_load() 150 BPF_ATOMIC_OP(BPF_DW, BPF_ADD, BPF_REG_9, BPF_REG_1, in prog_load() 154 BPF_ATOMIC_OP(BPF_DW, BPF_ADD, BPF_REG_9, BPF_REG_1, in prog_load()
|
D | sock_example.c | 53 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), /* r2 = fp - 4 */ in test_sock() 58 BPF_ATOMIC_OP(BPF_DW, BPF_ADD, BPF_REG_0, BPF_REG_1, 0), in test_sock()
|
/Linux-v6.6/arch/mips/net/ |
D | bpf_jit_comp64.c | 186 case BPF_ADD: in emit_alu_i64() 217 case BPF_ADD: in emit_alu_r64() 390 case BPF_ADD: in emit_atomic_r64() 391 case BPF_ADD | BPF_FETCH: in emit_atomic_r64() 694 case BPF_ALU | BPF_ADD | BPF_K: in build_insn() 729 case BPF_ALU | BPF_ADD | BPF_X: in build_insn() 768 case BPF_ALU64 | BPF_ADD | BPF_K: in build_insn() 797 case BPF_ALU64 | BPF_ADD | BPF_X: in build_insn() 852 case BPF_ADD: in build_insn() 853 case BPF_ADD | BPF_FETCH: in build_insn()
|
D | bpf_jit_comp32.c | 212 case BPF_ADD: in emit_alu_i64() 217 op = BPF_ADD; in emit_alu_i64() 227 case BPF_ADD: in emit_alu_i64() 272 case BPF_ADD: in emit_alu_r64() 738 case BPF_ADD: in emit_atomic_r32() 741 case BPF_ADD | BPF_FETCH: in emit_atomic_r32() 819 case BPF_ADD: in emit_atomic_r64() 822 case BPF_ADD | BPF_FETCH: in emit_atomic_r64() 1514 case BPF_ALU | BPF_ADD | BPF_K: in build_insn() 1544 case BPF_ALU | BPF_ADD | BPF_X: in build_insn() [all …]
|
/Linux-v6.6/kernel/bpf/ |
D | disasm.c | 74 [BPF_ADD >> 4] = "+=", 102 [BPF_ADD >> 4] = "add", 212 (insn->imm == BPF_ADD || insn->imm == BPF_AND || in print_bpf_insn() 221 (insn->imm == (BPF_ADD | BPF_FETCH) || in print_bpf_insn()
|