Lines Matching refs:skel

74 	struct user_ringbuf_success *skel;  in open_load_ringbuf_skel()  local
77 skel = user_ringbuf_success__open(); in open_load_ringbuf_skel()
78 if (!ASSERT_OK_PTR(skel, "skel_open")) in open_load_ringbuf_skel()
81 err = bpf_map__set_max_entries(skel->maps.user_ringbuf, c_ringbuf_size); in open_load_ringbuf_skel()
85 err = bpf_map__set_max_entries(skel->maps.kernel_ringbuf, c_ringbuf_size); in open_load_ringbuf_skel()
89 err = user_ringbuf_success__load(skel); in open_load_ringbuf_skel()
93 return skel; in open_load_ringbuf_skel()
96 user_ringbuf_success__destroy(skel); in open_load_ringbuf_skel()
105 struct user_ringbuf_success *skel; in test_user_ringbuf_mappings() local
107 skel = open_load_ringbuf_skel(); in test_user_ringbuf_mappings()
108 if (!skel) in test_user_ringbuf_mappings()
111 rb_fd = bpf_map__fd(skel->maps.user_ringbuf); in test_user_ringbuf_mappings()
140 user_ringbuf_success__destroy(skel); in test_user_ringbuf_mappings()
148 struct user_ringbuf_success *skel; in load_skel_create_ringbufs() local
153 skel = open_load_ringbuf_skel(); in load_skel_create_ringbufs()
154 if (!skel) in load_skel_create_ringbufs()
158 skel->bss->pid = getpid(); in load_skel_create_ringbufs()
161 rb_fd = bpf_map__fd(skel->maps.kernel_ringbuf); in load_skel_create_ringbufs()
162 kern_ringbuf = ring_buffer__new(rb_fd, callback, skel, NULL); in load_skel_create_ringbufs()
170 rb_fd = bpf_map__fd(skel->maps.user_ringbuf); in load_skel_create_ringbufs()
176 ASSERT_EQ(skel->bss->read, 0, "no_reads_after_load"); in load_skel_create_ringbufs()
179 err = user_ringbuf_success__attach(skel); in load_skel_create_ringbufs()
183 *skel_out = skel; in load_skel_create_ringbufs()
193 user_ringbuf_success__destroy(skel); in load_skel_create_ringbufs()
203 static void manually_write_test_invalid_sample(struct user_ringbuf_success *skel, in manually_write_test_invalid_sample() argument
210 rb_fd = bpf_map__fd(skel->maps.user_ringbuf); in manually_write_test_invalid_sample()
212 ASSERT_EQ(skel->bss->read, 0, "num_samples_before_bad_sample"); in manually_write_test_invalid_sample()
230 ASSERT_EQ(skel->bss->read, 0, "num_samples_after_bad_sample"); in manually_write_test_invalid_sample()
231 ASSERT_EQ(skel->bss->err, err, "err_after_bad_sample"); in manually_write_test_invalid_sample()
239 struct user_ringbuf_success *skel; in test_user_ringbuf_post_misaligned() local
244 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_post_misaligned()
248 manually_write_test_invalid_sample(skel, size, size, -EINVAL); in test_user_ringbuf_post_misaligned()
250 user_ringbuf_success__destroy(skel); in test_user_ringbuf_post_misaligned()
255 struct user_ringbuf_success *skel; in test_user_ringbuf_post_producer_wrong_offset() local
260 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_post_producer_wrong_offset()
264 manually_write_test_invalid_sample(skel, size, size - 8, -EINVAL); in test_user_ringbuf_post_producer_wrong_offset()
266 user_ringbuf_success__destroy(skel); in test_user_ringbuf_post_producer_wrong_offset()
271 struct user_ringbuf_success *skel; in test_user_ringbuf_post_larger_than_ringbuf_sz() local
276 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_post_larger_than_ringbuf_sz()
280 manually_write_test_invalid_sample(skel, size, size, -E2BIG); in test_user_ringbuf_post_larger_than_ringbuf_sz()
282 user_ringbuf_success__destroy(skel); in test_user_ringbuf_post_larger_than_ringbuf_sz()
287 struct user_ringbuf_success *skel; in test_user_ringbuf_basic() local
291 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_basic()
295 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_basic()
301 ASSERT_EQ(skel->bss->read, 2, "num_samples_read_after"); in test_user_ringbuf_basic()
305 user_ringbuf_success__destroy(skel); in test_user_ringbuf_basic()
310 struct user_ringbuf_success *skel; in test_user_ringbuf_sample_full_ring_buffer() local
315 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_sample_full_ring_buffer()
324 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_sample_full_ring_buffer()
326 ASSERT_EQ(skel->bss->read, 1, "num_samples_read_after"); in test_user_ringbuf_sample_full_ring_buffer()
330 user_ringbuf_success__destroy(skel); in test_user_ringbuf_sample_full_ring_buffer()
335 struct user_ringbuf_success *skel; in test_user_ringbuf_post_alignment_autoadjust() local
340 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_post_alignment_autoadjust()
349 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_post_alignment_autoadjust()
351 ASSERT_EQ(skel->bss->read, 1, "num_samples_read_after"); in test_user_ringbuf_post_alignment_autoadjust()
354 user_ringbuf_success__destroy(skel); in test_user_ringbuf_post_alignment_autoadjust()
359 struct user_ringbuf_success *skel; in test_user_ringbuf_overfill() local
363 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_overfill()
369 ASSERT_EQ(skel->bss->read, c_max_entries, "max_entries"); in test_user_ringbuf_overfill()
372 user_ringbuf_success__destroy(skel); in test_user_ringbuf_overfill()
377 struct user_ringbuf_success *skel; in test_user_ringbuf_discards_properly_ignored() local
382 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_discards_properly_ignored()
386 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_discards_properly_ignored()
402 ASSERT_EQ(skel->bss->read, 0, "num_pre_kick"); in test_user_ringbuf_discards_properly_ignored()
404 ASSERT_EQ(skel->bss->read, 0, "num_post_kick"); in test_user_ringbuf_discards_properly_ignored()
417 user_ringbuf_success__destroy(skel); in test_user_ringbuf_discards_properly_ignored()
422 struct user_ringbuf_success *skel; in test_user_ringbuf_loop() local
429 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_loop()
448 ASSERT_EQ(skel->bss->read, total_samples - remaining_samples, in test_user_ringbuf_loop()
451 ASSERT_EQ(skel->bss->read, total_samples, "total_batched_entries"); in test_user_ringbuf_loop()
455 user_ringbuf_success__destroy(skel); in test_user_ringbuf_loop()
502 struct user_ringbuf_success *skel = ctx; in handle_kernel_msg() local
507 skel->bss->user_mutated += msg->operand_64; in handle_kernel_msg()
510 skel->bss->user_mutated += msg->operand_32; in handle_kernel_msg()
513 skel->bss->user_mutated *= msg->operand_64; in handle_kernel_msg()
516 skel->bss->user_mutated *= msg->operand_32; in handle_kernel_msg()
525 struct user_ringbuf_success *skel) in drain_kernel_messages_buffer() argument
531 ASSERT_OK(skel->bss->err, "consume_kern_ringbuf_err"); in drain_kernel_messages_buffer()
536 struct user_ringbuf_success *skel; in test_user_ringbuf_msg_protocol() local
542 err = load_skel_create_ringbufs(&skel, &kern_ringbuf, handle_kernel_msg, &user_ringbuf); in test_user_ringbuf_msg_protocol()
578 ASSERT_EQ(skel->bss->kern_mutated, expected_kern, "expected_kern"); in test_user_ringbuf_msg_protocol()
579 ASSERT_EQ(skel->bss->err, 0, "bpf_prog_err"); in test_user_ringbuf_msg_protocol()
580 drain_kernel_messages_buffer(kern_ringbuf, skel); in test_user_ringbuf_msg_protocol()
587 user_ringbuf_success__destroy(skel); in test_user_ringbuf_msg_protocol()
609 struct user_ringbuf_success *skel; in test_user_ringbuf_blocking_reserve() local
614 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_blocking_reserve()
618 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_blocking_reserve()
636 ASSERT_EQ(skel->bss->read, 0, "num_pre_kick"); in test_user_ringbuf_blocking_reserve()
656 ASSERT_GT(skel->bss->read, 0, "num_post_kill"); in test_user_ringbuf_blocking_reserve()
657 ASSERT_LE(skel->bss->read, num_written, "num_post_kill"); in test_user_ringbuf_blocking_reserve()
658 ASSERT_EQ(skel->bss->err, 0, "err_post_poll"); in test_user_ringbuf_blocking_reserve()
663 user_ringbuf_success__destroy(skel); in test_user_ringbuf_blocking_reserve()
705 struct user_ringbuf_fail *skel; in verify_fail() local
712 skel = user_ringbuf_fail__open_opts(&opts); in verify_fail()
713 if (!ASSERT_OK_PTR(skel, "dynptr_fail__open_opts")) in verify_fail()
716 prog = bpf_object__find_program_by_name(skel->obj, prog_name); in verify_fail()
722 bpf_map__set_max_entries(skel->maps.user_ringbuf, getpagesize()); in verify_fail()
724 err = user_ringbuf_fail__load(skel); in verify_fail()
734 user_ringbuf_fail__destroy(skel); in verify_fail()