Lines Matching refs:skel
72 struct user_ringbuf_success *skel; in open_load_ringbuf_skel() local
75 skel = user_ringbuf_success__open(); in open_load_ringbuf_skel()
76 if (!ASSERT_OK_PTR(skel, "skel_open")) in open_load_ringbuf_skel()
79 err = bpf_map__set_max_entries(skel->maps.user_ringbuf, c_ringbuf_size); in open_load_ringbuf_skel()
83 err = bpf_map__set_max_entries(skel->maps.kernel_ringbuf, c_ringbuf_size); in open_load_ringbuf_skel()
87 err = user_ringbuf_success__load(skel); in open_load_ringbuf_skel()
91 return skel; in open_load_ringbuf_skel()
94 user_ringbuf_success__destroy(skel); in open_load_ringbuf_skel()
103 struct user_ringbuf_success *skel; in test_user_ringbuf_mappings() local
105 skel = open_load_ringbuf_skel(); in test_user_ringbuf_mappings()
106 if (!skel) in test_user_ringbuf_mappings()
109 rb_fd = bpf_map__fd(skel->maps.user_ringbuf); in test_user_ringbuf_mappings()
138 user_ringbuf_success__destroy(skel); in test_user_ringbuf_mappings()
146 struct user_ringbuf_success *skel; in load_skel_create_ringbufs() local
151 skel = open_load_ringbuf_skel(); in load_skel_create_ringbufs()
152 if (!skel) in load_skel_create_ringbufs()
156 skel->bss->pid = getpid(); in load_skel_create_ringbufs()
159 rb_fd = bpf_map__fd(skel->maps.kernel_ringbuf); in load_skel_create_ringbufs()
160 kern_ringbuf = ring_buffer__new(rb_fd, callback, skel, NULL); in load_skel_create_ringbufs()
168 rb_fd = bpf_map__fd(skel->maps.user_ringbuf); in load_skel_create_ringbufs()
174 ASSERT_EQ(skel->bss->read, 0, "no_reads_after_load"); in load_skel_create_ringbufs()
177 err = user_ringbuf_success__attach(skel); in load_skel_create_ringbufs()
181 *skel_out = skel; in load_skel_create_ringbufs()
191 user_ringbuf_success__destroy(skel); in load_skel_create_ringbufs()
201 static void manually_write_test_invalid_sample(struct user_ringbuf_success *skel, in manually_write_test_invalid_sample() argument
208 rb_fd = bpf_map__fd(skel->maps.user_ringbuf); in manually_write_test_invalid_sample()
210 ASSERT_EQ(skel->bss->read, 0, "num_samples_before_bad_sample"); in manually_write_test_invalid_sample()
228 ASSERT_EQ(skel->bss->read, 0, "num_samples_after_bad_sample"); in manually_write_test_invalid_sample()
229 ASSERT_EQ(skel->bss->err, err, "err_after_bad_sample"); in manually_write_test_invalid_sample()
237 struct user_ringbuf_success *skel; in test_user_ringbuf_post_misaligned() local
242 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_post_misaligned()
246 manually_write_test_invalid_sample(skel, size, size, -EINVAL); in test_user_ringbuf_post_misaligned()
248 user_ringbuf_success__destroy(skel); in test_user_ringbuf_post_misaligned()
253 struct user_ringbuf_success *skel; in test_user_ringbuf_post_producer_wrong_offset() local
258 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_post_producer_wrong_offset()
262 manually_write_test_invalid_sample(skel, size, size - 8, -EINVAL); in test_user_ringbuf_post_producer_wrong_offset()
264 user_ringbuf_success__destroy(skel); in test_user_ringbuf_post_producer_wrong_offset()
269 struct user_ringbuf_success *skel; in test_user_ringbuf_post_larger_than_ringbuf_sz() local
274 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_post_larger_than_ringbuf_sz()
278 manually_write_test_invalid_sample(skel, size, size, -E2BIG); in test_user_ringbuf_post_larger_than_ringbuf_sz()
280 user_ringbuf_success__destroy(skel); in test_user_ringbuf_post_larger_than_ringbuf_sz()
285 struct user_ringbuf_success *skel; in test_user_ringbuf_basic() local
289 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_basic()
293 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_basic()
299 ASSERT_EQ(skel->bss->read, 2, "num_samples_read_after"); in test_user_ringbuf_basic()
303 user_ringbuf_success__destroy(skel); in test_user_ringbuf_basic()
308 struct user_ringbuf_success *skel; in test_user_ringbuf_sample_full_ring_buffer() local
313 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_sample_full_ring_buffer()
322 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_sample_full_ring_buffer()
324 ASSERT_EQ(skel->bss->read, 1, "num_samples_read_after"); in test_user_ringbuf_sample_full_ring_buffer()
328 user_ringbuf_success__destroy(skel); in test_user_ringbuf_sample_full_ring_buffer()
333 struct user_ringbuf_success *skel; in test_user_ringbuf_post_alignment_autoadjust() local
338 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_post_alignment_autoadjust()
347 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_post_alignment_autoadjust()
349 ASSERT_EQ(skel->bss->read, 1, "num_samples_read_after"); in test_user_ringbuf_post_alignment_autoadjust()
352 user_ringbuf_success__destroy(skel); in test_user_ringbuf_post_alignment_autoadjust()
357 struct user_ringbuf_success *skel; in test_user_ringbuf_overfill() local
361 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_overfill()
367 ASSERT_EQ(skel->bss->read, c_max_entries, "max_entries"); in test_user_ringbuf_overfill()
370 user_ringbuf_success__destroy(skel); in test_user_ringbuf_overfill()
375 struct user_ringbuf_success *skel; in test_user_ringbuf_discards_properly_ignored() local
380 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_discards_properly_ignored()
384 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_discards_properly_ignored()
400 ASSERT_EQ(skel->bss->read, 0, "num_pre_kick"); in test_user_ringbuf_discards_properly_ignored()
402 ASSERT_EQ(skel->bss->read, 0, "num_post_kick"); in test_user_ringbuf_discards_properly_ignored()
415 user_ringbuf_success__destroy(skel); in test_user_ringbuf_discards_properly_ignored()
420 struct user_ringbuf_success *skel; in test_user_ringbuf_loop() local
427 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_loop()
446 ASSERT_EQ(skel->bss->read, total_samples - remaining_samples, in test_user_ringbuf_loop()
449 ASSERT_EQ(skel->bss->read, total_samples, "total_batched_entries"); in test_user_ringbuf_loop()
453 user_ringbuf_success__destroy(skel); in test_user_ringbuf_loop()
500 struct user_ringbuf_success *skel = ctx; in handle_kernel_msg() local
505 skel->bss->user_mutated += msg->operand_64; in handle_kernel_msg()
508 skel->bss->user_mutated += msg->operand_32; in handle_kernel_msg()
511 skel->bss->user_mutated *= msg->operand_64; in handle_kernel_msg()
514 skel->bss->user_mutated *= msg->operand_32; in handle_kernel_msg()
523 struct user_ringbuf_success *skel) in drain_kernel_messages_buffer() argument
529 ASSERT_OK(skel->bss->err, "consume_kern_ringbuf_err"); in drain_kernel_messages_buffer()
534 struct user_ringbuf_success *skel; in test_user_ringbuf_msg_protocol() local
540 err = load_skel_create_ringbufs(&skel, &kern_ringbuf, handle_kernel_msg, &user_ringbuf); in test_user_ringbuf_msg_protocol()
576 ASSERT_EQ(skel->bss->kern_mutated, expected_kern, "expected_kern"); in test_user_ringbuf_msg_protocol()
577 ASSERT_EQ(skel->bss->err, 0, "bpf_prog_err"); in test_user_ringbuf_msg_protocol()
578 drain_kernel_messages_buffer(kern_ringbuf, skel); in test_user_ringbuf_msg_protocol()
585 user_ringbuf_success__destroy(skel); in test_user_ringbuf_msg_protocol()
607 struct user_ringbuf_success *skel; in test_user_ringbuf_blocking_reserve() local
612 err = load_skel_create_user_ringbuf(&skel, &ringbuf); in test_user_ringbuf_blocking_reserve()
616 ASSERT_EQ(skel->bss->read, 0, "num_samples_read_before"); in test_user_ringbuf_blocking_reserve()
634 ASSERT_EQ(skel->bss->read, 0, "num_pre_kick"); in test_user_ringbuf_blocking_reserve()
654 ASSERT_GT(skel->bss->read, 0, "num_post_kill"); in test_user_ringbuf_blocking_reserve()
655 ASSERT_LE(skel->bss->read, num_written, "num_post_kill"); in test_user_ringbuf_blocking_reserve()
656 ASSERT_EQ(skel->bss->err, 0, "err_post_poll"); in test_user_ringbuf_blocking_reserve()
661 user_ringbuf_success__destroy(skel); in test_user_ringbuf_blocking_reserve()