| /linux/tools/testing/selftests/bpf/prog_tests/ |
| A D | tracing_struct.c | 10 struct tracing_struct *skel; in test_struct_args() local 13 skel = tracing_struct__open_and_load(); in test_struct_args() 17 err = tracing_struct__attach(skel); in test_struct_args() 25 ASSERT_EQ(skel->bss->t1_b, 1, "t1:b"); in test_struct_args() 26 ASSERT_EQ(skel->bss->t1_c, 4, "t1:c"); in test_struct_args() 35 ASSERT_EQ(skel->bss->t2_a, 1, "t2:a"); in test_struct_args() 38 ASSERT_EQ(skel->bss->t2_c, 4, "t2:c"); in test_struct_args() 41 ASSERT_EQ(skel->bss->t3_a, 1, "t3:a"); in test_struct_args() 42 ASSERT_EQ(skel->bss->t3_b, 4, "t3:b"); in test_struct_args() 48 ASSERT_EQ(skel->bss->t4_b, 1, "t4:b"); in test_struct_args() [all …]
|
| A D | attach_probe.c | 98 skel->links.handle_uprobe_byname = in test_attach_probe_manual() 130 skel->links.handle_kprobe_auto = bpf_program__attach(skel->progs.handle_kprobe_auto); in test_attach_probe_auto() 133 skel->links.handle_kretprobe_auto = bpf_program__attach(skel->progs.handle_kretprobe_auto); in test_attach_probe_auto() 254 skel->links.handle_kprobe_sleepable = bpf_program__attach(skel->progs.handle_kprobe_sleepable); in test_kprobe_sleepable() 268 skel->links.handle_uprobe_byname3 = bpf_program__attach(skel->progs.handle_uprobe_byname3); in test_uprobe_sleepable() 280 skel->bss->user_ptr = test_data; in test_uprobe_sleepable() 295 struct test_attach_probe *skel; in test_attach_probe() local 316 test_attach_probe_auto(skel); in test_attach_probe() 320 test_uprobe_lib(skel); in test_attach_probe() 322 test_uprobe_sleepable(skel); in test_attach_probe() [all …]
|
| A D | atomics.c | 13 prog_fd = skel->progs.add.prog_fd; in test_add() 167 struct atomics_lskel *skel; in test_atomics() local 173 if (skel->data->skip_tests) { in test_atomics() 179 skel->bss->pid = getpid(); in test_atomics() 182 test_add(skel); in test_atomics() 184 test_sub(skel); in test_atomics() 186 test_and(skel); in test_atomics() 188 test_or(skel); in test_atomics() 190 test_xor(skel); in test_atomics() 192 test_cmpxchg(skel); in test_atomics() [all …]
|
| A D | test_strncmp.c | 12 cmp = skel->bss->cmp_ret; in trigger_strncmp() 28 char *str = skel->bss->str; in strncmp_full_str_cmp() 37 got = trigger_strncmp(skel); in strncmp_full_str_cmp() 46 struct strncmp_test *skel; in test_strncmp_ret() local 49 skel = strncmp_test__open(); in test_strncmp_ret() 66 skel->bss->str[0] = '\0'; in test_strncmp_ret() 71 memcpy(skel->bss->str, skel->rodata->target, sizeof(skel->bss->str)); in test_strncmp_ret() 76 memcpy(skel->bss->str, skel->rodata->target, sizeof(skel->bss->str)); in test_strncmp_ret() 77 skel->bss->str[sizeof(skel->bss->str) - 1] = 'A'; in test_strncmp_ret() 89 struct strncmp_test *skel; in test_strncmp_bad_not_const_str_size() local [all …]
|
| A D | test_struct_ops_module.c | 60 struct struct_ops_module *skel; in test_struct_ops_load() local 65 skel = struct_ops_module__open(); in test_struct_ops_load() 70 skel->struct_ops.testmod_1->test_2 = skel->progs.test_3; in test_struct_ops_load() 94 if (!attach_ops_and_check(skel, skel->maps.testmod_1, 20)) in test_struct_ops_load() 96 if (!attach_ops_and_check(skel, skel->maps.testmod_2, 12)) in test_struct_ops_load() 100 struct_ops_module__destroy(skel); in test_struct_ops_load() 105 struct struct_ops_module *skel; in test_struct_ops_not_zeroed() local 146 skel->struct_ops.testmod_zeroed->zeroed_op = skel->progs.test_3; in test_struct_ops_not_zeroed() 161 struct struct_ops_module *skel; in test_struct_ops_incompatible() local 239 skel->struct_ops.ops->test_1 = skel->progs.test_1_forgotten; /* not anymore */ in test_struct_ops_forgotten_cb() [all …]
|
| A D | arena_atomics.c | 167 struct arena_atomics *skel; in test_arena_atomics() local 170 skel = arena_atomics__open(); in test_arena_atomics() 183 skel->bss->pid = getpid(); in test_arena_atomics() 186 test_add(skel); in test_arena_atomics() 188 test_sub(skel); in test_arena_atomics() 190 test_and(skel); in test_arena_atomics() 192 test_or(skel); in test_arena_atomics() 194 test_xor(skel); in test_arena_atomics() 196 test_cmpxchg(skel); in test_arena_atomics() 198 test_xchg(skel); in test_arena_atomics() [all …]
|
| A D | cgrp_local_storage.c | 25 #define CGROUP_MODE_SET(skel) \ argument 39 struct cgrp_ls_tp_btf *skel; in test_tp_btf() local 47 CGROUP_MODE_SET(skel); in test_tp_btf() 75 skel->bss->target_pid = 0; in test_tp_btf() 82 cgrp_ls_tp_btf__destroy(skel); in test_tp_btf() 150 CGROUP_MODE_SET(skel); in test_recursion() 165 struct cgrp_ls_negative *skel; in test_negative() local 187 CGROUP_MODE_SET(skel); in test_cgroup_iter_sleepable() 226 CGROUP_MODE_SET(skel); in test_yes_rcu_lock() 254 CGROUP_MODE_SET(skel); in test_no_rcu_lock() [all …]
|
| A D | bpf_loop.c | 17 skel->bss->nr_loops = 0; in check_nr_loops() 21 ASSERT_EQ(skel->bss->nr_loops_returned, skel->bss->nr_loops, in check_nr_loops() 25 skel->bss->nr_loops = 500; in check_nr_loops() 29 ASSERT_EQ(skel->bss->nr_loops_returned, skel->bss->nr_loops, in check_nr_loops() 34 skel->bss->nr_loops = -1; in check_nr_loops() 57 ASSERT_EQ(skel->bss->nr_loops_returned, skel->data->stop_index + 1, in check_callback_fn_stop() 74 skel->bss->nr_loops = 10; in check_null_callback_ctx() 78 ASSERT_EQ(skel->bss->nr_loops_returned, skel->bss->nr_loops, in check_null_callback_ctx() 183 struct bpf_loop *skel; in test_bpf_loop() local 192 check_nr_loops(skel); in test_bpf_loop() [all …]
|
| A D | struct_ops_autocreate.c | 9 struct struct_ops_autocreate *skel; in cant_load_full_object() local 13 skel = struct_ops_autocreate__open(); in cant_load_full_object() 24 err = struct_ops_autocreate__load(skel); in cant_load_full_object() 34 struct_ops_autocreate__destroy(skel); in cant_load_full_object() 54 struct struct_ops_autocreate *skel; in can_load_partial_object() local 57 skel = struct_ops_autocreate__open(); in can_load_partial_object() 75 check_test_1_link(skel, skel->maps.testmod_1); in can_load_partial_object() 78 struct_ops_autocreate__destroy(skel); in can_load_partial_object() 83 struct struct_ops_autocreate *skel; in optional_maps() local 105 check_test_1_link(skel, skel->maps.optional_map2); in optional_maps() [all …]
|
| A D | bpf_cookie.c | 404 skel->bss->tp_res = 0; in tp_subtest() 476 skel->bss->pe_res = 0; in pe_subtest() 553 skel->bss->lsm_res = 0; in lsm_subtest() 696 kprobe_subtest(skel); in test_bpf_cookie() 702 uprobe_subtest(skel); in test_bpf_cookie() 706 tp_subtest(skel); in test_bpf_cookie() 708 pe_subtest(skel); in test_bpf_cookie() 710 tracing_subtest(skel); in test_bpf_cookie() 712 lsm_subtest(skel); in test_bpf_cookie() 714 tp_btf_subtest(skel); in test_bpf_cookie() [all …]
|
| A D | bpf_iter.c | 104 *skel = NULL; in do_read_map_iter_fd() 229 skel->bss->tid = gettid(); in test_task_common_nocheck() 386 skel->bss->last_tgid = 0; in test_task_file() 387 skel->bss->count = 0; in test_task_file() 455 bss = skel->bss; in test_task_btf() 457 ret = do_btf_read(skel); in test_task_btf() 693 skel->rodata->ret1 = ret1; in test_overflow() 758 ASSERT_EQ(skel->bss->map2_seqnum1, skel->bss->map2_seqnum2, "map2_seqnum"); in test_overflow() 1044 do_read_map_iter_fd(&skel->skeleton, skel->progs.dump_bpf_array_map, in test_bpf_array_map_iter_fd() 1246 do_read_map_iter_fd(&skel->skeleton, skel->progs.rw_bpf_sk_storage_map, in test_bpf_sk_storage_map_iter_fd() [all …]
|
| A D | map_ops.c | 50 if (!skel) in setup() 72 if (skel && *skel) in teardown() 78 struct test_map_ops *skel; in map_ops_update_delete_subtest() local 80 if (setup(&skel)) in map_ops_update_delete_subtest() 98 teardown(&skel); in map_ops_update_delete_subtest() 103 struct test_map_ops *skel; in map_ops_push_peek_pop_subtest() local 105 if (setup(&skel)) in map_ops_push_peek_pop_subtest() 130 teardown(&skel); in map_ops_push_peek_pop_subtest() 135 struct test_map_ops *skel; in map_ops_for_each_subtest() local 137 if (setup(&skel)) in map_ops_for_each_subtest() [all …]
|
| A D | iters.c | 27 struct iters_num *skel; in subtest_num_iters() local 34 err = iters_num__attach(skel); in subtest_num_iters() 39 iters_num__detach(skel); in subtest_num_iters() 67 iters_num__destroy(skel); in subtest_num_iters() 109 struct iters_task_vma *skel; in subtest_task_vma_iters() local 126 iters_task_vma__detach(skel); in subtest_task_vma_iters() 193 iters_task__detach(skel); in subtest_task_iters() 202 iters_task__destroy(skel); in subtest_task_iters() 229 skel->bss->cg_id = cg_id; in subtest_css_task_iters() 288 iters_css__detach(skel); in subtest_css_iters() [all …]
|
| A D | find_vma.c | 19 skel->bss->found_vm_exec = 0; in test_and_reset_skel() 20 skel->data->find_addr_ret = -1; in test_and_reset_skel() 21 skel->data->find_zero_ret = -1; in test_and_reset_skel() 22 skel->bss->d_iname[0] = 0; in test_and_reset_skel() 84 err = find_vma__attach(skel); in test_find_vma_kprobe() 94 struct find_vma_fail1 *skel; in test_illegal_write_vma() local 103 struct find_vma_fail2 *skel; in test_illegal_write_task() local 112 struct find_vma *skel; in serial_test_find_vma() local 121 test_find_vma_pe(skel); in serial_test_find_vma() 122 test_find_vma_kprobe(skel); in serial_test_find_vma() [all …]
|
| A D | tc_links.c | 22 struct test_tc_link *skel; in serial_test_tc_links_basic() local 45 skel->links.tc1 = link; in serial_test_tc_links_basic() 80 skel->links.tc2 = link; in serial_test_tc_links_basic() 158 skel->links.tc1 = link; in test_tc_links_before_target() 168 skel->links.tc2 = link; in test_tc_links_before_target() 211 skel->links.tc3 = link; in test_tc_links_before_target() 224 skel->links.tc4 = link; in test_tc_links_before_target() 312 skel->links.tc1 = link; in test_tc_links_after_target() 759 err = bpf_link__update_program(skel->links.tc2, skel->progs.tc3); in test_tc_links_replace_target() 817 err = bpf_link__update_program(skel->links.tc1, skel->progs.tc1); in test_tc_links_replace_target() [all …]
|
| A D | rcu_read_lock.c | 17 struct rcu_read_lock *skel; in test_success() local 20 skel = rcu_read_lock__open(); in test_success() 36 err = rcu_read_lock__load(skel); in test_success() 49 rcu_read_lock__destroy(skel); in test_success() 54 struct rcu_read_lock *skel; in test_rcuptr_acquire() local 57 skel = rcu_read_lock__open(); in test_rcuptr_acquire() 71 rcu_read_lock__destroy(skel); in test_rcuptr_acquire() 88 struct rcu_read_lock *skel; in test_inproper_region() local 93 skel = rcu_read_lock__open(); in test_inproper_region() 104 rcu_read_lock__destroy(skel); in test_inproper_region() [all …]
|
| A D | cgroup1_hierarchy.c | 108 skel->bss->target_hid = hid; in test_cgroup1_hierarchy() 112 bpf_cgroup1(skel); in test_cgroup1_hierarchy() 118 bpf_cgroup1(skel); in test_cgroup1_hierarchy() 124 bpf_cgroup1_invalid_id(skel); in test_cgroup1_hierarchy() 129 bpf_cgroup1_invalid_id(skel); in test_cgroup1_hierarchy() 135 skel->bss->target_hid = -1; in test_cgroup1_hierarchy() 136 bpf_cgroup1_invalid_id(skel); in test_cgroup1_hierarchy() 142 bpf_cgroup1_invalid_id(skel); in test_cgroup1_hierarchy() 148 bpf_cgroup1_invalid_id(skel); in test_cgroup1_hierarchy() 152 skel->bss->target_hid = hid; in test_cgroup1_hierarchy() [all …]
|
| A D | btf_tag.c | 17 struct test_btf_decl_tag *skel; in test_btf_decl_tag() local 23 if (skel->rodata->skip_tests) { in test_btf_decl_tag() 28 test_btf_decl_tag__destroy(skel); in test_btf_decl_tag() 33 struct btf_type_tag *skel; in test_btf_type_tag() local 39 if (skel->rodata->skip_tests) { in test_btf_type_tag() 44 btf_type_tag__destroy(skel); in test_btf_type_tag() 113 struct btf_type_tag_user *skel; in test_btf_type_tag_mod_user() local 119 skel = btf_type_tag_user__open(); in test_btf_type_tag_mod_user() 132 btf_type_tag_user__destroy(skel); in test_btf_type_tag_mod_user() 141 struct btf_type_tag_user *skel; in test_btf_type_tag_vmlinux_user() local [all …]
|
| A D | percpu_alloc.c | 9 struct percpu_alloc_array *skel; in test_array() local 13 skel = percpu_alloc_array__open(); in test_array() 22 skel->bss->my_pid = getpid(); in test_array() 25 err = percpu_alloc_array__load(skel); in test_array() 40 percpu_alloc_array__destroy(skel); in test_array() 45 struct percpu_alloc_array *skel; in test_array_sleepable() local 49 skel = percpu_alloc_array__open(); in test_array_sleepable() 55 skel->bss->my_pid = getpid(); in test_array_sleepable() 58 err = percpu_alloc_array__load(skel); in test_array_sleepable() 73 percpu_alloc_array__destroy(skel); in test_array_sleepable() [all …]
|
| A D | rbtree.c | 19 struct rbtree *skel; in test_rbtree_add_nodes() local 31 rbtree__destroy(skel); in test_rbtree_add_nodes() 41 struct rbtree *skel; in test_rbtree_add_nodes_nested() local 53 rbtree__destroy(skel); in test_rbtree_add_nodes_nested() 63 struct rbtree *skel; in test_rbtree_add_and_remove() local 75 rbtree__destroy(skel); in test_rbtree_add_and_remove() 85 struct rbtree *skel; in test_rbtree_add_and_remove_array() local 96 rbtree__destroy(skel); in test_rbtree_add_and_remove_array() 106 struct rbtree *skel; in test_rbtree_first_and_remove() local 120 rbtree__destroy(skel); in test_rbtree_first_and_remove() [all …]
|
| A D | global_map_resize.c | 35 skel->bss->array[0] = 1; in global_map_resize_bss_subtest() 38 map = skel->maps.bss; in global_map_resize_bss_subtest() 50 array_len = (desired_sz - sizeof(skel->bss->sum)) / sizeof(skel->bss->array[0]); in global_map_resize_bss_subtest() 54 skel->bss = bpf_map__initial_value(skel->maps.bss, &actual_sz); in global_map_resize_bss_subtest() 64 skel->bss->array[i] = 1; in global_map_resize_bss_subtest() 67 skel->rodata->pid = getpid(); in global_map_resize_bss_subtest() 108 map = skel->maps.data_custom; in global_map_resize_data_subtest() 120 array_len = (desired_sz - sizeof(skel->bss->sum)) / sizeof(skel->data_custom->my_array[0]); in global_map_resize_data_subtest() 124 skel->data_custom = bpf_map__initial_value(skel->maps.data_custom, &actual_sz); in global_map_resize_data_subtest() 137 skel->rodata->pid = getpid(); in global_map_resize_data_subtest() [all …]
|
| /linux/kernel/bpf/preload/iterators/ |
| A D | iterators.lskel-big-endian.h | 64 if (!skel) in iterators_bpf__destroy() 66 iterators_bpf__detach(skel); in iterators_bpf__destroy() 70 skel_free(skel); in iterators_bpf__destroy() 75 struct iterators_bpf *skel; in iterators_bpf__open() local 77 skel = skel_alloc(sizeof(*skel)); in iterators_bpf__open() 78 if (!skel) in iterators_bpf__open() 80 skel->ctx.sz = (void *)&skel->links - (void *)skel; in iterators_bpf__open() 81 return skel; in iterators_bpf__open() 83 iterators_bpf__destroy(skel); in iterators_bpf__open() 399 if (!skel) in iterators_bpf__open_and_load() [all …]
|
| A D | iterators.lskel-little-endian.h | 64 if (!skel) in iterators_bpf__destroy() 66 iterators_bpf__detach(skel); in iterators_bpf__destroy() 70 skel_free(skel); in iterators_bpf__destroy() 75 struct iterators_bpf *skel; in iterators_bpf__open() local 77 skel = skel_alloc(sizeof(*skel)); in iterators_bpf__open() 78 if (!skel) in iterators_bpf__open() 80 skel->ctx.sz = (void *)&skel->links - (void *)skel; in iterators_bpf__open() 81 return skel; in iterators_bpf__open() 83 iterators_bpf__destroy(skel); in iterators_bpf__open() 415 if (!skel) in iterators_bpf__open_and_load() [all …]
|
| /linux/samples/v4l/ |
| A D | v4l2-pci-skeleton.c | 160 skel->field = skel->format.field; in queue_setup() 404 if (skel->input) in skeleton_s_std() 424 skel->std = std; in skeleton_s_std() 427 skeleton_fill_pix_format(skel, &skel->format); in skeleton_s_std() 436 if (skel->input) in skeleton_g_std() 455 if (skel->input) in skeleton_querystd() 512 skeleton_fill_pix_format(skel, &skel->format); in skeleton_s_dv_timings() 625 skel->input = i; in skeleton_s_input() 634 skeleton_fill_pix_format(skel, &skel->format); in skeleton_s_input() 766 if (!skel) { in skeleton_probe() [all …]
|
| /linux/tools/sched_ext/ |
| A D | scx_qmap.c | 57 struct scx_qmap *skel; in main() local 95 if (skel->rodata->disallow_tgid < 0) in main() 102 skel->rodata->suppress_dump = true; in main() 125 skel->bss->nr_reenqueued, skel->bss->nr_dequeued, in main() 127 skel->bss->nr_ddsp_from_enq); in main() 129 skel->bss->nr_expedited_local, in main() 135 skel->bss->cpuperf_min, in main() 136 skel->bss->cpuperf_avg, in main() 137 skel->bss->cpuperf_max, in main() 146 UEI_REPORT(skel, uei); in main() [all …]
|