| /tools/testing/selftests/bpf/benchs/ |
| A D | run_bench_ringbufs.sh | 10 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 15 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 20 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 43 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
|
| A D | bench_ringbufs.c | 451 struct perf_buffer *pb; member 478 struct perf_buffer *pb = ctx->perfbuf; in perfbuf_custom_consumer() local 481 size_t mmap_mask = pb->mmap_size - 1; in perfbuf_custom_consumer() 491 cnt = epoll_wait(pb->epoll_fd, pb->events, pb->cpu_cnt, -1); in perfbuf_custom_consumer() 498 cpu_buf = pb->events[i].data.ptr; in perfbuf_custom_consumer() 500 base = ((void *)header) + pb->page_size; in perfbuf_custom_consumer()
|
| /tools/testing/selftests/bpf/prog_tests/ |
| A D | perf_buffer.c | 52 struct perf_buffer *pb; in serial_test_perf_buffer() local 84 pb = perf_buffer__new(bpf_map__fd(skel->maps.perf_buf_map), 1, in serial_test_perf_buffer() 86 if (!ASSERT_OK_PTR(pb, "perf_buf__new")) in serial_test_perf_buffer() 89 CHECK(perf_buffer__epoll_fd(pb) < 0, "epoll_fd", in serial_test_perf_buffer() 90 "bad fd: %d\n", perf_buffer__epoll_fd(pb)); in serial_test_perf_buffer() 105 err = perf_buffer__poll(pb, 100); in serial_test_perf_buffer() 113 if (CHECK(perf_buffer__buffer_cnt(pb) != nr_on_cpus, "buf_cnt", in serial_test_perf_buffer() 121 fd = perf_buffer__buffer_fd(pb, j); in serial_test_perf_buffer() 125 err = perf_buffer__consume_buffer(pb, j); in serial_test_perf_buffer() 133 err = perf_buffer__consume_buffer(pb, j); in serial_test_perf_buffer() [all …]
|
| A D | xdp_bpf2bpf.c | 44 static void run_xdp_bpf2bpf_pkt_size(int pkt_fd, struct perf_buffer *pb, in run_xdp_bpf2bpf_pkt_size() argument 90 err = perf_buffer__poll(pb, 100); in run_xdp_bpf2bpf_pkt_size() 112 struct perf_buffer *pb = NULL; in test_xdp_bpf2bpf() local 149 pb = perf_buffer__new(bpf_map__fd(ftrace_skel->maps.perf_buf_map), 8, in test_xdp_bpf2bpf() 151 if (!ASSERT_OK_PTR(pb, "perf_buf__new")) in test_xdp_bpf2bpf() 155 run_xdp_bpf2bpf_pkt_size(pkt_fd, pb, ftrace_skel, in test_xdp_bpf2bpf() 158 perf_buffer__free(pb); in test_xdp_bpf2bpf()
|
| A D | get_stack_raw_tp.c | 91 struct perf_buffer *pb = NULL; in test_get_stack_raw_tp() local 129 pb = perf_buffer__new(bpf_map__fd(map), 8, get_stack_print_output, in test_get_stack_raw_tp() 131 if (!ASSERT_OK_PTR(pb, "perf_buf__new")) in test_get_stack_raw_tp() 139 err = perf_buffer__poll(pb, 100); in test_get_stack_raw_tp() 147 perf_buffer__free(pb); in test_get_stack_raw_tp()
|
| A D | kfree_skb.c | 66 struct perf_buffer *pb = NULL; in serial_test_kfree_skb() local 98 pb = perf_buffer__new(bpf_map__fd(skel->maps.perf_buf_map), 1, in serial_test_kfree_skb() 100 if (!ASSERT_OK_PTR(pb, "perf_buf__new")) in serial_test_kfree_skb() 109 err = perf_buffer__poll(pb, 100); in serial_test_kfree_skb() 125 perf_buffer__free(pb); in serial_test_kfree_skb()
|
| A D | uprobe_syscall.c | 99 unsigned long *pb = (unsigned long *) &before; in test_uretprobe_regs_equal() local 116 cnt = sizeof(before)/sizeof(*pb); in test_uretprobe_regs_equal() 130 if (!ASSERT_EQ(pb[i], pa[i], "register before-after value check")) in test_uretprobe_regs_equal() 181 unsigned long *pb = (unsigned long *) &before; in test_uretprobe_regs_change() local 183 unsigned long cnt = sizeof(before)/sizeof(*pb); in test_uretprobe_regs_change() 212 if (!ASSERT_EQ(pa[i], pb[i], "register before-after value check")) in test_uretprobe_regs_change()
|
| A D | xdp_attach.c | 108 struct perf_buffer *pb = NULL; in test_xdp_attach_fail() local 123 pb = perf_buffer__new(bpf_map__fd(skel->maps.xdp_errmsg_pb), 1, in test_xdp_attach_fail() 125 if (!ASSERT_OK_PTR(pb, "perf_buffer__new")) in test_xdp_attach_fail() 138 err = perf_buffer__poll(pb, 100); in test_xdp_attach_fail() 146 perf_buffer__free(pb); in test_xdp_attach_fail()
|
| /tools/testing/selftests/bpf/ |
| A D | test_tcpnotify_user.c | 44 void tcp_notifier_poller(struct perf_buffer *pb) in tcp_notifier_poller() argument 49 err = perf_buffer__poll(pb, 100); in tcp_notifier_poller() 59 struct perf_buffer *pb = arg; in poller_thread() local 61 tcp_notifier_poller(pb); in poller_thread() 75 struct perf_buffer *pb = NULL; in main() local 118 pb = perf_buffer__new(bpf_map__fd(perf_map), 8, dummyfn, NULL, NULL, NULL); in main() 119 if (!pb) in main() 122 pthread_create(&tid, NULL, poller_thread, pb); in main() 166 perf_buffer__free(pb); in main()
|
| /tools/bpf/runqslower/ |
| A D | runqslower.c | 115 struct perf_buffer *pb = NULL; in main() local 153 pb = perf_buffer__new(bpf_map__fd(obj->maps.events), 64, in main() 155 err = libbpf_get_error(pb); in main() 157 pb = NULL; in main() 162 while ((err = perf_buffer__poll(pb, 100)) >= 0) in main() 167 perf_buffer__free(pb); in main()
|
| /tools/bpf/bpftool/ |
| A D | map_perf_ring.c | 126 struct perf_buffer *pb; in do_event_pipe() local 190 pb = perf_buffer__new_raw(map_fd, MMAP_PAGE_CNT, &perf_attr, in do_event_pipe() 192 if (!pb) { in do_event_pipe() 206 err = perf_buffer__poll(pb, 200); in do_event_pipe() 217 perf_buffer__free(pb); in do_event_pipe() 223 perf_buffer__free(pb); in do_event_pipe()
|
| A D | btf.c | 502 static int cmp_kfuncs(const void *pa, const void *pb, void *ctx) in cmp_kfuncs() argument 506 const struct btf_type *b = *(void **)pb; in cmp_kfuncs()
|
| /tools/testing/ktest/ |
| A D | config-bisect.pl | 320 my ($pa, $pb) = @_; 324 my %b = %{$pb}; 339 my ($pa, $pb) = @_; 345 my %b = %{$pb}; 361 my ($pa, $pb) = @_; 367 my %b = %{$pb};
|
| /tools/memory-model/ |
| A D | linux-kernel.cat | 117 let pb = prop ; strong-fence ; hb* ; [Marked] 118 acyclic pb as propagation 144 let rcu-link = po? ; hb* ; pb* ; prop ; po 168 (* rb orders instructions just as pb does *) 169 let rb = prop ; rcu-fence ; hb* ; pb* ; [Marked] 178 * let xb = hb | pb | rb 192 let xbstar = (hb | pb | rb)*
|
| /tools/lib/bpf/ |
| A D | libbpf.c | 13360 cpu_buf->pb = pb; in perf_buffer__open_cpu_buf() 13497 pb = calloc(1, sizeof(*pb)); in __perf_buffer__new() 13498 if (!pb) in __perf_buffer__new() 13507 pb->mmap_size = pb->page_size * page_cnt; in __perf_buffer__new() 13530 pb->events = calloc(pb->cpu_cnt, sizeof(*pb->events)); in __perf_buffer__new() 13536 pb->cpu_bufs = calloc(pb->cpu_cnt, sizeof(*pb->cpu_bufs)); in __perf_buffer__new() 13599 if (pb) in __perf_buffer__new() 13621 struct perf_buffer *pb = cpu_buf->pb; in perf_buffer__process_record() local 13626 return pb->event_cb(pb->ctx, cpu_buf->cpu, e); in perf_buffer__process_record() 13640 pb->lost_cb(pb->ctx, cpu_buf->cpu, s->lost); in perf_buffer__process_record() [all …]
|
| A D | libbpf.h | 1635 LIBBPF_API void perf_buffer__free(struct perf_buffer *pb); 1636 LIBBPF_API int perf_buffer__epoll_fd(const struct perf_buffer *pb); 1637 LIBBPF_API int perf_buffer__poll(struct perf_buffer *pb, int timeout_ms); 1638 LIBBPF_API int perf_buffer__consume(struct perf_buffer *pb); 1639 LIBBPF_API int perf_buffer__consume_buffer(struct perf_buffer *pb, size_t buf_idx); 1640 LIBBPF_API size_t perf_buffer__buffer_cnt(const struct perf_buffer *pb); 1641 LIBBPF_API int perf_buffer__buffer_fd(const struct perf_buffer *pb, size_t buf_idx); 1656 LIBBPF_API int perf_buffer__buffer(struct perf_buffer *pb, int buf_idx, void **buf,
|
| /tools/bpf/resolve_btfids/ |
| A D | main.c | 672 static int cmp_id(const void *pa, const void *pb) in cmp_id() argument 674 const int *a = pa, *b = pb; in cmp_id()
|
| /tools/memory-model/Documentation/ |
| A D | explanation.txt | 29 21. THE PROPAGATES-BEFORE RELATION: pb 1433 THE PROPAGATES-BEFORE RELATION: pb 1436 The propagates-before (pb) relation capitalizes on the special 1463 The existence of a pb link from E to F implies that E must execute 1495 If r0 = 0 at the end then there is a pb link from P0's load to P1's 1499 Note that this pb link is not included in hb as an instance of prop, 1508 In summary, the fact that the pb relation links events in the order 1695 rcu-fence. This is done in essentially the same way as the pb 1698 before F, just as E ->pb F does (and for much the same reasons). 2265 if they can be connected by a sequence of hb, pb, and rb links [all …]
|