/tools/testing/selftests/bpf/prog_tests/ |
D | perf_buffer.c | 52 struct perf_buffer *pb; in test_perf_buffer() local 82 pb = perf_buffer__new(bpf_map__fd(skel->maps.perf_buf_map), 1, &pb_opts); in test_perf_buffer() 83 if (!ASSERT_OK_PTR(pb, "perf_buf__new")) in test_perf_buffer() 86 CHECK(perf_buffer__epoll_fd(pb) < 0, "epoll_fd", in test_perf_buffer() 87 "bad fd: %d\n", perf_buffer__epoll_fd(pb)); in test_perf_buffer() 102 err = perf_buffer__poll(pb, 100); in test_perf_buffer() 110 if (CHECK(perf_buffer__buffer_cnt(pb) != nr_on_cpus, "buf_cnt", in test_perf_buffer() 111 "got %zu, expected %d\n", perf_buffer__buffer_cnt(pb), nr_on_cpus)) in test_perf_buffer() 118 fd = perf_buffer__buffer_fd(pb, i); in test_perf_buffer() 122 err = perf_buffer__consume_buffer(pb, i); in test_perf_buffer() [all …]
|
D | xdp_bpf2bpf.c | 51 struct perf_buffer *pb = NULL; in test_xdp_bpf2bpf() local 91 pb = perf_buffer__new(bpf_map__fd(ftrace_skel->maps.perf_buf_map), in test_xdp_bpf2bpf() 93 if (!ASSERT_OK_PTR(pb, "perf_buf__new")) in test_xdp_bpf2bpf() 109 err = perf_buffer__poll(pb, 100); in test_xdp_bpf2bpf() 125 if (pb) in test_xdp_bpf2bpf() 126 perf_buffer__free(pb); in test_xdp_bpf2bpf()
|
D | get_stack_raw_tp.c | 89 struct perf_buffer *pb = NULL; in test_get_stack_raw_tp() local 128 pb = perf_buffer__new(bpf_map__fd(map), 8, &pb_opts); in test_get_stack_raw_tp() 129 if (!ASSERT_OK_PTR(pb, "perf_buf__new")) in test_get_stack_raw_tp() 137 err = perf_buffer__poll(pb, 100); in test_get_stack_raw_tp() 145 perf_buffer__free(pb); in test_get_stack_raw_tp()
|
D | kfree_skb.c | 69 struct perf_buffer *pb = NULL; in test_kfree_skb() local 116 pb = perf_buffer__new(bpf_map__fd(perf_buf_map), 1, &pb_opts); in test_kfree_skb() 117 if (!ASSERT_OK_PTR(pb, "perf_buf__new")) in test_kfree_skb() 128 err = perf_buffer__poll(pb, 100); in test_kfree_skb() 144 perf_buffer__free(pb); in test_kfree_skb()
|
/tools/testing/selftests/bpf/benchs/ |
D | run_bench_ringbufs.sh | 34 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 39 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 44 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 67 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
|
D | bench_ringbufs.c | 455 struct perf_buffer *pb; member 482 struct perf_buffer *pb = ctx->perfbuf; in perfbuf_custom_consumer() local 485 size_t mmap_mask = pb->mmap_size - 1; in perfbuf_custom_consumer() 495 cnt = epoll_wait(pb->epoll_fd, pb->events, pb->cpu_cnt, -1); in perfbuf_custom_consumer() 502 cpu_buf = pb->events[i].data.ptr; in perfbuf_custom_consumer() 504 base = ((void *)header) + pb->page_size; in perfbuf_custom_consumer()
|
/tools/testing/selftests/bpf/ |
D | test_tcpnotify_user.c | 44 void tcp_notifier_poller(struct perf_buffer *pb) in tcp_notifier_poller() argument 49 err = perf_buffer__poll(pb, 100); in tcp_notifier_poller() 59 struct perf_buffer *pb = arg; in poller_thread() local 61 tcp_notifier_poller(pb); in poller_thread() 76 struct perf_buffer *pb = NULL; in main() local 120 pb = perf_buffer__new(bpf_map__fd(perf_map), 8, &pb_opts); in main() 121 if (!pb) in main() 124 pthread_create(&tid, NULL, poller_thread, pb); in main() 168 perf_buffer__free(pb); in main()
|
/tools/bpf/runqslower/ |
D | runqslower.c | 127 struct perf_buffer *pb = NULL; in main() local 170 pb = perf_buffer__new(bpf_map__fd(obj->maps.events), 64, &pb_opts); in main() 171 err = libbpf_get_error(pb); in main() 173 pb = NULL; in main() 178 while ((err = perf_buffer__poll(pb, 100)) >= 0) in main() 183 perf_buffer__free(pb); in main()
|
/tools/bpf/bpftool/ |
D | map_perf_ring.c | 134 struct perf_buffer *pb; in do_event_pipe() local 201 pb = perf_buffer__new_raw(map_fd, MMAP_PAGE_CNT, &opts); in do_event_pipe() 202 err = libbpf_get_error(pb); in do_event_pipe() 217 err = perf_buffer__poll(pb, 200); in do_event_pipe() 228 perf_buffer__free(pb); in do_event_pipe() 234 perf_buffer__free(pb); in do_event_pipe()
|
/tools/testing/ktest/ |
D | config-bisect.pl | 320 my ($pa, $pb) = @_; 324 my %b = %{$pb}; 339 my ($pa, $pb) = @_; 345 my %b = %{$pb}; 361 my ($pa, $pb) = @_; 367 my %b = %{$pb};
|
/tools/memory-model/ |
D | linux-kernel.cat | 94 let pb = prop ; strong-fence ; hb* ; [Marked] 95 acyclic pb as propagation 121 let rcu-link = po? ; hb* ; pb* ; prop ; po 145 (* rb orders instructions just as pb does *) 146 let rb = prop ; rcu-fence ; hb* ; pb* ; [Marked] 155 * let xb = hb | pb | rb 169 let xbstar = (hb | pb | rb)*
|
/tools/lib/bpf/ |
D | libbpf.c | 9834 struct perf_buffer *pb; member 9858 static void perf_buffer__free_cpu_buf(struct perf_buffer *pb, in perf_buffer__free_cpu_buf() argument 9864 munmap(cpu_buf->base, pb->mmap_size + pb->page_size)) in perf_buffer__free_cpu_buf() 9874 void perf_buffer__free(struct perf_buffer *pb) in perf_buffer__free() argument 9878 if (IS_ERR_OR_NULL(pb)) in perf_buffer__free() 9880 if (pb->cpu_bufs) { in perf_buffer__free() 9881 for (i = 0; i < pb->cpu_cnt; i++) { in perf_buffer__free() 9882 struct perf_cpu_buf *cpu_buf = pb->cpu_bufs[i]; in perf_buffer__free() 9887 bpf_map_delete_elem(pb->map_fd, &cpu_buf->map_key); in perf_buffer__free() 9888 perf_buffer__free_cpu_buf(pb, cpu_buf); in perf_buffer__free() [all …]
|
D | libbpf.h | 718 LIBBPF_API void perf_buffer__free(struct perf_buffer *pb); 719 LIBBPF_API int perf_buffer__epoll_fd(const struct perf_buffer *pb); 720 LIBBPF_API int perf_buffer__poll(struct perf_buffer *pb, int timeout_ms); 721 LIBBPF_API int perf_buffer__consume(struct perf_buffer *pb); 722 LIBBPF_API int perf_buffer__consume_buffer(struct perf_buffer *pb, size_t buf_idx); 723 LIBBPF_API size_t perf_buffer__buffer_cnt(const struct perf_buffer *pb); 724 LIBBPF_API int perf_buffer__buffer_fd(const struct perf_buffer *pb, size_t buf_idx);
|
/tools/bpf/resolve_btfids/ |
D | main.c | 599 static int cmp_id(const void *pa, const void *pb) in cmp_id() argument 601 const int *a = pa, *b = pb; in cmp_id()
|
/tools/memory-model/Documentation/ |
D | explanation.txt | 29 21. THE PROPAGATES-BEFORE RELATION: pb 1350 THE PROPAGATES-BEFORE RELATION: pb 1353 The propagates-before (pb) relation capitalizes on the special 1380 The existence of a pb link from E to F implies that E must execute 1389 A good example illustrating how pb works is the SB pattern with strong 1412 If r0 = 0 at the end then there is a pb link from P0's load to P1's 1416 Note that this pb link is not included in hb as an instance of prop, 1419 Similarly, if r1 = 0 at the end then there is a pb link from P1's load 1421 cycle in pb, which is not possible since an instruction cannot execute 1425 In summary, the fact that the pb relation links events in the order [all …]
|
/tools/lib/traceevent/ |
D | event-parse.c | 681 const struct printk_map *pb = b; in printk_cmp() local 683 if (pa->addr < pb->addr) in printk_cmp() 685 if (pa->addr > pb->addr) in printk_cmp()
|