/tools/testing/selftests/bpf/verifier/ |
D | ctx_skb.c | 396 offsetof(struct __sk_buff, cb[0])), 398 offsetof(struct __sk_buff, cb[0]) + 1), 400 offsetof(struct __sk_buff, cb[0]) + 2), 402 offsetof(struct __sk_buff, cb[0]) + 3), 404 offsetof(struct __sk_buff, cb[1])), 406 offsetof(struct __sk_buff, cb[1]) + 1), 408 offsetof(struct __sk_buff, cb[1]) + 2), 410 offsetof(struct __sk_buff, cb[1]) + 3), 412 offsetof(struct __sk_buff, cb[2])), 414 offsetof(struct __sk_buff, cb[2]) + 1), [all …]
|
D | runtime_jit.c | 58 offsetof(struct __sk_buff, cb[0])), 60 offsetof(struct __sk_buff, cb[0])), 80 offsetof(struct __sk_buff, cb[0])), 82 offsetof(struct __sk_buff, cb[0])), 102 offsetof(struct __sk_buff, cb[0])), 104 offsetof(struct __sk_buff, cb[0])), 124 offsetof(struct __sk_buff, cb[0])), 126 offsetof(struct __sk_buff, cb[0])), 146 offsetof(struct __sk_buff, cb[0])), 148 offsetof(struct __sk_buff, cb[0])), [all …]
|
D | leak_ptr.c | 6 offsetof(struct __sk_buff, cb[0])), 9 offsetof(struct __sk_buff, cb[0])), 23 offsetof(struct __sk_buff, cb[0])), 25 offsetof(struct __sk_buff, cb[0])), 39 offsetof(struct __sk_buff, cb[0])),
|
D | cgroup_skb.c | 61 offsetof(struct __sk_buff, cb[0])), 63 offsetof(struct __sk_buff, cb[1])), 65 offsetof(struct __sk_buff, cb[2])), 67 offsetof(struct __sk_buff, cb[3])), 69 offsetof(struct __sk_buff, cb[4])), 73 offsetof(struct __sk_buff, cb[0])), 75 offsetof(struct __sk_buff, cb[1])), 77 offsetof(struct __sk_buff, cb[2])), 79 offsetof(struct __sk_buff, cb[3])), 81 offsetof(struct __sk_buff, cb[4])),
|
/tools/testing/selftests/bpf/prog_tests/ |
D | skb_ctx.c | 8 .cb[0] = 1, in test_skb_ctx() 9 .cb[1] = 2, in test_skb_ctx() 10 .cb[2] = 3, in test_skb_ctx() 11 .cb[3] = 4, in test_skb_ctx() 12 .cb[4] = 5, in test_skb_ctx() 88 CHECK_ATTR(skb.cb[i] != i + 2, in test_skb_ctx() 91 skb.cb[i], i + 2); in test_skb_ctx()
|
D | kfree_skb.c | 14 } cb = { variable 31 if (CHECK(meta->cb8_0 != cb.cb8[0], "check_cb8_0", "cb8_0 %x != %x\n", in on_sample() 32 meta->cb8_0, cb.cb8[0])) in on_sample() 34 if (CHECK(meta->cb32_0 != cb.cb32[0], "check_cb32_0", in on_sample() 36 meta->cb32_0, cb.cb32[0])) in on_sample() 120 memcpy(skb.cb, &cb, sizeof(cb)); in test_kfree_skb()
|
D | perf_branches.c | 61 void (*cb)(struct test_perf_branches *)) in test_perf_branches_common() 93 cb(skel); in test_perf_branches_common()
|
D | fexit_bpf2bpf.c | 58 test_cb cb) in test_fexit_bpf2bpf_common() argument 130 if (cb) { in test_fexit_bpf2bpf_common() 131 err = cb(obj); in test_fexit_bpf2bpf_common()
|
/tools/perf/util/ |
D | sideband_evlist.c | 16 evsel__sb_cb_t cb, void *data) in evlist__add_sb_event() argument 29 evsel->side_band.cb = cb; in evlist__add_sb_event() 67 if (evsel && evsel->side_band.cb) in perf_evlist__poll_thread() 68 evsel->side_band.cb(event, evsel->side_band.data); in perf_evlist__poll_thread() 84 void evlist__set_cb(struct evlist *evlist, evsel__sb_cb_t cb, void *data) in evlist__set_cb() argument 92 evsel->side_band.cb = cb; in evlist__set_cb()
|
D | unwind.h | 24 int (*get_entries)(unwind_entry_cb_t cb, void *arg, 30 int unwind__get_entries(unwind_entry_cb_t cb, void *arg, 64 unwind__get_entries(unwind_entry_cb_t cb __maybe_unused, in unwind__get_entries()
|
D | unwind-libunwind.c | 81 int unwind__get_entries(unwind_entry_cb_t cb, void *arg, in unwind__get_entries() argument 86 return thread->maps->unwind_libunwind_ops->get_entries(cb, arg, thread, data, max_stack); in unwind__get_entries()
|
D | unwind-libdw.h | 19 unwind_entry_cb_t cb; member
|
D | color_config.c | 40 void *cb __maybe_unused) in perf_color_default_config()
|
D | unwind-libdw.c | 222 int unwind__get_entries(unwind_entry_cb_t cb, void *arg, in unwind__get_entries() argument 231 .cb = cb, in unwind__get_entries() 277 err = ui->entries[j].ip ? ui->cb(&ui->entries[j], ui->arg) : 0; in unwind__get_entries()
|
D | unwind-libunwind-local.c | 573 unwind_entry_cb_t cb, void *arg) in entry() argument 588 return cb(&e, arg); in entry() 641 static int get_entries(struct unwind_info *ui, unwind_entry_cb_t cb, in get_entries() argument 699 ret = ips[j] ? entry(ips[j], ui->thread, cb, arg) : 0; in get_entries() 705 static int _unwind__get_entries(unwind_entry_cb_t cb, void *arg, in _unwind__get_entries() argument 721 return get_entries(&ui, cb, arg, max_stack); in _unwind__get_entries()
|
D | color.h | 34 int perf_color_default_config(const char *var, const char *value, void *cb);
|
/tools/testing/selftests/bpf/progs/ |
D | test_skb_ctx.c | 14 if (skb->cb[i] != i + 1) in process() 16 skb->cb[i]++; in process()
|
D | kfree_skb.c | 48 char cb[48]; member 80 cb8 = (__u8 *)&skb->cb; in BPF_PROG() 81 cb32 = (__u32 *)&skb->cb; in BPF_PROG()
|
D | btf_dump_test_case_ordering.c | 55 struct callback_head cb; member
|
D | netif_receive_skb.c | 234 { .cb = {1, 2, 3, 4, 5,},}); in BPF_PROG() 237 { .cb = { 1, 2, 3, 4, 5},}); in BPF_PROG() 241 { .cb = { 0, 0, 1, 0, 0},}); in BPF_PROG()
|
D | test_verif_scale3.c | 24 ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \ in balancer_ingress()
|
D | test_verif_scale1.c | 24 ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \ in balancer_ingress()
|
D | test_verif_scale2.c | 24 ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \ in balancer_ingress()
|
/tools/perf/jvmti/ |
D | libjvmti.c | 340 jvmtiEventCallbacks cb; in Agent_OnLoad() local 385 memset(&cb, 0, sizeof(cb)); in Agent_OnLoad() 387 cb.CompiledMethodLoad = compiled_method_load_cb; in Agent_OnLoad() 388 cb.DynamicCodeGenerated = code_generated_cb; in Agent_OnLoad() 390 ret = (*jvmti)->SetEventCallbacks(jvmti, &cb, sizeof(cb)); in Agent_OnLoad()
|
/tools/build/feature/ |
D | test-jvmti.c | 7 jvmtiEventCallbacks cb __attribute__((unused)); in main() local
|