/tools/testing/selftests/netfilter/ |
D | nft_meta.sh | 30 counter iifcount {} 31 counter iifnamecount {} 32 counter iifgroupcount {} 33 counter iiftypecount {} 34 counter infproto4count {} 35 counter il4protocounter {} 36 counter imarkcounter {} 37 counter icpu0counter {} 38 counter ilastyearcounter {} 39 counter icurrentyearcounter {} [all …]
|
D | conntrack_icmp_related.sh | 55 cnt=$(ip netns exec $ns nft list counter inet filter "$name" | grep -q "$expect") 58 ip netns exec $ns nft list counter inet filter "$name" 1>&2 134 counter unknown { } 135 counter related { } 138 meta l4proto icmpv6 icmpv6 type "packet-too-big" ct state "related" counter name "related" accept 139 …meta l4proto icmp icmp type "destination-unreachable" ct state "related" counter name "related" ac… 141 counter name "unknown" drop 149 counter unknown { } 150 counter related { } 151 counter redir4 { } [all …]
|
D | conntrack_vrf.sh | 113 iif { veth0, tvrf } counter meta nftrace set 1 114 iif veth0 counter ct zone set 1 counter return 115 iif tvrf counter ct zone set 2 counter return 116 ip protocol icmp counter 117 notrack counter 123 oif veth0 counter ct zone set 1 counter return 124 oif tvrf counter ct zone set 2 counter return 125 notrack counter 167 oif tvrf ct state untracked counter 172 oif tvrf ct state untracked counter [all …]
|
D | nft_nat.sh | 81 local counter=$2 85 echo "ERROR: $counter counter in $ns has unexpected value (expected $expect) at $tag" 1>&2 86 ip netns exec $ns nft list counter inet filter $counter 1>&2 94 cnt=$(ip netns exec $ns nft list counter inet filter ns0in | grep -q "packets 1 bytes 84") 99 cnt=$(ip netns exec $ns nft list counter inet filter ns0out | grep -q "packets 1 bytes 84") 106 cnt=$(ip netns exec $ns nft list counter inet filter ns0in6 | grep -q "$expect") 111 cnt=$(ip netns exec $ns nft list counter inet filter ns0out6 | grep -q "$expect") 125 cnt=$(ip netns exec "$ns0" nft list counter inet filter ns0in | grep -q "packets 0 bytes 0") 131 cnt=$(ip netns exec "$ns0" nft list counter inet filter ns0in6 | grep -q "packets 0 bytes 0") 137 cnt=$(ip netns exec "$ns0" nft list counter inet filter ns0out | grep -q "packets 0 bytes 0") [all …]
|
D | conntrack_tcp_unreplied.sh | 47 cnt=$(ip netns exec $ns2 nft list counter inet filter "$name" | grep -q "$expect") 50 ip netns exec $ns2 nft list counter inet filter "$name" 1>&2 93 counter connreq { } 94 counter redir { } 97 ct state new tcp flags syn ip daddr 10.99.99.99 tcp dport 80 counter name "connreq" accept 98 ct state new ct status dnat tcp dport 8080 counter name "redir" accept
|
/tools/perf/util/ |
D | bpf_counter.c | 40 struct bpf_counter *counter; in bpf_counter_alloc() local 42 counter = zalloc(sizeof(*counter)); in bpf_counter_alloc() 43 if (counter) in bpf_counter_alloc() 44 INIT_LIST_HEAD(&counter->list); in bpf_counter_alloc() 45 return counter; in bpf_counter_alloc() 50 struct bpf_counter *counter, *tmp; in bpf_program_profiler__destroy() local 52 list_for_each_entry_safe(counter, tmp, in bpf_program_profiler__destroy() 54 list_del_init(&counter->list); in bpf_program_profiler__destroy() 55 bpf_prog_profiler_bpf__destroy(counter->skel); in bpf_program_profiler__destroy() 56 free(counter); in bpf_program_profiler__destroy() [all …]
|
D | stat-display.c | 728 static bool is_mixed_hw_group(struct evsel *counter) in is_mixed_hw_group() argument 730 struct evlist *evlist = counter->evlist; in is_mixed_hw_group() 731 u32 pmu_type = counter->core.attr.type; in is_mixed_hw_group() 734 if (counter->core.nr_members < 2) in is_mixed_hw_group() 775 struct evsel *counter = os->evsel; in printout() local 781 os->nfields = 4 + (counter->cgrp ? 1 : 0); in printout() 792 if (run == 0 || ena == 0 || counter->counts->scaled == -1) { in printout() 800 if (counter->supported) { in printout() 801 if (!evlist__has_hybrid(counter->evlist)) { in printout() 803 if (is_mixed_hw_group(counter)) in printout() [all …]
|
D | sideband_evlist.c | 99 struct evsel *counter; in evlist__start_sb_thread() local 110 evlist__for_each_entry(evlist, counter) in evlist__start_sb_thread() 111 evsel__set_sample_id(counter, can_sample_identifier); in evlist__start_sb_thread() 116 evlist__for_each_entry(evlist, counter) { in evlist__start_sb_thread() 117 if (evsel__open(counter, evlist->core.user_requested_cpus, in evlist__start_sb_thread() 125 evlist__for_each_entry(evlist, counter) { in evlist__start_sb_thread() 126 if (evsel__enable(counter)) in evlist__start_sb_thread()
|
D | stat.c | 304 static int check_per_pkg(struct evsel *counter, struct perf_counts_values *vals, in check_per_pkg() argument 307 struct hashmap *mask = counter->per_pkg_mask; in check_per_pkg() 308 struct perf_cpu_map *cpus = evsel__cpus(counter); in check_per_pkg() 315 if (!counter->per_pkg) in check_per_pkg() 326 counter->per_pkg_mask = mask; in check_per_pkg() 462 struct evsel *counter) in process_counter_maps() argument 464 int nthreads = perf_thread_map__nr(counter->core.threads); in process_counter_maps() 465 int ncpus = evsel__nr_cpus(counter); in process_counter_maps() 470 if (process_counter_values(config, counter, idx, thread, in process_counter_maps() 471 perf_counts(counter->counts, idx, thread))) in process_counter_maps() [all …]
|
/tools/testing/selftests/drivers/net/mlxsw/ |
D | qos_lib.sh | 21 local counter=$1; shift # Counter to use for measurement 35 local t0=$(ethtool_stats_get $host_in $counter) 36 local u0=$(ethtool_stats_get $sw_in $counter) 38 local t1=$(ethtool_stats_get $host_in $counter) 39 local u1=$(ethtool_stats_get $sw_in $counter)
|
/tools/arch/x86/include/asm/ |
D | atomic.h | 29 return READ_ONCE((v)->counter); in atomic_read() 41 v->counter = i; in atomic_set() 53 : "+m" (v->counter)); in atomic_inc() 66 GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, "%0", "e"); in atomic_dec_and_test() 71 return cmpxchg(&v->counter, old, new); in atomic_cmpxchg()
|
/tools/include/asm-generic/ |
D | atomic-gcc.h | 26 return READ_ONCE((v)->counter); in atomic_read() 38 v->counter = i; in atomic_set() 49 __sync_add_and_fetch(&v->counter, 1); in atomic_inc() 62 return __sync_sub_and_fetch(&v->counter, 1) == 0; in atomic_dec_and_test() 70 return cmpxchg(&(v)->counter, oldval, newval); in atomic_cmpxchg()
|
/tools/perf/ |
D | design.txt | 13 hardware capabilities. It provides per task and per CPU counters, counter 19 There's one file descriptor per virtual counter used. 29 VFS system calls: read() can be used to read the counter, fcntl() 35 When creating a new counter fd, 'perf_event_attr' is: 40 * specific (raw) counter configuration data, if unset, the next 71 The 'config' field specifies what the counter should count. It 78 If 'raw_type' is 1, then the counter will count a hardware event 82 If 'raw_type' is 0, then the 'type' field says what kind of counter 91 A counter of PERF_TYPE_HARDWARE will count the hardware event 95 * Generalized performance counter event types, used by the hw_event.event_id [all …]
|
D | builtin-stat.c | 268 static int evsel__write_stat_event(struct evsel *counter, int cpu_map_idx, u32 thread, in evsel__write_stat_event() argument 271 struct perf_sample_id *sid = SID(counter, cpu_map_idx, thread); in evsel__write_stat_event() 272 struct perf_cpu cpu = perf_cpu_map__cpu(evsel__cpus(counter), cpu_map_idx); in evsel__write_stat_event() 278 static int read_single_counter(struct evsel *counter, int cpu_map_idx, in read_single_counter() argument 281 switch(counter->tool_event) { in read_single_counter() 285 perf_counts(counter->counts, cpu_map_idx, thread); in read_single_counter() 294 perf_counts(counter->counts, cpu_map_idx, thread); in read_single_counter() 295 if (counter->tool_event == PERF_TOOL_USER_TIME) in read_single_counter() 305 return evsel__read_counter(counter, cpu_map_idx, thread); in read_single_counter() 316 static int read_counter_cpu(struct evsel *counter, struct timespec *rs, int cpu_map_idx) in read_counter_cpu() argument [all …]
|
/tools/testing/selftests/sync/ |
D | sync_stress_parallelism.c | 37 int counter; member 61 ASSERT(test_data_two_threads.counter == i * 2 + thread_id, in test_stress_two_threads_shared_timeline_thread() 63 test_data_two_threads.counter++; in test_stress_two_threads_shared_timeline_thread() 85 test_data_two_threads.counter = 0; in test_stress_two_threads_shared_timeline() 104 ASSERT(test_data_two_threads.counter == in test_stress_two_threads_shared_timeline()
|
/tools/perf/util/bpf_skel/ |
D | bpf_prog_profiler.bpf.c | 60 if (before && before->counter) { in fexit_update_maps() 63 diff.counter = after->counter - before->counter; in fexit_update_maps() 69 accum->counter += diff.counter; in fexit_update_maps()
|
D | bperf_cgroup.bpf.c | 170 val.counter = val.enabled = val.running = 0; in bperf_cgroup_count() 185 delta.counter = val.counter - prev_val->counter; in bperf_cgroup_count() 199 cgrp_val->counter += delta.counter; in bperf_cgroup_count()
|
D | bperf_leader.bpf.c | 48 diff_val->counter = val.counter - prev_val->counter; in BPF_PROG()
|
/tools/bpf/bpftool/skeleton/ |
D | profiler.bpf.c | 8 __u64 counter; member 83 if (before && before->counter) { in fexit_update_maps() 86 diff.counter = after->counter - before->counter; in fexit_update_maps() 92 accum->counter += diff.counter; in fexit_update_maps()
|
/tools/counter/ |
D | Makefile | 28 $(OUTPUT)include/linux/counter.h: ../../include/uapi/linux/counter.h 30 ln -sf $(CURDIR)/../../include/uapi/linux/counter.h $@ 32 prepare: $(OUTPUT)include/linux/counter.h 42 rm -rf $(OUTPUT)include/linux/counter.h
|
/tools/testing/selftests/bpf/progs/ |
D | bpf_iter_netlink.c | 45 s->sk_rmem_alloc.counter, in dump_netlink() 46 s->sk_wmem_alloc.refs.counter - 1, in dump_netlink() 47 nlk->cb_running, s->sk_refcnt.refs.counter); in dump_netlink() 60 BPF_SEQ_PRINTF(seq, "%-8u %-8lu\n", s->sk_drops.counter, ino); in dump_netlink()
|
D | bpf_iter_udp4.c | 55 rqueue = inet->sk.sk_rmem_alloc.counter - udp_sk->forward_deficit; in dump_udp4() 62 inet->sk.sk_wmem_alloc.refs.counter - 1, in dump_udp4() 66 inet->sk.sk_refcnt.refs.counter, udp_sk, in dump_udp4() 67 inet->sk.sk_drops.counter); in dump_udp4()
|
D | sk_storage_omem_uncharge.c | 33 if (sk->sk_cookie.counter != cookie) in BPF_PROG() 37 omem = sk->sk_omem_alloc.counter; in BPF_PROG() 48 if (!cookie || sk->sk_cookie.counter != cookie) in BPF_PROG()
|
D | bpf_iter_udp6.c | 57 rqueue = inet->sk.sk_rmem_alloc.counter - udp_sk->forward_deficit; in dump_udp6() 70 inet->sk.sk_wmem_alloc.refs.counter - 1, in dump_udp6() 74 inet->sk.sk_refcnt.refs.counter, udp_sk, in dump_udp6() 75 inet->sk.sk_drops.counter); in dump_udp6()
|
D | bpf_iter_bpf_map.c | 23 BPF_SEQ_PRINTF(seq, "%8u %8ld %8ld %10lu\n", map->id, map->refcnt.counter, in dump_bpf_map() 24 map->usercnt.counter, in dump_bpf_map()
|