Home
last modified time | relevance | path

Searched refs:stats (Results 1 – 25 of 142) sorted by relevance

123456

/net/wireless/
Dwext-proc.c33 struct iw_statistics *stats = get_wireless_stats(dev); in wireless_seq_printf_stats() local
37 if (!stats) { in wireless_seq_printf_stats()
40 stats = &nullstats; in wireless_seq_printf_stats()
44 stats = &nullstats; in wireless_seq_printf_stats()
48 if (stats) { in wireless_seq_printf_stats()
51 dev->name, stats->status, stats->qual.qual, in wireless_seq_printf_stats()
52 stats->qual.updated & IW_QUAL_QUAL_UPDATED in wireless_seq_printf_stats()
54 ((__s32) stats->qual.level) - in wireless_seq_printf_stats()
55 ((stats->qual.updated & IW_QUAL_DBM) ? 0x100 : 0), in wireless_seq_printf_stats()
56 stats->qual.updated & IW_QUAL_LEVEL_UPDATED in wireless_seq_printf_stats()
[all …]
/net/sunrpc/
Dstats.c126 struct rpc_iostats *stats; in rpc_alloc_iostats() local
129 stats = kcalloc(clnt->cl_maxproc, sizeof(*stats), GFP_KERNEL); in rpc_alloc_iostats()
130 if (stats) { in rpc_alloc_iostats()
132 spin_lock_init(&stats[i].om_lock); in rpc_alloc_iostats()
134 return stats; in rpc_alloc_iostats()
143 void rpc_free_iostats(struct rpc_iostats *stats) in rpc_free_iostats() argument
145 kfree(stats); in rpc_free_iostats()
200 void rpc_count_iostats(const struct rpc_task *task, struct rpc_iostats *stats) in rpc_count_iostats() argument
203 &stats[task->tk_msg.rpc_proc->p_statidx]); in rpc_count_iostats()
231 static void _print_rpc_iostats(struct seq_file *seq, struct rpc_iostats *stats, in _print_rpc_iostats() argument
[all …]
/net/smc/
Dsmc_stats.c46 struct smc_stats *stats, int tech, in smc_nl_fill_stats_rmb_data() argument
53 stats_rmb_cnt = &stats->smc[tech].rmb_tx; in smc_nl_fill_stats_rmb_data()
55 stats_rmb_cnt = &stats->smc[tech].rmb_rx; in smc_nl_fill_stats_rmb_data()
99 struct smc_stats *stats, int tech, in smc_nl_fill_stats_bufsize_data() argument
106 stats_pload = &stats->smc[tech].tx_pd; in smc_nl_fill_stats_bufsize_data()
108 stats_pload = &stats->smc[tech].rx_pd; in smc_nl_fill_stats_bufsize_data()
110 stats_pload = &stats->smc[tech].tx_rmbsize; in smc_nl_fill_stats_bufsize_data()
112 stats_pload = &stats->smc[tech].rx_rmbsize; in smc_nl_fill_stats_bufsize_data()
166 struct smc_stats *stats, int tech) in smc_nl_fill_stats_tech_data() argument
171 smc_tech = &stats->smc[tech]; in smc_nl_fill_stats_tech_data()
[all …]
/net/ethtool/
Dfec.c17 u64 stats[1 + ETHTOOL_MAX_LANES]; member
72 fec_stats_recalc(struct fec_stat_grp *grp, struct ethtool_fec_stat *stats) in fec_stats_recalc() argument
76 if (stats->lanes[0] == ETHTOOL_STAT_NOT_SET) { in fec_stats_recalc()
77 grp->stats[0] = stats->total; in fec_stats_recalc()
78 grp->cnt = stats->total != ETHTOOL_STAT_NOT_SET; in fec_stats_recalc()
83 grp->stats[0] = 0; in fec_stats_recalc()
85 if (stats->lanes[i] == ETHTOOL_STAT_NOT_SET) in fec_stats_recalc()
88 grp->stats[0] += stats->lanes[i]; in fec_stats_recalc()
89 grp->stats[grp->cnt++] = stats->lanes[i]; in fec_stats_recalc()
113 struct ethtool_fec_stats stats; in fec_prepare_data() local
[all …]
Dmm.c15 struct ethtool_mm_stats stats; member
42 ethtool_stats_init((u64 *)&data->stats, in mm_prepare_data()
43 sizeof(data->stats) / sizeof(u64)); in mm_prepare_data()
54 ops->get_mm_stats(dev, &data->stats); in mm_prepare_data()
94 const struct ethtool_mm_stats *stats) in mm_put_stats() argument
102 if (mm_put_stat(skb, stats->MACMergeFrameAssErrorCount, in mm_put_stats()
104 mm_put_stat(skb, stats->MACMergeFrameSmdErrorCount, in mm_put_stats()
106 mm_put_stat(skb, stats->MACMergeFrameAssOkCount, in mm_put_stats()
108 mm_put_stat(skb, stats->MACMergeFragCountRx, in mm_put_stats()
110 mm_put_stat(skb, stats->MACMergeFragCountTx, in mm_put_stats()
[all …]
/net/core/
Dnet-procfs.c80 const struct rtnl_link_stats64 *stats = dev_get_stats(dev, &temp); in dev_seq_printf_stats() local
84 dev->name, stats->rx_bytes, stats->rx_packets, in dev_seq_printf_stats()
85 stats->rx_errors, in dev_seq_printf_stats()
86 stats->rx_dropped + stats->rx_missed_errors, in dev_seq_printf_stats()
87 stats->rx_fifo_errors, in dev_seq_printf_stats()
88 stats->rx_length_errors + stats->rx_over_errors + in dev_seq_printf_stats()
89 stats->rx_crc_errors + stats->rx_frame_errors, in dev_seq_printf_stats()
90 stats->rx_compressed, stats->multicast, in dev_seq_printf_stats()
91 stats->tx_bytes, stats->tx_packets, in dev_seq_printf_stats()
92 stats->tx_errors, stats->tx_dropped, in dev_seq_printf_stats()
[all …]
/net/mpls/
Dinternal.h19 struct mpls_link_stats stats; member
26 struct mpls_pcpu_stats __percpu *stats; member
36 __typeof__(*(mdev)->stats) *ptr = \
37 raw_cpu_ptr((mdev)->stats); \
40 ptr->stats.pkts_field++; \
41 ptr->stats.bytes_field += (len); \
48 __typeof__(*(mdev)->stats) *ptr = \
49 raw_cpu_ptr((mdev)->stats); \
52 ptr->stats.field++; \
61 this_cpu_inc((mdev)->stats->stats.pkts_field); \
[all …]
/net/tipc/
Dnetlink_compat.c503 struct nlattr *prop[], struct nlattr *stats[]) in __fill_bc_link_stat() argument
510 nla_get_u32(stats[TIPC_NLA_STATS_RX_INFO]), in __fill_bc_link_stat()
511 nla_get_u32(stats[TIPC_NLA_STATS_RX_FRAGMENTS]), in __fill_bc_link_stat()
512 nla_get_u32(stats[TIPC_NLA_STATS_RX_FRAGMENTED]), in __fill_bc_link_stat()
513 nla_get_u32(stats[TIPC_NLA_STATS_RX_BUNDLES]), in __fill_bc_link_stat()
514 nla_get_u32(stats[TIPC_NLA_STATS_RX_BUNDLED])); in __fill_bc_link_stat()
518 nla_get_u32(stats[TIPC_NLA_STATS_TX_INFO]), in __fill_bc_link_stat()
519 nla_get_u32(stats[TIPC_NLA_STATS_TX_FRAGMENTS]), in __fill_bc_link_stat()
520 nla_get_u32(stats[TIPC_NLA_STATS_TX_FRAGMENTED]), in __fill_bc_link_stat()
521 nla_get_u32(stats[TIPC_NLA_STATS_TX_BUNDLES]), in __fill_bc_link_stat()
[all …]
Dlink.c227 struct tipc_stats stats; member
795 l->stats.accu_queue_sz += skb_queue_len(&l->transmq); in link_profile_stats()
796 l->stats.queue_sz_counts++; in link_profile_stats()
809 l->stats.msg_lengths_total += length; in link_profile_stats()
810 l->stats.msg_length_counts++; in link_profile_stats()
812 l->stats.msg_length_profile[0]++; in link_profile_stats()
814 l->stats.msg_length_profile[1]++; in link_profile_stats()
816 l->stats.msg_length_profile[2]++; in link_profile_stats()
818 l->stats.msg_length_profile[3]++; in link_profile_stats()
820 l->stats.msg_length_profile[4]++; in link_profile_stats()
[all …]
/net/netrom/
Dnr_dev.c42 struct net_device_stats *stats = &dev->stats; in nr_rx_ip() local
45 stats->rx_dropped++; in nr_rx_ip()
49 stats->rx_packets++; in nr_rx_ip()
50 stats->rx_bytes += skb->len; in nr_rx_ip()
141 struct net_device_stats *stats = &dev->stats; in nr_xmit() local
146 stats->tx_errors++; in nr_xmit()
150 stats->tx_packets++; in nr_xmit()
151 stats->tx_bytes += len; in nr_xmit()
/net/kcm/
Dkcmproc.c115 kcm->stats.rx_msgs, in kcm_format_sock()
116 kcm->stats.rx_bytes, in kcm_format_sock()
117 kcm->stats.tx_msgs, in kcm_format_sock()
118 kcm->stats.tx_bytes, in kcm_format_sock()
145 psock->strp.stats.msgs, in kcm_format_psock()
146 psock->strp.stats.bytes, in kcm_format_psock()
147 psock->stats.tx_msgs, in kcm_format_psock()
148 psock->stats.tx_bytes, in kcm_format_psock()
196 mux->stats.rx_msgs, in kcm_format_mux()
197 mux->stats.rx_bytes, in kcm_format_mux()
[all …]
/net/openvswitch/
Dvport.c284 void ovs_vport_get_stats(struct vport *vport, struct ovs_vport_stats *stats) in ovs_vport_get_stats() argument
290 stats->rx_errors = dev_stats->rx_errors; in ovs_vport_get_stats()
291 stats->tx_errors = dev_stats->tx_errors; in ovs_vport_get_stats()
292 stats->tx_dropped = dev_stats->tx_dropped; in ovs_vport_get_stats()
293 stats->rx_dropped = dev_stats->rx_dropped; in ovs_vport_get_stats()
295 stats->rx_bytes = dev_stats->rx_bytes; in ovs_vport_get_stats()
296 stats->rx_packets = dev_stats->rx_packets; in ovs_vport_get_stats()
297 stats->tx_bytes = dev_stats->tx_bytes; in ovs_vport_get_stats()
298 stats->tx_packets = dev_stats->tx_packets; in ovs_vport_get_stats()
320 const struct vport_upcall_stats_percpu *stats; in ovs_vport_get_upcall_stats() local
[all …]
Dflow.c64 struct sw_flow_stats *stats; in ovs_flow_stats_update() local
68 stats = rcu_dereference(flow->stats[cpu]); in ovs_flow_stats_update()
71 if (likely(stats)) { in ovs_flow_stats_update()
72 spin_lock(&stats->lock); in ovs_flow_stats_update()
77 stats = rcu_dereference(flow->stats[0]); /* Pre-allocated. */ in ovs_flow_stats_update()
78 spin_lock(&stats->lock); in ovs_flow_stats_update()
90 likely(!rcu_access_pointer(flow->stats[cpu]))) { in ovs_flow_stats_update()
108 rcu_assign_pointer(flow->stats[cpu], in ovs_flow_stats_update()
119 stats->used = jiffies; in ovs_flow_stats_update()
120 stats->packet_count++; in ovs_flow_stats_update()
[all …]
Dflow_table.c75 struct sw_flow_stats *stats; in ovs_flow_alloc() local
82 flow->cpu_used_mask = (struct cpumask *)&flow->stats[nr_cpu_ids]; in ovs_flow_alloc()
85 stats = kmem_cache_alloc_node(flow_stats_cache, in ovs_flow_alloc()
88 if (!stats) in ovs_flow_alloc()
91 spin_lock_init(&stats->lock); in ovs_flow_alloc()
93 RCU_INIT_POINTER(flow->stats[0], stats); in ovs_flow_alloc()
120 if (flow->stats[cpu]) in flow_free()
122 (struct sw_flow_stats __force *)flow->stats[cpu]); in flow_free()
203 struct mask_array_stats *stats; in tbl_mask_array_reset_counters() local
207 stats = per_cpu_ptr(ma->masks_usage_stats, cpu); in tbl_mask_array_reset_counters()
[all …]
/net/l2tp/
Dl2tp_ppp.c243 atomic_long_inc(&session->stats.rx_errors); in pppol2tp_recv()
1021 const struct l2tp_stats *stats) in pppol2tp_copy_stats() argument
1025 dest->tx_packets = atomic_long_read(&stats->tx_packets); in pppol2tp_copy_stats()
1026 dest->tx_bytes = atomic_long_read(&stats->tx_bytes); in pppol2tp_copy_stats()
1027 dest->tx_errors = atomic_long_read(&stats->tx_errors); in pppol2tp_copy_stats()
1028 dest->rx_packets = atomic_long_read(&stats->rx_packets); in pppol2tp_copy_stats()
1029 dest->rx_bytes = atomic_long_read(&stats->rx_bytes); in pppol2tp_copy_stats()
1030 dest->rx_seq_discards = atomic_long_read(&stats->rx_seq_discards); in pppol2tp_copy_stats()
1031 dest->rx_oos_packets = atomic_long_read(&stats->rx_oos_packets); in pppol2tp_copy_stats()
1032 dest->rx_errors = atomic_long_read(&stats->rx_errors); in pppol2tp_copy_stats()
[all …]
Dl2tp_debugfs.c169 atomic_long_read(&tunnel->stats.tx_packets), in l2tp_dfs_seq_tunnel_show()
170 atomic_long_read(&tunnel->stats.tx_bytes), in l2tp_dfs_seq_tunnel_show()
171 atomic_long_read(&tunnel->stats.tx_errors), in l2tp_dfs_seq_tunnel_show()
172 atomic_long_read(&tunnel->stats.rx_packets), in l2tp_dfs_seq_tunnel_show()
173 atomic_long_read(&tunnel->stats.rx_bytes), in l2tp_dfs_seq_tunnel_show()
174 atomic_long_read(&tunnel->stats.rx_errors)); in l2tp_dfs_seq_tunnel_show()
220 atomic_long_read(&session->stats.tx_packets), in l2tp_dfs_seq_session_show()
221 atomic_long_read(&session->stats.tx_bytes), in l2tp_dfs_seq_session_show()
222 atomic_long_read(&session->stats.tx_errors), in l2tp_dfs_seq_session_show()
223 atomic_long_read(&session->stats.rx_packets), in l2tp_dfs_seq_session_show()
[all …]
/net/sched/
Dsch_codel.c61 struct codel_stats stats; member
95 &q->stats, qdisc_pkt_len, codel_get_enqueue_time, in codel_qdisc_dequeue()
101 if (q->stats.drop_count && sch->q.qlen) { in codel_qdisc_dequeue()
102 qdisc_tree_reduce_backlog(sch, q->stats.drop_count, q->stats.drop_len); in codel_qdisc_dequeue()
103 q->stats.drop_count = 0; in codel_qdisc_dequeue()
104 q->stats.drop_len = 0; in codel_qdisc_dequeue()
195 codel_stats_init(&q->stats); in codel_init()
246 .maxpacket = q->stats.maxpacket, in codel_dump_stats()
252 .ecn_mark = q->stats.ecn_mark, in codel_dump_stats()
253 .ce_mark = q->stats.ce_mark, in codel_dump_stats()
Dsch_gred.c44 struct red_stats stats; member
229 q->stats.prob_drop++; in gred_enqueue()
233 q->stats.prob_mark++; in gred_enqueue()
240 q->stats.forced_drop++; in gred_enqueue()
243 q->stats.forced_mark++; in gred_enqueue()
252 q->stats.pdrop++; in gred_enqueue()
371 gnet_stats_basic_sync_init(&hw_stats->stats.bstats[i]); in gred_offload_dump_stats()
373 hw_stats->stats.xstats[i] = &table->tab[i]->stats; in gred_offload_dump_stats()
384 table->tab[i]->packetsin += u64_stats_read(&hw_stats->stats.bstats[i].packets); in gred_offload_dump_stats()
385 table->tab[i]->bytesin += u64_stats_read(&hw_stats->stats.bstats[i].bytes); in gred_offload_dump_stats()
[all …]
Dsch_red.c47 struct red_stats stats; member
92 q->stats.prob_drop++; in red_enqueue()
97 q->stats.prob_mark++; in red_enqueue()
102 q->stats.prob_drop++; in red_enqueue()
112 q->stats.forced_drop++; in red_enqueue()
117 q->stats.forced_mark++; in red_enqueue()
122 q->stats.forced_drop++; in red_enqueue()
136 q->stats.pdrop++; in red_enqueue()
401 .stats.bstats = &sch->bstats, in red_dump_offload_stats()
402 .stats.qstats = &sch->qstats, in red_dump_offload_stats()
[all …]
Dsch_choke.c64 } stats; member
232 q->stats.matched++; in choke_enqueue()
244 q->stats.forced_drop++; in choke_enqueue()
248 q->stats.forced_mark++; in choke_enqueue()
256 q->stats.prob_drop++; in choke_enqueue()
260 q->stats.prob_mark++; in choke_enqueue()
275 q->stats.pdrop++; in choke_enqueue()
463 .early = q->stats.prob_drop + q->stats.forced_drop, in choke_dump_stats()
464 .marked = q->stats.prob_mark + q->stats.forced_mark, in choke_dump_stats()
465 .pdrop = q->stats.pdrop, in choke_dump_stats()
[all …]
/net/strparser/
Dstrparser.c115 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv()
120 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv()
135 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv()
154 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv()
173 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv()
196 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv()
223 STRP_STATS_INCR(strp->stats.need_more_hdr); in __strp_recv()
237 STRP_STATS_INCR(strp->stats.msg_too_big); in __strp_recv()
245 STRP_STATS_INCR(strp->stats.bad_hdr_len); in __strp_recv()
276 STRP_STATS_ADD(strp->stats.bytes, cand_len); in __strp_recv()
[all …]
/net/devlink/
Dtrap.c48 struct devlink_stats __percpu *stats; member
68 struct devlink_stats __percpu *stats; member
157 struct devlink_stats *stats) in devlink_trap_stats_read() argument
161 memset(stats, 0, sizeof(*stats)); in devlink_trap_stats_read()
174 u64_stats_add(&stats->rx_packets, rx_packets); in devlink_trap_stats_read()
175 u64_stats_add(&stats->rx_bytes, rx_bytes); in devlink_trap_stats_read()
183 struct devlink_stats stats; in devlink_trap_group_stats_put() local
186 devlink_trap_stats_read(trap_stats, &stats); in devlink_trap_group_stats_put()
193 u64_stats_read(&stats.rx_packets), in devlink_trap_group_stats_put()
198 u64_stats_read(&stats.rx_bytes), in devlink_trap_group_stats_put()
[all …]
/net/bridge/
Dbr_vlan.c193 free_percpu(v->stats); in br_master_vlan_rcu_free()
194 v->stats = NULL; in br_master_vlan_rcu_free()
224 free_percpu(v->stats); in nbp_vlan_rcu_free()
225 v->stats = NULL; in nbp_vlan_rcu_free()
310 v->stats = in __vlan_add()
312 if (!v->stats) { in __vlan_add()
318 v->stats = masterv->stats; in __vlan_add()
369 if (v->stats && masterv->stats != v->stats) in __vlan_add()
370 free_percpu(v->stats); in __vlan_add()
371 v->stats = NULL; in __vlan_add()
[all …]
/net/rds/
Dtcp_stats.c54 struct rds_tcp_statistics stats = {0, }; in rds_tcp_stats_info_copy() local
65 sum = (uint64_t *)&stats; in rds_tcp_stats_info_copy()
66 for (i = 0; i < sizeof(stats) / sizeof(uint64_t); i++) in rds_tcp_stats_info_copy()
70 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_tcp_stat_names, in rds_tcp_stats_info_copy()
Dib_stats.c87 struct rds_ib_statistics stats = {0, }; in rds_ib_stats_info_copy() local
98 sum = (uint64_t *)&stats; in rds_ib_stats_info_copy()
99 for (i = 0; i < sizeof(stats) / sizeof(uint64_t); i++) in rds_ib_stats_info_copy()
103 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_ib_stat_names, in rds_ib_stats_info_copy()

123456