/net/wireless/ |
D | wext-proc.c | 33 struct iw_statistics *stats = get_wireless_stats(dev); in wireless_seq_printf_stats() local 37 if (!stats) { in wireless_seq_printf_stats() 40 stats = &nullstats; in wireless_seq_printf_stats() 44 stats = &nullstats; in wireless_seq_printf_stats() 48 if (stats) { in wireless_seq_printf_stats() 51 dev->name, stats->status, stats->qual.qual, in wireless_seq_printf_stats() 52 stats->qual.updated & IW_QUAL_QUAL_UPDATED in wireless_seq_printf_stats() 54 ((__s32) stats->qual.level) - in wireless_seq_printf_stats() 55 ((stats->qual.updated & IW_QUAL_DBM) ? 0x100 : 0), in wireless_seq_printf_stats() 56 stats->qual.updated & IW_QUAL_LEVEL_UPDATED in wireless_seq_printf_stats() [all …]
|
/net/sunrpc/ |
D | stats.c | 126 struct rpc_iostats *stats; in rpc_alloc_iostats() local 129 stats = kcalloc(clnt->cl_maxproc, sizeof(*stats), GFP_KERNEL); in rpc_alloc_iostats() 130 if (stats) { in rpc_alloc_iostats() 132 spin_lock_init(&stats[i].om_lock); in rpc_alloc_iostats() 134 return stats; in rpc_alloc_iostats() 143 void rpc_free_iostats(struct rpc_iostats *stats) in rpc_free_iostats() argument 145 kfree(stats); in rpc_free_iostats() 200 void rpc_count_iostats(const struct rpc_task *task, struct rpc_iostats *stats) in rpc_count_iostats() argument 203 &stats[task->tk_msg.rpc_proc->p_statidx]); in rpc_count_iostats() 231 static void _print_rpc_iostats(struct seq_file *seq, struct rpc_iostats *stats, in _print_rpc_iostats() argument [all …]
|
/net/smc/ |
D | smc_stats.c | 46 struct smc_stats *stats, int tech, in smc_nl_fill_stats_rmb_data() argument 53 stats_rmb_cnt = &stats->smc[tech].rmb_tx; in smc_nl_fill_stats_rmb_data() 55 stats_rmb_cnt = &stats->smc[tech].rmb_rx; in smc_nl_fill_stats_rmb_data() 99 struct smc_stats *stats, int tech, in smc_nl_fill_stats_bufsize_data() argument 106 stats_pload = &stats->smc[tech].tx_pd; in smc_nl_fill_stats_bufsize_data() 108 stats_pload = &stats->smc[tech].rx_pd; in smc_nl_fill_stats_bufsize_data() 110 stats_pload = &stats->smc[tech].tx_rmbsize; in smc_nl_fill_stats_bufsize_data() 112 stats_pload = &stats->smc[tech].rx_rmbsize; in smc_nl_fill_stats_bufsize_data() 166 struct smc_stats *stats, int tech) in smc_nl_fill_stats_tech_data() argument 171 smc_tech = &stats->smc[tech]; in smc_nl_fill_stats_tech_data() [all …]
|
/net/ethtool/ |
D | fec.c | 17 u64 stats[1 + ETHTOOL_MAX_LANES]; member 72 fec_stats_recalc(struct fec_stat_grp *grp, struct ethtool_fec_stat *stats) in fec_stats_recalc() argument 76 if (stats->lanes[0] == ETHTOOL_STAT_NOT_SET) { in fec_stats_recalc() 77 grp->stats[0] = stats->total; in fec_stats_recalc() 78 grp->cnt = stats->total != ETHTOOL_STAT_NOT_SET; in fec_stats_recalc() 83 grp->stats[0] = 0; in fec_stats_recalc() 85 if (stats->lanes[i] == ETHTOOL_STAT_NOT_SET) in fec_stats_recalc() 88 grp->stats[0] += stats->lanes[i]; in fec_stats_recalc() 89 grp->stats[grp->cnt++] = stats->lanes[i]; in fec_stats_recalc() 113 struct ethtool_fec_stats stats; in fec_prepare_data() local [all …]
|
D | mm.c | 15 struct ethtool_mm_stats stats; member 42 ethtool_stats_init((u64 *)&data->stats, in mm_prepare_data() 43 sizeof(data->stats) / sizeof(u64)); in mm_prepare_data() 54 ops->get_mm_stats(dev, &data->stats); in mm_prepare_data() 94 const struct ethtool_mm_stats *stats) in mm_put_stats() argument 102 if (mm_put_stat(skb, stats->MACMergeFrameAssErrorCount, in mm_put_stats() 104 mm_put_stat(skb, stats->MACMergeFrameSmdErrorCount, in mm_put_stats() 106 mm_put_stat(skb, stats->MACMergeFrameAssOkCount, in mm_put_stats() 108 mm_put_stat(skb, stats->MACMergeFragCountRx, in mm_put_stats() 110 mm_put_stat(skb, stats->MACMergeFragCountTx, in mm_put_stats() [all …]
|
/net/core/ |
D | net-procfs.c | 80 const struct rtnl_link_stats64 *stats = dev_get_stats(dev, &temp); in dev_seq_printf_stats() local 84 dev->name, stats->rx_bytes, stats->rx_packets, in dev_seq_printf_stats() 85 stats->rx_errors, in dev_seq_printf_stats() 86 stats->rx_dropped + stats->rx_missed_errors, in dev_seq_printf_stats() 87 stats->rx_fifo_errors, in dev_seq_printf_stats() 88 stats->rx_length_errors + stats->rx_over_errors + in dev_seq_printf_stats() 89 stats->rx_crc_errors + stats->rx_frame_errors, in dev_seq_printf_stats() 90 stats->rx_compressed, stats->multicast, in dev_seq_printf_stats() 91 stats->tx_bytes, stats->tx_packets, in dev_seq_printf_stats() 92 stats->tx_errors, stats->tx_dropped, in dev_seq_printf_stats() [all …]
|
/net/mpls/ |
D | internal.h | 19 struct mpls_link_stats stats; member 26 struct mpls_pcpu_stats __percpu *stats; member 36 __typeof__(*(mdev)->stats) *ptr = \ 37 raw_cpu_ptr((mdev)->stats); \ 40 ptr->stats.pkts_field++; \ 41 ptr->stats.bytes_field += (len); \ 48 __typeof__(*(mdev)->stats) *ptr = \ 49 raw_cpu_ptr((mdev)->stats); \ 52 ptr->stats.field++; \ 61 this_cpu_inc((mdev)->stats->stats.pkts_field); \ [all …]
|
/net/tipc/ |
D | netlink_compat.c | 503 struct nlattr *prop[], struct nlattr *stats[]) in __fill_bc_link_stat() argument 510 nla_get_u32(stats[TIPC_NLA_STATS_RX_INFO]), in __fill_bc_link_stat() 511 nla_get_u32(stats[TIPC_NLA_STATS_RX_FRAGMENTS]), in __fill_bc_link_stat() 512 nla_get_u32(stats[TIPC_NLA_STATS_RX_FRAGMENTED]), in __fill_bc_link_stat() 513 nla_get_u32(stats[TIPC_NLA_STATS_RX_BUNDLES]), in __fill_bc_link_stat() 514 nla_get_u32(stats[TIPC_NLA_STATS_RX_BUNDLED])); in __fill_bc_link_stat() 518 nla_get_u32(stats[TIPC_NLA_STATS_TX_INFO]), in __fill_bc_link_stat() 519 nla_get_u32(stats[TIPC_NLA_STATS_TX_FRAGMENTS]), in __fill_bc_link_stat() 520 nla_get_u32(stats[TIPC_NLA_STATS_TX_FRAGMENTED]), in __fill_bc_link_stat() 521 nla_get_u32(stats[TIPC_NLA_STATS_TX_BUNDLES]), in __fill_bc_link_stat() [all …]
|
D | link.c | 227 struct tipc_stats stats; member 795 l->stats.accu_queue_sz += skb_queue_len(&l->transmq); in link_profile_stats() 796 l->stats.queue_sz_counts++; in link_profile_stats() 809 l->stats.msg_lengths_total += length; in link_profile_stats() 810 l->stats.msg_length_counts++; in link_profile_stats() 812 l->stats.msg_length_profile[0]++; in link_profile_stats() 814 l->stats.msg_length_profile[1]++; in link_profile_stats() 816 l->stats.msg_length_profile[2]++; in link_profile_stats() 818 l->stats.msg_length_profile[3]++; in link_profile_stats() 820 l->stats.msg_length_profile[4]++; in link_profile_stats() [all …]
|
/net/netrom/ |
D | nr_dev.c | 42 struct net_device_stats *stats = &dev->stats; in nr_rx_ip() local 45 stats->rx_dropped++; in nr_rx_ip() 49 stats->rx_packets++; in nr_rx_ip() 50 stats->rx_bytes += skb->len; in nr_rx_ip() 141 struct net_device_stats *stats = &dev->stats; in nr_xmit() local 146 stats->tx_errors++; in nr_xmit() 150 stats->tx_packets++; in nr_xmit() 151 stats->tx_bytes += len; in nr_xmit()
|
/net/kcm/ |
D | kcmproc.c | 115 kcm->stats.rx_msgs, in kcm_format_sock() 116 kcm->stats.rx_bytes, in kcm_format_sock() 117 kcm->stats.tx_msgs, in kcm_format_sock() 118 kcm->stats.tx_bytes, in kcm_format_sock() 145 psock->strp.stats.msgs, in kcm_format_psock() 146 psock->strp.stats.bytes, in kcm_format_psock() 147 psock->stats.tx_msgs, in kcm_format_psock() 148 psock->stats.tx_bytes, in kcm_format_psock() 196 mux->stats.rx_msgs, in kcm_format_mux() 197 mux->stats.rx_bytes, in kcm_format_mux() [all …]
|
/net/openvswitch/ |
D | vport.c | 284 void ovs_vport_get_stats(struct vport *vport, struct ovs_vport_stats *stats) in ovs_vport_get_stats() argument 290 stats->rx_errors = dev_stats->rx_errors; in ovs_vport_get_stats() 291 stats->tx_errors = dev_stats->tx_errors; in ovs_vport_get_stats() 292 stats->tx_dropped = dev_stats->tx_dropped; in ovs_vport_get_stats() 293 stats->rx_dropped = dev_stats->rx_dropped; in ovs_vport_get_stats() 295 stats->rx_bytes = dev_stats->rx_bytes; in ovs_vport_get_stats() 296 stats->rx_packets = dev_stats->rx_packets; in ovs_vport_get_stats() 297 stats->tx_bytes = dev_stats->tx_bytes; in ovs_vport_get_stats() 298 stats->tx_packets = dev_stats->tx_packets; in ovs_vport_get_stats() 320 const struct vport_upcall_stats_percpu *stats; in ovs_vport_get_upcall_stats() local [all …]
|
D | flow.c | 64 struct sw_flow_stats *stats; in ovs_flow_stats_update() local 68 stats = rcu_dereference(flow->stats[cpu]); in ovs_flow_stats_update() 71 if (likely(stats)) { in ovs_flow_stats_update() 72 spin_lock(&stats->lock); in ovs_flow_stats_update() 77 stats = rcu_dereference(flow->stats[0]); /* Pre-allocated. */ in ovs_flow_stats_update() 78 spin_lock(&stats->lock); in ovs_flow_stats_update() 90 likely(!rcu_access_pointer(flow->stats[cpu]))) { in ovs_flow_stats_update() 108 rcu_assign_pointer(flow->stats[cpu], in ovs_flow_stats_update() 119 stats->used = jiffies; in ovs_flow_stats_update() 120 stats->packet_count++; in ovs_flow_stats_update() [all …]
|
D | flow_table.c | 75 struct sw_flow_stats *stats; in ovs_flow_alloc() local 82 flow->cpu_used_mask = (struct cpumask *)&flow->stats[nr_cpu_ids]; in ovs_flow_alloc() 85 stats = kmem_cache_alloc_node(flow_stats_cache, in ovs_flow_alloc() 88 if (!stats) in ovs_flow_alloc() 91 spin_lock_init(&stats->lock); in ovs_flow_alloc() 93 RCU_INIT_POINTER(flow->stats[0], stats); in ovs_flow_alloc() 120 if (flow->stats[cpu]) in flow_free() 122 (struct sw_flow_stats __force *)flow->stats[cpu]); in flow_free() 203 struct mask_array_stats *stats; in tbl_mask_array_reset_counters() local 207 stats = per_cpu_ptr(ma->masks_usage_stats, cpu); in tbl_mask_array_reset_counters() [all …]
|
/net/l2tp/ |
D | l2tp_ppp.c | 243 atomic_long_inc(&session->stats.rx_errors); in pppol2tp_recv() 1021 const struct l2tp_stats *stats) in pppol2tp_copy_stats() argument 1025 dest->tx_packets = atomic_long_read(&stats->tx_packets); in pppol2tp_copy_stats() 1026 dest->tx_bytes = atomic_long_read(&stats->tx_bytes); in pppol2tp_copy_stats() 1027 dest->tx_errors = atomic_long_read(&stats->tx_errors); in pppol2tp_copy_stats() 1028 dest->rx_packets = atomic_long_read(&stats->rx_packets); in pppol2tp_copy_stats() 1029 dest->rx_bytes = atomic_long_read(&stats->rx_bytes); in pppol2tp_copy_stats() 1030 dest->rx_seq_discards = atomic_long_read(&stats->rx_seq_discards); in pppol2tp_copy_stats() 1031 dest->rx_oos_packets = atomic_long_read(&stats->rx_oos_packets); in pppol2tp_copy_stats() 1032 dest->rx_errors = atomic_long_read(&stats->rx_errors); in pppol2tp_copy_stats() [all …]
|
D | l2tp_debugfs.c | 169 atomic_long_read(&tunnel->stats.tx_packets), in l2tp_dfs_seq_tunnel_show() 170 atomic_long_read(&tunnel->stats.tx_bytes), in l2tp_dfs_seq_tunnel_show() 171 atomic_long_read(&tunnel->stats.tx_errors), in l2tp_dfs_seq_tunnel_show() 172 atomic_long_read(&tunnel->stats.rx_packets), in l2tp_dfs_seq_tunnel_show() 173 atomic_long_read(&tunnel->stats.rx_bytes), in l2tp_dfs_seq_tunnel_show() 174 atomic_long_read(&tunnel->stats.rx_errors)); in l2tp_dfs_seq_tunnel_show() 220 atomic_long_read(&session->stats.tx_packets), in l2tp_dfs_seq_session_show() 221 atomic_long_read(&session->stats.tx_bytes), in l2tp_dfs_seq_session_show() 222 atomic_long_read(&session->stats.tx_errors), in l2tp_dfs_seq_session_show() 223 atomic_long_read(&session->stats.rx_packets), in l2tp_dfs_seq_session_show() [all …]
|
/net/sched/ |
D | sch_codel.c | 61 struct codel_stats stats; member 95 &q->stats, qdisc_pkt_len, codel_get_enqueue_time, in codel_qdisc_dequeue() 101 if (q->stats.drop_count && sch->q.qlen) { in codel_qdisc_dequeue() 102 qdisc_tree_reduce_backlog(sch, q->stats.drop_count, q->stats.drop_len); in codel_qdisc_dequeue() 103 q->stats.drop_count = 0; in codel_qdisc_dequeue() 104 q->stats.drop_len = 0; in codel_qdisc_dequeue() 195 codel_stats_init(&q->stats); in codel_init() 246 .maxpacket = q->stats.maxpacket, in codel_dump_stats() 252 .ecn_mark = q->stats.ecn_mark, in codel_dump_stats() 253 .ce_mark = q->stats.ce_mark, in codel_dump_stats()
|
D | sch_gred.c | 44 struct red_stats stats; member 229 q->stats.prob_drop++; in gred_enqueue() 233 q->stats.prob_mark++; in gred_enqueue() 240 q->stats.forced_drop++; in gred_enqueue() 243 q->stats.forced_mark++; in gred_enqueue() 252 q->stats.pdrop++; in gred_enqueue() 371 gnet_stats_basic_sync_init(&hw_stats->stats.bstats[i]); in gred_offload_dump_stats() 373 hw_stats->stats.xstats[i] = &table->tab[i]->stats; in gred_offload_dump_stats() 384 table->tab[i]->packetsin += u64_stats_read(&hw_stats->stats.bstats[i].packets); in gred_offload_dump_stats() 385 table->tab[i]->bytesin += u64_stats_read(&hw_stats->stats.bstats[i].bytes); in gred_offload_dump_stats() [all …]
|
D | sch_red.c | 47 struct red_stats stats; member 92 q->stats.prob_drop++; in red_enqueue() 97 q->stats.prob_mark++; in red_enqueue() 102 q->stats.prob_drop++; in red_enqueue() 112 q->stats.forced_drop++; in red_enqueue() 117 q->stats.forced_mark++; in red_enqueue() 122 q->stats.forced_drop++; in red_enqueue() 136 q->stats.pdrop++; in red_enqueue() 401 .stats.bstats = &sch->bstats, in red_dump_offload_stats() 402 .stats.qstats = &sch->qstats, in red_dump_offload_stats() [all …]
|
D | sch_choke.c | 64 } stats; member 232 q->stats.matched++; in choke_enqueue() 244 q->stats.forced_drop++; in choke_enqueue() 248 q->stats.forced_mark++; in choke_enqueue() 256 q->stats.prob_drop++; in choke_enqueue() 260 q->stats.prob_mark++; in choke_enqueue() 275 q->stats.pdrop++; in choke_enqueue() 463 .early = q->stats.prob_drop + q->stats.forced_drop, in choke_dump_stats() 464 .marked = q->stats.prob_mark + q->stats.forced_mark, in choke_dump_stats() 465 .pdrop = q->stats.pdrop, in choke_dump_stats() [all …]
|
/net/strparser/ |
D | strparser.c | 115 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv() 120 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv() 135 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv() 154 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv() 173 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv() 196 STRP_STATS_INCR(strp->stats.mem_fail); in __strp_recv() 223 STRP_STATS_INCR(strp->stats.need_more_hdr); in __strp_recv() 237 STRP_STATS_INCR(strp->stats.msg_too_big); in __strp_recv() 245 STRP_STATS_INCR(strp->stats.bad_hdr_len); in __strp_recv() 276 STRP_STATS_ADD(strp->stats.bytes, cand_len); in __strp_recv() [all …]
|
/net/devlink/ |
D | trap.c | 48 struct devlink_stats __percpu *stats; member 68 struct devlink_stats __percpu *stats; member 157 struct devlink_stats *stats) in devlink_trap_stats_read() argument 161 memset(stats, 0, sizeof(*stats)); in devlink_trap_stats_read() 174 u64_stats_add(&stats->rx_packets, rx_packets); in devlink_trap_stats_read() 175 u64_stats_add(&stats->rx_bytes, rx_bytes); in devlink_trap_stats_read() 183 struct devlink_stats stats; in devlink_trap_group_stats_put() local 186 devlink_trap_stats_read(trap_stats, &stats); in devlink_trap_group_stats_put() 193 u64_stats_read(&stats.rx_packets), in devlink_trap_group_stats_put() 198 u64_stats_read(&stats.rx_bytes), in devlink_trap_group_stats_put() [all …]
|
/net/bridge/ |
D | br_vlan.c | 193 free_percpu(v->stats); in br_master_vlan_rcu_free() 194 v->stats = NULL; in br_master_vlan_rcu_free() 224 free_percpu(v->stats); in nbp_vlan_rcu_free() 225 v->stats = NULL; in nbp_vlan_rcu_free() 310 v->stats = in __vlan_add() 312 if (!v->stats) { in __vlan_add() 318 v->stats = masterv->stats; in __vlan_add() 369 if (v->stats && masterv->stats != v->stats) in __vlan_add() 370 free_percpu(v->stats); in __vlan_add() 371 v->stats = NULL; in __vlan_add() [all …]
|
/net/rds/ |
D | tcp_stats.c | 54 struct rds_tcp_statistics stats = {0, }; in rds_tcp_stats_info_copy() local 65 sum = (uint64_t *)&stats; in rds_tcp_stats_info_copy() 66 for (i = 0; i < sizeof(stats) / sizeof(uint64_t); i++) in rds_tcp_stats_info_copy() 70 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_tcp_stat_names, in rds_tcp_stats_info_copy()
|
D | ib_stats.c | 87 struct rds_ib_statistics stats = {0, }; in rds_ib_stats_info_copy() local 98 sum = (uint64_t *)&stats; in rds_ib_stats_info_copy() 99 for (i = 0; i < sizeof(stats) / sizeof(uint64_t); i++) in rds_ib_stats_info_copy() 103 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_ib_stat_names, in rds_ib_stats_info_copy()
|