/net/core/ |
D | gen_stats.c | 286 __gnet_stats_copy_queue_cpu(struct gnet_stats_queue *qstats, in __gnet_stats_copy_queue_cpu() argument 294 qstats->qlen = 0; in __gnet_stats_copy_queue_cpu() 295 qstats->backlog += qcpu->backlog; in __gnet_stats_copy_queue_cpu() 296 qstats->drops += qcpu->drops; in __gnet_stats_copy_queue_cpu() 297 qstats->requeues += qcpu->requeues; in __gnet_stats_copy_queue_cpu() 298 qstats->overlimits += qcpu->overlimits; in __gnet_stats_copy_queue_cpu() 302 void __gnet_stats_copy_queue(struct gnet_stats_queue *qstats, in __gnet_stats_copy_queue() argument 308 __gnet_stats_copy_queue_cpu(qstats, cpu); in __gnet_stats_copy_queue() 310 qstats->qlen = q->qlen; in __gnet_stats_copy_queue() 311 qstats->backlog = q->backlog; in __gnet_stats_copy_queue() [all …]
|
/net/sched/ |
D | sch_skbprio.c | 35 struct gnet_stats_queue qstats[SKBPRIO_MAX_PRIORITY]; member 85 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() 101 q->qstats[prio].drops++; in skbprio_enqueue() 102 q->qstats[prio].overlimits++; in skbprio_enqueue() 108 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() 117 q->qstats[lp].backlog -= qdisc_pkt_len(to_drop); in skbprio_enqueue() 118 q->qstats[lp].drops++; in skbprio_enqueue() 119 q->qstats[lp].overlimits++; in skbprio_enqueue() 152 q->qstats[q->highest_prio].backlog -= qdisc_pkt_len(skb); in skbprio_dequeue() 189 memset(&q->qstats, 0, sizeof(q->qstats)); in skbprio_init() [all …]
|
D | sch_mq.c | 45 .qstats = &sch->qstats, in mq_offload_stats() 160 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mq_dump() 176 __gnet_stats_copy_queue(&sch->qstats, in mq_dump() 178 &qdisc->qstats, qlen); in mq_dump() 184 sch->qstats.qlen += qdisc->qstats.qlen; in mq_dump() 185 sch->qstats.backlog += qdisc->qstats.backlog; in mq_dump() 186 sch->qstats.drops += qdisc->qstats.drops; in mq_dump() 187 sch->qstats.requeues += qdisc->qstats.requeues; in mq_dump() 188 sch->qstats.overlimits += qdisc->qstats.overlimits; in mq_dump()
|
D | sch_mqprio.c | 458 memset(&sch->qstats, 0, sizeof(sch->qstats)); in mqprio_dump() 475 __gnet_stats_copy_queue(&sch->qstats, in mqprio_dump() 477 &qdisc->qstats, qlen); in mqprio_dump() 483 sch->qstats.backlog += qdisc->qstats.backlog; in mqprio_dump() 484 sch->qstats.drops += qdisc->qstats.drops; in mqprio_dump() 485 sch->qstats.requeues += qdisc->qstats.requeues; in mqprio_dump() 486 sch->qstats.overlimits += qdisc->qstats.overlimits; in mqprio_dump() 580 struct gnet_stats_queue qstats = {0}; in mqprio_dump_class_stats() local 605 __gnet_stats_copy_queue(&qstats, in mqprio_dump_class_stats() 607 &qdisc->qstats, in mqprio_dump_class_stats() [all …]
|
D | sch_gred.c | 117 return sch->qstats.backlog; in gred_backlog() 181 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= in gred_enqueue() 278 if (!sch->qstats.backlog) in gred_dequeue() 346 opt.set.qstats = &sch->qstats; in gred_offload() 380 table->tab[i]->backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats() 385 sch->qstats.qlen += hw_stats->stats.qstats[i].qlen; in gred_offload_dump_stats() 386 sch->qstats.backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats() 387 sch->qstats.drops += hw_stats->stats.qstats[i].drops; in gred_offload_dump_stats() 388 sch->qstats.requeues += hw_stats->stats.qstats[i].requeues; in gred_offload_dump_stats() 389 sch->qstats.overlimits += hw_stats->stats.qstats[i].overlimits; in gred_offload_dump_stats()
|
D | sch_fifo.c | 22 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= sch->limit)) in bfifo_enqueue() 45 prev_backlog = sch->qstats.backlog; in pfifo_tail_enqueue() 51 qdisc_tree_reduce_backlog(sch, 0, prev_backlog - sch->qstats.backlog); in pfifo_tail_enqueue() 91 qopt.stats.qstats = &sch->qstats; in fifo_offload_dump()
|
D | sch_prio.c | 88 sch->qstats.backlog += len; in prio_enqueue() 156 opt.replace_params.qstats = &sch->qstats; in prio_offload() 254 .qstats = &sch->qstats, in prio_dump_offload()
|
D | sch_red.c | 80 child->qstats.backlog); in red_enqueue() 133 sch->qstats.backlog += len; in red_enqueue() 205 opt.set.qstats = &sch->qstats; in red_offload() 402 .stats.qstats = &sch->qstats, in red_dump_offload_stats()
|
D | sch_ets.c | 45 struct gnet_stats_queue qstats; member 125 qopt.replace_params.qstats = &sch->qstats; in ets_offload_change() 185 qopt.stats.qstats = &sch->qstats; in ets_offload_dump() 439 cl->qstats.drops++; in ets_qdisc_enqueue() 450 sch->qstats.backlog += len; in ets_qdisc_enqueue()
|
D | sch_tbf.c | 155 qopt.replace_params.qstats = &sch->qstats; in tbf_offload_change() 182 qopt.stats.qstats = &sch->qstats; in tbf_offload_dump() 244 sch->qstats.backlog += len; in tbf_enqueue()
|
D | sch_fq_codel.c | 178 sch->qstats.drops += i; in fq_codel_drop() 179 sch->qstats.backlog -= len; in fq_codel_drop() 220 prev_backlog = sch->qstats.backlog; in fq_codel_enqueue() 233 prev_backlog -= sch->qstats.backlog; in fq_codel_enqueue() 268 sch->qstats.backlog -= qdisc_pkt_len(skb); in dequeue_func() 303 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->cparams, in fq_codel_dequeue()
|
D | sch_drr.c | 23 struct gnet_stats_queue qstats; member 273 gnet_stats_copy_queue(d, cl_q->cpu_qstats, &cl_q->qstats, qlen) < 0) in drr_dump_class_stats() 362 cl->qstats.drops++; in drr_enqueue() 373 sch->qstats.backlog += len; in drr_enqueue()
|
D | sch_hhf.c | 405 prev_backlog = sch->qstats.backlog; in hhf_enqueue() 414 qdisc_tree_reduce_backlog(sch, 1, prev_backlog - sch->qstats.backlog); in hhf_enqueue() 564 prev_backlog = sch->qstats.backlog; in hhf_change() 571 prev_backlog - sch->qstats.backlog); in hhf_change()
|
D | sch_pie.c | 96 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, in pie_qdisc_enqueue() 430 pie_calculate_probability(&q->params, &q->vars, sch->qstats.backlog); in pie_timer() 525 pie_process_dequeue(skb, &q->params, &q->vars, sch->qstats.backlog); in pie_qdisc_dequeue()
|
D | sch_codel.c | 75 sch->qstats.backlog -= qdisc_pkt_len(skb); in dequeue_func() 94 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->params, &q->vars, in codel_qdisc_dequeue()
|
D | sch_sfb.c | 407 sch->qstats.backlog += len; in sfb_enqueue() 583 sch->qstats.backlog = q->qdisc->qstats.backlog; in sfb_dump()
|
D | sch_atm.c | 56 struct gnet_stats_queue qstats; member 445 flow->qstats.drops++; in atm_tc_enqueue() 659 gnet_stats_copy_queue(d, NULL, &flow->qstats, flow->q->q.qlen) < 0) in atm_tc_dump_class_stats()
|
D | sch_plug.c | 92 if (likely(sch->qstats.backlog + skb->len <= q->limit)) { in plug_enqueue()
|
D | sch_hfsc.c | 115 struct gnet_stats_queue qstats; member 1343 qdisc_qstats_qlen_backlog(cl->qdisc, &qlen, &cl->qstats.backlog); in hfsc_dump_class_stats() 1351 gnet_stats_copy_queue(d, NULL, &cl->qstats, qlen) < 0) in hfsc_dump_class_stats() 1568 cl->qstats.drops++; in hfsc_enqueue() 1589 sch->qstats.backlog += len; in hfsc_enqueue()
|
D | sch_cbq.c | 120 struct gnet_stats_queue qstats; member 388 cl->qstats.drops++; in cbq_enqueue() 666 this_cl->qstats.overlimits++; in cbq_under_limit() 1381 qdisc_qstats_qlen_backlog(cl->q, &qlen, &cl->qstats.backlog); in cbq_dump_class_stats() 1389 gnet_stats_copy_queue(d, NULL, &cl->qstats, qlen) < 0) in cbq_dump_class_stats()
|
D | sch_qfq.c | 136 struct gnet_stats_queue qstats; member 1242 cl->qstats.drops++; in qfq_enqueue() 1253 cl->qstats.drops++; in qfq_enqueue() 1261 sch->qstats.backlog += len; in qfq_enqueue()
|
D | sch_cake.c | 1540 sch->qstats.backlog -= len; in cake_drop() 1545 sch->qstats.drops++; in cake_drop() 1732 sch->qstats.overlimits++; in cake_enqueue() 1769 sch->qstats.backlog += slen; in cake_enqueue() 1785 sch->qstats.drops++; in cake_enqueue() 1804 sch->qstats.backlog += len; in cake_enqueue() 1923 sch->qstats.backlog -= len; in cake_dequeue_one() 1969 sch->qstats.overlimits++; in cake_dequeue()
|
D | sch_fq_pie.c | 183 sch->qstats.backlog += pkt_len; in fq_pie_qdisc_enqueue() 262 sch->qstats.backlog -= pkt_len; in fq_pie_qdisc_dequeue()
|
D | sch_dsmark.c | 278 sch->qstats.backlog += len; in dsmark_enqueue()
|
D | sch_cbs.c | 99 sch->qstats.backlog += len; in cbs_child_enqueue()
|