/net/sched/ |
D | sch_sfq.c | 103 sfq_index qlen; /* number of skbs in skblist */ member 207 int qlen = slot->qlen; in sfq_link() local 209 p = qlen + SFQ_MAX_FLOWS; in sfq_link() 210 n = q->dep[qlen].next; in sfq_link() 215 q->dep[qlen].next = x; /* sfq_dep_head(q, p)->next = x */ in sfq_link() 235 d = q->slots[x].qlen--; in sfq_dec() 248 d = ++q->slots[x].qlen; in sfq_inc() 310 sch->q.qlen--; in sfq_drop() 350 sfq_index x, qlen; in sfq_enqueue() local 424 if (slot->qlen >= q->maxdepth) { in sfq_enqueue() [all …]
|
D | sch_sfb.c | 40 u16 qlen; /* length of virtual queue */ member 132 if (b[hash].qlen < 0xFFFF) in increment_one_qlen() 133 b[hash].qlen++; in increment_one_qlen() 161 if (b[hash].qlen > 0) in decrement_one_qlen() 162 b[hash].qlen--; in decrement_one_qlen() 201 u32 qlen = 0, prob = 0, totalpm = 0; in sfb_compute_qlen() local 205 if (qlen < b->qlen) in sfb_compute_qlen() 206 qlen = b->qlen; in sfb_compute_qlen() 214 return qlen; in sfb_compute_qlen() 293 if (unlikely(sch->q.qlen >= q->limit)) { in sfb_enqueue() [all …]
|
D | sch_mq.c | 133 __u32 qlen = 0; in mq_dump() local 135 sch->q.qlen = 0; in mq_dump() 149 qlen = qdisc_qlen_sum(qdisc); in mq_dump() 155 &qdisc->qstats, qlen); in mq_dump() 156 sch->q.qlen += qlen; in mq_dump() 158 sch->q.qlen += qdisc->q.qlen; in mq_dump() 161 sch->qstats.qlen += qdisc->qstats.qlen; in mq_dump()
|
D | sch_pie.c | 204 unsigned int qlen, dropped = 0; in pie_change() local 251 qlen = sch->q.qlen; in pie_change() 252 while (sch->q.qlen > sch->limit) { in pie_change() 259 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped); in pie_change() 268 int qlen = sch->qstats.backlog; /* current queue size in bytes */ in pie_process_dequeue() local 274 if (qlen >= QUEUE_THRESHOLD && q->vars.dq_count == DQCOUNT_INVALID) { in pie_process_dequeue() 313 if (qlen < QUEUE_THRESHOLD) { in pie_process_dequeue() 333 u32 qlen = sch->qstats.backlog; /* queue size in bytes */ in calculate_probability() local 345 qdelay = (qlen << PIE_SCALE) / q->vars.avg_dq_rate; in calculate_probability() 352 if (qdelay == 0 && qlen != 0) in calculate_probability() [all …]
|
D | sch_codel.c | 101 if (q->stats.drop_count && sch->q.qlen) { in codel_qdisc_dequeue() 138 unsigned int qlen, dropped = 0; in codel_change() local 175 qlen = sch->q.qlen; in codel_change() 176 while (sch->q.qlen > sch->limit) { in codel_change() 183 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped); in codel_change()
|
D | sch_skbprio.c | 82 if (sch->q.qlen < sch->limit) { in skbprio_enqueue() 94 sch->q.qlen++; in skbprio_enqueue() 125 BUG_ON(sch->q.qlen != 1); in skbprio_enqueue() 148 sch->q.qlen--; in skbprio_dequeue() 157 BUG_ON(sch->q.qlen); in skbprio_dequeue() 214 sch->q.qlen = 0; in skbprio_reset() 255 q->qstats[cl - 1].qlen) < 0) in skbprio_dump_class_stats()
|
D | sch_choke.c | 130 --sch->q.qlen; in choke_drop_by_idx() 227 q->vars.qavg = red_calc_qavg(p, &q->vars, sch->q.qlen); in choke_enqueue() 274 if (sch->q.qlen < q->limit) { in choke_enqueue() 277 ++sch->q.qlen; in choke_enqueue() 304 --sch->q.qlen; in choke_dequeue() 324 sch->q.qlen = 0; in choke_reset() 387 unsigned int oqlen = sch->q.qlen, tail = 0; in choke_change() 402 --sch->q.qlen; in choke_change() 405 qdisc_tree_reduce_backlog(sch, oqlen - sch->q.qlen, dropped); in choke_change()
|
D | sch_drr.c | 261 __u32 qlen = qdisc_qlen_sum(cl->qdisc); in drr_dump_class_stats() local 266 if (qlen) in drr_dump_class_stats() 272 gnet_stats_copy_queue(d, cl_q->cpu_qstats, &cl_q->qstats, qlen) < 0) in drr_dump_class_stats() 357 first = !cl->qdisc->q.qlen; in drr_enqueue() 373 sch->q.qlen++; in drr_enqueue() 400 if (cl->qdisc->q.qlen == 0) in drr_dequeue() 406 sch->q.qlen--; in drr_dequeue() 441 if (cl->qdisc->q.qlen) in drr_reset_qdisc() 447 sch->q.qlen = 0; in drr_reset_qdisc()
|
D | sch_mqprio.c | 392 sch->q.qlen = 0; in mqprio_dump() 406 __u32 qlen = qdisc_qlen_sum(qdisc); in mqprio_dump() local 413 &qdisc->qstats, qlen); in mqprio_dump() 414 sch->q.qlen += qlen; in mqprio_dump() 416 sch->q.qlen += qdisc->q.qlen; in mqprio_dump() 515 __u32 qlen = 0; in mqprio_dump_class_stats() local 541 qlen = qdisc_qlen_sum(qdisc); in mqprio_dump_class_stats() 547 qlen); in mqprio_dump_class_stats() 555 gnet_stats_copy_queue(d, NULL, &qstats, qlen) < 0) in mqprio_dump_class_stats()
|
D | sch_hhf.c | 362 sch->q.qlen--; in hhf_drop() 402 if (++sch->q.qlen <= sch->limit) in hhf_enqueue() 445 sch->q.qlen--; in hhf_dequeue() 513 unsigned int qlen, prev_backlog; in hhf_change() local 563 qlen = sch->q.qlen; in hhf_change() 565 while (sch->q.qlen > sch->limit) { in hhf_change() 570 qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, in hhf_change()
|
D | sch_fq_codel.c | 181 sch->q.qlen -= i; in fq_codel_drop() 218 if (++sch->q.qlen <= sch->limit && !memory_limited) in fq_codel_enqueue() 222 prev_qlen = sch->q.qlen; in fq_codel_enqueue() 233 prev_qlen -= sch->q.qlen; in fq_codel_enqueue() 268 sch->q.qlen--; in dequeue_func() 321 if (q->cstats.drop_count && sch->q.qlen) { in fq_codel_dequeue() 351 sch->q.qlen = 0; in fq_codel_reset() 425 while (sch->q.qlen > sch->limit || in fq_codel_change() 647 qs.qlen++; in fq_codel_dump_class_stats() 655 if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0) in fq_codel_dump_class_stats()
|
D | sch_fq.c | 77 int qlen; /* number of packets in flow queue */ member 379 flow->qlen--; in fq_dequeue_head() 381 sch->q.qlen--; in fq_dequeue_head() 426 if (unlikely(sch->q.qlen >= sch->limit)) in fq_enqueue() 430 if (unlikely(f->qlen >= q->flow_plimit && f != &q->internal)) { in fq_enqueue() 435 f->qlen++; in fq_enqueue() 450 sch->q.qlen++; in fq_enqueue() 492 if (!sch->q.qlen) in fq_dequeue() 615 flow->qlen = 0; in fq_flow_purge() 626 sch->q.qlen = 0; in fq_reset() [all …]
|
D | sch_generic.c | 73 q->q.qlen--; in __skb_dequeue_bad_txq() 113 q->q.qlen++; in qdisc_enqueue_skb_bad_txq() 142 q->q.qlen++; in dev_requeue_skb() 241 q->q.qlen--; in dequeue_skb() 329 dev->name, ret, q->q.qlen); in sch_direct_xmit() 565 .qlen = 0, 571 .qlen = 0, 706 q->qlen = 0; in pfifo_fast_reset() 727 unsigned int qlen = qdisc_dev(qdisc)->tx_queue_len; in pfifo_fast_init() local 732 if (!qlen) in pfifo_fast_init() [all …]
|
D | sch_teql.c | 81 if (q->q.qlen < dev->tx_queue_len) { in teql_enqueue() 110 sch->q.qlen = dat->q.qlen + q->q.qlen; in teql_dequeue() 127 sch->q.qlen = 0; in teql_reset()
|
D | sch_etf.c | 191 sch->q.qlen++; in etf_enqueue_timesortedlist() 224 sch->q.qlen--; in timesortedlist_drop() 249 sch->q.qlen--; in timesortedlist_remove() 431 sch->q.qlen--; in timesortedlist_clear() 448 sch->q.qlen = 0; in etf_reset()
|
D | sch_cbq.c | 378 sch->q.qlen++; in cbq_enqueue() 533 if (cl->q->q.qlen > 1) { in cbq_update_toplevel() 696 if (cl->q->q.qlen && in cbq_dequeue_prio() 740 if (cl->q->q.qlen == 0 || prio != cl->cpriority) { in cbq_dequeue_prio() 757 if (cl->q->q.qlen) in cbq_dequeue_prio() 764 if (cl->q->q.qlen) in cbq_dequeue_prio() 819 sch->q.qlen--; in cbq_dequeue() 853 if (sch->q.qlen) { in cbq_dequeue() 1056 sch->q.qlen = 0; in cbq_reset() 1378 __u32 qlen; in cbq_dump_class_stats() local [all …]
|
D | sch_hfsc.c | 749 if (cl->qdisc->q.qlen == 0 && cl->cl_flags & HFSC_FSC) in update_vf() 984 if (cl->qdisc->q.qlen != 0) { in hfsc_change_class() 1322 __u32 qlen; in hfsc_dump_class_stats() local 1324 qdisc_qstats_qlen_backlog(cl->qdisc, &qlen, &cl->qstats.backlog); in hfsc_dump_class_stats() 1332 gnet_stats_copy_queue(d, NULL, &cl->qstats, qlen) < 0) in hfsc_dump_class_stats() 1488 sch->q.qlen = 0; in hfsc_reset_qdisc() 1547 first = !cl->qdisc->q.qlen; in hfsc_enqueue() 1573 sch->q.qlen++; in hfsc_enqueue() 1588 if (sch->q.qlen == 0) in hfsc_dequeue() 1626 if (cl->qdisc->q.qlen != 0) { in hfsc_dequeue() [all …]
|
D | sch_htb.c | 557 WARN_ON(cl->level || !cl->leaf.q || !cl->leaf.q->q.qlen); in htb_activate() 589 if (q->direct_queue.qlen < q->direct_qlen) { in htb_enqueue() 614 sch->q.qlen++; in htb_enqueue() 844 if (unlikely(cl->leaf.q->q.qlen == 0)) { in htb_dequeue_tree() 882 if (!cl->leaf.q->q.qlen) in htb_dequeue_tree() 903 sch->q.qlen--; in htb_dequeue() 907 if (!sch->q.qlen) in htb_dequeue() 969 sch->q.qlen = 0; in htb_reset() 1128 __u32 qlen = 0; in htb_dump_class_stats() local 1131 qdisc_qstats_qlen_backlog(cl->leaf.q, &qlen, &qs.backlog); in htb_dump_class_stats() [all …]
|
D | sch_red.c | 100 sch->q.qlen++; in red_enqueue() 122 sch->q.qlen--; in red_dequeue() 144 sch->q.qlen = 0; in red_reset() 249 if (!q->qdisc->q.qlen) in red_change()
|
D | sch_tbf.c | 172 sch->q.qlen += nb; in tbf_segment() 200 sch->q.qlen++; in tbf_enqueue() 245 sch->q.qlen--; in tbf_dequeue() 275 sch->q.qlen = 0; in tbf_reset()
|
D | sch_qfq.c | 318 if (cl->qdisc->q.qlen > 0) { /* adding an active class */ in qfq_add_to_agg() 367 if (cl->qdisc->q.qlen > 0) /* class is active */ in qfq_deact_rm_from_agg() 981 if (cl->qdisc->q.qlen == 0) /* no more packets, remove from list */ in agg_dequeue() 1117 } else if (sch->q.qlen == 0) { /* no aggregate to serve */ in qfq_dequeue() 1133 sch->q.qlen--; in qfq_dequeue() 1227 first = !cl->qdisc->q.qlen; in qfq_enqueue() 1241 ++sch->q.qlen; in qfq_enqueue() 1459 if (cl->qdisc->q.qlen > 0) in qfq_reset_qdisc() 1466 sch->q.qlen = 0; in qfq_reset_qdisc()
|
/net/core/ |
D | gen_stats.c | 290 qstats->qlen = 0; in __gnet_stats_copy_queue_cpu() 301 __u32 qlen) in __gnet_stats_copy_queue() argument 306 qstats->qlen = q->qlen; in __gnet_stats_copy_queue() 313 qstats->qlen = qlen; in __gnet_stats_copy_queue() 334 struct gnet_stats_queue *q, __u32 qlen) in gnet_stats_copy_queue() argument 338 __gnet_stats_copy_queue(&qstats, cpu_q, q, qlen); in gnet_stats_copy_queue() 342 d->tc_stats.qlen = qstats.qlen; in gnet_stats_copy_queue()
|
D | request_sock.c | 41 queue->fastopenq.qlen = 0; in reqsk_queue_alloc() 101 fastopenq->qlen--; in reqsk_fastopen_remove() 129 fastopenq->qlen++; in reqsk_fastopen_remove()
|
/net/nfc/hci/ |
D | llc_shdlc.c | 324 if (shdlc->send_q.qlen == 0) { in llc_shdlc_rcv_s_frame() 468 if (shdlc->rcv_q.qlen) in llc_shdlc_handle_rcv_queue() 469 pr_debug("rcvQlen=%d\n", shdlc->rcv_q.qlen); in llc_shdlc_handle_rcv_queue() 524 if (shdlc->send_q.qlen) in llc_shdlc_handle_send_queue() 527 shdlc->send_q.qlen, shdlc->ns, shdlc->dnr, in llc_shdlc_handle_send_queue() 530 shdlc->ack_pending_q.qlen); in llc_shdlc_handle_send_queue() 532 while (shdlc->send_q.qlen && shdlc->ack_pending_q.qlen < shdlc->w && in llc_shdlc_handle_send_queue()
|
/net/kcm/ |
D | kcmproc.c | 119 kcm->sk.sk_receive_queue.qlen, in kcm_format_sock() 121 kcm->sk.sk_write_queue.qlen, in kcm_format_sock() 149 psock->sk->sk_receive_queue.qlen, in kcm_format_psock() 151 psock->sk->sk_write_queue.qlen, in kcm_format_psock() 167 if (psock->sk->sk_receive_queue.qlen) { in kcm_format_psock()
|