| /kernel/linux/linux-4.19/net/sched/ |
| D | sch_drr.c | 2 * net/sched/sch_drr.c Deficit Round Robin scheduler 32 u32 deficit; member 277 xstats.deficit = cl->deficit; in drr_dump_class_stats() 376 cl->deficit = cl->quantum; in drr_enqueue() 402 if (len <= cl->deficit) { in drr_dequeue() 403 cl->deficit -= len; in drr_dequeue() 417 cl->deficit += cl->quantum; in drr_dequeue()
|
| D | sch_hhf.c | 21 * The buckets are dequeued by a Weighted Deficit Round Robin (WDRR) scheduler, 89 * - The Deficit Round Robin engine is taken from fq_codel implementation 118 /* Weighted Deficit Round Robin (WDRR) scheduler */ 123 int deficit; member 399 bucket->deficit = weight * q->quantum; in hhf_enqueue() 433 if (bucket->deficit <= 0) { in hhf_dequeue() 437 bucket->deficit += weight * q->quantum; in hhf_dequeue() 457 bucket->deficit -= qdisc_pkt_len(skb); in hhf_dequeue()
|
| D | sch_fq_codel.c | 51 int deficit; member 217 flow->deficit = q->quantum; in fq_codel_enqueue() 304 if (flow->deficit <= 0) { in fq_codel_dequeue() 305 flow->deficit += q->quantum; in fq_codel_dequeue() 329 flow->deficit -= qdisc_pkt_len(skb); in fq_codel_dequeue() 641 xstats.class_stats.deficit = flow->deficit; in fq_codel_dump_class_stats()
|
| D | sch_cbq.c | 122 long deficit; /* Saved deficit for WRR */ member 689 int deficit; in cbq_dequeue_prio() local 695 deficit = 0; in cbq_dequeue_prio() 705 if (cl->deficit <= 0) { in cbq_dequeue_prio() 709 deficit = 1; in cbq_dequeue_prio() 710 cl->deficit += cl->quantum; in cbq_dequeue_prio() 723 cl->deficit -= qdisc_pkt_len(skb); in cbq_dequeue_prio() 737 if (cl->deficit <= 0) { in cbq_dequeue_prio() 740 cl->deficit += cl->quantum; in cbq_dequeue_prio() 779 } while (deficit); in cbq_dequeue_prio() [all …]
|
| D | sch_qfq.c | 143 int deficit; /* DRR deficit counter. */ member 481 cl->deficit = lmax; in qfq_change_class() 992 cl->deficit -= (int) len; in agg_dequeue() 996 else if (cl->deficit < qdisc_pkt_len(cl->qdisc->ops->peek(cl->qdisc))) { in agg_dequeue() 997 cl->deficit += agg->lmax; in agg_dequeue() 1257 == cl && cl->deficit < qdisc_pkt_len(skb)) in qfq_enqueue() 1264 cl->deficit = agg->lmax; in qfq_enqueue()
|
| D | sch_cake.c | 19 * equipment and bloated MACs. This operates in deficit mode (as in sch_fq), 130 s32 deficit; member 1825 flow->deficit = (b->flow_quantum * in cake_enqueue() 2010 if (flow->deficit <= 0) { in cake_dequeue() 2014 flow->deficit += (b->flow_quantum * quantum_div[host_load] + in cake_dequeue() 2029 * correct deficit accounting but we still want in cake_dequeue() 2092 flow->deficit -= len; in cake_dequeue() 2116 flow->deficit -= len; in cake_dequeue() 2963 PUT_STAT_S32(DEFICIT, flow->deficit); in cake_dump_class_stats()
|
| /kernel/linux/linux-5.10/net/sched/ |
| D | sch_drr.c | 3 * net/sched/sch_drr.c Deficit Round Robin scheduler 29 u32 deficit; member 267 xstats.deficit = cl->deficit; in drr_dump_class_stats() 369 cl->deficit = cl->quantum; in drr_enqueue() 395 if (len <= cl->deficit) { in drr_dequeue() 396 cl->deficit -= len; in drr_dequeue() 410 cl->deficit += cl->quantum; in drr_dequeue()
|
| D | sch_ets.c | 24 * ones are tried next. Each band is assigned a deficit counter, initialized to 27 * head of the list if the packet size is smaller or equal to the deficit 43 u32 deficit; member 447 cl->deficit = cl->quantum; in ets_qdisc_enqueue() 491 if (len <= cl->deficit) { in ets_qdisc_dequeue() 492 cl->deficit -= len; in ets_qdisc_dequeue() 501 cl->deficit += cl->quantum; in ets_qdisc_dequeue() 667 q->classes[i].deficit = quanta[i]; in ets_qdisc_change()
|
| D | sch_fq_pie.c | 37 * @deficit: number of remaining byte credits 46 s32 deficit; member 188 sel_flow->deficit = q->quantum; in fq_pie_qdisc_enqueue() 246 if (flow->deficit <= 0) { in fq_pie_qdisc_dequeue() 247 flow->deficit += q->quantum; in fq_pie_qdisc_dequeue() 270 flow->deficit -= pkt_len; in fq_pie_qdisc_dequeue()
|
| D | sch_hhf.c | 22 * The buckets are dequeued by a Weighted Deficit Round Robin (WDRR) scheduler, 90 * - The Deficit Round Robin engine is taken from fq_codel implementation 119 /* Weighted Deficit Round Robin (WDRR) scheduler */ 124 int deficit; member 400 bucket->deficit = weight * q->quantum; in hhf_enqueue() 434 if (bucket->deficit <= 0) { in hhf_dequeue() 438 bucket->deficit += weight * q->quantum; in hhf_dequeue() 458 bucket->deficit -= qdisc_pkt_len(skb); in hhf_dequeue()
|
| D | sch_fq_codel.c | 46 int deficit; member 212 flow->deficit = q->quantum; in fq_codel_enqueue() 297 if (flow->deficit <= 0) { in fq_codel_dequeue() 298 flow->deficit += q->quantum; in fq_codel_dequeue() 316 flow->deficit -= qdisc_pkt_len(skb); in fq_codel_dequeue() 636 xstats.class_stats.deficit = flow->deficit; in fq_codel_dump_class_stats()
|
| D | sch_cbq.c | 117 long deficit; /* Saved deficit for WRR */ member 684 int deficit; in cbq_dequeue_prio() local 690 deficit = 0; in cbq_dequeue_prio() 700 if (cl->deficit <= 0) { in cbq_dequeue_prio() 704 deficit = 1; in cbq_dequeue_prio() 705 cl->deficit += cl->quantum; in cbq_dequeue_prio() 718 cl->deficit -= qdisc_pkt_len(skb); in cbq_dequeue_prio() 732 if (cl->deficit <= 0) { in cbq_dequeue_prio() 735 cl->deficit += cl->quantum; in cbq_dequeue_prio() 774 } while (deficit); in cbq_dequeue_prio() [all …]
|
| D | sch_qfq.c | 140 int deficit; /* DRR deficit counter. */ member 469 cl->deficit = lmax; in qfq_change_class() 977 cl->deficit -= (int) len; in agg_dequeue() 981 else if (cl->deficit < qdisc_pkt_len(cl->qdisc->ops->peek(cl->qdisc))) { in agg_dequeue() 982 cl->deficit += agg->lmax; in agg_dequeue() 1246 == cl && cl->deficit < len) in qfq_enqueue() 1253 cl->deficit = agg->lmax; in qfq_enqueue()
|
| D | sch_cake.c | 19 * equipment and bloated MACs. This operates in deficit mode (as in sch_fq), 130 s32 deficit; member 1874 flow->deficit = (b->flow_quantum * in cake_enqueue() 2061 if (flow->deficit <= 0) { in cake_dequeue() 2080 * correct deficit accounting but we still want in cake_dequeue() 2098 flow->deficit += (b->flow_quantum * quantum_div[host_load] + in cake_dequeue() 2171 flow->deficit -= len; in cake_dequeue() 2195 flow->deficit -= len; in cake_dequeue() 3031 PUT_STAT_S32(DEFICIT, flow->deficit); in cake_dump_class_stats()
|
| /kernel/linux/linux-5.10/include/net/ |
| D | fq.h | 20 * (deficit round robin) based round robin queuing similar to the one 27 * @deficit: used for DRR++ 35 int deficit; member
|
| D | fq_impl.h | 82 if (flow->deficit <= 0) { in fq_tin_dequeue() 83 flow->deficit += fq->quantum; in fq_tin_dequeue() 102 flow->deficit -= skb->len; in fq_tin_dequeue() 179 flow->deficit = fq->quantum; in fq_tin_enqueue()
|
| /kernel/linux/linux-4.19/include/net/ |
| D | fq.h | 21 * (deficit round robin) based round robin queuing similar to the one 28 * @deficit: used for DRR++ 36 int deficit; member
|
| D | fq_impl.h | 83 if (flow->deficit <= 0) { in fq_tin_dequeue() 84 flow->deficit += fq->quantum; in fq_tin_dequeue() 103 flow->deficit -= skb->len; in fq_tin_dequeue() 178 flow->deficit = fq->quantum; in fq_tin_enqueue()
|
| /kernel/linux/linux-4.19/drivers/net/ethernet/intel/ixgbe/ |
| D | ixgbe_dcb_82598.h | 10 #define IXGBE_DPMCS_TDPAC 0x00000001 /* 0 Round Robin, 1 DFP - Deficit Fixed Priority */ 33 #define IXGBE_PDPMCS_TPPAC 0x00000020 /* 0 Round Robin, 1 for DFP - Deficit Fixed Priority */
|
| /kernel/linux/linux-5.10/drivers/net/ethernet/intel/ixgbe/ |
| D | ixgbe_dcb_82598.h | 10 #define IXGBE_DPMCS_TDPAC 0x00000001 /* 0 Round Robin, 1 DFP - Deficit Fixed Priority */ 33 #define IXGBE_PDPMCS_TPPAC 0x00000020 /* 0 Round Robin, 1 for DFP - Deficit Fixed Priority */
|
| /kernel/linux/linux-5.10/net/mac80211/ |
| D | debugfs_sta.c | 204 s64 deficit[IEEE80211_NUM_ACS]; in sta_airtime_read() local 215 deficit[ac] = sta->airtime[ac].deficit; in sta_airtime_read() 221 "Deficit: VO: %lld us VI: %lld us BE: %lld us BK: %lld us\n", in sta_airtime_read() 223 deficit[0], deficit[1], deficit[2], deficit[3]); in sta_airtime_read() 241 sta->airtime[ac].deficit = sta->airtime_weight; in sta_airtime_write()
|
| /kernel/linux/linux-4.19/kernel/sched/ |
| D | features.h | 3 * Only give sleepers 50% of their service deficit. This allows
|
| /kernel/linux/linux-5.10/kernel/sched/ |
| D | features.h | 3 * Only give sleepers 50% of their service deficit. This allows
|
| /kernel/linux/patches/linux-4.19/prebuilts/usr/include/linux/ |
| D | pkt_sched.h | 471 __u32 deficit; member 604 __s32 deficit; member
|
| /kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/mt7615/ |
| D | sdio_mcu.c | 35 sdio->sched.deficit = txdwcnt << 2; in mt7663s_mcu_init_sched()
|