Home
last modified time | relevance | path

Searched refs:sch (Results 1 – 25 of 37) sorted by relevance

12

/net/sched/
Dsch_mq.c24 static int mq_offload(struct Qdisc *sch, enum tc_mq_command cmd) in mq_offload() argument
26 struct net_device *dev = qdisc_dev(sch); in mq_offload()
29 .handle = sch->handle, in mq_offload()
38 static int mq_offload_stats(struct Qdisc *sch) in mq_offload_stats() argument
42 .handle = sch->handle, in mq_offload_stats()
44 .bstats = &sch->bstats, in mq_offload_stats()
45 .qstats = &sch->qstats, in mq_offload_stats()
49 return qdisc_offload_dump_helper(sch, TC_SETUP_QDISC_MQ, &opt); in mq_offload_stats()
52 static void mq_destroy(struct Qdisc *sch) in mq_destroy() argument
54 struct net_device *dev = qdisc_dev(sch); in mq_destroy()
[all …]
Dsch_prio.c31 prio_classify(struct sk_buff *skb, struct Qdisc *sch, int *qerr) in prio_classify() argument
33 struct prio_sched_data *q = qdisc_priv(sch); in prio_classify()
40 if (TC_H_MAJ(skb->priority) != sch->handle) { in prio_classify()
69 prio_enqueue(struct sk_buff *skb, struct Qdisc *sch, struct sk_buff **to_free) in prio_enqueue() argument
75 qdisc = prio_classify(skb, sch, &ret); in prio_enqueue()
80 qdisc_qstats_drop(sch); in prio_enqueue()
88 sch->qstats.backlog += len; in prio_enqueue()
89 sch->q.qlen++; in prio_enqueue()
93 qdisc_qstats_drop(sch); in prio_enqueue()
97 static struct sk_buff *prio_peek(struct Qdisc *sch) in prio_peek() argument
[all …]
Dsch_red.c40 struct Qdisc *sch; member
57 static int red_enqueue(struct sk_buff *skb, struct Qdisc *sch, in red_enqueue() argument
60 struct red_sched_data *q = qdisc_priv(sch); in red_enqueue()
77 qdisc_qstats_overlimit(sch); in red_enqueue()
87 qdisc_qstats_overlimit(sch); in red_enqueue()
101 sch->qstats.backlog += len; in red_enqueue()
102 sch->q.qlen++; in red_enqueue()
105 qdisc_qstats_drop(sch); in red_enqueue()
110 qdisc_drop(skb, sch, to_free); in red_enqueue()
114 static struct sk_buff *red_dequeue(struct Qdisc *sch) in red_dequeue() argument
[all …]
Dsch_ingress.c23 static struct Qdisc *ingress_leaf(struct Qdisc *sch, unsigned long arg) in ingress_leaf() argument
28 static unsigned long ingress_find(struct Qdisc *sch, u32 classid) in ingress_find() argument
33 static unsigned long ingress_bind_filter(struct Qdisc *sch, in ingress_bind_filter() argument
36 return ingress_find(sch, classid); in ingress_bind_filter()
39 static void ingress_unbind_filter(struct Qdisc *sch, unsigned long cl) in ingress_unbind_filter() argument
43 static void ingress_walk(struct Qdisc *sch, struct qdisc_walker *walker) in ingress_walk() argument
47 static struct tcf_block *ingress_tcf_block(struct Qdisc *sch, unsigned long cl, in ingress_tcf_block() argument
50 struct ingress_sched_data *q = qdisc_priv(sch); in ingress_tcf_block()
62 static void ingress_ingress_block_set(struct Qdisc *sch, u32 block_index) in ingress_ingress_block_set() argument
64 struct ingress_sched_data *q = qdisc_priv(sch); in ingress_ingress_block_set()
[all …]
Dsch_fifo.c18 static int bfifo_enqueue(struct sk_buff *skb, struct Qdisc *sch, in bfifo_enqueue() argument
21 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= sch->limit)) in bfifo_enqueue()
22 return qdisc_enqueue_tail(skb, sch); in bfifo_enqueue()
24 return qdisc_drop(skb, sch, to_free); in bfifo_enqueue()
27 static int pfifo_enqueue(struct sk_buff *skb, struct Qdisc *sch, in pfifo_enqueue() argument
30 if (likely(sch->q.qlen < sch->limit)) in pfifo_enqueue()
31 return qdisc_enqueue_tail(skb, sch); in pfifo_enqueue()
33 return qdisc_drop(skb, sch, to_free); in pfifo_enqueue()
36 static int pfifo_tail_enqueue(struct sk_buff *skb, struct Qdisc *sch, in pfifo_tail_enqueue() argument
41 if (likely(sch->q.qlen < sch->limit)) in pfifo_tail_enqueue()
[all …]
Dsch_codel.c71 struct Qdisc *sch = ctx; in dequeue_func() local
72 struct sk_buff *skb = __qdisc_dequeue_head(&sch->q); in dequeue_func()
75 sch->qstats.backlog -= qdisc_pkt_len(skb); in dequeue_func()
83 struct Qdisc *sch = ctx; in drop_func() local
86 qdisc_qstats_drop(sch); in drop_func()
89 static struct sk_buff *codel_qdisc_dequeue(struct Qdisc *sch) in codel_qdisc_dequeue() argument
91 struct codel_sched_data *q = qdisc_priv(sch); in codel_qdisc_dequeue()
94 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->params, &q->vars, in codel_qdisc_dequeue()
101 if (q->stats.drop_count && sch->q.qlen) { in codel_qdisc_dequeue()
102 qdisc_tree_reduce_backlog(sch, q->stats.drop_count, q->stats.drop_len); in codel_qdisc_dequeue()
[all …]
Dsch_multiq.c30 multiq_classify(struct sk_buff *skb, struct Qdisc *sch, int *qerr) in multiq_classify() argument
32 struct multiq_sched_data *q = qdisc_priv(sch); in multiq_classify()
60 multiq_enqueue(struct sk_buff *skb, struct Qdisc *sch, in multiq_enqueue() argument
66 qdisc = multiq_classify(skb, sch, &ret); in multiq_enqueue()
71 qdisc_qstats_drop(sch); in multiq_enqueue()
79 sch->q.qlen++; in multiq_enqueue()
83 qdisc_qstats_drop(sch); in multiq_enqueue()
87 static struct sk_buff *multiq_dequeue(struct Qdisc *sch) in multiq_dequeue() argument
89 struct multiq_sched_data *q = qdisc_priv(sch); in multiq_dequeue()
104 netdev_get_tx_queue(qdisc_dev(sch), q->curband))) { in multiq_dequeue()
[all …]
Dsch_drr.c39 static struct drr_class *drr_find_class(struct Qdisc *sch, u32 classid) in drr_find_class() argument
41 struct drr_sched *q = qdisc_priv(sch); in drr_find_class()
54 static int drr_change_class(struct Qdisc *sch, u32 classid, u32 parentid, in drr_change_class() argument
58 struct drr_sched *q = qdisc_priv(sch); in drr_change_class()
82 quantum = psched_mtu(qdisc_dev(sch)); in drr_change_class()
89 qdisc_root_sleeping_running(sch), in drr_change_class()
97 sch_tree_lock(sch); in drr_change_class()
100 sch_tree_unlock(sch); in drr_change_class()
111 cl->qdisc = qdisc_create_dflt(sch->dev_queue, in drr_change_class()
122 qdisc_root_sleeping_running(sch), in drr_change_class()
[all …]
Dsch_etf.c75 static bool is_packet_valid(struct Qdisc *sch, struct sk_buff *nskb) in is_packet_valid() argument
77 struct etf_sched_data *q = qdisc_priv(sch); in is_packet_valid()
108 static struct sk_buff *etf_peek_timesortedlist(struct Qdisc *sch) in etf_peek_timesortedlist() argument
110 struct etf_sched_data *q = qdisc_priv(sch); in etf_peek_timesortedlist()
120 static void reset_watchdog(struct Qdisc *sch) in reset_watchdog() argument
122 struct etf_sched_data *q = qdisc_priv(sch); in reset_watchdog()
123 struct sk_buff *skb = etf_peek_timesortedlist(sch); in reset_watchdog()
162 static int etf_enqueue_timesortedlist(struct sk_buff *nskb, struct Qdisc *sch, in etf_enqueue_timesortedlist() argument
165 struct etf_sched_data *q = qdisc_priv(sch); in etf_enqueue_timesortedlist()
170 if (!is_packet_valid(sch, nskb)) { in etf_enqueue_timesortedlist()
[all …]
Dsch_fq_codel.c77 static unsigned int fq_codel_classify(struct sk_buff *skb, struct Qdisc *sch, in fq_codel_classify() argument
80 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_classify()
85 if (TC_H_MAJ(skb->priority) == sch->handle && in fq_codel_classify()
138 static unsigned int fq_codel_drop(struct Qdisc *sch, unsigned int max_packets, in fq_codel_drop() argument
141 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_drop()
179 sch->qstats.drops += i; in fq_codel_drop()
180 sch->qstats.backlog -= len; in fq_codel_drop()
181 sch->q.qlen -= i; in fq_codel_drop()
185 static int fq_codel_enqueue(struct sk_buff *skb, struct Qdisc *sch, in fq_codel_enqueue() argument
188 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_enqueue()
[all …]
Dsch_skbprio.c68 static int skbprio_enqueue(struct sk_buff *skb, struct Qdisc *sch, in skbprio_enqueue() argument
72 struct skbprio_sched_data *q = qdisc_priv(sch); in skbprio_enqueue()
82 if (sch->q.qlen < sch->limit) { in skbprio_enqueue()
84 qdisc_qstats_backlog_inc(sch, skb); in skbprio_enqueue()
94 sch->q.qlen++; in skbprio_enqueue()
103 return qdisc_drop(skb, sch, to_free); in skbprio_enqueue()
107 qdisc_qstats_backlog_inc(sch, skb); in skbprio_enqueue()
114 qdisc_qstats_backlog_dec(sch, to_drop); in skbprio_enqueue()
115 qdisc_drop(to_drop, sch, to_free); in skbprio_enqueue()
125 BUG_ON(sch->q.qlen != 1); in skbprio_enqueue()
[all …]
Dsch_cbs.c80 int (*enqueue)(struct sk_buff *skb, struct Qdisc *sch,
82 struct sk_buff *(*dequeue)(struct Qdisc *sch);
87 static int cbs_child_enqueue(struct sk_buff *skb, struct Qdisc *sch, in cbs_child_enqueue() argument
98 sch->qstats.backlog += len; in cbs_child_enqueue()
99 sch->q.qlen++; in cbs_child_enqueue()
104 static int cbs_enqueue_offload(struct sk_buff *skb, struct Qdisc *sch, in cbs_enqueue_offload() argument
107 struct cbs_sched_data *q = qdisc_priv(sch); in cbs_enqueue_offload()
110 return cbs_child_enqueue(skb, sch, qdisc, to_free); in cbs_enqueue_offload()
113 static int cbs_enqueue_soft(struct sk_buff *skb, struct Qdisc *sch, in cbs_enqueue_soft() argument
116 struct cbs_sched_data *q = qdisc_priv(sch); in cbs_enqueue_soft()
[all …]
Dsch_pie.c68 struct Qdisc *sch; member
92 static bool drop_early(struct Qdisc *sch, u32 packet_size) in drop_early() argument
94 struct pie_sched_data *q = qdisc_priv(sch); in drop_early()
97 u32 mtu = psched_mtu(qdisc_dev(sch)); in drop_early()
113 if (sch->qstats.backlog < 2 * mtu) in drop_early()
151 static int pie_qdisc_enqueue(struct sk_buff *skb, struct Qdisc *sch, in pie_qdisc_enqueue() argument
154 struct pie_sched_data *q = qdisc_priv(sch); in pie_qdisc_enqueue()
157 if (unlikely(qdisc_qlen(sch) >= sch->limit)) { in pie_qdisc_enqueue()
162 if (!drop_early(sch, skb->len)) { in pie_qdisc_enqueue()
176 if (qdisc_qlen(sch) > q->stats.maxq) in pie_qdisc_enqueue()
[all …]
Dsch_tbf.c143 static int tbf_segment(struct sk_buff *skb, struct Qdisc *sch, in tbf_segment() argument
146 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_segment()
155 return qdisc_drop(skb, sch, to_free); in tbf_segment()
166 qdisc_qstats_drop(sch); in tbf_segment()
172 sch->q.qlen += nb; in tbf_segment()
174 qdisc_tree_reduce_backlog(sch, 1 - nb, prev_len - len); in tbf_segment()
179 static int tbf_enqueue(struct sk_buff *skb, struct Qdisc *sch, in tbf_enqueue() argument
182 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_enqueue()
189 return tbf_segment(skb, sch, to_free); in tbf_enqueue()
190 return qdisc_drop(skb, sch, to_free); in tbf_enqueue()
[all …]
Dsch_choke.c114 static void choke_drop_by_idx(struct Qdisc *sch, unsigned int idx, in choke_drop_by_idx() argument
117 struct choke_sched_data *q = qdisc_priv(sch); in choke_drop_by_idx()
127 qdisc_qstats_backlog_dec(sch, skb); in choke_drop_by_idx()
128 qdisc_tree_reduce_backlog(sch, 1, qdisc_pkt_len(skb)); in choke_drop_by_idx()
129 qdisc_drop(skb, sch, to_free); in choke_drop_by_idx()
130 --sch->q.qlen; in choke_drop_by_idx()
219 static int choke_enqueue(struct sk_buff *skb, struct Qdisc *sch, in choke_enqueue() argument
222 struct choke_sched_data *q = qdisc_priv(sch); in choke_enqueue()
227 q->vars.qavg = red_calc_qavg(p, &q->vars, sch->q.qlen); in choke_enqueue()
240 choke_drop_by_idx(sch, idx, to_free); in choke_enqueue()
[all …]
Dsch_mqprio.c30 static void mqprio_destroy(struct Qdisc *sch) in mqprio_destroy() argument
32 struct net_device *dev = qdisc_dev(sch); in mqprio_destroy()
33 struct mqprio_sched *priv = qdisc_priv(sch); in mqprio_destroy()
133 static int mqprio_parse_nlattr(struct Qdisc *sch, struct tc_mqprio_qopt *qopt, in mqprio_parse_nlattr() argument
137 struct mqprio_sched *priv = qdisc_priv(sch); in mqprio_parse_nlattr()
224 static int mqprio_init(struct Qdisc *sch, struct nlattr *opt, in mqprio_init() argument
227 struct net_device *dev = qdisc_dev(sch); in mqprio_init()
228 struct mqprio_sched *priv = qdisc_priv(sch); in mqprio_init()
238 if (sch->parent != TC_H_ROOT) in mqprio_init()
257 err = mqprio_parse_nlattr(sch, qopt, opt, extack); in mqprio_init()
[all …]
Dsch_gred.c91 static inline int gred_wred_mode_check(struct Qdisc *sch) in gred_wred_mode_check() argument
93 struct gred_sched *table = qdisc_priv(sch); in gred_wred_mode_check()
114 struct Qdisc *sch) in gred_backlog() argument
117 return sch->qstats.backlog; in gred_backlog()
164 static int gred_enqueue(struct sk_buff *skb, struct Qdisc *sch, in gred_enqueue() argument
168 struct gred_sched *t = qdisc_priv(sch); in gred_enqueue()
181 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= in gred_enqueue()
182 sch->limit)) in gred_enqueue()
183 return qdisc_enqueue_tail(skb, sch); in gred_enqueue()
213 gred_backlog(t, q, sch)); in gred_enqueue()
[all …]
Dsch_sfq.c144 struct Qdisc *sch; member
163 static unsigned int sfq_classify(struct sk_buff *skb, struct Qdisc *sch, in sfq_classify() argument
166 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_classify()
171 if (TC_H_MAJ(skb->priority) == sch->handle && in sfq_classify()
293 static unsigned int sfq_drop(struct Qdisc *sch, struct sk_buff **to_free) in sfq_drop() argument
295 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_drop()
310 sch->q.qlen--; in sfq_drop()
311 qdisc_qstats_backlog_dec(sch, skb); in sfq_drop()
312 qdisc_drop(skb, sch, to_free); in sfq_drop()
346 sfq_enqueue(struct sk_buff *skb, struct Qdisc *sch, struct sk_buff **to_free) in sfq_enqueue() argument
[all …]
Dsch_sfb.c279 static int sfb_enqueue(struct sk_buff *skb, struct Qdisc *sch, in sfb_enqueue() argument
283 struct sfb_sched_data *q = qdisc_priv(sch); in sfb_enqueue()
295 if (unlikely(sch->q.qlen >= q->limit)) { in sfb_enqueue()
296 qdisc_qstats_overlimit(sch); in sfb_enqueue()
349 qdisc_qstats_overlimit(sch); in sfb_enqueue()
375 qdisc_qstats_overlimit(sch); in sfb_enqueue()
407 sch->qstats.backlog += len; in sfb_enqueue()
408 sch->q.qlen++; in sfb_enqueue()
412 qdisc_qstats_drop(sch); in sfb_enqueue()
417 qdisc_drop(skb, sch, to_free); in sfb_enqueue()
[all …]
Dsch_hhf.c247 static enum wdrr_bucket_idx hhf_classify(struct sk_buff *skb, struct Qdisc *sch) in hhf_classify() argument
249 struct hhf_sched_data *q = qdisc_priv(sch); in hhf_classify()
349 static unsigned int hhf_drop(struct Qdisc *sch, struct sk_buff **to_free) in hhf_drop() argument
351 struct hhf_sched_data *q = qdisc_priv(sch); in hhf_drop()
362 sch->q.qlen--; in hhf_drop()
363 qdisc_qstats_backlog_dec(sch, skb); in hhf_drop()
364 qdisc_drop(skb, sch, to_free); in hhf_drop()
371 static int hhf_enqueue(struct sk_buff *skb, struct Qdisc *sch, in hhf_enqueue() argument
374 struct hhf_sched_data *q = qdisc_priv(sch); in hhf_enqueue()
379 idx = hhf_classify(skb, sch); in hhf_enqueue()
[all …]
Dsch_hfsc.c832 qdisc_peek_len(struct Qdisc *sch) in qdisc_peek_len() argument
837 skb = sch->ops->peek(sch); in qdisc_peek_len()
839 qdisc_warn_nonwc("qdisc_peek_len", sch); in qdisc_peek_len()
864 hfsc_find_class(u32 classid, struct Qdisc *sch) in hfsc_find_class() argument
866 struct hfsc_sched *q = qdisc_priv(sch); in hfsc_find_class()
921 hfsc_change_class(struct Qdisc *sch, u32 classid, u32 parentid, in hfsc_change_class() argument
925 struct hfsc_sched *q = qdisc_priv(sch); in hfsc_change_class()
976 qdisc_root_sleeping_running(sch), in hfsc_change_class()
982 sch_tree_lock(sch); in hfsc_change_class()
1009 sch_tree_unlock(sch); in hfsc_change_class()
[all …]
Dsch_plug.c87 static int plug_enqueue(struct sk_buff *skb, struct Qdisc *sch, in plug_enqueue() argument
90 struct plug_sched_data *q = qdisc_priv(sch); in plug_enqueue()
92 if (likely(sch->qstats.backlog + skb->len <= q->limit)) { in plug_enqueue()
95 return qdisc_enqueue_tail(skb, sch); in plug_enqueue()
98 return qdisc_drop(skb, sch, to_free); in plug_enqueue()
101 static struct sk_buff *plug_dequeue(struct Qdisc *sch) in plug_dequeue() argument
103 struct plug_sched_data *q = qdisc_priv(sch); in plug_dequeue()
119 return qdisc_dequeue_head(sch); in plug_dequeue()
122 static int plug_init(struct Qdisc *sch, struct nlattr *opt, in plug_init() argument
125 struct plug_sched_data *q = qdisc_priv(sch); in plug_init()
[all …]
Dsch_taprio.c80 struct sk_buff *(*dequeue)(struct Qdisc *sch);
81 struct sk_buff *(*peek)(struct Qdisc *sch);
190 struct Qdisc *sch, in find_entry_to_transmit() argument
201 struct taprio_sched *q = qdisc_priv(sch); in find_entry_to_transmit()
202 struct net_device *dev = qdisc_dev(sch); in find_entry_to_transmit()
266 static bool is_valid_interval(struct sk_buff *skb, struct Qdisc *sch) in is_valid_interval() argument
268 struct taprio_sched *q = qdisc_priv(sch); in is_valid_interval()
277 entry = find_entry_to_transmit(skb, sch, sched, admin, skb->tstamp, in is_valid_interval()
347 static long get_packet_txtime(struct sk_buff *skb, struct Qdisc *sch) in get_packet_txtime() argument
350 struct taprio_sched *q = qdisc_priv(sch); in get_packet_txtime()
[all …]
Dsch_fq.c360 static void fq_erase_head(struct Qdisc *sch, struct fq_flow *flow, in fq_erase_head() argument
367 skb->dev = qdisc_dev(sch); in fq_erase_head()
372 static struct sk_buff *fq_dequeue_head(struct Qdisc *sch, struct fq_flow *flow) in fq_dequeue_head() argument
377 fq_erase_head(sch, flow, skb); in fq_dequeue_head()
380 qdisc_qstats_backlog_dec(sch, skb); in fq_dequeue_head()
381 sch->q.qlen--; in fq_dequeue_head()
420 static int fq_enqueue(struct sk_buff *skb, struct Qdisc *sch, in fq_enqueue() argument
423 struct fq_sched_data *q = qdisc_priv(sch); in fq_enqueue()
426 if (unlikely(sch->q.qlen >= sch->limit)) in fq_enqueue()
427 return qdisc_drop(skb, sch, to_free); in fq_enqueue()
[all …]
Dsch_qfq.c206 static struct qfq_class *qfq_find_class(struct Qdisc *sch, u32 classid) in qfq_find_class() argument
208 struct qfq_sched *q = qdisc_priv(sch); in qfq_find_class()
374 static int qfq_change_agg(struct Qdisc *sch, struct qfq_class *cl, u32 weight, in qfq_change_agg() argument
377 struct qfq_sched *q = qdisc_priv(sch); in qfq_change_agg()
397 static int qfq_change_class(struct Qdisc *sch, u32 classid, u32 parentid, in qfq_change_class() argument
401 struct qfq_sched *q = qdisc_priv(sch); in qfq_change_class()
432 lmax = psched_mtu(qdisc_dev(sch)); in qfq_change_class()
460 qdisc_root_sleeping_running(sch), in qfq_change_class()
477 cl->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in qfq_change_class()
486 qdisc_root_sleeping_running(sch), in qfq_change_class()
[all …]

12