Home
last modified time | relevance | path

Searched refs:backlog (Results 1 – 25 of 193) sorted by relevance

12345678

/kernel/linux/linux-5.10/drivers/crypto/ccp/
Dccp-crypto-main.c55 struct list_head *backlog; member
97 struct ccp_crypto_cmd *crypto_cmd, struct ccp_crypto_cmd **backlog) in ccp_crypto_cmd_complete() argument
102 *backlog = NULL; in ccp_crypto_cmd_complete()
121 if (req_queue.backlog != &req_queue.cmds) { in ccp_crypto_cmd_complete()
123 if (req_queue.backlog == &crypto_cmd->entry) in ccp_crypto_cmd_complete()
124 req_queue.backlog = crypto_cmd->entry.next; in ccp_crypto_cmd_complete()
126 *backlog = container_of(req_queue.backlog, in ccp_crypto_cmd_complete()
128 req_queue.backlog = req_queue.backlog->next; in ccp_crypto_cmd_complete()
131 if (req_queue.backlog == &crypto_cmd->entry) in ccp_crypto_cmd_complete()
132 req_queue.backlog = crypto_cmd->entry.next; in ccp_crypto_cmd_complete()
[all …]
Dccp-dev.c315 list_add_tail(&cmd->entry, &ccp->backlog); in ccp_enqueue_cmd()
378 struct ccp_cmd *backlog = NULL; in ccp_dequeue_cmd() local
403 if (!list_empty(&ccp->backlog)) { in ccp_dequeue_cmd()
404 backlog = list_first_entry(&ccp->backlog, struct ccp_cmd, in ccp_dequeue_cmd()
406 list_del(&backlog->entry); in ccp_dequeue_cmd()
411 if (backlog) { in ccp_dequeue_cmd()
412 INIT_WORK(&backlog->work, ccp_do_cmd_backlog); in ccp_dequeue_cmd()
413 schedule_work(&backlog->work); in ccp_dequeue_cmd()
488 INIT_LIST_HEAD(&ccp->backlog); in ccp_alloc_struct()
/kernel/linux/linux-5.10/tools/perf/ui/tui/
Dhelpline.c34 static int backlog; in tui_helpline__show() local
37 ret = vscnprintf(ui_helpline__last_msg + backlog, in tui_helpline__show()
38 sizeof(ui_helpline__last_msg) - backlog, format, ap); in tui_helpline__show()
39 backlog += ret; in tui_helpline__show()
43 if (ui_helpline__last_msg[backlog - 1] == '\n') { in tui_helpline__show()
46 backlog = 0; in tui_helpline__show()
/kernel/linux/linux-5.10/tools/perf/ui/gtk/
Dhelpline.c32 static int backlog; in gtk_helpline_show() local
34 ret = vscnprintf(ui_helpline__current + backlog, in gtk_helpline_show()
35 sizeof(ui_helpline__current) - backlog, fmt, ap); in gtk_helpline_show()
36 backlog += ret; in gtk_helpline_show()
40 if (ptr && (ptr - ui_helpline__current) <= backlog) { in gtk_helpline_show()
43 backlog = 0; in gtk_helpline_show()
/kernel/linux/linux-5.10/tools/testing/selftests/net/forwarding/
Dsch_red.sh164 qdisc_stats_get $swp3 11: .backlog
258 local backlog
266 backlog=$(build_backlog $((2 * limit / 3)) udp)
269 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected == 0."
276 backlog=$(build_backlog $((3 * limit / 2)) tcp tos=0x01)
279 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected >= 95."
332 local backlog
342 backlog=$(build_backlog $((2 * limit / 3)) tcp tos=0x01)
345 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected == 0."
350 backlog=$(build_backlog $((3 * limit / 2)) tcp tos=0x01)
[all …]
/kernel/linux/linux-5.10/tools/testing/selftests/drivers/net/mlxsw/
Dsch_red_core.sh316 qdisc_stats_get $swp3 $(get_qdisc_handle $vlan) .backlog
409 local backlog
417 backlog=$(build_backlog $vlan $((2 * limit / 3)) udp)
420 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected == 0."
427 backlog=$(build_backlog $vlan $((3 * limit / 2)) tcp tos=0x01)
430 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected >= 95."
486 local backlog
496 backlog=$(build_backlog $vlan $((2 * limit / 3)) tcp tos=0x01)
499 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected == 0."
504 backlog=$(build_backlog $vlan $((3 * limit / 2)) tcp tos=0x01)
[all …]
/kernel/linux/linux-5.10/include/net/
Dfq_impl.h22 flow->backlog -= skb->len; in fq_adjust_removal()
23 fq->backlog--; in fq_adjust_removal()
31 if (flow->backlog == 0) { in fq_rejigger_backlog()
37 if (i->backlog < flow->backlog) in fq_rejigger_backlog()
150 if (i->backlog > flow->backlog) in fq_recalc_backlog()
170 flow->backlog += skb->len; in fq_tin_enqueue()
174 fq->backlog++; in fq_tin_enqueue()
186 while (fq->backlog > fq->limit || oom) { in fq_tin_enqueue()
264 WARN_ON_ONCE(flow->backlog); in fq_flow_reset()
Dcodel_impl.h107 u32 *backlog, in codel_should_drop() argument
125 *backlog <= params->mtu) { in codel_should_drop()
143 u32 *backlog, in codel_dequeue() argument
162 skb_len_func, skb_time_func, backlog, now); in codel_dequeue()
198 backlog, now)) { in codel_dequeue()
223 skb_time_func, backlog, now); in codel_dequeue()
Dpie.h127 struct pie_vars *vars, u32 backlog, u32 packet_size);
130 struct pie_vars *vars, u32 backlog);
133 u32 backlog);
Dsch_generic.h883 sch->qstats.backlog -= qdisc_pkt_len(skb); in qdisc_qstats_backlog_dec()
889 this_cpu_sub(sch->cpu_qstats->backlog, qdisc_pkt_len(skb)); in qdisc_qstats_cpu_backlog_dec()
895 sch->qstats.backlog += qdisc_pkt_len(skb); in qdisc_qstats_backlog_inc()
901 this_cpu_add(sch->cpu_qstats->backlog, qdisc_pkt_len(skb)); in qdisc_qstats_cpu_backlog_inc()
957 __u32 *backlog) in qdisc_qstats_qlen_backlog() argument
964 *backlog = qstats.backlog; in qdisc_qstats_qlen_backlog()
969 __u32 qlen, backlog; in qdisc_tree_flush_backlog() local
971 qdisc_qstats_qlen_backlog(sch, &qlen, &backlog); in qdisc_tree_flush_backlog()
972 qdisc_tree_reduce_backlog(sch, qlen, backlog); in qdisc_tree_flush_backlog()
977 __u32 qlen, backlog; in qdisc_purge_queue() local
[all …]
Dfq.h34 u32 backlog; member
76 u32 backlog; member
Dred.h340 unsigned int backlog) in red_calc_qavg_no_idle_time() argument
351 return v->qavg + (backlog - (v->qavg >> p->Wlog)); in red_calc_qavg_no_idle_time()
356 unsigned int backlog) in red_calc_qavg() argument
359 return red_calc_qavg_no_idle_time(p, v, backlog); in red_calc_qavg()
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlxsw/
Dspectrum_qdisc.c65 u64 backlog; member
264 return xstats->backlog[tclass_num] + in mlxsw_sp_xstats_backlog()
265 xstats->backlog[tclass_num + 8]; in mlxsw_sp_xstats_backlog()
318 u64 drops, u64 backlog, in mlxsw_sp_qdisc_update_stats() argument
326 backlog -= stats_base->backlog; in mlxsw_sp_qdisc_update_stats()
330 stats_ptr->qstats->backlog += mlxsw_sp_cells_bytes(mlxsw_sp, backlog); in mlxsw_sp_qdisc_update_stats()
332 stats_base->backlog += backlog; in mlxsw_sp_qdisc_update_stats()
345 u64 backlog = 0; in mlxsw_sp_qdisc_get_tc_stats() local
350 &drops, &backlog); in mlxsw_sp_qdisc_get_tc_stats()
352 tx_bytes, tx_packets, drops, backlog, in mlxsw_sp_qdisc_get_tc_stats()
[all …]
/kernel/linux/linux-5.10/net/sched/
Dsch_pie.c34 struct pie_vars *vars, u32 backlog, u32 packet_size) in pie_drop_early() argument
54 if (backlog < 2 * mtu) in pie_drop_early()
96 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, in pie_qdisc_enqueue()
209 struct pie_vars *vars, u32 backlog) in pie_process_dequeue() argument
225 if (backlog == 0) in pie_process_dequeue()
238 if (backlog >= QUEUE_THRESHOLD && vars->dq_count == DQCOUNT_INVALID) { in pie_process_dequeue()
277 if (backlog < QUEUE_THRESHOLD) { in pie_process_dequeue()
301 u32 backlog) in pie_calculate_probability() argument
316 qdelay = (backlog << PIE_SCALE) / vars->avg_dq_rate; in pie_calculate_probability()
327 if (qdelay == 0 && backlog != 0) in pie_calculate_probability()
[all …]
Dsch_gred.c39 u32 backlog; /* bytes on the virtualQ */ member
117 return sch->qstats.backlog; in gred_backlog()
119 return q->backlog; in gred_backlog()
181 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= in gred_enqueue()
247 q->backlog += qdisc_pkt_len(skb); in gred_enqueue()
275 q->backlog -= qdisc_pkt_len(skb); in gred_dequeue()
278 if (!sch->qstats.backlog) in gred_dequeue()
281 if (!q->backlog) in gred_dequeue()
306 q->backlog = 0; in gred_reset()
344 opt.set.tab[i].backlog = &q->backlog; in gred_offload()
[all …]
Dsch_fq_pie.c47 u32 backlog; member
163 sel_flow->backlog, skb->len)) { in fq_pie_qdisc_enqueue()
182 sch->qstats.backlog += pkt_len; in fq_pie_qdisc_enqueue()
190 sel_flow->backlog = 0; in fq_pie_qdisc_enqueue()
193 sel_flow->backlog += pkt_len; in fq_pie_qdisc_enqueue()
255 sch->qstats.backlog -= pkt_len; in fq_pie_qdisc_dequeue()
271 flow->backlog -= pkt_len; in fq_pie_qdisc_dequeue()
273 pie_process_dequeue(skb, &q->p_params, &flow->vars, flow->backlog); in fq_pie_qdisc_dequeue()
384 q->flows[idx].backlog); in fq_pie_timer()
Dsch_sfq.c109 unsigned int backlog; member
308 slot->backlog -= len; in sfq_drop()
374 slot->backlog = 0; /* should already be 0 anyway... */ in sfq_enqueue()
381 slot->backlog); in sfq_enqueue()
432 sch->qstats.backlog -= delta; in sfq_enqueue()
433 slot->backlog -= delta; in sfq_enqueue()
443 slot->backlog += qdisc_pkt_len(skb); in sfq_enqueue()
504 slot->backlog -= qdisc_pkt_len(skb); in sfq_dequeue()
556 slot->backlog = 0; in sfq_rehash()
587 slot->backlog); in sfq_rehash()
[all …]
/kernel/linux/linux-5.10/drivers/crypto/qce/
Dcore.c80 struct crypto_async_request *async_req, *backlog; in qce_handle_queue() local
95 backlog = crypto_get_backlog(&qce->queue); in qce_handle_queue()
105 if (backlog) { in qce_handle_queue()
107 backlog->complete(backlog, -EINPROGRESS); in qce_handle_queue()
/kernel/linux/linux-5.10/net/core/
Dgen_stats.c295 qstats->backlog += qcpu->backlog; in __gnet_stats_copy_queue_cpu()
311 qstats->backlog = q->backlog; in __gnet_stats_copy_queue()
347 d->tc_stats.backlog = qstats.backlog; in gnet_stats_copy_queue()
/kernel/linux/linux-5.10/net/sunrpc/
Dstats.c153 ktime_t backlog, execute, now; in rpc_count_iostats_metrics() local
169 backlog = 0; in rpc_count_iostats_metrics()
171 backlog = ktime_sub(req->rq_xtime, task->tk_start); in rpc_count_iostats_metrics()
172 op_metrics->om_queue = ktime_add(op_metrics->om_queue, backlog); in rpc_count_iostats_metrics()
184 trace_rpc_stats_latency(req->rq_task, backlog, req->rq_rtt, execute); in rpc_count_iostats_metrics()
/kernel/linux/linux-5.10/net/rxrpc/
Dcall_accept.c165 struct rxrpc_backlog *b = rx->backlog; in rxrpc_service_prealloc()
171 rx->backlog = b; in rxrpc_service_prealloc()
182 struct rxrpc_backlog *b = rx->backlog; in rxrpc_discard_prealloc()
188 rx->backlog = NULL; in rxrpc_discard_prealloc()
268 struct rxrpc_backlog *b = rx->backlog; in rxrpc_alloc_incoming_call()
450 struct rxrpc_backlog *b = rx->backlog; in rxrpc_user_charge_accept()
482 struct rxrpc_backlog *b = rx->backlog; in rxrpc_kernel_charge_accept()
/kernel/linux/linux-5.10/crypto/
Dcrypto_engine.c73 struct crypto_async_request *async_req, *backlog; in crypto_pump_requests() local
118 backlog = crypto_get_backlog(&engine->queue); in crypto_pump_requests()
131 if (backlog) in crypto_pump_requests()
132 backlog->complete(backlog, -EINPROGRESS); in crypto_pump_requests()
/kernel/linux/linux-5.10/drivers/crypto/marvell/cesa/
Dtdma.c136 struct crypto_async_request *backlog = NULL; in mv_cesa_tdma_process() local
149 &backlog); in mv_cesa_tdma_process()
170 if (backlog) in mv_cesa_tdma_process()
171 backlog->complete(backlog, -EINPROGRESS); in mv_cesa_tdma_process()
Dcesa.c39 struct crypto_async_request **backlog) in mv_cesa_dequeue_req_locked() argument
43 *backlog = crypto_get_backlog(&engine->queue); in mv_cesa_dequeue_req_locked()
54 struct crypto_async_request *req = NULL, *backlog = NULL; in mv_cesa_rearm_engine() local
60 req = mv_cesa_dequeue_req_locked(engine, &backlog); in mv_cesa_rearm_engine()
68 if (backlog) in mv_cesa_rearm_engine()
69 backlog->complete(backlog, -EINPROGRESS); in mv_cesa_rearm_engine()
/kernel/linux/linux-5.10/drivers/crypto/
Dmxs-dcp.c394 struct crypto_async_request *backlog; in dcp_chan_thread_aes() local
403 backlog = crypto_get_backlog(&sdcp->queue[chan]); in dcp_chan_thread_aes()
407 if (!backlog && !arq) { in dcp_chan_thread_aes()
414 if (backlog) in dcp_chan_thread_aes()
415 backlog->complete(backlog, -EINPROGRESS); in dcp_chan_thread_aes()
691 struct crypto_async_request *backlog; in dcp_chan_thread_sha() local
699 backlog = crypto_get_backlog(&sdcp->queue[chan]); in dcp_chan_thread_sha()
703 if (!backlog && !arq) { in dcp_chan_thread_sha()
710 if (backlog) in dcp_chan_thread_sha()
711 backlog->complete(backlog, -EINPROGRESS); in dcp_chan_thread_sha()

12345678