Home
last modified time | relevance | path

Searched refs:flows (Results 1 – 25 of 79) sorted by relevance

1234

/kernel/linux/linux-5.10/drivers/crypto/allwinner/sun8i-ss/
Dsun8i-ss-core.c74 ss->flows[flow].stat_req++; in sun8i_ss_run_task()
130 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_run_task()
131 ss->flows[flow].status = 0; in sun8i_ss_run_task()
136 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_run_task()
138 if (ss->flows[flow].status == 0) { in sun8i_ss_run_task()
157 ss->flows[flow].status = 1; in ss_irq_handler()
158 complete(&ss->flows[flow].complete); in ss_irq_handler()
422 seq_printf(seq, "Channel %d: nreq %lu\n", i, ss->flows[i].stat_req); in sun8i_ss_debugfs_show()
457 crypto_engine_exit(ss->flows[i].engine); in sun8i_ss_free_flows()
469 ss->flows = devm_kcalloc(ss->dev, MAXFLOW, sizeof(struct sun8i_ss_flow), in allocate_flows()
[all …]
Dsun8i-ss-prng.c131 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_prng_generate()
132 ss->flows[flow].status = 0; in sun8i_ss_prng_generate()
138 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_prng_generate()
140 if (ss->flows[flow].status == 0) { in sun8i_ss_prng_generate()
Dsun8i-ss-hash.c208 ss->flows[flow].stat_req++; in sun8i_ss_run_hash_task()
241 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_run_hash_task()
242 ss->flows[flow].status = 0; in sun8i_ss_run_hash_task()
247 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_run_hash_task()
249 if (ss->flows[flow].status == 0) { in sun8i_ss_run_hash_task()
310 engine = ss->flows[e].engine; in sun8i_ss_hash_digest()
344 result = ss->flows[rctx->flow].result; in sun8i_ss_hash_run()
345 pad = ss->flows[rctx->flow].pad; in sun8i_ss_hash_run()
Dsun8i-ss-cipher.c106 struct sun8i_ss_flow *sf = &ss->flows[rctx->flow]; in sun8i_ss_setup_ivs()
167 struct sun8i_ss_flow *sf = &ss->flows[rctx->flow]; in sun8i_ss_cipher()
337 engine = op->ss->flows[e].engine; in sun8i_ss_skdecrypt()
358 engine = op->ss->flows[e].engine; in sun8i_ss_skencrypt()
/kernel/linux/linux-5.10/include/net/
Dfq_impl.h125 flow = &fq->flows[idx]; in fq_flow_classify()
133 tin->flows++; in fq_flow_classify()
315 fq->flows = kvcalloc(fq->flows_cnt, sizeof(fq->flows[0]), GFP_KERNEL); in fq_init()
316 if (!fq->flows) in fq_init()
320 fq_flow_init(&fq->flows[i]); in fq_init()
331 fq_flow_reset(fq, &fq->flows[i], free_func); in fq_reset()
333 kvfree(fq->flows); in fq_reset()
334 fq->flows = NULL; in fq_reset()
Dfq.h54 u32 flows; member
68 struct fq_flow *flows; member
/kernel/linux/linux-5.10/samples/bpf/
Ddo_hbm_test.sh78 flows=1
140 -f=*|--flows=*)
141 flows="${i#*=}"
267 while [ $flow_cnt -le $flows ] ; do
309 while [ $flow_cnt -le $flows ] ; do
335 iperf3 -c $host -p $port -i 0 -P $flows -f m -t $dur > iperf.$id
355 while [ $flow_cnt -le $flows ] ; do
375 while [ $flow_cnt -le $flows ] ; do
/kernel/linux/linux-5.10/net/sched/
Dsch_fq_codel.c53 struct fq_codel_flow *flows; /* Flows table [flows_cnt] */ member
164 flow = &q->flows[idx]; in fq_codel_drop()
204 flow = &q->flows[idx]; in fq_codel_enqueue()
265 q->backlogs[flow - q->flows] -= qdisc_pkt_len(skb); in dequeue_func()
343 struct fq_codel_flow *flow = q->flows + i; in fq_codel_reset()
381 if (q->flows) in fq_codel_change()
452 kvfree(q->flows); in fq_codel_destroy()
484 if (!q->flows) { in fq_codel_init()
485 q->flows = kvcalloc(q->flows_cnt, in fq_codel_init()
488 if (!q->flows) { in fq_codel_init()
[all …]
Dsch_fq_pie.c57 struct fq_pie_flow *flows; member
149 sel_flow = &q->flows[idx]; in fq_pie_qdisc_enqueue()
307 if (q->flows) { in fq_pie_change()
389 pie_calculate_probability(&q->p_params, &q->flows[idx].vars, in fq_pie_timer()
390 q->flows[idx].backlog); in fq_pie_timer()
430 q->flows = kvcalloc(q->flows_cnt, sizeof(struct fq_pie_flow), in fq_pie_init()
432 if (!q->flows) { in fq_pie_init()
437 struct fq_pie_flow *flow = q->flows + idx; in fq_pie_init()
521 struct fq_pie_flow *flow = q->flows + idx; in fq_pie_reset()
539 kvfree(q->flows); in fq_pie_destroy()
Dsch_atm.c66 struct list_head flows; /* NB: "link" is also on this member
78 list_for_each_entry(flow, &p->flows, list) { in lookup_flow()
355 list_for_each_entry(flow, &p->flows, list) { in atm_tc_walk()
393 list_for_each_entry(flow, &p->flows, list) { in atm_tc_enqueue()
480 list_for_each_entry(flow, &p->flows, list) { in sch_atm_dequeue()
551 INIT_LIST_HEAD(&p->flows); in atm_tc_init()
553 list_add(&p->link.list, &p->flows); in atm_tc_init()
579 list_for_each_entry(flow, &p->flows, list) in atm_tc_reset()
589 list_for_each_entry(flow, &p->flows, list) { in atm_tc_destroy()
594 list_for_each_entry_safe(flow, tmp, &p->flows, list) { in atm_tc_destroy()
Dsch_fq.c120 u32 flows; member
254 q->flows -= fcnt; in fq_gc()
304 if (q->flows >= (2U << q->fq_trees_log) && in fq_classify()
305 q->inactive_flows > q->flows/2) in fq_classify()
359 q->flows++; in fq_classify()
692 q->flows = 0; in fq_reset()
737 q->flows -= fcnt; in fq_rehash()
1027 st.flows = q->flows; in fq_dump_stats()
Dsch_cake.c150 struct cake_flow flows[CAKE_QUEUES]; member
743 q->flows[reduced_hash].set)) { in cake_hash()
761 if (!q->flows[outer_hash + k].set) { in cake_hash()
776 if (!q->flows[outer_hash + k].set) { in cake_hash()
788 if (q->flows[outer_hash + k].set == CAKE_SET_BULK) { in cake_hash()
789 q->hosts[q->flows[reduced_hash].srchost].srchost_bulk_flow_count--; in cake_hash()
790 q->hosts[q->flows[reduced_hash].dsthost].dsthost_bulk_flow_count--; in cake_hash()
817 if (q->flows[reduced_hash].set == CAKE_SET_BULK) in cake_hash()
819 q->flows[reduced_hash].srchost = srchost_idx; in cake_hash()
840 if (q->flows[reduced_hash].set == CAKE_SET_BULK) in cake_hash()
[all …]
/kernel/linux/linux-5.10/drivers/dma/ti/
Dk3-udma-glue.c79 struct k3_udma_glue_rx_flow *flows; member
533 struct k3_udma_glue_rx_flow *flow = &rx_chn->flows[flow_num]; in k3_udma_glue_release_rx_flow()
553 struct k3_udma_glue_rx_flow *flow = &rx_chn->flows[flow_idx]; in k3_udma_glue_cfg_rx_flow()
786 rx_chn->flows = devm_kcalloc(dev, rx_chn->flow_num, in k3_udma_glue_request_rx_chn_priv()
787 sizeof(*rx_chn->flows), GFP_KERNEL); in k3_udma_glue_request_rx_chn_priv()
788 if (!rx_chn->flows) { in k3_udma_glue_request_rx_chn_priv()
798 rx_chn->flows[i].udma_rflow_id = rx_chn->flow_id_base + i; in k3_udma_glue_request_rx_chn_priv()
880 rx_chn->flows = devm_kcalloc(dev, rx_chn->flow_num, in k3_udma_glue_request_remote_rx_chn()
881 sizeof(*rx_chn->flows), GFP_KERNEL); in k3_udma_glue_request_remote_rx_chn()
882 if (!rx_chn->flows) { in k3_udma_glue_request_remote_rx_chn()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
Dtid_rdma.c752 u32 generation = rcd->flows[flow_idx].generation; in kern_setup_hw_flow()
770 rcd->flows[flow_idx].generation = in kern_clear_hw_flow()
771 kern_flow_generation_next(rcd->flows[flow_idx].generation); in kern_clear_hw_flow()
799 rcd->flows[fs->index].generation = fs->generation; in hfi1_kern_setup_hw_flow()
848 rcd->flows[i].generation = mask_generation(prandom_u32()); in hfi1_kern_init_ctxt_generations()
1458 struct tid_rdma_flow *flow = &req->flows[req->setup_head]; in hfi1_kern_exp_rcv_setup()
1549 struct tid_rdma_flow *flow = &req->flows[req->clear_tail]; in hfi1_kern_exp_rcv_clear()
1605 kfree(req->flows); in hfi1_kern_exp_rcv_free_flows()
1606 req->flows = NULL; in hfi1_kern_exp_rcv_free_flows()
1627 struct tid_rdma_flow *flows; in hfi1_kern_exp_rcv_alloc_flows() local
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/
Den_rep.h164 struct list_head flows; member
183 struct list_head flows; member
Deswitch_offloads.c925 struct mlx5_flow_handle **flows; in esw_add_fdb_peer_miss_rules() local
939 flows = kvzalloc(nvports * sizeof(*flows), GFP_KERNEL); in esw_add_fdb_peer_miss_rules()
940 if (!flows) { in esw_add_fdb_peer_miss_rules()
959 flows[MLX5_VPORT_PF] = flow; in esw_add_fdb_peer_miss_rules()
970 flows[mlx5_eswitch_ecpf_idx(esw)] = flow; in esw_add_fdb_peer_miss_rules()
984 flows[i] = flow; in esw_add_fdb_peer_miss_rules()
987 esw->fdb_table.offloads.peer_miss_rules = flows; in esw_add_fdb_peer_miss_rules()
995 mlx5_del_flow_rules(flows[i]); in esw_add_fdb_peer_miss_rules()
998 mlx5_del_flow_rules(flows[mlx5_eswitch_ecpf_idx(esw)]); in esw_add_fdb_peer_miss_rules()
1001 mlx5_del_flow_rules(flows[MLX5_VPORT_PF]); in esw_add_fdb_peer_miss_rules()
[all …]
/kernel/linux/linux-5.10/Documentation/userspace-api/media/mediactl/
Dmedia-controller-model.rst26 by an entity flows from the entity's output to one or more entity
31 pads, either on the same entity or on different entities. Data flows
/kernel/linux/linux-5.10/Documentation/admin-guide/pm/
Dsystem-wide.rst11 suspend-flows
Dsuspend-flows.rst25 The kernel code flows associated with the suspend and resume transitions for
27 significant differences between the :ref:`suspend-to-idle <s2idle>` code flows
28 and the code flows related to the :ref:`suspend-to-RAM <s2ram>` and
35 available. Apart from that, the suspend and resume code flows for these sleep
/kernel/linux/linux-5.10/net/core/
Dpktgen.c399 struct flow_state *flows; member
2202 return !!(pkt_dev->flows[flow].flags & F_INIT); in f_seen()
2210 if (pkt_dev->flows[flow].count >= pkt_dev->lflow) { in f_pick()
2212 pkt_dev->flows[flow].count = 0; in f_pick()
2213 pkt_dev->flows[flow].flags = 0; in f_pick()
2222 if (pkt_dev->flows[flow].count > pkt_dev->lflow) { in f_pick()
2223 pkt_dev->flows[flow].count = 0; in f_pick()
2224 pkt_dev->flows[flow].flags = 0; in f_pick()
2239 struct xfrm_state *x = pkt_dev->flows[flow].x; in get_ipsec_sa()
2258 pkt_dev->flows[flow].x = x; in get_ipsec_sa()
[all …]
/kernel/linux/linux-5.10/Documentation/networking/
Dscaling.rst31 of logical flows. Packets for each flow are steered to a separate receive
188 to the same CPU is CPU load imbalance if flows vary in packet rate.
194 Flow Limit is an optional RPS feature that prioritizes small flows
195 during CPU contention by dropping packets from large flows slightly
196 ahead of those from small flows. It is active only when an RPS or RFS
202 new packet is dropped. Packets from other flows are still only
206 even large flows maintain connectivity.
224 identification of large flows and fewer false positives. The default
261 flows to the CPUs where those flows are being processed. The flow hash
266 same CPU. Indeed, with many flows and few CPUs, it is very likely that
[all …]
Dopenvswitch.rst16 table" that userspace populates with "flows" that map from keys based
104 A wildcarded flow can represent a group of exact match flows. Each '1' bit
108 by reduce the number of new flows need to be processed by the user space program.
120 two possible approaches: reactively install flows as they miss the kernel
130 The behavior when using overlapping wildcarded flows is undefined. It is the
133 performs best-effort detection of overlapping wildcarded flows and may reject
146 future operations. The kernel is not required to index flows by the original
Dpktgen.rst97 flows: 0 flowlen: 0
112 flows: 0
269 pktgen.conf-1-1-flows # 1 CPU 1 dev multiple flows.
287 pgset "flows 1"
381 flows
/kernel/linux/linux-5.10/Documentation/admin-guide/blockdev/drbd/
Dfigures.rst5 Data flows that Relate some functions, and write packets
/kernel/linux/linux-5.10/net/dccp/ccids/
DKconfig20 be reasonably fair when competing for bandwidth with TCP-like flows,

1234