| /drivers/crypto/allwinner/sun8i-ss/ |
| D | sun8i-ss-core.c | 76 ss->flows[flow].stat_req++; in sun8i_ss_run_task() 132 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_run_task() 133 ss->flows[flow].status = 0; in sun8i_ss_run_task() 138 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_run_task() 140 if (ss->flows[flow].status == 0) { in sun8i_ss_run_task() 159 ss->flows[flow].status = 1; in ss_irq_handler() 160 complete(&ss->flows[flow].complete); in ss_irq_handler() 478 ss->flows[i].stat_req); in sun8i_ss_debugfs_show() 536 crypto_engine_exit(ss->flows[i].engine); in sun8i_ss_free_flows() 548 ss->flows = devm_kcalloc(ss->dev, MAXFLOW, sizeof(struct sun8i_ss_flow), in allocate_flows() [all …]
|
| D | sun8i-ss-prng.c | 134 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_prng_generate() 135 ss->flows[flow].status = 0; in sun8i_ss_prng_generate() 141 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_prng_generate() 143 if (ss->flows[flow].status == 0) { in sun8i_ss_prng_generate()
|
| D | sun8i-ss-hash.c | 290 ss->flows[flow].stat_req++; in sun8i_ss_run_hash_task() 323 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_run_hash_task() 324 ss->flows[flow].status = 0; in sun8i_ss_run_hash_task() 329 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_run_hash_task() 331 if (ss->flows[flow].status == 0) { in sun8i_ss_run_hash_task() 407 engine = ss->flows[e].engine; in sun8i_ss_hash_digest() 505 result = ss->flows[rctx->flow].result; in sun8i_ss_hash_run() 506 pad = ss->flows[rctx->flow].pad; in sun8i_ss_hash_run()
|
| D | sun8i-ss-cipher.c | 130 struct sun8i_ss_flow *sf = &ss->flows[rctx->flow]; in sun8i_ss_setup_ivs() 191 struct sun8i_ss_flow *sf = &ss->flows[rctx->flow]; in sun8i_ss_cipher() 360 engine = op->ss->flows[e].engine; in sun8i_ss_skdecrypt() 381 engine = op->ss->flows[e].engine; in sun8i_ss_skencrypt()
|
| D | sun8i-ss.h | 163 struct sun8i_ss_flow *flows; member
|
| /drivers/dma/ti/ |
| D | k3-udma-glue.c | 84 struct k3_udma_glue_rx_flow *flows; member 715 struct k3_udma_glue_rx_flow *flow = &rx_chn->flows[flow_num]; in k3_udma_glue_release_rx_flow() 735 struct k3_udma_glue_rx_flow *flow = &rx_chn->flows[flow_idx]; in k3_udma_glue_cfg_rx_flow() 1032 rx_chn->flows = devm_kcalloc(dev, rx_chn->flow_num, in k3_udma_glue_request_rx_chn_priv() 1033 sizeof(*rx_chn->flows), GFP_KERNEL); in k3_udma_glue_request_rx_chn_priv() 1034 if (!rx_chn->flows) { in k3_udma_glue_request_rx_chn_priv() 1044 rx_chn->flows[i].udma_rflow_id = rx_chn->flow_id_base + i; in k3_udma_glue_request_rx_chn_priv() 1084 rx_chn->flows = devm_kcalloc(dev, rx_chn->flow_num, in k3_udma_glue_request_remote_rx_chn_common() 1085 sizeof(*rx_chn->flows), GFP_KERNEL); in k3_udma_glue_request_remote_rx_chn_common() 1086 if (!rx_chn->flows) in k3_udma_glue_request_remote_rx_chn_common() [all …]
|
| /drivers/media/platform/amphion/ |
| D | vpu_dbg.c | 207 for (i = 0; i < ARRAY_SIZE(inst->flows); i++) { in vpu_dbg_instance() 208 u32 idx = (inst->flow_idx + i) % (ARRAY_SIZE(inst->flows)); in vpu_dbg_instance() 210 if (!inst->flows[idx]) in vpu_dbg_instance() 213 inst->flows[idx] >= VPU_MSG_ID_NOOP ? "M" : "C", in vpu_dbg_instance() 214 vpu_id_name(inst->flows[idx])); in vpu_dbg_instance() 506 inst->flows[inst->flow_idx] = flow; in vpu_inst_record_flow() 507 inst->flow_idx = (inst->flow_idx + 1) % (ARRAY_SIZE(inst->flows)); in vpu_inst_record_flow()
|
| D | vpu.h | 274 u32 flows[16]; member
|
| /drivers/infiniband/hw/hfi1/ |
| D | tid_rdma.c | 757 u32 generation = rcd->flows[flow_idx].generation; in kern_setup_hw_flow() 775 rcd->flows[flow_idx].generation = in kern_clear_hw_flow() 776 kern_flow_generation_next(rcd->flows[flow_idx].generation); in kern_clear_hw_flow() 804 rcd->flows[fs->index].generation = fs->generation; in hfi1_kern_setup_hw_flow() 853 rcd->flows[i].generation = mask_generation(get_random_u32()); in hfi1_kern_init_ctxt_generations() 1465 struct tid_rdma_flow *flow = &req->flows[req->setup_head]; in hfi1_kern_exp_rcv_setup() 1556 struct tid_rdma_flow *flow = &req->flows[req->clear_tail]; in hfi1_kern_exp_rcv_clear() 1612 kfree(req->flows); in hfi1_kern_exp_rcv_free_flows() 1613 req->flows = NULL; in hfi1_kern_exp_rcv_free_flows() 1634 struct tid_rdma_flow *flows; in hfi1_kern_exp_rcv_alloc_flows() local [all …]
|
| D | tid_rdma.h | 105 struct tid_rdma_flow *flows; /* array of tid flows */ member
|
| D | rc.c | 783 &req->flows[req->setup_head]; in hfi1_make_rc_req() 1072 flow = &req->flows[req->flow_idx]; in hfi1_make_rc_req()
|
| D | hfi.h | 310 struct tid_flow_state flows[RXE_NUM_TID_FLOWS]; member
|
| /drivers/net/ethernet/ti/ |
| D | am65-cpsw-nuss.c | 410 flow = &rx_chn->flows[id]; in am65_cpsw_destroy_xdp_rxqs() 447 flow = &rx_chn->flows[id]; in am65_cpsw_create_xdp_rxqs() 547 am65_cpsw_put_page(&rx_chn->flows[flow_id], page, false); in am65_cpsw_nuss_rx_cleanup() 700 flow = &rx_chn->flows[flow_idx]; in am65_cpsw_nuss_common_open() 728 napi_enable(&rx_chn->flows[i].napi_rx); in am65_cpsw_nuss_common_open() 729 if (rx_chn->flows[i].irq_disabled) { in am65_cpsw_nuss_common_open() 730 rx_chn->flows[i].irq_disabled = false; in am65_cpsw_nuss_common_open() 731 enable_irq(rx_chn->flows[i].irq); in am65_cpsw_nuss_common_open() 757 flow = &rx_chn->flows[flow_idx]; in am65_cpsw_nuss_common_open() 823 napi_disable(&rx_chn->flows[i].napi_rx); in am65_cpsw_nuss_common_stop() [all …]
|
| D | am65-cpsw-nuss.h | 119 struct am65_cpsw_rx_flow flows[AM65_CPSW_MAX_QUEUES]; member
|
| D | am65-cpsw-ethtool.c | 928 rx_flow = &common->rx_chns.flows[queue]; in am65_cpsw_get_per_queue_coalesce() 957 rx_flow = &common->rx_chns.flows[queue]; in am65_cpsw_set_per_queue_coalesce()
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| D | en_rep.h | 183 struct list_head flows; member 208 struct list_head flows; member
|
| D | eswitch_offloads.c | 1187 struct mlx5_flow_handle **flows; in esw_add_fdb_peer_miss_rules() local 1205 flows = kvcalloc(peer_esw->total_vports, sizeof(*flows), GFP_KERNEL); in esw_add_fdb_peer_miss_rules() 1206 if (!flows) { in esw_add_fdb_peer_miss_rules() 1226 flows[peer_vport->index] = flow; in esw_add_fdb_peer_miss_rules() 1238 flows[peer_vport->index] = flow; in esw_add_fdb_peer_miss_rules() 1253 flows[peer_vport->index] = flow; in esw_add_fdb_peer_miss_rules() 1268 flows[peer_vport->index] = flow; in esw_add_fdb_peer_miss_rules() 1279 esw->fdb_table.offloads.peer_miss_rules[pfindex] = flows; in esw_add_fdb_peer_miss_rules() 1287 if (!flows[peer_vport->index]) in esw_add_fdb_peer_miss_rules() 1289 mlx5_del_flow_rules(flows[peer_vport->index]); in esw_add_fdb_peer_miss_rules() [all …]
|
| D | en_tc.c | 589 struct list_head flows; member 996 WARN_ON(!list_empty(&hpe->flows)); in mlx5e_hairpin_put() 1145 INIT_LIST_HEAD(&hpe->flows); in mlx5e_hairpin_flow_add() 1205 list_add(&flow->hairpin, &hpe->flows); in mlx5e_hairpin_flow_add()
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/ |
| D | tc_tun_encap.c | 301 list_for_each_entry(efi, &e->flows, list) { in mlx5e_take_all_encap_flows() 422 list_for_each_entry_safe(efi, tmp, &e->flows, list) { in mlx5e_tc_update_neigh_used_value() 467 WARN_ON(!list_empty(&e->flows)); in mlx5e_encap_dealloc() 484 WARN_ON(!list_empty(&d->flows)); in mlx5e_decap_dealloc() 892 INIT_LIST_HEAD(&e->flows); in mlx5e_attach_encap() 926 list_add(&flow->encaps[out_index].list, &e->flows); in mlx5e_attach_encap() 991 INIT_LIST_HEAD(&d->flows); in mlx5e_attach_decap() 1014 list_add(&flow->l3_to_l2_reformat, &d->flows); in mlx5e_attach_decap()
|
| /drivers/net/ |
| D | tun.c | 201 struct hlist_head flows[TUN_NUM_FLOW_ENTRIES]; member 420 hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) in tun_flow_flush() 435 hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) { in tun_flow_delete_by_queue() 456 hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) { in tun_flow_cleanup() 483 head = &tun->flows[tun_hashfn(rxhash)]; in tun_flow_update() 533 e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq); in tun_automq_select_queue() 1046 e = tun_flow_find(&tun->flows[tun_hashfn(rxhash)], rxhash); in tun_automq_xmit() 1353 INIT_HLIST_HEAD(&tun->flows[i]); in tun_flow_init()
|
| /drivers/infiniband/core/ |
| D | uverbs_cmd.c | 3274 flow_attr = kzalloc(struct_size(flow_attr, flows, in ib_uverbs_ex_create_flow()
|