Home
last modified time | relevance | path

Searched refs:flow_rule (Results 1 – 25 of 41) sorted by relevance

12

/drivers/net/ethernet/netronome/nfp/flower/
Dmain.h424 struct flow_rule *rule);
434 struct flow_rule *rule);
438 struct flow_rule *rule,
443 struct flow_rule *rule);
447 struct flow_rule *rule);
450 struct nfp_flower_ipv4 *msk, struct flow_rule *rule);
453 struct nfp_flower_ipv6 *msk, struct flow_rule *rule);
455 nfp_flower_compile_geneve_opt(u8 *ext, u8 *msk, struct flow_rule *rule);
459 struct flow_rule *rule);
463 struct flow_rule *rule);
[all …]
Dmatch.c25 struct flow_rule *rule) in nfp_flower_compile_tci()
54 struct flow_rule *rule, u8 key_type, bool qinq_sup) in nfp_flower_compile_meta_tci()
97 struct flow_rule *rule) in nfp_flower_compile_mac()
119 struct flow_rule *rule, in nfp_flower_compile_mpls()
174 struct flow_rule *rule, in nfp_flower_compile_mac_mpls()
188 struct flow_rule *rule) in nfp_flower_compile_tport()
203 struct nfp_flower_ip_ext *msk, struct flow_rule *rule) in nfp_flower_compile_ip_ext()
308 struct flow_rule *rule) in nfp_flower_compile_vlan()
324 struct nfp_flower_ipv4 *msk, struct flow_rule *rule) in nfp_flower_compile_ipv4()
341 struct nfp_flower_ipv6 *msk, struct flow_rule *rule) in nfp_flower_compile_ipv6()
[all …]
Dconntrack.c23 static struct flow_action_entry *get_flow_act(struct flow_rule *rule,
67 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in is_post_ct_flow()
256 struct flow_rule *rule) in nfp_ct_check_mangle_merge()
475 static int nfp_fl_merge_actions_offload(struct flow_rule **rules, in nfp_fl_merge_actions_offload()
482 struct flow_rule *a_rule; in nfp_fl_merge_actions_offload()
551 struct flow_rule *rules[_CT_TYPE_MAX]; in nfp_fl_ct_add_offload()
1353 static struct flow_action_entry *get_flow_act(struct flow_rule *rule, in get_flow_act()
1471 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in nfp_fl_ct_handle_post_ct()
Daction.c265 nfp_flower_tun_is_gre(struct flow_rule *rule, int start_idx) in nfp_flower_tun_is_gre()
282 struct flow_rule *rule, in nfp_fl_get_tun_from_act()
791 nfp_fl_commit_mangle(struct flow_rule *rule, char *nfp_action, in nfp_fl_commit_mangle()
979 struct flow_rule *rule, in nfp_flower_loop_action()
1197 struct flow_rule *rule, in nfp_flower_compile_action()
Dconntrack.h127 struct flow_rule *rule;
Doffload.c139 static bool nfp_flower_check_higher_than_mac(struct flow_rule *rule) in nfp_flower_check_higher_than_mac()
147 static bool nfp_flower_check_higher_than_l3(struct flow_rule *rule) in nfp_flower_check_higher_than_l3()
241 struct flow_rule *rule, in nfp_flower_calculate_key_layers()
1273 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in offload_pre_check()
1299 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in nfp_flower_add_offload()
/drivers/net/ethernet/mellanox/mlx5/core/
Drdma.c26 struct mlx5_flow_handle *flow_rule = NULL; in mlx5_rdma_enable_roce_steering() local
90 flow_rule = mlx5_add_flow_rules(ft, spec, &flow_act, NULL, 0); in mlx5_rdma_enable_roce_steering()
91 if (IS_ERR(flow_rule)) { in mlx5_rdma_enable_roce_steering()
92 err = PTR_ERR(flow_rule); in mlx5_rdma_enable_roce_steering()
102 roce->allow_rule = flow_rule; in mlx5_rdma_enable_roce_steering()
Den_rep.c335 struct mlx5_flow_handle *flow_rule; in mlx5e_sqs2vport_start() local
357 flow_rule = mlx5_eswitch_add_send_to_vport_rule(esw, esw, rep, in mlx5e_sqs2vport_start()
359 if (IS_ERR(flow_rule)) { in mlx5e_sqs2vport_start()
360 err = PTR_ERR(flow_rule); in mlx5e_sqs2vport_start()
364 rep_sq->send_to_vport_rule = flow_rule; in mlx5e_sqs2vport_start()
368 flow_rule = mlx5_eswitch_add_send_to_vport_rule(peer_esw, esw, in mlx5e_sqs2vport_start()
370 if (IS_ERR(flow_rule)) { in mlx5e_sqs2vport_start()
371 err = PTR_ERR(flow_rule); in mlx5e_sqs2vport_start()
376 rep_sq->send_to_vport_rule_peer = flow_rule; in mlx5e_sqs2vport_start()
769 struct mlx5_flow_handle *flow_rule; in mlx5e_create_rep_vport_rx_rule() local
[all …]
Deswitch.c61 struct mlx5_flow_handle *flow_rule; member
195 struct mlx5_flow_handle *flow_rule = NULL; in __esw_fdb_set_vport_rule() local
238 flow_rule = in __esw_fdb_set_vport_rule()
241 if (IS_ERR(flow_rule)) { in __esw_fdb_set_vport_rule()
244 dmac_v, dmac_c, vport, PTR_ERR(flow_rule)); in __esw_fdb_set_vport_rule()
245 flow_rule = NULL; in __esw_fdb_set_vport_rule()
249 return flow_rule; in __esw_fdb_set_vport_rule()
313 vaddr->flow_rule = esw_fdb_set_vport_rule(esw, mac, vport); in esw_add_uc_addr()
316 vport, mac, vaddr->flow_rule); in esw_add_uc_addr()
341 if (vaddr->flow_rule) in esw_del_uc_addr()
[all …]
Deswitch_offloads.c931 struct mlx5_flow_handle *flow_rule; in mlx5_eswitch_add_send_to_vport_rule() local
937 flow_rule = ERR_PTR(-ENOMEM); in mlx5_eswitch_add_send_to_vport_rule()
967 flow_rule = mlx5_add_flow_rules(on_esw->fdb_table.offloads.slow_fdb, in mlx5_eswitch_add_send_to_vport_rule()
969 if (IS_ERR(flow_rule)) in mlx5_eswitch_add_send_to_vport_rule()
971 PTR_ERR(flow_rule)); in mlx5_eswitch_add_send_to_vport_rule()
974 return flow_rule; in mlx5_eswitch_add_send_to_vport_rule()
1003 struct mlx5_flow_handle *flow_rule; in mlx5_eswitch_add_send_to_vport_meta_rules() local
1038 flow_rule = mlx5_add_flow_rules(esw->fdb_table.offloads.slow_fdb, in mlx5_eswitch_add_send_to_vport_meta_rules()
1040 if (IS_ERR(flow_rule)) { in mlx5_eswitch_add_send_to_vport_meta_rules()
1041 err = PTR_ERR(flow_rule); in mlx5_eswitch_add_send_to_vport_meta_rules()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/en/
Dtc_ct.c80 struct mlx5_flow_handle *flow_rule; member
212 mlx5_tc_ct_rule_to_tuple(struct mlx5_ct_tuple *tuple, struct flow_rule *rule) in mlx5_tc_ct_rule_to_tuple()
262 struct flow_rule *rule) in mlx5_tc_ct_rule_to_tuple_nat()
324 struct flow_rule *rule) in mlx5_tc_ct_set_tuple_match()
474 mlx5_tc_ct_get_ct_metadata_action(struct flow_rule *flow_rule) in mlx5_tc_ct_get_ct_metadata_action() argument
476 struct flow_action *flow_action = &flow_rule->action; in mlx5_tc_ct_get_ct_metadata_action()
606 struct flow_rule *flow_rule, in mlx5_tc_ct_entry_create_nat() argument
609 struct flow_action *flow_action = &flow_rule->action; in mlx5_tc_ct_entry_create_nat()
644 struct flow_rule *flow_rule, in mlx5_tc_ct_entry_create_mod_hdr() argument
653 meta = mlx5_tc_ct_get_ct_metadata_action(flow_rule); in mlx5_tc_ct_entry_create_mod_hdr()
[all …]
Dtc_tun_vxlan.c21 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_check_udp_dport_vxlan()
109 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_vxlan()
Dtc_tun_geneve.c25 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_check_udp_dport_geneve()
127 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_geneve_vni()
161 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_geneve_options()
Dtc_tun_gre.c63 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_gretap()
Dtc_tun_mplsoudp.c62 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in parse_tunnel()
Dtc_ct.h158 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5_tc_ct_match_add()
/drivers/net/ethernet/mellanox/mlx5/core/esw/
Dlegacy.c255 struct mlx5_flow_handle *flow_rule; in _mlx5_eswitch_set_vepa_locked() local
283 flow_rule = mlx5_add_flow_rules(esw->fdb_table.legacy.vepa_fdb, spec, in _mlx5_eswitch_set_vepa_locked()
285 if (IS_ERR(flow_rule)) { in _mlx5_eswitch_set_vepa_locked()
286 err = PTR_ERR(flow_rule); in _mlx5_eswitch_set_vepa_locked()
289 esw->fdb_table.legacy.vepa_uplink_rule = flow_rule; in _mlx5_eswitch_set_vepa_locked()
297 flow_rule = mlx5_add_flow_rules(esw->fdb_table.legacy.vepa_fdb, NULL, in _mlx5_eswitch_set_vepa_locked()
299 if (IS_ERR(flow_rule)) { in _mlx5_eswitch_set_vepa_locked()
300 err = PTR_ERR(flow_rule); in _mlx5_eswitch_set_vepa_locked()
303 esw->fdb_table.legacy.vepa_star_rule = flow_rule; in _mlx5_eswitch_set_vepa_locked()
/drivers/net/ethernet/mellanox/mlxsw/
Dspectrum_flower.c243 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_meta()
335 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_ports()
362 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_tcp()
393 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_ip()
426 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse()
/drivers/net/ethernet/stmicro/stmmac/
Dstmmac_tc.c458 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_basic_flow()
475 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_ip4_flow()
510 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_ports_flow()
586 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_flow()
660 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_vlan_flow()
/drivers/net/ethernet/freescale/dpaa2/
Ddpaa2-switch-flower.c14 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_flower_parse_key()
493 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_cls_flower_replace_acl()
537 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_flower_parse_mirror_key()
645 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_cls_flower_replace()
/drivers/net/ethernet/chelsio/cxgb4/
Dcxgb4_tc_flower.h140 int cxgb4_flow_rule_replace(struct net_device *dev, struct flow_rule *rule,
Dcxgb4_tc_flower.c163 struct flow_rule *rule, in cxgb4_process_flow_match()
309 struct flow_rule *rule) in cxgb4_validate_flow_match()
855 int cxgb4_flow_rule_replace(struct net_device *dev, struct flow_rule *rule, in cxgb4_flow_rule_replace()
929 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in cxgb4_tc_flower_replace()
/drivers/net/ethernet/qlogic/qede/
Dqede_filter.c1703 qede_flow_parse_ports(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_ports()
1724 qede_flow_parse_v6_common(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_v6_common()
1756 qede_flow_parse_v4_common(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_v4_common()
1780 qede_flow_parse_tcp_v6(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_tcp_v6()
1790 qede_flow_parse_tcp_v4(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_tcp_v4()
1800 qede_flow_parse_udp_v6(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_udp_v6()
1810 qede_flow_parse_udp_v4(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_udp_v4()
1821 struct flow_rule *rule, struct qede_arfs_tuple *tuple) in qede_parse_flow_attr()
/drivers/net/ethernet/marvell/prestera/
Dprestera_flower.c51 struct flow_rule *f_rule = flow_cls_offload_flow_rule(f); in prestera_flower_parse_meta()
90 struct flow_rule *f_rule = flow_cls_offload_flow_rule(f); in prestera_flower_parse()
/drivers/net/dsa/sja1105/
Dsja1105_flower.c200 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in sja1105_flower_parse_key()
306 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in sja1105_cls_flower_add()

12