/drivers/net/ethernet/netronome/nfp/flower/ |
D | main.h | 424 struct flow_rule *rule); 434 struct flow_rule *rule); 438 struct flow_rule *rule, 443 struct flow_rule *rule); 447 struct flow_rule *rule); 450 struct nfp_flower_ipv4 *msk, struct flow_rule *rule); 453 struct nfp_flower_ipv6 *msk, struct flow_rule *rule); 455 nfp_flower_compile_geneve_opt(u8 *ext, u8 *msk, struct flow_rule *rule); 459 struct flow_rule *rule); 463 struct flow_rule *rule); [all …]
|
D | match.c | 25 struct flow_rule *rule) in nfp_flower_compile_tci() 54 struct flow_rule *rule, u8 key_type, bool qinq_sup) in nfp_flower_compile_meta_tci() 97 struct flow_rule *rule) in nfp_flower_compile_mac() 119 struct flow_rule *rule, in nfp_flower_compile_mpls() 174 struct flow_rule *rule, in nfp_flower_compile_mac_mpls() 188 struct flow_rule *rule) in nfp_flower_compile_tport() 203 struct nfp_flower_ip_ext *msk, struct flow_rule *rule) in nfp_flower_compile_ip_ext() 308 struct flow_rule *rule) in nfp_flower_compile_vlan() 324 struct nfp_flower_ipv4 *msk, struct flow_rule *rule) in nfp_flower_compile_ipv4() 341 struct nfp_flower_ipv6 *msk, struct flow_rule *rule) in nfp_flower_compile_ipv6() [all …]
|
D | conntrack.c | 23 static struct flow_action_entry *get_flow_act(struct flow_rule *rule, 67 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in is_post_ct_flow() 256 struct flow_rule *rule) in nfp_ct_check_mangle_merge() 475 static int nfp_fl_merge_actions_offload(struct flow_rule **rules, in nfp_fl_merge_actions_offload() 482 struct flow_rule *a_rule; in nfp_fl_merge_actions_offload() 551 struct flow_rule *rules[_CT_TYPE_MAX]; in nfp_fl_ct_add_offload() 1353 static struct flow_action_entry *get_flow_act(struct flow_rule *rule, in get_flow_act() 1471 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in nfp_fl_ct_handle_post_ct()
|
D | action.c | 265 nfp_flower_tun_is_gre(struct flow_rule *rule, int start_idx) in nfp_flower_tun_is_gre() 282 struct flow_rule *rule, in nfp_fl_get_tun_from_act() 791 nfp_fl_commit_mangle(struct flow_rule *rule, char *nfp_action, in nfp_fl_commit_mangle() 979 struct flow_rule *rule, in nfp_flower_loop_action() 1197 struct flow_rule *rule, in nfp_flower_compile_action()
|
D | conntrack.h | 127 struct flow_rule *rule;
|
D | offload.c | 139 static bool nfp_flower_check_higher_than_mac(struct flow_rule *rule) in nfp_flower_check_higher_than_mac() 147 static bool nfp_flower_check_higher_than_l3(struct flow_rule *rule) in nfp_flower_check_higher_than_l3() 241 struct flow_rule *rule, in nfp_flower_calculate_key_layers() 1273 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in offload_pre_check() 1299 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in nfp_flower_add_offload()
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | rdma.c | 26 struct mlx5_flow_handle *flow_rule = NULL; in mlx5_rdma_enable_roce_steering() local 90 flow_rule = mlx5_add_flow_rules(ft, spec, &flow_act, NULL, 0); in mlx5_rdma_enable_roce_steering() 91 if (IS_ERR(flow_rule)) { in mlx5_rdma_enable_roce_steering() 92 err = PTR_ERR(flow_rule); in mlx5_rdma_enable_roce_steering() 102 roce->allow_rule = flow_rule; in mlx5_rdma_enable_roce_steering()
|
D | en_rep.c | 335 struct mlx5_flow_handle *flow_rule; in mlx5e_sqs2vport_start() local 357 flow_rule = mlx5_eswitch_add_send_to_vport_rule(esw, esw, rep, in mlx5e_sqs2vport_start() 359 if (IS_ERR(flow_rule)) { in mlx5e_sqs2vport_start() 360 err = PTR_ERR(flow_rule); in mlx5e_sqs2vport_start() 364 rep_sq->send_to_vport_rule = flow_rule; in mlx5e_sqs2vport_start() 368 flow_rule = mlx5_eswitch_add_send_to_vport_rule(peer_esw, esw, in mlx5e_sqs2vport_start() 370 if (IS_ERR(flow_rule)) { in mlx5e_sqs2vport_start() 371 err = PTR_ERR(flow_rule); in mlx5e_sqs2vport_start() 376 rep_sq->send_to_vport_rule_peer = flow_rule; in mlx5e_sqs2vport_start() 769 struct mlx5_flow_handle *flow_rule; in mlx5e_create_rep_vport_rx_rule() local [all …]
|
D | eswitch.c | 61 struct mlx5_flow_handle *flow_rule; member 195 struct mlx5_flow_handle *flow_rule = NULL; in __esw_fdb_set_vport_rule() local 238 flow_rule = in __esw_fdb_set_vport_rule() 241 if (IS_ERR(flow_rule)) { in __esw_fdb_set_vport_rule() 244 dmac_v, dmac_c, vport, PTR_ERR(flow_rule)); in __esw_fdb_set_vport_rule() 245 flow_rule = NULL; in __esw_fdb_set_vport_rule() 249 return flow_rule; in __esw_fdb_set_vport_rule() 313 vaddr->flow_rule = esw_fdb_set_vport_rule(esw, mac, vport); in esw_add_uc_addr() 316 vport, mac, vaddr->flow_rule); in esw_add_uc_addr() 341 if (vaddr->flow_rule) in esw_del_uc_addr() [all …]
|
D | eswitch_offloads.c | 931 struct mlx5_flow_handle *flow_rule; in mlx5_eswitch_add_send_to_vport_rule() local 937 flow_rule = ERR_PTR(-ENOMEM); in mlx5_eswitch_add_send_to_vport_rule() 967 flow_rule = mlx5_add_flow_rules(on_esw->fdb_table.offloads.slow_fdb, in mlx5_eswitch_add_send_to_vport_rule() 969 if (IS_ERR(flow_rule)) in mlx5_eswitch_add_send_to_vport_rule() 971 PTR_ERR(flow_rule)); in mlx5_eswitch_add_send_to_vport_rule() 974 return flow_rule; in mlx5_eswitch_add_send_to_vport_rule() 1003 struct mlx5_flow_handle *flow_rule; in mlx5_eswitch_add_send_to_vport_meta_rules() local 1038 flow_rule = mlx5_add_flow_rules(esw->fdb_table.offloads.slow_fdb, in mlx5_eswitch_add_send_to_vport_meta_rules() 1040 if (IS_ERR(flow_rule)) { in mlx5_eswitch_add_send_to_vport_meta_rules() 1041 err = PTR_ERR(flow_rule); in mlx5_eswitch_add_send_to_vport_meta_rules() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | tc_ct.c | 80 struct mlx5_flow_handle *flow_rule; member 212 mlx5_tc_ct_rule_to_tuple(struct mlx5_ct_tuple *tuple, struct flow_rule *rule) in mlx5_tc_ct_rule_to_tuple() 262 struct flow_rule *rule) in mlx5_tc_ct_rule_to_tuple_nat() 324 struct flow_rule *rule) in mlx5_tc_ct_set_tuple_match() 474 mlx5_tc_ct_get_ct_metadata_action(struct flow_rule *flow_rule) in mlx5_tc_ct_get_ct_metadata_action() argument 476 struct flow_action *flow_action = &flow_rule->action; in mlx5_tc_ct_get_ct_metadata_action() 606 struct flow_rule *flow_rule, in mlx5_tc_ct_entry_create_nat() argument 609 struct flow_action *flow_action = &flow_rule->action; in mlx5_tc_ct_entry_create_nat() 644 struct flow_rule *flow_rule, in mlx5_tc_ct_entry_create_mod_hdr() argument 653 meta = mlx5_tc_ct_get_ct_metadata_action(flow_rule); in mlx5_tc_ct_entry_create_mod_hdr() [all …]
|
D | tc_tun_vxlan.c | 21 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_check_udp_dport_vxlan() 109 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_vxlan()
|
D | tc_tun_geneve.c | 25 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_check_udp_dport_geneve() 127 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_geneve_vni() 161 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_geneve_options()
|
D | tc_tun_gre.c | 63 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_gretap()
|
D | tc_tun_mplsoudp.c | 62 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in parse_tunnel()
|
D | tc_ct.h | 158 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5_tc_ct_match_add()
|
/drivers/net/ethernet/mellanox/mlx5/core/esw/ |
D | legacy.c | 255 struct mlx5_flow_handle *flow_rule; in _mlx5_eswitch_set_vepa_locked() local 283 flow_rule = mlx5_add_flow_rules(esw->fdb_table.legacy.vepa_fdb, spec, in _mlx5_eswitch_set_vepa_locked() 285 if (IS_ERR(flow_rule)) { in _mlx5_eswitch_set_vepa_locked() 286 err = PTR_ERR(flow_rule); in _mlx5_eswitch_set_vepa_locked() 289 esw->fdb_table.legacy.vepa_uplink_rule = flow_rule; in _mlx5_eswitch_set_vepa_locked() 297 flow_rule = mlx5_add_flow_rules(esw->fdb_table.legacy.vepa_fdb, NULL, in _mlx5_eswitch_set_vepa_locked() 299 if (IS_ERR(flow_rule)) { in _mlx5_eswitch_set_vepa_locked() 300 err = PTR_ERR(flow_rule); in _mlx5_eswitch_set_vepa_locked() 303 esw->fdb_table.legacy.vepa_star_rule = flow_rule; in _mlx5_eswitch_set_vepa_locked()
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_flower.c | 243 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_meta() 335 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_ports() 362 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_tcp() 393 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_ip() 426 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse()
|
/drivers/net/ethernet/stmicro/stmmac/ |
D | stmmac_tc.c | 458 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_basic_flow() 475 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_ip4_flow() 510 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_ports_flow() 586 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_flow() 660 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_vlan_flow()
|
/drivers/net/ethernet/freescale/dpaa2/ |
D | dpaa2-switch-flower.c | 14 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_flower_parse_key() 493 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_cls_flower_replace_acl() 537 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_flower_parse_mirror_key() 645 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_cls_flower_replace()
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_tc_flower.h | 140 int cxgb4_flow_rule_replace(struct net_device *dev, struct flow_rule *rule,
|
D | cxgb4_tc_flower.c | 163 struct flow_rule *rule, in cxgb4_process_flow_match() 309 struct flow_rule *rule) in cxgb4_validate_flow_match() 855 int cxgb4_flow_rule_replace(struct net_device *dev, struct flow_rule *rule, in cxgb4_flow_rule_replace() 929 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in cxgb4_tc_flower_replace()
|
/drivers/net/ethernet/qlogic/qede/ |
D | qede_filter.c | 1703 qede_flow_parse_ports(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_ports() 1724 qede_flow_parse_v6_common(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_v6_common() 1756 qede_flow_parse_v4_common(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_v4_common() 1780 qede_flow_parse_tcp_v6(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_tcp_v6() 1790 qede_flow_parse_tcp_v4(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_tcp_v4() 1800 qede_flow_parse_udp_v6(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_udp_v6() 1810 qede_flow_parse_udp_v4(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_udp_v4() 1821 struct flow_rule *rule, struct qede_arfs_tuple *tuple) in qede_parse_flow_attr()
|
/drivers/net/ethernet/marvell/prestera/ |
D | prestera_flower.c | 51 struct flow_rule *f_rule = flow_cls_offload_flow_rule(f); in prestera_flower_parse_meta() 90 struct flow_rule *f_rule = flow_cls_offload_flow_rule(f); in prestera_flower_parse()
|
/drivers/net/dsa/sja1105/ |
D | sja1105_flower.c | 200 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in sja1105_flower_parse_key() 306 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in sja1105_cls_flower_add()
|