/net/mac80211/ |
D | link.c | 185 unsigned long add = new_links & ~old_links; in ieee80211_vif_update_links() local 206 for_each_set_bit(link_id, &add, IEEE80211_MLD_MAX_NUM_LINKS) { in ieee80211_vif_update_links() 239 for_each_set_bit(link_id, &add, IEEE80211_MLD_MAX_NUM_LINKS) { in ieee80211_vif_update_links() 330 unsigned long add = active_links & ~old_active; in _ieee80211_set_active_links() local 352 if (add) { in _ieee80211_set_active_links() 388 ret = ieee80211_key_switch_links(sdata, rem, add); in _ieee80211_set_active_links() 412 for_each_set_bit(link_id, &add, IEEE80211_MLD_MAX_NUM_LINKS) { in _ieee80211_set_active_links()
|
/net/sched/ |
D | cls_u32.c | 1221 bool add, flow_setup_cb_t *cb, void *cb_priv, in u32_reoffload_hnode() argument 1228 cls_u32.command = add ? TC_CLSU32_NEW_HNODE : TC_CLSU32_DELETE_HNODE; in u32_reoffload_hnode() 1234 if (err && add && tc_skip_sw(ht->flags)) in u32_reoffload_hnode() 1241 bool add, flow_setup_cb_t *cb, void *cb_priv, in u32_reoffload_knode() argument 1249 cls_u32.command = add ? in u32_reoffload_knode() 1253 if (add) { in u32_reoffload_knode() 1269 return tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSU32, in u32_reoffload_knode() 1274 static int u32_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb, in u32_reoffload() argument 1293 if (add && !tc_skip_hw(ht->flags)) { in u32_reoffload() 1294 err = u32_reoffload_hnode(tp, ht, add, cb, cb_priv, in u32_reoffload() [all …]
|
D | cls_matchall.c | 273 static int mall_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb, in mall_reoffload() argument 289 cls_mall.command = add ? in mall_reoffload() 298 return add && tc_skip_sw(head->flags) ? err : 0; in mall_reoffload() 301 err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSMATCHALL, in mall_reoffload()
|
D | cls_bpf.c | 650 static int cls_bpf_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb, in cls_bpf_reoffload() argument 667 cls_bpf.prog = add ? prog->filter : NULL; in cls_bpf_reoffload() 668 cls_bpf.oldprog = add ? NULL : prog->filter; in cls_bpf_reoffload() 672 err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSBPF, in cls_bpf_reoffload()
|
D | cls_api.c | 1525 void *cb_priv, bool add, bool offload_in_use, in tcf_block_playback_offloads() argument 1539 if (chain->tmplt_ops && add) in tcf_block_playback_offloads() 1547 err = tp->ops->reoffload(tp, add, cb, cb_priv, in tcf_block_playback_offloads() 1549 if (err && add) in tcf_block_playback_offloads() 1551 } else if (add && offload_in_use) { in tcf_block_playback_offloads() 1557 if (chain->tmplt_ops && !add) in tcf_block_playback_offloads() 3450 u32 *flags, u32 diff, bool add) in tc_cls_offload_cnt_update() argument 3455 if (add) { in tc_cls_offload_cnt_update() 3673 bool add, flow_setup_cb_t *cb, in tc_setup_cb_reoffload() argument 3680 if (add && tc_skip_sw(*flags)) in tc_setup_cb_reoffload() [all …]
|
D | cls_flower.c | 2525 fl_get_next_hw_filter(struct tcf_proto *tp, struct cls_fl_filter *f, bool add) in fl_get_next_hw_filter() argument 2539 if (!(add && f->deleted) && refcount_inc_not_zero(&f->refcnt)) { in fl_get_next_hw_filter() 2549 static int fl_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb, in fl_reoffload() argument 2563 while ((f = fl_get_next_hw_filter(tp, f, add))) { in fl_reoffload() 2573 cls_flower.command = add ? in fl_reoffload() 2593 err = tc_setup_cb_reoffload(block, tp, add, cb, in fl_reoffload() 2727 static void fl_tmplt_reoffload(struct tcf_chain *chain, bool add, in fl_tmplt_reoffload() argument 2738 cls_flower.command = add ? FLOW_CLS_TMPLT_CREATE : in fl_tmplt_reoffload()
|
D | act_api.c | 1836 void *cb_priv, bool add) in tcf_action_reoffload_cb() argument 1870 if (add) { in tcf_action_reoffload_cb()
|
/net/caif/ |
D | cfpkt_skbuff.c | 299 struct sk_buff *add = pkt_to_skb(addpkt); in cfpkt_append() local 300 u16 addlen = skb_headlen(add); in cfpkt_append() 326 skb_put_data(dst, add->data, skb_headlen(add)); in cfpkt_append()
|
/net/ipv4/ |
D | ip_sockglue.c | 729 int omode, add, err; in do_mcast_group_source() local 747 add = 1; in do_mcast_group_source() 750 add = 0; in do_mcast_group_source() 763 add = 1; in do_mcast_group_source() 766 add = 0; in do_mcast_group_source() 768 return ip_mc_source(add, omode, sk, &mreqs, greqs.gsr_interface); in do_mcast_group_source() 1278 int omode, add; in do_ip_setsockopt() local 1288 add = 1; in do_ip_setsockopt() 1291 add = 0; in do_ip_setsockopt() 1302 add = 1; in do_ip_setsockopt() [all …]
|
D | route.c | 2569 goto add; in __mkroute_output() 2578 goto add; in __mkroute_output() 2587 add: in __mkroute_output()
|
/net/netfilter/ |
D | xt_recent.c | 568 bool add, succ; in recent_mt_proc_write() local 587 add = false; in recent_mt_proc_write() 590 add = true; in recent_mt_proc_write() 613 if (add) in recent_mt_proc_write() 616 if (add) in recent_mt_proc_write()
|
D | Kconfig | 800 Using this target and match, you can add/delete and match 1140 Workaround: activate this option and add a rule to your firewall 1310 It will also add a "tos" match, which allows you to match packets
|
/net/bridge/ |
D | br_mrp_switchdev.c | 9 const struct switchdev_obj *obj, bool add) in br_mrp_switchdev_port_obj() argument 13 if (add) in br_mrp_switchdev_port_obj()
|
/net/psample/ |
D | Kconfig | 10 Say Y here to add support for packet-sampling netlink channel
|
/net/ipv6/ |
D | ipv6_sockglue.c | 168 int omode, add; in do_ipv6_mcast_group_source() local 181 add = 1; in do_ipv6_mcast_group_source() 184 add = 0; in do_ipv6_mcast_group_source() 197 add = 1; in do_ipv6_mcast_group_source() 200 add = 0; in do_ipv6_mcast_group_source() 202 return ip6_mc_source(add, omode, sk, &greqs); in do_ipv6_mcast_group_source()
|
D | ip6_fib.c | 1081 int add = (!info->nlh || in fib6_add_rt2node() local 1209 if (!add) in fib6_add_rt2node() 1212 add: in fib6_add_rt2node() 1267 if (add) in fib6_add_rt2node() 1268 goto add; in fib6_add_rt2node()
|
/net/ife/ |
D | Kconfig | 10 Say Y here to add support of IFE encapsulation protocol
|
/net/rds/ |
D | rdma.c | 885 rm->atomic.op_m_fadd.add = args->fadd.add; in rds_cmsg_atomic() 890 rm->atomic.op_m_fadd.add = args->m_fadd.add; in rds_cmsg_atomic()
|
D | ib.c | 290 .add = rds_ib_add_one,
|
D | rds.h | 438 uint64_t add; member
|
D | ib_send.c | 795 send->s_atomic_wr.compare_add = op->op_m_fadd.add; in rds_ib_xmit_atomic()
|
/net/l2tp/ |
D | Kconfig | 104 interface, or add the interface to a bridge.
|
/net/smc/ |
D | smc_ism.c | 39 .add = smcd_register_dev,
|
D | smc_ib.c | 1010 .add = smc_ib_add_dev,
|
/net/netfilter/ipvs/ |
D | ip_vs_ctl.c | 968 struct ip_vs_dest_user_kern *udest, int add) in __ip_vs_update_dest() argument 976 BUG_ON(!add && udest->af != dest->af); in __ip_vs_update_dest() 978 if (add && udest->af != svc->af) in __ip_vs_update_dest() 982 if (add || udest->weight != 0) in __ip_vs_update_dest() 1040 if (add) { in __ip_vs_update_dest()
|