/net/core/ |
D | neighbour.c | 60 static int pneigh_ifdown_and_unlock(struct neigh_table *tbl, 125 struct neighbour __rcu **np, struct neigh_table *tbl) in neigh_del() argument 134 lockdep_is_held(&tbl->lock)); in neigh_del() 145 bool neigh_remove_one(struct neighbour *ndel, struct neigh_table *tbl) in neigh_remove_one() argument 153 nht = rcu_dereference_protected(tbl->nht, in neigh_remove_one() 154 lockdep_is_held(&tbl->lock)); in neigh_remove_one() 155 hash_val = tbl->hash(pkey, ndel->dev, nht->hash_rnd); in neigh_remove_one() 160 lockdep_is_held(&tbl->lock)))) { in neigh_remove_one() 162 return neigh_del(n, 0, np, tbl); in neigh_remove_one() 168 static int neigh_forced_gc(struct neigh_table *tbl) in neigh_forced_gc() argument [all …]
|
D | sysctl_net_core.c | 214 struct ctl_table tbl = { in set_default_qdisc() local 222 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); in set_default_qdisc() 550 struct ctl_table *tbl; in sysctl_core_net_init() local 552 tbl = netns_core_table; in sysctl_core_net_init() 554 tbl = kmemdup(tbl, sizeof(netns_core_table), GFP_KERNEL); in sysctl_core_net_init() 555 if (tbl == NULL) in sysctl_core_net_init() 558 tbl[0].data = &net->core.sysctl_somaxconn; in sysctl_core_net_init() 562 tbl[0].procname = NULL; in sysctl_core_net_init() 566 net->core.sysctl_hdr = register_net_sysctl(net, "net/core", tbl); in sysctl_core_net_init() 573 if (tbl != netns_core_table) in sysctl_core_net_init() [all …]
|
/net/netfilter/ipvs/ |
D | ip_vs_lblc.c | 171 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) in ip_vs_lblc_hash() argument 175 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblc_hash() 176 atomic_inc(&tbl->entries); in ip_vs_lblc_hash() 182 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, in ip_vs_lblc_get() argument 188 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblc_get() 201 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblc_new() argument 206 en = ip_vs_lblc_get(af, tbl, daddr); in ip_vs_lblc_new() 223 ip_vs_lblc_hash(tbl, en); in ip_vs_lblc_new() 234 struct ip_vs_lblc_table *tbl = svc->sched_data; in ip_vs_lblc_flush() local 240 tbl->dead = 1; in ip_vs_lblc_flush() [all …]
|
D | ip_vs_lblcr.c | 334 ip_vs_lblcr_hash(struct ip_vs_lblcr_table *tbl, struct ip_vs_lblcr_entry *en) in ip_vs_lblcr_hash() argument 338 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblcr_hash() 339 atomic_inc(&tbl->entries); in ip_vs_lblcr_hash() 345 ip_vs_lblcr_get(int af, struct ip_vs_lblcr_table *tbl, in ip_vs_lblcr_get() argument 351 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblcr_get() 364 ip_vs_lblcr_new(struct ip_vs_lblcr_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblcr_new() argument 369 en = ip_vs_lblcr_get(af, tbl, daddr); in ip_vs_lblcr_new() 385 ip_vs_lblcr_hash(tbl, en); in ip_vs_lblcr_new() 400 struct ip_vs_lblcr_table *tbl = svc->sched_data; in ip_vs_lblcr_flush() local 406 tbl->dead = 1; in ip_vs_lblcr_flush() [all …]
|
D | ip_vs_ctl.c | 3924 struct ctl_table *tbl; in ip_vs_control_net_init_sysctl() local 3932 tbl = kmemdup(vs_vars, sizeof(vs_vars), GFP_KERNEL); in ip_vs_control_net_init_sysctl() 3933 if (tbl == NULL) in ip_vs_control_net_init_sysctl() 3938 tbl[0].procname = NULL; in ip_vs_control_net_init_sysctl() 3940 tbl = vs_vars; in ip_vs_control_net_init_sysctl() 3943 if (tbl[idx].proc_handler == proc_do_defense_mode) in ip_vs_control_net_init_sysctl() 3944 tbl[idx].extra2 = ipvs; in ip_vs_control_net_init_sysctl() 3948 tbl[idx++].data = &ipvs->sysctl_amemthresh; in ip_vs_control_net_init_sysctl() 3950 tbl[idx++].data = &ipvs->sysctl_am_droprate; in ip_vs_control_net_init_sysctl() 3951 tbl[idx++].data = &ipvs->sysctl_drop_entry; in ip_vs_control_net_init_sysctl() [all …]
|
/net/netfilter/ |
D | xt_repldata.h | 24 } *tbl; \ 26 size_t term_offset = (offsetof(typeof(*tbl), entries[nhooks]) + \ 28 tbl = kzalloc(term_offset + sizeof(*term), GFP_KERNEL); \ 29 if (tbl == NULL) \ 31 term = (struct type##_error *)&(((char *)tbl)[term_offset]); \ 32 strncpy(tbl->repl.name, info->name, sizeof(tbl->repl.name)); \ 34 tbl->repl.valid_hooks = hook_mask; \ 35 tbl->repl.num_entries = nhooks + 1; \ 36 tbl->repl.size = nhooks * sizeof(struct type##_standard) + \ 41 tbl->repl.hook_entry[hooknum] = bytes; \ [all …]
|
/net/mac80211/ |
D | mesh_pathtbl.c | 21 static void mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath); 48 struct mesh_table *tbl = tblptr; in mesh_path_rht_free() local 50 mesh_path_free_rcu(tbl, mpath); in mesh_path_rht_free() 68 static void mesh_table_free(struct mesh_table *tbl) in mesh_table_free() argument 70 rhashtable_free_and_destroy(&tbl->rhead, in mesh_table_free() 71 mesh_path_rht_free, tbl); in mesh_table_free() 72 kfree(tbl); in mesh_table_free() 213 static struct mesh_path *mpath_lookup(struct mesh_table *tbl, const u8 *dst, in mpath_lookup() argument 218 mpath = rhashtable_lookup_fast(&tbl->rhead, dst, mesh_rht_params); in mpath_lookup() 250 __mesh_path_lookup_by_idx(struct mesh_table *tbl, int idx) in __mesh_path_lookup_by_idx() argument [all …]
|
/net/sctp/ |
D | sysctl.c | 334 struct ctl_table tbl; in proc_sctp_do_hmac_alg() local 340 memset(&tbl, 0, sizeof(struct ctl_table)); in proc_sctp_do_hmac_alg() 343 tbl.data = tmp; in proc_sctp_do_hmac_alg() 344 tbl.maxlen = sizeof(tmp); in proc_sctp_do_hmac_alg() 346 tbl.data = net->sctp.sctp_hmac_alg ? : none; in proc_sctp_do_hmac_alg() 347 tbl.maxlen = strlen(tbl.data); in proc_sctp_do_hmac_alg() 350 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); in proc_sctp_do_hmac_alg() 382 struct ctl_table tbl; in proc_sctp_do_rto_min() local 385 memset(&tbl, 0, sizeof(struct ctl_table)); in proc_sctp_do_rto_min() 386 tbl.maxlen = sizeof(unsigned int); in proc_sctp_do_rto_min() [all …]
|
/net/ipv4/ |
D | sysctl_net_ipv4.c | 223 struct ctl_table tbl = { in proc_tcp_congestion_control() local 231 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); in proc_tcp_congestion_control() 242 struct ctl_table tbl = { .maxlen = TCP_CA_BUF_MAX, }; in proc_tcp_available_congestion_control() local 245 tbl.data = kmalloc(tbl.maxlen, GFP_USER); in proc_tcp_available_congestion_control() 246 if (!tbl.data) in proc_tcp_available_congestion_control() 248 tcp_get_available_congestion_control(tbl.data, TCP_CA_BUF_MAX); in proc_tcp_available_congestion_control() 249 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); in proc_tcp_available_congestion_control() 250 kfree(tbl.data); in proc_tcp_available_congestion_control() 259 struct ctl_table tbl = { .maxlen = TCP_CA_BUF_MAX }; in proc_allowed_congestion_control() local 262 tbl.data = kmalloc(tbl.maxlen, GFP_USER); in proc_allowed_congestion_control() [all …]
|
D | udp_diag.c | 34 static int udp_dump_one(struct udp_table *tbl, struct sk_buff *in_skb, in udp_dump_one() argument 48 req->id.idiag_if, 0, tbl, NULL); in udp_dump_one() 56 req->id.idiag_if, 0, tbl, NULL); in udp_dump_one() 174 struct udp_table *tbl) in __udp_diag_destroy() argument 186 req->id.idiag_if, 0, tbl, NULL); in __udp_diag_destroy() 194 req->id.idiag_if, 0, tbl, NULL); in __udp_diag_destroy() 202 req->id.idiag_if, 0, tbl, NULL); in __udp_diag_destroy()
|
D | fib_rules.c | 112 struct fib_table *tbl; in fib4_rule_action() local 133 tbl = fib_get_table(rule->fr_net, tb_id); in fib4_rule_action() 134 if (tbl) in fib4_rule_action() 135 err = fib_table_lookup(tbl, &flp->u.ip4, in fib4_rule_action()
|
D | devinet.c | 2383 struct ctl_table *tbl = ctl_forward_entry; in devinet_init_net() local 2401 tbl = kmemdup(tbl, sizeof(ctl_forward_entry), GFP_KERNEL); in devinet_init_net() 2402 if (!tbl) in devinet_init_net() 2405 tbl[0].data = &all->data[IPV4_DEVCONF_FORWARDING - 1]; in devinet_init_net() 2406 tbl[0].extra1 = all; in devinet_init_net() 2407 tbl[0].extra2 = net; in devinet_init_net() 2422 forw_hdr = register_net_sysctl(net, "net/ipv4", tbl); in devinet_init_net() 2438 if (tbl != ctl_forward_entry) in devinet_init_net() 2439 kfree(tbl); in devinet_init_net() 2454 struct ctl_table *tbl; in devinet_exit_net() local [all …]
|
D | route.c | 3012 struct ctl_table *tbl; in sysctl_route_net_init() local 3014 tbl = ipv4_route_flush_table; in sysctl_route_net_init() 3016 tbl = kmemdup(tbl, sizeof(ipv4_route_flush_table), GFP_KERNEL); in sysctl_route_net_init() 3017 if (!tbl) in sysctl_route_net_init() 3022 tbl[0].procname = NULL; in sysctl_route_net_init() 3024 tbl[0].extra1 = net; in sysctl_route_net_init() 3026 net->ipv4.route_hdr = register_net_sysctl(net, "net/ipv4/route", tbl); in sysctl_route_net_init() 3032 if (tbl != ipv4_route_flush_table) in sysctl_route_net_init() 3033 kfree(tbl); in sysctl_route_net_init() 3040 struct ctl_table *tbl; in sysctl_route_net_exit() local [all …]
|
D | arp.c | 165 .tbl = &arp_tbl, 1121 struct neigh_table *tbl = &arp_tbl; in arp_invalidate() local 1128 write_lock_bh(&tbl->lock); in arp_invalidate() 1130 neigh_remove_one(neigh, tbl); in arp_invalidate() 1131 write_unlock_bh(&tbl->lock); in arp_invalidate()
|
/net/openvswitch/ |
D | flow_table.c | 486 struct sw_flow *ovs_flow_tbl_lookup_stats(struct flow_table *tbl, in ovs_flow_tbl_lookup_stats() argument 490 struct table_instance *ti = rcu_dereference_ovsl(tbl->ti); in ovs_flow_tbl_lookup_stats() 495 list_for_each_entry_rcu(mask, &tbl->mask_list, list) { in ovs_flow_tbl_lookup_stats() 504 struct sw_flow *ovs_flow_tbl_lookup(struct flow_table *tbl, in ovs_flow_tbl_lookup() argument 509 return ovs_flow_tbl_lookup_stats(tbl, key, &n_mask_hit); in ovs_flow_tbl_lookup() 512 struct sw_flow *ovs_flow_tbl_lookup_exact(struct flow_table *tbl, in ovs_flow_tbl_lookup_exact() argument 515 struct table_instance *ti = rcu_dereference_ovsl(tbl->ti); in ovs_flow_tbl_lookup_exact() 520 list_for_each_entry(mask, &tbl->mask_list, list) { in ovs_flow_tbl_lookup_exact() 551 struct sw_flow *ovs_flow_tbl_lookup_ufid(struct flow_table *tbl, in ovs_flow_tbl_lookup_ufid() argument 554 struct table_instance *ti = rcu_dereference_ovsl(tbl->ufid_ti); in ovs_flow_tbl_lookup_ufid() [all …]
|
D | flow_table.h | 81 struct sw_flow *ovs_flow_tbl_lookup_exact(struct flow_table *tbl,
|
/net/decnet/ |
D | dn_rules.c | 73 struct dn_fib_table *tbl; in dn_fib_rule_action() local 93 tbl = dn_fib_get_table(rule->table, 0); in dn_fib_rule_action() 94 if (tbl == NULL) in dn_fib_rule_action() 97 err = tbl->lookup(tbl, fld, (struct dn_fib_res *)arg->result); in dn_fib_rule_action()
|
/net/rds/ |
D | tcp.c | 422 struct ctl_table *tbl; in rds_tcp_init_net() local 431 tbl = rds_tcp_sysctl_table; in rds_tcp_init_net() 433 tbl = kmemdup(rds_tcp_sysctl_table, in rds_tcp_init_net() 435 if (!tbl) { in rds_tcp_init_net() 439 rtn->ctl_table = tbl; in rds_tcp_init_net() 441 tbl[RDS_TCP_SNDBUF].data = &rtn->sndbuf_size; in rds_tcp_init_net() 442 tbl[RDS_TCP_RCVBUF].data = &rtn->rcvbuf_size; in rds_tcp_init_net() 443 rtn->rds_tcp_sysctl = register_net_sysctl(net, "net/rds/tcp", tbl); in rds_tcp_init_net() 462 kfree(tbl); in rds_tcp_init_net()
|
/net/netlabel/ |
D | netlabel_domainhash.c | 50 struct list_head *tbl; member 159 bkt_list = &netlbl_domhsh_rcu_deref(netlbl_domhsh)->tbl[bkt]; in netlbl_domhsh_search() 386 hsh_tbl->tbl = kcalloc(hsh_tbl->size, in netlbl_domhsh_init() 389 if (hsh_tbl->tbl == NULL) { in netlbl_domhsh_init() 394 INIT_LIST_HEAD(&hsh_tbl->tbl[iter]); in netlbl_domhsh_init() 450 &rcu_dereference(netlbl_domhsh)->tbl[bkt]); in netlbl_domhsh_add() 967 iter_list = &rcu_dereference(netlbl_domhsh)->tbl[iter_bkt]; in netlbl_domhsh_walk()
|
D | netlabel_unlabeled.c | 77 struct list_head *tbl; member 223 bkt_list = &netlbl_unlhsh_rcu_deref(netlbl_unlhsh)->tbl[bkt]; in netlbl_unlhsh_search_iface() 345 &netlbl_unlhsh_rcu_deref(netlbl_unlhsh)->tbl[bkt]); in netlbl_unlhsh_add_iface() 1199 iter_list = &rcu_dereference(netlbl_unlhsh)->tbl[iter_bkt]; in netlbl_unlabel_staticlist() 1434 hsh_tbl->tbl = kcalloc(hsh_tbl->size, in netlbl_unlabel_init() 1437 if (hsh_tbl->tbl == NULL) { in netlbl_unlabel_init() 1442 INIT_LIST_HEAD(&hsh_tbl->tbl[iter]); in netlbl_unlabel_init()
|
/net/netlink/ |
D | diag.c | 92 struct netlink_table *tbl = &nl_table[protocol]; in __netlink_diag_dump() local 116 rhashtable_walk_enter(&tbl->hash, hti); in __netlink_diag_dump() 159 sk_for_each_bound(sk, &tbl->mc_list) { in __netlink_diag_dump()
|
D | af_netlink.c | 513 struct netlink_table *tbl = &nl_table[sk->sk_protocol]; in netlink_update_listeners() local 518 listeners = nl_deref_protected(tbl->listeners); in netlink_update_listeners() 522 for (i = 0; i < NLGRPLONGS(tbl->groups); i++) { in netlink_update_listeners() 524 sk_for_each_bound(sk, &tbl->mc_list) { in netlink_update_listeners() 2073 struct netlink_table *tbl = &nl_table[sk->sk_protocol]; in __netlink_change_ngroups() local 2078 if (NLGRPSZ(tbl->groups) < NLGRPSZ(groups)) { in __netlink_change_ngroups() 2082 old = nl_deref_protected(tbl->listeners); in __netlink_change_ngroups() 2083 memcpy(new->masks, old->masks, NLGRPSZ(tbl->groups)); in __netlink_change_ngroups() 2084 rcu_assign_pointer(tbl->listeners, new); in __netlink_change_ngroups() 2088 tbl->groups = groups; in __netlink_change_ngroups() [all …]
|
/net/ipv6/ |
D | ip6_fib.c | 2101 struct fib6_table *tbl; member 2152 iter->w.root = &iter->tbl->tb6_root; in ipv6_route_seq_setup_walk() 2161 static struct fib6_table *ipv6_route_seq_next_table(struct fib6_table *tbl, in ipv6_route_seq_next_table() argument 2167 if (tbl) { in ipv6_route_seq_next_table() 2168 h = (tbl->tb6_id & (FIB6_TABLE_HASHSZ - 1)) + 1; in ipv6_route_seq_next_table() 2169 node = rcu_dereference_bh(hlist_next_rcu(&tbl->tb6_hlist)); in ipv6_route_seq_next_table() 2211 read_lock(&iter->tbl->tb6_lock); in ipv6_route_seq_next() 2213 read_unlock(&iter->tbl->tb6_lock); in ipv6_route_seq_next() 2224 iter->tbl = ipv6_route_seq_next_table(iter->tbl, net); in ipv6_route_seq_next() 2225 if (!iter->tbl) in ipv6_route_seq_next() [all …]
|
/net/bridge/ |
D | br_vlan_tunnel.c | 42 static struct net_bridge_vlan *br_vlan_tunnel_lookup(struct rhashtable *tbl, in br_vlan_tunnel_lookup() argument 45 return rhashtable_lookup_fast(tbl, &tunnel_id, in br_vlan_tunnel_lookup()
|
/net/802/ |
D | hippi.c | 142 if (p->tbl->family != AF_INET6) in hippi_neigh_setup_dev()
|