/net/sched/ |
D | sch_gred.c | 62 static inline int gred_wred_mode(struct gred_sched *table) in gred_wred_mode() argument 64 return test_bit(GRED_WRED_MODE, &table->flags); in gred_wred_mode() 67 static inline void gred_enable_wred_mode(struct gred_sched *table) in gred_enable_wred_mode() argument 69 __set_bit(GRED_WRED_MODE, &table->flags); in gred_enable_wred_mode() 72 static inline void gred_disable_wred_mode(struct gred_sched *table) in gred_disable_wred_mode() argument 74 __clear_bit(GRED_WRED_MODE, &table->flags); in gred_disable_wred_mode() 77 static inline int gred_rio_mode(struct gred_sched *table) in gred_rio_mode() argument 79 return test_bit(GRED_RIO_MODE, &table->flags); in gred_rio_mode() 82 static inline void gred_enable_rio_mode(struct gred_sched *table) in gred_enable_rio_mode() argument 84 __set_bit(GRED_RIO_MODE, &table->flags); in gred_enable_rio_mode() [all …]
|
/net/devlink/ |
D | dpipe.c | 87 static int devlink_dpipe_matches_put(struct devlink_dpipe_table *table, in devlink_dpipe_matches_put() argument 97 if (table->table_ops->matches_dump(table->priv, skb)) in devlink_dpipe_matches_put() 135 static int devlink_dpipe_actions_put(struct devlink_dpipe_table *table, in devlink_dpipe_actions_put() argument 145 if (table->table_ops->actions_dump(table->priv, skb)) in devlink_dpipe_actions_put() 157 struct devlink_dpipe_table *table) in devlink_dpipe_table_put() argument 162 table_size = table->table_ops->size_get(table->priv); in devlink_dpipe_table_put() 167 if (nla_put_string(skb, DEVLINK_ATTR_DPIPE_TABLE_NAME, table->name) || in devlink_dpipe_table_put() 172 table->counters_enabled)) in devlink_dpipe_table_put() 175 if (table->resource_valid) { in devlink_dpipe_table_put() 177 table->resource_id, DEVLINK_ATTR_PAD) || in devlink_dpipe_table_put() [all …]
|
/net/xfrm/ |
D | xfrm_sysctl.c | 46 struct ctl_table *table; in xfrm_sysctl_init() local 51 table = kmemdup(xfrm_table, sizeof(xfrm_table), GFP_KERNEL); in xfrm_sysctl_init() 52 if (!table) in xfrm_sysctl_init() 54 table[0].data = &net->xfrm.sysctl_aevent_etime; in xfrm_sysctl_init() 55 table[1].data = &net->xfrm.sysctl_aevent_rseqth; in xfrm_sysctl_init() 56 table[2].data = &net->xfrm.sysctl_larval_drop; in xfrm_sysctl_init() 57 table[3].data = &net->xfrm.sysctl_acq_expires; in xfrm_sysctl_init() 61 table[0].procname = NULL; in xfrm_sysctl_init() 65 net->xfrm.sysctl_hdr = register_net_sysctl_sz(net, "net/core", table, in xfrm_sysctl_init() 72 kfree(table); in xfrm_sysctl_init() [all …]
|
/net/netfilter/ |
D | nf_tables_api.c | 74 struct nft_table *table; member 109 static void nft_validate_state_update(struct nft_table *table, u8 new_validate_state) in nft_validate_state_update() argument 111 switch (table->validate_state) { in nft_validate_state_update() 122 table->validate_state = new_validate_state; in nft_validate_state_update() 135 struct nft_table *table, in nft_ctx_init() argument 142 ctx->table = table; in nft_ctx_init() 319 const struct nft_table *table, in nf_tables_register_hook() argument 325 if (table->flags & NFT_TABLE_F_DORMANT || in nf_tables_register_hook() 335 if (nft_base_chain_netdev(table->family, basechain->ops.hooknum)) in nf_tables_register_hook() 342 const struct nft_table *table, in __nf_tables_unregister_hook() argument [all …]
|
D | nf_hooks_lwtunnel.c | 31 int nf_hooks_lwtunnel_sysctl_handler(struct ctl_table *table, int write, in nf_hooks_lwtunnel_sysctl_handler() argument 36 .procname = table->procname, in nf_hooks_lwtunnel_sysctl_handler() 39 .mode = table->mode, in nf_hooks_lwtunnel_sysctl_handler() 70 struct ctl_table *table; in nf_lwtunnel_net_init() local 72 table = nf_lwtunnel_sysctl_table; in nf_lwtunnel_net_init() 74 table = kmemdup(nf_lwtunnel_sysctl_table, in nf_lwtunnel_net_init() 77 if (!table) in nf_lwtunnel_net_init() 81 hdr = register_net_sysctl_sz(net, "net/netfilter", table, in nf_lwtunnel_net_init() 91 kfree(table); in nf_lwtunnel_net_init() 98 const struct ctl_table *table; in nf_lwtunnel_net_exit() local [all …]
|
D | nf_conntrack_standalone.c | 527 nf_conntrack_hash_sysctl(struct ctl_table *table, int write, in nf_conntrack_hash_sysctl() argument 535 ret = proc_dointvec(table, write, buffer, lenp, ppos); in nf_conntrack_hash_sysctl() 960 struct ctl_table *table) in nf_conntrack_standalone_init_tcp_sysctl() argument 965 table[NF_SYSCTL_CT_PROTO_TIMEOUT_TCP_ ## XNAME].data = \ in nf_conntrack_standalone_init_tcp_sysctl() 980 table[NF_SYSCTL_CT_PROTO_TCP_ ## XNAME].data = (rval) in nf_conntrack_standalone_init_tcp_sysctl() 989 table[NF_SYSCTL_CT_PROTO_TIMEOUT_TCP_OFFLOAD].data = &tn->offload_timeout; in nf_conntrack_standalone_init_tcp_sysctl() 995 struct ctl_table *table) in nf_conntrack_standalone_init_sctp_sysctl() argument 1001 table[NF_SYSCTL_CT_PROTO_TIMEOUT_SCTP_ ## XNAME].data = \ in nf_conntrack_standalone_init_sctp_sysctl() 1017 struct ctl_table *table) in nf_conntrack_standalone_init_dccp_sysctl() argument 1023 table[NF_SYSCTL_CT_PROTO_TIMEOUT_DCCP_ ## XNAME].data = \ in nf_conntrack_standalone_init_dccp_sysctl() [all …]
|
D | x_tables.c | 497 if (par->match->table != NULL && in xt_check_match() 498 strcmp(par->match->table, par->table) != 0) { in xt_check_match() 501 par->match->table, par->table); in xt_check_match() 1012 if (par->target->table != NULL && in xt_check_target() 1013 strcmp(par->target->table, par->table) != 0) { in xt_check_target() 1016 par->target->table, par->table); in xt_check_target() 1299 void xt_table_unlock(struct xt_table *table) in xt_table_unlock() argument 1301 mutex_unlock(&xt[table->af].mutex); in xt_table_unlock() 1384 xt_replace_table(struct xt_table *table, in xt_replace_table() argument 1401 private = table->private; in xt_replace_table() [all …]
|
/net/openvswitch/ |
D | flow_table.c | 103 int ovs_flow_tbl_count(const struct flow_table *table) in ovs_flow_tbl_count() argument 105 return table->count; in ovs_flow_tbl_count() 392 int ovs_flow_tbl_masks_cache_resize(struct flow_table *table, u32 size) in ovs_flow_tbl_masks_cache_resize() argument 394 struct mask_cache *mc = rcu_dereference_ovsl(table->mask_cache); in ovs_flow_tbl_masks_cache_resize() 408 rcu_assign_pointer(table->mask_cache, new); in ovs_flow_tbl_masks_cache_resize() 414 int ovs_flow_tbl_init(struct flow_table *table) in ovs_flow_tbl_init() argument 436 rcu_assign_pointer(table->ti, ti); in ovs_flow_tbl_init() 437 rcu_assign_pointer(table->ufid_ti, ufid_ti); in ovs_flow_tbl_init() 438 rcu_assign_pointer(table->mask_array, ma); in ovs_flow_tbl_init() 439 rcu_assign_pointer(table->mask_cache, mc); in ovs_flow_tbl_init() [all …]
|
D | flow_table.h | 81 int ovs_flow_tbl_count(const struct flow_table *table); 82 void ovs_flow_tbl_destroy(struct flow_table *table); 85 int ovs_flow_tbl_insert(struct flow_table *table, struct sw_flow *flow, 87 void ovs_flow_tbl_remove(struct flow_table *table, struct sw_flow *flow); 88 int ovs_flow_tbl_num_masks(const struct flow_table *table); 89 u32 ovs_flow_tbl_masks_cache_size(const struct flow_table *table); 90 int ovs_flow_tbl_masks_cache_resize(struct flow_table *table, u32 size); 91 struct sw_flow *ovs_flow_tbl_dump_next(struct table_instance *table, 110 void ovs_flow_masks_rebalance(struct flow_table *table); 111 void table_instance_flow_flush(struct flow_table *table,
|
/net/unix/ |
D | sysctl_net_unix.c | 27 struct ctl_table *table; in unix_sysctl_register() local 30 table = unix_table; in unix_sysctl_register() 32 table = kmemdup(unix_table, sizeof(unix_table), GFP_KERNEL); in unix_sysctl_register() 33 if (!table) in unix_sysctl_register() 36 table[0].data = &net->unx.sysctl_max_dgram_qlen; in unix_sysctl_register() 39 net->unx.ctl = register_net_sysctl_sz(net, "net/unix", table, in unix_sysctl_register() 48 kfree(table); in unix_sysctl_register() 55 struct ctl_table *table; in unix_sysctl_unregister() local 57 table = net->unx.ctl->ctl_table_arg; in unix_sysctl_unregister() 60 kfree(table); in unix_sysctl_unregister()
|
/net/mptcp/ |
D | ctrl.c | 187 struct ctl_table *table; in mptcp_pernet_new_table() local 189 table = mptcp_sysctl_table; in mptcp_pernet_new_table() 191 table = kmemdup(table, sizeof(mptcp_sysctl_table), GFP_KERNEL); in mptcp_pernet_new_table() 192 if (!table) in mptcp_pernet_new_table() 196 table[0].data = &pernet->mptcp_enabled; in mptcp_pernet_new_table() 197 table[1].data = &pernet->add_addr_timeout; in mptcp_pernet_new_table() 198 table[2].data = &pernet->checksum_enabled; in mptcp_pernet_new_table() 199 table[3].data = &pernet->allow_join_initial_addr_port; in mptcp_pernet_new_table() 200 table[4].data = &pernet->stale_loss_cnt; in mptcp_pernet_new_table() 201 table[5].data = &pernet->pm_type; in mptcp_pernet_new_table() [all …]
|
/net/smc/ |
D | smc_sysctl.c | 76 struct ctl_table *table; in smc_sysctl_net_init() local 78 table = smc_table; in smc_sysctl_net_init() 82 table = kmemdup(table, sizeof(smc_table), GFP_KERNEL); in smc_sysctl_net_init() 83 if (!table) in smc_sysctl_net_init() 87 table[i].data += (void *)net - (void *)&init_net; in smc_sysctl_net_init() 90 net->smc.smc_hdr = register_net_sysctl_sz(net, "net/smc", table, in smc_sysctl_net_init() 105 kfree(table); in smc_sysctl_net_init() 112 struct ctl_table *table; in smc_sysctl_net_exit() local 114 table = net->smc.smc_hdr->ctl_table_arg; in smc_sysctl_net_exit() 117 kfree(table); in smc_sysctl_net_exit()
|
/net/ipv6/netfilter/ |
D | ip6table_raw.c | 39 const struct xt_table *table = &packet_raw; in ip6table_raw_table_init() local 43 table = &packet_raw_before_defrag; in ip6table_raw_table_init() 45 repl = ip6t_alloc_initial_table(table); in ip6table_raw_table_init() 48 ret = ip6t_register_table(net, table, repl, rawtable_ops); in ip6table_raw_table_init() 70 const struct xt_table *table = &packet_raw; in ip6table_raw_init() local 74 table = &packet_raw_before_defrag; in ip6table_raw_init() 78 ret = xt_register_template(table, ip6table_raw_table_init); in ip6table_raw_init() 83 rawtable_ops = xt_hook_ops_alloc(table, ip6t_do_table); in ip6table_raw_init() 85 xt_unregister_template(table); in ip6table_raw_init() 92 xt_unregister_template(table); in ip6table_raw_init()
|
/net/ipv4/netfilter/ |
D | iptable_raw.c | 40 const struct xt_table *table = &packet_raw; in iptable_raw_table_init() local 44 table = &packet_raw_before_defrag; in iptable_raw_table_init() 46 repl = ipt_alloc_initial_table(table); in iptable_raw_table_init() 49 ret = ipt_register_table(net, table, repl, rawtable_ops); in iptable_raw_table_init() 72 const struct xt_table *table = &packet_raw; in iptable_raw_init() local 75 table = &packet_raw_before_defrag; in iptable_raw_init() 80 ret = xt_register_template(table, in iptable_raw_init() 85 rawtable_ops = xt_hook_ops_alloc(table, ipt_do_table); in iptable_raw_init() 87 xt_unregister_template(table); in iptable_raw_init() 93 xt_unregister_template(table); in iptable_raw_init()
|
/net/ipv4/ |
D | udp_tunnel_nic.c | 145 const struct udp_tunnel_nic_table_info *table; in udp_tunnel_nic_should_replay() local 152 table = &dev->udp_tunnel_nic_info->tables[i]; in udp_tunnel_nic_should_replay() 156 for (j = 0; j < table->n_entries; j++) in udp_tunnel_nic_should_replay() 165 __udp_tunnel_nic_get_port(struct net_device *dev, unsigned int table, in __udp_tunnel_nic_get_port() argument 172 entry = &utn->entries[table][idx]; in __udp_tunnel_nic_get_port() 179 __udp_tunnel_nic_set_port_priv(struct net_device *dev, unsigned int table, in __udp_tunnel_nic_set_port_priv() argument 182 dev->udp_tunnel_nic->entries[table][idx].hw_priv = priv; in __udp_tunnel_nic_set_port_priv() 211 unsigned int table, unsigned int idx) in udp_tunnel_nic_device_sync_one() argument 217 entry = &utn->entries[table][idx]; in udp_tunnel_nic_device_sync_one() 223 err = dev->udp_tunnel_nic_info->set_port(dev, table, idx, &ti); in udp_tunnel_nic_device_sync_one() [all …]
|
D | sysctl_net_ipv4.c | 68 static int ipv4_local_port_range(struct ctl_table *table, int write, in ipv4_local_port_range() argument 72 container_of(table->data, struct net, ipv4.ip_local_ports.range); in ipv4_local_port_range() 78 .mode = table->mode, in ipv4_local_port_range() 103 static int ipv4_privileged_ports(struct ctl_table *table, int write, in ipv4_privileged_ports() argument 106 struct net *net = container_of(table->data, struct net, in ipv4_privileged_ports() 114 .mode = table->mode, in ipv4_privileged_ports() 137 static void inet_get_ping_group_range_table(struct ctl_table *table, kgid_t *low, kgid_t *high) in inet_get_ping_group_range_table() argument 139 kgid_t *data = table->data; in inet_get_ping_group_range_table() 141 container_of(table->data, struct net, ipv4.ping_group_range.range); in inet_get_ping_group_range_table() 152 static void set_ping_group_range(struct ctl_table *table, kgid_t low, kgid_t high) in set_ping_group_range() argument [all …]
|
D | xfrm4_policy.c | 160 struct ctl_table *table; in xfrm4_net_sysctl_init() local 163 table = xfrm4_policy_table; in xfrm4_net_sysctl_init() 165 table = kmemdup(table, sizeof(xfrm4_policy_table), GFP_KERNEL); in xfrm4_net_sysctl_init() 166 if (!table) in xfrm4_net_sysctl_init() 169 table[0].data = &net->xfrm.xfrm4_dst_ops.gc_thresh; in xfrm4_net_sysctl_init() 172 hdr = register_net_sysctl_sz(net, "net/ipv4", table, in xfrm4_net_sysctl_init() 182 kfree(table); in xfrm4_net_sysctl_init() 189 struct ctl_table *table; in xfrm4_net_sysctl_exit() local 194 table = net->ipv4.xfrm4_hdr->ctl_table_arg; in xfrm4_net_sysctl_exit() 197 kfree(table); in xfrm4_net_sysctl_exit()
|
/net/ |
D | sysctl_net.c | 43 struct ctl_table *table) in net_ctl_permissions() argument 49 int mode = (table->mode >> 6) & 7; in net_ctl_permissions() 53 return table->mode; in net_ctl_permissions() 57 struct ctl_table *table, in net_ctl_set_ownership() argument 125 struct ctl_table *table, size_t table_size) in ensure_safe_net_sysctl() argument 130 ent = table; in ensure_safe_net_sysctl() 166 struct ctl_table *table, in register_net_sysctl_sz() argument 173 ensure_safe_net_sysctl(net, path, table, table_size); in register_net_sysctl_sz() 175 entry = table; in register_net_sysctl_sz() 179 return __register_sysctl_table(&net->sysctls, path, table, count); in register_net_sysctl_sz()
|
/net/bridge/netfilter/ |
D | ebtables.c | 195 struct ebt_table *table = priv; in ebt_do_table() local 211 read_lock_bh(&table->lock); in ebt_do_table() 212 private = table->private; in ebt_do_table() 233 read_unlock_bh(&table->lock); in ebt_do_table() 254 read_unlock_bh(&table->lock); in ebt_do_table() 258 read_unlock_bh(&table->lock); in ebt_do_table() 282 read_unlock_bh(&table->lock); in ebt_do_table() 294 read_unlock_bh(&table->lock); in ebt_do_table() 312 read_unlock_bh(&table->lock); in ebt_do_table() 315 read_unlock_bh(&table->lock); in ebt_do_table() [all …]
|
/net/ipv6/ |
D | ip6_fib.c | 58 struct fib6_table *table, 61 struct fib6_table *table, 212 static void fib6_free_table(struct fib6_table *table) in fib6_free_table() argument 214 inetpeer_invalidate_tree(&table->tb6_peers); in fib6_free_table() 215 kfree(table); in fib6_free_table() 240 struct fib6_table *table; in fib6_alloc_table() local 242 table = kzalloc(sizeof(*table), GFP_ATOMIC); in fib6_alloc_table() 243 if (table) { in fib6_alloc_table() 244 table->tb6_id = id; in fib6_alloc_table() 245 rcu_assign_pointer(table->tb6_root.leaf, in fib6_alloc_table() [all …]
|
/net/6lowpan/ |
D | debugfs.c | 70 container_of(ctx, struct lowpan_iphc_ctx_table, table[ctx->id]); in lowpan_ctx_plen_set() 86 container_of(ctx, struct lowpan_iphc_ctx_table, table[ctx->id]); in lowpan_ctx_plen_get() 101 container_of(ctx, struct lowpan_iphc_ctx_table, table[ctx->id]); in lowpan_ctx_pfx_show() 131 container_of(ctx, struct lowpan_iphc_ctx_table, table[ctx->id]); in lowpan_ctx_pfx_write() 180 debugfs_create_file("active", 0644, root, &ldev->ctx.table[id], in lowpan_dev_debugfs_ctx_init() 183 debugfs_create_file("compression", 0644, root, &ldev->ctx.table[id], in lowpan_dev_debugfs_ctx_init() 186 debugfs_create_file("prefix", 0644, root, &ldev->ctx.table[id], in lowpan_dev_debugfs_ctx_init() 189 debugfs_create_file("prefix_len", 0644, root, &ldev->ctx.table[id], in lowpan_dev_debugfs_ctx_init() 203 if (!lowpan_iphc_ctx_is_active(&t->table[i])) in lowpan_context_show() 206 seq_printf(file, "%3d|%39pI6c/%-3d|%d\n", t->table[i].id, in lowpan_context_show() [all …]
|
/net/core/ |
D | sysctl_net_core.c | 95 static int rps_default_mask_sysctl(struct ctl_table *table, int write, in rps_default_mask_sysctl() argument 98 struct net *net = (struct net *)table->data; in rps_default_mask_sysctl() 126 static int rps_sock_flow_sysctl(struct ctl_table *table, int write, in rps_sock_flow_sysctl() argument 134 .mode = table->mode in rps_sock_flow_sysctl() 195 static int flow_limit_cpu_sysctl(struct ctl_table *table, int write, in flow_limit_cpu_sysctl() argument 252 static int flow_limit_table_len_sysctl(struct ctl_table *table, int write, in flow_limit_table_len_sysctl() argument 260 ptr = table->data; in flow_limit_table_len_sysctl() 262 ret = proc_dointvec(table, write, buffer, lenp, ppos); in flow_limit_table_len_sysctl() 274 static int set_default_qdisc(struct ctl_table *table, int write, in set_default_qdisc() argument 293 static int proc_do_dev_weight(struct ctl_table *table, int write, in proc_do_dev_weight() argument [all …]
|
/net/ax25/ |
D | sysctl_net_ax25.c | 152 struct ctl_table *table; in ax25_register_dev_sysctl() local 154 table = kmemdup(ax25_param_table, sizeof(ax25_param_table), GFP_KERNEL); in ax25_register_dev_sysctl() 155 if (!table) in ax25_register_dev_sysctl() 159 table[k].data = &ax25_dev->values[k]; in ax25_register_dev_sysctl() 162 ax25_dev->sysheader = register_net_sysctl_sz(&init_net, path, table, in ax25_register_dev_sysctl() 165 kfree(table); in ax25_register_dev_sysctl() 174 struct ctl_table *table; in ax25_unregister_dev_sysctl() local 178 table = header->ctl_table_arg; in ax25_unregister_dev_sysctl() 180 kfree(table); in ax25_unregister_dev_sysctl()
|
/net/netfilter/ipvs/ |
D | Kconfig | 46 int "IPVS connection table size (the Nth power of 2)" 51 The IPVS connection hash table uses the chaining scheme to handle 52 hash collisions. Using a big IPVS connection hash table will greatly 54 in the hash table. 56 Note the table size must be power of 2. The table size will be the 59 which means the table size is 4096. Don't input the number too 61 table size yourself, according to your virtual server application. 62 It is good to set the table size not far less than the number of 64 connection in the table. For example, your virtual server gets 200 66 average in the connection table, the table size should be not far [all …]
|
/net/ieee802154/6lowpan/ |
D | reassembly.c | 359 struct ctl_table *table; in lowpan_frags_ns_sysctl_register() local 365 table = lowpan_frags_ns_ctl_table; in lowpan_frags_ns_sysctl_register() 367 table = kmemdup(table, sizeof(lowpan_frags_ns_ctl_table), in lowpan_frags_ns_sysctl_register() 369 if (table == NULL) in lowpan_frags_ns_sysctl_register() 374 table[0].procname = NULL; in lowpan_frags_ns_sysctl_register() 379 table[0].data = &ieee802154_lowpan->fqdir->high_thresh; in lowpan_frags_ns_sysctl_register() 380 table[0].extra1 = &ieee802154_lowpan->fqdir->low_thresh; in lowpan_frags_ns_sysctl_register() 381 table[1].data = &ieee802154_lowpan->fqdir->low_thresh; in lowpan_frags_ns_sysctl_register() 382 table[1].extra2 = &ieee802154_lowpan->fqdir->high_thresh; in lowpan_frags_ns_sysctl_register() 383 table[2].data = &ieee802154_lowpan->fqdir->timeout; in lowpan_frags_ns_sysctl_register() [all …]
|