/net/tipc/ |
D | ref.c | 65 struct reference *entries; member 118 tipc_ref_table.entries = table; in tipc_ref_table_init() 135 if (!tipc_ref_table.entries) in tipc_ref_table_stop() 138 vfree(tipc_ref_table.entries); in tipc_ref_table_stop() 139 tipc_ref_table.entries = NULL; in tipc_ref_table_stop() 166 if (!tipc_ref_table.entries) { in tipc_ref_acquire() 176 entry = &(tipc_ref_table.entries[index]); in tipc_ref_acquire() 189 entry = &(tipc_ref_table.entries[index]); in tipc_ref_acquire() 218 if (!tipc_ref_table.entries) { in tipc_ref_discard() 225 entry = &(tipc_ref_table.entries[index]); in tipc_ref_discard() [all …]
|
/net/bridge/netfilter/ |
D | ebtables.c | 182 base = private->entries; in ebt_do_table() 410 struct ebt_entry *e = (void *)newinfo->entries + offset; in ebt_verify_pointers() 419 repl->entries + offset) in ebt_verify_pointers() 820 if (newinfo->hook_entry[i] != (struct ebt_entries *)newinfo->entries) { in translate_table() 842 ret = EBT_ENTRY_ITERATE(newinfo->entries, newinfo->entries_size, in translate_table() 884 EBT_ENTRY_ITERATE(newinfo->entries, newinfo->entries_size, in translate_table() 898 cl_s, udc_cnt, i, newinfo->entries)) { in translate_table() 915 ret = EBT_ENTRY_ITERATE(newinfo->entries, newinfo->entries_size, in translate_table() 918 EBT_ENTRY_ITERATE(newinfo->entries, newinfo->entries_size, in translate_table() 986 newinfo->entries = vmalloc(tmp.entries_size); in do_replace() [all …]
|
/net/netfilter/ipvs/ |
D | ip_vs_lblc.c | 101 atomic_t entries; /* number of entries */ member 165 atomic_inc(&tbl->entries); in ip_vs_lblc_hash() 235 atomic_dec(&tbl->entries); in ip_vs_lblc_flush() 258 atomic_dec(&tbl->entries); in ip_vs_lblc_full_check() 293 if (atomic_read(&tbl->entries) <= tbl->max_size) { in ip_vs_lblc_check_expire() 298 goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; in ip_vs_lblc_check_expire() 311 atomic_dec(&tbl->entries); in ip_vs_lblc_check_expire()
|
D | ip_vs_lblcr.c | 280 atomic_t entries; /* number of entries */ member 340 atomic_inc(&tbl->entries); in ip_vs_lblcr_hash() 435 atomic_dec(&tbl->entries); in ip_vs_lblcr_full_check() 470 if (atomic_read(&tbl->entries) <= tbl->max_size) { in ip_vs_lblcr_check_expire() 475 goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; in ip_vs_lblcr_check_expire() 488 atomic_dec(&tbl->entries); in ip_vs_lblcr_check_expire()
|
/net/ipv4/netfilter/ |
D | arp_tables.c | 242 table_base = (void *)private->entries[smp_processor_id()]; in arpt_do_table() 656 if (newinfo->entries[i] && newinfo->entries[i] != entry0) in translate_table() 657 memcpy(newinfo->entries[i], entry0, newinfo->size); in translate_table() 699 ARPT_ENTRY_ITERATE(t->entries[curcpu], in get_counters() 709 ARPT_ENTRY_ITERATE(t->entries[cpu], in get_counters() 756 loc_cpu_entry = private->entries[raw_smp_processor_id()]; in copy_entries_to_user() 850 memcpy(newinfo, info, offsetof(struct xt_table_info, entries)); in compat_table_info() 852 loc_cpu_entry = info->entries[raw_smp_processor_id()]; in compat_table_info() 1010 loc_cpu_old_entry = oldinfo->entries[raw_smp_processor_id()]; in __do_replace() 1050 loc_cpu_entry = newinfo->entries[raw_smp_processor_id()]; in do_replace() [all …]
|
D | iptable_filter.c | 29 struct ipt_standard entries[3]; member 48 .entries = { 160 initial_table.entries[1].target.verdict = -forward - 1; in iptable_filter_init()
|
D | ip_tables.c | 293 table_base = (void *)private->entries[smp_processor_id()]; in trace_packet() 353 table_base = (void *)private->entries[smp_processor_id()]; in ipt_do_table() 865 if (newinfo->entries[i] && newinfo->entries[i] != entry0) in translate_table() 866 memcpy(newinfo->entries[i], entry0, newinfo->size); in translate_table() 911 IPT_ENTRY_ITERATE(t->entries[curcpu], in get_counters() 921 IPT_ENTRY_ITERATE(t->entries[cpu], in get_counters() 972 loc_cpu_entry = private->entries[raw_smp_processor_id()]; in copy_entries_to_user() 1090 memcpy(newinfo, info, offsetof(struct xt_table_info, entries)); in compat_table_info() 1092 loc_cpu_entry = info->entries[raw_smp_processor_id()]; in compat_table_info() 1247 loc_cpu_old_entry = oldinfo->entries[raw_smp_processor_id()]; in __do_replace() [all …]
|
D | iptable_raw.c | 15 struct ipt_standard entries[2]; member 32 .entries = {
|
D | arptable_filter.c | 21 struct arpt_standard entries[3]; member 40 .entries = {
|
D | iptable_security.c | 33 struct ipt_standard entries[3]; member 52 .entries = {
|
D | nf_nat_rule.c | 34 struct ipt_standard entries[3]; member 53 .entries = {
|
D | iptable_mangle.c | 34 struct ipt_standard entries[5]; member 57 .entries = {
|
D | ipt_CLUSTERIP.c | 40 atomic_t entries; /* number of entries/rules member 86 if (atomic_dec_and_test(&c->entries)) { in clusterip_config_entry_put() 130 atomic_inc(&c->entries); in clusterip_config_find_get() 164 atomic_set(&c->entries, 1); in clusterip_config_init()
|
/net/ipv6/netfilter/ |
D | ip6table_filter.c | 27 struct ip6t_standard entries[3]; member 46 .entries = { 153 initial_table.entries[1].target.verdict = -forward - 1; in ip6table_filter_init()
|
D | ip6_tables.c | 325 table_base = (void *)private->entries[smp_processor_id()]; in trace_packet() 379 table_base = (void *)private->entries[smp_processor_id()]; in ip6t_do_table() 894 if (newinfo->entries[i] && newinfo->entries[i] != entry0) in translate_table() 895 memcpy(newinfo->entries[i], entry0, newinfo->size); in translate_table() 940 IP6T_ENTRY_ITERATE(t->entries[curcpu], in get_counters() 950 IP6T_ENTRY_ITERATE(t->entries[cpu], in get_counters() 1001 loc_cpu_entry = private->entries[raw_smp_processor_id()]; in copy_entries_to_user() 1119 memcpy(newinfo, info, offsetof(struct xt_table_info, entries)); in compat_table_info() 1121 loc_cpu_entry = info->entries[raw_smp_processor_id()]; in compat_table_info() 1277 loc_cpu_old_entry = oldinfo->entries[raw_smp_processor_id()]; in __do_replace() [all …]
|
D | ip6table_raw.c | 14 struct ip6t_standard entries[2]; member 31 .entries = {
|
D | ip6table_security.c | 32 struct ip6t_standard entries[3]; member 51 .entries = {
|
D | ip6table_mangle.c | 27 struct ip6t_standard entries[5]; member 50 .entries = {
|
/net/mac80211/ |
D | mesh.c | 173 int entries = 0; in mesh_rmc_check() local 181 ++entries; in mesh_rmc_check() 183 (entries == RMC_QUEUE_MAX_LEN)) { in mesh_rmc_check() 186 --entries; in mesh_rmc_check() 315 atomic_set(&newtbl->entries, 0); in mesh_table_alloc() 342 atomic_dec(&tbl->entries); in mesh_table_free() 366 if (atomic_read(&tbl->entries) in mesh_table_grow() 377 atomic_set(&newtbl->entries, atomic_read(&tbl->entries)); in mesh_table_grow()
|
/net/xfrm/ |
D | xfrm_algo.c | 472 int entries; member 479 .entries = ARRAY_SIZE(aead_list), 486 .entries = ARRAY_SIZE(aalg_list), 493 .entries = ARRAY_SIZE(ealg_list), 500 .entries = ARRAY_SIZE(calg_list), 513 for (i = 0; i < algo_list->entries; i++) { in xfrm_find_algo()
|
/net/can/ |
D | af_can.c | 457 d->entries++; in can_rx_register() 552 d->entries--; in can_rx_unregister() 558 if (d->remove_on_zero_entries && !d->entries) in can_rx_unregister() 590 if (d->entries == 0) in can_rcv_filter() 806 if (d->entries) { in can_notifier()
|
D | af_can.h | 73 int entries; member
|
/net/core/ |
D | dst.c | 167 if (ops->gc && atomic_read(&ops->entries) > ops->gc_thresh) { in dst_alloc() 182 atomic_inc(&ops->entries); in dst_alloc() 234 atomic_dec(&dst->ops->entries); in dst_destroy()
|
D | neighbour.c | 264 int entries; in neigh_alloc() local 266 entries = atomic_inc_return(&tbl->entries) - 1; in neigh_alloc() 267 if (entries >= tbl->gc_thresh3 || in neigh_alloc() 268 (entries >= tbl->gc_thresh2 && in neigh_alloc() 271 entries >= tbl->gc_thresh3) in neigh_alloc() 295 atomic_dec(&tbl->entries); in neigh_alloc() 299 static struct neighbour **neigh_hash_alloc(unsigned int entries) in neigh_hash_alloc() argument 301 unsigned long size = entries * sizeof(struct neighbour *); in neigh_hash_alloc() 313 static void neigh_hash_free(struct neighbour **hash, unsigned int entries) in neigh_hash_free() argument 315 unsigned long size = entries * sizeof(struct neighbour *); in neigh_hash_free() [all …]
|
/net/netfilter/ |
D | xt_recent.c | 76 unsigned int entries; member 142 t->entries--; in recent_entry_remove() 151 if (t->entries >= ip_list_tot) { in recent_entry_init() 170 t->entries++; in recent_entry_init()
|