/net/hsr/ |
D | hsr_framereg.c | 62 struct hsr_node *node; in hsr_addr_is_self() local 64 node = list_first_or_null_rcu(&hsr->self_node_db, struct hsr_node, in hsr_addr_is_self() 66 if (!node) { in hsr_addr_is_self() 71 if (ether_addr_equal(addr, node->MacAddressA)) in hsr_addr_is_self() 73 if (ether_addr_equal(addr, node->MacAddressB)) in hsr_addr_is_self() 84 struct hsr_node *node; in find_node_by_AddrA() local 86 list_for_each_entry_rcu(node, node_db, mac_list) { in find_node_by_AddrA() 87 if (ether_addr_equal(node->MacAddressA, addr)) in find_node_by_AddrA() 88 return node; in find_node_by_AddrA() 102 struct hsr_node *node, *oldnode; in hsr_create_self_node() local [all …]
|
/net/tipc/ |
D | name_distr.c | 46 u32 node; member 88 struct tipc_node *node; in named_cluster_distribute() local 92 list_for_each_entry_rcu(node, &tn->node_list, list) { in named_cluster_distribute() 93 dnode = node->addr; in named_cluster_distribute() 96 if (!tipc_node_is_up(node)) in named_cluster_distribute() 229 struct tipc_node *node; in tipc_publ_subscribe() local 234 node = tipc_node_find(net, addr); in tipc_publ_subscribe() 235 if (!node) { in tipc_publ_subscribe() 241 tipc_node_lock(node); in tipc_publ_subscribe() 242 list_add_tail(&publ->nodesub_list, &node->publ_list); in tipc_publ_subscribe() [all …]
|
D | node.c | 75 static void tipc_node_delete(struct tipc_node *node); 111 void tipc_node_put(struct tipc_node *node) in tipc_node_put() argument 113 kref_put(&node->kref, tipc_node_kref_release); in tipc_node_put() 116 static void tipc_node_get(struct tipc_node *node) in tipc_node_get() argument 118 kref_get(&node->kref); in tipc_node_get() 127 struct tipc_node *node; in tipc_node_find() local 133 hlist_for_each_entry_rcu(node, &tn->node_htable[tipc_hashfn(addr)], in tipc_node_find() 135 if (node->addr == addr) { in tipc_node_find() 136 tipc_node_get(node); in tipc_node_find() 138 return node; in tipc_node_find() [all …]
|
D | name_table.c | 121 u32 scope, u32 node, u32 port_ref, in publ_create() argument 134 publ->node = node; in publ_create() 236 u32 node, u32 port, u32 key) in tipc_nameseq_insert_publ() argument 258 (!publ->node || (publ->node == node))) in tipc_nameseq_insert_publ() 314 publ = publ_create(type, lower, upper, scope, node, port, key); in tipc_nameseq_insert_publ() 321 if (in_own_cluster(net, node)) { in tipc_nameseq_insert_publ() 326 if (in_own_node(net, node)) { in tipc_nameseq_insert_publ() 335 publ->node, created_subseq); in tipc_nameseq_insert_publ() 353 u32 inst, u32 node, in tipc_nameseq_remove_publ() argument 371 (!publ->node || (publ->node == node))) in tipc_nameseq_remove_publ() [all …]
|
D | link.c | 1619 struct tipc_node *node; in tipc_nl_link_set() local 1643 node = tipc_link_find_owner(net, name, &bearer_id); in tipc_nl_link_set() 1644 if (!node) in tipc_nl_link_set() 1647 tipc_node_lock(node); in tipc_nl_link_set() 1649 link = node->links[bearer_id].link; in tipc_nl_link_set() 1688 tipc_node_unlock(node); in tipc_nl_link_set() 1689 tipc_bearer_xmit(net, bearer_id, &xmitq, &node->links[bearer_id].maddr); in tipc_nl_link_set() 1831 struct tipc_node *node, u32 *prev_link) in __tipc_nl_add_node_links() argument 1839 if (!node->links[i].link) in __tipc_nl_add_node_links() 1843 node->links[i].link, NLM_F_MULTI); in __tipc_nl_add_node_links() [all …]
|
/net/netfilter/ |
D | xt_qtaguid_print.c | 294 struct rb_node *node; in prdebug_sock_tag_tree() local 310 for (node = rb_first(sock_tag_tree); in prdebug_sock_tag_tree() 311 node; in prdebug_sock_tag_tree() 312 node = rb_next(node)) { in prdebug_sock_tag_tree() 313 sock_tag_entry = rb_entry(node, struct sock_tag, sock_node); in prdebug_sock_tag_tree() 355 struct rb_node *node; in prdebug_proc_qtu_data_tree() local 370 for (node = rb_first(proc_qtu_data_tree); in prdebug_proc_qtu_data_tree() 371 node; in prdebug_proc_qtu_data_tree() 372 node = rb_next(node)) { in prdebug_proc_qtu_data_tree() 373 proc_qtu_data_entry = rb_entry(node, in prdebug_proc_qtu_data_tree() [all …]
|
D | nft_rbtree.c | 28 struct rb_node node; member 45 rbe = rb_entry(parent, struct nft_rbtree_elem, node); in nft_rbtree_lookup() 92 rbe = rb_entry(parent, struct nft_rbtree_elem, node); in __nft_rbtree_insert() 106 rb_link_node(&new->node, parent, p); in __nft_rbtree_insert() 107 rb_insert_color(&new->node, &priv->root); in __nft_rbtree_insert() 131 rb_erase(&rbe->node, &priv->root); in nft_rbtree_remove() 153 rbe = rb_entry(parent, struct nft_rbtree_elem, node); in nft_rbtree_deactivate() 180 struct rb_node *node; in nft_rbtree_walk() local 184 for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) { in nft_rbtree_walk() 185 rbe = rb_entry(node, struct nft_rbtree_elem, node); in nft_rbtree_walk() [all …]
|
D | xt_qtaguid.c | 179 struct rb_node *node = root->rb_node; in tag_node_tree_search() local 181 while (node) { in tag_node_tree_search() 182 struct tag_node *data = rb_entry(node, struct tag_node, node); in tag_node_tree_search() 185 " node=%p data=%p\n", tag, node, data); in tag_node_tree_search() 191 node = node->rb_left; in tag_node_tree_search() 193 node = node->rb_right; in tag_node_tree_search() 207 node); in tag_node_tree_insert() 223 rb_link_node(&data->node, parent, new); in tag_node_tree_insert() 224 rb_insert_color(&data->node, root); in tag_node_tree_insert() 234 struct tag_node *node = tag_node_tree_search(root, tag); in tag_stat_tree_search() local [all …]
|
D | xt_connlimit.c | 47 struct hlist_node node; member 53 struct rb_node node; member 130 hlist_add_head(&conn->node, head); in add_hlist() 150 hlist_for_each_entry_safe(conn, n, head, node) { in check_hlist() 153 hlist_del(&conn->node); in check_hlist() 173 hlist_del(&conn->node); in check_hlist() 195 rb_erase(&rbconn->node, root); in tree_nodes_free() 221 rbconn = container_of(*rbnode, struct xt_connlimit_rb, node); in count_tree() 281 hlist_add_head(&conn->node, &rbconn->hhead); in count_tree() 283 rb_link_node(&rbconn->node, parent, rbnode); in count_tree() [all …]
|
/net/mac80211/ |
D | mesh_pathtbl.c | 76 #define for_each_mesh_entry(tbl, node, i) \ argument 78 hlist_for_each_entry_rcu(node, &tbl->hash_buckets[i], list) 340 struct mpath_node *node; in mpath_lookup() local 343 hlist_for_each_entry_rcu(node, bucket, list) { in mpath_lookup() 344 mpath = node->mpath; in mpath_lookup() 393 struct mpath_node *node; in mesh_path_lookup_by_idx() local 397 for_each_mesh_entry(tbl, node, i) { in mesh_path_lookup_by_idx() 398 if (sdata && node->mpath->sdata != sdata) in mesh_path_lookup_by_idx() 401 if (mpath_expired(node->mpath)) { in mesh_path_lookup_by_idx() 402 spin_lock_bh(&node->mpath->state_lock); in mesh_path_lookup_by_idx() [all …]
|
/net/batman-adv/ |
D | hash.h | 70 struct hlist_node *node, *node_tmp; in batadv_hash_delete() local 79 hlist_for_each_safe(node, node_tmp, head) { in batadv_hash_delete() 80 hlist_del_rcu(node); in batadv_hash_delete() 83 free_cb(node, arg); in batadv_hash_delete() 111 struct hlist_node *node; in batadv_hash_add() local 123 hlist_for_each(node, head) { in batadv_hash_add() 124 if (!compare(node, data)) in batadv_hash_add() 153 struct hlist_node *node; in batadv_hash_remove() local 161 hlist_for_each(node, head) { in batadv_hash_remove() 162 if (!compare(node, data)) in batadv_hash_remove() [all …]
|
D | multicast.c | 611 struct hlist_node *node = &orig->mcast_want_all_unsnoopables_node; in batadv_mcast_want_unsnoop_update() local 623 WARN_ON(!hlist_unhashed(node)); in batadv_mcast_want_unsnoop_update() 625 hlist_add_head_rcu(node, head); in batadv_mcast_want_unsnoop_update() 634 WARN_ON(hlist_unhashed(node)); in batadv_mcast_want_unsnoop_update() 636 hlist_del_init_rcu(node); in batadv_mcast_want_unsnoop_update() 656 struct hlist_node *node = &orig->mcast_want_all_ipv4_node; in batadv_mcast_want_ipv4_update() local 668 WARN_ON(!hlist_unhashed(node)); in batadv_mcast_want_ipv4_update() 670 hlist_add_head_rcu(node, head); in batadv_mcast_want_ipv4_update() 679 WARN_ON(hlist_unhashed(node)); in batadv_mcast_want_ipv4_update() 681 hlist_del_init_rcu(node); in batadv_mcast_want_ipv4_update() [all …]
|
/net/ipx/ |
D | ipx_proc.c | 47 i = list_entry(v, struct ipx_interface, node); in ipx_seq_interface_show() 88 rt = list_entry(v, struct ipx_route, node); in ipx_seq_route_show() 108 list_for_each_entry(i, &ipx_interfaces, node) { in ipx_get_socket_idx() 163 if (i->node.next == &ipx_interfaces) in ipx_seq_socket_next() 165 i = list_entry(i->node.next, struct ipx_interface, node); in ipx_seq_socket_next() 199 ipxs->node[0], ipxs->node[1], ipxs->node[2], ipxs->node[3], in ipx_seq_socket_show() 200 ipxs->node[4], ipxs->node[5], ntohs(ipxs->port)); in ipx_seq_socket_show() 210 ipxs->dest_addr.node[0], ipxs->dest_addr.node[1], in ipx_seq_socket_show() 211 ipxs->dest_addr.node[2], ipxs->dest_addr.node[3], in ipx_seq_socket_show() 212 ipxs->dest_addr.node[4], ipxs->dest_addr.node[5], in ipx_seq_socket_show()
|
D | ipx_route.c | 34 list_for_each_entry(r, &ipx_routes, node) in ipxrtr_lookup() 49 unsigned char *node) in ipxrtr_add_route() argument 65 list_add(&rt->node, &ipx_routes); in ipxrtr_add_route() 75 if (!node) { in ipxrtr_add_route() 79 memcpy(rt->ir_router_node, node, IPX_NODE_LEN); in ipxrtr_add_route() 95 list_for_each_entry_safe(r, tmp, &ipx_routes, node) in ipxrtr_del_routes() 97 list_del(&r->node); in ipxrtr_del_routes() 124 list_for_each_entry_safe(r, tmp, &ipx_routes, node) in ipxrtr_delete() 130 list_del(&r->node); in ipxrtr_delete() 157 r->ir_router_node : ipx->ipx_dest.node); in ipxrtr_route_skb() [all …]
|
/net/ipv4/ |
D | inetpeer.c | 246 struct inet_peer *node, *l, *r; in peer_avl_rebalance() local 251 node = rcu_deref_locked(*nodep, base); in peer_avl_rebalance() 252 l = rcu_deref_locked(node->avl_left, base); in peer_avl_rebalance() 253 r = rcu_deref_locked(node->avl_right, base); in peer_avl_rebalance() 263 RCU_INIT_POINTER(node->avl_left, lr); /* lr: RH or RH+1 */ in peer_avl_rebalance() 264 RCU_INIT_POINTER(node->avl_right, r); /* r: RH */ in peer_avl_rebalance() 265 node->avl_height = lrh + 1; /* RH+1 or RH+2 */ in peer_avl_rebalance() 267 RCU_INIT_POINTER(l->avl_right, node); /* node: RH+1 or RH+2 */ in peer_avl_rebalance() 268 l->avl_height = node->avl_height + 1; in peer_avl_rebalance() 273 RCU_INIT_POINTER(node->avl_left, lrr); /* lrr: RH or RH-1 */ in peer_avl_rebalance() [all …]
|
/net/6lowpan/ |
D | nhc.c | 31 node); in lowpan_nhc_insert() 55 rb_link_node(&nhc->node, parent, new); in lowpan_nhc_insert() 56 rb_insert_color(&nhc->node, &rb_root); in lowpan_nhc_insert() 63 rb_erase(&nhc->node, &rb_root); in lowpan_nhc_remove() 68 struct rb_node *node = rb_root.rb_node; in lowpan_nhc_by_nhcid() local 71 while (node) { in lowpan_nhc_by_nhcid() 72 struct lowpan_nhc *nhc = container_of(node, struct lowpan_nhc, in lowpan_nhc_by_nhcid() 73 node); in lowpan_nhc_by_nhcid() 87 node = node->rb_left; in lowpan_nhc_by_nhcid() 89 node = node->rb_right; in lowpan_nhc_by_nhcid()
|
/net/decnet/ |
D | sysctl_net_decnet.c | 91 __u16 area, node; in parse_addr() local 110 node = *str++ - '0'; in parse_addr() 112 node *= 10; in parse_addr() 113 node += (*str++ - '0'); in parse_addr() 116 node *= 10; in parse_addr() 117 node += (*str++ - '0'); in parse_addr() 120 node *= 10; in parse_addr() 121 node += (*str++ - '0'); in parse_addr() 124 if ((node > 1023) || (area > 63)) in parse_addr() 130 *addr = cpu_to_le16((area << 10) | node); in parse_addr()
|
/net/802/ |
D | garp.c | 158 attr = rb_entry(parent, struct garp_attr, node); in garp_attr_lookup() 179 attr = rb_entry(parent, struct garp_attr, node); in garp_attr_create() 198 rb_link_node(&attr->node, parent, p); in garp_attr_create() 199 rb_insert_color(&attr->node, &app->gid); in garp_attr_create() 205 rb_erase(&attr->node, &app->gid); in garp_attr_destroy() 211 struct rb_node *node, *next; in garp_attr_destroy_all() local 214 for (node = rb_first(&app->gid); in garp_attr_destroy_all() 215 next = node ? rb_next(node) : NULL, node != NULL; in garp_attr_destroy_all() 216 node = next) { in garp_attr_destroy_all() 217 attr = rb_entry(node, struct garp_attr, node); in garp_attr_destroy_all() [all …]
|
D | mrp.c | 247 attr = rb_entry(parent, struct mrp_attr, node); in mrp_attr_lookup() 268 attr = rb_entry(parent, struct mrp_attr, node); in mrp_attr_create() 287 rb_link_node(&attr->node, parent, p); in mrp_attr_create() 288 rb_insert_color(&attr->node, &app->mad); in mrp_attr_create() 294 rb_erase(&attr->node, &app->mad); in mrp_attr_destroy() 300 struct rb_node *node, *next; in mrp_attr_destroy_all() local 303 for (node = rb_first(&app->mad); in mrp_attr_destroy_all() 304 next = node ? rb_next(node) : NULL, node != NULL; in mrp_attr_destroy_all() 305 node = next) { in mrp_attr_destroy_all() 306 attr = rb_entry(node, struct mrp_attr, node); in mrp_attr_destroy_all() [all …]
|
/net/x25/ |
D | x25_route.c | 41 rt = list_entry(entry, struct x25_route, node); in x25_add_route() 60 list_add(&rt->node, &x25_route_list); in x25_add_route() 76 if (rt->node.next) { in __x25_remove_route() 77 list_del(&rt->node); in __x25_remove_route() 92 rt = list_entry(entry, struct x25_route, node); in x25_del_route() 117 rt = list_entry(entry, struct x25_route, node); in x25_route_device_down() 162 rt = list_entry(entry, struct x25_route, node); in x25_get_route() 222 rt = list_entry(entry, struct x25_route, node); in x25_route_free()
|
D | x25_forward.c | 55 x25_frwd = list_entry(entry, struct x25_forward, node); in x25_forward_call() 74 list_add(&new_frwd->node, &x25_forward_list); in x25_forward_call() 108 frwd = list_entry(entry, struct x25_forward, node); in x25_forward_data() 145 fwd = list_entry(entry, struct x25_forward, node); in x25_clear_forward_by_lci() 147 list_del(&fwd->node); in x25_clear_forward_by_lci() 163 fwd = list_entry(entry, struct x25_forward, node); in x25_clear_forward_by_dev() 165 list_del(&fwd->node); in x25_clear_forward_by_dev()
|
/net/openvswitch/ |
D | flow_table.c | 82 int node; in ovs_flow_alloc() local 105 for_each_node(node) in ovs_flow_alloc() 106 if (node != 0) in ovs_flow_alloc() 107 RCU_INIT_POINTER(flow->stats[node], NULL); in ovs_flow_alloc() 145 int node; in flow_free() local 151 for_each_node(node) in flow_free() 152 if (flow->stats[node]) in flow_free() 154 (struct flow_stats __force *)flow->stats[node]); in flow_free() 262 hlist_for_each_entry_safe(flow, n, head, flow_table.node[ver]) { in table_instance_destroy() 263 hlist_del_rcu(&flow->flow_table.node[ver]); in table_instance_destroy() [all …]
|
/net/irda/ |
D | discovery.c | 62 discovery_t *discovery, *node; in irlmp_add_discovery() local 78 node = discovery; in irlmp_add_discovery() 83 if ((node->data.saddr == new->data.saddr) && in irlmp_add_discovery() 84 ((node->data.daddr == new->data.daddr) || in irlmp_add_discovery() 85 (strcmp(node->data.info, new->data.info) == 0))) in irlmp_add_discovery() 90 hashbin_remove_this(cachelog, (irda_queue_t *) node); in irlmp_add_discovery() 92 if (get_unaligned((__u16 *)node->data.hints) == get_unaligned((__u16 *)new->data.hints)) in irlmp_add_discovery() 94 new->firststamp = node->firststamp; in irlmp_add_discovery() 95 kfree(node); in irlmp_add_discovery()
|
/net/ipv6/ |
D | inet6_hashtables.c | 60 const struct hlist_nulls_node *node; in __inet6_lookup_established() local 72 sk_nulls_for_each_rcu(sk, node, &head->chain) { in __inet6_lookup_established() 86 if (get_nulls_value(node) != slot) in __inet6_lookup_established() 129 const struct hlist_nulls_node *node; in inet6_lookup_listener() local 140 sk_nulls_for_each(sk, node, &ilb->head) { in inet6_lookup_listener() 163 if (get_nulls_value(node) != hash + LISTENING_NULLS_BASE) in inet6_lookup_listener() 210 const struct hlist_nulls_node *node; in __inet6_check_established() local 215 sk_nulls_for_each(sk2, node, &head->chain) { in __inet6_check_established()
|
/net/sunrpc/ |
D | svc.c | 112 unsigned int node; in svc_pool_map_choose_mode() local 122 node = first_online_node; in svc_pool_map_choose_mode() 123 if (nr_cpus_node(node) > 2) { in svc_pool_map_choose_mode() 197 unsigned int node; in svc_pool_map_init_pernode() local 204 for_each_node_with_cpus(node) { in svc_pool_map_init_pernode() 207 m->to_pool[node] = pidx; in svc_pool_map_init_pernode() 208 m->pool_to[pidx] = node; in svc_pool_map_init_pernode() 305 unsigned int node = m->pool_to[pidx]; in svc_pool_map_set_cpumask() local 318 set_cpus_allowed_ptr(task, cpumask_of(node)); in svc_pool_map_set_cpumask() 323 set_cpus_allowed_ptr(task, cpumask_of_node(node)); in svc_pool_map_set_cpumask() [all …]
|