/net/sctp/ |
D | transport.c | 261 bool sctp_transport_pl_send(struct sctp_transport *t) in sctp_transport_pl_send() argument 263 if (t->pl.probe_count < SCTP_MAX_PROBES) in sctp_transport_pl_send() 266 t->pl.last_rtx_chunks = t->asoc->rtx_data_chunks; in sctp_transport_pl_send() 267 t->pl.probe_count = 0; in sctp_transport_pl_send() 268 if (t->pl.state == SCTP_PL_BASE) { in sctp_transport_pl_send() 269 if (t->pl.probe_size == SCTP_BASE_PLPMTU) { /* BASE_PLPMTU Confirmation Failed */ in sctp_transport_pl_send() 270 t->pl.state = SCTP_PL_ERROR; /* Base -> Error */ in sctp_transport_pl_send() 272 t->pl.pmtu = SCTP_BASE_PLPMTU; in sctp_transport_pl_send() 273 t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t); in sctp_transport_pl_send() 274 sctp_assoc_sync_pmtu(t->asoc); in sctp_transport_pl_send() [all …]
|
D | sm_sideeffect.c | 231 void sctp_generate_t3_rtx_event(struct timer_list *t) in sctp_generate_t3_rtx_event() argument 234 from_timer(transport, t, T3_rtx_timer); in sctp_generate_t3_rtx_event() 308 static void sctp_generate_t1_cookie_event(struct timer_list *t) in sctp_generate_t1_cookie_event() argument 311 from_timer(asoc, t, timers[SCTP_EVENT_TIMEOUT_T1_COOKIE]); in sctp_generate_t1_cookie_event() 316 static void sctp_generate_t1_init_event(struct timer_list *t) in sctp_generate_t1_init_event() argument 319 from_timer(asoc, t, timers[SCTP_EVENT_TIMEOUT_T1_INIT]); in sctp_generate_t1_init_event() 324 static void sctp_generate_t2_shutdown_event(struct timer_list *t) in sctp_generate_t2_shutdown_event() argument 327 from_timer(asoc, t, timers[SCTP_EVENT_TIMEOUT_T2_SHUTDOWN]); in sctp_generate_t2_shutdown_event() 332 static void sctp_generate_t4_rto_event(struct timer_list *t) in sctp_generate_t4_rto_event() argument 335 from_timer(asoc, t, timers[SCTP_EVENT_TIMEOUT_T4_RTO]); in sctp_generate_t4_rto_event() [all …]
|
D | input.c | 295 struct sctp_transport *t = chunk->transport; in sctp_backlog_rcv() local 353 sctp_transport_put(t); in sctp_backlog_rcv() 365 struct sctp_transport *t = chunk->transport; in sctp_add_backlog() local 376 sctp_transport_hold(t); in sctp_add_backlog() 388 struct sctp_transport *t, __u32 pmtu) in sctp_icmp_frag_needed() argument 390 if (!t || in sctp_icmp_frag_needed() 391 (t->pathmtu <= pmtu && in sctp_icmp_frag_needed() 392 t->pl.probe_size + sctp_transport_pl_hlen(t) <= pmtu)) in sctp_icmp_frag_needed() 396 atomic_set(&t->mtu_info, pmtu); in sctp_icmp_frag_needed() 398 t->pmtu_pending = 1; in sctp_icmp_frag_needed() [all …]
|
/net/ipv6/ |
D | ip6_gre.c | 78 static void ip6gre_tunnel_link(struct ip6gre_net *ign, struct ip6_tnl *t); 79 static void ip6gre_tnl_link_config(struct ip6_tnl *t, int set_mtu); 80 static void ip6erspan_tnl_link_config(struct ip6_tnl *t, int set_mtu); 123 struct ip6_tnl *t, *cand = NULL; in ip6gre_tunnel_lookup() local 132 for_each_ip_tunnel_rcu(t, ign->tunnels_r_l[h0 ^ h1]) { in ip6gre_tunnel_lookup() 133 if (!ipv6_addr_equal(local, &t->parms.laddr) || in ip6gre_tunnel_lookup() 134 !ipv6_addr_equal(remote, &t->parms.raddr) || in ip6gre_tunnel_lookup() 135 key != t->parms.i_key || in ip6gre_tunnel_lookup() 136 !(t->dev->flags & IFF_UP)) in ip6gre_tunnel_lookup() 139 if (t->dev->type != ARPHRD_IP6GRE && in ip6gre_tunnel_lookup() [all …]
|
D | ip6_tunnel.c | 98 for (t = rcu_dereference(start); t; t = rcu_dereference(t->next)) 118 struct ip6_tnl *t, *cand = NULL; in ip6_tnl_lookup() local 123 if (!ipv6_addr_equal(local, &t->parms.laddr) || in ip6_tnl_lookup() 124 !ipv6_addr_equal(remote, &t->parms.raddr) || in ip6_tnl_lookup() 125 !(t->dev->flags & IFF_UP)) in ip6_tnl_lookup() 128 if (link == t->parms.link) in ip6_tnl_lookup() 129 return t; in ip6_tnl_lookup() 131 cand = t; in ip6_tnl_lookup() 137 if (!ipv6_addr_equal(local, &t->parms.laddr) || in ip6_tnl_lookup() 138 !ipv6_addr_any(&t->parms.raddr) || in ip6_tnl_lookup() [all …]
|
D | ip6_vti.c | 75 for (t = rcu_dereference(start); t; t = rcu_dereference(t->next)) 93 struct ip6_tnl *t; in vti6_tnl_lookup() local 98 if (ipv6_addr_equal(local, &t->parms.laddr) && in vti6_tnl_lookup() 99 ipv6_addr_equal(remote, &t->parms.raddr) && in vti6_tnl_lookup() 100 (t->dev->flags & IFF_UP)) in vti6_tnl_lookup() 101 return t; in vti6_tnl_lookup() 107 if (ipv6_addr_equal(local, &t->parms.laddr) && in vti6_tnl_lookup() 108 (t->dev->flags & IFF_UP)) in vti6_tnl_lookup() 109 return t; in vti6_tnl_lookup() 114 if (ipv6_addr_equal(remote, &t->parms.raddr) && in vti6_tnl_lookup() [all …]
|
D | sit.c | 88 struct ip_tunnel *t = netdev_priv(dev); in dev_to_sit_net() local 90 return net_generic(t->net, sit_net_id); in dev_to_sit_net() 103 struct ip_tunnel *t; in ipip6_tunnel_lookup() local 107 for_each_ip_tunnel_rcu(t, sitn->tunnels_r_l[h0 ^ h1]) { in ipip6_tunnel_lookup() 108 if (local == t->parms.iph.saddr && in ipip6_tunnel_lookup() 109 remote == t->parms.iph.daddr && in ipip6_tunnel_lookup() 110 (!dev || !t->parms.link || ifindex == t->parms.link || in ipip6_tunnel_lookup() 111 sifindex == t->parms.link) && in ipip6_tunnel_lookup() 112 (t->dev->flags & IFF_UP)) in ipip6_tunnel_lookup() 113 return t; in ipip6_tunnel_lookup() [all …]
|
D | ipcomp6.c | 77 struct xfrm_state *t = NULL; in ipcomp6_tunnel_create() local 79 t = xfrm_state_alloc(net); in ipcomp6_tunnel_create() 80 if (!t) in ipcomp6_tunnel_create() 83 t->id.proto = IPPROTO_IPV6; in ipcomp6_tunnel_create() 84 t->id.spi = xfrm6_tunnel_alloc_spi(net, (xfrm_address_t *)&x->props.saddr); in ipcomp6_tunnel_create() 85 if (!t->id.spi) in ipcomp6_tunnel_create() 88 memcpy(t->id.daddr.a6, x->id.daddr.a6, sizeof(struct in6_addr)); in ipcomp6_tunnel_create() 89 memcpy(&t->sel, &x->sel, sizeof(t->sel)); in ipcomp6_tunnel_create() 90 t->props.family = AF_INET6; in ipcomp6_tunnel_create() 91 t->props.mode = x->props.mode; in ipcomp6_tunnel_create() [all …]
|
/net/ipv4/ |
D | ipcomp.c | 60 struct xfrm_state *t; in ipcomp_tunnel_create() local 62 t = xfrm_state_alloc(net); in ipcomp_tunnel_create() 63 if (!t) in ipcomp_tunnel_create() 66 t->id.proto = IPPROTO_IPIP; in ipcomp_tunnel_create() 67 t->id.spi = x->props.saddr.a4; in ipcomp_tunnel_create() 68 t->id.daddr.a4 = x->id.daddr.a4; in ipcomp_tunnel_create() 69 memcpy(&t->sel, &x->sel, sizeof(t->sel)); in ipcomp_tunnel_create() 70 t->props.family = AF_INET; in ipcomp_tunnel_create() 71 t->props.mode = x->props.mode; in ipcomp_tunnel_create() 72 t->props.saddr.a4 = x->props.saddr.a4; in ipcomp_tunnel_create() [all …]
|
D | ip_tunnel.c | 88 struct ip_tunnel *t, *cand = NULL; in ip_tunnel_lookup() local 96 hlist_for_each_entry_rcu(t, head, hash_node) { in ip_tunnel_lookup() 97 if (local != t->parms.iph.saddr || in ip_tunnel_lookup() 98 remote != t->parms.iph.daddr || in ip_tunnel_lookup() 99 !(t->dev->flags & IFF_UP)) in ip_tunnel_lookup() 102 if (!ip_tunnel_key_match(&t->parms, flags, key)) in ip_tunnel_lookup() 105 if (t->parms.link == link) in ip_tunnel_lookup() 106 return t; in ip_tunnel_lookup() 108 cand = t; in ip_tunnel_lookup() 111 hlist_for_each_entry_rcu(t, head, hash_node) { in ip_tunnel_lookup() [all …]
|
D | ip_gre.c | 144 struct ip_tunnel *t; in ipgre_err() local 155 t = ip_tunnel_lookup(itn, skb->dev->ifindex, tpi->flags, in ipgre_err() 158 if (!t) in ipgre_err() 198 if (t->parms.iph.daddr == 0 || in ipgre_err() 199 ipv4_is_multicast(t->parms.iph.daddr)) in ipgre_err() 202 if (t->parms.iph.ttl == 0 && type == ICMP_TIME_EXCEEDED) in ipgre_err() 205 if (time_before(jiffies, t->err_time + IPTUNNEL_ERR_TIMEO)) in ipgre_err() 206 t->err_count++; in ipgre_err() 208 t->err_count = 1; in ipgre_err() 209 t->err_time = jiffies; in ipgre_err() [all …]
|
/net/netfilter/ |
D | xt_recent.c | 144 static void recent_entry_remove(struct recent_table *t, struct recent_entry *e) in recent_entry_remove() argument 149 t->entries--; in recent_entry_remove() 155 static void recent_entry_reap(struct recent_table *t, unsigned long time, in recent_entry_reap() argument 163 e = list_entry(t->lru_list.next, struct recent_entry, lru_list); in recent_entry_reap() 175 recent_entry_remove(t, e); in recent_entry_reap() 179 recent_entry_init(struct recent_table *t, const union nf_inet_addr *addr, in recent_entry_init() argument 183 unsigned int nstamps_max = t->nstamps_max_mask; in recent_entry_init() 185 if (t->entries >= ip_list_tot) { in recent_entry_init() 186 e = list_entry(t->lru_list.next, struct recent_entry, lru_list); in recent_entry_init() 187 recent_entry_remove(t, e); in recent_entry_init() [all …]
|
D | xt_HMARK.c | 77 hmark_ct_set_htuple(const struct sk_buff *skb, struct hmark_tuple *t, in hmark_ct_set_htuple() argument 92 t->src = hmark_addr_mask(otuple->src.l3num, otuple->src.u3.ip6, in hmark_ct_set_htuple() 94 t->dst = hmark_addr_mask(otuple->src.l3num, rtuple->src.u3.ip6, in hmark_ct_set_htuple() 100 t->proto = nf_ct_protonum(ct); in hmark_ct_set_htuple() 101 if (t->proto != IPPROTO_ICMP) { in hmark_ct_set_htuple() 102 t->uports.b16.src = otuple->src.u.all; in hmark_ct_set_htuple() 103 t->uports.b16.dst = rtuple->src.u.all; in hmark_ct_set_htuple() 104 hmark_swap_ports(&t->uports, info); in hmark_ct_set_htuple() 116 hmark_hash(struct hmark_tuple *t, const struct xt_hmark_info *info) in hmark_hash() argument 119 u32 src = ntohl(t->src); in hmark_hash() [all …]
|
D | x_tables.c | 248 struct xt_target *t; in xt_find_target() local 255 list_for_each_entry(t, &xt[af].target, list) { in xt_find_target() 256 if (strcmp(t->name, name) == 0) { in xt_find_target() 257 if (t->revision == revision) { in xt_find_target() 258 if (try_module_get(t->me)) { in xt_find_target() 260 return t; in xt_find_target() 340 int xt_target_to_user(const struct xt_entry_target *t, in xt_target_to_user() argument 343 return XT_OBJ_TO_USER(u, t, target, 0) || in xt_target_to_user() 344 XT_DATA_TO_USER(u, t, target); in xt_target_to_user() 372 const struct xt_target *t; in target_revfn() local [all …]
|
/net/netfilter/ipset/ |
D | ip_set_hash_gen.h | 363 mtype_ahash_memsize(const struct htype *h, const struct htable *t) in mtype_ahash_memsize() argument 365 return sizeof(*h) + sizeof(*t) + ahash_sizeof_regions(t->htable_bits); in mtype_ahash_memsize() 387 struct htable *t; in mtype_flush() local 391 t = ipset_dereference_nfnl(h->table); in mtype_flush() 392 for (r = 0; r < ahash_numof_locks(t->htable_bits); r++) { in mtype_flush() 393 spin_lock_bh(&t->hregion[r].lock); in mtype_flush() 394 for (i = ahash_bucket_start(r, t->htable_bits); in mtype_flush() 395 i < ahash_bucket_end(r, t->htable_bits); i++) { in mtype_flush() 396 n = __ipset_dereference(hbucket(t, i)); in mtype_flush() 402 rcu_assign_pointer(hbucket(t, i), NULL); in mtype_flush() [all …]
|
/net/ipv4/netfilter/ |
D | arp_tables.c | 221 const struct xt_entry_target *t; in arpt_do_table() local 232 t = arpt_get_target_c(e); in arpt_do_table() 235 if (!t->u.kernel.target->target) { in arpt_do_table() 238 v = ((struct xt_standard_target *)t)->verdict; in arpt_do_table() 267 acpar.target = t->u.kernel.target; in arpt_do_table() 268 acpar.targinfo = t->data; in arpt_do_table() 269 verdict = t->u.kernel.target->target(skb, &acpar); in arpt_do_table() 321 const struct xt_standard_target *t in mark_source_chains() local 333 (strcmp(t->target.u.user.name, in mark_source_chains() 335 t->verdict < 0) || visited) { in mark_source_chains() [all …]
|
D | ip_tables.c | 162 const struct xt_standard_target *t = (void *)ipt_get_target_c(s); in get_chainname_rulenum() local 164 if (strcmp(t->target.u.kernel.target->name, XT_ERROR_TARGET) == 0) { in get_chainname_rulenum() 166 *chainname = t->target.data; in get_chainname_rulenum() 172 strcmp(t->target.u.kernel.target->name, in get_chainname_rulenum() 174 t->verdict < 0) { in get_chainname_rulenum() 279 const struct xt_entry_target *t; in ipt_do_table() local 301 t = ipt_get_target_c(e); in ipt_do_table() 302 WARN_ON(!t->u.kernel.target); in ipt_do_table() 311 if (!t->u.kernel.target->target) { in ipt_do_table() 314 v = ((struct xt_standard_target *)t)->verdict; in ipt_do_table() [all …]
|
/net/bridge/ |
D | br_stp.c | 101 int t; in br_should_become_root_port() local 116 t = memcmp(&p->designated_root, &rp->designated_root, 8); in br_should_become_root_port() 117 if (t < 0) in br_should_become_root_port() 119 else if (t > 0) in br_should_become_root_port() 129 t = memcmp(&p->designated_bridge, &rp->designated_bridge, 8); in br_should_become_root_port() 130 if (t < 0) in br_should_become_root_port() 132 else if (t > 0) in br_should_become_root_port() 285 int t; in br_should_become_designated_port() local 299 t = memcmp(&br->bridge_id, &p->designated_bridge, 8); in br_should_become_designated_port() 300 if (t < 0) in br_should_become_designated_port() [all …]
|
/net/ipv6/netfilter/ |
D | ip6_tables.c | 187 const struct xt_standard_target *t = (void *)ip6t_get_target_c(s); in get_chainname_rulenum() local 189 if (strcmp(t->target.u.kernel.target->name, XT_ERROR_TARGET) == 0) { in get_chainname_rulenum() 191 *chainname = t->target.data; in get_chainname_rulenum() 197 strcmp(t->target.u.kernel.target->name, in get_chainname_rulenum() 199 t->verdict < 0) { in get_chainname_rulenum() 302 const struct xt_entry_target *t; in ip6t_do_table() local 325 t = ip6t_get_target_c(e); in ip6t_do_table() 326 WARN_ON(!t->u.kernel.target); in ip6t_do_table() 335 if (!t->u.kernel.target->target) { in ip6t_do_table() 338 v = ((struct xt_standard_target *)t)->verdict; in ip6t_do_table() [all …]
|
/net/bridge/netfilter/ |
D | ebtables.c | 200 const struct ebt_entry_target *t; in ebt_do_table() local 244 t = ebt_get_target_c(point); in ebt_do_table() 246 if (!t->u.target->target) in ebt_do_table() 247 verdict = ((struct ebt_standard_target *)t)->verdict; in ebt_do_table() 249 acpar.target = t->u.target; in ebt_do_table() 250 acpar.targinfo = t->data; in ebt_do_table() 251 verdict = t->u.target->target(skb, &acpar); in ebt_do_table() 659 struct ebt_entry_target *t; in ebt_cleanup_entry() local 668 t = ebt_get_target(e); in ebt_cleanup_entry() 671 par.target = t->u.target; in ebt_cleanup_entry() [all …]
|
/net/6lowpan/ |
D | debugfs.c | 69 struct lowpan_iphc_ctx_table *t = in lowpan_ctx_plen_set() local 75 spin_lock_bh(&t->lock); in lowpan_ctx_plen_set() 77 spin_unlock_bh(&t->lock); in lowpan_ctx_plen_set() 85 struct lowpan_iphc_ctx_table *t = in lowpan_ctx_plen_get() local 88 spin_lock_bh(&t->lock); in lowpan_ctx_plen_get() 90 spin_unlock_bh(&t->lock); in lowpan_ctx_plen_get() 100 struct lowpan_iphc_ctx_table *t = in lowpan_ctx_pfx_show() local 103 spin_lock_bh(&t->lock); in lowpan_ctx_pfx_show() 113 spin_unlock_bh(&t->lock); in lowpan_ctx_pfx_show() 130 struct lowpan_iphc_ctx_table *t = in lowpan_ctx_pfx_write() local [all …]
|
/net/9p/ |
D | mod.c | 94 struct p9_trans_module *t, *found = NULL; in v9fs_get_trans_by_name() local 98 list_for_each_entry(t, &v9fs_trans_list, list) in v9fs_get_trans_by_name() 99 if (strcmp(t->name, s) == 0 && in v9fs_get_trans_by_name() 100 try_module_get(t->owner)) { in v9fs_get_trans_by_name() 101 found = t; in v9fs_get_trans_by_name() 117 struct p9_trans_module *t, *found = NULL; in v9fs_get_default_trans() local 121 list_for_each_entry(t, &v9fs_trans_list, list) in v9fs_get_default_trans() 122 if (t->def && try_module_get(t->owner)) { in v9fs_get_default_trans() 123 found = t; in v9fs_get_default_trans() 128 list_for_each_entry(t, &v9fs_trans_list, list) in v9fs_get_default_trans() [all …]
|
/net/core/ |
D | pktgen.c | 235 #define if_lock(t) mutex_lock(&(t->if_lock)); argument 236 #define if_unlock(t) mutex_unlock(&(t->if_lock)); argument 478 static int pktgen_remove_device(struct pktgen_thread *t, struct pktgen_dev *i); 479 static int pktgen_add_device(struct pktgen_thread *t, const char *ifname); 480 static struct pktgen_dev *pktgen_find_dev(struct pktgen_thread *t, 487 static void pktgen_stop(struct pktgen_thread *t); 1826 struct pktgen_thread *t = seq->private; in pktgen_thread_show() local 1829 BUG_ON(!t); in pktgen_thread_show() 1834 list_for_each_entry_rcu(pkt_dev, &t->if_list, list) in pktgen_thread_show() 1840 list_for_each_entry_rcu(pkt_dev, &t->if_list, list) in pktgen_thread_show() [all …]
|
/net/rxrpc/ |
D | call_event.c | 292 unsigned long now, next, t; in rxrpc_process_call() local 319 t = READ_ONCE(call->expect_rx_by); in rxrpc_process_call() 320 if (time_after_eq(now, t)) { in rxrpc_process_call() 325 t = READ_ONCE(call->expect_req_by); in rxrpc_process_call() 327 time_after_eq(now, t)) { in rxrpc_process_call() 332 t = READ_ONCE(call->expect_term_by); in rxrpc_process_call() 333 if (time_after_eq(now, t)) { in rxrpc_process_call() 338 t = READ_ONCE(call->ack_at); in rxrpc_process_call() 339 if (time_after_eq(now, t)) { in rxrpc_process_call() 341 cmpxchg(&call->ack_at, t, now + MAX_JIFFY_OFFSET); in rxrpc_process_call() [all …]
|
/net/sched/ |
D | act_ipt.c | 33 static int ipt_init_target(struct net *net, struct xt_entry_target *t, in ipt_init_target() argument 41 target = xt_request_find_target(AF_INET, t->u.user.name, in ipt_init_target() 42 t->u.user.revision); in ipt_init_target() 46 t->u.kernel.target = target; in ipt_init_target() 52 par.targinfo = t->data; in ipt_init_target() 56 ret = xt_check_target(&par, t->u.target_size - sizeof(*t), 0, false); in ipt_init_target() 58 module_put(t->u.kernel.target->me); in ipt_init_target() 64 static void ipt_destroy_target(struct xt_entry_target *t, struct net *net) in ipt_destroy_target() argument 67 .target = t->u.kernel.target, in ipt_destroy_target() 68 .targinfo = t->data, in ipt_destroy_target() [all …]
|