/net/rxrpc/ |
D | peer_object.c | 76 static long rxrpc_peer_cmp_key(const struct rxrpc_peer *peer, in rxrpc_peer_cmp_key() argument 83 diff = ((peer->hash_key - hash_key) ?: in rxrpc_peer_cmp_key() 84 ((unsigned long)peer->local - (unsigned long)local) ?: in rxrpc_peer_cmp_key() 85 (peer->srx.transport_type - srx->transport_type) ?: in rxrpc_peer_cmp_key() 86 (peer->srx.transport_len - srx->transport_len) ?: in rxrpc_peer_cmp_key() 87 (peer->srx.transport.family - srx->transport.family)); in rxrpc_peer_cmp_key() 93 return ((u16 __force)peer->srx.transport.sin.sin_port - in rxrpc_peer_cmp_key() 95 memcmp(&peer->srx.transport.sin.sin_addr, in rxrpc_peer_cmp_key() 100 return ((u16 __force)peer->srx.transport.sin6.sin6_port - in rxrpc_peer_cmp_key() 102 memcmp(&peer->srx.transport.sin6.sin6_addr, in rxrpc_peer_cmp_key() [all …]
|
D | rtt.c | 18 static u32 rxrpc_rto_min_us(struct rxrpc_peer *peer) in rxrpc_rto_min_us() argument 23 static u32 __rxrpc_set_rto(const struct rxrpc_peer *peer) in __rxrpc_set_rto() argument 25 return usecs_to_jiffies((peer->srtt_us >> 3) + peer->rttvar_us); in __rxrpc_set_rto() 43 static void rxrpc_rtt_estimator(struct rxrpc_peer *peer, long sample_rtt_us) in rxrpc_rtt_estimator() argument 46 u32 srtt = peer->srtt_us; in rxrpc_rtt_estimator() 69 m -= (peer->mdev_us >> 2); /* similar update on mdev */ in rxrpc_rtt_estimator() 81 m -= (peer->mdev_us >> 2); /* similar update on mdev */ in rxrpc_rtt_estimator() 84 peer->mdev_us += m; /* mdev = 3/4 mdev + 1/4 new */ in rxrpc_rtt_estimator() 85 if (peer->mdev_us > peer->mdev_max_us) { in rxrpc_rtt_estimator() 86 peer->mdev_max_us = peer->mdev_us; in rxrpc_rtt_estimator() [all …]
|
D | peer_event.c | 161 struct rxrpc_peer *peer; in rxrpc_encap_err_rcv() local 179 peer = rxrpc_lookup_peer_icmp_rcu(local, skb, udp_offset, in rxrpc_encap_err_rcv() 184 peer = rxrpc_lookup_peer_icmp6_rcu(local, skb, udp_offset, in rxrpc_encap_err_rcv() 193 if (peer && !rxrpc_get_peer_maybe(peer)) in rxrpc_encap_err_rcv() 194 peer = NULL; in rxrpc_encap_err_rcv() 195 if (!peer) { in rxrpc_encap_err_rcv() 208 rxrpc_adjust_mtu(peer, info); in rxrpc_encap_err_rcv() 210 rxrpc_put_peer(peer); in rxrpc_encap_err_rcv() 244 rxrpc_adjust_mtu(peer, info); in rxrpc_encap_err_rcv() 246 rxrpc_put_peer(peer); in rxrpc_encap_err_rcv() [all …]
|
D | conn_service.c | 27 struct rxrpc_connection *rxrpc_find_service_conn_rcu(struct rxrpc_peer *peer, in rxrpc_find_service_conn_rcu() argument 45 read_seqbegin_or_lock(&peer->service_conn_lock, &seq); in rxrpc_find_service_conn_rcu() 47 p = rcu_dereference_raw(peer->service_conns.rb_node); in rxrpc_find_service_conn_rcu() 59 } while (need_seqretry(&peer->service_conn_lock, seq)); in rxrpc_find_service_conn_rcu() 61 done_seqretry(&peer->service_conn_lock, seq); in rxrpc_find_service_conn_rcu() 70 static void rxrpc_publish_service_conn(struct rxrpc_peer *peer, in rxrpc_publish_service_conn() argument 77 write_seqlock_bh(&peer->service_conn_lock); in rxrpc_publish_service_conn() 79 pp = &peer->service_conns.rb_node; in rxrpc_publish_service_conn() 95 rb_insert_color(&conn->service_node, &peer->service_conns); in rxrpc_publish_service_conn() 98 write_sequnlock_bh(&peer->service_conn_lock); in rxrpc_publish_service_conn() [all …]
|
D | call_accept.c | 73 struct rxrpc_peer *peer = rxrpc_alloc_peer(rx->local, gfp); in rxrpc_service_prealloc_one() local 74 if (!peer) in rxrpc_service_prealloc_one() 76 b->peer_backlog[head] = peer; in rxrpc_service_prealloc_one() 199 struct rxrpc_peer *peer = b->peer_backlog[tail]; in rxrpc_discard_prealloc() local 200 rxrpc_put_local(peer->local); in rxrpc_discard_prealloc() 201 kfree(peer); in rxrpc_discard_prealloc() 249 if (call->peer->rtt_count < 3 || in rxrpc_send_ping() 250 ktime_before(ktime_add_ms(call->peer->rtt_last_req, 1000), now)) in rxrpc_send_ping() 262 struct rxrpc_peer *peer, in rxrpc_alloc_incoming_call() argument 290 if (peer && !rxrpc_get_peer_maybe(peer)) in rxrpc_alloc_incoming_call() [all …]
|
D | proc.c | 53 struct rxrpc_peer *peer; in rxrpc_call_seq_show() local 82 peer = call->peer; in rxrpc_call_seq_show() 83 if (peer) in rxrpc_call_seq_show() 84 sprintf(rbuff, "%pISpc", &peer->srx.transport); in rxrpc_call_seq_show() 176 sprintf(rbuff, "%pISpc", &conn->params.peer->srx.transport); in rxrpc_connection_seq_show() 211 struct rxrpc_peer *peer; in rxrpc_peer_seq_show() local 224 peer = list_entry(v, struct rxrpc_peer, hash_link); in rxrpc_peer_seq_show() 226 sprintf(lbuff, "%pISpc", &peer->local->srx.transport); in rxrpc_peer_seq_show() 228 sprintf(rbuff, "%pISpc", &peer->srx.transport); in rxrpc_peer_seq_show() 236 refcount_read(&peer->ref), in rxrpc_peer_seq_show() [all …]
|
D | output.c | 119 mtu = conn->params.peer->if_mtu; in rxrpc_fill_out_ack() 120 mtu -= conn->params.peer->hdrsize; in rxrpc_fill_out_ack() 201 msg.msg_name = &call->peer->srx.transport; in rxrpc_send_ack_packet() 202 msg.msg_namelen = call->peer->srx.transport_len; in rxrpc_send_ack_packet() 257 conn->params.peer->last_tx_at = ktime_get_seconds(); in rxrpc_send_ack_packet() 310 msg.msg_name = &call->peer->srx.transport; in rxrpc_send_abort_packet() 311 msg.msg_namelen = call->peer->srx.transport_len; in rxrpc_send_abort_packet() 336 conn->params.peer->last_tx_at = ktime_get_seconds(); in rxrpc_send_abort_packet() 365 spin_lock_bh(&call->peer->lock); in rxrpc_send_data_packet() 366 hlist_add_head_rcu(&call->error_link, &call->peer->error_targets); in rxrpc_send_data_packet() [all …]
|
D | conn_object.c | 78 struct rxrpc_peer *peer; in rxrpc_find_connection_rcu() local 102 peer = rxrpc_lookup_peer_rcu(local, &srx); in rxrpc_find_connection_rcu() 103 if (!peer) in rxrpc_find_connection_rcu() 105 *_peer = peer; in rxrpc_find_connection_rcu() 106 conn = rxrpc_find_service_conn_rcu(peer, skb); in rxrpc_find_connection_rcu() 126 peer = conn->params.peer; in rxrpc_find_connection_rcu() 129 if (peer->srx.transport.sin.sin_port != in rxrpc_find_connection_rcu() 131 peer->srx.transport.sin.sin_addr.s_addr != in rxrpc_find_connection_rcu() 137 if (peer->srx.transport.sin6.sin6_port != in rxrpc_find_connection_rcu() 139 memcmp(&peer->srx.transport.sin6.sin6_addr, in rxrpc_find_connection_rcu() [all …]
|
/net/tipc/ |
D | monitor.c | 166 static struct tipc_peer *peer_prev(struct tipc_peer *peer) in peer_prev() argument 168 return list_last_entry(&peer->list, struct tipc_peer, list); in peer_prev() 171 static struct tipc_peer *peer_nxt(struct tipc_peer *peer) in peer_nxt() argument 173 return list_first_entry(&peer->list, struct tipc_peer, list); in peer_nxt() 176 static struct tipc_peer *peer_head(struct tipc_peer *peer) in peer_head() argument 178 while (!peer->is_head) in peer_head() 179 peer = peer_prev(peer); in peer_head() 180 return peer; in peer_head() 185 struct tipc_peer *peer; in get_peer() local 188 hlist_for_each_entry(peer, &mon->peers[thash], hash) { in get_peer() [all …]
|
/net/sctp/ |
D | associola.c | 170 asoc->peer.rwnd = SCTP_DEFAULT_MAXWINDOW; in sctp_association_init() 206 INIT_LIST_HEAD(&asoc->peer.transport_addr_list); in sctp_association_init() 219 asoc->peer.sack_needed = 1; in sctp_association_init() 220 asoc->peer.sack_generation = 1; in sctp_association_init() 242 asoc->peer.ipv4_address = 1; in sctp_association_init() 244 asoc->peer.ipv6_address = 1; in sctp_association_init() 349 sctp_tsnmap_free(&asoc->peer.tsn_map); in sctp_association_free() 371 kfree(asoc->peer.cookie); in sctp_association_free() 372 kfree(asoc->peer.peer_random); in sctp_association_free() 373 kfree(asoc->peer.peer_chunks); in sctp_association_free() [all …]
|
D | sm_make_chunk.c | 423 if (asoc->peer.ecn_capable) in sctp_make_init_ack() 426 if (asoc->peer.prsctp_capable) in sctp_make_init_ack() 429 if (asoc->peer.asconf_capable) { in sctp_make_init_ack() 435 if (asoc->peer.reconf_capable) { in sctp_make_init_ack() 443 if (asoc->peer.intl_capable) { in sctp_make_init_ack() 448 if (asoc->peer.auth_capable) { in sctp_make_init_ack() 494 if (asoc->peer.ecn_capable) in sctp_make_init_ack() 502 if (asoc->peer.prsctp_capable) in sctp_make_init_ack() 512 if (asoc->peer.auth_capable) { in sctp_make_init_ack() 574 cookie = asoc->peer.cookie; in sctp_make_cookie_echo() [all …]
|
D | transport.c | 41 struct sctp_transport *peer, in sctp_transport_init() argument 46 peer->af_specific = sctp_get_af_specific(addr->sa.sa_family); in sctp_transport_init() 47 memcpy(&peer->ipaddr, addr, peer->af_specific->sockaddr_len); in sctp_transport_init() 48 memset(&peer->saddr, 0, sizeof(union sctp_addr)); in sctp_transport_init() 50 peer->sack_generation = 0; in sctp_transport_init() 58 peer->rto = msecs_to_jiffies(net->sctp.rto_initial); in sctp_transport_init() 60 peer->last_time_heard = 0; in sctp_transport_init() 61 peer->last_time_ecne_reduced = jiffies; in sctp_transport_init() 63 peer->param_flags = SPP_HB_DISABLE | in sctp_transport_init() 68 peer->pathmaxrxt = net->sctp.max_retrans_path; in sctp_transport_init() [all …]
|
D | sm_sideeffect.c | 141 struct sctp_transport *trans = asoc->peer.last_data_from; in sctp_gen_sack() 149 asoc->peer.sack_needed = 1; in sctp_gen_sack() 151 ctsn = sctp_tsnmap_get_ctsn(&asoc->peer.tsn_map); in sctp_gen_sack() 152 max_tsn_seen = sctp_tsnmap_get_max_tsn_seen(&asoc->peer.tsn_map); in sctp_gen_sack() 165 asoc->peer.sack_needed = 1; in sctp_gen_sack() 175 if (!asoc->peer.sack_needed) { in sctp_gen_sack() 176 asoc->peer.sack_cnt++; in sctp_gen_sack() 185 if (asoc->peer.sack_cnt >= trans->sackfreq - 1) in sctp_gen_sack() 186 asoc->peer.sack_needed = 1; in sctp_gen_sack() 192 if (asoc->peer.sack_cnt >= asoc->sackfreq - 1) in sctp_gen_sack() [all …]
|
D | outqueue.c | 213 list_for_each_entry(transport, &q->asoc->peer.transport_addr_list, in __sctp_outq_teardown() 300 if (chunk->asoc->peer.prsctp_capable && in sctp_outq_tail() 428 if (!asoc->peer.prsctp_capable || !asoc->sent_cnt_removable) in sctp_prsctp_prune() 437 list_for_each_entry(transport, &asoc->peer.transport_addr_list, in sctp_prsctp_prune() 477 q->asoc->peer.rwnd += sctp_data_size(chunk); in sctp_retransmit_mark() 497 q->asoc->peer.rwnd += sctp_data_size(chunk); in sctp_retransmit_mark() 549 if (transport == transport->asoc->peer.retran_path) in sctp_retransmit() 769 const __u16 dport = asoc->peer.port; in sctp_packet_singleton() 770 const __u32 vtag = asoc->peer.i.init_tag; in sctp_packet_singleton() 825 new_transport = ctx->asoc->peer.active_path; in sctp_outq_select_transport() [all …]
|
D | proc.c | 81 struct sctp_transport *peer; in sctp_seq_dump_local_addrs() local 88 peer = asoc->peer.primary_path; in sctp_seq_dump_local_addrs() 89 if (unlikely(peer == NULL)) { in sctp_seq_dump_local_addrs() 94 primary = &peer->saddr; in sctp_seq_dump_local_addrs() 119 primary = &assoc->peer.primary_addr; in sctp_seq_dump_remote_addrs() 120 list_for_each_entry_rcu(transport, &assoc->peer.transport_addr_list, in sctp_seq_dump_remote_addrs() 275 assoc->peer.port); in sctp_assocs_seq_show() 316 list_for_each_entry_rcu(tsp, &assoc->peer.transport_addr_list, in sctp_remaddr_seq_show()
|
D | auth.c | 236 return sctp_auth_make_key_vector(asoc->peer.peer_random, in sctp_auth_make_peer_vector() 237 asoc->peer.peer_chunks, in sctp_auth_make_peer_vector() 238 asoc->peer.peer_hmacs, in sctp_auth_make_peer_vector() 392 if (!asoc->peer.auth_capable) in sctp_auth_asoc_init_active_key() 538 hmacs = asoc->peer.peer_hmacs; in sctp_auth_asoc_get_hmac() 679 if (!asoc->peer.auth_capable) in sctp_auth_send_cid() 682 return __sctp_auth_cid(chunk, asoc->peer.peer_chunks); in sctp_auth_send_cid() 691 if (!asoc->peer.auth_capable) in sctp_auth_recv_cid() 830 if (!asoc->peer.auth_capable) in sctp_auth_set_key() 890 if (!asoc->peer.auth_capable) in sctp_auth_set_active_key() [all …]
|
/net/bluetooth/ |
D | 6lowpan.c | 84 struct lowpan_peer *peer) in peer_add() argument 86 list_add_rcu(&peer->list, &dev->peers); in peer_add() 91 struct lowpan_peer *peer) in peer_del() argument 93 list_del_rcu(&peer->list); in peer_del() 94 kfree_rcu(peer, rcu); in peer_del() 109 struct lowpan_peer *peer; in __peer_lookup_chan() local 111 list_for_each_entry_rcu(peer, &dev->peers, list) { in __peer_lookup_chan() 112 if (peer->chan == chan) in __peer_lookup_chan() 113 return peer; in __peer_lookup_chan() 122 struct lowpan_peer *peer; in __peer_lookup_conn() local [all …]
|
/net/smc/ |
D | smc_cdc.h | 195 static inline void smc_host_cursor_to_cdc(union smc_cdc_cursor *peer, in smc_host_cursor_to_cdc() argument 201 peer->count = htonl(save->count); in smc_host_cursor_to_cdc() 202 peer->wrap = htons(save->wrap); in smc_host_cursor_to_cdc() 206 static inline void smc_host_msg_to_cdc(struct smc_cdc_msg *peer, in smc_host_msg_to_cdc() argument 212 peer->common.type = local->common.type; in smc_host_msg_to_cdc() 213 peer->len = local->len; in smc_host_msg_to_cdc() 214 peer->seqno = htons(local->seqno); in smc_host_msg_to_cdc() 215 peer->token = htonl(local->token); in smc_host_msg_to_cdc() 216 smc_host_cursor_to_cdc(&peer->prod, &local->prod, save, conn); in smc_host_msg_to_cdc() 217 smc_host_cursor_to_cdc(&peer->cons, &local->cons, save, conn); in smc_host_msg_to_cdc() [all …]
|
/net/core/ |
D | net_namespace.c | 188 static int alloc_netid(struct net *net, struct net *peer, int reqid) in alloc_netid() argument 197 return idr_alloc(&net->netns_ids, peer, min, max, GFP_ATOMIC); in alloc_netid() 206 static int net_eq_idr(int id, void *net, void *peer) in net_eq_idr() argument 208 if (net_eq(net, peer)) in net_eq_idr() 214 static int __peernet2id(const struct net *net, struct net *peer) in __peernet2id() argument 216 int id = idr_for_each(&net->netns_ids, net_eq_idr, peer); in __peernet2id() 232 int peernet2id_alloc(struct net *net, struct net *peer, gfp_t gfp) in peernet2id_alloc() argument 240 id = __peernet2id(net, peer); in peernet2id_alloc() 251 if (!maybe_get_net(peer)) { in peernet2id_alloc() 256 id = alloc_netid(net, peer, -1); in peernet2id_alloc() [all …]
|
/net/mac80211/ |
D | tdls.c | 202 struct sk_buff *skb, const u8 *peer, in ieee80211_tdls_add_link_ie() argument 210 rsp_addr = peer; in ieee80211_tdls_add_link_ie() 212 init_addr = peer; in ieee80211_tdls_add_link_ie() 362 struct sk_buff *skb, const u8 *peer, in ieee80211_tdls_add_setup_start_ies() argument 432 sta = sta_info_get(sdata, peer); in ieee80211_tdls_add_setup_start_ies() 474 ieee80211_tdls_add_link_ie(sdata, skb, peer, initiator); in ieee80211_tdls_add_setup_start_ies() 543 struct sk_buff *skb, const u8 *peer, in ieee80211_tdls_add_setup_cfm_ies() argument 560 sta = sta_info_get(sdata, peer); in ieee80211_tdls_add_setup_cfm_ies() 617 ieee80211_tdls_add_link_ie(sdata, skb, peer, initiator); in ieee80211_tdls_add_setup_cfm_ies() 645 struct sk_buff *skb, const u8 *peer, in ieee80211_tdls_add_chan_switch_req_ies() argument [all …]
|
/net/unix/ |
D | diag.c | 44 struct sock *peer; in sk_diag_dump_peer() local 47 peer = unix_peer_get(sk); in sk_diag_dump_peer() 48 if (peer) { in sk_diag_dump_peer() 49 unix_state_lock(peer); in sk_diag_dump_peer() 50 ino = sock_i_ino(peer); in sk_diag_dump_peer() 51 unix_state_unlock(peer); in sk_diag_dump_peer() 52 sock_put(peer); in sk_diag_dump_peer() 78 struct sock *req, *peer; in sk_diag_dump_icons() local 87 peer = unix_sk(req)->peer; in sk_diag_dump_icons() 88 buf[i++] = (peer ? sock_i_ino(peer) : 0); in sk_diag_dump_icons()
|
/net/ipv4/ |
D | inetpeer.c | 270 bool inet_peer_xrlim_allow(struct inet_peer *peer, int timeout) in inet_peer_xrlim_allow() argument 275 if (!peer) in inet_peer_xrlim_allow() 278 token = peer->rate_tokens; in inet_peer_xrlim_allow() 280 token += now - peer->rate_last; in inet_peer_xrlim_allow() 281 peer->rate_last = now; in inet_peer_xrlim_allow() 288 peer->rate_tokens = token; in inet_peer_xrlim_allow() 298 struct inet_peer *peer = rb_entry(p, struct inet_peer, rb_node); in inetpeer_invalidate_tree() local 301 rb_erase(&peer->rb_node, &base->rb_root); in inetpeer_invalidate_tree() 302 inet_putpeer(peer); in inetpeer_invalidate_tree()
|
/net/wireless/ |
D | pmsr.c | 187 struct nlattr *peer, in pmsr_parse_peer() argument 197 nla_parse_nested_deprecated(tb, NL80211_PMSR_PEER_ATTR_MAX, peer, in pmsr_parse_peer() 203 NL_SET_ERR_MSG_ATTR(info->extack, peer, in pmsr_parse_peer() 268 struct nlattr *peers, *peer; in nl80211_pmsr_start() local 283 nla_for_each_nested(peer, peers, rem) { in nl80211_pmsr_start() 287 NL_SET_ERR_MSG_ATTR(info->extack, peer, in nl80211_pmsr_start() 319 nla_for_each_nested(peer, peers, rem) { in nl80211_pmsr_start() 321 err = pmsr_parse_peer(rdev, peer, &req->peers[idx], info); in nl80211_pmsr_start() 488 struct nlattr *pmsr, *peers, *peer, *resp, *data, *typedata; in nl80211_pmsr_send_result() local 498 peer = nla_nest_start_noflag(msg, 1); in nl80211_pmsr_send_result() [all …]
|
D | trace.h | 1653 const u8 *peer, const struct cfg80211_bitrate_mask *mask), 1654 TP_ARGS(wiphy, netdev, link_id, peer, mask), 1659 MAC_ENTRY(peer) 1665 MAC_ASSIGN(peer, peer); 1669 MAC_PR_ARG(peer)) 1786 u8 *peer, u8 action_code, u8 dialog_token, 1789 TP_ARGS(wiphy, netdev, peer, action_code, dialog_token, status_code, 1794 MAC_ENTRY(peer) 1805 MAC_ASSIGN(peer, peer); 1816 WIPHY_PR_ARG, NETDEV_PR_ARG, MAC_PR_ARG(peer), [all …]
|
/net/x25/ |
D | x25_forward.c | 93 struct net_device *peer = NULL; in x25_forward_data() local 103 peer = frwd->dev2; in x25_forward_data() 105 peer = frwd->dev1; in x25_forward_data() 112 if ( (nb = x25_get_neigh(peer)) == NULL) in x25_forward_data()
|