/net/rxrpc/ |
D | input.c | 420 u8 ack = 0; in rxrpc_input_data() local 463 ack = RXRPC_ACK_NOSPACE; in rxrpc_input_data() 465 goto ack; in rxrpc_input_data() 504 ack = RXRPC_ACK_DUPLICATE; in rxrpc_input_data() 512 if (ack != RXRPC_ACK_DUPLICATE) { in rxrpc_input_data() 513 ack = RXRPC_ACK_DUPLICATE; in rxrpc_input_data() 521 ack = RXRPC_ACK_EXCEEDS_WINDOW; in rxrpc_input_data() 530 goto ack; in rxrpc_input_data() 533 if (flags & RXRPC_REQUEST_ACK && !ack) { in rxrpc_input_data() 534 ack = RXRPC_ACK_REQUESTED; in rxrpc_input_data() [all …]
|
D | conn_event.c | 34 struct rxrpc_ackpacket ack; in rxrpc_conn_retransmit_call() member 91 pkt.ack.bufferSpace = 0; in rxrpc_conn_retransmit_call() 92 pkt.ack.maxSkew = htons(skb ? skb->priority : 0); in rxrpc_conn_retransmit_call() 93 pkt.ack.firstPacket = htonl(chan->last_seq + 1); in rxrpc_conn_retransmit_call() 94 pkt.ack.previousPacket = htonl(chan->last_seq); in rxrpc_conn_retransmit_call() 95 pkt.ack.serial = htonl(skb ? sp->hdr.serial : 0); in rxrpc_conn_retransmit_call() 96 pkt.ack.reason = skb ? RXRPC_ACK_DUPLICATE : RXRPC_ACK_IDLE; in rxrpc_conn_retransmit_call() 97 pkt.ack.nAcks = 0; in rxrpc_conn_retransmit_call() 104 iov[0].iov_len += sizeof(pkt.ack); in rxrpc_conn_retransmit_call() 105 len += sizeof(pkt.ack) + 3 + sizeof(ack_info); in rxrpc_conn_retransmit_call() [all …]
|
D | output.c | 20 struct rxrpc_ackpacket ack; member 89 pkt->ack.bufferSpace = htons(8); in rxrpc_fill_out_ack() 90 pkt->ack.maxSkew = htons(0); in rxrpc_fill_out_ack() 91 pkt->ack.firstPacket = htonl(hard_ack + 1); in rxrpc_fill_out_ack() 92 pkt->ack.previousPacket = htonl(call->ackr_prev_seq); in rxrpc_fill_out_ack() 93 pkt->ack.serial = htonl(serial); in rxrpc_fill_out_ack() 94 pkt->ack.reason = reason; in rxrpc_fill_out_ack() 95 pkt->ack.nAcks = top - hard_ack; in rxrpc_fill_out_ack() 189 iov[0].iov_len = sizeof(pkt->whdr) + sizeof(pkt->ack) + n; in rxrpc_send_ack_packet() 197 ntohl(pkt->ack.firstPacket), in rxrpc_send_ack_packet() [all …]
|
/net/decnet/ |
D | dn_nsp_in.c | 92 static void dn_ack(struct sock *sk, struct sk_buff *skb, unsigned short ack) in dn_ack() argument 95 unsigned short type = ((ack >> 12) & 0x0003); in dn_ack() 100 if (dn_after(ack, scp->ackrcv_dat)) { in dn_ack() 101 scp->ackrcv_dat = ack & 0x0fff; in dn_ack() 104 ack); in dn_ack() 110 if (dn_after(ack, scp->ackrcv_oth)) { in dn_ack() 111 scp->ackrcv_oth = ack & 0x0fff; in dn_ack() 114 ack); in dn_ack() 132 unsigned short ack; in dn_process_ack() local 137 if ((ack = le16_to_cpu(*ptr)) & 0x8000) { in dn_process_ack() [all …]
|
D | dn_nsp_out.c | 373 struct sk_buff *skb2, *n, *ack = NULL; in dn_nsp_check_xmit_queue() local 385 ack = skb2; in dn_nsp_check_xmit_queue() 389 if (ack == NULL) in dn_nsp_check_xmit_queue() 403 skb_unlink(ack, q); in dn_nsp_check_xmit_queue() 404 kfree_skb(ack); in dn_nsp_check_xmit_queue() 405 ack = NULL; in dn_nsp_check_xmit_queue()
|
/net/netfilter/ |
D | nf_conntrack_proto_tcp.c | 286 else if (tcph->syn) return (tcph->ack ? TCP_SYNACK_SET : TCP_SYN_SET); in get_conntrack_index() 288 else if (tcph->ack) return TCP_ACK_SET; in get_conntrack_index() 474 __u32 seq, ack, sack, end, win, swin; in tcp_in_window() local 483 ack = sack = ntohl(tcph->ack_seq); in tcp_in_window() 492 receiver_offset = nf_ct_seq_offset(ct, !dir, ack - 1); in tcp_in_window() 493 ack -= receiver_offset; in tcp_in_window() 500 seq, ack, receiver_offset, sack, receiver_offset, win, end); in tcp_in_window() 531 if (!tcph->ack) in tcp_in_window() 569 if (!(tcph->ack)) { in tcp_in_window() 573 ack = sack = receiver->td_end; in tcp_in_window() [all …]
|
D | nfnetlink.c | 381 goto ack; in nfnetlink_rcv_batch() 395 goto ack; in nfnetlink_rcv_batch() 403 goto ack; in nfnetlink_rcv_batch() 409 goto ack; in nfnetlink_rcv_batch() 422 goto ack; in nfnetlink_rcv_batch() 430 goto ack; in nfnetlink_rcv_batch() 447 ack: in nfnetlink_rcv_batch()
|
/net/ipv4/ |
D | tcp_vegas.c | 165 static void tcp_vegas_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_vegas_cong_avoid() argument 171 tcp_reno_cong_avoid(sk, ack, acked); in tcp_vegas_cong_avoid() 175 if (after(ack, vegas->beg_snd_nxt)) { in tcp_vegas_cong_avoid() 196 tcp_reno_cong_avoid(sk, ack, acked); in tcp_vegas_cong_avoid()
|
D | tcp_veno.c | 119 static void tcp_veno_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_veno_cong_avoid() argument 125 tcp_reno_cong_avoid(sk, ack, acked); in tcp_veno_cong_avoid() 138 tcp_reno_cong_avoid(sk, ack, acked); in tcp_veno_cong_avoid()
|
D | tcp_input.c | 2980 static void tcp_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_cong_avoid() argument 2984 icsk->icsk_ca_ops->cong_avoid(sk, ack, acked); in tcp_cong_avoid() 3323 static void tcp_cong_control(struct sock *sk, u32 ack, u32 acked_sacked, in tcp_cong_control() argument 3338 tcp_cong_avoid(sk, ack, acked_sacked); in tcp_cong_control() 3347 const u32 ack, const u32 ack_seq, in tcp_may_update_window() argument 3350 return after(ack, tp->snd_una) || in tcp_may_update_window() 3356 static void tcp_snd_una_update(struct tcp_sock *tp, u32 ack) in tcp_snd_una_update() argument 3358 u32 delta = ack - tp->snd_una; in tcp_snd_una_update() 3362 tp->snd_una = ack; in tcp_snd_una_update() 3380 static int tcp_ack_update_window(struct sock *sk, const struct sk_buff *skb, u32 ack, in tcp_ack_update_window() argument [all …]
|
D | tcp_hybla.c | 90 static void hybla_cong_avoid(struct sock *sk, u32 ack, u32 acked) in hybla_cong_avoid() argument 107 tcp_reno_cong_avoid(sk, ack, acked); in hybla_cong_avoid()
|
D | tcp_yeah.c | 72 static void tcp_yeah_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_yeah_cong_avoid() argument 122 if (after(ack, yeah->vegas.beg_snd_nxt)) { in tcp_yeah_cong_avoid()
|
D | tcp_cdg.c | 262 static void tcp_cdg_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_cdg_cong_avoid() argument 272 if (after(ack, ca->rtt_seq) && ca->rtt.v64) { in tcp_cdg_cong_avoid() 293 tcp_reno_cong_avoid(sk, ack, acked); in tcp_cdg_cong_avoid()
|
D | tcp_lp.c | 119 static void tcp_lp_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_lp_cong_avoid() argument 124 tcp_reno_cong_avoid(sk, ack, acked); in tcp_lp_cong_avoid()
|
D | tcp_illinois.c | 260 static void tcp_illinois_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_illinois_cong_avoid() argument 265 if (after(ack, ca->end_seq)) in tcp_illinois_cong_avoid()
|
D | tcp_scalable.c | 19 static void tcp_scalable_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_scalable_cong_avoid() argument
|
/net/sctp/ |
D | associola.c | 1659 struct sctp_chunk *ack; in sctp_assoc_free_asconf_acks() local 1662 list_for_each_entry_safe(ack, tmp, &asoc->asconf_ack_list, in sctp_assoc_free_asconf_acks() 1664 list_del_init(&ack->transmitted_list); in sctp_assoc_free_asconf_acks() 1665 sctp_chunk_free(ack); in sctp_assoc_free_asconf_acks() 1672 struct sctp_chunk *ack; in sctp_assoc_clean_asconf_ack_cache() local 1678 list_for_each_entry_safe(ack, tmp, &asoc->asconf_ack_list, in sctp_assoc_clean_asconf_ack_cache() 1680 if (ack->subh.addip_hdr->serial == in sctp_assoc_clean_asconf_ack_cache() 1684 list_del_init(&ack->transmitted_list); in sctp_assoc_clean_asconf_ack_cache() 1685 sctp_chunk_free(ack); in sctp_assoc_clean_asconf_ack_cache() 1694 struct sctp_chunk *ack; in sctp_assoc_lookup_asconf_ack() local [all …]
|
/net/ipv6/netfilter/ |
D | ip6t_SYNPROXY.c | 31 if (th->syn && !(th->ack || th->fin || th->rst)) { in synproxy_tg6() 52 } else if (th->ack && !(th->fin || th->rst || th->syn)) { in synproxy_tg6()
|
D | nf_reject_ipv6.c | 101 if (oth->ack) { in nf_reject_ip6_tcphdr_put() 115 tcph->ack = needs_ack; in nf_reject_ip6_tcphdr_put()
|
/net/ipv4/netfilter/ |
D | ipt_SYNPROXY.c | 31 if (th->syn && !(th->ack || th->fin || th->rst)) { in synproxy_tg4() 51 } else if (th->ack && !(th->fin || th->rst || th->syn)) { in synproxy_tg4()
|
D | nf_reject_ipv4.c | 81 if (oth->ack) { in nf_reject_ip_tcphdr_put() 87 tcph->ack = 1; in nf_reject_ip_tcphdr_put()
|
/net/tipc/ |
D | link.c | 356 u16 ack = snd_l->snd_nxt - 1; in tipc_link_remove_bc_peer() local 361 tipc_link_bc_ack_rcv(rcv_l, ack, xmitq); in tipc_link_remove_bc_peer() 946 u16 ack = l->rcv_nxt - 1; in tipc_link_xmit() local 982 msg_set_ack(hdr, ack); in tipc_link_xmit() 1031 u16 ack = l->rcv_nxt - 1; in tipc_link_advance_backlog() local 1056 msg_set_ack(hdr, ack); in tipc_link_advance_backlog() 1130 u16 ack = l->rcv_nxt - 1; in tipc_link_bc_retrans() local 1158 msg_set_ack(hdr, ack); in tipc_link_bc_retrans() 1363 ga->gacks[n].ack = htons(expect - 1); in tipc_build_gap_ack_blks() 1377 ga->gacks[n].ack = htons(seqno); in tipc_build_gap_ack_blks() [all …]
|
D | group.c | 388 void tipc_group_update_bc_members(struct tipc_group *grp, int len, bool ack) in tipc_group_update_bc_members() argument 405 if (ack) in tipc_group_update_bc_members() 493 bool ack, deliver, update, leave = false; in tipc_group_filter_msg() local 526 ack = false; in tipc_group_filter_msg() 542 ack = msg_grp_bc_ack_req(hdr); in tipc_group_filter_msg() 563 if (ack) in tipc_group_filter_msg()
|
/net/rds/ |
D | tcp_send.c | 176 static int rds_tcp_is_acked(struct rds_message *rm, uint64_t ack) in rds_tcp_is_acked() argument 180 return (__s32)((u32)rm->m_ack_seq - (u32)ack) < 0; in rds_tcp_is_acked()
|
D | send.c | 486 static inline int rds_send_is_acked(struct rds_message *rm, u64 ack, in rds_send_is_acked() argument 490 return is_acked(rm, ack); in rds_send_is_acked() 491 return be64_to_cpu(rm->m_inc.i_hdr.h_sequence) <= ack; in rds_send_is_acked() 686 void rds_send_path_drop_acked(struct rds_conn_path *cp, u64 ack, in rds_send_path_drop_acked() argument 696 if (!rds_send_is_acked(rm, ack, is_acked)) in rds_send_path_drop_acked() 714 void rds_send_drop_acked(struct rds_connection *conn, u64 ack, in rds_send_drop_acked() argument 718 rds_send_path_drop_acked(&conn->c_path[0], ack, is_acked); in rds_send_drop_acked()
|