/net/rxrpc/ |
D | input.c | 419 u8 ack = 0; in rxrpc_input_data() local 461 ack = RXRPC_ACK_NOSPACE; in rxrpc_input_data() 463 goto ack; in rxrpc_input_data() 502 ack = RXRPC_ACK_DUPLICATE; in rxrpc_input_data() 510 if (ack != RXRPC_ACK_DUPLICATE) { in rxrpc_input_data() 511 ack = RXRPC_ACK_DUPLICATE; in rxrpc_input_data() 519 ack = RXRPC_ACK_EXCEEDS_WINDOW; in rxrpc_input_data() 528 goto ack; in rxrpc_input_data() 531 if (flags & RXRPC_REQUEST_ACK && !ack) { in rxrpc_input_data() 532 ack = RXRPC_ACK_REQUESTED; in rxrpc_input_data() [all …]
|
D | conn_event.c | 34 struct rxrpc_ackpacket ack; in rxrpc_conn_retransmit_call() member 46 &pkt.ack, sizeof(pkt.ack)) < 0) in rxrpc_conn_retransmit_call() 48 if (pkt.ack.reason == RXRPC_ACK_PING_RESPONSE) in rxrpc_conn_retransmit_call() 99 pkt.ack.bufferSpace = 0; in rxrpc_conn_retransmit_call() 100 pkt.ack.maxSkew = htons(skb ? skb->priority : 0); in rxrpc_conn_retransmit_call() 101 pkt.ack.firstPacket = htonl(chan->last_seq + 1); in rxrpc_conn_retransmit_call() 102 pkt.ack.previousPacket = htonl(chan->last_seq); in rxrpc_conn_retransmit_call() 103 pkt.ack.serial = htonl(skb ? sp->hdr.serial : 0); in rxrpc_conn_retransmit_call() 104 pkt.ack.reason = skb ? RXRPC_ACK_DUPLICATE : RXRPC_ACK_IDLE; in rxrpc_conn_retransmit_call() 105 pkt.ack.nAcks = 0; in rxrpc_conn_retransmit_call() [all …]
|
D | output.c | 20 struct rxrpc_ackpacket ack; member 96 pkt->ack.bufferSpace = htons(0); in rxrpc_fill_out_ack() 97 pkt->ack.maxSkew = htons(0); in rxrpc_fill_out_ack() 98 pkt->ack.firstPacket = htonl(hard_ack + 1); in rxrpc_fill_out_ack() 99 pkt->ack.previousPacket = htonl(call->ackr_highest_seq); in rxrpc_fill_out_ack() 100 pkt->ack.serial = htonl(serial); in rxrpc_fill_out_ack() 101 pkt->ack.reason = reason; in rxrpc_fill_out_ack() 102 pkt->ack.nAcks = top - hard_ack; in rxrpc_fill_out_ack() 239 iov[0].iov_len = sizeof(pkt->whdr) + sizeof(pkt->ack) + n; in rxrpc_send_ack_packet() 247 ntohl(pkt->ack.firstPacket), in rxrpc_send_ack_packet() [all …]
|
/net/netfilter/ |
D | nf_conntrack_proto_tcp.c | 272 else if (tcph->syn) return (tcph->ack ? TCP_SYNACK_SET : TCP_SYN_SET); in get_conntrack_index() 274 else if (tcph->ack) return TCP_ACK_SET; in get_conntrack_index() 463 __u32 seq, ack, sack, end, win, swin; in tcp_in_window() local 472 ack = sack = ntohl(tcph->ack_seq); in tcp_in_window() 481 receiver_offset = nf_ct_seq_offset(ct, !dir, ack - 1); in tcp_in_window() 482 ack -= receiver_offset; in tcp_in_window() 489 seq, ack, receiver_offset, sack, receiver_offset, win, end); in tcp_in_window() 520 if (!tcph->ack) in tcp_in_window() 565 if (!(tcph->ack)) { in tcp_in_window() 569 ack = sack = receiver->td_end; in tcp_in_window() [all …]
|
D | nfnetlink.c | 450 goto ack; in nfnetlink_rcv_batch() 464 goto ack; in nfnetlink_rcv_batch() 472 goto ack; in nfnetlink_rcv_batch() 478 goto ack; in nfnetlink_rcv_batch() 483 goto ack; in nfnetlink_rcv_batch() 504 goto ack; in nfnetlink_rcv_batch() 512 goto ack; in nfnetlink_rcv_batch() 525 ack: in nfnetlink_rcv_batch()
|
D | nf_synproxy_core.c | 700 if (!th->syn || th->ack || in ipv4_synproxy_hook() 716 if (!th->syn && th->ack && in ipv4_synproxy_hook() 739 if (!th->syn || !th->ack) in ipv4_synproxy_hook() 1123 if (!th->syn || th->ack || in ipv6_synproxy_hook() 1139 if (!th->syn && th->ack && in ipv6_synproxy_hook() 1162 if (!th->syn || !th->ack) in ipv6_synproxy_hook()
|
/net/ipv4/ |
D | tcp_vegas.c | 165 static void tcp_vegas_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_vegas_cong_avoid() argument 171 tcp_reno_cong_avoid(sk, ack, acked); in tcp_vegas_cong_avoid() 175 if (after(ack, vegas->beg_snd_nxt)) { in tcp_vegas_cong_avoid() 196 tcp_reno_cong_avoid(sk, ack, acked); in tcp_vegas_cong_avoid()
|
D | tcp_veno.c | 119 static void tcp_veno_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_veno_cong_avoid() argument 125 tcp_reno_cong_avoid(sk, ack, acked); in tcp_veno_cong_avoid() 138 tcp_reno_cong_avoid(sk, ack, acked); in tcp_veno_cong_avoid()
|
D | tcp_hybla.c | 90 static void hybla_cong_avoid(struct sock *sk, u32 ack, u32 acked) in hybla_cong_avoid() argument 107 tcp_reno_cong_avoid(sk, ack, acked); in hybla_cong_avoid()
|
D | tcp_yeah.c | 58 static void tcp_yeah_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_yeah_cong_avoid() argument 103 if (after(ack, yeah->vegas.beg_snd_nxt)) { in tcp_yeah_cong_avoid()
|
D | tcp_input.c | 3155 static void tcp_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_cong_avoid() argument 3159 icsk->icsk_ca_ops->cong_avoid(sk, ack, acked); in tcp_cong_avoid() 3503 static void tcp_cong_control(struct sock *sk, u32 ack, u32 acked_sacked, in tcp_cong_control() argument 3518 tcp_cong_avoid(sk, ack, acked_sacked); in tcp_cong_control() 3527 const u32 ack, const u32 ack_seq, in tcp_may_update_window() argument 3530 return after(ack, tp->snd_una) || in tcp_may_update_window() 3536 static void tcp_snd_una_update(struct tcp_sock *tp, u32 ack) in tcp_snd_una_update() argument 3538 u32 delta = ack - tp->snd_una; in tcp_snd_una_update() 3542 tp->snd_una = ack; in tcp_snd_una_update() 3560 static int tcp_ack_update_window(struct sock *sk, const struct sk_buff *skb, u32 ack, in tcp_ack_update_window() argument [all …]
|
D | tcp_cdg.c | 262 static void tcp_cdg_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_cdg_cong_avoid() argument 272 if (after(ack, ca->rtt_seq) && ca->rtt.v64) { in tcp_cdg_cong_avoid() 293 tcp_reno_cong_avoid(sk, ack, acked); in tcp_cdg_cong_avoid()
|
D | tcp_lp.c | 121 static void tcp_lp_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_lp_cong_avoid() argument 126 tcp_reno_cong_avoid(sk, ack, acked); in tcp_lp_cong_avoid()
|
D | tcp_illinois.c | 260 static void tcp_illinois_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_illinois_cong_avoid() argument 265 if (after(ack, ca->end_seq)) in tcp_illinois_cong_avoid()
|
D | tcp_scalable.c | 18 static void tcp_scalable_cong_avoid(struct sock *sk, u32 ack, u32 acked) in tcp_scalable_cong_avoid() argument
|
D | tcp_highspeed.c | 112 static void hstcp_cong_avoid(struct sock *sk, u32 ack, u32 acked) in hstcp_cong_avoid() argument
|
/net/sctp/ |
D | associola.c | 1666 struct sctp_chunk *ack; in sctp_assoc_free_asconf_acks() local 1669 list_for_each_entry_safe(ack, tmp, &asoc->asconf_ack_list, in sctp_assoc_free_asconf_acks() 1671 list_del_init(&ack->transmitted_list); in sctp_assoc_free_asconf_acks() 1672 sctp_chunk_free(ack); in sctp_assoc_free_asconf_acks() 1679 struct sctp_chunk *ack; in sctp_assoc_clean_asconf_ack_cache() local 1685 list_for_each_entry_safe(ack, tmp, &asoc->asconf_ack_list, in sctp_assoc_clean_asconf_ack_cache() 1687 if (ack->subh.addip_hdr->serial == in sctp_assoc_clean_asconf_ack_cache() 1691 list_del_init(&ack->transmitted_list); in sctp_assoc_clean_asconf_ack_cache() 1692 sctp_chunk_free(ack); in sctp_assoc_clean_asconf_ack_cache() 1701 struct sctp_chunk *ack; in sctp_assoc_lookup_asconf_ack() local [all …]
|
/net/ipv4/netfilter/ |
D | ipt_SYNPROXY.c | 31 if (th->syn && !(th->ack || th->fin || th->rst)) { in synproxy_tg4() 51 } else if (th->ack && !(th->fin || th->rst || th->syn)) { in synproxy_tg4()
|
D | nf_reject_ipv4.c | 203 if (oth->ack) { in nf_reject_ip_tcphdr_put() 209 tcph->ack = 1; in nf_reject_ip_tcphdr_put()
|
/net/ipv6/netfilter/ |
D | ip6t_SYNPROXY.c | 31 if (th->syn && !(th->ack || th->fin || th->rst)) { in synproxy_tg6() 52 } else if (th->ack && !(th->fin || th->rst || th->syn)) { in synproxy_tg6()
|
/net/tipc/ |
D | link.c | 401 u16 ack = snd_l->snd_nxt - 1; in tipc_link_remove_bc_peer() local 406 tipc_link_bc_ack_rcv(rcv_l, ack, 0, NULL, xmitq, NULL); in tipc_link_remove_bc_peer() 1043 u16 ack = l->rcv_nxt - 1; in tipc_link_xmit() local 1086 msg_set_ack(hdr, ack); in tipc_link_xmit() 1176 u16 ack = l->rcv_nxt - 1; in tipc_link_advance_backlog() local 1201 msg_set_ack(hdr, ack); in tipc_link_advance_backlog() 1483 gacks[n].ack = htons(expect - 1); in __tipc_build_gap_ack_blks() 1499 gacks[n].ack = htons(seqno); in __tipc_build_gap_ack_blks() 1568 u16 ack = l->rcv_nxt - 1; in tipc_link_advance_transmq() local 1613 end = ntohs(last_ga->gacks[si].ack); in tipc_link_advance_transmq() [all …]
|
D | group.c | 395 void tipc_group_update_bc_members(struct tipc_group *grp, int len, bool ack) in tipc_group_update_bc_members() argument 412 if (ack) in tipc_group_update_bc_members() 500 bool ack, deliver, update, leave = false; in tipc_group_filter_msg() local 533 ack = false; in tipc_group_filter_msg() 549 ack = msg_grp_bc_ack_req(hdr); in tipc_group_filter_msg() 570 if (ack) in tipc_group_filter_msg()
|
D | group.h | 68 void tipc_group_update_bc_members(struct tipc_group *grp, int len, bool ack);
|
/net/rds/ |
D | tcp_send.c | 171 static int rds_tcp_is_acked(struct rds_message *rm, uint64_t ack) in rds_tcp_is_acked() argument 175 return (__s32)((u32)rm->m_ack_seq - (u32)ack) < 0; in rds_tcp_is_acked()
|
D | send.c | 486 static inline int rds_send_is_acked(struct rds_message *rm, u64 ack, in rds_send_is_acked() argument 490 return is_acked(rm, ack); in rds_send_is_acked() 491 return be64_to_cpu(rm->m_inc.i_hdr.h_sequence) <= ack; in rds_send_is_acked() 686 void rds_send_path_drop_acked(struct rds_conn_path *cp, u64 ack, in rds_send_path_drop_acked() argument 696 if (!rds_send_is_acked(rm, ack, is_acked)) in rds_send_path_drop_acked() 714 void rds_send_drop_acked(struct rds_connection *conn, u64 ack, in rds_send_drop_acked() argument 718 rds_send_path_drop_acked(&conn->c_path[0], ack, is_acked); in rds_send_drop_acked()
|