/net/ipv6/ |
D | inet6_hashtables.c | 117 struct sk_buff *skb, int doff, in lookup_reuseport() argument 128 reuse_sk = reuseport_select_sock(sk, phash, skb, doff); in lookup_reuseport() 136 struct sk_buff *skb, int doff, in inet6_lhash2_lookup() argument 149 result = lookup_reuseport(net, sk, skb, doff, in inet6_lhash2_lookup() 164 struct sk_buff *skb, int doff, in inet6_lookup_run_bpf() argument 181 reuse_sk = lookup_reuseport(net, sk, skb, doff, saddr, sport, daddr, hnum); in inet6_lookup_run_bpf() 189 struct sk_buff *skb, int doff, in inet6_lookup_listener() argument 200 result = inet6_lookup_run_bpf(net, hashinfo, skb, doff, in inet6_lookup_listener() 209 result = inet6_lhash2_lookup(net, ilb2, skb, doff, in inet6_lookup_listener() 219 result = inet6_lhash2_lookup(net, ilb2, skb, doff, in inet6_lookup_listener() [all …]
|
D | tcp_ipv6.c | 705 if (tcp_v6_md5_hash_headers(hp, daddr, saddr, th, th->doff << 2)) in tcp_v6_md5_hash_hdr() 752 if (tcp_md5_hash_skb_data(hp, skb, th->doff << 2)) in tcp_v6_md5_hash_skb() 920 t1->doff = tot_len / 4; in tcp_v6_send_response() 1078 (th->doff << 2); in tcp_v6_send_reset() 1590 skb->len - th->doff*4); in tcp_v6_fill_cb() 1625 if (unlikely(th->doff < sizeof(struct tcphdr)/4)) in tcp_v6_rcv() 1627 if (!pskb_may_pull(skb, th->doff*4)) in tcp_v6_rcv() 1835 if (th->doff < sizeof(struct tcphdr) / 4) in tcp_v6_early_demux()
|
/net/ipv6/netfilter/ |
D | nf_socket_ipv6.c | 78 nf_socket_get_sock_v6(struct net *net, struct sk_buff *skb, int doff, in nf_socket_get_sock_v6() argument 86 return inet6_lookup(net, &tcp_hashinfo, skb, doff, in nf_socket_get_sock_v6() 104 int doff = 0; in nf_sk_lookup_slow_v6() local 127 doff = tproto == IPPROTO_TCP ? in nf_sk_lookup_slow_v6() 139 return nf_socket_get_sock_v6(net, data_skb, doff, tproto, saddr, daddr, in nf_sk_lookup_slow_v6()
|
D | nf_reject_ipv6.c | 97 tcph->doff = sizeof(struct tcphdr)/4; in nf_reject_ip6_tcphdr_put() 108 otcplen - (oth->doff<<2)); in nf_reject_ip6_tcphdr_put()
|
/net/ipv4/netfilter/ |
D | nf_socket_ipv4.c | 66 nf_socket_get_sock_v4(struct net *net, struct sk_buff *skb, const int doff, in nf_socket_get_sock_v4() argument 74 return inet_lookup(net, &tcp_hashinfo, skb, doff, in nf_socket_get_sock_v4() 96 int doff = 0; in nf_sk_lookup_slow_v4() local 114 doff = iph->protocol == IPPROTO_TCP ? in nf_sk_lookup_slow_v4() 146 return nf_socket_get_sock_v4(net, data_skb, doff, protocol, saddr, in nf_sk_lookup_slow_v4()
|
D | nf_reject_ipv4.c | 79 tcph->doff = sizeof(struct tcphdr) / 4; in nf_reject_ip_tcphdr_put() 86 (oth->doff << 2)); in nf_reject_ip_tcphdr_put()
|
/net/netfilter/ |
D | nf_synproxy_core.c | 28 synproxy_parse_options(const struct sk_buff *skb, unsigned int doff, in synproxy_parse_options() argument 31 int length = (th->doff * 4) - sizeof(*th); in synproxy_parse_options() 37 ptr = skb_header_pointer(skb, doff + sizeof(*th), length, buf); in synproxy_parse_options() 197 optend = protoff + th->doff * 4; in synproxy_tstamp_adjust() 500 nth->doff = tcp_hdr_size / 4; in synproxy_send_client_synack() 546 nth->doff = tcp_hdr_size / 4; in synproxy_send_server_syn() 586 nth->doff = tcp_hdr_size / 4; in synproxy_send_server_ack() 624 nth->doff = tcp_hdr_size / 4; in synproxy_send_client_ack() 913 nth->doff = tcp_hdr_size / 4; in synproxy_send_client_synack_ipv6() 960 nth->doff = tcp_hdr_size / 4; in synproxy_send_server_syn_ipv6() [all …]
|
D | xt_tcpmss.c | 41 if (th->doff*4 < sizeof(*th)) in tcpmss_mt() 44 optlen = th->doff*4 - sizeof(*th); in tcpmss_mt()
|
D | xt_tcpudp.c | 108 if (th->doff * 4 < sizeof(_tcph)) { in tcp_mt() 113 th->doff*4 - sizeof(_tcph), in tcp_mt()
|
D | nf_log_common.c | 107 if ((logflags & NF_LOG_TCPOPT) && th->doff*4 > sizeof(struct tcphdr)) { in nf_log_dump_tcp_header() 111 unsigned int optsize = th->doff*4 - sizeof(struct tcphdr); in nf_log_dump_tcp_header()
|
D | xt_TCPMSS.c | 97 tcp_hdrlen = tcph->doff * 4; in tcpmss_mangle_packet() 189 tcph->doff += TCPOLEN_MSS/4; in tcpmss_mangle_packet()
|
D | nfnetlink_osf.c | 184 if (tcp->doff * 4 > sizeof(struct tcphdr)) { in nf_osf_hdr_ctx_init() 185 ctx->optsize = tcp->doff * 4 - sizeof(struct tcphdr); in nf_osf_hdr_ctx_init()
|
D | xt_TCPOPTSTRIP.c | 48 tcp_hdrlen = tcph->doff * 4; in tcpoptstrip_mangle_packet()
|
D | nf_conntrack_proto_tcp.c | 328 return (seq + len - dataoff - tcph->doff*4 in segment_seq_plus_len() 348 int length = (tcph->doff*4) - sizeof(struct tcphdr); in tcp_options() 403 int length = (tcph->doff*4) - sizeof(struct tcphdr); in tcp_sack() 738 if (th->doff*4 < sizeof(struct tcphdr) || tcplen < th->doff*4) { in tcp_error()
|
D | nf_conntrack_sane.c | 90 dataoff = protoff + th->doff * 4; in help()
|
D | nf_nat_helper.c | 109 mangle_contents(skb, protoff + tcph->doff*4, in __nf_nat_mangle_tcp_packet()
|
D | nf_conntrack_pptp.c | 548 nexthdr_off += tcph->doff * 4; in conntrack_pptp_help() 549 datalen = tcplen - tcph->doff * 4; in conntrack_pptp_help()
|
D | nf_conntrack_irc.c | 139 dataoff = protoff + th->doff*4; in help()
|
D | nf_conntrack_seqadj.c | 127 optend = protoff + tcph->doff * 4; in nf_ct_sack_adjust()
|
/net/ipv4/ |
D | inet_hashtables.c | 256 struct sk_buff *skb, int doff, in lookup_reuseport() argument 265 reuse_sk = reuseport_select_sock(sk, phash, skb, doff); in lookup_reuseport() 280 struct sk_buff *skb, int doff, in inet_lhash2_lookup() argument 293 result = lookup_reuseport(net, sk, skb, doff, in inet_lhash2_lookup() 308 struct sk_buff *skb, int doff, in inet_lookup_run_bpf() argument 323 reuse_sk = lookup_reuseport(net, sk, skb, doff, saddr, sport, daddr, hnum); in inet_lookup_run_bpf() 331 struct sk_buff *skb, int doff, in __inet_lookup_listener() argument 342 result = inet_lookup_run_bpf(net, hashinfo, skb, doff, in __inet_lookup_listener() 351 result = inet_lhash2_lookup(net, ilb2, skb, doff, in __inet_lookup_listener() 361 result = inet_lhash2_lookup(net, ilb2, skb, doff, in __inet_lookup_listener()
|
D | tcp_ipv4.c | 695 rep.th.doff = sizeof(struct tcphdr) / 4; in tcp_v4_send_reset() 703 skb->len - (th->doff << 2)); in tcp_v4_send_reset() 768 rep.th.doff = arg.iov[0].iov_len / 4; in tcp_v4_send_reset() 863 rep.th.doff = arg.iov[0].iov_len / 4; in tcp_v4_send_ack() 878 rep.th.doff = arg.iov[0].iov_len/4; in tcp_v4_send_ack() 1317 if (tcp_v4_md5_hash_headers(hp, daddr, saddr, th, th->doff << 2)) in tcp_v4_md5_hash_hdr() 1363 if (tcp_md5_hash_skb_data(hp, skb, th->doff << 2)) in tcp_v4_md5_hash_skb() 1750 if (th->doff < sizeof(struct tcphdr) / 4) in tcp_v4_early_demux() 1808 hdrlen = th->doff * 4; in tcp_add_backlog() 1827 thtail->doff != th->doff || in tcp_add_backlog() [all …]
|
D | tcp_offload.c | 70 thlen = th->doff * 4; in tcp_gso_segment() 204 thlen = th->doff * 4; in tcp_gro_receive()
|
D | tcp_minisocks.c | 100 if (th->doff > (sizeof(*th) >> 2) && tcptw->tw_ts_recent_stamp) { in tcp_timewait_state_process() 585 if (th->doff > (sizeof(struct tcphdr)>>2)) { in tcp_check_req()
|
/net/netfilter/ipvs/ |
D | ip_vs_ftp.c | 70 if ((th->doff << 2) < sizeof(struct tcphdr)) in ip_vs_ftp_data_ptr() 73 return (char *)th + (th->doff << 2); in ip_vs_ftp_data_ptr()
|
/net/core/ |
D | flow_dissector.c | 1642 const u8 *doff; in __skb_get_poff() local 1645 doff = __skb_header_pointer(skb, poff + 12, sizeof(_doff), in __skb_get_poff() 1647 if (!doff) in __skb_get_poff() 1650 poff += max_t(u32, sizeof(struct tcphdr), (*doff & 0xF0) >> 2); in __skb_get_poff()
|