/net/ipv4/ |
D | ip_input.c | 151 u8 protocol = ip_hdr(skb)->protocol; in ip_call_ra_chain() 165 if (ip_is_fragment(ip_hdr(skb))) { in ip_call_ra_chain() 231 ip_protocol_deliver_rcu(net, skb, ip_hdr(skb)->protocol); in ip_local_deliver_finish() 247 if (ip_is_fragment(ip_hdr(skb))) { in ip_local_deliver() 274 iph = ip_hdr(skb); in ip_rcv_options() 310 const struct iphdr *iph = ip_hdr(skb); in ip_rcv_finish_core() 324 iph = ip_hdr(skb); in ip_rcv_finish_core() 334 iph = ip_hdr(skb); in ip_rcv_finish_core() 449 iph = ip_hdr(skb); in ip_rcv_core() 475 iph = ip_hdr(skb); in ip_rcv_core() [all …]
|
D | xfrm4_input.c | 36 const struct iphdr *iph = ip_hdr(skb); in xfrm4_rcv_encap_finish() 55 struct iphdr *iph = ip_hdr(skb); in xfrm4_transport_finish() 150 iph = ip_hdr(skb); in xfrm4_udp_encap_rcv() 175 return xfrm4_rcv_spi(skb, ip_hdr(skb)->protocol, 0); in xfrm4_rcv()
|
D | tcp_ipv4.c | 95 return secure_tcp_seq(ip_hdr(skb)->daddr, in tcp_v4_init_seq() 96 ip_hdr(skb)->saddr, in tcp_v4_init_seq() 103 return secure_tcp_ts_off(net, ip_hdr(skb)->daddr, ip_hdr(skb)->saddr); in tcp_v4_init_ts_off() 709 &ip_hdr(skb)->saddr, AF_INET); in tcp_v4_send_reset() 719 ip_hdr(skb)->saddr, in tcp_v4_send_reset() 720 th->source, ip_hdr(skb)->daddr, in tcp_v4_send_reset() 728 &ip_hdr(skb)->saddr, AF_INET); in tcp_v4_send_reset() 749 key, ip_hdr(skb)->saddr, in tcp_v4_send_reset() 750 ip_hdr(skb)->daddr, &rep.th); in tcp_v4_send_reset() 753 arg.csum = csum_tcpudp_nofold(ip_hdr(skb)->daddr, in tcp_v4_send_reset() [all …]
|
D | xfrm4_output.c | 24 if (!(ip_hdr(skb)->frag_off & htons(IP_DF)) || skb->ignore_df) in xfrm4_tunnel_check_size() 52 XFRM_MODE_SKB_CB(skb)->protocol = ip_hdr(skb)->protocol; in xfrm4_extract_output() 102 hdr = skb->encapsulation ? inner_ip_hdr(skb) : ip_hdr(skb); in xfrm4_local_error()
|
D | syncookies.c | 174 const struct iphdr *iph = ip_hdr(skb); in cookie_v4_init_sequence() 301 mss = __cookie_v4_check(ip_hdr(skb), th, cookie); in cookie_v4_check() 315 ip_hdr(skb)->daddr, in cookie_v4_check() 316 ip_hdr(skb)->saddr); in cookie_v4_check() 338 sk_rcv_saddr_set(req_to_sk(req), ip_hdr(skb)->daddr); in cookie_v4_check() 339 sk_daddr_set(req_to_sk(req), ip_hdr(skb)->saddr); in cookie_v4_check()
|
D | udp_offload.c | 198 iph = ip_hdr(seg); in __udpv4_gso_segment_csum() 222 iph = ip_hdr(seg); in __udpv4_gso_segment_list_csum() 226 (ip_hdr(seg)->daddr == ip_hdr(seg->next)->daddr) && in __udpv4_gso_segment_list_csum() 227 (ip_hdr(seg)->saddr == ip_hdr(seg->next)->saddr)) in __udpv4_gso_segment_list_csum() 232 iph2 = ip_hdr(seg); in __udpv4_gso_segment_list_csum() 399 iph = ip_hdr(skb); in udp4_ufo_fragment() 654 const struct iphdr *iph = ip_hdr(skb); in udp4_gro_complete()
|
D | ip_forward.c | 48 if (unlikely((ip_hdr(skb)->frag_off & htons(IP_DF)) == 0)) in ip_exceeds_mtu() 118 if (ip_hdr(skb)->ttl <= 1) in ip_forward() 141 iph = ip_hdr(skb); in ip_forward()
|
D | esp4_offload.c | 55 (xfrm_address_t *)&ip_hdr(skb)->daddr, in esp4_gro_receive() 95 struct iphdr *iph = ip_hdr(skb); in esp4_gso_encap() 274 ip_hdr(skb)->tot_len = htons(skb->len); in esp_xmit() 275 ip_send_check(ip_hdr(skb)); in esp_xmit()
|
D | xfrm4_tunnel.c | 24 return ip_hdr(skb)->protocol; in ipip_xfrm_rcv() 56 return xfrm4_rcv_spi(skb, IPPROTO_IPIP, ip_hdr(skb)->saddr); in xfrm_tunnel_rcv()
|
D | ip_fragment.c | 176 iph = ip_hdr(head); in ip_expire() 293 ecn = ip4_frag_ecn(ip_hdr(skb)->tos); in ip_frag_queue() 294 offset = ntohs(ip_hdr(skb)->frag_off); in ip_frag_queue() 363 if (ip_hdr(skb)->frag_off & htons(IP_DF) && in ip_frag_queue() 436 iph = ip_hdr(skb); in ip_frag_reasm() 485 qp = ip_find(net, ip_hdr(skb), user, vif); in ip_defrag()
|
D | icmp.c | 434 inet->tos = ip_hdr(skb)->tos; in icmp_reply() 436 daddr = ipc.addr = ip_hdr(skb)->saddr; in icmp_reply() 449 fl4.flowi4_tos = RT_TOS(ip_hdr(skb)->tos); in icmp_reply() 623 iph = ip_hdr(skb_in); in __icmp_send() 811 orig_ip = ip_hdr(skb_in)->saddr; in icmp_ndo_send() 812 ip_hdr(skb_in)->saddr = ct->tuplehash[0].tuple.src.u3.ip; in icmp_ndo_send() 814 ip_hdr(skb_in)->saddr = orig_ip; in icmp_ndo_send() 950 &ip_hdr(skb)->saddr, in icmp_unreach()
|
D | route.c | 450 pkey = skb ? ip_hdr(skb)->daddr : *((__be32 *) daddr); in ipv4_neigh_lookup() 557 const struct iphdr *iph = ip_hdr(skb); in build_skb_flow_key() 761 __be32 old_gw = ip_hdr(skb)->saddr; in __ip_do_redirect() 913 peer = inet_getpeer_v4(net->ipv4.peers, ip_hdr(skb)->saddr, vif, 1); in ip_rt_send_redirect() 916 rt_nexthop(rt, ip_hdr(skb)->daddr)); in ip_rt_send_redirect() 943 __be32 gw = rt_nexthop(rt, ip_hdr(skb)->daddr); in ip_rt_send_redirect() 952 &ip_hdr(skb)->saddr, inet_iif(skb), in ip_rt_send_redirect() 953 &ip_hdr(skb)->daddr, &gw); in ip_rt_send_redirect() 1013 peer = inet_getpeer_v4(net->ipv4.peers, ip_hdr(skb)->saddr, in ip_error() 1232 ip_hdr(skb)->version != 4 || ip_hdr(skb)->ihl < 5) in ipv4_send_dest_unreach() [all …]
|
/net/ipv4/netfilter/ |
D | nf_reject_ipv4.c | 21 if (ip_hdr(oldskb)->frag_off & htons(IP_OFFSET)) in nf_reject_ip_tcphdr_get() 24 if (ip_hdr(oldskb)->protocol != IPPROTO_TCP) in nf_reject_ip_tcphdr_get() 48 struct iphdr *niph, *oiph = ip_hdr(oldskb); in nf_reject_iphdr_put() 72 struct iphdr *niph = ip_hdr(nskb); in nf_reject_ip_tcphdr_put() 133 niph = ip_hdr(nskb); in nf_send_reset() 172 struct iphdr *iph = ip_hdr(skb_in); in nf_send_unreach()
|
D | ipt_ECN.c | 28 struct iphdr *iph = ip_hdr(skb); in set_ect_ip() 34 iph = ip_hdr(skb); in set_ect_ip() 63 tcph = (void *)ip_hdr(skb) + ip_hdrlen(skb); in set_ect_tcp() 86 ip_hdr(skb)->protocol == IPPROTO_TCP) in ecn_tg()
|
D | nf_dup_ipv4.c | 26 const struct iphdr *iph = ip_hdr(skb); in nf_dup_ipv4_route() 80 iph = ip_hdr(skb); in nf_dup_ipv4()
|
D | iptable_mangle.c | 51 iph = ip_hdr(skb); in ipt_mangle_out() 59 iph = ip_hdr(skb); in ipt_mangle_out()
|
/net/netfilter/ |
D | xt_DSCP.c | 31 u_int8_t dscp = ipv4_get_dsfield(ip_hdr(skb)) >> XT_DSCP_SHIFT; in dscp_tg() 37 ipv4_change_dsfield(ip_hdr(skb), in dscp_tg() 75 struct iphdr *iph = ip_hdr(skb); in tos_tg() 84 iph = ip_hdr(skb); in tos_tg()
|
D | nf_flow_table_ip.c | 145 struct iphdr *iph = ip_hdr(skb); in nf_flow_nat_ip() 174 iph = ip_hdr(skb); in nf_flow_tuple_ip() 192 iph = ip_hdr(skb); in nf_flow_tuple_ip() 272 thoff = ip_hdr(skb)->ihl * 4; in nf_flow_offload_ip_hook() 273 if (nf_flow_state_check(flow, ip_hdr(skb)->protocol, skb, thoff)) in nf_flow_offload_ip_hook() 285 iph = ip_hdr(skb); in nf_flow_offload_ip_hook()
|
D | utils.c | 13 const struct iphdr *iph = ip_hdr(skb); in nf_ip_checksum() 47 const struct iphdr *iph = ip_hdr(skb); in nf_ip_checksum_partial() 188 const struct iphdr *iph = ip_hdr(skb); in nf_ip_reroute()
|
D | xt_dscp.c | 28 u_int8_t dscp = ipv4_get_dsfield(ip_hdr(skb)) >> XT_DSCP_SHIFT; in dscp_mt() 57 return ((ip_hdr(skb)->tos & info->tos_mask) == in tos_mt()
|
/net/dccp/ |
D | ipv4.c | 383 return secure_dccp_sequence_number(ip_hdr(skb)->daddr, in dccp_v4_init_sequence() 384 ip_hdr(skb)->saddr, in dccp_v4_init_sequence() 420 newinet->mc_ttl = ip_hdr(skb)->ttl; in dccp_v4_request_recv_sock() 458 const struct iphdr *iph = ip_hdr(skb); in dccp_v4_route_skb() 534 rxiph = ip_hdr(rxskb); in dccp_v4_ctl_send_reset() 615 sk_rcv_saddr_set(req_to_sk(req), ip_hdr(skb)->daddr); in dccp_v4_conn_request() 616 sk_daddr_set(req_to_sk(req), ip_hdr(skb)->saddr); in dccp_v4_conn_request() 785 iph = ip_hdr(skb); in dccp_v4_rcv()
|
/net/batman-adv/ |
D | main.c | 406 struct iphdr ip_hdr_tmp, *ip_hdr; in batadv_skb_set_priority() local 430 ip_hdr = skb_header_pointer(skb, offset + sizeof(*ethhdr), in batadv_skb_set_priority() 431 sizeof(*ip_hdr), &ip_hdr_tmp); in batadv_skb_set_priority() 432 if (!ip_hdr) in batadv_skb_set_priority() 434 prio = (ipv4_get_dsfield(ip_hdr) & 0xfc) >> 5; in batadv_skb_set_priority()
|
/net/bridge/netfilter/ |
D | nft_reject_bridge.c | 50 iph = ip_hdr(skb); in nft_bridge_iphdr_validate() 119 if (ip_hdr(oldskb)->frag_off & htons(IP_OFFSET)) in nft_reject_br_send_v4_unreach() 128 if (pskb_trim_rcsum(oldskb, ntohs(ip_hdr(oldskb)->tot_len))) in nft_reject_br_send_v4_unreach() 131 proto = ip_hdr(oldskb)->protocol; in nft_reject_br_send_v4_unreach()
|
D | nf_conntrack_bridge.c | 46 iph = ip_hdr(skb); in nf_br_ip_fragment() 153 if (!ip_is_fragment(ip_hdr(skb))) in nf_ct_br_defrag4() 209 iph = ip_hdr(skb); in nf_ct_br_ip_check() 258 len = ntohs(ip_hdr(skb)->tot_len); in nf_ct_bridge_pre()
|
/net/netfilter/ipvs/ |
D | ip_vs_xmit.c | 187 source_is_loopback = ipv4_is_loopback(ip_hdr(skb)->saddr); in crosses_local_route_boundary() 242 if (unlikely(ip_hdr(skb)->frag_off & htons(IP_DF) && in ensure_mtu_is_adequate() 248 &ip_hdr(skb)->saddr); in ensure_mtu_is_adequate() 287 if (ip_hdr(skb)->ttl <= 1) { in decrement_ttl() 299 ip_decrease_ttl(ip_hdr(skb)); in decrement_ttl() 713 struct iphdr *iph = ip_hdr(skb); in ip_vs_bypass_xmit() 834 ip_hdr(skb)->daddr = cp->daddr.ip; in ip_vs_nat_xmit() 835 ip_send_check(ip_hdr(skb)); in ip_vs_nat_xmit() 992 old_iph = ip_hdr(skb); in ip_vs_prepare_tunneled_skb() 1257 iph = ip_hdr(skb); in ip_vs_tunnel_xmit() [all …]
|