/net/bridge/ |
D | br_netfilter.c | 45 #define skb_origaddr(skb) (((struct bridge_skb_cb *) \ argument 46 (skb->nf_bridge->data))->daddr.ipv4) 47 #define store_orig_dstaddr(skb) (skb_origaddr(skb) = ip_hdr(skb)->daddr) argument 48 #define dnat_took_place(skb) (skb_origaddr(skb) != ip_hdr(skb)->daddr) argument 67 #define IS_IP(skb) \ argument 68 (!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_IP)) 70 #define IS_IPV6(skb) \ argument 71 (!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_IPV6)) 73 #define IS_ARP(skb) \ argument 74 (!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_ARP)) [all …]
|
D | br_forward.c | 25 struct sk_buff *skb, 27 struct sk_buff *skb)); 31 const struct sk_buff *skb) in should_deliver() argument 33 return ((p->flags & BR_HAIRPIN_MODE) || skb->dev != p->dev) && in should_deliver() 34 br_allowed_egress(p->br, nbp_get_vlan_info(p), skb) && in should_deliver() 38 int br_dev_queue_push_xmit(struct sk_buff *skb) in br_dev_queue_push_xmit() argument 41 if (nf_bridge_maybe_copy_header(skb) || in br_dev_queue_push_xmit() 42 !is_skb_forwardable(skb->dev, skb)) { in br_dev_queue_push_xmit() 43 kfree_skb(skb); in br_dev_queue_push_xmit() 45 skb_push(skb, ETH_HLEN); in br_dev_queue_push_xmit() [all …]
|
D | br_input.c | 27 static int br_pass_frame_up(struct sk_buff *skb) in br_pass_frame_up() argument 29 struct net_device *indev, *brdev = BR_INPUT_SKB_CB(skb)->brdev; in br_pass_frame_up() 36 brstats->rx_bytes += skb->len; in br_pass_frame_up() 45 !br_allowed_egress(br, pv, skb)) { in br_pass_frame_up() 46 kfree_skb(skb); in br_pass_frame_up() 50 indev = skb->dev; in br_pass_frame_up() 51 skb->dev = brdev; in br_pass_frame_up() 52 skb = br_handle_vlan(br, pv, skb); in br_pass_frame_up() 53 if (!skb) in br_pass_frame_up() 56 return NF_HOOK(NFPROTO_BRIDGE, NF_BR_LOCAL_IN, skb, indev, NULL, in br_pass_frame_up() [all …]
|
/net/openvswitch/ |
D | actions.c | 41 static int do_execute_actions(struct datapath *dp, struct sk_buff *skb, 46 struct sk_buff *skb; member 92 static struct deferred_action *add_deferred_actions(struct sk_buff *skb, in add_deferred_actions() argument 102 da->skb = skb; in add_deferred_actions() 110 static int make_writable(struct sk_buff *skb, int write_len) in make_writable() argument 112 if (!pskb_may_pull(skb, write_len)) in make_writable() 115 if (!skb_cloned(skb) || skb_clone_writable(skb, write_len)) in make_writable() 118 return pskb_expand_head(skb, 0, 0, GFP_ATOMIC); in make_writable() 122 static int __pop_vlan_tci(struct sk_buff *skb, __be16 *current_tci) in __pop_vlan_tci() argument 127 err = make_writable(skb, VLAN_ETH_HLEN); in __pop_vlan_tci() [all …]
|
/net/core/ |
D | skbuff.c | 95 static void skb_panic(struct sk_buff *skb, unsigned int sz, void *addr, in skb_panic() argument 99 msg, addr, skb->len, sz, skb->head, skb->data, in skb_panic() 100 (unsigned long)skb->tail, (unsigned long)skb->end, in skb_panic() 101 skb->dev ? skb->dev->name : "<NULL>"); in skb_panic() 105 static void skb_over_panic(struct sk_buff *skb, unsigned int sz, void *addr) in skb_over_panic() argument 107 skb_panic(skb, sz, addr, __func__); in skb_over_panic() 110 static void skb_under_panic(struct sk_buff *skb, unsigned int sz, void *addr) in skb_under_panic() argument 112 skb_panic(skb, sz, addr, __func__); in skb_under_panic() 160 struct sk_buff *skb; in __alloc_skb_head() local 163 skb = kmem_cache_alloc_node(skbuff_head_cache, in __alloc_skb_head() [all …]
|
/net/ipv6/ |
D | xfrm6_output.c | 23 int xfrm6_find_1stfragopt(struct xfrm_state *x, struct sk_buff *skb, in xfrm6_find_1stfragopt() argument 26 return ip6_find_1stfragopt(skb, prevhdr); in xfrm6_find_1stfragopt() 30 static int xfrm6_local_dontfrag(struct sk_buff *skb) in xfrm6_local_dontfrag() argument 33 struct sock *sk = skb->sk; in xfrm6_local_dontfrag() 47 static void xfrm6_local_rxpmtu(struct sk_buff *skb, u32 mtu) in xfrm6_local_rxpmtu() argument 50 struct sock *sk = skb->sk; in xfrm6_local_rxpmtu() 53 fl6.daddr = ipv6_hdr(skb)->daddr; in xfrm6_local_rxpmtu() 58 void xfrm6_local_error(struct sk_buff *skb, u32 mtu) in xfrm6_local_error() argument 62 struct sock *sk = skb->sk; in xfrm6_local_error() 64 hdr = skb->encapsulation ? inner_ipv6_hdr(skb) : ipv6_hdr(skb); in xfrm6_local_error() [all …]
|
D | exthdrs.c | 62 bool (*func)(struct sk_buff *skb, int offset); 71 static bool ip6_tlvopt_unknown(struct sk_buff *skb, int optoff) in ip6_tlvopt_unknown() argument 73 switch ((skb_network_header(skb)[optoff] & 0xC0) >> 6) { in ip6_tlvopt_unknown() 84 if (ipv6_addr_is_multicast(&ipv6_hdr(skb)->daddr)) in ip6_tlvopt_unknown() 87 icmpv6_param_prob(skb, ICMPV6_UNK_OPTION, optoff); in ip6_tlvopt_unknown() 91 kfree_skb(skb); in ip6_tlvopt_unknown() 97 static bool ip6_parse_tlv(const struct tlvtype_proc *procs, struct sk_buff *skb) in ip6_parse_tlv() argument 100 const unsigned char *nh = skb_network_header(skb); in ip6_parse_tlv() 101 int off = skb_network_header_len(skb); in ip6_parse_tlv() 102 int len = (skb_transport_header(skb)[1] + 1) << 3; in ip6_parse_tlv() [all …]
|
D | ip6_input.c | 50 int ip6_rcv_finish(struct sk_buff *skb) in ip6_rcv_finish() argument 52 if (sysctl_ip_early_demux && !skb_dst(skb) && skb->sk == NULL) { in ip6_rcv_finish() 55 ipprot = rcu_dereference(inet6_protos[ipv6_hdr(skb)->nexthdr]); in ip6_rcv_finish() 57 ipprot->early_demux(skb); in ip6_rcv_finish() 59 if (!skb_dst(skb)) in ip6_rcv_finish() 60 ip6_route_input(skb); in ip6_rcv_finish() 62 return dst_input(skb); in ip6_rcv_finish() 65 int ipv6_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device… in ipv6_rcv() argument 70 struct net *net = dev_net(skb->dev); in ipv6_rcv() 72 if (skb->pkt_type == PACKET_OTHERHOST) { in ipv6_rcv() [all …]
|
D | ip6_output.c | 59 static int ip6_finish_output2(struct sk_buff *skb) in ip6_finish_output2() argument 61 struct dst_entry *dst = skb_dst(skb); in ip6_finish_output2() 67 skb->protocol = htons(ETH_P_IPV6); in ip6_finish_output2() 68 skb->dev = dev; in ip6_finish_output2() 70 if (ipv6_addr_is_multicast(&ipv6_hdr(skb)->daddr)) { in ip6_finish_output2() 71 struct inet6_dev *idev = ip6_dst_idev(skb_dst(skb)); in ip6_finish_output2() 73 if (!(dev->flags & IFF_LOOPBACK) && sk_mc_loop(skb->sk) && in ip6_finish_output2() 74 ((mroute6_socket(dev_net(dev), skb) && in ip6_finish_output2() 75 !(IP6CB(skb)->flags & IP6SKB_FORWARDED)) || in ip6_finish_output2() 76 ipv6_chk_mcast_addr(dev, &ipv6_hdr(skb)->daddr, in ip6_finish_output2() [all …]
|
D | ip6_offload.c | 22 static int ipv6_gso_pull_exthdrs(struct sk_buff *skb, int proto) in ipv6_gso_pull_exthdrs() argument 40 if (unlikely(!pskb_may_pull(skb, 8))) in ipv6_gso_pull_exthdrs() 43 opth = (void *)skb->data; in ipv6_gso_pull_exthdrs() 46 if (unlikely(!pskb_may_pull(skb, len))) in ipv6_gso_pull_exthdrs() 49 opth = (void *)skb->data; in ipv6_gso_pull_exthdrs() 51 __skb_pull(skb, len); in ipv6_gso_pull_exthdrs() 57 static struct sk_buff *ipv6_gso_segment(struct sk_buff *skb, in ipv6_gso_segment() argument 70 if (unlikely(skb_shinfo(skb)->gso_type & in ipv6_gso_segment() 86 skb_reset_network_header(skb); in ipv6_gso_segment() 87 nhoff = skb_network_header(skb) - skb_mac_header(skb); in ipv6_gso_segment() [all …]
|
/net/lapb/ |
D | lapb_subr.c | 53 struct sk_buff *skb; in lapb_frames_acked() local 63 skb = skb_dequeue(&lapb->ack_queue); in lapb_frames_acked() 64 kfree_skb(skb); in lapb_frames_acked() 71 struct sk_buff *skb, *skb_prev = NULL; in lapb_requeue_frames() local 78 while ((skb = skb_dequeue(&lapb->ack_queue)) != NULL) { in lapb_requeue_frames() 80 skb_queue_head(&lapb->write_queue, skb); in lapb_requeue_frames() 82 skb_append(skb_prev, skb, &lapb->write_queue); in lapb_requeue_frames() 83 skb_prev = skb; in lapb_requeue_frames() 111 int lapb_decode(struct lapb_cb *lapb, struct sk_buff *skb, in lapb_decode() argument 118 skb->data[0], skb->data[1], skb->data[2]); in lapb_decode() [all …]
|
/net/ipv4/ |
D | tcp_offload.c | 17 static void tcp_gso_tstamp(struct sk_buff *skb, unsigned int ts_seq, in tcp_gso_tstamp() argument 20 while (skb) { in tcp_gso_tstamp() 22 skb_shinfo(skb)->tx_flags |= SKBTX_SW_TSTAMP; in tcp_gso_tstamp() 23 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp() 27 skb = skb->next; in tcp_gso_tstamp() 32 struct sk_buff *tcp4_gso_segment(struct sk_buff *skb, in tcp4_gso_segment() argument 35 if (!pskb_may_pull(skb, sizeof(struct tcphdr))) in tcp4_gso_segment() 38 if (unlikely(skb->ip_summed != CHECKSUM_PARTIAL)) { in tcp4_gso_segment() 39 const struct iphdr *iph = ip_hdr(skb); in tcp4_gso_segment() 40 struct tcphdr *th = tcp_hdr(skb); in tcp4_gso_segment() [all …]
|
D | udp_offload.c | 28 static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb, in __skb_udp_tunnel_segment() argument 30 struct sk_buff *(*gso_inner_segment)(struct sk_buff *skb, in __skb_udp_tunnel_segment() argument 35 u16 mac_offset = skb->mac_header; in __skb_udp_tunnel_segment() 36 int mac_len = skb->mac_len; in __skb_udp_tunnel_segment() 37 int tnl_hlen = skb_inner_mac_header(skb) - skb_transport_header(skb); in __skb_udp_tunnel_segment() 38 __be16 protocol = skb->protocol; in __skb_udp_tunnel_segment() 44 oldlen = (u16)~skb->len; in __skb_udp_tunnel_segment() 46 if (unlikely(!pskb_may_pull(skb, tnl_hlen))) in __skb_udp_tunnel_segment() 49 skb->encapsulation = 0; in __skb_udp_tunnel_segment() 50 __skb_pull(skb, tnl_hlen); in __skb_udp_tunnel_segment() [all …]
|
D | ip_output.c | 94 int __ip_local_out(struct sk_buff *skb) in __ip_local_out() argument 96 struct iphdr *iph = ip_hdr(skb); in __ip_local_out() 98 iph->tot_len = htons(skb->len); in __ip_local_out() 100 return nf_hook(NFPROTO_IPV4, NF_INET_LOCAL_OUT, skb, NULL, in __ip_local_out() 101 skb_dst(skb)->dev, dst_output); in __ip_local_out() 104 int ip_local_out_sk(struct sock *sk, struct sk_buff *skb) in ip_local_out_sk() argument 108 err = __ip_local_out(skb); in ip_local_out_sk() 110 err = dst_output_sk(sk, skb); in ip_local_out_sk() 129 int ip_build_and_send_pkt(struct sk_buff *skb, struct sock *sk, in ip_build_and_send_pkt() argument 133 struct rtable *rt = skb_rtable(skb); in ip_build_and_send_pkt() [all …]
|
D | ip_input.c | 153 bool ip_call_ra_chain(struct sk_buff *skb) in ip_call_ra_chain() argument 156 u8 protocol = ip_hdr(skb)->protocol; in ip_call_ra_chain() 158 struct net_device *dev = skb->dev; in ip_call_ra_chain() 170 if (ip_is_fragment(ip_hdr(skb))) { in ip_call_ra_chain() 171 if (ip_defrag(skb, IP_DEFRAG_CALL_RA_CHAIN)) in ip_call_ra_chain() 175 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_call_ra_chain() 184 raw_rcv(last, skb); in ip_call_ra_chain() 190 static int ip_local_deliver_finish(struct sk_buff *skb) in ip_local_deliver_finish() argument 192 struct net *net = dev_net(skb->dev); in ip_local_deliver_finish() 194 __skb_pull(skb, skb_network_header_len(skb)); in ip_local_deliver_finish() [all …]
|
D | gre_offload.c | 18 static struct sk_buff *gre_gso_segment(struct sk_buff *skb, in gre_gso_segment() argument 25 u16 mac_offset = skb->mac_header; in gre_gso_segment() 26 int mac_len = skb->mac_len; in gre_gso_segment() 27 __be16 protocol = skb->protocol; in gre_gso_segment() 31 if (unlikely(skb_shinfo(skb)->gso_type & in gre_gso_segment() 43 if (!skb->encapsulation) in gre_gso_segment() 46 if (unlikely(!pskb_may_pull(skb, sizeof(*greh)))) in gre_gso_segment() 49 greh = (struct gre_base_hdr *)skb_transport_header(skb); in gre_gso_segment() 51 ghl = skb_inner_mac_header(skb) - skb_transport_header(skb); in gre_gso_segment() 57 skb->encap_hdr_csum = 1; in gre_gso_segment() [all …]
|
D | ip_forward.c | 42 static bool ip_may_fragment(const struct sk_buff *skb) in ip_may_fragment() argument 44 return unlikely((ip_hdr(skb)->frag_off & htons(IP_DF)) == 0) || in ip_may_fragment() 45 skb->ignore_df; in ip_may_fragment() 48 static bool ip_exceeds_mtu(const struct sk_buff *skb, unsigned int mtu) in ip_exceeds_mtu() argument 50 if (skb->len <= mtu) in ip_exceeds_mtu() 53 if (skb_is_gso(skb) && skb_gso_network_seglen(skb) <= mtu) in ip_exceeds_mtu() 60 static int ip_forward_finish(struct sk_buff *skb) in ip_forward_finish() argument 62 struct ip_options *opt = &(IPCB(skb)->opt); in ip_forward_finish() 64 IP_INC_STATS_BH(dev_net(skb_dst(skb)->dev), IPSTATS_MIB_OUTFORWDATAGRAMS); in ip_forward_finish() 65 IP_ADD_STATS_BH(dev_net(skb_dst(skb)->dev), IPSTATS_MIB_OUTOCTETS, skb->len); in ip_forward_finish() [all …]
|
D | xfrm4_output.c | 21 static int xfrm4_tunnel_check_size(struct sk_buff *skb) in xfrm4_tunnel_check_size() argument 25 if (IPCB(skb)->flags & IPSKB_XFRM_TUNNEL_SIZE) in xfrm4_tunnel_check_size() 28 if (!(ip_hdr(skb)->frag_off & htons(IP_DF)) || skb->ignore_df) in xfrm4_tunnel_check_size() 31 mtu = dst_mtu(skb_dst(skb)); in xfrm4_tunnel_check_size() 32 if (skb->len > mtu) { in xfrm4_tunnel_check_size() 33 if (skb->sk) in xfrm4_tunnel_check_size() 34 xfrm_local_error(skb, mtu); in xfrm4_tunnel_check_size() 36 icmp_send(skb, ICMP_DEST_UNREACH, in xfrm4_tunnel_check_size() 44 int xfrm4_extract_output(struct xfrm_state *x, struct sk_buff *skb) in xfrm4_extract_output() argument 48 err = xfrm4_tunnel_check_size(skb); in xfrm4_extract_output() [all …]
|
D | netfilter.c | 20 int ip_route_me_harder(struct sk_buff *skb, unsigned int addr_type) in ip_route_me_harder() argument 22 struct net *net = dev_net(skb_dst(skb)->dev); in ip_route_me_harder() 23 const struct iphdr *iph = ip_hdr(skb); in ip_route_me_harder() 27 __u8 flags = skb->sk ? inet_sk_flowi_flags(skb->sk) : 0; in ip_route_me_harder() 43 fl4.flowi4_oif = skb->sk ? skb->sk->sk_bound_dev_if : 0; in ip_route_me_harder() 44 fl4.flowi4_mark = skb->mark; in ip_route_me_harder() 51 skb_dst_drop(skb); in ip_route_me_harder() 52 skb_dst_set(skb, &rt->dst); in ip_route_me_harder() 54 if (skb_dst(skb)->error) in ip_route_me_harder() 55 return skb_dst(skb)->error; in ip_route_me_harder() [all …]
|
/net/ax25/ |
D | ax25_in.c | 38 static int ax25_rx_fragment(ax25_cb *ax25, struct sk_buff *skb) in ax25_rx_fragment() argument 43 if (!(*skb->data & AX25_SEG_FIRST)) { in ax25_rx_fragment() 44 if ((ax25->fragno - 1) == (*skb->data & AX25_SEG_REM)) { in ax25_rx_fragment() 46 ax25->fragno = *skb->data & AX25_SEG_REM; in ax25_rx_fragment() 47 skb_pull(skb, 1); /* skip fragno */ in ax25_rx_fragment() 48 ax25->fraglen += skb->len; in ax25_rx_fragment() 49 skb_queue_tail(&ax25->frag_queue, skb); in ax25_rx_fragment() 86 if (*skb->data & AX25_SEG_FIRST) { in ax25_rx_fragment() 88 ax25->fragno = *skb->data & AX25_SEG_REM; in ax25_rx_fragment() 89 skb_pull(skb, 1); /* skip fragno */ in ax25_rx_fragment() [all …]
|
/net/netfilter/ipvs/ |
D | ip_vs_xmit.c | 108 __mtu_check_toobig_v6(const struct sk_buff *skb, u32 mtu) in __mtu_check_toobig_v6() argument 110 if (IP6CB(skb)->frag_max_size) { in __mtu_check_toobig_v6() 114 if (IP6CB(skb)->frag_max_size > mtu) in __mtu_check_toobig_v6() 117 else if (skb->len > mtu && !skb_is_gso(skb)) { in __mtu_check_toobig_v6() 166 static inline bool crosses_local_route_boundary(int skb_af, struct sk_buff *skb, in crosses_local_route_boundary() argument 178 int addr_type = ipv6_addr_type(&ipv6_hdr(skb)->saddr); in crosses_local_route_boundary() 181 (!skb->dev || skb->dev->flags & IFF_LOOPBACK) && in crosses_local_route_boundary() 184 (struct rt6_info *)skb_dst(skb)); in crosses_local_route_boundary() 188 source_is_loopback = ipv4_is_loopback(ip_hdr(skb)->saddr); in crosses_local_route_boundary() 189 old_rt_is_local = skb_rtable(skb)->rt_flags & RTCF_LOCAL; in crosses_local_route_boundary() [all …]
|
/net/llc/ |
D | llc_c_ev.c | 78 struct sk_buff *skb; in llc_util_nr_inside_tx_window() local 88 skb = skb_peek(&llc->pdu_unack_q); in llc_util_nr_inside_tx_window() 89 pdu = llc_pdu_sn_hdr(skb); in llc_util_nr_inside_tx_window() 91 skb = skb_peek_tail(&llc->pdu_unack_q); in llc_util_nr_inside_tx_window() 92 pdu = llc_pdu_sn_hdr(skb); in llc_util_nr_inside_tx_window() 99 int llc_conn_ev_conn_req(struct sock *sk, struct sk_buff *skb) in llc_conn_ev_conn_req() argument 101 const struct llc_conn_state_ev *ev = llc_conn_ev(skb); in llc_conn_ev_conn_req() 107 int llc_conn_ev_data_req(struct sock *sk, struct sk_buff *skb) in llc_conn_ev_data_req() argument 109 const struct llc_conn_state_ev *ev = llc_conn_ev(skb); in llc_conn_ev_data_req() 115 int llc_conn_ev_disc_req(struct sock *sk, struct sk_buff *skb) in llc_conn_ev_disc_req() argument [all …]
|
/net/decnet/ |
D | dn_nsp_in.c | 81 static void dn_log_martian(struct sk_buff *skb, const char *msg) in dn_log_martian() argument 84 char *devname = skb->dev ? skb->dev->name : "???"; in dn_log_martian() 85 struct dn_skb_cb *cb = DN_SKB_CB(skb); in dn_log_martian() 100 static void dn_ack(struct sock *sk, struct sk_buff *skb, unsigned short ack) in dn_ack() argument 110 wakeup |= dn_nsp_check_xmit_queue(sk, skb, in dn_ack() 120 wakeup |= dn_nsp_check_xmit_queue(sk, skb, in dn_ack() 136 static int dn_process_ack(struct sock *sk, struct sk_buff *skb, int oth) in dn_process_ack() argument 138 __le16 *ptr = (__le16 *)skb->data; in dn_process_ack() 142 if (skb->len < 2) in dn_process_ack() 146 skb_pull(skb, 2); in dn_process_ack() [all …]
|
/net/x25/ |
D | x25_dev.c | 31 static int x25_receive_data(struct sk_buff *skb, struct x25_neigh *nb) in x25_receive_data() argument 37 if (!pskb_may_pull(skb, X25_STD_MIN_LEN)) in x25_receive_data() 40 frametype = skb->data[2]; in x25_receive_data() 41 lci = ((skb->data[0] << 8) & 0xF00) + ((skb->data[1] << 0) & 0x0FF); in x25_receive_data() 48 x25_link_control(skb, nb, frametype); in x25_receive_data() 58 skb_reset_transport_header(skb); in x25_receive_data() 61 queued = x25_process_rx_frame(sk, skb); in x25_receive_data() 63 queued = !sk_add_backlog(sk, skb, sk->sk_rcvbuf); in x25_receive_data() 74 return x25_rx_call_request(skb, nb, lci); in x25_receive_data() 81 if (x25_forward_data(lci, nb, skb)) { in x25_receive_data() [all …]
|
/net/dsa/ |
D | tag_edsa.c | 19 static netdev_tx_t edsa_xmit(struct sk_buff *skb, struct net_device *dev) in edsa_xmit() argument 25 dev->stats.tx_bytes += skb->len; in edsa_xmit() 33 if (skb->protocol == htons(ETH_P_8021Q)) { in edsa_xmit() 34 if (skb_cow_head(skb, DSA_HLEN) < 0) in edsa_xmit() 36 skb_push(skb, DSA_HLEN); in edsa_xmit() 38 memmove(skb->data, skb->data + DSA_HLEN, 2 * ETH_ALEN); in edsa_xmit() 43 edsa_header = skb->data + 2 * ETH_ALEN; in edsa_xmit() 59 if (skb_cow_head(skb, EDSA_HLEN) < 0) in edsa_xmit() 61 skb_push(skb, EDSA_HLEN); in edsa_xmit() 63 memmove(skb->data, skb->data + EDSA_HLEN, 2 * ETH_ALEN); in edsa_xmit() [all …]
|