/net/mctp/test/ |
D | route-test.c | 162 struct sk_buff *skb2; in mctp_test_fragment() local 169 skb2 = skb_dequeue(&rt->pkts); in mctp_test_fragment() 171 if (!skb2) in mctp_test_fragment() 174 hdr2 = mctp_hdr(skb2); in mctp_test_fragment() 200 KUNIT_EXPECT_EQ(test, skb2->len, mtu); in mctp_test_fragment() 202 KUNIT_EXPECT_LE(test, skb2->len, mtu); in mctp_test_fragment() 204 kfree_skb(skb2); in mctp_test_fragment() 334 struct sk_buff *skb, *skb2; in mctp_test_route_input_sk() local 355 skb2 = skb_recv_datagram(sock->sk, MSG_DONTWAIT, &rc); in mctp_test_route_input_sk() 356 KUNIT_EXPECT_NOT_ERR_OR_NULL(test, skb2); in mctp_test_route_input_sk() [all …]
|
/net/netfilter/ipset/ |
D | ip_set_core.c | 1759 struct sk_buff *skb2; in call_ad() local 1768 skb2 = nlmsg_new(payload, GFP_KERNEL); in call_ad() 1769 if (!skb2) in call_ad() 1771 rep = nlmsg_put(skb2, NETLINK_CB(skb).portid, in call_ad() 1785 nlmsg_free(skb2); in call_ad() 1792 nfnetlink_unicast(skb2, net, NETLINK_CB(skb).portid); in call_ad() 1910 struct sk_buff *skb2; in ip_set_header() local 1921 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in ip_set_header() 1922 if (!skb2) in ip_set_header() 1925 nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, info->nlh->nlmsg_seq, 0, in ip_set_header() [all …]
|
/net/xfrm/ |
D | xfrm_device.c | 119 struct sk_buff *skb2, *nskb, *pskb = NULL; in validate_xmit_xfrm() local 199 skb_list_walk_safe(skb, skb2, nskb) { in validate_xmit_xfrm() 201 skb_mark_not_on_list(skb2); in validate_xmit_xfrm() 203 xo = xfrm_offload(skb2); in validate_xmit_xfrm() 206 xfrm_outer_mode_prep(x, skb2); in validate_xmit_xfrm() 208 err = x->type_offload->xmit(x, skb2, esp_features); in validate_xmit_xfrm() 210 skb2->next = nskb; in validate_xmit_xfrm() 213 skb2->next = nskb; in validate_xmit_xfrm() 214 kfree_skb_list(skb2); in validate_xmit_xfrm() 217 if (skb == skb2) in validate_xmit_xfrm() [all …]
|
/net/mac802154/ |
D | rx.c | 313 struct sk_buff *skb2; in __ieee802154_rx_handle_packet() local 335 skb2 = skb_clone(skb, GFP_ATOMIC); in __ieee802154_rx_handle_packet() 336 if (skb2) { in __ieee802154_rx_handle_packet() 337 skb2->dev = sdata->dev; in __ieee802154_rx_handle_packet() 338 ieee802154_subif_frame(sdata, skb2, &hdr); in __ieee802154_rx_handle_packet() 346 struct sk_buff *skb2; in ieee802154_monitors_rx() local 361 skb2 = skb_clone(skb, GFP_ATOMIC); in ieee802154_monitors_rx() 362 if (skb2) { in ieee802154_monitors_rx() 363 skb2->dev = sdata->dev; in ieee802154_monitors_rx() 364 ieee802154_deliver_skb(skb2); in ieee802154_monitors_rx()
|
/net/netfilter/ |
D | nfnetlink_cttimeout.c | 284 struct sk_buff *skb2; in cttimeout_get_timeout() local 289 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in cttimeout_get_timeout() 290 if (skb2 == NULL) { in cttimeout_get_timeout() 295 ret = ctnl_timeout_fill_info(skb2, NETLINK_CB(skb).portid, in cttimeout_get_timeout() 300 kfree_skb(skb2); in cttimeout_get_timeout() 304 ret = nfnetlink_unicast(skb2, info->net, NETLINK_CB(skb).portid); in cttimeout_get_timeout() 439 struct sk_buff *skb2; in cttimeout_default_get() local 494 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in cttimeout_default_get() 495 if (!skb2) in cttimeout_default_get() 498 ret = cttimeout_default_fill_info(info->net, skb2, in cttimeout_default_get() [all …]
|
D | nfnetlink_acct.c | 298 struct sk_buff *skb2; in nfnl_acct_get() local 303 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in nfnl_acct_get() 304 if (skb2 == NULL) { in nfnl_acct_get() 309 ret = nfnl_acct_fill_info(skb2, NETLINK_CB(skb).portid, in nfnl_acct_get() 314 kfree_skb(skb2); in nfnl_acct_get() 318 ret = nfnetlink_unicast(skb2, info->net, NETLINK_CB(skb).portid); in nfnl_acct_get()
|
D | nfnetlink_cthelper.c | 621 struct sk_buff *skb2; in nfnl_cthelper_get() local 659 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in nfnl_cthelper_get() 660 if (skb2 == NULL) { in nfnl_cthelper_get() 665 ret = nfnl_cthelper_fill_info(skb2, NETLINK_CB(skb).portid, in nfnl_cthelper_get() 670 kfree_skb(skb2); in nfnl_cthelper_get() 674 ret = nfnetlink_unicast(skb2, info->net, NETLINK_CB(skb).portid); in nfnl_cthelper_get()
|
D | nft_compat.c | 687 struct sk_buff *skb2; in nfnl_compat_get_rcu() local 727 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in nfnl_compat_get_rcu() 728 if (skb2 == NULL) { in nfnl_compat_get_rcu() 734 if (nfnl_compat_fill_info(skb2, NETLINK_CB(skb).portid, in nfnl_compat_get_rcu() 739 kfree_skb(skb2); in nfnl_compat_get_rcu() 743 ret = nfnetlink_unicast(skb2, info->net, NETLINK_CB(skb).portid); in nfnl_compat_get_rcu()
|
D | nf_tables_api.c | 1112 struct sk_buff *skb2; in nf_tables_gettable() local 1130 skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); in nf_tables_gettable() 1131 if (!skb2) in nf_tables_gettable() 1134 err = nf_tables_fill_table_info(skb2, net, NETLINK_CB(skb).portid, in nf_tables_gettable() 1140 return nfnetlink_unicast(skb2, net, NETLINK_CB(skb).portid); in nf_tables_gettable() 1143 kfree_skb(skb2); in nf_tables_gettable() 1957 struct sk_buff *skb2; in nf_tables_getchain() local 1981 skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); in nf_tables_getchain() 1982 if (!skb2) in nf_tables_getchain() 1985 err = nf_tables_fill_chain_info(skb2, net, NETLINK_CB(skb).portid, in nf_tables_getchain() [all …]
|
D | nf_conntrack_netlink.c | 1650 struct sk_buff *skb2; in ctnetlink_get_conntrack() local 1687 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in ctnetlink_get_conntrack() 1688 if (!skb2) { in ctnetlink_get_conntrack() 1693 err = ctnetlink_fill_info(skb2, NETLINK_CB(skb).portid, in ctnetlink_get_conntrack() 1699 kfree_skb(skb2); in ctnetlink_get_conntrack() 1703 return nfnetlink_unicast(skb2, info->net, NETLINK_CB(skb).portid); in ctnetlink_get_conntrack() 2598 struct sk_buff *skb2; in ctnetlink_stat_ct() local 2601 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in ctnetlink_stat_ct() 2602 if (skb2 == NULL) in ctnetlink_stat_ct() 2605 err = ctnetlink_stat_ct_fill_info(skb2, NETLINK_CB(skb).portid, in ctnetlink_stat_ct() [all …]
|
/net/ipv6/ |
D | icmp.c | 657 struct sk_buff *skb2; in ip6_err_gen_icmpv6_unreach() local 667 skb2 = data_len ? skb_copy(skb, GFP_ATOMIC) : skb_clone(skb, GFP_ATOMIC); in ip6_err_gen_icmpv6_unreach() 669 if (!skb2) in ip6_err_gen_icmpv6_unreach() 672 skb_dst_drop(skb2); in ip6_err_gen_icmpv6_unreach() 673 skb_pull(skb2, nhs); in ip6_err_gen_icmpv6_unreach() 674 skb_reset_network_header(skb2); in ip6_err_gen_icmpv6_unreach() 676 rt = rt6_lookup(dev_net(skb->dev), &ipv6_hdr(skb2)->saddr, NULL, 0, in ip6_err_gen_icmpv6_unreach() 680 skb2->dev = rt->dst.dev; in ip6_err_gen_icmpv6_unreach() 688 __skb_push(skb2, nhs); in ip6_err_gen_icmpv6_unreach() 689 skb_reset_network_header(skb2); in ip6_err_gen_icmpv6_unreach() [all …]
|
D | netfilter.c | 220 struct sk_buff *skb2; in br_ip6_fragment() local 222 skb2 = ip6_frag_next(skb, &state); in br_ip6_fragment() 223 if (IS_ERR(skb2)) { in br_ip6_fragment() 224 err = PTR_ERR(skb2); in br_ip6_fragment() 228 skb_set_delivery_time(skb2, tstamp, mono_delivery_time); in br_ip6_fragment() 229 err = output(net, sk, data, skb2); in br_ip6_fragment()
|
D | ip6_tunnel.c | 565 struct sk_buff *skb2; in ip4ip6_err() local 600 skb2 = skb_clone(skb, GFP_ATOMIC); in ip4ip6_err() 601 if (!skb2) in ip4ip6_err() 604 skb_dst_drop(skb2); in ip4ip6_err() 606 skb_pull(skb2, offset); in ip4ip6_err() 607 skb_reset_network_header(skb2); in ip4ip6_err() 608 eiph = ip_hdr(skb2); in ip4ip6_err() 616 skb2->dev = rt->dst.dev; in ip4ip6_err() 629 skb_dst_set(skb2, &rt->dst); in ip4ip6_err() 631 if (ip_route_input(skb2, eiph->daddr, eiph->saddr, eiph->tos, in ip4ip6_err() [all …]
|
D | ip6_input.c | 539 struct sk_buff *skb2; in ip6_mc_input() local 573 skb2 = skb_clone(skb, GFP_ATOMIC); in ip6_mc_input() 575 skb2 = skb; in ip6_mc_input() 579 if (skb2) { in ip6_mc_input() 580 ip6_mr_input(skb2); in ip6_mc_input()
|
D | ip6mr.c | 2189 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip6_mr_forward() local 2190 if (skb2) in ip6_mr_forward() 2191 ip6mr_forward2(net, mrt, skb2, psend); in ip6_mr_forward() 2296 struct sk_buff *skb2; in ip6mr_get_route() local 2308 skb2 = alloc_skb(sizeof(struct ipv6hdr), GFP_ATOMIC); in ip6mr_get_route() 2309 if (!skb2) { in ip6mr_get_route() 2314 NETLINK_CB(skb2).portid = portid; in ip6mr_get_route() 2315 skb_reset_transport_header(skb2); in ip6mr_get_route() 2317 skb_put(skb2, sizeof(struct ipv6hdr)); in ip6mr_get_route() 2318 skb_reset_network_header(skb2); in ip6mr_get_route() [all …]
|
/net/mctp/ |
D | route.c | 783 struct sk_buff *skb2; in mctp_do_fragment_route() local 806 skb2 = alloc_skb(headroom + hlen + size, GFP_KERNEL); in mctp_do_fragment_route() 807 if (!skb2) { in mctp_do_fragment_route() 813 skb2->protocol = skb->protocol; in mctp_do_fragment_route() 814 skb2->priority = skb->priority; in mctp_do_fragment_route() 815 skb2->dev = skb->dev; in mctp_do_fragment_route() 816 memcpy(skb2->cb, skb->cb, sizeof(skb2->cb)); in mctp_do_fragment_route() 819 skb_set_owner_w(skb2, skb->sk); in mctp_do_fragment_route() 822 skb_reserve(skb2, headroom); in mctp_do_fragment_route() 823 skb_reset_network_header(skb2); in mctp_do_fragment_route() [all …]
|
/net/ipv4/ |
D | ip_output.c | 682 struct sk_buff *skb2; in ip_frag_next() local 695 skb2 = alloc_skb(len + state->hlen + state->ll_rs, GFP_ATOMIC); in ip_frag_next() 696 if (!skb2) in ip_frag_next() 703 ip_copy_metadata(skb2, skb); in ip_frag_next() 704 skb_reserve(skb2, state->ll_rs); in ip_frag_next() 705 skb_put(skb2, len + state->hlen); in ip_frag_next() 706 skb_reset_network_header(skb2); in ip_frag_next() 707 skb2->transport_header = skb2->network_header + state->hlen; in ip_frag_next() 715 skb_set_owner_w(skb2, skb->sk); in ip_frag_next() 721 skb_copy_from_linear_data(skb, skb_network_header(skb2), state->hlen); in ip_frag_next() [all …]
|
D | ipmr.c | 2048 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_mr_forward() local 2050 if (skb2) in ip_mr_forward() 2052 skb2, psend); in ip_mr_forward() 2060 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_mr_forward() local 2062 if (skb2) in ip_mr_forward() 2063 ipmr_queue_xmit(net, mrt, true_vifi, skb2, in ip_mr_forward() 2172 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_mr_input() local 2174 if (!skb2) in ip_mr_input() 2176 skb = skb2; in ip_mr_input() 2277 struct sk_buff *skb2; in ipmr_get_route() local [all …]
|
D | ip_input.c | 170 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_call_ra_chain() local 171 if (skb2) in ip_call_ra_chain() 172 raw_rcv(last, skb2); in ip_call_ra_chain()
|
/net/caif/ |
D | cfpkt_skbuff.c | 333 struct sk_buff *skb2; in cfpkt_split() local 352 skb2 = pkt_to_skb(tmppkt); in cfpkt_split() 355 if (skb2 == NULL) in cfpkt_split() 358 skb_put_data(skb2, split, len2nd); in cfpkt_split() 363 skb2->priority = skb->priority; in cfpkt_split() 364 return skb_to_pkt(skb2); in cfpkt_split()
|
/net/bridge/netfilter/ |
D | nf_conntrack_bridge.c | 108 struct sk_buff *skb2; in nf_br_ip_fragment() local 110 skb2 = ip_frag_next(skb, &state); in nf_br_ip_fragment() 111 if (IS_ERR(skb2)) { in nf_br_ip_fragment() 112 err = PTR_ERR(skb2); in nf_br_ip_fragment() 116 skb_set_delivery_time(skb2, tstamp, mono_delivery_time); in nf_br_ip_fragment() 117 err = output(net, sk, data, skb2); in nf_br_ip_fragment()
|
/net/sched/ |
D | sch_choke.c | 149 struct sk_buff *skb2) in choke_match_flow() argument 153 if (skb1->protocol != skb2->protocol) in choke_match_flow() 162 if (!choke_skb_cb(skb2)->keys_valid) { in choke_match_flow() 163 choke_skb_cb(skb2)->keys_valid = 1; in choke_match_flow() 164 skb_flow_dissect_flow_keys(skb2, &temp, 0); in choke_match_flow() 165 make_flow_keys_digest(&choke_skb_cb(skb2)->keys, &temp); in choke_match_flow() 169 &choke_skb_cb(skb2)->keys, in choke_match_flow()
|
D | sch_netem.c | 449 struct sk_buff *skb2 = NULL; in netem_enqueue() local 485 skb2 = skb_clone(skb, GFP_ATOMIC); in netem_enqueue() 524 if (skb2) in netem_enqueue() 525 __qdisc_drop(skb2, to_free); in netem_enqueue() 534 if (skb2) { in netem_enqueue() 539 rootq->enqueue(skb2, rootq, to_free); in netem_enqueue() 541 skb2 = NULL; in netem_enqueue() 612 if (skb2) in netem_enqueue() 613 __qdisc_drop(skb2, to_free); in netem_enqueue() 623 skb2 = segs->next; in netem_enqueue() [all …]
|
/net/atm/ |
D | lec.c | 142 struct sk_buff *skb2; in lec_handle_bridge() local 145 skb2 = alloc_skb(sizeof(struct atmlec_msg), GFP_ATOMIC); in lec_handle_bridge() 146 if (skb2 == NULL) in lec_handle_bridge() 148 skb2->len = sizeof(struct atmlec_msg); in lec_handle_bridge() 149 mesg = (struct atmlec_msg *)skb2->data; in lec_handle_bridge() 156 atm_force_charge(priv->lecd, skb2->truesize); in lec_handle_bridge() 158 skb_queue_tail(&sk->sk_receive_queue, skb2); in lec_handle_bridge() 207 struct sk_buff *skb2; in lec_start_xmit() local 236 skb2 = skb_realloc_headroom(skb, LEC_HEADER_LEN); in lec_start_xmit() 237 if (unlikely(!skb2)) { in lec_start_xmit() [all …]
|
/net/netlink/ |
D | af_netlink.c | 1435 struct sk_buff *skb, *skb2; member 1471 if (p->skb2 == NULL) { in do_one_broadcast() 1473 p->skb2 = skb_clone(p->skb, p->allocation); in do_one_broadcast() 1475 p->skb2 = skb_get(p->skb); in do_one_broadcast() 1480 skb_orphan(p->skb2); in do_one_broadcast() 1483 if (p->skb2 == NULL) { in do_one_broadcast() 1492 if (p->tx_filter && p->tx_filter(sk, p->skb2, p->tx_data)) { in do_one_broadcast() 1493 kfree_skb(p->skb2); in do_one_broadcast() 1494 p->skb2 = NULL; in do_one_broadcast() 1498 if (sk_filter(sk, p->skb2)) { in do_one_broadcast() [all …]
|