/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | tls_rxtx.c | 144 struct sk_buff *nskb, u32 tcp_seq, in mlx5e_tls_complete_sync_skb() argument 153 nskb->dev = skb->dev; in mlx5e_tls_complete_sync_skb() 154 skb_reset_mac_header(nskb); in mlx5e_tls_complete_sync_skb() 155 skb_set_network_header(nskb, skb_network_offset(skb)); in mlx5e_tls_complete_sync_skb() 156 skb_set_transport_header(nskb, skb_transport_offset(skb)); in mlx5e_tls_complete_sync_skb() 157 memcpy(nskb->data, skb->data, headln); in mlx5e_tls_complete_sync_skb() 158 memcpy(nskb->data + headln, &rcd_sn, sizeof(rcd_sn)); in mlx5e_tls_complete_sync_skb() 160 iph = ip_hdr(nskb); in mlx5e_tls_complete_sync_skb() 161 iph->tot_len = htons(nskb->len - skb_network_offset(nskb)); in mlx5e_tls_complete_sync_skb() 162 th = tcp_hdr(nskb); in mlx5e_tls_complete_sync_skb() [all …]
|
/drivers/bluetooth/ |
D | hci_bcsp.c | 166 struct sk_buff *nskb; in bcsp_prepare_pkt() local 218 nskb = alloc_skb((len + 6) * 2 + 2, GFP_ATOMIC); in bcsp_prepare_pkt() 219 if (!nskb) in bcsp_prepare_pkt() 222 hci_skb_pkt_type(nskb) = pkt_type; in bcsp_prepare_pkt() 224 bcsp_slip_msgdelim(nskb); in bcsp_prepare_pkt() 245 bcsp_slip_one_byte(nskb, hdr[i]); in bcsp_prepare_pkt() 253 bcsp_slip_one_byte(nskb, data[i]); in bcsp_prepare_pkt() 262 bcsp_slip_one_byte(nskb, (u8)((bcsp_txmsg_crc >> 8) & 0x00ff)); in bcsp_prepare_pkt() 263 bcsp_slip_one_byte(nskb, (u8)(bcsp_txmsg_crc & 0x00ff)); in bcsp_prepare_pkt() 266 bcsp_slip_msgdelim(nskb); in bcsp_prepare_pkt() [all …]
|
D | hci_h5.c | 113 struct sk_buff *nskb; in h5_link_control() local 115 nskb = alloc_skb(3, GFP_ATOMIC); in h5_link_control() 116 if (!nskb) in h5_link_control() 119 hci_skb_pkt_type(nskb) = HCI_3WIRE_LINK_PKT; in h5_link_control() 121 skb_put_data(nskb, data, len); in h5_link_control() 123 skb_queue_tail(&h5->unrel, nskb); in h5_link_control() 651 struct sk_buff *nskb; in h5_prepare_pkt() local 666 nskb = alloc_skb((len + 6) * 2 + 2, GFP_ATOMIC); in h5_prepare_pkt() 667 if (!nskb) in h5_prepare_pkt() 670 hci_skb_pkt_type(nskb) = pkt_type; in h5_prepare_pkt() [all …]
|
D | bfusb.c | 453 struct sk_buff *nskb; in bfusb_send_frame() local 478 nskb = bt_skb_alloc(count + 32, GFP_KERNEL); in bfusb_send_frame() 479 if (!nskb) { in bfusb_send_frame() 484 nskb->dev = (void *) data; in bfusb_send_frame() 493 skb_put_data(nskb, buf, 3); in bfusb_send_frame() 494 skb_copy_from_linear_data_offset(skb, sent, skb_put(nskb, size), size); in bfusb_send_frame() 501 if ((nskb->len % data->bulk_pkt_size) == 0) { in bfusb_send_frame() 504 skb_put_data(nskb, buf, 2); in bfusb_send_frame() 509 skb_queue_tail(&data->transmit_q, nskb); in bfusb_send_frame()
|
/drivers/net/wireless/ath/ath9k/ |
D | hif_usb.c | 313 struct sk_buff *nskb = NULL; in __hif_usb_tx() local 333 nskb = __skb_dequeue(&hif_dev->tx.tx_skb_queue); in __hif_usb_tx() 336 BUG_ON(!nskb); in __hif_usb_tx() 343 *hdr++ = cpu_to_le16(nskb->len); in __hif_usb_tx() 346 memcpy(buf, nskb->data, nskb->len); in __hif_usb_tx() 347 tx_buf->len = nskb->len + 4; in __hif_usb_tx() 355 __skb_queue_tail(&tx_buf->skb_queue, nskb); in __hif_usb_tx() 532 struct sk_buff *nskb, *skb_pool[MAX_PKT_NUM_IN_TRANSFER]; in ath9k_hif_usb_rx_stream() local 598 nskb = __dev_alloc_skb(pkt_len + 32, GFP_ATOMIC); in ath9k_hif_usb_rx_stream() 599 if (!nskb) { in ath9k_hif_usb_rx_stream() [all …]
|
/drivers/isdn/mISDN/ |
D | dsp_core.c | 283 struct sk_buff *nskb; in dsp_control_req() local 589 nskb = _alloc_mISDN_skb(PH_CONTROL_IND, MISDN_ID_ANY, in dsp_control_req() 591 if (nskb) { in dsp_control_req() 593 if (dsp->up->send(dsp->up, nskb)) in dsp_control_req() 594 dev_kfree_skb(nskb); in dsp_control_req() 596 dev_kfree_skb(nskb); in dsp_control_req() 741 struct sk_buff *nskb; in dsp_function() local 747 nskb = _alloc_mISDN_skb(PH_CONTROL_IND, in dsp_function() 750 if (nskb) { in dsp_function() 753 dsp->up, nskb)) in dsp_function() [all …]
|
D | dsp_cmx.c | 1315 struct sk_buff *nskb, *txskb; local 1354 nskb = mI_alloc_skb(len + preload, GFP_ATOMIC); 1355 if (!nskb) { 1361 hh = mISDN_HEAD_P(nskb); 1370 d = skb_put(nskb, preload + len); /* result */ 1584 skb_queue_tail(&dsp->sendq, nskb); 1598 skb_put_data(txskb, nskb->data + preload, len); 1608 dsp_change_volume(nskb, dsp->tx_volume); 1611 dsp_pipeline_process_tx(&dsp->pipeline, nskb->data, 1612 nskb->len); [all …]
|
D | dsp_tones.c | 440 struct sk_buff *nskb; in dsp_tone_hw_message() local 443 nskb = _alloc_mISDN_skb(PH_CONTROL_REQ, in dsp_tone_hw_message() 446 if (nskb) { in dsp_tone_hw_message() 448 if (dsp->ch.recv(dsp->ch.peer, nskb)) in dsp_tone_hw_message() 449 dev_kfree_skb(nskb); in dsp_tone_hw_message() 451 dev_kfree_skb(nskb); in dsp_tone_hw_message()
|
D | layer2.c | 238 struct sk_buff *nskb = skb; in ph_data_confirm() local 243 nskb = skb_dequeue(&l2->down_queue); in ph_data_confirm() 244 if (nskb) { in ph_data_confirm() 245 l2->down_id = mISDN_HEAD_ID(nskb); in ph_data_confirm() 246 if (l2down_skb(l2, nskb)) { in ph_data_confirm() 247 dev_kfree_skb(nskb); in ph_data_confirm() 263 nskb = skb_dequeue(&l2->down_queue); in ph_data_confirm() 264 if (nskb) { in ph_data_confirm() 265 l2->down_id = mISDN_HEAD_ID(nskb); in ph_data_confirm() 266 if (l2down_skb(l2, nskb)) { in ph_data_confirm() [all …]
|
D | l1oip_core.c | 353 struct sk_buff *nskb; in l1oip_socket_recv() local 384 nskb = mI_alloc_skb((remotecodec == 3) ? (len << 1) : len, GFP_ATOMIC); in l1oip_socket_recv() 385 if (!nskb) { in l1oip_socket_recv() 389 p = skb_put(nskb, (remotecodec == 3) ? (len << 1) : len); in l1oip_socket_recv() 402 dch->rx_skb = nskb; in l1oip_socket_recv() 429 swap(hc->chan[channel].disorder_skb, nskb); in l1oip_socket_recv() 433 if (nskb) in l1oip_socket_recv() 435 queue_ch_frame(&bch->ch, PH_DATA_IND, rx_counter, nskb); in l1oip_socket_recv()
|
/drivers/net/ipvlan/ |
D | ipvlan_core.c | 230 struct sk_buff *skb, *nskb; in ipvlan_process_multicast() local 267 nskb = skb_clone(skb, GFP_ATOMIC); in ipvlan_process_multicast() 269 if (nskb) { in ipvlan_process_multicast() 271 nskb->pkt_type = pkt_type; in ipvlan_process_multicast() 272 nskb->dev = ipvlan->dev; in ipvlan_process_multicast() 274 ret = dev_forward_skb(ipvlan->dev, nskb); in ipvlan_process_multicast() 276 ret = netif_rx(nskb); in ipvlan_process_multicast() 708 struct sk_buff *nskb = skb_clone(skb, GFP_ATOMIC); in ipvlan_handle_mode_l2() local 716 if (nskb) { in ipvlan_handle_mode_l2() 717 ipvlan_skb_crossing_ns(nskb, NULL); in ipvlan_handle_mode_l2() [all …]
|
/drivers/net/ethernet/sun/ |
D | sunvnet_common.c | 1110 struct sk_buff *nskb; in vnet_skb_shape() local 1136 nskb = alloc_and_align_skb(skb->dev, len); in vnet_skb_shape() 1137 if (!nskb) { in vnet_skb_shape() 1141 skb_reserve(nskb, VNET_PACKET_SKIP); in vnet_skb_shape() 1143 nskb->protocol = skb->protocol; in vnet_skb_shape() 1145 skb_set_mac_header(nskb, offset); in vnet_skb_shape() 1147 skb_set_network_header(nskb, offset); in vnet_skb_shape() 1149 skb_set_transport_header(nskb, offset); in vnet_skb_shape() 1152 nskb->csum_offset = skb->csum_offset; in vnet_skb_shape() 1153 nskb->ip_summed = skb->ip_summed; in vnet_skb_shape() [all …]
|
/drivers/net/xen-netback/ |
D | netback.c | 365 struct sk_buff *nskb) in xenvif_get_requests() argument 389 shinfo = skb_shinfo(nskb); in xenvif_get_requests() 402 skb_shinfo(skb)->frag_list = nskb; in xenvif_get_requests() 794 struct sk_buff *skb, *nskb; in xenvif_tx_build_gops() local 921 nskb = NULL; in xenvif_tx_build_gops() 926 nskb = xenvif_alloc_skb(0); in xenvif_tx_build_gops() 927 if (unlikely(nskb == NULL)) { in xenvif_tx_build_gops() 946 kfree_skb(nskb); in xenvif_tx_build_gops() 1014 frag_overflow, nskb); in xenvif_tx_build_gops() 1038 struct sk_buff *nskb = skb_shinfo(skb)->frag_list; in xenvif_handle_frag_list() local [all …]
|
/drivers/net/ |
D | macvlan.c | 264 struct sk_buff *nskb; in macvlan_broadcast() local 282 nskb = skb_clone(skb, GFP_ATOMIC); in macvlan_broadcast() 283 if (likely(nskb)) in macvlan_broadcast() 285 nskb, vlan, eth, in macvlan_broadcast() 287 netif_rx_ni(nskb); in macvlan_broadcast() 344 struct sk_buff *nskb; in macvlan_broadcast_enqueue() local 347 nskb = skb_clone(skb, GFP_ATOMIC); in macvlan_broadcast_enqueue() 348 if (!nskb) in macvlan_broadcast_enqueue() 351 MACVLAN_SKB_CB(nskb)->src = src; in macvlan_broadcast_enqueue() 357 __skb_queue_tail(&port->bc_queue, nskb); in macvlan_broadcast_enqueue() [all …]
|
D | xen-netfront.c | 580 struct sk_buff *nskb; in xennet_start_xmit() local 614 nskb = skb_copy(skb, GFP_ATOMIC); in xennet_start_xmit() 615 if (!nskb) in xennet_start_xmit() 618 skb = nskb; in xennet_start_xmit() 895 struct sk_buff *nskb; in xennet_fill_frags() local 897 while ((nskb = __skb_dequeue(list))) { in xennet_fill_frags() 900 skb_frag_t *nfrag = &skb_shinfo(nskb)->frags[0]; in xennet_fill_frags() 910 kfree_skb(nskb); in xennet_fill_frags() 918 skb_shinfo(nskb)->nr_frags = 0; in xennet_fill_frags() 919 kfree_skb(nskb); in xennet_fill_frags()
|
D | veth.c | 618 struct sk_buff *nskb; in veth_xdp_rcv_skb() local 639 nskb = veth_build_skb(head, in veth_xdp_rcv_skb() 642 if (!nskb) { in veth_xdp_rcv_skb() 647 skb_copy_header(nskb, skb); in veth_xdp_rcv_skb() 648 head_off = skb_headroom(nskb) - skb_headroom(skb); in veth_xdp_rcv_skb() 649 skb_headers_offset_update(nskb, head_off); in veth_xdp_rcv_skb() 651 skb = nskb; in veth_xdp_rcv_skb()
|
D | macsec.c | 676 struct sk_buff *nskb = skb_copy_expand(skb, in macsec_encrypt() local 680 if (likely(nskb)) { in macsec_encrypt() 682 skb = nskb; in macsec_encrypt() 1045 struct sk_buff *nskb; in handle_not_macsec() local 1056 nskb = skb_clone(skb, GFP_ATOMIC); in handle_not_macsec() 1057 if (!nskb) in handle_not_macsec() 1060 nskb->dev = macsec->secy.netdev; in handle_not_macsec() 1062 if (netif_rx(nskb) == NET_RX_SUCCESS) { in handle_not_macsec() 1267 struct sk_buff *nskb; in macsec_handle_frame() local 1285 nskb = skb_clone(skb, GFP_ATOMIC); in macsec_handle_frame() [all …]
|
D | tap.c | 356 struct sk_buff *nskb = segs->next; in tap_handle_frame() local 361 kfree_skb_list(nskb); in tap_handle_frame() 364 segs = nskb; in tap_handle_frame()
|
/drivers/isdn/capi/ |
D | capi.c | 386 struct sk_buff *nskb; in gen_data_b3_resp_for() local 387 nskb = alloc_skb(CAPI_DATA_B3_RESP_LEN, GFP_KERNEL); in gen_data_b3_resp_for() 388 if (nskb) { in gen_data_b3_resp_for() 390 unsigned char *s = skb_put(nskb, CAPI_DATA_B3_RESP_LEN); in gen_data_b3_resp_for() 399 return nskb; in gen_data_b3_resp_for() 406 struct sk_buff *nskb; in handle_recv_skb() local 440 nskb = gen_data_b3_resp_for(mp, skb); in handle_recv_skb() 441 if (!nskb) { in handle_recv_skb() 448 errcode = capi20_put_message(mp->ap, nskb); in handle_recv_skb() 458 kfree_skb(nskb); in handle_recv_skb()
|
/drivers/s390/net/ |
D | netiucv.c | 1172 struct sk_buff *nskb = skb; in netiucv_transmit_skb() local 1181 nskb = alloc_skb(skb->len + NETIUCV_HDRLEN + in netiucv_transmit_skb() 1183 if (!nskb) { in netiucv_transmit_skb() 1188 skb_reserve(nskb, NETIUCV_HDRLEN); in netiucv_transmit_skb() 1189 skb_put_data(nskb, skb->data, skb->len); in netiucv_transmit_skb() 1196 header.next = nskb->len + NETIUCV_HDRLEN; in netiucv_transmit_skb() 1197 memcpy(skb_push(nskb, NETIUCV_HDRLEN), &header, NETIUCV_HDRLEN); in netiucv_transmit_skb() 1199 skb_put_data(nskb, &header, NETIUCV_HDRLEN); in netiucv_transmit_skb() 1207 nskb->data, nskb->len); in netiucv_transmit_skb() 1221 dev_kfree_skb(nskb); in netiucv_transmit_skb() [all …]
|
D | ctcm_main.c | 472 struct sk_buff *nskb; in ctcm_transmit_skb() local 519 nskb = alloc_skb(skb->len, GFP_ATOMIC | GFP_DMA); in ctcm_transmit_skb() 520 if (!nskb) { in ctcm_transmit_skb() 526 skb_put_data(nskb, skb->data, skb->len); in ctcm_transmit_skb() 527 refcount_inc(&nskb->users); in ctcm_transmit_skb() 530 skb = nskb; in ctcm_transmit_skb() 670 struct sk_buff *nskb; in ctcmpc_transmit_skb() local 727 nskb = __dev_alloc_skb(skb->len, GFP_ATOMIC | GFP_DMA); in ctcmpc_transmit_skb() 728 if (!nskb) { in ctcmpc_transmit_skb() 731 skb_put_data(nskb, skb->data, skb->len); in ctcmpc_transmit_skb() [all …]
|
/drivers/scsi/fcoe/ |
D | fcoe_transport.c | 348 struct sk_buff *nskb; in fcoe_start_io() local 351 nskb = skb_clone(skb, GFP_ATOMIC); in fcoe_start_io() 352 if (!nskb) in fcoe_start_io() 354 rc = dev_queue_xmit(nskb); in fcoe_start_io()
|
/drivers/net/ethernet/broadcom/ |
D | bcm63xx_enet.c | 370 struct sk_buff *nskb; in bcm_enet_receive_queue() local 372 nskb = napi_alloc_skb(&priv->napi, len); in bcm_enet_receive_queue() 373 if (!nskb) { in bcm_enet_receive_queue() 381 memcpy(nskb->data, skb->data, len); in bcm_enet_receive_queue() 384 skb = nskb; in bcm_enet_receive_queue() 587 struct sk_buff *nskb; in bcm_enet_start_xmit() local 589 nskb = skb_copy_expand(skb, 0, needed, GFP_ATOMIC); in bcm_enet_start_xmit() 590 if (!nskb) { in bcm_enet_start_xmit() 595 skb = nskb; in bcm_enet_start_xmit()
|
/drivers/net/wireless/ |
D | mac80211_hwsim.c | 1285 struct sk_buff *nskb; in mac80211_hwsim_tx_frame_no_nl() local 1323 nskb = dev_alloc_skb(128); in mac80211_hwsim_tx_frame_no_nl() 1324 if (!nskb) { in mac80211_hwsim_tx_frame_no_nl() 1330 skb_add_rx_frag(nskb, 0, page, 0, skb->len, skb->len); in mac80211_hwsim_tx_frame_no_nl() 1332 nskb = skb_copy(skb, GFP_ATOMIC); in mac80211_hwsim_tx_frame_no_nl() 1333 if (!nskb) in mac80211_hwsim_tx_frame_no_nl() 1342 memcpy(IEEE80211_SKB_RXCB(nskb), &rx_status, sizeof(rx_status)); in mac80211_hwsim_tx_frame_no_nl() 1344 mac80211_hwsim_add_vendor_rtap(nskb); in mac80211_hwsim_tx_frame_no_nl() 1347 data2->rx_bytes += nskb->len; in mac80211_hwsim_tx_frame_no_nl() 1348 ieee80211_rx_irqsafe(data2->hw, nskb); in mac80211_hwsim_tx_frame_no_nl()
|
/drivers/net/wireless/ath/ath6kl/ |
D | txrx.c | 489 struct sk_buff *nskb; in ath6kl_data_tx() local 491 nskb = skb_copy_expand(skb, HTC_HDR_LENGTH, 0, GFP_ATOMIC); in ath6kl_data_tx() 492 if (nskb == NULL) in ath6kl_data_tx() 495 skb = nskb; in ath6kl_data_tx()
|