Home
last modified time | relevance | path

Searched refs:nskb (Results 1 – 25 of 33) sorted by relevance

12

/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
Dtls_rxtx.c144 struct sk_buff *nskb, u32 tcp_seq, in mlx5e_tls_complete_sync_skb() argument
153 nskb->dev = skb->dev; in mlx5e_tls_complete_sync_skb()
154 skb_reset_mac_header(nskb); in mlx5e_tls_complete_sync_skb()
155 skb_set_network_header(nskb, skb_network_offset(skb)); in mlx5e_tls_complete_sync_skb()
156 skb_set_transport_header(nskb, skb_transport_offset(skb)); in mlx5e_tls_complete_sync_skb()
157 memcpy(nskb->data, skb->data, headln); in mlx5e_tls_complete_sync_skb()
158 memcpy(nskb->data + headln, &rcd_sn, sizeof(rcd_sn)); in mlx5e_tls_complete_sync_skb()
160 iph = ip_hdr(nskb); in mlx5e_tls_complete_sync_skb()
161 iph->tot_len = htons(nskb->len - skb_network_offset(nskb)); in mlx5e_tls_complete_sync_skb()
162 th = tcp_hdr(nskb); in mlx5e_tls_complete_sync_skb()
[all …]
/drivers/bluetooth/
Dhci_bcsp.c166 struct sk_buff *nskb; in bcsp_prepare_pkt() local
218 nskb = alloc_skb((len + 6) * 2 + 2, GFP_ATOMIC); in bcsp_prepare_pkt()
219 if (!nskb) in bcsp_prepare_pkt()
222 hci_skb_pkt_type(nskb) = pkt_type; in bcsp_prepare_pkt()
224 bcsp_slip_msgdelim(nskb); in bcsp_prepare_pkt()
245 bcsp_slip_one_byte(nskb, hdr[i]); in bcsp_prepare_pkt()
253 bcsp_slip_one_byte(nskb, data[i]); in bcsp_prepare_pkt()
262 bcsp_slip_one_byte(nskb, (u8)((bcsp_txmsg_crc >> 8) & 0x00ff)); in bcsp_prepare_pkt()
263 bcsp_slip_one_byte(nskb, (u8)(bcsp_txmsg_crc & 0x00ff)); in bcsp_prepare_pkt()
266 bcsp_slip_msgdelim(nskb); in bcsp_prepare_pkt()
[all …]
Dhci_h5.c113 struct sk_buff *nskb; in h5_link_control() local
115 nskb = alloc_skb(3, GFP_ATOMIC); in h5_link_control()
116 if (!nskb) in h5_link_control()
119 hci_skb_pkt_type(nskb) = HCI_3WIRE_LINK_PKT; in h5_link_control()
121 skb_put_data(nskb, data, len); in h5_link_control()
123 skb_queue_tail(&h5->unrel, nskb); in h5_link_control()
651 struct sk_buff *nskb; in h5_prepare_pkt() local
666 nskb = alloc_skb((len + 6) * 2 + 2, GFP_ATOMIC); in h5_prepare_pkt()
667 if (!nskb) in h5_prepare_pkt()
670 hci_skb_pkt_type(nskb) = pkt_type; in h5_prepare_pkt()
[all …]
Dbfusb.c453 struct sk_buff *nskb; in bfusb_send_frame() local
478 nskb = bt_skb_alloc(count + 32, GFP_KERNEL); in bfusb_send_frame()
479 if (!nskb) { in bfusb_send_frame()
484 nskb->dev = (void *) data; in bfusb_send_frame()
493 skb_put_data(nskb, buf, 3); in bfusb_send_frame()
494 skb_copy_from_linear_data_offset(skb, sent, skb_put(nskb, size), size); in bfusb_send_frame()
501 if ((nskb->len % data->bulk_pkt_size) == 0) { in bfusb_send_frame()
504 skb_put_data(nskb, buf, 2); in bfusb_send_frame()
509 skb_queue_tail(&data->transmit_q, nskb); in bfusb_send_frame()
/drivers/net/wireless/ath/ath9k/
Dhif_usb.c313 struct sk_buff *nskb = NULL; in __hif_usb_tx() local
333 nskb = __skb_dequeue(&hif_dev->tx.tx_skb_queue); in __hif_usb_tx()
336 BUG_ON(!nskb); in __hif_usb_tx()
343 *hdr++ = cpu_to_le16(nskb->len); in __hif_usb_tx()
346 memcpy(buf, nskb->data, nskb->len); in __hif_usb_tx()
347 tx_buf->len = nskb->len + 4; in __hif_usb_tx()
355 __skb_queue_tail(&tx_buf->skb_queue, nskb); in __hif_usb_tx()
532 struct sk_buff *nskb, *skb_pool[MAX_PKT_NUM_IN_TRANSFER]; in ath9k_hif_usb_rx_stream() local
598 nskb = __dev_alloc_skb(pkt_len + 32, GFP_ATOMIC); in ath9k_hif_usb_rx_stream()
599 if (!nskb) { in ath9k_hif_usb_rx_stream()
[all …]
/drivers/isdn/mISDN/
Ddsp_core.c283 struct sk_buff *nskb; in dsp_control_req() local
589 nskb = _alloc_mISDN_skb(PH_CONTROL_IND, MISDN_ID_ANY, in dsp_control_req()
591 if (nskb) { in dsp_control_req()
593 if (dsp->up->send(dsp->up, nskb)) in dsp_control_req()
594 dev_kfree_skb(nskb); in dsp_control_req()
596 dev_kfree_skb(nskb); in dsp_control_req()
741 struct sk_buff *nskb; in dsp_function() local
747 nskb = _alloc_mISDN_skb(PH_CONTROL_IND, in dsp_function()
750 if (nskb) { in dsp_function()
753 dsp->up, nskb)) in dsp_function()
[all …]
Ddsp_cmx.c1315 struct sk_buff *nskb, *txskb; local
1354 nskb = mI_alloc_skb(len + preload, GFP_ATOMIC);
1355 if (!nskb) {
1361 hh = mISDN_HEAD_P(nskb);
1370 d = skb_put(nskb, preload + len); /* result */
1584 skb_queue_tail(&dsp->sendq, nskb);
1598 skb_put_data(txskb, nskb->data + preload, len);
1608 dsp_change_volume(nskb, dsp->tx_volume);
1611 dsp_pipeline_process_tx(&dsp->pipeline, nskb->data,
1612 nskb->len);
[all …]
Ddsp_tones.c440 struct sk_buff *nskb; in dsp_tone_hw_message() local
443 nskb = _alloc_mISDN_skb(PH_CONTROL_REQ, in dsp_tone_hw_message()
446 if (nskb) { in dsp_tone_hw_message()
448 if (dsp->ch.recv(dsp->ch.peer, nskb)) in dsp_tone_hw_message()
449 dev_kfree_skb(nskb); in dsp_tone_hw_message()
451 dev_kfree_skb(nskb); in dsp_tone_hw_message()
Dlayer2.c238 struct sk_buff *nskb = skb; in ph_data_confirm() local
243 nskb = skb_dequeue(&l2->down_queue); in ph_data_confirm()
244 if (nskb) { in ph_data_confirm()
245 l2->down_id = mISDN_HEAD_ID(nskb); in ph_data_confirm()
246 if (l2down_skb(l2, nskb)) { in ph_data_confirm()
247 dev_kfree_skb(nskb); in ph_data_confirm()
263 nskb = skb_dequeue(&l2->down_queue); in ph_data_confirm()
264 if (nskb) { in ph_data_confirm()
265 l2->down_id = mISDN_HEAD_ID(nskb); in ph_data_confirm()
266 if (l2down_skb(l2, nskb)) { in ph_data_confirm()
[all …]
Dl1oip_core.c353 struct sk_buff *nskb; in l1oip_socket_recv() local
384 nskb = mI_alloc_skb((remotecodec == 3) ? (len << 1) : len, GFP_ATOMIC); in l1oip_socket_recv()
385 if (!nskb) { in l1oip_socket_recv()
389 p = skb_put(nskb, (remotecodec == 3) ? (len << 1) : len); in l1oip_socket_recv()
402 dch->rx_skb = nskb; in l1oip_socket_recv()
429 swap(hc->chan[channel].disorder_skb, nskb); in l1oip_socket_recv()
433 if (nskb) in l1oip_socket_recv()
435 queue_ch_frame(&bch->ch, PH_DATA_IND, rx_counter, nskb); in l1oip_socket_recv()
/drivers/net/ipvlan/
Dipvlan_core.c230 struct sk_buff *skb, *nskb; in ipvlan_process_multicast() local
267 nskb = skb_clone(skb, GFP_ATOMIC); in ipvlan_process_multicast()
269 if (nskb) { in ipvlan_process_multicast()
271 nskb->pkt_type = pkt_type; in ipvlan_process_multicast()
272 nskb->dev = ipvlan->dev; in ipvlan_process_multicast()
274 ret = dev_forward_skb(ipvlan->dev, nskb); in ipvlan_process_multicast()
276 ret = netif_rx(nskb); in ipvlan_process_multicast()
708 struct sk_buff *nskb = skb_clone(skb, GFP_ATOMIC); in ipvlan_handle_mode_l2() local
716 if (nskb) { in ipvlan_handle_mode_l2()
717 ipvlan_skb_crossing_ns(nskb, NULL); in ipvlan_handle_mode_l2()
[all …]
/drivers/net/ethernet/sun/
Dsunvnet_common.c1110 struct sk_buff *nskb; in vnet_skb_shape() local
1136 nskb = alloc_and_align_skb(skb->dev, len); in vnet_skb_shape()
1137 if (!nskb) { in vnet_skb_shape()
1141 skb_reserve(nskb, VNET_PACKET_SKIP); in vnet_skb_shape()
1143 nskb->protocol = skb->protocol; in vnet_skb_shape()
1145 skb_set_mac_header(nskb, offset); in vnet_skb_shape()
1147 skb_set_network_header(nskb, offset); in vnet_skb_shape()
1149 skb_set_transport_header(nskb, offset); in vnet_skb_shape()
1152 nskb->csum_offset = skb->csum_offset; in vnet_skb_shape()
1153 nskb->ip_summed = skb->ip_summed; in vnet_skb_shape()
[all …]
/drivers/net/xen-netback/
Dnetback.c365 struct sk_buff *nskb) in xenvif_get_requests() argument
389 shinfo = skb_shinfo(nskb); in xenvif_get_requests()
402 skb_shinfo(skb)->frag_list = nskb; in xenvif_get_requests()
794 struct sk_buff *skb, *nskb; in xenvif_tx_build_gops() local
921 nskb = NULL; in xenvif_tx_build_gops()
926 nskb = xenvif_alloc_skb(0); in xenvif_tx_build_gops()
927 if (unlikely(nskb == NULL)) { in xenvif_tx_build_gops()
946 kfree_skb(nskb); in xenvif_tx_build_gops()
1014 frag_overflow, nskb); in xenvif_tx_build_gops()
1038 struct sk_buff *nskb = skb_shinfo(skb)->frag_list; in xenvif_handle_frag_list() local
[all …]
/drivers/net/
Dmacvlan.c264 struct sk_buff *nskb; in macvlan_broadcast() local
282 nskb = skb_clone(skb, GFP_ATOMIC); in macvlan_broadcast()
283 if (likely(nskb)) in macvlan_broadcast()
285 nskb, vlan, eth, in macvlan_broadcast()
287 netif_rx_ni(nskb); in macvlan_broadcast()
344 struct sk_buff *nskb; in macvlan_broadcast_enqueue() local
347 nskb = skb_clone(skb, GFP_ATOMIC); in macvlan_broadcast_enqueue()
348 if (!nskb) in macvlan_broadcast_enqueue()
351 MACVLAN_SKB_CB(nskb)->src = src; in macvlan_broadcast_enqueue()
357 __skb_queue_tail(&port->bc_queue, nskb); in macvlan_broadcast_enqueue()
[all …]
Dxen-netfront.c580 struct sk_buff *nskb; in xennet_start_xmit() local
614 nskb = skb_copy(skb, GFP_ATOMIC); in xennet_start_xmit()
615 if (!nskb) in xennet_start_xmit()
618 skb = nskb; in xennet_start_xmit()
895 struct sk_buff *nskb; in xennet_fill_frags() local
897 while ((nskb = __skb_dequeue(list))) { in xennet_fill_frags()
900 skb_frag_t *nfrag = &skb_shinfo(nskb)->frags[0]; in xennet_fill_frags()
910 kfree_skb(nskb); in xennet_fill_frags()
918 skb_shinfo(nskb)->nr_frags = 0; in xennet_fill_frags()
919 kfree_skb(nskb); in xennet_fill_frags()
Dveth.c618 struct sk_buff *nskb; in veth_xdp_rcv_skb() local
639 nskb = veth_build_skb(head, in veth_xdp_rcv_skb()
642 if (!nskb) { in veth_xdp_rcv_skb()
647 skb_copy_header(nskb, skb); in veth_xdp_rcv_skb()
648 head_off = skb_headroom(nskb) - skb_headroom(skb); in veth_xdp_rcv_skb()
649 skb_headers_offset_update(nskb, head_off); in veth_xdp_rcv_skb()
651 skb = nskb; in veth_xdp_rcv_skb()
Dmacsec.c676 struct sk_buff *nskb = skb_copy_expand(skb, in macsec_encrypt() local
680 if (likely(nskb)) { in macsec_encrypt()
682 skb = nskb; in macsec_encrypt()
1045 struct sk_buff *nskb; in handle_not_macsec() local
1056 nskb = skb_clone(skb, GFP_ATOMIC); in handle_not_macsec()
1057 if (!nskb) in handle_not_macsec()
1060 nskb->dev = macsec->secy.netdev; in handle_not_macsec()
1062 if (netif_rx(nskb) == NET_RX_SUCCESS) { in handle_not_macsec()
1267 struct sk_buff *nskb; in macsec_handle_frame() local
1285 nskb = skb_clone(skb, GFP_ATOMIC); in macsec_handle_frame()
[all …]
Dtap.c356 struct sk_buff *nskb = segs->next; in tap_handle_frame() local
361 kfree_skb_list(nskb); in tap_handle_frame()
364 segs = nskb; in tap_handle_frame()
/drivers/isdn/capi/
Dcapi.c386 struct sk_buff *nskb; in gen_data_b3_resp_for() local
387 nskb = alloc_skb(CAPI_DATA_B3_RESP_LEN, GFP_KERNEL); in gen_data_b3_resp_for()
388 if (nskb) { in gen_data_b3_resp_for()
390 unsigned char *s = skb_put(nskb, CAPI_DATA_B3_RESP_LEN); in gen_data_b3_resp_for()
399 return nskb; in gen_data_b3_resp_for()
406 struct sk_buff *nskb; in handle_recv_skb() local
440 nskb = gen_data_b3_resp_for(mp, skb); in handle_recv_skb()
441 if (!nskb) { in handle_recv_skb()
448 errcode = capi20_put_message(mp->ap, nskb); in handle_recv_skb()
458 kfree_skb(nskb); in handle_recv_skb()
/drivers/s390/net/
Dnetiucv.c1172 struct sk_buff *nskb = skb; in netiucv_transmit_skb() local
1181 nskb = alloc_skb(skb->len + NETIUCV_HDRLEN + in netiucv_transmit_skb()
1183 if (!nskb) { in netiucv_transmit_skb()
1188 skb_reserve(nskb, NETIUCV_HDRLEN); in netiucv_transmit_skb()
1189 skb_put_data(nskb, skb->data, skb->len); in netiucv_transmit_skb()
1196 header.next = nskb->len + NETIUCV_HDRLEN; in netiucv_transmit_skb()
1197 memcpy(skb_push(nskb, NETIUCV_HDRLEN), &header, NETIUCV_HDRLEN); in netiucv_transmit_skb()
1199 skb_put_data(nskb, &header, NETIUCV_HDRLEN); in netiucv_transmit_skb()
1207 nskb->data, nskb->len); in netiucv_transmit_skb()
1221 dev_kfree_skb(nskb); in netiucv_transmit_skb()
[all …]
Dctcm_main.c472 struct sk_buff *nskb; in ctcm_transmit_skb() local
519 nskb = alloc_skb(skb->len, GFP_ATOMIC | GFP_DMA); in ctcm_transmit_skb()
520 if (!nskb) { in ctcm_transmit_skb()
526 skb_put_data(nskb, skb->data, skb->len); in ctcm_transmit_skb()
527 refcount_inc(&nskb->users); in ctcm_transmit_skb()
530 skb = nskb; in ctcm_transmit_skb()
670 struct sk_buff *nskb; in ctcmpc_transmit_skb() local
727 nskb = __dev_alloc_skb(skb->len, GFP_ATOMIC | GFP_DMA); in ctcmpc_transmit_skb()
728 if (!nskb) { in ctcmpc_transmit_skb()
731 skb_put_data(nskb, skb->data, skb->len); in ctcmpc_transmit_skb()
[all …]
/drivers/scsi/fcoe/
Dfcoe_transport.c348 struct sk_buff *nskb; in fcoe_start_io() local
351 nskb = skb_clone(skb, GFP_ATOMIC); in fcoe_start_io()
352 if (!nskb) in fcoe_start_io()
354 rc = dev_queue_xmit(nskb); in fcoe_start_io()
/drivers/net/ethernet/broadcom/
Dbcm63xx_enet.c370 struct sk_buff *nskb; in bcm_enet_receive_queue() local
372 nskb = napi_alloc_skb(&priv->napi, len); in bcm_enet_receive_queue()
373 if (!nskb) { in bcm_enet_receive_queue()
381 memcpy(nskb->data, skb->data, len); in bcm_enet_receive_queue()
384 skb = nskb; in bcm_enet_receive_queue()
587 struct sk_buff *nskb; in bcm_enet_start_xmit() local
589 nskb = skb_copy_expand(skb, 0, needed, GFP_ATOMIC); in bcm_enet_start_xmit()
590 if (!nskb) { in bcm_enet_start_xmit()
595 skb = nskb; in bcm_enet_start_xmit()
/drivers/net/wireless/
Dmac80211_hwsim.c1285 struct sk_buff *nskb; in mac80211_hwsim_tx_frame_no_nl() local
1323 nskb = dev_alloc_skb(128); in mac80211_hwsim_tx_frame_no_nl()
1324 if (!nskb) { in mac80211_hwsim_tx_frame_no_nl()
1330 skb_add_rx_frag(nskb, 0, page, 0, skb->len, skb->len); in mac80211_hwsim_tx_frame_no_nl()
1332 nskb = skb_copy(skb, GFP_ATOMIC); in mac80211_hwsim_tx_frame_no_nl()
1333 if (!nskb) in mac80211_hwsim_tx_frame_no_nl()
1342 memcpy(IEEE80211_SKB_RXCB(nskb), &rx_status, sizeof(rx_status)); in mac80211_hwsim_tx_frame_no_nl()
1344 mac80211_hwsim_add_vendor_rtap(nskb); in mac80211_hwsim_tx_frame_no_nl()
1347 data2->rx_bytes += nskb->len; in mac80211_hwsim_tx_frame_no_nl()
1348 ieee80211_rx_irqsafe(data2->hw, nskb); in mac80211_hwsim_tx_frame_no_nl()
/drivers/net/wireless/ath/ath6kl/
Dtxrx.c489 struct sk_buff *nskb; in ath6kl_data_tx() local
491 nskb = skb_copy_expand(skb, HTC_HDR_LENGTH, 0, GFP_ATOMIC); in ath6kl_data_tx()
492 if (nskb == NULL) in ath6kl_data_tx()
495 skb = nskb; in ath6kl_data_tx()

12