• Home
  • Raw
  • Download

Lines Matching refs:skb

144 	struct sk_buff *skb;  in i2400m_wake_tx_work()  local
148 skb = i2400m->wake_tx_skb; in i2400m_wake_tx_work()
152 d_fnstart(3, dev, "(ws %p i2400m %p skb %p)\n", ws, i2400m, skb); in i2400m_wake_tx_work()
154 if (skb == NULL) { in i2400m_wake_tx_work()
184 result = i2400m_tx(i2400m, skb->data, skb->len, I2400M_PT_DATA); in i2400m_wake_tx_work()
188 kfree_skb(skb); /* refcount transferred by _hard_start_xmit() */ in i2400m_wake_tx_work()
192 ws, i2400m, skb, result); in i2400m_wake_tx_work()
205 void i2400m_tx_prep_header(struct sk_buff *skb) in i2400m_tx_prep_header() argument
208 skb_pull(skb, ETH_HLEN); in i2400m_tx_prep_header()
209 pl_hdr = skb_push(skb, sizeof(*pl_hdr)); in i2400m_tx_prep_header()
261 struct sk_buff *skb) in i2400m_net_wake_tx() argument
267 d_fnstart(3, dev, "(skb %p net_dev %p)\n", skb, net_dev); in i2400m_net_wake_tx()
271 skb, skb->len); in i2400m_net_wake_tx()
272 d_dump(4, dev, skb->data, skb->len); in i2400m_net_wake_tx()
282 i2400m->wake_tx_skb = skb_get(skb); /* transfer ref count */ in i2400m_net_wake_tx()
283 i2400m_tx_prep_header(skb); in i2400m_net_wake_tx()
296 skb, netif_queue_stopped(net_dev)); in i2400m_net_wake_tx()
299 d_fnend(3, dev, "(skb %p net_dev %p) = %d\n", skb, net_dev, result); in i2400m_net_wake_tx()
314 struct sk_buff *skb) in i2400m_net_tx() argument
320 i2400m, net_dev, skb); in i2400m_net_tx()
323 i2400m_tx_prep_header(skb); in i2400m_net_tx()
325 skb, skb->len); in i2400m_net_tx()
326 d_dump(4, dev, skb->data, skb->len); in i2400m_net_tx()
327 result = i2400m_tx(i2400m, skb->data, skb->len, I2400M_PT_DATA); in i2400m_net_tx()
329 i2400m, net_dev, skb, result); in i2400m_net_tx()
353 netdev_tx_t i2400m_hard_start_xmit(struct sk_buff *skb, in i2400m_hard_start_xmit() argument
360 d_fnstart(3, dev, "(skb %p net_dev %p)\n", skb, net_dev); in i2400m_hard_start_xmit()
362 if (skb_cow_head(skb, 0)) in i2400m_hard_start_xmit()
366 result = i2400m_net_wake_tx(i2400m, net_dev, skb); in i2400m_hard_start_xmit()
368 result = i2400m_net_tx(i2400m, net_dev, skb); in i2400m_hard_start_xmit()
374 net_dev->stats.tx_bytes += skb->len; in i2400m_hard_start_xmit()
376 dev_kfree_skb(skb); in i2400m_hard_start_xmit()
377 d_fnend(3, dev, "(skb %p net_dev %p) = %d\n", skb, net_dev, result); in i2400m_hard_start_xmit()
458 struct sk_buff *skb; in i2400m_net_rx() local
463 skb = skb_get(skb_rx); in i2400m_net_rx()
464 d_printf(2, dev, "RX: reusing first payload skb %p\n", skb); in i2400m_net_rx()
465 skb_pull(skb, buf - (void *) skb->data); in i2400m_net_rx()
466 skb_trim(skb, (void *) skb_end_pointer(skb) - buf); in i2400m_net_rx()
470 skb = __netdev_alloc_skb(net_dev, buf_len, GFP_KERNEL); in i2400m_net_rx()
471 if (skb == NULL) { in i2400m_net_rx()
476 skb_put_data(skb, buf, buf_len); in i2400m_net_rx()
479 skb->data - ETH_HLEN, in i2400m_net_rx()
481 skb_set_mac_header(skb, -ETH_HLEN); in i2400m_net_rx()
482 skb->dev = i2400m->wimax_dev.net_dev; in i2400m_net_rx()
483 skb->protocol = htons(ETH_P_IP); in i2400m_net_rx()
489 netif_rx_ni(skb); /* see notes in function header */ in i2400m_net_rx()
518 void i2400m_net_erx(struct i2400m *i2400m, struct sk_buff *skb, in i2400m_net_erx() argument
525 i2400m, skb, skb->len, cs); in i2400m_net_erx()
530 skb->data - ETH_HLEN, in i2400m_net_erx()
532 skb_set_mac_header(skb, -ETH_HLEN); in i2400m_net_erx()
533 skb->dev = i2400m->wimax_dev.net_dev; in i2400m_net_erx()
534 skb->protocol = htons(ETH_P_IP); in i2400m_net_erx()
536 net_dev->stats.rx_bytes += skb->len; in i2400m_net_erx()
544 skb->len); in i2400m_net_erx()
545 d_dump(4, dev, skb->data, skb->len); in i2400m_net_erx()
546 netif_rx_ni(skb); /* see notes in function header */ in i2400m_net_erx()
549 i2400m, skb, skb->len, cs); in i2400m_net_erx()