Home
last modified time | relevance | path

Searched refs:skbn (Results 1 – 18 of 18) sorted by relevance

/net/netrom/
Dnr_out.c37 struct sk_buff *skbn; in nr_output() local
49 if ((skbn = sock_alloc_send_skb(sk, frontlen + NR_MAX_PACKET_SIZE, 0, &err)) == NULL) in nr_output()
52 skb_reserve(skbn, frontlen); in nr_output()
57 skb_copy_from_linear_data(skb, skb_put(skbn, len), len); in nr_output()
61 skb_push(skbn, NR_TRANSPORT_LEN); in nr_output()
62 skb_copy_to_linear_data(skbn, transport, in nr_output()
65 skbn->data[4] |= NR_MORE_FLAG; in nr_output()
67 skb_queue_tail(&sk->sk_write_queue, skbn); /* Throw it on the queue */ in nr_output()
102 struct sk_buff *skb, *skbn; in nr_send_nak_frame() local
108 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) in nr_send_nak_frame()
[all …]
Dnr_in.c34 struct sk_buff *skbo, *skbn = skb; in nr_queue_rx_frame() local
51 if ((skbn = alloc_skb(nr->fraglen, GFP_ATOMIC)) == NULL) in nr_queue_rx_frame()
54 skb_reset_transport_header(skbn); in nr_queue_rx_frame()
58 skb_put(skbn, skbo->len), in nr_queue_rx_frame()
66 return sock_queue_rcv_skb(sk, skbn); in nr_queue_rx_frame()
153 struct sk_buff *skbn; in nr_state3_machine() local
236 while ((skbn = skb_dequeue(&nrom->reseq_queue)) != NULL) { in nr_state3_machine()
237 ns = skbn->data[17]; in nr_state3_machine()
239 if (nr_queue_rx_frame(sk, skbn, frametype & NR_MORE_FLAG) == 0) { in nr_state3_machine()
243 skb_queue_tail(&temp_queue, skbn); in nr_state3_machine()
[all …]
Dnr_loopback.c35 struct sk_buff *skbn; in nr_loopback_queue() local
37 if ((skbn = alloc_skb(skb->len, GFP_ATOMIC)) != NULL) { in nr_loopback_queue()
38 skb_copy_from_linear_data(skb, skb_put(skbn, skb->len), skb->len); in nr_loopback_queue()
39 skb_reset_transport_header(skbn); in nr_loopback_queue()
41 skb_queue_tail(&loopback_queue, skbn); in nr_loopback_queue()
Dnr_subr.c216 struct sk_buff *skbn; in __nr_transmit_reply() local
222 if ((skbn = alloc_skb(len, GFP_ATOMIC)) == NULL) in __nr_transmit_reply()
225 skb_reserve(skbn, 0); in __nr_transmit_reply()
227 dptr = skb_put(skbn, NR_NETWORK_LEN + NR_TRANSPORT_LEN); in __nr_transmit_reply()
258 if (!nr_route_frame(skbn, NULL)) in __nr_transmit_reply()
259 kfree_skb(skbn); in __nr_transmit_reply()
Dnr_route.c784 struct sk_buff *skbn; in nr_route_frame() local
837 if ((skbn=skb_copy_expand(skb, dev->hard_header_len, 0, GFP_ATOMIC)) == NULL) { in nr_route_frame()
844 skb=skbn; in nr_route_frame()
/net/ax25/
Dax25_out.c122 struct sk_buff *skbn; in ax25_output() local
148 if ((skbn = alloc_skb(paclen + 2 + frontlen, GFP_ATOMIC)) == NULL) { in ax25_output()
155 skb_set_owner_w(skbn, skb->sk); in ax25_output()
162 skb_reserve(skbn, frontlen + 2); in ax25_output()
163 skb_set_network_header(skbn, in ax25_output()
165 skb_copy_from_linear_data(skb, skb_put(skbn, len), len); in ax25_output()
166 p = skb_push(skbn, 2); in ax25_output()
176 skb_reserve(skbn, frontlen + 1); in ax25_output()
177 skb_set_network_header(skbn, in ax25_output()
179 skb_copy_from_linear_data(skb, skb_put(skbn, len), len); in ax25_output()
[all …]
Dax25_in.c40 struct sk_buff *skbn, *skbo; in ax25_rx_fragment() local
53 skbn = alloc_skb(AX25_MAX_HEADER_LEN + in ax25_rx_fragment()
56 if (!skbn) { in ax25_rx_fragment()
61 skb_reserve(skbn, AX25_MAX_HEADER_LEN); in ax25_rx_fragment()
63 skbn->dev = ax25->ax25_dev->dev; in ax25_rx_fragment()
64 skb_reset_network_header(skbn); in ax25_rx_fragment()
65 skb_reset_transport_header(skbn); in ax25_rx_fragment()
70 skb_put(skbn, skbo->len), in ax25_rx_fragment()
77 if (ax25_rx_iframe(ax25, skbn) == 0) in ax25_rx_fragment()
78 kfree_skb(skbn); in ax25_rx_fragment()
[all …]
Dax25_route.c462 struct sk_buff *skbn; in ax25_rt_build_path() local
469 if ((skbn = skb_realloc_headroom(skb, len)) == NULL) { in ax25_rt_build_path()
475 skb_set_owner_w(skbn, skb->sk); in ax25_rt_build_path()
479 skb = skbn; in ax25_rt_build_path()
/net/x25/
Dx25_out.c54 struct sk_buff *skbn; in x25_output() local
72 skbn = sock_alloc_send_skb(sk, frontlen + max_len, in x25_output()
75 if (!skbn) { in x25_output()
86 skb_reserve(skbn, frontlen); in x25_output()
91 skb_copy_from_linear_data(skb, skb_put(skbn, len), len); in x25_output()
95 skb_push(skbn, header_len); in x25_output()
96 skb_copy_to_linear_data(skbn, header, header_len); in x25_output()
100 skbn->data[3] |= X25_EXT_M_BIT; in x25_output()
102 skbn->data[2] |= X25_STD_M_BIT; in x25_output()
105 skb_queue_tail(&sk->sk_write_queue, skbn); in x25_output()
[all …]
Dx25_forward.c29 struct sk_buff *skbn; in x25_forward_call() local
79 if ( (skbn = skb_clone(skb, GFP_ATOMIC)) == NULL){ in x25_forward_call()
82 x25_transmit_link(skbn, neigh_new); in x25_forward_call()
103 struct sk_buff *skbn; in x25_forward_data() local
124 if ( (skbn = pskb_copy(skb, GFP_ATOMIC)) == NULL){ in x25_forward_data()
128 x25_transmit_link(skbn, nb); in x25_forward_data()
Dx25_in.c39 struct sk_buff *skbo, *skbn = skb; in x25_queue_rx_frame() local
52 if ((skbn = alloc_skb(len, GFP_ATOMIC)) == NULL){ in x25_queue_rx_frame()
59 skb_reset_transport_header(skbn); in x25_queue_rx_frame()
62 skb_copy_from_linear_data(skbo, skb_put(skbn, skbo->len), in x25_queue_rx_frame()
71 skb_put(skbn, skbo->len), in x25_queue_rx_frame()
79 skb_set_owner_r(skbn, sk); in x25_queue_rx_frame()
80 skb_queue_tail(&sk->sk_receive_queue, skbn); in x25_queue_rx_frame()
Dx25_link.c77 struct sk_buff *skbn; in x25_link_control() local
110 while ((skbn = skb_dequeue(&nb->queue)) != NULL) in x25_link_control()
111 x25_send_frame(skbn, nb); in x25_link_control()
/net/rose/
Drose_dev.c62 struct sk_buff *skbn; in rose_rebuild_header() local
69 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) { in rose_rebuild_header()
75 skb_set_owner_w(skbn, skb->sk); in rose_rebuild_header()
79 len = skbn->len; in rose_rebuild_header()
81 if (!rose_route_frame(skbn, NULL)) { in rose_rebuild_header()
82 kfree_skb(skbn); in rose_rebuild_header()
Drose_out.c51 struct sk_buff *skb, *skbn; in rose_kick() local
79 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) { in rose_kick()
84 skb_set_owner_w(skbn, sk); in rose_kick()
89 rose_send_iframe(sk, skbn); in rose_kick()
Drose_loopback.c37 struct sk_buff *skbn; in rose_loopback_queue() local
39 skbn = skb_clone(skb, GFP_ATOMIC); in rose_loopback_queue()
43 if (skbn != NULL) { in rose_loopback_queue()
44 skb_queue_tail(&loopback_queue, skbn); in rose_loopback_queue()
Drose_link.c147 struct sk_buff *skbn; in rose_link_rx_restart() local
173 while ((skbn = skb_dequeue(&neigh->queue)) != NULL) in rose_link_rx_restart()
174 if (!rose_send_frame(skbn, neigh)) in rose_link_rx_restart()
175 kfree_skb(skbn); in rose_link_rx_restart()
Daf_rose.c1162 struct sk_buff *skbn; in rose_sendmsg() local
1173 if ((skbn = sock_alloc_send_skb(sk, frontlen + ROSE_PACLEN, 0, &err)) == NULL) { in rose_sendmsg()
1178 skbn->sk = sk; in rose_sendmsg()
1179 skbn->free = 1; in rose_sendmsg()
1180 skbn->arp = 1; in rose_sendmsg()
1182 skb_reserve(skbn, frontlen); in rose_sendmsg()
1187 skb_copy_from_linear_data(skb, skb_put(skbn, lg), lg); in rose_sendmsg()
1191 skb_push(skbn, ROSE_MIN_LEN); in rose_sendmsg()
1192 skb_copy_to_linear_data(skbn, header, ROSE_MIN_LEN); in rose_sendmsg()
1195 skbn->data[2] |= M_BIT; in rose_sendmsg()
[all …]
/net/lapb/
Dlapb_out.c73 struct sk_buff *skb, *skbn; in lapb_kick() local
90 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) { in lapb_kick()
96 skb_set_owner_w(skbn, skb->sk); in lapb_kick()
101 lapb_send_iframe(lapb, skbn, LAPB_POLLOFF); in lapb_kick()