Home
last modified time | relevance | path

Searched refs:buff (Results 1 – 25 of 28) sorted by relevance

12

/net/ax25/
Dax25_ip.c49 unsigned char *buff; in ax25_hard_header() local
56 buff = skb_push(skb, AX25_HEADER_LEN); in ax25_hard_header()
57 *buff++ = 0x00; /* KISS DATA */ in ax25_hard_header()
60 memcpy(buff, daddr, dev->addr_len); /* Address specified */ in ax25_hard_header()
62 buff[6] &= ~AX25_CBIT; in ax25_hard_header()
63 buff[6] &= ~AX25_EBIT; in ax25_hard_header()
64 buff[6] |= AX25_SSSID_SPARE; in ax25_hard_header()
65 buff += AX25_ADDR_LEN; in ax25_hard_header()
68 memcpy(buff, saddr, dev->addr_len); in ax25_hard_header()
70 memcpy(buff, dev->dev_addr, dev->addr_len); in ax25_hard_header()
[all …]
/net/netrom/
Dnr_dev.c69 unsigned char *buff = skb_push(skb, NR_NETWORK_LEN + NR_TRANSPORT_LEN); in nr_header() local
71 memcpy(buff, (saddr != NULL) ? saddr : dev->dev_addr, dev->addr_len); in nr_header()
72 buff[6] &= ~AX25_CBIT; in nr_header()
73 buff[6] &= ~AX25_EBIT; in nr_header()
74 buff[6] |= AX25_SSSID_SPARE; in nr_header()
75 buff += AX25_ADDR_LEN; in nr_header()
78 memcpy(buff, daddr, dev->addr_len); in nr_header()
79 buff[6] &= ~AX25_CBIT; in nr_header()
80 buff[6] |= AX25_EBIT; in nr_header()
81 buff[6] |= AX25_SSSID_SPARE; in nr_header()
[all …]
/net/batman-adv/
Dgateway_common.c36 bool batadv_parse_throughput(struct net_device *net_dev, char *buff, in batadv_parse_throughput() argument
44 if (strlen(buff) > 4) { in batadv_parse_throughput()
45 tmp_ptr = buff + strlen(buff) - 4; in batadv_parse_throughput()
55 ret = kstrtou64(buff, 10, &lthroughput); in batadv_parse_throughput()
59 description, buff); in batadv_parse_throughput()
69 description, buff); in batadv_parse_throughput()
84 description, buff); in batadv_parse_throughput()
103 static bool batadv_parse_gw_bandwidth(struct net_device *net_dev, char *buff, in batadv_parse_gw_bandwidth() argument
109 slash_ptr = strchr(buff, '/'); in batadv_parse_gw_bandwidth()
113 ret = batadv_parse_throughput(net_dev, buff, "download gateway speed", in batadv_parse_gw_bandwidth()
[all …]
Dgateway_common.h30 ssize_t batadv_gw_bandwidth_set(struct net_device *net_dev, char *buff,
35 bool batadv_parse_throughput(struct net_device *net_dev, char *buff,
Dbat_v.c516 char *buff, size_t count) in batadv_v_store_sel_class() argument
520 if (!batadv_parse_throughput(bat_priv->soft_iface, buff, in batadv_v_store_sel_class()
Dtypes.h2203 ssize_t (*store_sel_class)(struct batadv_priv *bat_priv, char *buff,
/net/ipv4/
Dproc.c404 unsigned long buff[TCPUDP_MIB_MAX]; in snmp_seq_show_tcp_udp() local
408 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp()
415 snmp_get_cpu_field_batch(buff, snmp4_tcp_list, in snmp_seq_show_tcp_udp()
420 seq_printf(seq, " %ld", buff[i]); in snmp_seq_show_tcp_udp()
422 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp()
425 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp()
427 snmp_get_cpu_field_batch(buff, snmp4_udp_list, in snmp_seq_show_tcp_udp()
434 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp()
436 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp()
440 snmp_get_cpu_field_batch(buff, snmp4_udp_list, in snmp_seq_show_tcp_udp()
[all …]
Dtcp_output.c1514 struct sk_buff *buff, in tcp_insert_write_queue_after() argument
1519 __skb_queue_after(&sk->sk_write_queue, skb, buff); in tcp_insert_write_queue_after()
1521 tcp_rbtree_insert(&sk->tcp_rtx_queue, buff); in tcp_insert_write_queue_after()
1534 struct sk_buff *buff; in tcp_fragment() local
1565 buff = tcp_stream_alloc_skb(sk, nsize, gfp, true); in tcp_fragment()
1566 if (!buff) in tcp_fragment()
1568 skb_copy_decrypted(buff, skb); in tcp_fragment()
1569 mptcp_skb_ext_copy(buff, skb); in tcp_fragment()
1571 sk_wmem_queued_add(sk, buff->truesize); in tcp_fragment()
1572 sk_mem_charge(sk, buff->truesize); in tcp_fragment()
[all …]
Dicmp.c1033 char buff[IFNAMSIZ]; in icmp_build_probe() local
1073 memset(buff, 0, sizeof(buff)); in icmp_build_probe()
1074 memcpy(buff, &iio->ident.name, ident_len); in icmp_build_probe()
1075 dev = dev_get_by_name(net, buff); in icmp_build_probe()
/net/rose/
Drose_dev.c39 unsigned char *buff = skb_push(skb, ROSE_MIN_LEN + 2); in rose_header() local
42 memcpy(buff + 7, daddr, dev->addr_len); in rose_header()
44 *buff++ = ROSE_GFI | ROSE_Q_BIT; in rose_header()
45 *buff++ = 0x00; in rose_header()
46 *buff++ = ROSE_DATA; in rose_header()
47 *buff++ = 0x7F; in rose_header()
48 *buff++ = AX25_P_IP; in rose_header()
/net/atm/
Dmpoa_proc.c51 static ssize_t proc_mpc_write(struct file *file, const char __user *buff,
54 static int parse_qos(const char *buff);
206 static ssize_t proc_mpc_write(struct file *file, const char __user *buff, in proc_mpc_write() argument
223 if (get_user(*p, buff++)) { in proc_mpc_write()
242 static int parse_qos(const char *buff) in parse_qos() argument
254 if (sscanf(buff, "del %hhu.%hhu.%hhu.%hhu", in parse_qos()
260 if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=tx", in parse_qos()
264 } else if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=%d,%d", in parse_qos()
Dlec.c131 char *buff; in lec_handle_bridge() local
139 buff = skb->data + skb->dev->hard_header_len; in lec_handle_bridge()
140 if (*buff++ == 0x42 && *buff++ == 0x42 && *buff++ == 0x03) { in lec_handle_bridge()
151 buff += 4; in lec_handle_bridge()
152 mesg->content.normal.flag = *buff & 0x01; in lec_handle_bridge()
Dmpc.c496 char *buff; in send_via_shortcut() local
507 buff = skb->data + mpc->dev->hard_header_len; in send_via_shortcut()
508 iph = (struct iphdr *)buff; in send_via_shortcut()
/net/netfilter/ipvs/
Dip_vs_sync.c545 struct ip_vs_sync_buff *buff; in ip_vs_sync_conn_v0() local
567 buff = ms->sync_buff; in ip_vs_sync_conn_v0()
570 if (buff) { in ip_vs_sync_conn_v0()
571 m = (struct ip_vs_sync_mesg_v0 *) buff->mesg; in ip_vs_sync_conn_v0()
573 if (buff->head + len > buff->end || !m->nr_conns) { in ip_vs_sync_conn_v0()
576 buff = NULL; in ip_vs_sync_conn_v0()
579 if (!buff) { in ip_vs_sync_conn_v0()
580 buff = ip_vs_sync_buff_create_v0(ipvs, len); in ip_vs_sync_conn_v0()
581 if (!buff) { in ip_vs_sync_conn_v0()
586 ms->sync_buff = buff; in ip_vs_sync_conn_v0()
[all …]
/net/xfrm/
Dxfrm_proc.c49 unsigned long buff[LINUX_MIB_XFRMMAX]; in xfrm_statistics_seq_show() local
53 memset(buff, 0, sizeof(unsigned long) * LINUX_MIB_XFRMMAX); in xfrm_statistics_seq_show()
55 snmp_get_cpu_field_batch(buff, xfrm_mib_list, in xfrm_statistics_seq_show()
59 buff[i]); in xfrm_statistics_seq_show()
/net/netfilter/
Dnft_exthdr.c195 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_eval() local
202 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_eval()
239 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_set_eval() local
245 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_set_eval()
319 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_strip_eval() local
325 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_strip_eval()
Dnf_conntrack_proto_tcp.c338 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_options() local
346 length, buff); in tcp_options()
394 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_sack() local
403 length, buff); in tcp_sack()
/net/ipv6/
Dproc.c185 unsigned long buff[SNMP_MIB_MAX]; in snmp6_seq_show_item() local
189 memset(buff, 0, sizeof(unsigned long) * SNMP_MIB_MAX); in snmp6_seq_show_item()
191 snmp_get_cpu_field_batch(buff, itemlist, pcpumib); in snmp6_seq_show_item()
194 itemlist[i].name, buff[i]); in snmp6_seq_show_item()
Dtcp_ipv6.c856 struct sk_buff *buff; in tcp_v6_send_response() local
881 buff = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC); in tcp_v6_send_response()
882 if (!buff) in tcp_v6_send_response()
885 skb_reserve(buff, MAX_TCP_HEADER); in tcp_v6_send_response()
887 t1 = skb_push(buff, tot_len); in tcp_v6_send_response()
888 skb_reset_transport_header(buff); in tcp_v6_send_response()
928 buff->ip_summed = CHECKSUM_PARTIAL; in tcp_v6_send_response()
930 __tcp_v6_send_check(buff, &fl6.saddr, &fl6.daddr); in tcp_v6_send_response()
947 skb_set_delivery_time(buff, tcp_transmit_time(sk), true); in tcp_v6_send_response()
951 skb_set_hash(buff, txhash, PKT_HASH_TYPE_L4); in tcp_v6_send_response()
[all …]
Dndisc.c1673 struct sk_buff *buff; in ndisc_send_redirect() local
1757 buff = ndisc_alloc_skb(dev, sizeof(*msg) + optlen); in ndisc_send_redirect()
1758 if (!buff) in ndisc_send_redirect()
1761 msg = skb_put(buff, sizeof(*msg)); in ndisc_send_redirect()
1775 ndisc_fill_redirect_addr_option(buff, ha, ops_data); in ndisc_send_redirect()
1782 ndisc_fill_redirect_hdr_option(buff, skb, rd_len); in ndisc_send_redirect()
1784 skb_dst_set(buff, dst); in ndisc_send_redirect()
1785 ndisc_send_skb(buff, &ipv6_hdr(skb)->saddr, &saddr_buf); in ndisc_send_redirect()
/net/sctp/
Dproc.c61 unsigned long buff[SCTP_MIB_MAX]; in sctp_snmp_seq_show() local
65 memset(buff, 0, sizeof(unsigned long) * SCTP_MIB_MAX); in sctp_snmp_seq_show()
67 snmp_get_cpu_field_batch(buff, sctp_snmp_list, in sctp_snmp_seq_show()
71 buff[i]); in sctp_snmp_seq_show()
/net/xdp/
Dxsk_buff_pool.c608 struct xdp_buff *buff; in xp_alloc_batch() local
611 buff = xp_alloc(pool); in xp_alloc_batch()
612 if (buff) in xp_alloc_batch()
613 *xdp = buff; in xp_alloc_batch()
614 return !!buff; in xp_alloc_batch()
/net/
Dsocket.c2105 int __sys_sendto(int fd, void __user *buff, size_t len, unsigned int flags, in __sys_sendto() argument
2115 err = import_single_range(ITER_SOURCE, buff, len, &iov, &msg.msg_iter); in __sys_sendto()
2146 SYSCALL_DEFINE6(sendto, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE6() argument
2150 return __sys_sendto(fd, buff, len, flags, addr, addr_len); in SYSCALL_DEFINE6()
2157 SYSCALL_DEFINE4(send, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE4() argument
2160 return __sys_sendto(fd, buff, len, flags, NULL, 0); in SYSCALL_DEFINE4()
/net/packet/
Daf_packet.c197 static void packet_increment_head(struct packet_ring_buffer *buff);
1201 static void packet_increment_head(struct packet_ring_buffer *buff) in packet_increment_head() argument
1203 buff->head = buff->head != buff->frame_max ? buff->head+1 : 0; in packet_increment_head()
/net/core/
Dneighbour.c1225 struct sk_buff *buff; in __neigh_event_send() local
1227 buff = __skb_dequeue(&neigh->arp_queue); in __neigh_event_send()
1228 if (!buff) in __neigh_event_send()
1230 neigh->arp_queue_len_bytes -= buff->truesize; in __neigh_event_send()
1231 kfree_skb_reason(buff, SKB_DROP_REASON_NEIGH_QUEUEFULL); in __neigh_event_send()

12