Home
last modified time | relevance | path

Searched refs:buff (Results 1 – 24 of 24) sorted by relevance

/net/ax25/
Dax25_ip.c49 unsigned char *buff; in ax25_hard_header() local
56 buff = skb_push(skb, AX25_HEADER_LEN); in ax25_hard_header()
57 *buff++ = 0x00; /* KISS DATA */ in ax25_hard_header()
60 memcpy(buff, daddr, dev->addr_len); /* Address specified */ in ax25_hard_header()
62 buff[6] &= ~AX25_CBIT; in ax25_hard_header()
63 buff[6] &= ~AX25_EBIT; in ax25_hard_header()
64 buff[6] |= AX25_SSSID_SPARE; in ax25_hard_header()
65 buff += AX25_ADDR_LEN; in ax25_hard_header()
68 memcpy(buff, saddr, dev->addr_len); in ax25_hard_header()
70 memcpy(buff, dev->dev_addr, dev->addr_len); in ax25_hard_header()
[all …]
/net/netrom/
Dnr_dev.c69 unsigned char *buff = skb_push(skb, NR_NETWORK_LEN + NR_TRANSPORT_LEN); in nr_header() local
71 memcpy(buff, (saddr != NULL) ? saddr : dev->dev_addr, dev->addr_len); in nr_header()
72 buff[6] &= ~AX25_CBIT; in nr_header()
73 buff[6] &= ~AX25_EBIT; in nr_header()
74 buff[6] |= AX25_SSSID_SPARE; in nr_header()
75 buff += AX25_ADDR_LEN; in nr_header()
78 memcpy(buff, daddr, dev->addr_len); in nr_header()
79 buff[6] &= ~AX25_CBIT; in nr_header()
80 buff[6] |= AX25_EBIT; in nr_header()
81 buff[6] |= AX25_SSSID_SPARE; in nr_header()
[all …]
/net/ipv4/
Dproc.c408 unsigned long buff[TCPUDP_MIB_MAX]; in snmp_seq_show_tcp_udp() local
412 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp()
419 snmp_get_cpu_field_batch(buff, snmp4_tcp_list, in snmp_seq_show_tcp_udp()
424 seq_printf(seq, " %ld", buff[i]); in snmp_seq_show_tcp_udp()
426 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp()
429 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp()
431 snmp_get_cpu_field_batch(buff, snmp4_udp_list, in snmp_seq_show_tcp_udp()
438 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp()
440 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp()
444 snmp_get_cpu_field_batch(buff, snmp4_udp_list, in snmp_seq_show_tcp_udp()
[all …]
Dtcp_output.c1536 struct sk_buff *buff, in tcp_insert_write_queue_after() argument
1541 __skb_queue_after(&sk->sk_write_queue, skb, buff); in tcp_insert_write_queue_after()
1543 tcp_rbtree_insert(&sk->tcp_rtx_queue, buff); in tcp_insert_write_queue_after()
1556 struct sk_buff *buff; in tcp_fragment() local
1585 buff = tcp_stream_alloc_skb(sk, gfp, true); in tcp_fragment()
1586 if (!buff) in tcp_fragment()
1588 skb_copy_decrypted(buff, skb); in tcp_fragment()
1589 mptcp_skb_ext_copy(buff, skb); in tcp_fragment()
1591 sk_wmem_queued_add(sk, buff->truesize); in tcp_fragment()
1592 sk_mem_charge(sk, buff->truesize); in tcp_fragment()
[all …]
Dicmp.c1039 char buff[IFNAMSIZ]; in icmp_build_probe() local
1079 memset(buff, 0, sizeof(buff)); in icmp_build_probe()
1080 memcpy(buff, &iio->ident.name, ident_len); in icmp_build_probe()
1081 dev = dev_get_by_name(net, buff); in icmp_build_probe()
/net/rose/
Drose_dev.c39 unsigned char *buff = skb_push(skb, ROSE_MIN_LEN + 2); in rose_header() local
42 memcpy(buff + 7, daddr, dev->addr_len); in rose_header()
44 *buff++ = ROSE_GFI | ROSE_Q_BIT; in rose_header()
45 *buff++ = 0x00; in rose_header()
46 *buff++ = ROSE_DATA; in rose_header()
47 *buff++ = 0x7F; in rose_header()
48 *buff++ = AX25_P_IP; in rose_header()
/net/atm/
Dmpoa_proc.c51 static ssize_t proc_mpc_write(struct file *file, const char __user *buff,
54 static int parse_qos(const char *buff);
206 static ssize_t proc_mpc_write(struct file *file, const char __user *buff, in proc_mpc_write() argument
223 if (get_user(*p, buff++)) { in proc_mpc_write()
242 static int parse_qos(const char *buff) in parse_qos() argument
254 if (sscanf(buff, "del %hhu.%hhu.%hhu.%hhu", in parse_qos()
260 if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=tx", in parse_qos()
264 } else if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=%d,%d", in parse_qos()
Dlec.c131 char *buff; in lec_handle_bridge() local
139 buff = skb->data + skb->dev->hard_header_len; in lec_handle_bridge()
140 if (*buff++ == 0x42 && *buff++ == 0x42 && *buff++ == 0x03) { in lec_handle_bridge()
151 buff += 4; in lec_handle_bridge()
152 mesg->content.normal.flag = *buff & 0x01; in lec_handle_bridge()
Dmpc.c496 char *buff; in send_via_shortcut() local
507 buff = skb->data + mpc->dev->hard_header_len; in send_via_shortcut()
508 iph = (struct iphdr *)buff; in send_via_shortcut()
/net/netfilter/ipvs/
Dip_vs_sync.c545 struct ip_vs_sync_buff *buff; in ip_vs_sync_conn_v0() local
567 buff = ms->sync_buff; in ip_vs_sync_conn_v0()
570 if (buff) { in ip_vs_sync_conn_v0()
571 m = (struct ip_vs_sync_mesg_v0 *) buff->mesg; in ip_vs_sync_conn_v0()
573 if (buff->head + len > buff->end || !m->nr_conns) { in ip_vs_sync_conn_v0()
576 buff = NULL; in ip_vs_sync_conn_v0()
579 if (!buff) { in ip_vs_sync_conn_v0()
580 buff = ip_vs_sync_buff_create_v0(ipvs, len); in ip_vs_sync_conn_v0()
581 if (!buff) { in ip_vs_sync_conn_v0()
586 ms->sync_buff = buff; in ip_vs_sync_conn_v0()
[all …]
/net/xfrm/
Dxfrm_proc.c49 unsigned long buff[LINUX_MIB_XFRMMAX]; in xfrm_statistics_seq_show() local
53 memset(buff, 0, sizeof(unsigned long) * LINUX_MIB_XFRMMAX); in xfrm_statistics_seq_show()
55 snmp_get_cpu_field_batch(buff, xfrm_mib_list, in xfrm_statistics_seq_show()
59 buff[i]); in xfrm_statistics_seq_show()
/net/netfilter/
Dnft_exthdr.c194 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_eval() local
201 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_eval()
238 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_set_eval() local
244 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_set_eval()
318 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_strip_eval() local
324 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_strip_eval()
Dnf_conntrack_proto_tcp.c338 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_options() local
346 length, buff); in tcp_options()
394 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_sack() local
403 length, buff); in tcp_sack()
/net/ipv6/
Dproc.c187 unsigned long buff[SNMP_MIB_MAX]; in snmp6_seq_show_item() local
191 memset(buff, 0, sizeof(unsigned long) * SNMP_MIB_MAX); in snmp6_seq_show_item()
193 snmp_get_cpu_field_batch(buff, itemlist, pcpumib); in snmp6_seq_show_item()
196 itemlist[i].name, buff[i]); in snmp6_seq_show_item()
Dtcp_ipv6.c851 struct sk_buff *buff; in tcp_v6_send_response() local
876 buff = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC); in tcp_v6_send_response()
877 if (!buff) in tcp_v6_send_response()
880 skb_reserve(buff, MAX_TCP_HEADER); in tcp_v6_send_response()
882 t1 = skb_push(buff, tot_len); in tcp_v6_send_response()
883 skb_reset_transport_header(buff); in tcp_v6_send_response()
923 buff->ip_summed = CHECKSUM_PARTIAL; in tcp_v6_send_response()
925 __tcp_v6_send_check(buff, &fl6.saddr, &fl6.daddr); in tcp_v6_send_response()
942 skb_set_delivery_time(buff, tcp_transmit_time(sk), true); in tcp_v6_send_response()
946 skb_set_hash(buff, txhash, PKT_HASH_TYPE_L4); in tcp_v6_send_response()
[all …]
Dndisc.c1671 struct sk_buff *buff; in ndisc_send_redirect() local
1755 buff = ndisc_alloc_skb(dev, sizeof(*msg) + optlen); in ndisc_send_redirect()
1756 if (!buff) in ndisc_send_redirect()
1759 msg = skb_put(buff, sizeof(*msg)); in ndisc_send_redirect()
1773 ndisc_fill_redirect_addr_option(buff, ha, ops_data); in ndisc_send_redirect()
1780 ndisc_fill_redirect_hdr_option(buff, skb, rd_len); in ndisc_send_redirect()
1782 skb_dst_set(buff, dst); in ndisc_send_redirect()
1783 ndisc_send_skb(buff, &ipv6_hdr(skb)->saddr, &saddr_buf); in ndisc_send_redirect()
Daddrconf.c5727 u64 buff[IPSTATS_MIB_MAX]; in __snmp6_fill_stats64() local
5732 memset(buff, 0, sizeof(buff)); in __snmp6_fill_stats64()
5733 buff[0] = IPSTATS_MIB_MAX; in __snmp6_fill_stats64()
5737 buff[i] += snmp_get_cpu_field64(mib, c, i, syncpoff); in __snmp6_fill_stats64()
5740 memcpy(stats, buff, IPSTATS_MIB_MAX * sizeof(u64)); in __snmp6_fill_stats64()
/net/sctp/
Dproc.c61 unsigned long buff[SCTP_MIB_MAX]; in sctp_snmp_seq_show() local
65 memset(buff, 0, sizeof(unsigned long) * SCTP_MIB_MAX); in sctp_snmp_seq_show()
67 snmp_get_cpu_field_batch(buff, sctp_snmp_list, in sctp_snmp_seq_show()
71 buff[i]); in sctp_snmp_seq_show()
/net/xdp/
Dxsk_buff_pool.c623 struct xdp_buff *buff; in xp_alloc_batch() local
626 buff = xp_alloc(pool); in xp_alloc_batch()
627 if (buff) in xp_alloc_batch()
628 *xdp = buff; in xp_alloc_batch()
629 return !!buff; in xp_alloc_batch()
/net/
Dsocket.c2166 int __sys_sendto(int fd, void __user *buff, size_t len, unsigned int flags, in __sys_sendto() argument
2176 err = import_single_range(ITER_SOURCE, buff, len, &iov, &msg.msg_iter); in __sys_sendto()
2207 SYSCALL_DEFINE6(sendto, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE6() argument
2211 return __sys_sendto(fd, buff, len, flags, addr, addr_len); in SYSCALL_DEFINE6()
2218 SYSCALL_DEFINE4(send, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE4() argument
2221 return __sys_sendto(fd, buff, len, flags, NULL, 0); in SYSCALL_DEFINE4()
/net/packet/
Daf_packet.c197 static void packet_increment_head(struct packet_ring_buffer *buff);
1259 static void packet_increment_head(struct packet_ring_buffer *buff) in packet_increment_head() argument
1261 buff->head = buff->head != buff->frame_max ? buff->head+1 : 0; in packet_increment_head()
/net/core/
Dneighbour.c1225 struct sk_buff *buff; in __neigh_event_send() local
1227 buff = __skb_dequeue(&neigh->arp_queue); in __neigh_event_send()
1228 if (!buff) in __neigh_event_send()
1230 neigh->arp_queue_len_bytes -= buff->truesize; in __neigh_event_send()
1231 kfree_skb_reason(buff, SKB_DROP_REASON_NEIGH_QUEUEFULL); in __neigh_event_send()
Dskbuff.c3528 static __wsum warn_crc32c_csum_update(const void *buff, int len, __wsum sum) in warn_crc32c_csum_update() argument
Dfilter.c1653 u8 buff[MAX_BPF_STACK]; member