/net/ax25/ |
D | ax25_ip.c | 49 unsigned char *buff; in ax25_hard_header() local 56 buff = skb_push(skb, AX25_HEADER_LEN); in ax25_hard_header() 57 *buff++ = 0x00; /* KISS DATA */ in ax25_hard_header() 60 memcpy(buff, daddr, dev->addr_len); /* Address specified */ in ax25_hard_header() 62 buff[6] &= ~AX25_CBIT; in ax25_hard_header() 63 buff[6] &= ~AX25_EBIT; in ax25_hard_header() 64 buff[6] |= AX25_SSSID_SPARE; in ax25_hard_header() 65 buff += AX25_ADDR_LEN; in ax25_hard_header() 68 memcpy(buff, saddr, dev->addr_len); in ax25_hard_header() 70 memcpy(buff, dev->dev_addr, dev->addr_len); in ax25_hard_header() [all …]
|
/net/netrom/ |
D | nr_dev.c | 69 unsigned char *buff = skb_push(skb, NR_NETWORK_LEN + NR_TRANSPORT_LEN); in nr_header() local 71 memcpy(buff, (saddr != NULL) ? saddr : dev->dev_addr, dev->addr_len); in nr_header() 72 buff[6] &= ~AX25_CBIT; in nr_header() 73 buff[6] &= ~AX25_EBIT; in nr_header() 74 buff[6] |= AX25_SSSID_SPARE; in nr_header() 75 buff += AX25_ADDR_LEN; in nr_header() 78 memcpy(buff, daddr, dev->addr_len); in nr_header() 79 buff[6] &= ~AX25_CBIT; in nr_header() 80 buff[6] |= AX25_EBIT; in nr_header() 81 buff[6] |= AX25_SSSID_SPARE; in nr_header() [all …]
|
/net/ipv4/ |
D | proc.c | 408 unsigned long buff[TCPUDP_MIB_MAX]; in snmp_seq_show_tcp_udp() local 412 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp() 419 snmp_get_cpu_field_batch(buff, snmp4_tcp_list, in snmp_seq_show_tcp_udp() 424 seq_printf(seq, " %ld", buff[i]); in snmp_seq_show_tcp_udp() 426 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp() 429 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp() 431 snmp_get_cpu_field_batch(buff, snmp4_udp_list, in snmp_seq_show_tcp_udp() 438 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp() 440 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp() 444 snmp_get_cpu_field_batch(buff, snmp4_udp_list, in snmp_seq_show_tcp_udp() [all …]
|
D | tcp_output.c | 1536 struct sk_buff *buff, in tcp_insert_write_queue_after() argument 1541 __skb_queue_after(&sk->sk_write_queue, skb, buff); in tcp_insert_write_queue_after() 1543 tcp_rbtree_insert(&sk->tcp_rtx_queue, buff); in tcp_insert_write_queue_after() 1556 struct sk_buff *buff; in tcp_fragment() local 1585 buff = tcp_stream_alloc_skb(sk, gfp, true); in tcp_fragment() 1586 if (!buff) in tcp_fragment() 1588 skb_copy_decrypted(buff, skb); in tcp_fragment() 1589 mptcp_skb_ext_copy(buff, skb); in tcp_fragment() 1591 sk_wmem_queued_add(sk, buff->truesize); in tcp_fragment() 1592 sk_mem_charge(sk, buff->truesize); in tcp_fragment() [all …]
|
D | icmp.c | 1039 char buff[IFNAMSIZ]; in icmp_build_probe() local 1079 memset(buff, 0, sizeof(buff)); in icmp_build_probe() 1080 memcpy(buff, &iio->ident.name, ident_len); in icmp_build_probe() 1081 dev = dev_get_by_name(net, buff); in icmp_build_probe()
|
/net/rose/ |
D | rose_dev.c | 39 unsigned char *buff = skb_push(skb, ROSE_MIN_LEN + 2); in rose_header() local 42 memcpy(buff + 7, daddr, dev->addr_len); in rose_header() 44 *buff++ = ROSE_GFI | ROSE_Q_BIT; in rose_header() 45 *buff++ = 0x00; in rose_header() 46 *buff++ = ROSE_DATA; in rose_header() 47 *buff++ = 0x7F; in rose_header() 48 *buff++ = AX25_P_IP; in rose_header()
|
/net/atm/ |
D | mpoa_proc.c | 51 static ssize_t proc_mpc_write(struct file *file, const char __user *buff, 54 static int parse_qos(const char *buff); 206 static ssize_t proc_mpc_write(struct file *file, const char __user *buff, in proc_mpc_write() argument 223 if (get_user(*p, buff++)) { in proc_mpc_write() 242 static int parse_qos(const char *buff) in parse_qos() argument 254 if (sscanf(buff, "del %hhu.%hhu.%hhu.%hhu", in parse_qos() 260 if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=tx", in parse_qos() 264 } else if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=%d,%d", in parse_qos()
|
D | lec.c | 131 char *buff; in lec_handle_bridge() local 139 buff = skb->data + skb->dev->hard_header_len; in lec_handle_bridge() 140 if (*buff++ == 0x42 && *buff++ == 0x42 && *buff++ == 0x03) { in lec_handle_bridge() 151 buff += 4; in lec_handle_bridge() 152 mesg->content.normal.flag = *buff & 0x01; in lec_handle_bridge()
|
D | mpc.c | 496 char *buff; in send_via_shortcut() local 507 buff = skb->data + mpc->dev->hard_header_len; in send_via_shortcut() 508 iph = (struct iphdr *)buff; in send_via_shortcut()
|
/net/netfilter/ipvs/ |
D | ip_vs_sync.c | 545 struct ip_vs_sync_buff *buff; in ip_vs_sync_conn_v0() local 567 buff = ms->sync_buff; in ip_vs_sync_conn_v0() 570 if (buff) { in ip_vs_sync_conn_v0() 571 m = (struct ip_vs_sync_mesg_v0 *) buff->mesg; in ip_vs_sync_conn_v0() 573 if (buff->head + len > buff->end || !m->nr_conns) { in ip_vs_sync_conn_v0() 576 buff = NULL; in ip_vs_sync_conn_v0() 579 if (!buff) { in ip_vs_sync_conn_v0() 580 buff = ip_vs_sync_buff_create_v0(ipvs, len); in ip_vs_sync_conn_v0() 581 if (!buff) { in ip_vs_sync_conn_v0() 586 ms->sync_buff = buff; in ip_vs_sync_conn_v0() [all …]
|
/net/xfrm/ |
D | xfrm_proc.c | 49 unsigned long buff[LINUX_MIB_XFRMMAX]; in xfrm_statistics_seq_show() local 53 memset(buff, 0, sizeof(unsigned long) * LINUX_MIB_XFRMMAX); in xfrm_statistics_seq_show() 55 snmp_get_cpu_field_batch(buff, xfrm_mib_list, in xfrm_statistics_seq_show() 59 buff[i]); in xfrm_statistics_seq_show()
|
/net/netfilter/ |
D | nft_exthdr.c | 194 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_eval() local 201 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_eval() 238 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_set_eval() local 244 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_set_eval() 318 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_strip_eval() local 324 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_strip_eval()
|
D | nf_conntrack_proto_tcp.c | 338 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_options() local 346 length, buff); in tcp_options() 394 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_sack() local 403 length, buff); in tcp_sack()
|
/net/ipv6/ |
D | proc.c | 187 unsigned long buff[SNMP_MIB_MAX]; in snmp6_seq_show_item() local 191 memset(buff, 0, sizeof(unsigned long) * SNMP_MIB_MAX); in snmp6_seq_show_item() 193 snmp_get_cpu_field_batch(buff, itemlist, pcpumib); in snmp6_seq_show_item() 196 itemlist[i].name, buff[i]); in snmp6_seq_show_item()
|
D | tcp_ipv6.c | 851 struct sk_buff *buff; in tcp_v6_send_response() local 876 buff = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC); in tcp_v6_send_response() 877 if (!buff) in tcp_v6_send_response() 880 skb_reserve(buff, MAX_TCP_HEADER); in tcp_v6_send_response() 882 t1 = skb_push(buff, tot_len); in tcp_v6_send_response() 883 skb_reset_transport_header(buff); in tcp_v6_send_response() 923 buff->ip_summed = CHECKSUM_PARTIAL; in tcp_v6_send_response() 925 __tcp_v6_send_check(buff, &fl6.saddr, &fl6.daddr); in tcp_v6_send_response() 942 skb_set_delivery_time(buff, tcp_transmit_time(sk), true); in tcp_v6_send_response() 946 skb_set_hash(buff, txhash, PKT_HASH_TYPE_L4); in tcp_v6_send_response() [all …]
|
D | ndisc.c | 1671 struct sk_buff *buff; in ndisc_send_redirect() local 1755 buff = ndisc_alloc_skb(dev, sizeof(*msg) + optlen); in ndisc_send_redirect() 1756 if (!buff) in ndisc_send_redirect() 1759 msg = skb_put(buff, sizeof(*msg)); in ndisc_send_redirect() 1773 ndisc_fill_redirect_addr_option(buff, ha, ops_data); in ndisc_send_redirect() 1780 ndisc_fill_redirect_hdr_option(buff, skb, rd_len); in ndisc_send_redirect() 1782 skb_dst_set(buff, dst); in ndisc_send_redirect() 1783 ndisc_send_skb(buff, &ipv6_hdr(skb)->saddr, &saddr_buf); in ndisc_send_redirect()
|
D | addrconf.c | 5727 u64 buff[IPSTATS_MIB_MAX]; in __snmp6_fill_stats64() local 5732 memset(buff, 0, sizeof(buff)); in __snmp6_fill_stats64() 5733 buff[0] = IPSTATS_MIB_MAX; in __snmp6_fill_stats64() 5737 buff[i] += snmp_get_cpu_field64(mib, c, i, syncpoff); in __snmp6_fill_stats64() 5740 memcpy(stats, buff, IPSTATS_MIB_MAX * sizeof(u64)); in __snmp6_fill_stats64()
|
/net/sctp/ |
D | proc.c | 61 unsigned long buff[SCTP_MIB_MAX]; in sctp_snmp_seq_show() local 65 memset(buff, 0, sizeof(unsigned long) * SCTP_MIB_MAX); in sctp_snmp_seq_show() 67 snmp_get_cpu_field_batch(buff, sctp_snmp_list, in sctp_snmp_seq_show() 71 buff[i]); in sctp_snmp_seq_show()
|
/net/xdp/ |
D | xsk_buff_pool.c | 623 struct xdp_buff *buff; in xp_alloc_batch() local 626 buff = xp_alloc(pool); in xp_alloc_batch() 627 if (buff) in xp_alloc_batch() 628 *xdp = buff; in xp_alloc_batch() 629 return !!buff; in xp_alloc_batch()
|
/net/ |
D | socket.c | 2166 int __sys_sendto(int fd, void __user *buff, size_t len, unsigned int flags, in __sys_sendto() argument 2176 err = import_single_range(ITER_SOURCE, buff, len, &iov, &msg.msg_iter); in __sys_sendto() 2207 SYSCALL_DEFINE6(sendto, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE6() argument 2211 return __sys_sendto(fd, buff, len, flags, addr, addr_len); in SYSCALL_DEFINE6() 2218 SYSCALL_DEFINE4(send, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE4() argument 2221 return __sys_sendto(fd, buff, len, flags, NULL, 0); in SYSCALL_DEFINE4()
|
/net/packet/ |
D | af_packet.c | 197 static void packet_increment_head(struct packet_ring_buffer *buff); 1259 static void packet_increment_head(struct packet_ring_buffer *buff) in packet_increment_head() argument 1261 buff->head = buff->head != buff->frame_max ? buff->head+1 : 0; in packet_increment_head()
|
/net/core/ |
D | neighbour.c | 1225 struct sk_buff *buff; in __neigh_event_send() local 1227 buff = __skb_dequeue(&neigh->arp_queue); in __neigh_event_send() 1228 if (!buff) in __neigh_event_send() 1230 neigh->arp_queue_len_bytes -= buff->truesize; in __neigh_event_send() 1231 kfree_skb_reason(buff, SKB_DROP_REASON_NEIGH_QUEUEFULL); in __neigh_event_send()
|
D | skbuff.c | 3528 static __wsum warn_crc32c_csum_update(const void *buff, int len, __wsum sum) in warn_crc32c_csum_update() argument
|
D | filter.c | 1653 u8 buff[MAX_BPF_STACK]; member
|