/net/ax25/ |
D | ax25_ip.c | 49 unsigned char *buff; in ax25_hard_header() local 56 buff = skb_push(skb, AX25_HEADER_LEN); in ax25_hard_header() 57 *buff++ = 0x00; /* KISS DATA */ in ax25_hard_header() 60 memcpy(buff, daddr, dev->addr_len); /* Address specified */ in ax25_hard_header() 62 buff[6] &= ~AX25_CBIT; in ax25_hard_header() 63 buff[6] &= ~AX25_EBIT; in ax25_hard_header() 64 buff[6] |= AX25_SSSID_SPARE; in ax25_hard_header() 65 buff += AX25_ADDR_LEN; in ax25_hard_header() 68 memcpy(buff, saddr, dev->addr_len); in ax25_hard_header() 70 memcpy(buff, dev->dev_addr, dev->addr_len); in ax25_hard_header() [all …]
|
/net/netrom/ |
D | nr_dev.c | 69 unsigned char *buff = skb_push(skb, NR_NETWORK_LEN + NR_TRANSPORT_LEN); in nr_header() local 71 memcpy(buff, (saddr != NULL) ? saddr : dev->dev_addr, dev->addr_len); in nr_header() 72 buff[6] &= ~AX25_CBIT; in nr_header() 73 buff[6] &= ~AX25_EBIT; in nr_header() 74 buff[6] |= AX25_SSSID_SPARE; in nr_header() 75 buff += AX25_ADDR_LEN; in nr_header() 78 memcpy(buff, daddr, dev->addr_len); in nr_header() 79 buff[6] &= ~AX25_CBIT; in nr_header() 80 buff[6] |= AX25_EBIT; in nr_header() 81 buff[6] |= AX25_SSSID_SPARE; in nr_header() [all …]
|
/net/batman-adv/ |
D | gateway_common.c | 36 bool batadv_parse_throughput(struct net_device *net_dev, char *buff, in batadv_parse_throughput() argument 44 if (strlen(buff) > 4) { in batadv_parse_throughput() 45 tmp_ptr = buff + strlen(buff) - 4; in batadv_parse_throughput() 55 ret = kstrtou64(buff, 10, <hroughput); in batadv_parse_throughput() 59 description, buff); in batadv_parse_throughput() 69 description, buff); in batadv_parse_throughput() 84 description, buff); in batadv_parse_throughput() 103 static bool batadv_parse_gw_bandwidth(struct net_device *net_dev, char *buff, in batadv_parse_gw_bandwidth() argument 109 slash_ptr = strchr(buff, '/'); in batadv_parse_gw_bandwidth() 113 ret = batadv_parse_throughput(net_dev, buff, "download gateway speed", in batadv_parse_gw_bandwidth() [all …]
|
D | gateway_common.h | 30 ssize_t batadv_gw_bandwidth_set(struct net_device *net_dev, char *buff, 35 bool batadv_parse_throughput(struct net_device *net_dev, char *buff,
|
D | bat_v.c | 516 char *buff, size_t count) in batadv_v_store_sel_class() argument 520 if (!batadv_parse_throughput(bat_priv->soft_iface, buff, in batadv_v_store_sel_class()
|
D | types.h | 2203 ssize_t (*store_sel_class)(struct batadv_priv *bat_priv, char *buff,
|
/net/ipv4/ |
D | proc.c | 404 unsigned long buff[TCPUDP_MIB_MAX]; in snmp_seq_show_tcp_udp() local 408 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp() 415 snmp_get_cpu_field_batch(buff, snmp4_tcp_list, in snmp_seq_show_tcp_udp() 420 seq_printf(seq, " %ld", buff[i]); in snmp_seq_show_tcp_udp() 422 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp() 425 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp() 427 snmp_get_cpu_field_batch(buff, snmp4_udp_list, in snmp_seq_show_tcp_udp() 434 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp() 436 memset(buff, 0, TCPUDP_MIB_MAX * sizeof(unsigned long)); in snmp_seq_show_tcp_udp() 440 snmp_get_cpu_field_batch(buff, snmp4_udp_list, in snmp_seq_show_tcp_udp() [all …]
|
D | tcp_output.c | 1514 struct sk_buff *buff, in tcp_insert_write_queue_after() argument 1519 __skb_queue_after(&sk->sk_write_queue, skb, buff); in tcp_insert_write_queue_after() 1521 tcp_rbtree_insert(&sk->tcp_rtx_queue, buff); in tcp_insert_write_queue_after() 1534 struct sk_buff *buff; in tcp_fragment() local 1565 buff = tcp_stream_alloc_skb(sk, nsize, gfp, true); in tcp_fragment() 1566 if (!buff) in tcp_fragment() 1568 skb_copy_decrypted(buff, skb); in tcp_fragment() 1569 mptcp_skb_ext_copy(buff, skb); in tcp_fragment() 1571 sk_wmem_queued_add(sk, buff->truesize); in tcp_fragment() 1572 sk_mem_charge(sk, buff->truesize); in tcp_fragment() [all …]
|
D | icmp.c | 1033 char buff[IFNAMSIZ]; in icmp_build_probe() local 1073 memset(buff, 0, sizeof(buff)); in icmp_build_probe() 1074 memcpy(buff, &iio->ident.name, ident_len); in icmp_build_probe() 1075 dev = dev_get_by_name(net, buff); in icmp_build_probe()
|
/net/rose/ |
D | rose_dev.c | 39 unsigned char *buff = skb_push(skb, ROSE_MIN_LEN + 2); in rose_header() local 42 memcpy(buff + 7, daddr, dev->addr_len); in rose_header() 44 *buff++ = ROSE_GFI | ROSE_Q_BIT; in rose_header() 45 *buff++ = 0x00; in rose_header() 46 *buff++ = ROSE_DATA; in rose_header() 47 *buff++ = 0x7F; in rose_header() 48 *buff++ = AX25_P_IP; in rose_header()
|
/net/atm/ |
D | mpoa_proc.c | 51 static ssize_t proc_mpc_write(struct file *file, const char __user *buff, 54 static int parse_qos(const char *buff); 206 static ssize_t proc_mpc_write(struct file *file, const char __user *buff, in proc_mpc_write() argument 223 if (get_user(*p, buff++)) { in proc_mpc_write() 242 static int parse_qos(const char *buff) in parse_qos() argument 254 if (sscanf(buff, "del %hhu.%hhu.%hhu.%hhu", in parse_qos() 260 if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=tx", in parse_qos() 264 } else if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=%d,%d", in parse_qos()
|
D | lec.c | 131 char *buff; in lec_handle_bridge() local 139 buff = skb->data + skb->dev->hard_header_len; in lec_handle_bridge() 140 if (*buff++ == 0x42 && *buff++ == 0x42 && *buff++ == 0x03) { in lec_handle_bridge() 151 buff += 4; in lec_handle_bridge() 152 mesg->content.normal.flag = *buff & 0x01; in lec_handle_bridge()
|
D | mpc.c | 496 char *buff; in send_via_shortcut() local 507 buff = skb->data + mpc->dev->hard_header_len; in send_via_shortcut() 508 iph = (struct iphdr *)buff; in send_via_shortcut()
|
/net/netfilter/ipvs/ |
D | ip_vs_sync.c | 545 struct ip_vs_sync_buff *buff; in ip_vs_sync_conn_v0() local 567 buff = ms->sync_buff; in ip_vs_sync_conn_v0() 570 if (buff) { in ip_vs_sync_conn_v0() 571 m = (struct ip_vs_sync_mesg_v0 *) buff->mesg; in ip_vs_sync_conn_v0() 573 if (buff->head + len > buff->end || !m->nr_conns) { in ip_vs_sync_conn_v0() 576 buff = NULL; in ip_vs_sync_conn_v0() 579 if (!buff) { in ip_vs_sync_conn_v0() 580 buff = ip_vs_sync_buff_create_v0(ipvs, len); in ip_vs_sync_conn_v0() 581 if (!buff) { in ip_vs_sync_conn_v0() 586 ms->sync_buff = buff; in ip_vs_sync_conn_v0() [all …]
|
/net/xfrm/ |
D | xfrm_proc.c | 49 unsigned long buff[LINUX_MIB_XFRMMAX]; in xfrm_statistics_seq_show() local 53 memset(buff, 0, sizeof(unsigned long) * LINUX_MIB_XFRMMAX); in xfrm_statistics_seq_show() 55 snmp_get_cpu_field_batch(buff, xfrm_mib_list, in xfrm_statistics_seq_show() 59 buff[i]); in xfrm_statistics_seq_show()
|
/net/netfilter/ |
D | nft_exthdr.c | 195 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_eval() local 202 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_eval() 239 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_set_eval() local 245 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_set_eval() 319 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_strip_eval() local 325 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_strip_eval()
|
D | nf_conntrack_proto_tcp.c | 338 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_options() local 346 length, buff); in tcp_options() 394 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_sack() local 403 length, buff); in tcp_sack()
|
/net/ipv6/ |
D | proc.c | 185 unsigned long buff[SNMP_MIB_MAX]; in snmp6_seq_show_item() local 189 memset(buff, 0, sizeof(unsigned long) * SNMP_MIB_MAX); in snmp6_seq_show_item() 191 snmp_get_cpu_field_batch(buff, itemlist, pcpumib); in snmp6_seq_show_item() 194 itemlist[i].name, buff[i]); in snmp6_seq_show_item()
|
D | tcp_ipv6.c | 856 struct sk_buff *buff; in tcp_v6_send_response() local 881 buff = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC); in tcp_v6_send_response() 882 if (!buff) in tcp_v6_send_response() 885 skb_reserve(buff, MAX_TCP_HEADER); in tcp_v6_send_response() 887 t1 = skb_push(buff, tot_len); in tcp_v6_send_response() 888 skb_reset_transport_header(buff); in tcp_v6_send_response() 928 buff->ip_summed = CHECKSUM_PARTIAL; in tcp_v6_send_response() 930 __tcp_v6_send_check(buff, &fl6.saddr, &fl6.daddr); in tcp_v6_send_response() 947 skb_set_delivery_time(buff, tcp_transmit_time(sk), true); in tcp_v6_send_response() 951 skb_set_hash(buff, txhash, PKT_HASH_TYPE_L4); in tcp_v6_send_response() [all …]
|
D | ndisc.c | 1673 struct sk_buff *buff; in ndisc_send_redirect() local 1757 buff = ndisc_alloc_skb(dev, sizeof(*msg) + optlen); in ndisc_send_redirect() 1758 if (!buff) in ndisc_send_redirect() 1761 msg = skb_put(buff, sizeof(*msg)); in ndisc_send_redirect() 1775 ndisc_fill_redirect_addr_option(buff, ha, ops_data); in ndisc_send_redirect() 1782 ndisc_fill_redirect_hdr_option(buff, skb, rd_len); in ndisc_send_redirect() 1784 skb_dst_set(buff, dst); in ndisc_send_redirect() 1785 ndisc_send_skb(buff, &ipv6_hdr(skb)->saddr, &saddr_buf); in ndisc_send_redirect()
|
/net/sctp/ |
D | proc.c | 61 unsigned long buff[SCTP_MIB_MAX]; in sctp_snmp_seq_show() local 65 memset(buff, 0, sizeof(unsigned long) * SCTP_MIB_MAX); in sctp_snmp_seq_show() 67 snmp_get_cpu_field_batch(buff, sctp_snmp_list, in sctp_snmp_seq_show() 71 buff[i]); in sctp_snmp_seq_show()
|
/net/xdp/ |
D | xsk_buff_pool.c | 608 struct xdp_buff *buff; in xp_alloc_batch() local 611 buff = xp_alloc(pool); in xp_alloc_batch() 612 if (buff) in xp_alloc_batch() 613 *xdp = buff; in xp_alloc_batch() 614 return !!buff; in xp_alloc_batch()
|
/net/ |
D | socket.c | 2105 int __sys_sendto(int fd, void __user *buff, size_t len, unsigned int flags, in __sys_sendto() argument 2115 err = import_single_range(ITER_SOURCE, buff, len, &iov, &msg.msg_iter); in __sys_sendto() 2146 SYSCALL_DEFINE6(sendto, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE6() argument 2150 return __sys_sendto(fd, buff, len, flags, addr, addr_len); in SYSCALL_DEFINE6() 2157 SYSCALL_DEFINE4(send, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE4() argument 2160 return __sys_sendto(fd, buff, len, flags, NULL, 0); in SYSCALL_DEFINE4()
|
/net/packet/ |
D | af_packet.c | 197 static void packet_increment_head(struct packet_ring_buffer *buff); 1201 static void packet_increment_head(struct packet_ring_buffer *buff) in packet_increment_head() argument 1203 buff->head = buff->head != buff->frame_max ? buff->head+1 : 0; in packet_increment_head()
|
/net/core/ |
D | neighbour.c | 1225 struct sk_buff *buff; in __neigh_event_send() local 1227 buff = __skb_dequeue(&neigh->arp_queue); in __neigh_event_send() 1228 if (!buff) in __neigh_event_send() 1230 neigh->arp_queue_len_bytes -= buff->truesize; in __neigh_event_send() 1231 kfree_skb_reason(buff, SKB_DROP_REASON_NEIGH_QUEUEFULL); in __neigh_event_send()
|