/net/netfilter/ |
D | xt_connbytes.c | 27 u_int64_t bytes = 0; in connbytes_mt() local 57 what = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); in connbytes_mt() 60 what = atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); in connbytes_mt() 63 what = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); in connbytes_mt() 64 what += atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); in connbytes_mt() 71 bytes = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); in connbytes_mt() 75 bytes = atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); in connbytes_mt() 79 bytes = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes) + in connbytes_mt() 80 atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); in connbytes_mt() 86 what = div64_u64(bytes, pkts); in connbytes_mt()
|
D | xt_qtaguid_print.c | 84 dc->bpc[0][IFS_RX][IFS_TCP].bytes, in pp_data_counters() 86 dc->bpc[0][IFS_RX][IFS_UDP].bytes, in pp_data_counters() 88 dc->bpc[0][IFS_RX][IFS_PROTO_OTHER].bytes, in pp_data_counters() 90 dc->bpc[0][IFS_TX][IFS_TCP].bytes, in pp_data_counters() 92 dc->bpc[0][IFS_TX][IFS_UDP].bytes, in pp_data_counters() 94 dc->bpc[0][IFS_TX][IFS_PROTO_OTHER].bytes, in pp_data_counters() 96 dc->bpc[1][IFS_RX][IFS_TCP].bytes, in pp_data_counters() 98 dc->bpc[1][IFS_RX][IFS_UDP].bytes, in pp_data_counters() 100 dc->bpc[1][IFS_RX][IFS_PROTO_OTHER].bytes, in pp_data_counters() 102 dc->bpc[1][IFS_TX][IFS_TCP].bytes, in pp_data_counters() [all …]
|
D | nfnetlink_acct.c | 34 atomic64_t bytes; member 70 atomic64_set(&matching->bytes, 0); in nfnl_acct_new() 83 atomic64_set(&nfacct->bytes, in nfnl_acct_new() 102 u64 pkts, bytes; in nfnl_acct_fill_info() local 119 bytes = atomic64_xchg(&acct->bytes, 0); in nfnl_acct_fill_info() 122 bytes = atomic64_read(&acct->bytes); in nfnl_acct_fill_info() 125 nla_put_be64(skb, NFACCT_BYTES, cpu_to_be64(bytes)) || in nfnl_acct_fill_info() 335 atomic64_add(skb->len, &nfacct->bytes); in nfnl_acct_update()
|
D | xt_repldata.h | 11 unsigned int bytes = 0, hooknum = 0, i = 0; \ 28 tbl->repl.hook_entry[hooknum] = bytes; \ 29 tbl->repl.underflow[hooknum] = bytes; \ 32 bytes += sizeof(struct type##_standard); \
|
D | xt_qtaguid.c | 169 int bytes, in dc_add_byte_packets() argument 172 counters->bpc[set][direction][ifs_proto].bytes += bytes; in dc_add_byte_packets() 690 cnts->bpc[cnt_set][IFS_RX][IFS_TCP].bytes, in pp_iface_stat_line() 692 cnts->bpc[cnt_set][IFS_RX][IFS_UDP].bytes, in pp_iface_stat_line() 694 cnts->bpc[cnt_set][IFS_RX][IFS_PROTO_OTHER].bytes, in pp_iface_stat_line() 696 cnts->bpc[cnt_set][IFS_TX][IFS_TCP].bytes, in pp_iface_stat_line() 698 cnts->bpc[cnt_set][IFS_TX][IFS_UDP].bytes, in pp_iface_stat_line() 700 cnts->bpc[cnt_set][IFS_TX][IFS_PROTO_OTHER].bytes, in pp_iface_stat_line() 765 iface_entry->totals_via_dev[IFS_RX].bytes, in iface_stat_fmt_proc_show() 767 iface_entry->totals_via_dev[IFS_TX].bytes, in iface_stat_fmt_proc_show() [all …]
|
D | xt_qtaguid_internal.h | 174 uint64_t bytes; member 186 return counters->bpc[set][direction][IFS_TCP].bytes in dc_sum_bytes() 187 + counters->bpc[set][direction][IFS_UDP].bytes in dc_sum_bytes() 188 + counters->bpc[set][direction][IFS_PROTO_OTHER].bytes; in dc_sum_bytes()
|
D | nf_conntrack_h323_asn1.c | 208 unsigned int v, l, shift, bytes; in get_bitmap() local 222 for (bytes = l >> 3, shift = 24, v = 0; bytes; in get_bitmap() 223 bytes--, shift -= 8) in get_bitmap()
|
/net/rds/ |
D | page.c | 57 void __user *ptr, unsigned long bytes, in rds_page_copy_user() argument 65 rds_stats_add(s_copy_to_user, bytes); in rds_page_copy_user() 66 ret = copy_to_user(ptr, addr + offset, bytes); in rds_page_copy_user() 68 rds_stats_add(s_copy_from_user, bytes); in rds_page_copy_user() 69 ret = copy_from_user(addr + offset, ptr, bytes); in rds_page_copy_user() 97 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes, in rds_page_remainder_alloc() argument 108 if (bytes >= PAGE_SIZE) { in rds_page_remainder_alloc() 124 if (rem->r_page && bytes > (PAGE_SIZE - rem->r_offset)) { in rds_page_remainder_alloc() 131 if (rem->r_page && bytes <= (PAGE_SIZE - rem->r_offset)) { in rds_page_remainder_alloc() 132 sg_set_page(scat, rem->r_page, bytes, rem->r_offset); in rds_page_remainder_alloc() [all …]
|
D | info.c | 116 unsigned long bytes) in rds_info_copy() argument 120 while (bytes) { in rds_info_copy() 124 this = min(bytes, PAGE_SIZE - iter->offset); in rds_info_copy() 128 iter->offset, this, data, bytes); in rds_info_copy() 133 bytes -= this; in rds_info_copy()
|
D | rdma.c | 58 if ((vec->addr + vec->bytes <= vec->addr) || in rds_pages_in_vec() 59 (vec->bytes > (u64)UINT_MAX)) in rds_pages_in_vec() 62 return ((vec->addr + vec->bytes + PAGE_SIZE - 1) >> PAGE_SHIFT) - in rds_pages_in_vec() 204 args->vec.addr, args->vec.bytes, nr_pages); in __rds_rdma_map() 653 rs->rs_user_bytes = iov->bytes; in rds_cmsg_rdma_args() 663 nr_bytes, nr, iov->bytes, iov->addr); in rds_cmsg_rdma_args() 665 nr_bytes += iov->bytes; in rds_cmsg_rdma_args() 673 min_t(unsigned int, iov->bytes, PAGE_SIZE - offset), in rds_cmsg_rdma_args() 677 sg->offset, sg->length, iov->addr, iov->bytes); in rds_cmsg_rdma_args() 680 iov->bytes -= sg->length; in rds_cmsg_rdma_args() [all …]
|
D | tcp_listen.c | 111 void rds_tcp_listen_data_ready(struct sock *sk, int bytes) in rds_tcp_listen_data_ready() argument 113 void (*ready)(struct sock *sk, int bytes); in rds_tcp_listen_data_ready() 135 ready(sk, bytes); in rds_tcp_listen_data_ready()
|
D | tcp_recv.c | 317 void rds_tcp_data_ready(struct sock *sk, int bytes) in rds_tcp_data_ready() argument 319 void (*ready)(struct sock *sk, int bytes); in rds_tcp_data_ready() 323 rdsdebug("data ready sk %p bytes %d\n", sk, bytes); in rds_tcp_data_ready() 340 ready(sk, bytes); in rds_tcp_data_ready()
|
D | tcp.h | 64 void rds_tcp_listen_data_ready(struct sock *sk, int bytes); 69 void rds_tcp_data_ready(struct sock *sk, int bytes);
|
D | rds.h | 691 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes, 694 void __user *ptr, unsigned long bytes, 696 #define rds_page_copy_to_user(page, offset, ptr, bytes) \ argument 697 rds_page_copy_user(page, offset, ptr, bytes, 1) 698 #define rds_page_copy_from_user(page, offset, ptr, bytes) \ argument 699 rds_page_copy_user(page, offset, ptr, bytes, 0)
|
D | info.h | 26 unsigned long bytes);
|
/net/mac80211/ |
D | led.h | 58 ieee80211_tpt_led_trig_tx(struct ieee80211_local *local, __le16 fc, int bytes) in ieee80211_tpt_led_trig_tx() argument 62 local->tpt_led_trigger->tx_bytes += bytes; in ieee80211_tpt_led_trig_tx() 67 ieee80211_tpt_led_trig_rx(struct ieee80211_local *local, __le16 fc, int bytes) in ieee80211_tpt_led_trig_rx() argument 71 local->tpt_led_trigger->rx_bytes += bytes; in ieee80211_tpt_led_trig_rx()
|
/net/core/ |
D | gen_stats.c | 115 d->tc_stats.bytes = b->bytes; in gnet_stats_copy_basic() 123 sb.bytes = b->bytes; in gnet_stats_copy_basic()
|
D | gen_estimator.c | 128 nbytes = e->bstats->bytes; in est_timer() 229 est->last_bytes = bstats->bytes; in gen_new_estimator()
|
/net/ipv4/ |
D | fib_semantics.c | 687 static struct hlist_head *fib_info_hash_alloc(int bytes) in fib_info_hash_alloc() argument 689 if (bytes <= PAGE_SIZE) in fib_info_hash_alloc() 690 return kzalloc(bytes, GFP_KERNEL); in fib_info_hash_alloc() 694 get_order(bytes)); in fib_info_hash_alloc() 697 static void fib_info_hash_free(struct hlist_head *hash, int bytes) in fib_info_hash_free() argument 702 if (bytes <= PAGE_SIZE) in fib_info_hash_free() 705 free_pages((unsigned long) hash, get_order(bytes)); in fib_info_hash_free() 714 unsigned int i, bytes; in fib_info_hash_move() local 759 bytes = old_size * sizeof(struct hlist_head *); in fib_info_hash_move() 760 fib_info_hash_free(old_info_hash, bytes); in fib_info_hash_move() [all …]
|
/net/irda/ |
D | wrapper.c | 91 __u8 bytes[2]; in async_wrap_skb() member 149 n += stuff_byte(fcs.bytes[0], tx_buff+n); in async_wrap_skb() 150 n += stuff_byte(fcs.bytes[1], tx_buff+n); in async_wrap_skb() 152 n += stuff_byte(fcs.bytes[1], tx_buff+n); in async_wrap_skb() 153 n += stuff_byte(fcs.bytes[0], tx_buff+n); in async_wrap_skb()
|
/net/x25/ |
D | x25_out.c | 35 int bytes = 1; in x25_pacsize_to_bytes() local 41 bytes *= 2; in x25_pacsize_to_bytes() 43 return bytes; in x25_pacsize_to_bytes()
|
/net/irda/irlan/ |
D | irlan_client.c | 437 __u8 *bytes; in irlan_check_response_param() local 522 bytes = value; in irlan_check_response_param() 523 IRDA_DEBUG(4, "Ethernet address = %pM\n", bytes); in irlan_check_response_param() 525 self->dev->dev_addr[i] = bytes[i]; in irlan_check_response_param()
|
/net/sched/ |
D | sch_mqprio.c | 235 sch->bstats.bytes += qdisc->bstats.bytes; in mqprio_dump() 343 bstats.bytes += qdisc->bstats.bytes; in mqprio_dump_class_stats()
|
D | sch_mq.c | 108 sch->bstats.bytes += qdisc->bstats.bytes; in mq_dump()
|
/net/ceph/ |
D | messenger.c | 821 size_t bytes) in ceph_msg_data_bio_advance() argument 838 BUG_ON(cursor->resid < bytes); in ceph_msg_data_bio_advance() 839 cursor->resid -= bytes; in ceph_msg_data_bio_advance() 840 cursor->vector_offset += bytes; in ceph_msg_data_bio_advance() 913 size_t bytes) in ceph_msg_data_pages_advance() argument 917 BUG_ON(cursor->page_offset + bytes > PAGE_SIZE); in ceph_msg_data_pages_advance() 921 cursor->resid -= bytes; in ceph_msg_data_pages_advance() 922 cursor->page_offset = (cursor->page_offset + bytes) & ~PAGE_MASK; in ceph_msg_data_pages_advance() 923 if (!bytes || cursor->page_offset) in ceph_msg_data_pages_advance() 990 size_t bytes) in ceph_msg_data_pagelist_advance() argument [all …]
|