/net/ipv4/ |
D | tcp_bic.c | 51 u32 cnt; /* increase cwnd by 1 after ACKs */ member 62 ca->cnt = 0; in bictcp_reset() 97 ca->cnt = cwnd; in bictcp_update() 108 ca->cnt = cwnd / max_increment; in bictcp_update() 111 ca->cnt = (cwnd * smooth_part) / BICTCP_B; in bictcp_update() 114 ca->cnt = cwnd / dist; in bictcp_update() 119 ca->cnt = (cwnd * smooth_part) / BICTCP_B; in bictcp_update() 122 ca->cnt = (cwnd * (BICTCP_B-1)) in bictcp_update() 126 ca->cnt = cwnd / max_increment; in bictcp_update() 131 if (ca->cnt > 20) /* increase cwnd 5% per RTT */ in bictcp_update() [all …]
|
D | tcp_cubic.c | 87 u32 cnt; /* increase cwnd by 1 after ACKs */ member 287 ca->cnt = cwnd / (bic_target - cwnd); in bictcp_update() 289 ca->cnt = 100 * cwnd; /* very small increment*/ in bictcp_update() 296 if (ca->last_max_cwnd == 0 && ca->cnt > 20) in bictcp_update() 297 ca->cnt = 20; /* increase cwnd 5% per RTT */ in bictcp_update() 313 if (ca->cnt > max_cnt) in bictcp_update() 314 ca->cnt = max_cnt; in bictcp_update() 321 ca->cnt = max(ca->cnt, 2U); in bictcp_update() 338 tcp_cong_avoid_ai(tp, ca->cnt, acked); in cubictcp_cong_avoid()
|
D | tcp_nv.c | 184 u32 cnt; in tcpnv_cong_avoid() local 200 cnt = tcp_snd_cwnd(tp) << -ca->cwnd_growth_factor; in tcpnv_cong_avoid() 201 tcp_cong_avoid_ai(tp, cnt, acked); in tcpnv_cong_avoid() 203 cnt = max(4U, tcp_snd_cwnd(tp) >> ca->cwnd_growth_factor); in tcpnv_cong_avoid() 204 tcp_cong_avoid_ai(tp, cnt, acked); in tcpnv_cong_avoid()
|
/net/xdp/ |
D | xsk_queue.h | 221 static inline void xskq_cons_release_n(struct xsk_queue *q, u32 cnt) in xskq_cons_release_n() argument 223 q->cached_cons += cnt; in xskq_cons_release_n() 306 static inline bool xskq_cons_has_entries(struct xsk_queue *q, u32 cnt) in xskq_cons_has_entries() argument 308 return xskq_cons_nb_entries(q, cnt) >= cnt; in xskq_cons_has_entries() 336 static inline void xskq_cons_cancel_n(struct xsk_queue *q, u32 cnt) in xskq_cons_cancel_n() argument 338 q->cached_cons -= cnt; in xskq_cons_cancel_n() 368 static inline void xskq_prod_cancel_n(struct xsk_queue *q, u32 cnt) in xskq_prod_cancel_n() argument 370 q->cached_prod -= cnt; in xskq_prod_cancel_n()
|
/net/ethtool/ |
D | fec.c | 18 u8 cnt; member 78 grp->cnt = stats->total != ETHTOOL_STAT_NOT_SET; in fec_stats_recalc() 82 grp->cnt = 1; in fec_stats_recalc() 89 grp->stats[grp->cnt++] = stats->lanes[i]; in fec_stats_recalc() 176 sizeof(u64) * data->corr.cnt, in fec_put_stats() 179 sizeof(u64) * data->uncorr.cnt, in fec_put_stats() 182 sizeof(u64) * data->corr_bits.cnt, in fec_put_stats()
|
/net/sched/ |
D | cls_flower.c | 2024 #define FL_KEY_SET(keys, cnt, id, member) \ argument 2026 keys[cnt].key_id = id; \ 2027 keys[cnt].offset = FL_KEY_MEMBER_OFFSET(member); \ 2028 cnt++; \ 2031 #define FL_KEY_SET_IF_MASKED(mask, keys, cnt, id, member) \ argument 2034 FL_KEY_SET(keys, cnt, id, member); \ 2041 size_t cnt = 0; in fl_init_dissector() local 2043 FL_KEY_SET_IF_MASKED(mask, keys, cnt, in fl_init_dissector() 2045 FL_KEY_SET(keys, cnt, FLOW_DISSECTOR_KEY_CONTROL, control); in fl_init_dissector() 2046 FL_KEY_SET(keys, cnt, FLOW_DISSECTOR_KEY_BASIC, basic); in fl_init_dissector() [all …]
|
/net/mac80211/ |
D | debugfs_key.c | 337 key->debugfs.cnt = keycount; in ieee80211_debugfs_key_add() 389 sprintf(buf, "../keys/%d", key->debugfs.cnt); in ieee80211_debugfs_key_update_default() 401 sprintf(buf, "../keys/%d", key->debugfs.cnt); in ieee80211_debugfs_key_update_default() 419 sprintf(buf, "../keys/%d", key->debugfs.cnt); in ieee80211_debugfs_key_add_mgmt_default() 448 sprintf(buf, "../keys/%d", key->debugfs.cnt); in ieee80211_debugfs_key_add_beacon_default()
|
D | cfg.c | 1085 for (i = 0; i < src->cnt; i++) { in ieee80211_copy_mbssid_beacon() 1091 dst->cnt = src->cnt; in ieee80211_copy_mbssid_beacon() 1102 for (i = 0; i < src->cnt; i++) { in ieee80211_copy_rnr_beacon() 1108 dst->cnt = src->cnt; in ieee80211_copy_rnr_beacon() 1153 size += struct_size(new->mbssid_ies, elem, mbssid->cnt); in ieee80211_assign_beacon() 1156 size += struct_size(new->rnr_ies, elem, rnr->cnt); in ieee80211_assign_beacon() 1159 mbssid->cnt); in ieee80211_assign_beacon() 1162 size += struct_size(new->mbssid_ies, elem, mbssid->cnt); in ieee80211_assign_beacon() 1165 size += struct_size(new->rnr_ies, elem, rnr->cnt); in ieee80211_assign_beacon() 1168 mbssid->cnt); in ieee80211_assign_beacon() [all …]
|
D | tx.c | 5230 if (!beacon->mbssid_ies || !beacon->mbssid_ies->cnt || in ieee80211_beacon_add_mbssid() 5231 i > beacon->mbssid_ies->cnt) in ieee80211_beacon_add_mbssid() 5234 if (i < beacon->mbssid_ies->cnt) { in ieee80211_beacon_add_mbssid() 5238 if (beacon->rnr_ies && beacon->rnr_ies->cnt) { in ieee80211_beacon_add_mbssid() 5242 for (i = beacon->mbssid_ies->cnt; i < beacon->rnr_ies->cnt; i++) in ieee80211_beacon_add_mbssid() 5250 for (i = 0; i < beacon->mbssid_ies->cnt; i++) in ieee80211_beacon_add_mbssid() 5332 if (!beacon->mbssid_ies || !beacon->mbssid_ies->cnt) in ieee80211_beacon_get_ap_ema_list() 5335 ema = kzalloc(struct_size(ema, bcn, beacon->mbssid_ies->cnt), in ieee80211_beacon_get_ap_ema_list() 5340 for (ema->cnt = 0; ema->cnt < beacon->mbssid_ies->cnt; ema->cnt++) { in ieee80211_beacon_get_ap_ema_list() 5341 ema->bcn[ema->cnt].skb = in ieee80211_beacon_get_ap_ema_list() [all …]
|
D | key.h | 126 int cnt; member
|
D | ieee80211_i.h | 1196 if (!elems || !elems->cnt || i > elems->cnt) in ieee80211_get_mbssid_beacon_len() 1199 if (i < elems->cnt) { in ieee80211_get_mbssid_beacon_len() 1203 for (i = elems->cnt; i < rnr_elems->cnt; i++) in ieee80211_get_mbssid_beacon_len() 1210 for (i = 0; i < elems->cnt; i++) in ieee80211_get_mbssid_beacon_len() 1214 for (i = 0; i < rnr_elems->cnt; i++) in ieee80211_get_mbssid_beacon_len()
|
/net/bridge/ |
D | br_vlan_options.c | 539 u32 cnt; in br_vlan_process_global_one_opts() local 541 cnt = nla_get_u32(tb[BRIDGE_VLANDB_GOPTS_MCAST_LAST_MEMBER_CNT]); in br_vlan_process_global_one_opts() 542 v->br_mcast_ctx.multicast_last_member_count = cnt; in br_vlan_process_global_one_opts() 546 u32 cnt; in br_vlan_process_global_one_opts() local 548 cnt = nla_get_u32(tb[BRIDGE_VLANDB_GOPTS_MCAST_STARTUP_QUERY_CNT]); in br_vlan_process_global_one_opts() 549 v->br_mcast_ctx.multicast_startup_query_count = cnt; in br_vlan_process_global_one_opts()
|
D | br_if.c | 234 u32 cnt = 0; in nbp_update_port_count() local 238 cnt++; in nbp_update_port_count() 240 if (br->auto_cnt != cnt) { in nbp_update_port_count() 241 br->auto_cnt = cnt; in nbp_update_port_count()
|
/net/netfilter/ipvs/ |
D | ip_vs_core.c | 135 u64_stats_inc(&s->cnt.inpkts); in ip_vs_in_stats() 136 u64_stats_add(&s->cnt.inbytes, skb->len); in ip_vs_in_stats() 142 u64_stats_inc(&s->cnt.inpkts); in ip_vs_in_stats() 143 u64_stats_add(&s->cnt.inbytes, skb->len); in ip_vs_in_stats() 148 u64_stats_inc(&s->cnt.inpkts); in ip_vs_in_stats() 149 u64_stats_add(&s->cnt.inbytes, skb->len); in ip_vs_in_stats() 171 u64_stats_inc(&s->cnt.outpkts); in ip_vs_out_stats() 172 u64_stats_add(&s->cnt.outbytes, skb->len); in ip_vs_out_stats() 178 u64_stats_inc(&s->cnt.outpkts); in ip_vs_out_stats() 179 u64_stats_add(&s->cnt.outbytes, skb->len); in ip_vs_out_stats() [all …]
|
D | ip_vs_est.c | 100 conns = u64_stats_read(&c->cnt.conns); in ip_vs_chain_estimation() 101 inpkts = u64_stats_read(&c->cnt.inpkts); in ip_vs_chain_estimation() 102 outpkts = u64_stats_read(&c->cnt.outpkts); in ip_vs_chain_estimation() 103 inbytes = u64_stats_read(&c->cnt.inbytes); in ip_vs_chain_estimation() 104 outbytes = u64_stats_read(&c->cnt.outbytes); in ip_vs_chain_estimation()
|
/net/sctp/ |
D | bind_addr.c | 341 int cnt = 0; in sctp_bind_addrs_check() local 352 cnt = 0; in sctp_bind_addrs_check() 355 cnt++; in sctp_bind_addrs_check() 359 return (cnt == cnt2) ? 0 : (exist ? -EEXIST : 1); in sctp_bind_addrs_check()
|
D | socket.c | 506 int cnt; in sctp_bindx_add() local 516 for (cnt = 0; cnt < addrcnt; cnt++) { in sctp_bindx_add() 535 if (cnt > 0) in sctp_bindx_add() 536 sctp_bindx_rem(sk, addrs, cnt); in sctp_bindx_add() 681 int cnt; in sctp_bindx_rem() local 692 for (cnt = 0; cnt < addrcnt; cnt++) { in sctp_bindx_rem() 737 if (cnt > 0) in sctp_bindx_rem() 738 sctp_bindx_add(sk, addrs, cnt); in sctp_bindx_rem() 6143 int cnt = 0; in sctp_getsockopt_peer_addrs() local 6177 cnt++; in sctp_getsockopt_peer_addrs() [all …]
|
/net/vmw_vsock/ |
D | virtio_transport.c | 194 int cnt = 0, ret; in virtio_transport_cancel_pkt() local 203 cnt = virtio_transport_purge_skbs(vsk, &vsock->send_pkt_queue); in virtio_transport_cancel_pkt() 205 if (cnt) { in virtio_transport_cancel_pkt() 209 new_cnt = atomic_sub_return(cnt, &vsock->queued_replies); in virtio_transport_cancel_pkt() 210 if (new_cnt + cnt >= virtqueue_get_vring_size(rx_vq) && in virtio_transport_cancel_pkt()
|
/net/bluetooth/ |
D | hci_core.c | 3302 int cnt, q; in hci_quote_sent() local 3313 cnt = hdev->acl_cnt; in hci_quote_sent() 3317 cnt = hdev->sco_cnt; in hci_quote_sent() 3320 cnt = hdev->le_mtu ? hdev->le_cnt : hdev->acl_cnt; in hci_quote_sent() 3323 cnt = hdev->iso_mtu ? hdev->iso_cnt : in hci_quote_sent() 3327 cnt = 0; in hci_quote_sent() 3331 q = cnt / num; in hci_quote_sent() 3512 static void __check_timeout(struct hci_dev *hdev, unsigned int cnt, u8 type) in __check_timeout() argument 3531 if (!cnt && time_after(jiffies, last_tx + HCI_ACL_TX_TIMEOUT)) in __check_timeout() 3585 unsigned int cnt = hdev->acl_cnt; in hci_sched_acl_pkt() local [all …]
|
/net/netlink/ |
D | genetlink.c | 215 int i, cnt = 0; in genl_op_from_split() local 219 if (family->split_ops[i + cnt].flags & GENL_CMD_CAP_DO) { in genl_op_from_split() 220 iter->doit = family->split_ops[i + cnt]; in genl_op_from_split() 222 cnt++; in genl_op_from_split() 227 if (i + cnt < family->n_split_ops && in genl_op_from_split() 228 family->split_ops[i + cnt].flags & GENL_CMD_CAP_DUMP) { in genl_op_from_split() 229 iter->dumpit = family->split_ops[i + cnt]; in genl_op_from_split() 231 cnt++; in genl_op_from_split() 236 WARN_ON(!cnt); in genl_op_from_split() 237 iter->entry_idx += cnt; in genl_op_from_split()
|
/net/bridge/netfilter/ |
D | ebtables.c | 394 unsigned int *cnt) in ebt_check_match() argument 425 (*cnt)++; in ebt_check_match() 431 unsigned int *cnt) in ebt_check_watcher() argument 462 (*cnt)++; in ebt_check_watcher() 535 unsigned int *n, unsigned int *cnt, in ebt_check_entry_size_and_hooks() argument 551 if (*n != *cnt) in ebt_check_entry_size_and_hooks() 566 *cnt = 0; in ebt_check_entry_size_and_hooks() 579 (*cnt)++; in ebt_check_entry_size_and_hooks() 656 ebt_cleanup_entry(struct ebt_entry *e, struct net *net, unsigned int *cnt) in ebt_cleanup_entry() argument 664 if (cnt && (*cnt)-- == 0) in ebt_cleanup_entry() [all …]
|
/net/smc/ |
D | smc_clc.c | 531 int cnt = 0; in smc_clc_prfx_set6_rcu() local 539 ipv6_addr_prefix(&ipv6_prfx[cnt].prefix, in smc_clc_prfx_set6_rcu() 541 ipv6_prfx[cnt].prefix_len = ifa->prefix_len; in smc_clc_prfx_set6_rcu() 542 cnt++; in smc_clc_prfx_set6_rcu() 543 if (cnt == SMC_CLC_MAX_V6_PREFIX) in smc_clc_prfx_set6_rcu() 546 prop->ipv6_prefixes_cnt = cnt; in smc_clc_prfx_set6_rcu() 547 if (cnt) in smc_clc_prfx_set6_rcu()
|
/net/rds/ |
D | af_rds.c | 804 unsigned int cnt = 0; in rds_sock_info() local 812 cnt = rds_sock_count; in rds_sock_info() 829 cnt++; in rds_sock_info() 833 lens->nr = cnt; in rds_sock_info()
|
/net/rose/ |
D | af_rose.c | 188 int i, cnt; in rose_kill_by_device() local 192 cnt = 0; in rose_kill_by_device() 197 if (cnt == ARRAY_SIZE(array)) { in rose_kill_by_device() 202 array[cnt++] = sk; in rose_kill_by_device() 207 for (i = 0; i < cnt; i++) { in rose_kill_by_device() 208 sk = array[cnt]; in rose_kill_by_device()
|
/net/llc/ |
D | llc_conn.c | 638 u16 cnt = 0; in llc_find_next_offset() local 643 ++cnt; in llc_find_next_offset() 644 return cnt; in llc_find_next_offset()
|