/net/netfilter/ipvs/ |
D | ip_vs_wrr.c | 77 int weight; in ip_vs_wrr_gcd_weight() local 81 weight = atomic_read(&dest->weight); in ip_vs_wrr_gcd_weight() 82 if (weight > 0) { in ip_vs_wrr_gcd_weight() 84 g = gcd(weight, g); in ip_vs_wrr_gcd_weight() 86 g = weight; in ip_vs_wrr_gcd_weight() 99 int new_weight, weight = 0; in ip_vs_wrr_max_weight() local 102 new_weight = atomic_read(&dest->weight); in ip_vs_wrr_max_weight() 103 if (new_weight > weight) in ip_vs_wrr_max_weight() 104 weight = new_weight; in ip_vs_wrr_max_weight() 107 return weight; in ip_vs_wrr_max_weight() [all …]
|
D | ip_vs_lblcr.c | 177 if ((atomic_read(&least->weight) > 0) in ip_vs_dest_set_min() 193 if (((__s64)loh * atomic_read(&dest->weight) > in ip_vs_dest_set_min() 194 (__s64)doh * atomic_read(&least->weight)) in ip_vs_dest_set_min() 208 atomic_read(&least->weight), loh); in ip_vs_dest_set_min() 226 if (atomic_read(&most->weight) > 0) { in ip_vs_dest_set_max() 239 if (((__s64)moh * atomic_read(&dest->weight) < in ip_vs_dest_set_max() 240 (__s64)doh * atomic_read(&most->weight)) in ip_vs_dest_set_max() 241 && (atomic_read(&dest->weight) > 0)) { in ip_vs_dest_set_max() 253 atomic_read(&most->weight), moh); in ip_vs_dest_set_max() 587 if (atomic_read(&dest->weight) > 0) { in __ip_vs_lblcr_schedule() [all …]
|
D | ip_vs_nq.c | 82 !atomic_read(&dest->weight)) in ip_vs_nq_schedule() 95 ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_nq_schedule() 96 (__s64)doh * atomic_read(&least->weight))) { in ip_vs_nq_schedule() 114 atomic_read(&least->weight), loh); in ip_vs_nq_schedule()
|
D | ip_vs_wlc.c | 57 atomic_read(&dest->weight) > 0) { in ip_vs_wlc_schedule() 74 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_wlc_schedule() 75 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_wlc_schedule() 87 atomic_read(&least->weight), loh); in ip_vs_wlc_schedule()
|
D | ip_vs_sed.c | 85 atomic_read(&dest->weight) > 0) { in ip_vs_sed_schedule() 102 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_sed_schedule() 103 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_sed_schedule() 115 atomic_read(&least->weight), loh); in ip_vs_sed_schedule()
|
D | ip_vs_fo.c | 39 atomic_read(&dest->weight) > hw) { in ip_vs_fo_schedule() 41 hw = atomic_read(&dest->weight); in ip_vs_fo_schedule() 50 atomic_read(&hweight->weight)); in ip_vs_fo_schedule()
|
D | ip_vs_lblc.c | 423 if (atomic_read(&dest->weight) > 0) { in __ip_vs_lblc_schedule() 440 if ((__s64)loh * atomic_read(&dest->weight) > in __ip_vs_lblc_schedule() 441 (__s64)doh * atomic_read(&least->weight)) { in __ip_vs_lblc_schedule() 453 atomic_read(&least->weight), loh); in __ip_vs_lblc_schedule() 466 if (atomic_read(&dest->activeconns) > atomic_read(&dest->weight)) { in is_overloaded() 471 < atomic_read(&d->weight)) { in is_overloaded() 510 atomic_read(&dest->weight) > 0 && !is_overloaded(dest, svc)) in ip_vs_lblc_schedule()
|
D | ip_vs_rr.c | 76 atomic_read(&dest->weight) > 0) in ip_vs_rr_schedule() 100 atomic_read(&dest->refcnt), atomic_read(&dest->weight)); in ip_vs_rr_schedule()
|
D | ip_vs_sh.c | 81 return atomic_read(&dest->weight) <= 0 || in is_unavailable() 196 atomic_read(&dest->weight)); in ip_vs_sh_reassign() 199 if (++d_count >= atomic_read(&dest->weight)) { in ip_vs_sh_reassign()
|
D | ip_vs_lc.c | 48 atomic_read(&dest->weight) == 0) in ip_vs_lc_schedule()
|
D | Kconfig | 150 normalized by the server weight. 159 connections to the server with the highest weight that is 172 overloaded (its active connection numbers is larger than its weight) 224 on the ith server and Ui is the fixed service rate (weight)
|
D | ip_vs_ctl.c | 789 atomic_set(&dest->weight, udest->weight); in __ip_vs_update_dest() 929 if (udest->weight < 0) { in ip_vs_add_dest() 993 if (udest->weight < 0) { in ip_vs_edit_dest() 2020 atomic_read(&dest->weight), in ip_vs_info_seq_show() 2031 atomic_read(&dest->weight), in ip_vs_info_seq_show() 2271 udest->weight = udest_compat->weight; in ip_vs_copy_udest_compat() 2522 entry.weight = atomic_read(&dest->weight); in __ip_vs_get_dest_entries() 3066 atomic_read(&dest->weight)) || in ip_vs_genl_fill_dest() 3195 udest->weight = nla_get_u32(nla_weight); in ip_vs_genl_parse_dest()
|
D | ip_vs_dh.c | 229 || atomic_read(&dest->weight) <= 0 in ip_vs_dh_schedule()
|
D | ip_vs_conn.c | 735 (atomic_read(&dest->weight) == 0); in expire_quiescent_template()
|
/net/ceph/crush/ |
D | mapper.c | 269 const __u32 *weight, int weight_max, in is_out() argument 274 if (weight[item] >= 0x10000) in is_out() 276 if (weight[item] == 0) in is_out() 279 < weight[item]) in is_out() 305 const __u32 *weight, int weight_max, in crush_choose_firstn() argument 407 weight, weight_max, in crush_choose_firstn() 428 reject = is_out(map, weight, in crush_choose_firstn() 483 const __u32 *weight, int weight_max, in crush_choose_indep() argument 594 weight, weight_max, in crush_choose_indep() 611 is_out(map, weight, weight_max, item, x)) in crush_choose_indep() [all …]
|
/net/dccp/ccids/lib/ |
D | tfrc.h | 53 static inline u32 tfrc_ewma(const u32 avg, const u32 newval, const u8 weight) in tfrc_ewma() argument 55 return avg ? (weight * avg + (10 - weight) * newval) / 10 : newval; in tfrc_ewma()
|
/net/sched/ |
D | sch_qfq.c | 267 u32 lmax, u32 weight) in qfq_init_agg() argument 273 agg->class_weight = weight; in qfq_init_agg() 277 u32 lmax, u32 weight) in qfq_find_agg() argument 282 if (agg->lmax == lmax && agg->class_weight == weight) in qfq_find_agg() 388 static int qfq_change_agg(struct Qdisc *sch, struct qfq_class *cl, u32 weight, in qfq_change_agg() argument 392 struct qfq_aggregate *new_agg = qfq_find_agg(q, lmax, weight); in qfq_change_agg() 398 qfq_init_agg(q, new_agg, lmax, weight); in qfq_change_agg() 414 u32 weight, lmax, inv_w; in qfq_change_class() local 428 weight = nla_get_u32(tb[TCA_QFQ_WEIGHT]); in qfq_change_class() 429 if (!weight || weight > (1UL << QFQ_MAX_WSHIFT)) { in qfq_change_class() [all …]
|
D | sch_hhf.c | 402 unsigned int weight; in hhf_enqueue() local 410 weight = 1; in hhf_enqueue() 413 weight = q->hhf_non_hh_weight; in hhf_enqueue() 416 bucket->deficit = weight * q->quantum; in hhf_enqueue() 451 int weight = (bucket - q->buckets == WDRR_BUCKET_FOR_HH) ? in hhf_dequeue() local 454 bucket->deficit += weight * q->quantum; in hhf_dequeue()
|
D | sch_cbq.c | 104 long weight; /* Relative allotment: see below */ member 1042 cl->quantum = (cl->weight*cl->allot*q->nclasses[prio])/ in cbq_normalize_quanta() 1253 q->quanta[cl->priority] -= cl->weight; in cbq_rmprio() 1260 q->quanta[cl->priority] += cl->weight; in cbq_addprio() 1270 if (wrr->weight) in cbq_set_wrr() 1271 cl->weight = wrr->weight; in cbq_set_wrr() 1382 q->link.weight = q->link.R_tab->rate.rate; in cbq_init() 1456 opt.weight = cl->weight; in cbq_dump_wrr() 1870 cl->weight = cl->R_tab->rate.rate; in cbq_change_class()
|
/net/mac80211/ |
D | rc80211_minstrel.h | 29 minstrel_ewma(int old, int new, int weight) in minstrel_ewma() argument 31 return (new * (EWMA_DIV - weight) + old * weight) / EWMA_DIV; in minstrel_ewma()
|
D | main.c | 1093 int weight) in ieee80211_napi_add() argument 1097 netif_napi_add(napi_dev, napi, poll, weight); in ieee80211_napi_add()
|
/net/ipv4/netfilter/ |
D | ipt_CLUSTERIP.c | 568 unsigned int weight; /* number of bits set == size */ member 576 unsigned int weight; in clusterip_seq_start() local 582 weight = hweight32(local_nodes); in clusterip_seq_start() 583 if (*pos >= weight) in clusterip_seq_start() 591 idx->weight = weight; in clusterip_seq_start() 604 if (*pos >= idx->weight) { in clusterip_seq_next() 628 if (idx->pos == idx->weight - 1) in clusterip_seq_show()
|
/net/ceph/ |
D | osdmap.c | 221 b->weight = ceph_decode_32(p); in crush_decode() 669 u32 *weight; in osdmap_set_max_osd() local 678 weight = krealloc(map->osd_weight, max*sizeof(*weight), GFP_NOFS); in osdmap_set_max_osd() 679 if (!weight) in osdmap_set_max_osd() 681 map->osd_weight = weight; in osdmap_set_max_osd() 1520 const __u32 *weight, int weight_max) in do_crush() argument 1528 weight, weight_max, map->crush_scratch_ary); in do_crush()
|
/net/core/ |
D | dev.c | 4360 napi->weight = weight_p; in process_backlog() 4502 int (*poll)(struct napi_struct *, int), int weight) in netif_napi_add() argument 4509 if (weight > NAPI_POLL_WEIGHT) in netif_napi_add() 4511 weight, dev->name); in netif_napi_add() 4512 napi->weight = weight; in netif_napi_add() 4545 int work, weight; in net_rx_action() local 4565 weight = n->weight; in net_rx_action() 4575 work = n->poll(n, weight); in net_rx_action() 4579 WARN_ON_ONCE(work > weight); in net_rx_action() 4590 if (unlikely(work == weight)) { in net_rx_action() [all …]
|
/net/batman-adv/ |
D | bat_iv_ogm.c | 1649 uint8_t *weight; in batadv_iv_ogm_process() local 1671 weight = &orig_neigh_node->bat_iv.bcast_own_sum[if_num]; in batadv_iv_ogm_process() 1672 *weight = bitmap_weight(word, in batadv_iv_ogm_process()
|