/net/core/ |
D | link_watch.c | 100 unsigned long delay = linkwatch_nextevent - jiffies; in linkwatch_schedule_work() local 109 delay = 0; in linkwatch_schedule_work() 113 if (delay > HZ) in linkwatch_schedule_work() 114 delay = 0; in linkwatch_schedule_work() 120 if (schedule_delayed_work(&linkwatch_work, delay) == !delay) in linkwatch_schedule_work()
|
/net/ipv4/ |
D | tcp_cubic.c | 345 static void hystart_update(struct sock *sk, u32 delay) in hystart_update() argument 362 if (ca->curr_rtt == 0 || ca->curr_rtt > delay) in hystart_update() 363 ca->curr_rtt = delay; in hystart_update() 388 u32 delay; in bictcp_acked() local 403 delay = usecs_to_jiffies(rtt_us) << 3; in bictcp_acked() 404 if (delay == 0) in bictcp_acked() 405 delay = 1; in bictcp_acked() 408 if (ca->delay_min == 0 || ca->delay_min > delay) in bictcp_acked() 409 ca->delay_min = delay; in bictcp_acked() 414 hystart_update(sk, delay); in bictcp_acked()
|
D | route.c | 896 void rt_cache_flush(struct net *net, int delay) in rt_cache_flush() argument 899 if (delay >= 0) in rt_cache_flush() 3042 int delay; in ipv4_sysctl_rtcache_flush_strategy() local 3046 if (get_user(delay, (int __user *)newval)) in ipv4_sysctl_rtcache_flush_strategy() 3049 rt_cache_flush(net, delay); in ipv4_sysctl_rtcache_flush_strategy()
|
D | Kconfig | 556 algorithm, which uses a mixed loss/delay approach to compute the 570 high speed long delay links. It uses round-trip-time to
|
D | igmp.c | 195 static void igmp_ifc_start_timer(struct in_device *in_dev, int delay) in igmp_ifc_start_timer() argument 197 int tv = net_random() % delay; in igmp_ifc_start_timer()
|
/net/sched/ |
D | sch_cbq.c | 415 psched_tdiff_t delay = cl->undertime - q->now; in cbq_ovl_classic() local 418 delay += cl->offtime; in cbq_ovl_classic() 428 delay -= (-cl->avgidle) - ((-cl->avgidle) >> cl->ewma_log); in cbq_ovl_classic() 431 if (delay <= 0) in cbq_ovl_classic() 432 delay = 1; in cbq_ovl_classic() 433 cl->undertime = q->now + delay; in cbq_ovl_classic() 438 if (q->wd_expires == 0 || q->wd_expires > delay) in cbq_ovl_classic() 439 q->wd_expires = delay; in cbq_ovl_classic() 450 delay = b->undertime - q->now; in cbq_ovl_classic() 451 if (delay < base_delay) { in cbq_ovl_classic() [all …]
|
D | sch_netem.c | 217 psched_tdiff_t delay; in netem_enqueue() local 219 delay = tabledist(q->latency, q->jitter, in netem_enqueue() 223 cb->time_to_send = now + delay; in netem_enqueue()
|
D | Kconfig | 11 delay, and which ones to drop. This is the job of the queueing 188 Say Y if you want to emulate network delay, loss, and packet
|
D | sch_htb.c | 267 struct htb_class *cl, long delay) in htb_add_to_wait_tree() argument 271 cl->pq_key = q->now + delay; in htb_add_to_wait_tree()
|
/net/rfkill/ |
D | rfkill-input.c | 211 const unsigned long delay = msecs_to_jiffies(RFKILL_OPS_DELAY); in rfkill_ratelimit() local 212 return (time_after(jiffies, last + delay)) ? 0 : delay; in rfkill_ratelimit()
|
/net/dccp/ |
D | output.c | 210 static int dccp_wait_for_ccid(struct sock *sk, struct sk_buff *skb, int delay) in dccp_wait_for_ccid() argument 218 dccp_pr_debug("delayed send by %d msec\n", delay); in dccp_wait_for_ccid() 219 jiffdelay = msecs_to_jiffies(delay); in dccp_wait_for_ccid() 235 } while ((delay = rc) > 0); in dccp_wait_for_ccid()
|
/net/dccp/ccids/ |
D | ccid3.c | 298 s64 delay; in ccid3_hc_tx_send_packet() local 348 delay = ktime_us_delta(hctx->ccid3hctx_t_nom, now); in ccid3_hc_tx_send_packet() 349 ccid3_pr_debug("delay=%ld\n", (long)delay); in ccid3_hc_tx_send_packet() 358 if (delay - (s64)hctx->ccid3hctx_delta >= 1000) in ccid3_hc_tx_send_packet() 359 return (u32)delay / 1000L; in ccid3_hc_tx_send_packet()
|
/net/netfilter/ipvs/ |
D | Kconfig | 200 tristate "shortest expected delay scheduling" 202 The shortest expected delay scheduling algorithm assigns network 203 connections to the server with the shortest expected delay. The 204 expected delay that the job will experience is (Ci + 1) / Ui if 219 that minimize its expected delay (The Shortest Expected Delay
|
/net/bridge/ |
D | br_sysfs_br.c | 62 unsigned long delay = clock_t_to_jiffies(val); in set_forward_delay() local 63 br->forward_delay = delay; in set_forward_delay() 65 br->bridge_forward_delay = delay; in set_forward_delay()
|
D | br_fdb.c | 122 unsigned long delay = hold_time(br); in br_fdb_cleanup() local 135 this_timer = f->ageing_timer + delay; in br_fdb_cleanup()
|
/net/ipv6/ |
D | mcast.c | 1057 static void mld_ifc_start_timer(struct inet6_dev *idev, int delay) in mld_ifc_start_timer() argument 1059 int tv = net_random() % delay; in mld_ifc_start_timer() 1071 unsigned long delay = resptime; in igmp6_group_queried() local 1080 delay = ma->mca_timer.expires - jiffies; in igmp6_group_queried() 1083 if (delay >= resptime) { in igmp6_group_queried() 1085 delay = net_random() % resptime; in igmp6_group_queried() 1087 delay = 1; in igmp6_group_queried() 1089 ma->mca_timer.expires = jiffies + delay; in igmp6_group_queried() 1090 if (!mod_timer(&ma->mca_timer, jiffies + delay)) in igmp6_group_queried() 2146 unsigned long delay; in igmp6_join_group() local [all …]
|
D | route.c | 2527 int delay = net->ipv6.sysctl.flush_delay; in ipv6_sysctl_rtcache_flush() local 2530 fib6_run_gc(delay <= 0 ? ~0UL : (unsigned long)delay, net); in ipv6_sysctl_rtcache_flush()
|
/net/decnet/ |
D | dn_route.c | 338 void dn_rt_cache_flush(int delay) in dn_rt_cache_flush() argument 343 if (delay < 0) in dn_rt_cache_flush() 344 delay = dn_rt_min_delay; in dn_rt_cache_flush() 348 if (del_timer(&dn_rt_flush_timer) && delay > 0 && dn_rt_deadline) { in dn_rt_cache_flush() 354 if (delay > tmo) in dn_rt_cache_flush() 355 delay = tmo; in dn_rt_cache_flush() 358 if (delay <= 0) { in dn_rt_cache_flush() 367 dn_rt_flush_timer.expires = now + delay; in dn_rt_cache_flush()
|
/net/sunrpc/ |
D | cache.c | 493 int delay = 5; in do_cache_clean() local 495 delay = 30*HZ; in do_cache_clean() 498 delay = 0; in do_cache_clean() 500 if (delay) in do_cache_clean() 501 schedule_delayed_work(&cache_cleaner, delay); in do_cache_clean()
|
D | sched.c | 562 void rpc_delay(struct rpc_task *task, unsigned long delay) in rpc_delay() argument 564 task->tk_timeout = delay; in rpc_delay()
|
/net/802/ |
D | garp.c | 398 unsigned long delay; in garp_join_timer_arm() local 400 delay = (u64)msecs_to_jiffies(garp_join_time) * net_random() >> 32; in garp_join_timer_arm() 401 mod_timer(&app->join_timer, jiffies + delay); in garp_join_timer_arm()
|