/net/netfilter/ |
D | xt_LED.c | 63 if ((ledinfo->delay > 0) && ledinfo->always_blink && in led_tg() 70 if (ledinfo->delay > 0) { in led_tg() 72 jiffies + msecs_to_jiffies(ledinfo->delay)); in led_tg() 75 } else if (ledinfo->delay == 0) { in led_tg() 143 if (ledinfo->delay > 0) in led_tg_check() 182 if (ledinfo->delay > 0) in led_tg_destroy()
|
/net/ipv4/ |
D | tcp_cubic.c | 358 static void hystart_update(struct sock *sk, u32 delay) in hystart_update() argument 375 if (ca->curr_rtt == 0 || ca->curr_rtt > delay) in hystart_update() 376 ca->curr_rtt = delay; in hystart_update() 401 u32 delay; in bictcp_acked() local 420 delay = (rtt_us << 3) / USEC_PER_MSEC; in bictcp_acked() 421 if (delay == 0) in bictcp_acked() 422 delay = 1; in bictcp_acked() 425 if (ca->delay_min == 0 || ca->delay_min > delay) in bictcp_acked() 426 ca->delay_min = delay; in bictcp_acked() 431 hystart_update(sk, delay); in bictcp_acked()
|
/net/core/ |
D | link_watch.c | 114 unsigned long delay = linkwatch_nextevent - jiffies; in linkwatch_schedule_work() local 123 delay = 0; in linkwatch_schedule_work() 127 if (delay > HZ) in linkwatch_schedule_work() 128 delay = 0; in linkwatch_schedule_work() 137 schedule_delayed_work(&linkwatch_work, delay); in linkwatch_schedule_work()
|
D | pktgen.c | 256 u64 delay; /* nano-seconds */ member 536 pkt_dev->nfrags, (unsigned long long) pkt_dev->delay, in pktgen_if_show() 972 pkt_dev->delay = ULLONG_MAX; in pktgen_if_write() 974 pkt_dev->delay = (u64)value; in pktgen_if_write() 977 (unsigned long long) pkt_dev->delay); in pktgen_if_write() 988 pkt_dev->delay = pkt_dev->min_pkt_size*8*NSEC_PER_USEC/value; in pktgen_if_write() 990 pr_info("Delay set at: %llu ns\n", pkt_dev->delay); in pktgen_if_write() 1003 pkt_dev->delay = NSEC_PER_SEC/value; in pktgen_if_write() 1005 pr_info("Delay set at: %llu ns\n", pkt_dev->delay); in pktgen_if_write() 2142 pkt_dev->next_tx = ktime_add_ns(spin_until, pkt_dev->delay); in spin() [all …]
|
/net/sched/ |
D | sch_cbq.c | 416 psched_tdiff_t delay = cl->undertime - q->now; in cbq_ovl_classic() local 419 delay += cl->offtime; in cbq_ovl_classic() 429 delay -= (-cl->avgidle) - ((-cl->avgidle) >> cl->ewma_log); in cbq_ovl_classic() 432 if (delay <= 0) in cbq_ovl_classic() 433 delay = 1; in cbq_ovl_classic() 434 cl->undertime = q->now + delay; in cbq_ovl_classic() 439 if (q->wd_expires == 0 || q->wd_expires > delay) in cbq_ovl_classic() 440 q->wd_expires = delay; in cbq_ovl_classic() 451 delay = b->undertime - q->now; in cbq_ovl_classic() 452 if (delay < base_delay) { in cbq_ovl_classic() [all …]
|
D | sch_netem.c | 431 psched_tdiff_t delay; in netem_enqueue() local 433 delay = tabledist(q->latency, q->jitter, in netem_enqueue() 447 delay -= netem_skb_cb(skb_peek_tail(list))->time_to_send - now; in netem_enqueue() 448 delay = max_t(psched_tdiff_t, 0, delay); in netem_enqueue() 452 delay += packet_len_2_sched_time(skb->len, q); in netem_enqueue() 455 cb->time_to_send = now + delay; in netem_enqueue()
|
D | sch_htb.c | 277 struct htb_class *cl, s64 delay) in htb_add_to_wait_tree() argument 281 cl->pq_key = q->now + delay; in htb_add_to_wait_tree()
|
D | Kconfig | 11 delay, and which ones to drop. This is the job of the queueing 199 Say Y if you want to emulate network delay, loss, and packet
|
/net/dccp/ |
D | output.c | 220 static int dccp_wait_for_ccid(struct sock *sk, unsigned long delay) in dccp_wait_for_ccid() argument 229 remaining = schedule_timeout(delay); in dccp_wait_for_ccid() 311 long delay, rc; in dccp_flush_write_queue() local 326 delay = msecs_to_jiffies(rc); in dccp_flush_write_queue() 327 if (delay > *time_budget) in dccp_flush_write_queue() 329 rc = dccp_wait_for_ccid(sk, delay); in dccp_flush_write_queue() 332 *time_budget -= (delay - rc); in dccp_flush_write_queue()
|
/net/rfkill/ |
D | input.c | 145 const unsigned long delay = msecs_to_jiffies(RFKILL_OPS_DELAY); in rfkill_ratelimit() local 146 return time_after(jiffies, last + delay) ? 0 : delay; in rfkill_ratelimit()
|
/net/dccp/ccids/ |
D | ccid3.c | 285 s64 delay; in ccid3_hc_tx_send_packet() local 332 delay = ktime_us_delta(hc->tx_t_nom, now); in ccid3_hc_tx_send_packet() 333 ccid3_pr_debug("delay=%ld\n", (long)delay); in ccid3_hc_tx_send_packet() 342 if (delay >= TFRC_T_DELTA) in ccid3_hc_tx_send_packet() 343 return (u32)delay / USEC_PER_MSEC; in ccid3_hc_tx_send_packet()
|
/net/batman-adv/ |
D | send.h | 32 unsigned long delay);
|
D | send.c | 178 unsigned long delay) in batadv_add_bcast_packet_to_list() argument 219 _batadv_add_bcast_packet_to_list(bat_priv, forw_packet, delay); in batadv_add_bcast_packet_to_list()
|
/net/netfilter/ipvs/ |
D | Kconfig | 208 tristate "shortest expected delay scheduling" 210 The shortest expected delay scheduling algorithm assigns network 211 connections to the server with the shortest expected delay. The 212 expected delay that the job will experience is (Ci + 1) / Ui if 227 that minimize its expected delay (The Shortest Expected Delay
|
/net/ipv6/ |
D | mcast.c | 994 static void mld_ifc_start_timer(struct inet6_dev *idev, int delay) in mld_ifc_start_timer() argument 996 int tv = net_random() % delay; in mld_ifc_start_timer() 1008 unsigned long delay = resptime; in igmp6_group_queried() local 1017 delay = ma->mca_timer.expires - jiffies; in igmp6_group_queried() 1020 if (delay >= resptime) { in igmp6_group_queried() 1022 delay = net_random() % resptime; in igmp6_group_queried() 1024 delay = 1; in igmp6_group_queried() 1026 ma->mca_timer.expires = jiffies + delay; in igmp6_group_queried() 1027 if (!mod_timer(&ma->mca_timer, jiffies + delay)) in igmp6_group_queried() 2101 unsigned long delay; in igmp6_join_group() local [all …]
|
D | route.c | 2800 int delay; in ipv6_sysctl_rtcache_flush() local 2805 delay = net->ipv6.sysctl.flush_delay; in ipv6_sysctl_rtcache_flush() 2807 fib6_run_gc(delay <= 0 ? ~0UL : (unsigned long)delay, net); in ipv6_sysctl_rtcache_flush()
|
/net/ceph/ |
D | mon_client.c | 173 unsigned int delay; in __schedule_delayed() local 176 delay = 10 * HZ; in __schedule_delayed() 178 delay = 20 * HZ; in __schedule_delayed() 179 dout("__schedule_delayed after %u\n", delay); in __schedule_delayed() 180 schedule_delayed_work(&monc->delayed_work, delay); in __schedule_delayed()
|
D | messenger.c | 687 con->delay = 0; /* reset backoff memory */ in ceph_con_open() 2082 con->delay = 0; /* reset backoff memory */ in process_connect() 2643 static int queue_con_delay(struct ceph_connection *con, unsigned long delay) in queue_con_delay() argument 2651 if (!queue_delayed_work(ceph_msgr_wq, &con->work, delay)) { in queue_con_delay() 2658 dout("%s %p %lu\n", __func__, con, delay); in queue_con_delay() 2704 ret = queue_con_delay(con, round_jiffies_relative(con->delay)); in con_backoff() 2707 con, con->delay); in con_backoff() 2840 if (con->delay == 0) in con_fault() 2841 con->delay = BASE_DELAY_INTERVAL; in con_fault() 2842 else if (con->delay < MAX_DELAY_INTERVAL) in con_fault() [all …]
|
/net/decnet/ |
D | dn_route.c | 394 void dn_rt_cache_flush(int delay) in dn_rt_cache_flush() argument 399 if (delay < 0) in dn_rt_cache_flush() 400 delay = dn_rt_min_delay; in dn_rt_cache_flush() 404 if (del_timer(&dn_rt_flush_timer) && delay > 0 && dn_rt_deadline) { in dn_rt_cache_flush() 410 if (delay > tmo) in dn_rt_cache_flush() 411 delay = tmo; in dn_rt_cache_flush() 414 if (delay <= 0) { in dn_rt_cache_flush() 423 dn_rt_flush_timer.expires = now + delay; in dn_rt_cache_flush()
|
/net/802/ |
D | garp.c | 398 unsigned long delay; in garp_join_timer_arm() local 400 delay = (u64)msecs_to_jiffies(garp_join_time) * net_random() >> 32; in garp_join_timer_arm() 401 mod_timer(&app->join_timer, jiffies + delay); in garp_join_timer_arm()
|
D | mrp.c | 579 unsigned long delay; in mrp_join_timer_arm() local 581 delay = (u64)msecs_to_jiffies(mrp_join_time) * net_random() >> 32; in mrp_join_timer_arm() 582 mod_timer(&app->join_timer, jiffies + delay); in mrp_join_timer_arm()
|
/net/sunrpc/ |
D | cache.c | 476 int delay = 5; in do_cache_clean() local 478 delay = round_jiffies_relative(30*HZ); in do_cache_clean() 481 delay = 0; in do_cache_clean() 483 if (delay) in do_cache_clean() 484 schedule_delayed_work(&cache_cleaner, delay); in do_cache_clean()
|
D | sched.c | 660 void rpc_delay(struct rpc_task *task, unsigned long delay) in rpc_delay() argument 662 task->tk_timeout = delay; in rpc_delay()
|
D | svc_xprt.c | 1022 int delay = 0; in svc_close_net() local 1028 msleep(delay++); in svc_close_net()
|
/net/bridge/ |
D | br_fdb.c | 175 unsigned long delay = hold_time(br); in br_fdb_cleanup() local 188 this_timer = f->updated + delay; in br_fdb_cleanup()
|