/net/sched/ |
D | sch_gred.c | 40 u32 backlog; /* bytes on the virtualQ */ member 118 return sch->qstats.backlog; in gred_backlog() 120 return q->backlog; in gred_backlog() 232 if (q->backlog + qdisc_pkt_len(skb) <= q->limit) { in gred_enqueue() 233 q->backlog += qdisc_pkt_len(skb); in gred_enqueue() 261 q->backlog -= qdisc_pkt_len(skb); in gred_dequeue() 264 if (!sch->qstats.backlog) in gred_dequeue() 267 if (!q->backlog) in gred_dequeue() 293 q->backlog -= len; in gred_drop() 297 if (!sch->qstats.backlog) in gred_drop() [all …]
|
D | sch_sfq.c | 113 unsigned int backlog; member 331 slot->backlog -= len; in sfq_drop() 336 sch->qstats.backlog -= len; in sfq_drop() 398 slot->backlog = 0; /* should already be 0 anyway... */ in sfq_enqueue() 405 slot->backlog); in sfq_enqueue() 456 sch->qstats.backlog -= delta; in sfq_enqueue() 457 slot->backlog -= delta; in sfq_enqueue() 465 sch->qstats.backlog += qdisc_pkt_len(skb); in sfq_enqueue() 466 slot->backlog += qdisc_pkt_len(skb); in sfq_enqueue() 524 sch->qstats.backlog -= qdisc_pkt_len(skb); in sfq_dequeue() [all …]
|
D | sch_mqprio.c | 238 sch->qstats.backlog += qdisc->qstats.backlog; in mqprio_dump() 346 qstats.backlog += qdisc->qstats.backlog; in mqprio_dump_class_stats()
|
D | sch_red.c | 67 child->qstats.backlog); in red_enqueue() 271 sch->qstats.backlog = q->qdisc->qstats.backlog; in red_dump()
|
D | sch_choke.c | 137 sch->qstats.backlog -= qdisc_pkt_len(skb); in choke_drop_by_idx() 330 sch->qstats.backlog += qdisc_pkt_len(skb); in choke_enqueue() 363 sch->qstats.backlog -= qdisc_pkt_len(skb); in choke_dequeue() 462 sch->qstats.backlog -= qdisc_pkt_len(skb); in choke_change()
|
D | sch_mq.c | 111 sch->qstats.backlog += qdisc->qstats.backlog; in mq_dump()
|
D | sch_fq_codel.c | 165 sch->qstats.backlog -= len; in fq_codel_drop() 190 sch->qstats.backlog += qdisc_pkt_len(skb); in fq_codel_enqueue() 552 qs.backlog = q->backlogs[idx]; in fq_codel_dump_class_stats()
|
D | sch_tbf.c | 347 sch->qstats.backlog = q->qdisc->qstats.backlog; in tbf_dump()
|
D | sch_plug.c | 93 if (likely(sch->qstats.backlog + skb->len <= q->limit)) { in plug_enqueue()
|
D | sch_fifo.c | 24 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= sch->limit)) in bfifo_enqueue()
|
D | sch_hfsc.c | 1372 cl->qstats.backlog = cl->qdisc->qstats.backlog; in hfsc_dump_class_stats() 1567 sch->qstats.backlog = 0; in hfsc_dump_qdisc() 1570 sch->qstats.backlog += cl->qdisc->qstats.backlog; in hfsc_dump_qdisc()
|
D | sch_codel.c | 152 sch->qstats.backlog -= qdisc_pkt_len(skb); in codel_change()
|
D | sch_sfb.c | 571 sch->qstats.backlog = q->qdisc->qstats.backlog; in sfb_dump()
|
/net/core/ |
D | gen_stats.c | 180 d->tc_stats.backlog = q->backlog; in gnet_stats_copy_queue()
|
/net/sunrpc/ |
D | xprt.c | 934 xprt->stat.bklog_u += xprt->backlog.qlen; in xprt_transmit() 954 rpc_sleep_on(&xprt->backlog, task, NULL); in xprt_add_backlog() 959 if (rpc_wake_up_next(&xprt->backlog) == NULL) in xprt_wake_up_backlog() 971 rpc_sleep_on(&xprt->backlog, task, NULL); in xprt_throttle_congested() 1268 rpc_init_priority_wait_queue(&xprt->backlog, "xprt_backlog"); in xprt_init() 1341 rpc_destroy_wait_queue(&xprt->backlog); in xprt_destroy()
|
/net/dccp/ |
D | proto.c | 234 static inline int dccp_listen_start(struct sock *sk, int backlog) in dccp_listen_start() argument 242 return inet_csk_listen_start(sk, backlog); in dccp_listen_start() 918 int inet_dccp_listen(struct socket *sock, int backlog) in inet_dccp_listen() argument 942 err = dccp_listen_start(sk, backlog); in inet_dccp_listen() 946 sk->sk_max_ack_backlog = backlog; in inet_dccp_listen()
|
/net/atm/ |
D | svc.c | 287 static int svc_listen(struct socket *sock, int backlog) in svc_listen() argument 319 sk->sk_max_ack_backlog = backlog > 0 ? backlog : ATM_BACKLOG_DEFAULT; in svc_listen()
|
/net/rxrpc/ |
D | af_rxrpc.c | 188 static int rxrpc_listen(struct socket *sock, int backlog) in rxrpc_listen() argument 194 _enter("%p,%d", rx, backlog); in rxrpc_listen() 209 sk->sk_max_ack_backlog = backlog; in rxrpc_listen()
|
/net/llc/ |
D | af_llc.c | 505 static int llc_ui_listen(struct socket *sock, int backlog) in llc_ui_listen() argument 520 if (!(unsigned int)backlog) /* BSDism */ in llc_ui_listen() 521 backlog = 1; in llc_ui_listen() 522 sk->sk_max_ack_backlog = backlog; in llc_ui_listen()
|
/net/nfc/ |
D | llcp_sock.c | 199 static int llcp_sock_listen(struct socket *sock, int backlog) in llcp_sock_listen() argument 204 pr_debug("sk %p backlog %d\n", sk, backlog); in llcp_sock_listen() 214 sk->sk_max_ack_backlog = backlog; in llcp_sock_listen()
|
/net/bluetooth/rfcomm/ |
D | sock.c | 416 static int rfcomm_sock_listen(struct socket *sock, int backlog) in rfcomm_sock_listen() argument 421 BT_DBG("sk %p backlog %d", sk, backlog); in rfcomm_sock_listen() 456 sk->sk_max_ack_backlog = backlog; in rfcomm_sock_listen()
|
/net/bluetooth/ |
D | sco.c | 531 static int sco_sock_listen(struct socket *sock, int backlog) in sco_sock_listen() argument 537 BT_DBG("sk %p backlog %d", sk, backlog); in sco_sock_listen() 558 sk->sk_max_ack_backlog = backlog; in sco_sock_listen()
|
D | l2cap_sock.c | 152 static int l2cap_sock_listen(struct socket *sock, int backlog) in l2cap_sock_listen() argument 158 BT_DBG("sk %p backlog %d", sk, backlog); in l2cap_sock_listen() 185 sk->sk_max_ack_backlog = backlog; in l2cap_sock_listen()
|
/net/ipv4/ |
D | af_inet.c | 209 int inet_listen(struct socket *sock, int backlog) in inet_listen() argument 239 err = fastopen_init_queue(sk, backlog); in inet_listen() 249 err = inet_csk_listen_start(sk, backlog); in inet_listen() 253 sk->sk_max_ack_backlog = backlog; in inet_listen()
|
/net/ |
D | socket.c | 1540 SYSCALL_DEFINE2(listen, int, fd, int, backlog) in SYSCALL_DEFINE2() argument 1549 if ((unsigned int)backlog > somaxconn) in SYSCALL_DEFINE2() 1550 backlog = somaxconn; in SYSCALL_DEFINE2() 1552 err = security_socket_listen(sock, backlog); in SYSCALL_DEFINE2() 1554 err = sock->ops->listen(sock, backlog); in SYSCALL_DEFINE2() 3356 int kernel_listen(struct socket *sock, int backlog) in kernel_listen() argument 3358 return sock->ops->listen(sock, backlog); in kernel_listen()
|