Lines Matching refs:txq
1946 static void netif_setup_tc(struct net_device *dev, unsigned int txq) in netif_setup_tc() argument
1952 if (tc->offset + tc->count > txq) { in netif_setup_tc()
1963 if (tc->offset + tc->count > txq) { in netif_setup_tc()
2200 int netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq) in netif_set_real_num_tx_queues() argument
2205 disabling = txq < dev->real_num_tx_queues; in netif_set_real_num_tx_queues()
2207 if (txq < 1 || txq > dev->num_tx_queues) in netif_set_real_num_tx_queues()
2215 txq); in netif_set_real_num_tx_queues()
2220 netif_setup_tc(dev, txq); in netif_set_real_num_tx_queues()
2222 dev->real_num_tx_queues = txq; in netif_set_real_num_tx_queues()
2226 qdisc_reset_all_tx_gt(dev, txq); in netif_set_real_num_tx_queues()
2228 netif_reset_xps_queues_gt(dev, txq); in netif_set_real_num_tx_queues()
2232 dev->real_num_tx_queues = txq; in netif_set_real_num_tx_queues()
2315 void netif_schedule_queue(struct netdev_queue *txq) in netif_schedule_queue() argument
2318 if (!(txq->state & QUEUE_STATE_ANY_XOFF)) { in netif_schedule_queue()
2319 struct Qdisc *q = rcu_dereference(txq->qdisc); in netif_schedule_queue()
2336 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); in netif_wake_subqueue() local
2338 if (test_and_clear_bit(__QUEUE_STATE_DRV_XOFF, &txq->state)) { in netif_wake_subqueue()
2342 q = rcu_dereference(txq->qdisc); in netif_wake_subqueue()
2945 struct netdev_queue *txq, bool more) in xmit_one() argument
2955 rc = netdev_start_xmit(skb, dev, txq, more); in xmit_one()
2962 struct netdev_queue *txq, int *ret) in dev_hard_start_xmit() argument
2971 rc = xmit_one(skb, dev, txq, next != NULL); in dev_hard_start_xmit()
2978 if (netif_xmit_stopped(txq) && skb) { in dev_hard_start_xmit()
3119 struct netdev_queue *txq) in __dev_xmit_skb() argument
3151 if (sch_direct_xmit(skb, q, dev, txq, root_lock, true)) { in __dev_xmit_skb()
3379 struct netdev_queue *txq; in __dev_queue_xmit() local
3414 txq = netdev_pick_tx(dev, skb, accel_priv); in __dev_queue_xmit()
3415 q = rcu_dereference_bh(txq->qdisc); in __dev_queue_xmit()
3419 rc = __dev_xmit_skb(skb, q, dev, txq); in __dev_queue_xmit()
3438 if (txq->xmit_lock_owner != cpu) { in __dev_queue_xmit()
3447 HARD_TX_LOCK(dev, txq, cpu); in __dev_queue_xmit()
3449 if (!netif_xmit_stopped(txq)) { in __dev_queue_xmit()
3451 skb = dev_hard_start_xmit(skb, dev, txq, &rc); in __dev_queue_xmit()
3454 HARD_TX_UNLOCK(dev, txq); in __dev_queue_xmit()
3458 HARD_TX_UNLOCK(dev, txq); in __dev_queue_xmit()
7182 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_tx_stop_all_queues() local
7183 netif_tx_stop_queue(txq); in netif_tx_stop_all_queues()