/kernel/linux/linux-5.10/net/sched/ |
D | sch_mq.c | 119 if (ntx < dev->real_num_tx_queues) in mq_attach() 135 for (i = new_real_tx; i < dev->real_num_tx_queues; i++) { in mq_change_real_num_tx() 143 for (i = dev->real_num_tx_queues; i < new_real_tx; i++) { in mq_change_real_num_tx()
|
D | sch_mqprio.c | 97 if (qopt->offset[i] >= dev->real_num_tx_queues || in mqprio_parse_opt() 99 last > dev->real_num_tx_queues) in mqprio_parse_opt() 302 if (ntx < dev->real_num_tx_queues) in mqprio_attach() 316 for (i = new_real_tx; i < dev->real_num_tx_queues; i++) { in mqprio_change_real_num_tx() 324 for (i = dev->real_num_tx_queues; i < new_real_tx; i++) { in mqprio_change_real_num_tx()
|
D | sch_multiq.c | 187 qopt->bands = qdisc_dev(sch)->real_num_tx_queues; in multiq_tune()
|
D | act_skbedit.c | 61 skb->dev->real_num_tx_queues > params->queue_mapping) in tcf_skbedit_act()
|
D | sch_taprio.c | 965 last > dev->real_num_tx_queues) { in taprio_parse_mqprio_opt() 1721 if (i < dev->real_num_tx_queues) in taprio_init()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/liquidio/ |
D | octeon_network.h | 563 for (i = 0; i < netdev->real_num_tx_queues; i++) in stop_txqs() 576 for (i = 0; i < netdev->real_num_tx_queues; i++) { in wake_txqs() 597 for (i = 0; i < netdev->real_num_tx_queues; i++) in start_txqs()
|
/kernel/linux/linux-5.10/drivers/net/ |
D | xen-netfront.c | 356 unsigned int num_queues = dev->real_num_tx_queues; in xennet_open() 583 unsigned int num_queues = dev->real_num_tx_queues; in xennet_select_queue() 643 unsigned int num_queues = dev->real_num_tx_queues; in xennet_xdp_xmit() 691 unsigned int num_queues = dev->real_num_tx_queues; in xennet_start_xmit() 836 unsigned int num_queues = dev->real_num_tx_queues; in xennet_close() 1525 unsigned int num_queues = dev->real_num_tx_queues; in xennet_poll_controller() 1586 bpf_prog_add(prog, dev->real_num_tx_queues); in xennet_xdp_set() 1588 for (i = 0; i < dev->real_num_tx_queues; ++i) in xennet_xdp_set() 1592 for (i = 0; i < dev->real_num_tx_queues; ++i) in xennet_xdp_set() 1745 unsigned int num_queues = info->netdev->real_num_tx_queues; in xennet_disconnect_backend() [all …]
|
D | net_failover.c | 144 if (unlikely(txq >= dev->real_num_tx_queues)) { in net_failover_select_queue() 146 txq -= dev->real_num_tx_queues; in net_failover_select_queue() 147 } while (txq >= dev->real_num_tx_queues); in net_failover_select_queue()
|
D | veth.c | 150 for (i = 0; i < dev->real_num_tx_queues; i++) { in veth_get_strings() 168 VETH_TQ_STATS_LEN * dev->real_num_tx_queues; in veth_get_sset_count() 209 tx_idx += (i % dev->real_num_tx_queues) * VETH_TQ_STATS_LEN; in veth_get_ethtool_stats() 1163 if (dev->real_num_rx_queues < peer->real_num_tx_queues) { in veth_xdp_set()
|
/kernel/linux/linux-5.10/drivers/s390/net/ |
D | qeth_ethtool.c | 206 channels->tx_count = dev->real_num_tx_queues; in qeth_get_channels() 235 channels->tx_count < dev->real_num_tx_queues) in qeth_set_channels()
|
/kernel/linux/linux-5.10/net/xdp/ |
D | xsk.c | 101 if (queue_id < dev->real_num_tx_queues) in xsk_get_pool_from_qid() 125 dev->real_num_tx_queues)) in xsk_reg_pool_at_qid() 130 if (queue_id < dev->real_num_tx_queues) in xsk_reg_pool_at_qid() 387 if (xs->queue_id >= xs->dev->real_num_tx_queues) in xsk_generic_xmit()
|
/kernel/linux/linux-5.10/net/core/ |
D | netpoll.c | 120 if (unlikely(q_index >= dev->real_num_tx_queues)) { in queue_process() 121 q_index = q_index % dev->real_num_tx_queues; in queue_process()
|
D | net-sysfs.c | 1748 real_tx = dev->real_num_tx_queues; in register_queue_kobjects() 1783 real_tx = ndev->real_num_tx_queues; in queue_change_owner() 1803 real_tx = dev->real_num_tx_queues; in remove_queue_kobjects() 1809 dev->real_num_tx_queues = 0; in remove_queue_kobjects()
|
D | dev.c | 2913 if ((offset + count) > dev->real_num_tx_queues) in netdev_bind_sb_channel_queue() 2959 disabling = txq < dev->real_num_tx_queues; in netif_set_real_num_tx_queues() 2968 rc = netdev_queue_update_kobjects(dev, dev->real_num_tx_queues, in netif_set_real_num_tx_queues() 2978 dev->real_num_tx_queues = txq; in netif_set_real_num_tx_queues() 2988 dev->real_num_tx_queues = txq; in netif_set_real_num_tx_queues() 3169 u16 qcount = dev->real_num_tx_queues; in skb_tx_hash() 3180 qcount = dev->real_num_tx_queues; in skb_tx_hash() 3943 if (unlikely(queue_index >= dev->real_num_tx_queues)) in __get_xps_queue_idx() 4002 return (u16)raw_smp_processor_id() % dev->real_num_tx_queues; in dev_pick_tx_cpu_id() 4015 queue_index >= dev->real_num_tx_queues) { in netdev_pick_tx() [all …]
|
/kernel/linux/linux-5.10/drivers/net/xen-netback/ |
D | interface.c | 210 dev->real_num_tx_queues; in xenvif_select_queue() 215 return skb_get_hash_raw(skb) % dev->real_num_tx_queues; in xenvif_select_queue()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/myricom/myri10ge/ |
D | myri10ge.c | 975 if (mgp->dev->real_num_tx_queues > 1) in myri10ge_reset() 983 if (status != 0 && mgp->dev->real_num_tx_queues > 1) { in myri10ge_reset() 1424 if ((ss->mgp->dev->real_num_tx_queues > 1) && in myri10ge_tx_done() 1549 if ((mgp->dev->real_num_tx_queues == 1) && (ss != mgp->ss)) { in myri10ge_intr() 2231 if (slice == 0 || (mgp->dev->real_num_tx_queues > 1)) { in myri10ge_get_txrx() 2307 if (mgp->dev->real_num_tx_queues > 1) in myri10ge_open() 2396 if (slice == 0 || mgp->dev->real_num_tx_queues > 1) in myri10ge_open() 2848 if ((mgp->dev->real_num_tx_queues > 1) && tx->queue_active == 0) { in myri10ge_xmit()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/pensando/ionic/ |
D | ionic_stats.c | 177 #define MAX_Q(lif) ((lif)->netdev->real_num_tx_queues)
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/i40e/ |
D | i40e_xsk.c | 55 qid >= netdev->real_num_tx_queues) in i40e_xsk_pool_enable()
|
/kernel/linux/linux-5.10/drivers/net/hyperv/ |
D | netvsc_drv.c | 330 if (q_idx < 0 || skb->ooo_okay || q_idx >= ndev->real_num_tx_queues) { in netvsc_pick_tx() 370 while (txq >= ndev->real_num_tx_queues) in netvsc_select_queue() 371 txq -= ndev->real_num_tx_queues; in netvsc_select_queue() 600 if (hash != 0 && net->real_num_tx_queues > 1) { in netvsc_xmit()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_xsk.c | 35 qid >= netdev->real_num_tx_queues) in ixgbe_xsk_pool_enable()
|
/kernel/linux/linux-5.10/drivers/net/team/ |
D | team.c | 1540 if (new_queue_id >= team->dev->real_num_tx_queues) in team_queue_id_option_set() 1744 if (unlikely(txq >= dev->real_num_tx_queues)) { in team_select_queue() 1746 txq -= dev->real_num_tx_queues; in team_select_queue() 1747 } while (txq >= dev->real_num_tx_queues); in team_select_queue()
|
/kernel/linux/linux-5.10/net/netfilter/ |
D | xt_set.c | 427 skb->dev->real_num_tx_queues > MOPT(map_opt, skbqueue)) in set_target_v3()
|
/kernel/linux/linux-5.10/include/net/ |
D | sch_generic.h | 632 return ntx < dev->real_num_tx_queues ? in get_default_qdisc_ops()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/ti/ |
D | netcp_core.c | 1895 if ((dev->real_num_tx_queues <= 1) || in netcp_setup_tc() 1896 (dev->real_num_tx_queues < num_tc)) in netcp_setup_tc()
|
/kernel/linux/linux-5.10/include/linux/ |
D | netdevice.h | 2066 unsigned int real_num_tx_queues; member 3576 if (unlikely(queue_index >= dev->real_num_tx_queues)) { in netdev_cap_txqueue() 3579 dev->real_num_tx_queues); in netdev_cap_txqueue()
|