/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_dcbnl.c | 38 u8 num_tc = 0; in fm10k_dcbnl_ieee_setets() local 47 if (ets->prio_tc[i] > num_tc) in fm10k_dcbnl_ieee_setets() 48 num_tc = ets->prio_tc[i]; in fm10k_dcbnl_ieee_setets() 52 if (num_tc) in fm10k_dcbnl_ieee_setets() 53 num_tc++; in fm10k_dcbnl_ieee_setets() 55 if (num_tc > IEEE_8021QAZ_MAX_TCS) in fm10k_dcbnl_ieee_setets() 59 if (num_tc != netdev_get_num_tc(dev)) { in fm10k_dcbnl_ieee_setets() 60 int err = fm10k_setup_tc(dev, num_tc); in fm10k_dcbnl_ieee_setets()
|
/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
D | hclge_dcb.c | 75 static int hclge_dcb_common_validate(struct hclge_dev *hdev, u8 num_tc, in hclge_dcb_common_validate() argument 80 if (num_tc > hdev->tc_max) { in hclge_dcb_common_validate() 83 num_tc, hdev->tc_max); in hclge_dcb_common_validate() 88 if (prio_tc[i] >= num_tc) { in hclge_dcb_common_validate() 91 i, prio_tc[i], num_tc); in hclge_dcb_common_validate() 96 if (num_tc > hdev->vport[0].alloc_tqps) { in hclge_dcb_common_validate() 99 num_tc, hdev->vport[0].alloc_tqps); in hclge_dcb_common_validate() 151 if (*tc != hdev->tm_info.num_tc) in hclge_ets_validate() 193 ret = client->ops->setup_tc(handle, hdev->tm_info.num_tc); in hclge_client_setup_tc() 229 u8 num_tc = 0; in hclge_ieee_setets() local [all …]
|
D | hclge_tm.c | 233 if (tc >= hdev->tm_info.num_tc) in hclge_fill_pri_array() 524 kinfo->num_tc = vport->vport_id ? 1 : in hclge_tm_vport_tc_info_update() 525 min_t(u16, vport->alloc_tqps, hdev->tm_info.num_tc); in hclge_tm_vport_tc_info_update() 526 vport->qs_offset = (vport->vport_id ? hdev->tm_info.num_tc : 0) + in hclge_tm_vport_tc_info_update() 530 vport->alloc_tqps / kinfo->num_tc); in hclge_tm_vport_tc_info_update() 547 kinfo->num_tc); in hclge_tm_vport_tc_info_update() 553 kinfo->num_tqps = kinfo->num_tc * kinfo->rss_size; in hclge_tm_vport_tc_info_update() 559 if (hdev->hw_tc_map & BIT(i) && i < kinfo->num_tc) { in hclge_tm_vport_tc_info_update() 593 for (i = 0; i < hdev->tm_info.num_tc; i++) { in hclge_tm_tc_info_init() 603 (i >= hdev->tm_info.num_tc) ? 0 : i; in hclge_tm_tc_info_init() [all …]
|
D | hclge_tm.h | 149 void hclge_tm_schd_info_update(struct hclge_dev *hdev, u8 num_tc);
|
/drivers/net/ethernet/sfc/falcon/ |
D | tx.c | 432 unsigned tc, num_tc; in ef4_setup_tc() local 438 num_tc = mqprio->num_tc; in ef4_setup_tc() 440 if (ef4_nic_rev(efx) < EF4_REV_FALCON_B0 || num_tc > EF4_MAX_TX_TC) in ef4_setup_tc() 445 if (num_tc == net_dev->num_tc) in ef4_setup_tc() 448 for (tc = 0; tc < num_tc; tc++) { in ef4_setup_tc() 453 if (num_tc > net_dev->num_tc) { in ef4_setup_tc() 472 net_dev->num_tc = num_tc; in ef4_setup_tc() 476 max_t(int, num_tc, 1) * in ef4_setup_tc() 487 net_dev->num_tc = num_tc; in ef4_setup_tc()
|
/drivers/net/ethernet/sfc/ |
D | tx.c | 689 unsigned tc, num_tc; in efx_setup_tc() local 695 num_tc = mqprio->num_tc; in efx_setup_tc() 697 if (num_tc > EFX_MAX_TX_TC) in efx_setup_tc() 702 if (num_tc == net_dev->num_tc) in efx_setup_tc() 705 for (tc = 0; tc < num_tc; tc++) { in efx_setup_tc() 710 if (num_tc > net_dev->num_tc) { in efx_setup_tc() 729 net_dev->num_tc = num_tc; in efx_setup_tc() 733 max_t(int, num_tc, 1) * in efx_setup_tc() 744 net_dev->num_tc = num_tc; in efx_setup_tc()
|
/drivers/net/ethernet/intel/ice/ |
D | ice_dcb_lib.c | 54 u8 i, num_tc, ena_tc = 1; in ice_dcb_get_ena_tc() local 56 num_tc = ice_dcb_get_num_tc(dcbcfg); in ice_dcb_get_ena_tc() 58 for (i = 0; i < num_tc; i++) in ice_dcb_get_ena_tc() 71 u8 num_tc = 0; in ice_dcb_get_num_tc() local 79 num_tc |= BIT(dcbcfg->etscfg.prio_table[i]); in ice_dcb_get_num_tc() 83 if (num_tc & BIT(i)) { in ice_dcb_get_num_tc()
|
/drivers/net/ethernet/intel/iavf/ |
D | iavf_main.c | 1107 adapter->num_tc) in iavf_alloc_queues() 1442 adapter->num_tc) in iavf_init_interrupt_scheme() 1444 adapter->num_tc); in iavf_init_interrupt_scheme() 2198 adapter->num_tc) { in iavf_reset_task() 2530 if (mqprio_qopt->qopt.num_tc > IAVF_MAX_TRAFFIC_CLASS || in iavf_validate_ch_config() 2531 mqprio_qopt->qopt.num_tc < 1) in iavf_validate_ch_config() 2534 for (i = 0; i <= mqprio_qopt->qopt.num_tc - 1; i++) { in iavf_validate_ch_config() 2590 u8 num_tc = 0, total_qps = 0; in __iavf_setup_tc() local 2596 num_tc = mqprio_qopt->qopt.num_tc; in __iavf_setup_tc() 2604 adapter->num_tc = 0; in __iavf_setup_tc() [all …]
|
D | iavf_virtchnl.c | 983 len = struct_size(vti, list, adapter->num_tc - 1); in iavf_enable_channels() 987 vti->num_tc = adapter->num_tc; in iavf_enable_channels() 988 for (i = 0; i < vti->num_tc; i++) { in iavf_enable_channels()
|
D | iavf.h | 342 u8 num_tc; member
|
/drivers/net/ethernet/qlogic/qede/ |
D | qede_main.c | 530 static int qede_setup_tc(struct net_device *ndev, u8 num_tc) in qede_setup_tc() argument 535 if (num_tc > edev->dev_info.num_tc) in qede_setup_tc() 539 netdev_set_num_tc(ndev, num_tc); in qede_setup_tc() 601 return qede_setup_tc(dev, mqprio->num_tc); in qede_setup_tc_offload() 696 info->num_queues * info->num_tc, in qede_alloc_etherdev() 906 fp->txq = kcalloc(edev->dev_info.num_tc, in qede_alloc_fp_array() 1715 edev->dev_info.num_tc); in qede_set_real_num_queues() 2245 u8 num_tc; in qede_load() local 2292 num_tc = netdev_get_num_tc(edev->ndev); in qede_load() 2293 num_tc = num_tc ? num_tc : edev->dev_info.num_tc; in qede_load() [all …]
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_virtchnl_pf.c | 260 for (i = 0; i < vf->num_tc; i++) { in i40e_get_real_pf_qid() 803 u32 reg, num_tc = 1; /* VF has at least one traffic class */ in i40e_map_pf_queues_to_vsi() local 808 num_tc = vf->num_tc; in i40e_map_pf_queues_to_vsi() 810 for (i = 0; i < num_tc; i++) { in i40e_map_pf_queues_to_vsi() 851 u32 qps, num_tc = 1; /* VF has at least one traffic class */ in i40e_map_pf_to_vf_queues() local 856 num_tc = vf->num_tc; in i40e_map_pf_to_vf_queues() 858 for (i = 0; i < num_tc; i++) { in i40e_map_pf_to_vf_queues() 963 for (j = 0; j < vf->num_tc; j++) { in i40e_free_vf_res() 1038 for (idx = 1; idx < vf->num_tc; idx++) { in i40e_alloc_vf_res() 1831 for (i = 1; i < vf->num_tc; i++) { in i40e_del_qch() [all …]
|
D | i40e_virtchnl_pf.h | 109 u8 num_tc; member
|
/drivers/net/ethernet/freescale/dpaa/ |
D | dpaa_eth.c | 354 u8 num_tc; in dpaa_setup_tc() local 361 num_tc = mqprio->num_tc; in dpaa_setup_tc() 363 if (num_tc == priv->num_tc) in dpaa_setup_tc() 366 if (!num_tc) { in dpaa_setup_tc() 371 if (num_tc > DPAA_TC_NUM) { in dpaa_setup_tc() 377 netdev_set_num_tc(net_dev, num_tc); in dpaa_setup_tc() 379 for (i = 0; i < num_tc; i++) in dpaa_setup_tc() 384 priv->num_tc = num_tc ? : 1; in dpaa_setup_tc() 385 netif_set_real_num_tx_queues(net_dev, priv->num_tc * DPAA_TC_TXQ_NUM); in dpaa_setup_tc() 2919 priv->num_tc = 1; in dpaa_eth_probe() [all …]
|
D | dpaa_eth.h | 162 u8 num_tc; member
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_txrx.c | 125 for (i = 0; i < c->num_tc; i++) in mlx5e_napi_poll() 168 for (i = 0; i < c->num_tc; i++) { in mlx5e_napi_poll()
|
D | en_main.c | 1668 for (tc = 0; tc < c->num_tc; tc++) { in mlx5e_open_tx_cqs() 1688 for (tc = 0; tc < c->num_tc; tc++) in mlx5e_close_tx_cqs() 1698 for (tc = 0; tc < params->num_tc; tc++) { in mlx5e_open_sqs() 1720 for (tc = 0; tc < c->num_tc; tc++) in mlx5e_close_sqs() 1972 c->num_tc = params->num_tc; in mlx5e_open_channel() 2016 for (tc = 0; tc < c->num_tc; tc++) in mlx5e_activate_channel() 2035 for (tc = 0; tc < c->num_tc; tc++) in mlx5e_deactivate_channel() 2863 int ntc = priv->channels.params.num_tc; in mlx5e_netdev_set_tcs() 2889 for (tc = 0; tc < priv->channels.params.num_tc; tc++) { in mlx5e_build_txq_maps() 2901 int num_txqs = priv->channels.num * priv->channels.params.num_tc; in mlx5e_activate_priv_channels() [all …]
|
/drivers/net/ethernet/freescale/enetc/ |
D | enetc.c | 1436 u8 num_tc; in enetc_setup_tc() local 1443 num_tc = mqprio->num_tc; in enetc_setup_tc() 1445 if (!num_tc) { in enetc_setup_tc() 1459 if (num_tc > priv->num_tx_rings) { in enetc_setup_tc() 1469 for (i = 0; i < num_tc; i++) { in enetc_setup_tc() 1475 netif_set_real_num_tx_queues(ndev, num_tc); in enetc_setup_tc() 1477 netdev_set_num_tc(ndev, num_tc); in enetc_setup_tc() 1480 for (i = 0; i < num_tc; i++) in enetc_setup_tc()
|
/drivers/net/ethernet/freescale/dpaa2/ |
D | dpaa2-eth.c | 819 if (net_dev->num_tc) { in dpaa2_eth_tx() 824 prio = net_dev->num_tc - prio - 1; in dpaa2_eth_tx() 1954 netdev_queues = (net_dev->num_tc ? : 1) * num_queues; in update_xps() 1980 u8 num_tc, num_queues; in dpaa2_eth_setup_tc() local 1988 num_tc = mqprio->num_tc; in dpaa2_eth_setup_tc() 1990 if (num_tc == net_dev->num_tc) in dpaa2_eth_setup_tc() 1993 if (num_tc > dpaa2_eth_tc_count(priv)) { in dpaa2_eth_setup_tc() 1999 if (!num_tc) { in dpaa2_eth_setup_tc() 2005 netdev_set_num_tc(net_dev, num_tc); in dpaa2_eth_setup_tc() 2006 netif_set_real_num_tx_queues(net_dev, num_tc * num_queues); in dpaa2_eth_setup_tc() [all …]
|
/drivers/dma/ti/ |
D | edma.c | 250 unsigned num_tc; member 705 if (eventq_no >= ecc->num_tc) in edma_assign_channel_eventq() 2007 ecc->num_tc = value + 1; in edma_setup_from_hw() 2016 dev_dbg(dev, "num_tc: %u\n", ecc->num_tc); in edma_setup_from_hw() 2033 queue_priority_map = devm_kcalloc(dev, ecc->num_tc + 1, sizeof(s8), in edma_setup_from_hw() 2038 for (i = 0; i < ecc->num_tc; i++) { in edma_setup_from_hw() 2221 dma_spec->args[1] < echan->ecc->num_tc) { in of_edma_xlate() 2404 ecc->tc_list = devm_kcalloc(dev, ecc->num_tc, in edma_probe() 2414 if (ret || i == ecc->num_tc) in edma_probe()
|
/drivers/net/ethernet/ti/ |
D | cpsw.c | 1413 static int cpsw_tc_to_fifo(int tc, int num_tc) in cpsw_tc_to_fifo() argument 1415 if (tc == num_tc - 1) in cpsw_tc_to_fifo() 1547 fifo = cpsw_tc_to_fifo(tc, ndev->num_tc); in cpsw_set_cbs() 2274 int fifo, num_tc, count, offset; in cpsw_set_mqprio() local 2279 num_tc = mqprio->qopt.num_tc; in cpsw_set_mqprio() 2280 if (num_tc > CPSW_TC_NUM) in cpsw_set_mqprio() 2292 if (num_tc) { in cpsw_set_mqprio() 2295 fifo = cpsw_tc_to_fifo(tc, num_tc); in cpsw_set_mqprio() 2299 netdev_set_num_tc(ndev, num_tc); in cpsw_set_mqprio() 2300 for (i = 0; i < num_tc; i++) { in cpsw_set_mqprio()
|
D | netcp_core.c | 1882 u8 num_tc; in netcp_setup_tc() local 1892 num_tc = mqprio->num_tc; in netcp_setup_tc() 1896 (dev->real_num_tx_queues < num_tc)) in netcp_setup_tc() 1900 if (num_tc) { in netcp_setup_tc() 1901 netdev_set_num_tc(dev, num_tc); in netcp_setup_tc() 1902 for (i = 0; i < num_tc; i++) in netcp_setup_tc()
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_sriov.c | 272 u8 num_tc; in ixgbe_pci_sriov_enable() local 294 num_tc = adapter->hw_tcs; in ixgbe_pci_sriov_enable() 297 limit = (num_tc > 4) ? IXGBE_MAX_VFS_8TC : in ixgbe_pci_sriov_enable() 298 (num_tc > 1) ? IXGBE_MAX_VFS_4TC : IXGBE_MAX_VFS_1TC; in ixgbe_pci_sriov_enable() 302 num_tc, num_rx_pools - 1, limit - num_rx_pools); in ixgbe_pci_sriov_enable()
|
/drivers/net/ethernet/hisilicon/hns3/hns3vf/ |
D | hclgevf_main.c | 390 kinfo->num_tc = 0; in hclgevf_knic_setup() 396 kinfo->num_tc++; in hclgevf_knic_setup() 399 = min_t(u16, hdev->rss_size_max, new_tqps / kinfo->num_tc); in hclgevf_knic_setup() 400 new_tqps = kinfo->rss_size * kinfo->num_tc; in hclgevf_knic_setup() 418 kinfo->rss_size = min_t(u16, kinfo->num_tqps / kinfo->num_tc, in hclgevf_knic_setup() 2828 hdev->num_tqps / kinfo->num_tc); in hclgevf_get_max_channels() 2871 hdev->num_tqps / kinfo->num_tc); in hclgevf_update_rss_size() 2883 kinfo->num_tqps = kinfo->num_tc * kinfo->rss_size; in hclgevf_update_rss_size() 2927 cur_tqps, kinfo->rss_size * kinfo->num_tc); in hclgevf_set_channels()
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | reporter_tx.c | 259 for (tc = 0; tc < priv->channels.params.num_tc; tc++) { in mlx5e_tx_reporter_diagnose()
|