Home
last modified time | relevance | path

Searched refs:num_vfs (Results 1 – 25 of 43) sorted by relevance

12

/drivers/net/ethernet/intel/fm10k/
Dfm10k_iov.c76 i = iov_data->num_vfs; in fm10k_iov_event()
87 } while (i != iov_data->num_vfs); in fm10k_iov_event()
95 i = iov_data->next_vf_mbx ? : iov_data->num_vfs; in fm10k_iov_event()
143 for (i = iov_data->next_vf_mbx ? : iov_data->num_vfs; i--;) { in fm10k_iov_mbx()
194 int num_vfs, i; in fm10k_iov_suspend() local
197 num_vfs = iov_data ? iov_data->num_vfs : 0; in fm10k_iov_suspend()
204 for (i = 0; i < num_vfs; i++) { in fm10k_iov_suspend()
218 int num_vfs, i; in fm10k_iov_resume() local
221 num_vfs = iov_data ? iov_data->num_vfs : 0; in fm10k_iov_resume()
228 hw->iov.ops.assign_resources(hw, num_vfs, num_vfs); in fm10k_iov_resume()
[all …]
Dfm10k_pf.c464 if (!hw->iov.num_vfs) in fm10k_update_int_moderator_pf()
602 u16 num_vfs = hw->iov.num_vfs; in fm10k_vf_queue_index() local
605 vf_q_idx -= fm10k_queues_per_pool(hw) * (num_vfs - vf_idx); in fm10k_vf_queue_index()
636 static s32 fm10k_iov_assign_resources_pf(struct fm10k_hw *hw, u16 num_vfs, in fm10k_iov_assign_resources_pf() argument
648 if ((num_vfs > num_pools) || (num_vfs > hw->iov.total_vfs)) in fm10k_iov_assign_resources_pf()
652 hw->iov.num_vfs = num_vfs; in fm10k_iov_assign_resources_pf()
656 qmap_stride = (num_vfs > 8) ? 32 : 256; in fm10k_iov_assign_resources_pf()
665 for (i = 0; i < num_vfs; i++) { in fm10k_iov_assign_resources_pf()
673 for (i = FM10K_VFMBMEM_LEN * num_vfs; i--;) in fm10k_iov_assign_resources_pf()
699 fm10k_vf_vector_index(hw, num_vfs - 1)); in fm10k_iov_assign_resources_pf()
[all …]
Dfm10k.h224 unsigned int num_vfs; member
486 int fm10k_iov_configure(struct pci_dev *pdev, int num_vfs);
/drivers/net/ethernet/intel/ixgbe/
Dixgbe_sriov.c55 e_info(probe, "SR-IOV enabled with %d VFs\n", adapter->num_vfs); in __ixgbe_enable_sriov()
61 adapter->ring_feature[RING_F_VMDQ].offset = adapter->num_vfs; in __ixgbe_enable_sriov()
64 (IXGBE_MAX_PF_MACVLANS + 1 + adapter->num_vfs); in __ixgbe_enable_sriov()
88 kcalloc(adapter->num_vfs, in __ixgbe_enable_sriov()
93 (adapter->num_vfs < 16)) { in __ixgbe_enable_sriov()
96 } else if (adapter->num_vfs < 32) { in __ixgbe_enable_sriov()
112 for (i = 0; i < adapter->num_vfs; i++) in __ixgbe_enable_sriov()
128 if (!pre_existing_vfs && !adapter->num_vfs) in ixgbe_enable_sriov()
138 adapter->num_vfs = pre_existing_vfs; in ixgbe_enable_sriov()
150 adapter->num_vfs = min_t(unsigned int, adapter->num_vfs, IXGBE_MAX_VFS_DRV_LIMIT); in ixgbe_enable_sriov()
[all …]
Dixgbe_sriov.h57 int ixgbe_pci_sriov_configure(struct pci_dev *dev, int num_vfs);
Dixgbe_main.c2129 if (adapter->num_vfs > 32) { in ixgbe_configure_msix()
2130 u32 eitrsel = (1 << (adapter->num_vfs - 32)) - 1; in ixgbe_configure_msix()
3152 if (adapter->num_vfs || (adapter->num_rx_queues > 1 && in ixgbe_set_rx_drop_en()
3475 hw->mac.ops.set_mac_anti_spoofing(hw, (adapter->num_vfs != 0), in ixgbe_configure_virtualization()
3476 adapter->num_vfs); in ixgbe_configure_virtualization()
3478 for (i = 0; i < adapter->num_vfs; i++) { in ixgbe_configure_virtualization()
4930 if (adapter->num_vfs) { in ixgbe_down()
4935 for (i = 0 ; i < adapter->num_vfs; i++) in ixgbe_down()
5138 adapter->num_vfs = 0; in ixgbe_sw_init()
5141 adapter->num_vfs = max_vfs; in ixgbe_sw_init()
[all …]
Dixgbe_dcb_nl.c674 for (vf = 0; vf < adapter->num_vfs; vf++) { in ixgbe_dcbnl_ieee_setapp()
719 for (vf = 0; vf < adapter->num_vfs; vf++) { in ixgbe_dcbnl_ieee_delapp()
/drivers/net/ethernet/qlogic/qlcnic/
Dqlcnic_sriov_pf.c75 u32 num_vfs, max, temp; in qlcnic_sriov_pf_cal_res_limit() local
81 num_vfs = sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit()
82 max = num_vfs + 1; in qlcnic_sriov_pf_cal_res_limit()
95 temp = res->num_rx_ucast_mac_filters - num_macs * num_vfs; in qlcnic_sriov_pf_cal_res_limit()
97 temp = res->num_tx_mac_filters - num_macs * num_vfs; in qlcnic_sriov_pf_cal_res_limit()
99 temp = num_macs * num_vfs * QLCNIC_SRIOV_VF_MAX_MAC; in qlcnic_sriov_pf_cal_res_limit()
103 info->max_tx_ques = res->num_tx_queues - sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit()
162 total_fn = sriov->num_vfs + 1; in qlcnic_sriov_set_vf_max_vlan()
412 for (i = 0; i < sriov->num_vfs; i++) in qlcnic_sriov_pf_del_flr_queue()
566 static int qlcnic_sriov_pf_enable(struct qlcnic_adapter *adapter, int num_vfs) in qlcnic_sriov_pf_enable() argument
[all …]
Dqlcnic_sriov_common.c135 int qlcnic_sriov_init(struct qlcnic_adapter *adapter, int num_vfs) in qlcnic_sriov_init() argument
152 sriov->num_vfs = num_vfs; in qlcnic_sriov_init()
155 num_vfs, GFP_KERNEL); in qlcnic_sriov_init()
181 for (i = 0; i < num_vfs; i++) { in qlcnic_sriov_init()
264 for (i = 0; i < sriov->num_vfs; i++) { in __qlcnic_sriov_cleanup()
273 for (i = 0; i < sriov->num_vfs; i++) in __qlcnic_sriov_cleanup()
679 for (i = 0; i < adapter->ahw->sriov->num_vfs; i++) { in qlcnic_sriov_func_to_index()
2150 for (i = 0; i < sriov->num_vfs; i++) { in qlcnic_sriov_alloc_vlans()
2163 for (i = 0; i < sriov->num_vfs; i++) { in qlcnic_sriov_free_vlans()
Dqlcnic_sriov.h175 u8 num_vfs; member
/drivers/net/ethernet/mellanox/mlx4/
Dport.c556 dev->num_vfs + 1); in mlx4_get_slave_num_gids()
558 vfs = bitmap_weight(slaves_pport.slaves, dev->num_vfs + 1) - 1; in mlx4_get_slave_num_gids()
593 dev->num_vfs + 1); in mlx4_get_base_gid_ix()
596 vfs = bitmap_weight(slaves_pport.slaves, dev->num_vfs + 1) - 1; in mlx4_get_base_gid_ix()
647 if (slave < 0 || slave > dev->num_vfs) in mlx4_reset_roce_gids()
1210 unsigned num_vfs; in mlx4_get_slave_from_roce_gid() local
1217 num_vfs = bitmap_weight(slaves_pport.slaves, dev->num_vfs + 1) - 1; in mlx4_get_slave_from_roce_gid()
1233 else if (found_ix < MLX4_ROCE_PF_GIDS + (vf_gids % num_vfs) * in mlx4_get_slave_from_roce_gid()
1234 (vf_gids / num_vfs + 1)) in mlx4_get_slave_from_roce_gid()
1236 (vf_gids / num_vfs + 1)) + 1; in mlx4_get_slave_from_roce_gid()
[all …]
Dmain.c80 static uint8_t num_vfs[3] = {0, 0, 0}; variable
82 module_param_array(num_vfs, byte , &num_vfs_argc, 0444);
83 MODULE_PARM_DESC(num_vfs, "enable #num_vfs functions if num_vfs > 0\n"
1568 (dev_cap->fs_max_num_qp_per_entry >= (dev->num_vfs + 1))) && in choose_steering_mode()
2333 dev->num_vfs = total_vfs; in mlx4_load_one()
2494 if (mlx4_is_master(dev) && dev->num_vfs) in mlx4_load_one()
2552 if (mlx4_is_master(dev) && dev->num_vfs) in mlx4_load_one()
2589 nvfs[param_map[num_vfs_argc - 1][i]] = num_vfs[i]; in __mlx4_init_one()
2797 dev->num_vfs = 0; in mlx4_unload_one()
2835 total_vfs = dev->num_vfs; in mlx4_restart_one()
Deq.c240 if (dev->num_vfs < slave) in mlx4_gen_guid_change_eqe()
258 if (dev->num_vfs < slave) in mlx4_gen_port_state_change_eqe()
313 for (i = 0; i < dev->num_vfs + 1; i++) in set_all_slave_state()
563 for (i = 0; i < dev->num_vfs + 1; i++) { in mlx4_eq_int()
599 for (i = 0; i < dev->num_vfs + 1; i++) { in mlx4_eq_int()
Dresource_tracker.c312 if (slave > dev->num_vfs) in mlx4_grant_resource()
317 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] : in mlx4_grant_resource()
355 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] += count; in mlx4_grant_resource()
379 if (slave > dev->num_vfs) in mlx4_release_resource()
385 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] : in mlx4_release_resource()
400 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] -= count; in mlx4_release_resource()
418 res_alloc->guaranteed[vf] = num_instances / (2 * (dev->num_vfs + 1)); in initialize_res_quotas()
489 res_alloc->quota = kmalloc((dev->num_vfs + 1) * sizeof(int), GFP_KERNEL); in mlx4_init_resource_tracker()
490 res_alloc->guaranteed = kmalloc((dev->num_vfs + 1) * sizeof(int), GFP_KERNEL); in mlx4_init_resource_tracker()
493 (dev->num_vfs + 1) * sizeof(int), in mlx4_init_resource_tracker()
[all …]
/drivers/pci/
Dpci-sysfs.c461 u16 num_vfs; in sriov_numvfs_store() local
463 ret = kstrtou16(buf, 0, &num_vfs); in sriov_numvfs_store()
467 if (num_vfs > pci_sriov_get_totalvfs(pdev)) in sriov_numvfs_store()
470 if (num_vfs == pdev->sriov->num_VFs) in sriov_numvfs_store()
479 if (num_vfs == 0) { in sriov_numvfs_store()
490 pdev->sriov->num_VFs, num_vfs); in sriov_numvfs_store()
494 ret = pdev->driver->sriov_configure(pdev, num_vfs); in sriov_numvfs_store()
498 if (ret != num_vfs) in sriov_numvfs_store()
500 num_vfs, ret); in sriov_numvfs_store()
/drivers/net/ethernet/emulex/benet/
Dbe_main.c33 static unsigned int num_vfs; variable
34 module_param(num_vfs, uint, S_IRUGO);
35 MODULE_PARM_DESC(num_vfs, "Number of PCI VFs to initialize");
1278 if (!is_valid_ether_addr(mac) || vf >= adapter->num_vfs) in be_set_vf_mac()
1318 if (vf >= adapter->num_vfs) in be_get_vf_config()
1341 if (vf >= adapter->num_vfs || vlan > 4095 || qos > 7) in be_set_vf_vlan()
1379 if (vf >= adapter->num_vfs) in be_set_vf_tx_rate()
1438 if (vf >= adapter->num_vfs) in be_set_vf_link_state()
3089 adapter->num_vfs = 0; in be_vf_clear()
3205 adapter->vf_cfg = kcalloc(adapter->num_vfs, sizeof(*vf_cfg), in be_vf_setup_init()
[all …]
Dbe_cmds.c3665 res->max_vfs = le16_to_cpu(pcie->num_vfs); in be_cmd_get_profile_config()
3751 pcie->num_vfs = 0xFFFF; in be_reset_pcie_desc()
3788 struct be_resources res, u16 num_vfs) in be_cmd_set_sriov_config() argument
3805 desc.pcie.sriov_state = num_vfs ? 1 : 0; in be_cmd_set_sriov_config()
3806 desc.pcie.num_vfs = cpu_to_le16(num_vfs); in be_cmd_set_sriov_config()
3817 if (num_vfs && res.vf_if_cap_flags & BE_IF_FLAGS_RSS) { in be_cmd_set_sriov_config()
3822 if (num_vfs < (be_max_vfs(adapter) - 8)) in be_cmd_set_sriov_config()
3823 vf_q_count = (res.max_rss_qs - 8) / num_vfs; in be_cmd_set_sriov_config()
3825 vf_q_count = res.max_rss_qs / num_vfs; in be_cmd_set_sriov_config()
/drivers/net/ethernet/intel/i40e/
Di40e_virtchnl_pf.c906 static int i40e_pci_sriov_enable(struct pci_dev *pdev, int num_vfs) in i40e_pci_sriov_enable() argument
913 dev_info(&pdev->dev, "Allocating %d VFs.\n", num_vfs); in i40e_pci_sriov_enable()
914 if (pre_existing_vfs && pre_existing_vfs != num_vfs) in i40e_pci_sriov_enable()
916 else if (pre_existing_vfs && pre_existing_vfs == num_vfs) in i40e_pci_sriov_enable()
919 if (num_vfs > pf->num_req_vfs) { in i40e_pci_sriov_enable()
924 err = i40e_alloc_vfs(pf, num_vfs); in i40e_pci_sriov_enable()
931 return num_vfs; in i40e_pci_sriov_enable()
947 int i40e_pci_sriov_configure(struct pci_dev *pdev, int num_vfs) in i40e_pci_sriov_configure() argument
951 if (num_vfs) in i40e_pci_sriov_configure()
952 return i40e_pci_sriov_enable(pdev, num_vfs); in i40e_pci_sriov_configure()
Di40e_virtchnl_pf.h108 int i40e_pci_sriov_configure(struct pci_dev *dev, int num_vfs);
/drivers/misc/genwqe/
Dcard_debugfs.c203 for (vf_num = 0; vf_num < cd->num_vfs; vf_num++) { in genwqe_jtimer_show()
223 for (vf_num = 0; vf_num < cd->num_vfs; vf_num++) { in genwqe_queue_working_time_show()
Dcard_base.c1132 cd->num_vfs = pci_sriov_get_totalvfs(pci_dev); in genwqe_pci_setup()
1133 if (cd->num_vfs < 0) in genwqe_pci_setup()
1134 cd->num_vfs = 0; in genwqe_pci_setup()
Dcard_base.h311 int num_vfs; member
/drivers/net/ethernet/cisco/enic/
Denic.h161 u16 num_vfs; member
Denic_pp.c43 if (vf < 0 || vf >= enic->num_vfs) { in enic_is_valid_pp_vf()
/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_sriov.h525 int bnx2x_sriov_configure(struct pci_dev *dev, int num_vfs);
591 static inline int bnx2x_sriov_configure(struct pci_dev *dev, int num_vfs) {return 0; } in bnx2x_sriov_configure() argument

12