/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_iov.c | 76 i = iov_data->num_vfs; in fm10k_iov_event() 87 } while (i != iov_data->num_vfs); in fm10k_iov_event() 95 i = iov_data->next_vf_mbx ? : iov_data->num_vfs; in fm10k_iov_event() 143 for (i = iov_data->next_vf_mbx ? : iov_data->num_vfs; i--;) { in fm10k_iov_mbx() 194 int num_vfs, i; in fm10k_iov_suspend() local 197 num_vfs = iov_data ? iov_data->num_vfs : 0; in fm10k_iov_suspend() 204 for (i = 0; i < num_vfs; i++) { in fm10k_iov_suspend() 218 int num_vfs, i; in fm10k_iov_resume() local 221 num_vfs = iov_data ? iov_data->num_vfs : 0; in fm10k_iov_resume() 228 hw->iov.ops.assign_resources(hw, num_vfs, num_vfs); in fm10k_iov_resume() [all …]
|
D | fm10k_pf.c | 464 if (!hw->iov.num_vfs) in fm10k_update_int_moderator_pf() 602 u16 num_vfs = hw->iov.num_vfs; in fm10k_vf_queue_index() local 605 vf_q_idx -= fm10k_queues_per_pool(hw) * (num_vfs - vf_idx); in fm10k_vf_queue_index() 636 static s32 fm10k_iov_assign_resources_pf(struct fm10k_hw *hw, u16 num_vfs, in fm10k_iov_assign_resources_pf() argument 648 if ((num_vfs > num_pools) || (num_vfs > hw->iov.total_vfs)) in fm10k_iov_assign_resources_pf() 652 hw->iov.num_vfs = num_vfs; in fm10k_iov_assign_resources_pf() 656 qmap_stride = (num_vfs > 8) ? 32 : 256; in fm10k_iov_assign_resources_pf() 665 for (i = 0; i < num_vfs; i++) { in fm10k_iov_assign_resources_pf() 673 for (i = FM10K_VFMBMEM_LEN * num_vfs; i--;) in fm10k_iov_assign_resources_pf() 699 fm10k_vf_vector_index(hw, num_vfs - 1)); in fm10k_iov_assign_resources_pf() [all …]
|
D | fm10k.h | 224 unsigned int num_vfs; member 486 int fm10k_iov_configure(struct pci_dev *pdev, int num_vfs);
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_sriov.c | 55 e_info(probe, "SR-IOV enabled with %d VFs\n", adapter->num_vfs); in __ixgbe_enable_sriov() 61 adapter->ring_feature[RING_F_VMDQ].offset = adapter->num_vfs; in __ixgbe_enable_sriov() 64 (IXGBE_MAX_PF_MACVLANS + 1 + adapter->num_vfs); in __ixgbe_enable_sriov() 88 kcalloc(adapter->num_vfs, in __ixgbe_enable_sriov() 93 (adapter->num_vfs < 16)) { in __ixgbe_enable_sriov() 96 } else if (adapter->num_vfs < 32) { in __ixgbe_enable_sriov() 112 for (i = 0; i < adapter->num_vfs; i++) in __ixgbe_enable_sriov() 128 if (!pre_existing_vfs && !adapter->num_vfs) in ixgbe_enable_sriov() 138 adapter->num_vfs = pre_existing_vfs; in ixgbe_enable_sriov() 150 adapter->num_vfs = min_t(unsigned int, adapter->num_vfs, IXGBE_MAX_VFS_DRV_LIMIT); in ixgbe_enable_sriov() [all …]
|
D | ixgbe_sriov.h | 57 int ixgbe_pci_sriov_configure(struct pci_dev *dev, int num_vfs);
|
D | ixgbe_main.c | 2129 if (adapter->num_vfs > 32) { in ixgbe_configure_msix() 2130 u32 eitrsel = (1 << (adapter->num_vfs - 32)) - 1; in ixgbe_configure_msix() 3152 if (adapter->num_vfs || (adapter->num_rx_queues > 1 && in ixgbe_set_rx_drop_en() 3475 hw->mac.ops.set_mac_anti_spoofing(hw, (adapter->num_vfs != 0), in ixgbe_configure_virtualization() 3476 adapter->num_vfs); in ixgbe_configure_virtualization() 3478 for (i = 0; i < adapter->num_vfs; i++) { in ixgbe_configure_virtualization() 4930 if (adapter->num_vfs) { in ixgbe_down() 4935 for (i = 0 ; i < adapter->num_vfs; i++) in ixgbe_down() 5138 adapter->num_vfs = 0; in ixgbe_sw_init() 5141 adapter->num_vfs = max_vfs; in ixgbe_sw_init() [all …]
|
D | ixgbe_dcb_nl.c | 674 for (vf = 0; vf < adapter->num_vfs; vf++) { in ixgbe_dcbnl_ieee_setapp() 719 for (vf = 0; vf < adapter->num_vfs; vf++) { in ixgbe_dcbnl_ieee_delapp()
|
/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_sriov_pf.c | 75 u32 num_vfs, max, temp; in qlcnic_sriov_pf_cal_res_limit() local 81 num_vfs = sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit() 82 max = num_vfs + 1; in qlcnic_sriov_pf_cal_res_limit() 95 temp = res->num_rx_ucast_mac_filters - num_macs * num_vfs; in qlcnic_sriov_pf_cal_res_limit() 97 temp = res->num_tx_mac_filters - num_macs * num_vfs; in qlcnic_sriov_pf_cal_res_limit() 99 temp = num_macs * num_vfs * QLCNIC_SRIOV_VF_MAX_MAC; in qlcnic_sriov_pf_cal_res_limit() 103 info->max_tx_ques = res->num_tx_queues - sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit() 162 total_fn = sriov->num_vfs + 1; in qlcnic_sriov_set_vf_max_vlan() 412 for (i = 0; i < sriov->num_vfs; i++) in qlcnic_sriov_pf_del_flr_queue() 566 static int qlcnic_sriov_pf_enable(struct qlcnic_adapter *adapter, int num_vfs) in qlcnic_sriov_pf_enable() argument [all …]
|
D | qlcnic_sriov_common.c | 135 int qlcnic_sriov_init(struct qlcnic_adapter *adapter, int num_vfs) in qlcnic_sriov_init() argument 152 sriov->num_vfs = num_vfs; in qlcnic_sriov_init() 155 num_vfs, GFP_KERNEL); in qlcnic_sriov_init() 181 for (i = 0; i < num_vfs; i++) { in qlcnic_sriov_init() 264 for (i = 0; i < sriov->num_vfs; i++) { in __qlcnic_sriov_cleanup() 273 for (i = 0; i < sriov->num_vfs; i++) in __qlcnic_sriov_cleanup() 679 for (i = 0; i < adapter->ahw->sriov->num_vfs; i++) { in qlcnic_sriov_func_to_index() 2150 for (i = 0; i < sriov->num_vfs; i++) { in qlcnic_sriov_alloc_vlans() 2163 for (i = 0; i < sriov->num_vfs; i++) { in qlcnic_sriov_free_vlans()
|
D | qlcnic_sriov.h | 175 u8 num_vfs; member
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | port.c | 556 dev->num_vfs + 1); in mlx4_get_slave_num_gids() 558 vfs = bitmap_weight(slaves_pport.slaves, dev->num_vfs + 1) - 1; in mlx4_get_slave_num_gids() 593 dev->num_vfs + 1); in mlx4_get_base_gid_ix() 596 vfs = bitmap_weight(slaves_pport.slaves, dev->num_vfs + 1) - 1; in mlx4_get_base_gid_ix() 647 if (slave < 0 || slave > dev->num_vfs) in mlx4_reset_roce_gids() 1210 unsigned num_vfs; in mlx4_get_slave_from_roce_gid() local 1217 num_vfs = bitmap_weight(slaves_pport.slaves, dev->num_vfs + 1) - 1; in mlx4_get_slave_from_roce_gid() 1233 else if (found_ix < MLX4_ROCE_PF_GIDS + (vf_gids % num_vfs) * in mlx4_get_slave_from_roce_gid() 1234 (vf_gids / num_vfs + 1)) in mlx4_get_slave_from_roce_gid() 1236 (vf_gids / num_vfs + 1)) + 1; in mlx4_get_slave_from_roce_gid() [all …]
|
D | main.c | 80 static uint8_t num_vfs[3] = {0, 0, 0}; variable 82 module_param_array(num_vfs, byte , &num_vfs_argc, 0444); 83 MODULE_PARM_DESC(num_vfs, "enable #num_vfs functions if num_vfs > 0\n" 1568 (dev_cap->fs_max_num_qp_per_entry >= (dev->num_vfs + 1))) && in choose_steering_mode() 2333 dev->num_vfs = total_vfs; in mlx4_load_one() 2494 if (mlx4_is_master(dev) && dev->num_vfs) in mlx4_load_one() 2552 if (mlx4_is_master(dev) && dev->num_vfs) in mlx4_load_one() 2589 nvfs[param_map[num_vfs_argc - 1][i]] = num_vfs[i]; in __mlx4_init_one() 2797 dev->num_vfs = 0; in mlx4_unload_one() 2835 total_vfs = dev->num_vfs; in mlx4_restart_one()
|
D | eq.c | 240 if (dev->num_vfs < slave) in mlx4_gen_guid_change_eqe() 258 if (dev->num_vfs < slave) in mlx4_gen_port_state_change_eqe() 313 for (i = 0; i < dev->num_vfs + 1; i++) in set_all_slave_state() 563 for (i = 0; i < dev->num_vfs + 1; i++) { in mlx4_eq_int() 599 for (i = 0; i < dev->num_vfs + 1; i++) { in mlx4_eq_int()
|
D | resource_tracker.c | 312 if (slave > dev->num_vfs) in mlx4_grant_resource() 317 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] : in mlx4_grant_resource() 355 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] += count; in mlx4_grant_resource() 379 if (slave > dev->num_vfs) in mlx4_release_resource() 385 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] : in mlx4_release_resource() 400 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] -= count; in mlx4_release_resource() 418 res_alloc->guaranteed[vf] = num_instances / (2 * (dev->num_vfs + 1)); in initialize_res_quotas() 489 res_alloc->quota = kmalloc((dev->num_vfs + 1) * sizeof(int), GFP_KERNEL); in mlx4_init_resource_tracker() 490 res_alloc->guaranteed = kmalloc((dev->num_vfs + 1) * sizeof(int), GFP_KERNEL); in mlx4_init_resource_tracker() 493 (dev->num_vfs + 1) * sizeof(int), in mlx4_init_resource_tracker() [all …]
|
/drivers/pci/ |
D | pci-sysfs.c | 461 u16 num_vfs; in sriov_numvfs_store() local 463 ret = kstrtou16(buf, 0, &num_vfs); in sriov_numvfs_store() 467 if (num_vfs > pci_sriov_get_totalvfs(pdev)) in sriov_numvfs_store() 470 if (num_vfs == pdev->sriov->num_VFs) in sriov_numvfs_store() 479 if (num_vfs == 0) { in sriov_numvfs_store() 490 pdev->sriov->num_VFs, num_vfs); in sriov_numvfs_store() 494 ret = pdev->driver->sriov_configure(pdev, num_vfs); in sriov_numvfs_store() 498 if (ret != num_vfs) in sriov_numvfs_store() 500 num_vfs, ret); in sriov_numvfs_store()
|
/drivers/net/ethernet/emulex/benet/ |
D | be_main.c | 33 static unsigned int num_vfs; variable 34 module_param(num_vfs, uint, S_IRUGO); 35 MODULE_PARM_DESC(num_vfs, "Number of PCI VFs to initialize"); 1278 if (!is_valid_ether_addr(mac) || vf >= adapter->num_vfs) in be_set_vf_mac() 1318 if (vf >= adapter->num_vfs) in be_get_vf_config() 1341 if (vf >= adapter->num_vfs || vlan > 4095 || qos > 7) in be_set_vf_vlan() 1379 if (vf >= adapter->num_vfs) in be_set_vf_tx_rate() 1438 if (vf >= adapter->num_vfs) in be_set_vf_link_state() 3089 adapter->num_vfs = 0; in be_vf_clear() 3205 adapter->vf_cfg = kcalloc(adapter->num_vfs, sizeof(*vf_cfg), in be_vf_setup_init() [all …]
|
D | be_cmds.c | 3665 res->max_vfs = le16_to_cpu(pcie->num_vfs); in be_cmd_get_profile_config() 3751 pcie->num_vfs = 0xFFFF; in be_reset_pcie_desc() 3788 struct be_resources res, u16 num_vfs) in be_cmd_set_sriov_config() argument 3805 desc.pcie.sriov_state = num_vfs ? 1 : 0; in be_cmd_set_sriov_config() 3806 desc.pcie.num_vfs = cpu_to_le16(num_vfs); in be_cmd_set_sriov_config() 3817 if (num_vfs && res.vf_if_cap_flags & BE_IF_FLAGS_RSS) { in be_cmd_set_sriov_config() 3822 if (num_vfs < (be_max_vfs(adapter) - 8)) in be_cmd_set_sriov_config() 3823 vf_q_count = (res.max_rss_qs - 8) / num_vfs; in be_cmd_set_sriov_config() 3825 vf_q_count = res.max_rss_qs / num_vfs; in be_cmd_set_sriov_config()
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_virtchnl_pf.c | 906 static int i40e_pci_sriov_enable(struct pci_dev *pdev, int num_vfs) in i40e_pci_sriov_enable() argument 913 dev_info(&pdev->dev, "Allocating %d VFs.\n", num_vfs); in i40e_pci_sriov_enable() 914 if (pre_existing_vfs && pre_existing_vfs != num_vfs) in i40e_pci_sriov_enable() 916 else if (pre_existing_vfs && pre_existing_vfs == num_vfs) in i40e_pci_sriov_enable() 919 if (num_vfs > pf->num_req_vfs) { in i40e_pci_sriov_enable() 924 err = i40e_alloc_vfs(pf, num_vfs); in i40e_pci_sriov_enable() 931 return num_vfs; in i40e_pci_sriov_enable() 947 int i40e_pci_sriov_configure(struct pci_dev *pdev, int num_vfs) in i40e_pci_sriov_configure() argument 951 if (num_vfs) in i40e_pci_sriov_configure() 952 return i40e_pci_sriov_enable(pdev, num_vfs); in i40e_pci_sriov_configure()
|
D | i40e_virtchnl_pf.h | 108 int i40e_pci_sriov_configure(struct pci_dev *dev, int num_vfs);
|
/drivers/misc/genwqe/ |
D | card_debugfs.c | 203 for (vf_num = 0; vf_num < cd->num_vfs; vf_num++) { in genwqe_jtimer_show() 223 for (vf_num = 0; vf_num < cd->num_vfs; vf_num++) { in genwqe_queue_working_time_show()
|
D | card_base.c | 1132 cd->num_vfs = pci_sriov_get_totalvfs(pci_dev); in genwqe_pci_setup() 1133 if (cd->num_vfs < 0) in genwqe_pci_setup() 1134 cd->num_vfs = 0; in genwqe_pci_setup()
|
D | card_base.h | 311 int num_vfs; member
|
/drivers/net/ethernet/cisco/enic/ |
D | enic.h | 161 u16 num_vfs; member
|
D | enic_pp.c | 43 if (vf < 0 || vf >= enic->num_vfs) { in enic_is_valid_pp_vf()
|
/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.h | 525 int bnx2x_sriov_configure(struct pci_dev *dev, int num_vfs); 591 static inline int bnx2x_sriov_configure(struct pci_dev *dev, int num_vfs) {return 0; } in bnx2x_sriov_configure() argument
|