Home
last modified time | relevance | path

Searched refs:sriov (Results 1 – 25 of 25) sorted by relevance

/drivers/infiniband/hw/mlx4/
Dalias_GUID.c86 guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid. in mlx4_ib_update_cache_on_guid_change()
102 memcpy(&dev->sriov.demux[port_index].guid_cache[slave_id], in mlx4_ib_update_cache_on_guid_change()
117 return *(__be64 *)&dev->sriov.demux[port - 1].guid_cache[index]; in get_cached_alias_guid()
136 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags); in mlx4_ib_slave_alias_guid_event()
137 if (dev->sriov.alias_guid.ports_guid[port_index].state_flags & in mlx4_ib_slave_alias_guid_event()
141 curr_guid = *(__be64 *)&dev->sriov. in mlx4_ib_slave_alias_guid_event()
154 *(__be64 *)&dev->sriov.alias_guid.ports_guid[port_index]. in mlx4_ib_slave_alias_guid_event()
157 dev->sriov.alias_guid.ports_guid[port_index]. in mlx4_ib_slave_alias_guid_event()
160 dev->sriov.alias_guid.ports_guid[port_index]. in mlx4_ib_slave_alias_guid_event()
164 dev->sriov.alias_guid.ports_guid[port_index]. in mlx4_ib_slave_alias_guid_event()
[all …]
Dcm.c146 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in id_map_find_by_sl_id()
173 struct mlx4_ib_sriov *sriov = &dev->sriov; in id_map_ent_timeout() local
174 struct rb_root *sl_id_map = &sriov->sl_id_map; in id_map_ent_timeout()
177 spin_lock(&sriov->id_map_lock); in id_map_ent_timeout()
178 db_ent = (struct id_map_entry *)idr_find(&sriov->pv_id_table, pv_id); in id_map_ent_timeout()
184 idr_remove(&sriov->pv_id_table, pv_id); in id_map_ent_timeout()
188 spin_unlock(&sriov->id_map_lock); in id_map_ent_timeout()
194 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_find_del() local
195 struct rb_root *sl_id_map = &sriov->sl_id_map; in id_map_find_del()
198 spin_lock(&sriov->id_map_lock); in id_map_find_del()
[all …]
Dmad.c276 if (!dev->sriov.is_going_down) in smp_snoop()
289 !dev->sriov.is_going_down) { in smp_snoop()
407 if (dev->sriov.demux[port - 1].guid_cache[i] == guid) in mlx4_ib_find_real_gid()
478 tun_ctx = dev->sriov.demux[port-1].tun[slave]; in mlx4_ib_send_to_slave()
975 if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) in handle_lid_change_event()
986 if (!dev->sriov.is_going_down) { in handle_client_rereg_event()
987 mlx4_ib_mcg_port_cleanup(&dev->sriov.demux[port_num - 1], 0); in handle_client_rereg_event()
1098 (u64)atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix), in handle_port_mgmt_change_event()
1100 atomic64_set(&dev->sriov.demux[port - 1].subnet_prefix, in handle_port_mgmt_change_event()
1117 if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) in handle_port_mgmt_change_event()
[all …]
Dsysfs.c87 spin_lock_irqsave(&mdev->sriov.alias_guid.ag_work_lock, flags); in store_admin_alias_guid()
89 *(__be64 *)&mdev->sriov.alias_guid.ports_guid[port->num - 1]. in store_admin_alias_guid()
95 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].status in store_admin_alias_guid()
102 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].guid_indexes in store_admin_alias_guid()
105 spin_unlock_irqrestore(&mdev->sriov.alias_guid.ag_work_lock, flags); in store_admin_alias_guid()
Dmain.c2650 spin_lock_irqsave(&ibdev->sriov.going_down_lock, flags); in do_slave_init()
2651 if (!ibdev->sriov.is_going_down) { in do_slave_init()
2653 queue_work(ibdev->sriov.demux[i].ud_wq, &dm[i]->work); in do_slave_init()
2654 spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); in do_slave_init()
2656 spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); in do_slave_init()
Dmcg.c891 struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1]; in mlx4_ib_mcg_demux_handler()
940 struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1]; in mlx4_ib_mcg_multiplex_handler()
Dmlx4_ib.h557 struct mlx4_ib_sriov sriov; member
Dqp.c2349 cpu_to_be64(atomic64_read(&(to_mdev(ib_dev)->sriov. in build_mlx_header()
2353 to_mdev(ib_dev)->sriov.demux[sqp->qp.port - 1]. in build_mlx_header()
/drivers/pci/
Diov.c26 return dev->bus->number + ((dev->devfn + dev->sriov->offset + in pci_iov_virtfn_bus()
27 dev->sriov->stride * vf_id) >> 8); in pci_iov_virtfn_bus()
34 return (dev->devfn + dev->sriov->offset + in pci_iov_virtfn_devfn()
35 dev->sriov->stride * vf_id) & 0xff; in pci_iov_virtfn_devfn()
46 struct pci_sriov *iov = dev->sriov; in pci_iov_set_numvfs()
62 struct pci_sriov *iov = dev->sriov; in compute_max_vf_buses()
113 return dev->sriov->barsz[resno - PCI_IOV_RESOURCES]; in pci_iov_resource_size()
124 struct pci_sriov *iov = dev->sriov; in virtfn_add()
127 mutex_lock(&iov->dev->sriov->lock); in virtfn_add()
162 mutex_unlock(&iov->dev->sriov->lock); in virtfn_add()
[all …]
Dpci-sysfs.c457 return sprintf(buf, "%u\n", pdev->sriov->num_VFs); in sriov_numvfs_show()
482 if (num_vfs == pdev->sriov->num_VFs) in sriov_numvfs_store()
500 if (pdev->sriov->num_VFs) { in sriov_numvfs_store()
502 pdev->sriov->num_VFs, num_vfs); in sriov_numvfs_store()
/drivers/net/ethernet/qlogic/qlcnic/
Dqlcnic_sriov_common.c140 struct qlcnic_sriov *sriov; in qlcnic_sriov_init() local
150 sriov = kzalloc(sizeof(struct qlcnic_sriov), GFP_KERNEL); in qlcnic_sriov_init()
151 if (!sriov) in qlcnic_sriov_init()
154 adapter->ahw->sriov = sriov; in qlcnic_sriov_init()
155 sriov->num_vfs = num_vfs; in qlcnic_sriov_init()
156 bc = &sriov->bc; in qlcnic_sriov_init()
157 sriov->vf_info = kzalloc(sizeof(struct qlcnic_vf_info) * in qlcnic_sriov_init()
159 if (!sriov->vf_info) { in qlcnic_sriov_init()
185 vf = &sriov->vf_info[i]; in qlcnic_sriov_init()
204 sriov->vf_info[i].vp = vp; in qlcnic_sriov_init()
[all …]
Dqlcnic_sriov_pf.c71 struct qlcnic_sriov *sriov = adapter->ahw->sriov; in qlcnic_sriov_pf_cal_res_limit() local
72 struct qlcnic_resources *res = &sriov->ff_max; in qlcnic_sriov_pf_cal_res_limit()
73 u16 num_macs = sriov->num_allowed_vlans + 1; in qlcnic_sriov_pf_cal_res_limit()
82 num_vfs = sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit()
104 info->max_tx_ques = res->num_tx_queues - sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit()
109 vp = sriov->vf_info[id].vp; in qlcnic_sriov_pf_cal_res_limit()
140 struct qlcnic_resources *ff_max = &adapter->ahw->sriov->ff_max; in qlcnic_sriov_pf_set_ff_max_res()
159 struct qlcnic_sriov *sriov = adapter->ahw->sriov; in qlcnic_sriov_set_vf_max_vlan() local
163 total_fn = sriov->num_vfs + 1; in qlcnic_sriov_set_vf_max_vlan()
166 sriov->num_allowed_vlans = temp - 1; in qlcnic_sriov_set_vf_max_vlan()
[all …]
Dqlcnic_sriov.h266 static inline void qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov, in qlcnic_sriov_pf_handle_flr() argument
Dqlcnic.h530 struct qlcnic_sriov *sriov; member
Dqlcnic_main.c1960 qlcnic_sriov_cleanup_async_list(&adapter->ahw->sriov->bc); in __qlcnic_down()
/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_sriov.h216 #define BNX2X_NR_VIRTFN(bp) ((bp)->vfdb->sriov.nr_virtfn)
238 #define GET_NUM_VFS_PER_PF(bp) ((bp)->vfdb ? (bp)->vfdb->sriov.total \
330 struct bnx2x_sriov sriov; member
538 return bp->vfdb->sriov.nr_virtfn * BNX2X_CIDS_PER_VF; in bnx2x_vf_headroom()
Dbnx2x_sriov.c1050 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_bus()
1059 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_devfn()
1068 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_set_bars()
1242 iov = &(bp->vfdb->sriov); in bnx2x_iov_init_one()
1320 for (vf_idx = 0; vf_idx < bp->vfdb->sriov.total; vf_idx++) { in bnx2x_iov_remove_one()
1323 bp->vfdb->sriov.first_vf_in_pf + in bnx2x_iov_remove_one()
1326 bp->vfdb->sriov.first_vf_in_pf + vf_idx); in bnx2x_iov_remove_one()
1370 tot_size = (BP_VFDB(bp)->sriov.first_vf_in_pf + BNX2X_NR_VIRTFN(bp)) * in bnx2x_iov_alloc_mem()
1556 DP(BNX2X_MSG_IOV, "num of vfs: %d\n", (bp)->vfdb->sriov.nr_virtfn); in bnx2x_iov_nic_init()
1565 int base_vf_cid = (BP_VFDB(bp)->sriov.first_vf_in_pf + vfid) * in bnx2x_iov_nic_init()
[all …]
Dbnx2x_vfpf.c2190 if (vfpf_event->vf_id - BP_VFDB(bp)->sriov.first_vf_in_pf > in bnx2x_vf_mbx_schedule()
Dbnx2x_main.c3403 ether_stat->vf_cnt = IS_SRIOV(bp) ? bp->vfdb->sriov.nr_virtfn : 0; in bnx2x_drv_info_ether_stat()
/drivers/net/ethernet/sfc/
DMakefile6 sfc-$(CONFIG_SFC_SRIOV) += sriov.o siena_sriov.o ef10_sriov.o
Dsiena_sriov.c1246 unsigned index, devfn, sriov, buftbl_base; in efx_siena_sriov_vfs_init() local
1251 sriov = pci_find_ext_capability(pci_dev, PCI_EXT_CAP_ID_SRIOV); in efx_siena_sriov_vfs_init()
1252 if (!sriov) in efx_siena_sriov_vfs_init()
1255 pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_OFFSET, &offset); in efx_siena_sriov_vfs_init()
1256 pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_STRIDE, &stride); in efx_siena_sriov_vfs_init()
/drivers/net/hyperv/
Dhyperv_net.h496 u64 sriov:1; member
Dnetvsc.c464 init_packet->msg.v2_msg.send_ndis_config.capability.sriov = 1; in negotiate_nvsp_ver()
/drivers/scsi/bfa/
Dbfa_defs.h776 u8 sriov; member
/drivers/net/ethernet/chelsio/cxgb4/
Dcxgb4_main.c4695 goto sriov; in init_one()
4953 sriov: in init_one()