/drivers/infiniband/hw/mlx4/ |
D | alias_GUID.c | 81 guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid. in mlx4_ib_update_cache_on_guid_change() 97 memcpy(&dev->sriov.demux[port_index].guid_cache[slave_id], in mlx4_ib_update_cache_on_guid_change() 112 return *(__be64 *)&dev->sriov.demux[port - 1].guid_cache[index]; in get_cached_alias_guid() 145 guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid. in mlx4_ib_notify_slaves_on_guid_change() 215 rec = &dev->sriov.alias_guid.ports_guid[port_index]. in aliasguid_query_handler() 235 rec = &dev->sriov.alias_guid.ports_guid[port_index]. in aliasguid_query_handler() 298 spin_lock_irqsave(&dev->sriov.going_down_lock, flags); in aliasguid_query_handler() 299 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1); in aliasguid_query_handler() 300 if (!dev->sriov.is_going_down) in aliasguid_query_handler() 301 queue_delayed_work(dev->sriov.alias_guid.ports_guid[port_index].wq, in aliasguid_query_handler() [all …]
|
D | cm.c | 107 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in id_map_find_by_sl_id() 134 struct mlx4_ib_sriov *sriov = &dev->sriov; in id_map_ent_timeout() local 135 struct rb_root *sl_id_map = &sriov->sl_id_map; in id_map_ent_timeout() 138 spin_lock(&sriov->id_map_lock); in id_map_ent_timeout() 139 db_ent = (struct id_map_entry *)idr_find(&sriov->pv_id_table, pv_id); in id_map_ent_timeout() 145 idr_remove(&sriov->pv_id_table, pv_id); in id_map_ent_timeout() 149 spin_unlock(&sriov->id_map_lock); in id_map_ent_timeout() 155 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_find_del() local 156 struct rb_root *sl_id_map = &sriov->sl_id_map; in id_map_find_del() 159 spin_lock(&sriov->id_map_lock); in id_map_find_del() [all …]
|
D | mad.c | 275 if (!dev->sriov.is_going_down) in smp_snoop() 288 !dev->sriov.is_going_down) { in smp_snoop() 405 if (dev->sriov.demux[port - 1].guid_cache[i] == guid) in mlx4_ib_find_real_gid() 474 tun_ctx = dev->sriov.demux[port-1].tun[slave]; in mlx4_ib_send_to_slave() 863 if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) in handle_lid_change_event() 874 if (!dev->sriov.is_going_down) { in handle_client_rereg_event() 875 mlx4_ib_mcg_port_cleanup(&dev->sriov.demux[port_num - 1], 0); in handle_client_rereg_event() 984 if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) in handle_port_mgmt_change_event() 992 else if (!dev->sriov.is_going_down) { in handle_port_mgmt_change_event() 1023 spin_lock_irqsave(&dev->sriov.going_down_lock, flags); in mlx4_ib_tunnel_comp_handler() [all …]
|
D | sysfs.c | 60 be64_to_cpu(*(__be64 *)&mdev->sriov.alias_guid. in show_admin_alias_guid() 91 *(__be64 *)&mdev->sriov.alias_guid.ports_guid[port->num - 1]. in store_admin_alias_guid() 97 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].status in store_admin_alias_guid() 100 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].method in store_admin_alias_guid() 105 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].method in store_admin_alias_guid() 107 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].ownership in store_admin_alias_guid() 112 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].ownership in store_admin_alias_guid() 117 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].ownership in store_admin_alias_guid() 123 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].guid_indexes in store_admin_alias_guid()
|
D | main.c | 1617 spin_lock_irqsave(&ibdev->sriov.going_down_lock, flags); in do_slave_init() 1618 if (!ibdev->sriov.is_going_down) in do_slave_init() 1619 queue_work(ibdev->sriov.demux[i].ud_wq, &dm[i]->work); in do_slave_init() 1620 spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); in do_slave_init()
|
D | mcg.c | 885 struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1]; in mlx4_ib_mcg_demux_handler() 934 struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1]; in mlx4_ib_mcg_multiplex_handler()
|
D | mlx4_ib.h | 478 struct mlx4_ib_sriov sriov; member
|
D | qp.c | 1774 to_mdev(ib_dev)->sriov.demux[sqp->qp.port - 1]. in build_mlx_header() 1777 to_mdev(ib_dev)->sriov.demux[sqp->qp.port - 1]. in build_mlx_header() 1810 to_mdev(ib_dev)->sriov.demux[sqp->qp.port - 1]. in build_mlx_header() 1813 to_mdev(ib_dev)->sriov.demux[sqp->qp.port - 1]. in build_mlx_header()
|
/drivers/pci/ |
D | iov.c | 24 return dev->bus->number + ((dev->devfn + dev->sriov->offset + in virtfn_bus() 25 dev->sriov->stride * id) >> 8); in virtfn_bus() 30 return (dev->devfn + dev->sriov->offset + in virtfn_devfn() 31 dev->sriov->stride * id) & 0xff; in virtfn_devfn() 77 struct pci_sriov *iov = dev->sriov; in virtfn_add() 83 mutex_lock(&iov->dev->sriov->lock); in virtfn_add() 87 mutex_unlock(&iov->dev->sriov->lock); in virtfn_add() 114 mutex_unlock(&iov->dev->sriov->lock); in virtfn_add() 136 mutex_lock(&iov->dev->sriov->lock); in virtfn_add() 139 mutex_unlock(&iov->dev->sriov->lock); in virtfn_add() [all …]
|
D | ats.c | 70 mutex_lock(&pdev->sriov->lock); in pci_enable_ats() 78 mutex_unlock(&pdev->sriov->lock); in pci_enable_ats() 119 mutex_lock(&pdev->sriov->lock); in pci_disable_ats() 123 mutex_unlock(&pdev->sriov->lock); in pci_disable_ats()
|
D | pci-sysfs.c | 421 return sprintf(buf, "%u\n", pdev->sriov->num_VFs); in sriov_numvfs_show() 446 if (num_vfs == pdev->sriov->num_VFs) in sriov_numvfs_store() 464 if (pdev->sriov->num_VFs) { in sriov_numvfs_store() 466 pdev->sriov->num_VFs, num_vfs); in sriov_numvfs_store()
|
/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_sriov_common.c | 127 struct qlcnic_sriov *sriov; in qlcnic_sriov_init() local 137 sriov = kzalloc(sizeof(struct qlcnic_sriov), GFP_KERNEL); in qlcnic_sriov_init() 138 if (!sriov) in qlcnic_sriov_init() 141 adapter->ahw->sriov = sriov; in qlcnic_sriov_init() 142 sriov->num_vfs = num_vfs; in qlcnic_sriov_init() 143 bc = &sriov->bc; in qlcnic_sriov_init() 144 sriov->vf_info = kzalloc(sizeof(struct qlcnic_vf_info) * in qlcnic_sriov_init() 146 if (!sriov->vf_info) { in qlcnic_sriov_init() 172 vf = &sriov->vf_info[i]; in qlcnic_sriov_init() 188 sriov->vf_info[i].vp = vp; in qlcnic_sriov_init() [all …]
|
D | qlcnic_sriov_pf.c | 65 struct qlcnic_sriov *sriov = adapter->ahw->sriov; in qlcnic_sriov_pf_cal_res_limit() local 66 struct qlcnic_resources *res = &sriov->ff_max; in qlcnic_sriov_pf_cal_res_limit() 75 num_vfs = sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit() 93 vp = sriov->vf_info[id].vp; in qlcnic_sriov_pf_cal_res_limit() 119 struct qlcnic_resources *ff_max = &adapter->ahw->sriov->ff_max; in qlcnic_sriov_pf_set_ff_max_res() 195 struct qlcnic_sriov *sriov = adapter->ahw->sriov; in qlcnic_sriov_pf_reset_vport_handle() local 200 sriov->vp_handle = 0; in qlcnic_sriov_pf_reset_vport_handle() 205 vp = sriov->vf_info[index].vp; in qlcnic_sriov_pf_reset_vport_handle() 213 struct qlcnic_sriov *sriov = adapter->ahw->sriov; in qlcnic_sriov_pf_set_vport_handle() local 218 sriov->vp_handle = vport_handle; in qlcnic_sriov_pf_set_vport_handle() [all …]
|
D | qlcnic_sriov.h | 252 static inline void qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov, in qlcnic_sriov_pf_handle_flr() argument
|
D | qlcnic.h | 461 struct qlcnic_sriov *sriov; member
|
D | qlcnic_main.c | 1543 qlcnic_sriov_cleanup_async_list(&adapter->ahw->sriov->bc); in __qlcnic_down()
|
/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.h | 280 #define BNX2X_NR_VIRTFN(bp) ((bp)->vfdb->sriov.nr_virtfn) 384 struct bnx2x_sriov sriov; member 761 return bp->vfdb->sriov.nr_virtfn * BNX2X_CLIENTS_PER_VF; in bnx2x_vf_headroom()
|
D | bnx2x_sriov.c | 1765 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_bus() 1774 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_devfn() 1783 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_set_bars() 1966 iov = &(bp->vfdb->sriov); in bnx2x_iov_init_one() 2070 tot_size = (BP_VFDB(bp)->sriov.first_vf_in_pf + BNX2X_NR_VIRTFN(bp)) * in bnx2x_iov_alloc_mem() 2177 DP(BNX2X_MSG_IOV, "num of vfs: %d\n", (bp)->vfdb->sriov.nr_virtfn); in bnx2x_iov_nic_init() 2183 int base_vf_cid = (BP_VFDB(bp)->sriov.first_vf_in_pf + vfid) * in bnx2x_iov_nic_init() 2675 int base_vf_cid = (BP_VFDB(bp)->sriov.first_vf_in_pf + vf->index) * in bnx2x_vf_acquire() 2988 bar_info->nr_bars = bp->vfdb->sriov.nres; in bnx2x_vf_get_bars()
|
D | bnx2x_vfpf.c | 1630 if (vfpf_event->vf_id - BP_VFDB(bp)->sriov.first_vf_in_pf > in bnx2x_vf_mbx()
|
/drivers/net/ethernet/sfc/ |
D | siena_sriov.c | 1224 unsigned index, devfn, sriov, buftbl_base; in efx_sriov_vfs_init() local 1229 sriov = pci_find_ext_capability(pci_dev, PCI_EXT_CAP_ID_SRIOV); in efx_sriov_vfs_init() 1230 if (!sriov) in efx_sriov_vfs_init() 1233 pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_OFFSET, &offset); in efx_sriov_vfs_init() 1234 pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_STRIDE, &stride); in efx_sriov_vfs_init()
|
/drivers/net/hyperv/ |
D | hyperv_net.h | 402 u64 sriov:1; member
|
/drivers/scsi/bfa/ |
D | bfa_defs.h | 749 u8 sriov; member
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_main.c | 5547 goto sriov; in init_one() 5749 sriov: in init_one()
|