Lines Matching refs:p_vf
30 static u8 qed_vf_calculate_legacy(struct qed_vf_info *p_vf) in qed_vf_calculate_legacy() argument
34 if (p_vf->acquire.vfdev_info.eth_fp_hsi_minor == in qed_vf_calculate_legacy()
38 if (!(p_vf->acquire.vfdev_info.capabilities & in qed_vf_calculate_legacy()
46 static int qed_sp_vf_start(struct qed_hwfn *p_hwfn, struct qed_vf_info *p_vf) in qed_sp_vf_start() argument
57 init_data.opaque_fid = p_vf->opaque_fid; in qed_sp_vf_start()
68 p_ramrod->vf_id = GET_FIELD(p_vf->concrete_fid, PXP_CONCRETE_FID_VFID); in qed_sp_vf_start()
69 p_ramrod->opaque_fid = cpu_to_le16(p_vf->opaque_fid); in qed_sp_vf_start()
86 fp_minor = p_vf->acquire.vfdev_info.eth_fp_hsi_minor; in qed_sp_vf_start()
92 p_vf->abs_vf_id, in qed_sp_vf_start()
103 p_vf->abs_vf_id, ETH_HSI_VER_MAJOR, fp_minor); in qed_sp_vf_start()
200 struct qed_vf_info *p_vf, in qed_iov_validate_queue_mode() argument
213 p_qcid = &p_vf->vf_queues[qid].cids[i]; in qed_iov_validate_queue_mode()
229 struct qed_vf_info *p_vf, in qed_iov_validate_rxq() argument
233 if (rx_qid >= p_vf->num_rxqs) { in qed_iov_validate_rxq()
237 p_vf->abs_vf_id, rx_qid, p_vf->num_rxqs); in qed_iov_validate_rxq()
241 return qed_iov_validate_queue_mode(p_hwfn, p_vf, rx_qid, mode, false); in qed_iov_validate_rxq()
245 struct qed_vf_info *p_vf, in qed_iov_validate_txq() argument
249 if (tx_qid >= p_vf->num_txqs) { in qed_iov_validate_txq()
253 p_vf->abs_vf_id, tx_qid, p_vf->num_txqs); in qed_iov_validate_txq()
257 return qed_iov_validate_queue_mode(p_hwfn, p_vf, tx_qid, mode, true); in qed_iov_validate_txq()
261 struct qed_vf_info *p_vf, u16 sb_idx) in qed_iov_validate_sb() argument
265 for (i = 0; i < p_vf->num_sbs; i++) in qed_iov_validate_sb()
266 if (p_vf->igu_sbs[i] == sb_idx) in qed_iov_validate_sb()
272 p_vf->abs_vf_id, sb_idx, p_vf->num_sbs); in qed_iov_validate_sb()
278 struct qed_vf_info *p_vf) in qed_iov_validate_active_rxq() argument
282 for (i = 0; i < p_vf->num_rxqs; i++) in qed_iov_validate_active_rxq()
283 if (qed_iov_validate_queue_mode(p_hwfn, p_vf, i, in qed_iov_validate_active_rxq()
292 struct qed_vf_info *p_vf) in qed_iov_validate_active_txq() argument
296 for (i = 0; i < p_vf->num_txqs; i++) in qed_iov_validate_active_txq()
297 if (qed_iov_validate_queue_mode(p_hwfn, p_vf, i, in qed_iov_validate_active_txq()
311 struct qed_vf_info *p_vf; in qed_iov_post_vf_bulletin() local
313 p_vf = qed_iov_get_vf_info(p_hwfn, (u16)vfid, true); in qed_iov_post_vf_bulletin()
314 if (!p_vf) in qed_iov_post_vf_bulletin()
317 if (!p_vf->vf_bulletin) in qed_iov_post_vf_bulletin()
320 p_bulletin = p_vf->bulletin.p_virt; in qed_iov_post_vf_bulletin()
325 p_vf->bulletin.size - crc_size); in qed_iov_post_vf_bulletin()
329 p_bulletin->version, p_vf->relative_vf_id, p_bulletin->crc); in qed_iov_post_vf_bulletin()
334 params.dst_vfid = p_vf->abs_vf_id; in qed_iov_post_vf_bulletin()
335 return qed_dmae_host2host(p_hwfn, p_ptt, p_vf->bulletin.phys, in qed_iov_post_vf_bulletin()
336 p_vf->vf_bulletin, p_vf->bulletin.size / 4, in qed_iov_post_vf_bulletin()
766 struct qed_vf_info *p_vf; in qed_iov_enable_vf_access_msix() local
768 p_vf = qed_iov_get_vf_info(p_hwfn, (u16)i, true); in qed_iov_enable_vf_access_msix()
769 if (!p_vf) in qed_iov_enable_vf_access_msix()
772 current_max = max_t(u8, current_max, p_vf->num_sbs); in qed_iov_enable_vf_access_msix()
944 struct qed_vf_info *p_vf = qed_iov_get_vf_info(p_hwfn, in qed_iov_set_link() local
949 if (!p_vf) in qed_iov_set_link()
952 p_bulletin = p_vf->bulletin.p_virt; in qed_iov_set_link()
1197 struct qed_vf_info *p_vf, in qed_iov_send_response() argument
1200 struct qed_iov_vf_mbx *mbx = &p_vf->vf_mbx; in qed_iov_send_response()
1208 eng_vf_id = p_vf->abs_vf_id; in qed_iov_send_response()
1259 struct qed_vf_info *p_vf, in qed_iov_prep_vp_update_resp_tlvs() argument
1290 p_vf->relative_vf_id, in qed_iov_prep_vp_update_resp_tlvs()
1349 struct qed_vf_info *p_vf) in qed_iov_vf_cleanup() argument
1353 p_vf->vf_bulletin = 0; in qed_iov_vf_cleanup()
1354 p_vf->vport_instance = 0; in qed_iov_vf_cleanup()
1355 p_vf->configured_features = 0; in qed_iov_vf_cleanup()
1358 p_vf->num_rxqs = p_vf->num_sbs; in qed_iov_vf_cleanup()
1359 p_vf->num_txqs = p_vf->num_sbs; in qed_iov_vf_cleanup()
1361 p_vf->num_active_rxqs = 0; in qed_iov_vf_cleanup()
1364 struct qed_vf_queue *p_queue = &p_vf->vf_queues[i]; in qed_iov_vf_cleanup()
1376 memset(&p_vf->shadow_config, 0, sizeof(p_vf->shadow_config)); in qed_iov_vf_cleanup()
1377 memset(&p_vf->acquire, 0, sizeof(p_vf->acquire)); in qed_iov_vf_cleanup()
1378 qed_iov_clean_vf(p_hwfn, p_vf->relative_vf_id); in qed_iov_vf_cleanup()
1395 struct qed_vf_info *p_vf, in qed_iov_vf_mbx_acquire_resc_cids() argument
1410 if (!(p_vf->acquire.vfdev_info.capabilities & in qed_iov_vf_mbx_acquire_resc_cids()
1418 if (p_vf->acquire.vfdev_info.capabilities & in qed_iov_vf_mbx_acquire_resc_cids()
1437 struct qed_vf_info *p_vf, in qed_iov_vf_mbx_acquire_resc() argument
1444 p_resp->num_rxqs = p_vf->num_rxqs; in qed_iov_vf_mbx_acquire_resc()
1445 p_resp->num_txqs = p_vf->num_txqs; in qed_iov_vf_mbx_acquire_resc()
1446 p_resp->num_sbs = p_vf->num_sbs; in qed_iov_vf_mbx_acquire_resc()
1449 p_resp->hw_sbs[i].hw_sb_id = p_vf->igu_sbs[i]; in qed_iov_vf_mbx_acquire_resc()
1457 qed_fw_l2_queue(p_hwfn, p_vf->vf_queues[i].fw_rx_qid, in qed_iov_vf_mbx_acquire_resc()
1463 p_resp->num_mac_filters = min_t(u8, p_vf->num_mac_filters, in qed_iov_vf_mbx_acquire_resc()
1465 p_resp->num_vlan_filters = min_t(u8, p_vf->num_vlan_filters, in qed_iov_vf_mbx_acquire_resc()
1468 qed_iov_vf_mbx_acquire_resc_cids(p_hwfn, p_ptt, p_vf, p_req, p_resp); in qed_iov_vf_mbx_acquire_resc()
1486 p_vf->abs_vf_id, in qed_iov_vf_mbx_acquire_resc()
1504 if ((p_vf->acquire.vfdev_info.eth_fp_hsi_minor == in qed_iov_vf_mbx_acquire_resc()
1506 (p_vf->acquire.vfdev_info.os_type == in qed_iov_vf_mbx_acquire_resc()
1687 struct qed_vf_info *p_vf, bool val) in __qed_iov_spoofchk_set() argument
1692 if (val == p_vf->spoof_chk) { in __qed_iov_spoofchk_set()
1699 params.opaque_fid = p_vf->opaque_fid; in __qed_iov_spoofchk_set()
1700 params.vport_id = p_vf->vport_id; in __qed_iov_spoofchk_set()
1706 p_vf->spoof_chk = val; in __qed_iov_spoofchk_set()
1707 p_vf->req_spoofchk_val = p_vf->spoof_chk; in __qed_iov_spoofchk_set()
1713 val, p_vf->relative_vf_id); in __qed_iov_spoofchk_set()
1720 struct qed_vf_info *p_vf) in qed_iov_reconfigure_unicast_vlan() argument
1729 filter.vport_to_add_to = p_vf->vport_id; in qed_iov_reconfigure_unicast_vlan()
1734 if (!p_vf->shadow_config.vlans[i].used) in qed_iov_reconfigure_unicast_vlan()
1738 filter.vlan = p_vf->shadow_config.vlans[i].vid; in qed_iov_reconfigure_unicast_vlan()
1741 filter.vlan, p_vf->relative_vf_id); in qed_iov_reconfigure_unicast_vlan()
1742 rc = qed_sp_eth_filter_ucast(p_hwfn, p_vf->opaque_fid, in qed_iov_reconfigure_unicast_vlan()
1747 filter.vlan, p_vf->relative_vf_id); in qed_iov_reconfigure_unicast_vlan()
1757 struct qed_vf_info *p_vf, u64 events) in qed_iov_reconfigure_unicast_shadow() argument
1762 !(p_vf->configured_features & (1 << VLAN_ADDR_FORCED))) in qed_iov_reconfigure_unicast_shadow()
1763 rc = qed_iov_reconfigure_unicast_vlan(p_hwfn, p_vf); in qed_iov_reconfigure_unicast_shadow()
1769 struct qed_vf_info *p_vf, u64 events) in qed_iov_configure_vport_forced() argument
1774 if (!p_vf->vport_instance) in qed_iov_configure_vport_forced()
1778 p_vf->p_vf_info.is_trusted_configured) { in qed_iov_configure_vport_forced()
1787 filter.vport_to_add_to = p_vf->vport_id; in qed_iov_configure_vport_forced()
1788 ether_addr_copy(filter.mac, p_vf->bulletin.p_virt->mac); in qed_iov_configure_vport_forced()
1790 rc = qed_sp_eth_filter_ucast(p_hwfn, p_vf->opaque_fid, in qed_iov_configure_vport_forced()
1797 if (p_vf->p_vf_info.is_trusted_configured) in qed_iov_configure_vport_forced()
1798 p_vf->configured_features |= in qed_iov_configure_vport_forced()
1801 p_vf->configured_features |= in qed_iov_configure_vport_forced()
1814 filter.vport_to_add_to = p_vf->vport_id; in qed_iov_configure_vport_forced()
1815 filter.vlan = p_vf->bulletin.p_virt->pvid; in qed_iov_configure_vport_forced()
1820 rc = qed_sp_eth_filter_ucast(p_hwfn, p_vf->opaque_fid, in qed_iov_configure_vport_forced()
1830 vport_update.opaque_fid = p_vf->opaque_fid; in qed_iov_configure_vport_forced()
1831 vport_update.vport_id = p_vf->vport_id; in qed_iov_configure_vport_forced()
1839 : p_vf->shadow_config.inner_vlan_removal; in qed_iov_configure_vport_forced()
1853 struct qed_vf_queue *p_queue = &p_vf->vf_queues[i]; in qed_iov_configure_vport_forced()
1875 p_vf->configured_features |= 1 << VLAN_ADDR_FORCED; in qed_iov_configure_vport_forced()
1877 p_vf->configured_features &= ~BIT(VLAN_ADDR_FORCED); in qed_iov_configure_vport_forced()
1884 qed_iov_reconfigure_unicast_shadow(p_hwfn, p_vf, events); in qed_iov_configure_vport_forced()
2051 struct qed_vf_info *p_vf, bool b_is_tx) in qed_iov_vf_mbx_qid() argument
2053 struct qed_iov_vf_mbx *p_mbx = &p_vf->vf_mbx; in qed_iov_vf_mbx_qid()
2057 if (!(p_vf->acquire.vfdev_info.capabilities & in qed_iov_vf_mbx_qid()
2071 p_vf->relative_vf_id); in qed_iov_vf_mbx_qid()
2079 p_vf->relative_vf_id, p_qid_tlv->qid); in qed_iov_vf_mbx_qid()
2300 struct qed_vf_info *p_vf) in qed_iov_vf_mbx_update_tunn_param() argument
2303 struct qed_iov_vf_mbx *mbx = &p_vf->vf_mbx; in qed_iov_vf_mbx_update_tunn_param()
2379 qed_iov_send_response(p_hwfn, p_ptt, p_vf, sizeof(*p_resp), status); in qed_iov_vf_mbx_update_tunn_param()
2384 struct qed_vf_info *p_vf, in qed_iov_vf_mbx_start_txq_resp() argument
2387 struct qed_iov_vf_mbx *mbx = &p_vf->vf_mbx; in qed_iov_vf_mbx_start_txq_resp()
2398 if (p_vf->acquire.vfdev_info.eth_fp_hsi_minor == in qed_iov_vf_mbx_start_txq_resp()
2416 qed_iov_send_response(p_hwfn, p_ptt, p_vf, length, status); in qed_iov_vf_mbx_start_txq_resp()
2764 struct qed_vf_info *p_vf, in qed_iov_vp_update_vlan_param() argument
2775 p_vf->shadow_config.inner_vlan_removal = p_vlan_tlv->remove_vlan; in qed_iov_vp_update_vlan_param()
2778 if (!(p_vf->configured_features & BIT(VLAN_ADDR_FORCED))) { in qed_iov_vp_update_vlan_param()
3108 struct qed_vf_info *p_vf, in qed_iov_vf_update_vlan_shadow() argument
3116 if (p_vf->shadow_config.vlans[i].used && in qed_iov_vf_update_vlan_shadow()
3117 p_vf->shadow_config.vlans[i].vid == in qed_iov_vf_update_vlan_shadow()
3119 p_vf->shadow_config.vlans[i].used = false; in qed_iov_vf_update_vlan_shadow()
3126 p_vf->relative_vf_id); in qed_iov_vf_update_vlan_shadow()
3132 p_vf->shadow_config.vlans[i].used = false; in qed_iov_vf_update_vlan_shadow()
3138 if (p_vf->bulletin.p_virt->valid_bitmap & BIT(VLAN_ADDR_FORCED)) in qed_iov_vf_update_vlan_shadow()
3144 if (p_vf->shadow_config.vlans[i].used) in qed_iov_vf_update_vlan_shadow()
3147 p_vf->shadow_config.vlans[i].used = true; in qed_iov_vf_update_vlan_shadow()
3148 p_vf->shadow_config.vlans[i].vid = p_params->vlan; in qed_iov_vf_update_vlan_shadow()
3156 p_vf->relative_vf_id, in qed_iov_vf_update_vlan_shadow()
3166 struct qed_vf_info *p_vf, in qed_iov_vf_update_mac_shadow() argument
3172 if (p_vf->bulletin.p_virt->valid_bitmap & BIT(MAC_ADDR_FORCED)) in qed_iov_vf_update_mac_shadow()
3176 if (p_vf->p_vf_info.is_trusted_configured) in qed_iov_vf_update_mac_shadow()
3182 if (ether_addr_equal(p_vf->shadow_config.macs[i], in qed_iov_vf_update_mac_shadow()
3184 eth_zero_addr(p_vf->shadow_config.macs[i]); in qed_iov_vf_update_mac_shadow()
3197 eth_zero_addr(p_vf->shadow_config.macs[i]); in qed_iov_vf_update_mac_shadow()
3206 if (is_zero_ether_addr(p_vf->shadow_config.macs[i])) { in qed_iov_vf_update_mac_shadow()
3207 ether_addr_copy(p_vf->shadow_config.macs[i], in qed_iov_vf_update_mac_shadow()
3225 struct qed_vf_info *p_vf, in qed_iov_vf_update_unicast_shadow() argument
3231 rc = qed_iov_vf_update_mac_shadow(p_hwfn, p_vf, p_params); in qed_iov_vf_update_unicast_shadow()
3237 rc = qed_iov_vf_update_vlan_shadow(p_hwfn, p_vf, p_params); in qed_iov_vf_update_unicast_shadow()
3389 struct qed_vf_info *p_vf) in qed_iov_vf_mbx_release() argument
3395 qed_iov_vf_cleanup(p_hwfn, p_vf); in qed_iov_vf_mbx_release()
3397 if (p_vf->state != VF_STOPPED && p_vf->state != VF_FREE) { in qed_iov_vf_mbx_release()
3399 rc = qed_sp_vf_stop(p_hwfn, p_vf->concrete_fid, in qed_iov_vf_mbx_release()
3400 p_vf->opaque_fid); in qed_iov_vf_mbx_release()
3408 p_vf->state = VF_STOPPED; in qed_iov_vf_mbx_release()
3411 qed_iov_prepare_resp(p_hwfn, p_ptt, p_vf, CHANNEL_TLV_RELEASE, in qed_iov_vf_mbx_release()
3417 struct qed_vf_info *p_vf) in qed_iov_vf_pf_get_coalesce() argument
3419 struct qed_iov_vf_mbx *mbx = &p_vf->vf_mbx; in qed_iov_vf_pf_get_coalesce()
3436 if (!qed_iov_validate_rxq(p_hwfn, p_vf, qid, in qed_iov_vf_pf_get_coalesce()
3440 p_vf->abs_vf_id, qid); in qed_iov_vf_pf_get_coalesce()
3444 p_cid = qed_iov_get_vf_rx_queue_cid(&p_vf->vf_queues[qid]); in qed_iov_vf_pf_get_coalesce()
3449 if (!qed_iov_validate_txq(p_hwfn, p_vf, qid, in qed_iov_vf_pf_get_coalesce()
3453 p_vf->abs_vf_id, qid); in qed_iov_vf_pf_get_coalesce()
3457 p_queue = &p_vf->vf_queues[qid]; in qed_iov_vf_pf_get_coalesce()
3481 qed_iov_send_response(p_hwfn, p_ptt, p_vf, sizeof(*p_resp), status); in qed_iov_vf_pf_get_coalesce()
3569 struct qed_vf_info *p_vf, struct qed_ptt *p_ptt) in qed_iov_vf_flr_poll_dorq() argument
3574 qed_fid_pretend(p_hwfn, p_ptt, (u16)p_vf->concrete_fid); in qed_iov_vf_flr_poll_dorq()
3587 p_vf->abs_vf_id, val); in qed_iov_vf_flr_poll_dorq()
3598 struct qed_vf_info *p_vf, struct qed_ptt *p_ptt) in qed_iov_vf_flr_poll_pbf() argument
3656 p_vf->abs_vf_id, (int)voq); in qed_iov_vf_flr_poll_pbf()
3668 struct qed_vf_info *p_vf, struct qed_ptt *p_ptt) in qed_iov_vf_flr_poll() argument
3672 rc = qed_iov_vf_flr_poll_dorq(p_hwfn, p_vf, p_ptt); in qed_iov_vf_flr_poll()
3676 rc = qed_iov_vf_flr_poll_pbf(p_hwfn, p_vf, p_ptt); in qed_iov_vf_flr_poll()
3688 struct qed_vf_info *p_vf; in qed_iov_execute_vf_flr_cleanup() local
3691 p_vf = qed_iov_get_vf_info(p_hwfn, rel_vf_id, false); in qed_iov_execute_vf_flr_cleanup()
3692 if (!p_vf) in qed_iov_execute_vf_flr_cleanup()
3697 u16 vfid = p_vf->abs_vf_id; in qed_iov_execute_vf_flr_cleanup()
3702 qed_iov_vf_cleanup(p_hwfn, p_vf); in qed_iov_execute_vf_flr_cleanup()
3705 if (!p_vf->b_init) in qed_iov_execute_vf_flr_cleanup()
3708 rc = qed_iov_vf_flr_poll(p_hwfn, p_vf, p_ptt); in qed_iov_execute_vf_flr_cleanup()
3728 p_vf->state = VF_STOPPED; in qed_iov_execute_vf_flr_cleanup()
3730 rc = qed_iov_enable_vf_access(p_hwfn, p_ptt, p_vf); in qed_iov_execute_vf_flr_cleanup()
3738 if (p_vf->state == VF_RESET) in qed_iov_execute_vf_flr_cleanup()
3739 p_vf->state = VF_STOPPED; in qed_iov_execute_vf_flr_cleanup()
3743 p_vf->vf_mbx.b_pending_msg = false; in qed_iov_execute_vf_flr_cleanup()
3789 struct qed_vf_info *p_vf; in qed_iov_mark_vf_flr() local
3792 p_vf = qed_iov_get_vf_info(p_hwfn, i, false); in qed_iov_mark_vf_flr()
3793 if (!p_vf) in qed_iov_mark_vf_flr()
3796 vfid = p_vf->abs_vf_id; in qed_iov_mark_vf_flr()
3799 u16 rel_vf_id = p_vf->relative_vf_id; in qed_iov_mark_vf_flr()
3805 p_vf->state = VF_RESET; in qed_iov_mark_vf_flr()
3826 struct qed_vf_info *p_vf = qed_iov_get_vf_info(p_hwfn, in qed_iov_get_link() local
3831 if (!p_vf) in qed_iov_get_link()
3834 p_bulletin = p_vf->bulletin.p_virt; in qed_iov_get_link()
3848 struct qed_vf_info *p_vf) in qed_iov_vf_pf_bulletin_update_mac() argument
3850 struct qed_bulletin_content *p_bulletin = p_vf->bulletin.p_virt; in qed_iov_vf_pf_bulletin_update_mac()
3851 struct qed_iov_vf_mbx *mbx = &p_vf->vf_mbx; in qed_iov_vf_pf_bulletin_update_mac()
3856 if (!p_vf->p_vf_info.is_trusted_configured) { in qed_iov_vf_pf_bulletin_update_mac()
3860 p_vf->abs_vf_id); in qed_iov_vf_pf_bulletin_update_mac()
3870 p_vf->abs_vf_id, p_req->mac); in qed_iov_vf_pf_bulletin_update_mac()
3873 qed_iov_prepare_resp(p_hwfn, p_ptt, p_vf, in qed_iov_vf_pf_bulletin_update_mac()
3883 struct qed_vf_info *p_vf; in qed_iov_process_mbx_req() local
3885 p_vf = qed_iov_get_vf_info(p_hwfn, (u16)vfid, true); in qed_iov_process_mbx_req()
3886 if (!p_vf) in qed_iov_process_mbx_req()
3889 mbx = &p_vf->vf_mbx; in qed_iov_process_mbx_req()
3895 p_vf->abs_vf_id); in qed_iov_process_mbx_req()
3904 p_vf->abs_vf_id, mbx->first_tlv.tl.type); in qed_iov_process_mbx_req()
3908 !p_vf->b_malicious) { in qed_iov_process_mbx_req()
3911 qed_iov_vf_mbx_acquire(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3914 qed_iov_vf_mbx_start_vport(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3917 qed_iov_vf_mbx_stop_vport(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3920 qed_iov_vf_mbx_start_rxq(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3923 qed_iov_vf_mbx_start_txq(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3926 qed_iov_vf_mbx_stop_rxqs(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3929 qed_iov_vf_mbx_stop_txqs(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3932 qed_iov_vf_mbx_update_rxqs(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3935 qed_iov_vf_mbx_vport_update(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3938 qed_iov_vf_mbx_ucast_filter(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3941 qed_iov_vf_mbx_close(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3944 qed_iov_vf_mbx_int_cleanup(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3947 qed_iov_vf_mbx_release(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3950 qed_iov_vf_mbx_update_tunn_param(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3953 qed_iov_vf_pf_set_coalesce(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3956 qed_iov_vf_pf_get_coalesce(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3959 qed_iov_vf_pf_bulletin_update_mac(p_hwfn, p_ptt, p_vf); in qed_iov_process_mbx_req()
3965 p_vf->abs_vf_id, mbx->first_tlv.tl.type); in qed_iov_process_mbx_req()
3967 qed_iov_prepare_resp(p_hwfn, p_ptt, p_vf, in qed_iov_process_mbx_req()
3980 p_vf->abs_vf_id, in qed_iov_process_mbx_req()
3988 if (p_vf->acquire.first_tlv.reply_address && in qed_iov_process_mbx_req()
3990 p_vf->acquire.first_tlv.reply_address)) { in qed_iov_process_mbx_req()
3991 qed_iov_prepare_resp(p_hwfn, p_ptt, p_vf, in qed_iov_process_mbx_req()
3999 p_vf->abs_vf_id); in qed_iov_process_mbx_req()
4011 struct qed_vf_info *p_vf; in qed_iov_pf_get_pending_events() local
4013 p_vf = &p_hwfn->pf_iov_info->vfs_array[i]; in qed_iov_pf_get_pending_events()
4014 if (p_vf->vf_mbx.b_pending_msg) in qed_iov_pf_get_pending_events()
4038 struct qed_vf_info *p_vf = qed_sriov_get_vf_from_absid(p_hwfn, in qed_sriov_vfpf_msg() local
4041 if (!p_vf) in qed_sriov_vfpf_msg()
4047 p_vf->vf_mbx.pending_req = HILO_64(vf_msg->hi, vf_msg->lo); in qed_sriov_vfpf_msg()
4050 p_vf->vf_mbx.b_pending_msg = true; in qed_sriov_vfpf_msg()
4059 struct qed_vf_info *p_vf; in qed_sriov_vfpf_malicious() local
4061 p_vf = qed_sriov_get_vf_from_absid(p_hwfn, qed_vf_from_entity_id in qed_sriov_vfpf_malicious()
4063 if (!p_vf) in qed_sriov_vfpf_malicious()
4066 if (!p_vf->b_malicious) { in qed_sriov_vfpf_malicious()
4069 p_vf->abs_vf_id, p_data->err_id); in qed_sriov_vfpf_malicious()
4071 p_vf->b_malicious = true; in qed_sriov_vfpf_malicious()
4075 p_vf->abs_vf_id, p_data->err_id); in qed_sriov_vfpf_malicious()
4325 struct qed_vf_info *p_vf; in qed_iov_bulletin_get_mac() local
4327 p_vf = qed_iov_get_vf_info(p_hwfn, rel_vf_id, true); in qed_iov_bulletin_get_mac()
4328 if (!p_vf || !p_vf->bulletin.p_virt) in qed_iov_bulletin_get_mac()
4331 if (!(p_vf->bulletin.p_virt->valid_bitmap & in qed_iov_bulletin_get_mac()
4335 return p_vf->bulletin.p_virt->mac; in qed_iov_bulletin_get_mac()
4341 struct qed_vf_info *p_vf; in qed_iov_bulletin_get_forced_mac() local
4343 p_vf = qed_iov_get_vf_info(p_hwfn, rel_vf_id, true); in qed_iov_bulletin_get_forced_mac()
4344 if (!p_vf || !p_vf->bulletin.p_virt) in qed_iov_bulletin_get_forced_mac()
4347 if (!(p_vf->bulletin.p_virt->valid_bitmap & BIT(MAC_ADDR_FORCED))) in qed_iov_bulletin_get_forced_mac()
4350 return p_vf->bulletin.p_virt->mac; in qed_iov_bulletin_get_forced_mac()
4356 struct qed_vf_info *p_vf; in qed_iov_bulletin_get_forced_vlan() local
4358 p_vf = qed_iov_get_vf_info(p_hwfn, rel_vf_id, true); in qed_iov_bulletin_get_forced_vlan()
4359 if (!p_vf || !p_vf->bulletin.p_virt) in qed_iov_bulletin_get_forced_vlan()
4362 if (!(p_vf->bulletin.p_virt->valid_bitmap & BIT(VLAN_ADDR_FORCED))) in qed_iov_bulletin_get_forced_vlan()
4365 return p_vf->bulletin.p_virt->pvid; in qed_iov_bulletin_get_forced_vlan()