Home
last modified time | relevance | path

Searched refs:hwfn (Results 1 – 10 of 10) sorted by relevance

/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qed/
Dqed_dcbx.c853 static void qed_dcbx_aen(struct qed_hwfn *hwfn, u32 mib_type) in qed_dcbx_aen() argument
855 struct qed_common_cb_ops *op = hwfn->cdev->protocol_ops.common; in qed_dcbx_aen()
856 void *cookie = hwfn->cdev->ops_cookie; in qed_dcbx_aen()
859 op->dcbx_aen(cookie, &hwfn->p_dcbx_info->get, mib_type); in qed_dcbx_aen()
1283 static struct qed_dcbx_get *qed_dcbnl_get_dcbx(struct qed_hwfn *hwfn, in qed_dcbnl_get_dcbx() argument
1292 if (qed_dcbx_query_params(hwfn, dcbx_info, type)) { in qed_dcbnl_get_dcbx()
1299 DP_INFO(hwfn, "DCBX is not enabled/operational\n"); in qed_dcbnl_get_dcbx()
1309 struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev); in qed_dcbnl_getstate() local
1313 dcbx_info = qed_dcbnl_get_dcbx(hwfn, QED_DCBX_OPERATIONAL_MIB); in qed_dcbnl_getstate()
1318 DP_VERBOSE(hwfn, QED_MSG_DCB, "DCB state = %d\n", enabled); in qed_dcbnl_getstate()
[all …]
Dqed_main.c704 struct qed_hwfn *hwfn = &cdev->hwfns[index % cdev->num_hwfns]; in qed_simd_handler_config() local
707 hwfn->simd_proto_handler[relative_idx].func = handler; in qed_simd_handler_config()
708 hwfn->simd_proto_handler[relative_idx].token = token; in qed_simd_handler_config()
713 struct qed_hwfn *hwfn = &cdev->hwfns[index % cdev->num_hwfns]; in qed_simd_handler_clean() local
716 memset(&hwfn->simd_proto_handler[relative_idx], 0, in qed_simd_handler_clean()
729 struct qed_hwfn *hwfn; in qed_single_int() local
740 hwfn = &cdev->hwfns[i]; in qed_single_int()
744 tasklet_schedule(&hwfn->sp_dpc); in qed_single_int()
753 &hwfn->simd_proto_handler[j]; in qed_single_int()
758 DP_NOTICE(hwfn, in qed_single_int()
[all …]
Dqed_init_ops.h69 #define STORE_RT_REG(hwfn, offset, val) \ argument
70 qed_init_store_rt_reg(hwfn, offset, val)
72 #define OVERWRITE_RT_REG(hwfn, offset, val) \ argument
73 qed_init_store_rt_reg(hwfn, offset, val)
89 #define STORE_RT_REG_AGG(hwfn, offset, val) \ argument
90 qed_init_store_rt_agg(hwfn, offset, (u32 *)&val, sizeof(val))
Dqed_sriov.c2977 static int qed_iov_pre_update_vport(struct qed_hwfn *hwfn, in qed_iov_pre_update_vport() argument
2997 vf_info = qed_iov_get_public_vf_info(hwfn, vfid, true); in qed_iov_pre_update_vport()
3240 static int qed_iov_chk_ucast(struct qed_hwfn *hwfn, in qed_iov_chk_ucast() argument
3245 vf = qed_iov_get_public_vf_info(hwfn, vfid, true); in qed_iov_chk_ucast()
3255 qed_iov_bulletin_set_mac(hwfn, vf->mac, vfid); in qed_iov_chk_ucast()
3258 qed_schedule_iov(hwfn, QED_IOV_WQ_BULLETIN_UPDATE_FLAG); in qed_iov_chk_ucast()
4408 void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag) in qed_schedule_iov() argument
4411 set_bit(flag, &hwfn->iov_task_flags); in qed_schedule_iov()
4413 DP_VERBOSE(hwfn, QED_MSG_IOV, "Scheduling iov task [Flag: %d]\n", flag); in qed_schedule_iov()
4414 queue_delayed_work(hwfn->iov_wq, &hwfn->iov_task, 0); in qed_schedule_iov()
[all …]
Dqed_sriov.h368 void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag);
371 void qed_inform_vf_link_state(struct qed_hwfn *hwfn);
429 static inline void qed_schedule_iov(struct qed_hwfn *hwfn, in qed_schedule_iov() argument
443 static inline void qed_inform_vf_link_state(struct qed_hwfn *hwfn) in qed_inform_vf_link_state() argument
Dqed_vf.c1631 static bool qed_vf_bulletin_get_forced_mac(struct qed_hwfn *hwfn, in qed_vf_bulletin_get_forced_mac() argument
1636 bulletin = &hwfn->vf_iov_info->bulletin_shadow; in qed_vf_bulletin_get_forced_mac()
1679 static void qed_handle_bulletin_change(struct qed_hwfn *hwfn) in qed_handle_bulletin_change() argument
1681 struct qed_eth_cb_ops *ops = hwfn->cdev->protocol_ops.eth; in qed_handle_bulletin_change()
1683 void *cookie = hwfn->cdev->ops_cookie; in qed_handle_bulletin_change()
1686 qed_vf_bulletin_get_udp_ports(hwfn, &vxlan_port, &geneve_port); in qed_handle_bulletin_change()
1687 is_mac_exist = qed_vf_bulletin_get_forced_mac(hwfn, mac, in qed_handle_bulletin_change()
1695 qed_link_update(hwfn, NULL); in qed_handle_bulletin_change()
1700 struct qed_hwfn *hwfn = container_of(work, struct qed_hwfn, in qed_iov_vf_task() local
1704 if (test_and_clear_bit(QED_IOV_WQ_STOP_WQ_FLAG, &hwfn->iov_task_flags)) in qed_iov_vf_task()
[all …]
Dqed.h979 void qed_link_update(struct qed_hwfn *hwfn, struct qed_ptt *ptt);
980 void qed_bw_update(struct qed_hwfn *hwfn, struct qed_ptt *ptt);
991 int qed_slowpath_irq_req(struct qed_hwfn *hwfn);
993 int qed_mfw_tlv_req(struct qed_hwfn *hwfn);
995 int qed_mfw_fill_tlv_data(struct qed_hwfn *hwfn,
Dqed_fcoe.c726 struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); in qed_fill_fcoe_dev_info() local
733 qed_fcoe_get_primary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_fcoe_dev_info()
735 qed_fcoe_get_secondary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_fcoe_dev_info()
737 info->wwpn = hwfn->mcp_info->func_info.wwn_port; in qed_fill_fcoe_dev_info()
738 info->wwnn = hwfn->mcp_info->func_info.wwn_node; in qed_fill_fcoe_dev_info()
740 info->num_cqs = FEAT_NUM(hwfn, QED_FCOE_CQ); in qed_fill_fcoe_dev_info()
Dqed_l2.c2205 struct qed_hwfn *hwfn = &cdev->hwfns[i]; in qed_fill_eth_dev_info() local
2206 u16 l2_queues = (u16)FEAT_NUM(hwfn, in qed_fill_eth_dev_info()
2210 cids = hwfn->pf_params.eth_pf_params.num_cons; in qed_fill_eth_dev_info()
2621 struct qed_hwfn *hwfn = &cdev->hwfns[i]; in qed_tunn_configure() local
2625 tun = &hwfn->cdev->tunnel; in qed_tunn_configure()
2627 p_ptt = qed_ptt_acquire(hwfn); in qed_tunn_configure()
2634 rc = qed_sp_pf_update_tunn_cfg(hwfn, p_ptt, &tunn_info, in qed_tunn_configure()
2638 qed_ptt_release(hwfn, p_ptt); in qed_tunn_configure()
2642 if (IS_PF_SRIOV(hwfn)) { in qed_tunn_configure()
2649 qed_for_each_vf(hwfn, j) { in qed_tunn_configure()
[all …]
Dqed_iscsi.c1036 struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); in qed_fill_iscsi_dev_info() local
1044 qed_iscsi_get_primary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_iscsi_dev_info()
1046 qed_iscsi_get_secondary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_iscsi_dev_info()
1048 info->num_cqs = FEAT_NUM(hwfn, QED_ISCSI_CQ); in qed_fill_iscsi_dev_info()