Home
last modified time | relevance | path

Searched refs:hwfn (Results 1 – 10 of 10) sorted by relevance

/drivers/net/ethernet/qlogic/qed/
Dqed_dcbx.c879 static void qed_dcbx_aen(struct qed_hwfn *hwfn, u32 mib_type) in qed_dcbx_aen() argument
881 struct qed_common_cb_ops *op = hwfn->cdev->protocol_ops.common; in qed_dcbx_aen()
882 void *cookie = hwfn->cdev->ops_cookie; in qed_dcbx_aen()
885 op->dcbx_aen(cookie, &hwfn->p_dcbx_info->get, mib_type); in qed_dcbx_aen()
1308 static struct qed_dcbx_get *qed_dcbnl_get_dcbx(struct qed_hwfn *hwfn, in qed_dcbnl_get_dcbx() argument
1317 if (qed_dcbx_query_params(hwfn, dcbx_info, type)) { in qed_dcbnl_get_dcbx()
1324 DP_INFO(hwfn, "DCBX is not enabled/operational\n"); in qed_dcbnl_get_dcbx()
1334 struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev); in qed_dcbnl_getstate() local
1338 dcbx_info = qed_dcbnl_get_dcbx(hwfn, QED_DCBX_OPERATIONAL_MIB); in qed_dcbnl_getstate()
1343 DP_VERBOSE(hwfn, QED_MSG_DCB, "DCB state = %d\n", enabled); in qed_dcbnl_getstate()
[all …]
Dqed_main.c640 struct qed_hwfn *hwfn = &cdev->hwfns[index % cdev->num_hwfns]; in qed_simd_handler_config() local
643 hwfn->simd_proto_handler[relative_idx].func = handler; in qed_simd_handler_config()
644 hwfn->simd_proto_handler[relative_idx].token = token; in qed_simd_handler_config()
649 struct qed_hwfn *hwfn = &cdev->hwfns[index % cdev->num_hwfns]; in qed_simd_handler_clean() local
652 memset(&hwfn->simd_proto_handler[relative_idx], 0, in qed_simd_handler_clean()
665 struct qed_hwfn *hwfn; in qed_single_int() local
676 hwfn = &cdev->hwfns[i]; in qed_single_int()
680 tasklet_schedule(hwfn->sp_dpc); in qed_single_int()
689 &hwfn->simd_proto_handler[j]; in qed_single_int()
694 DP_NOTICE(hwfn, in qed_single_int()
[all …]
Dqed_init_ops.h103 #define STORE_RT_REG(hwfn, offset, val) \ argument
104 qed_init_store_rt_reg(hwfn, offset, val)
106 #define OVERWRITE_RT_REG(hwfn, offset, val) \ argument
107 qed_init_store_rt_reg(hwfn, offset, val)
123 #define STORE_RT_REG_AGG(hwfn, offset, val) \ argument
124 qed_init_store_rt_agg(hwfn, offset, (u32 *)&val, sizeof(val))
Dqed_sriov.c2997 static int qed_iov_pre_update_vport(struct qed_hwfn *hwfn, in qed_iov_pre_update_vport() argument
3013 vf_info = qed_iov_get_public_vf_info(hwfn, vfid, true); in qed_iov_pre_update_vport()
3249 static int qed_iov_chk_ucast(struct qed_hwfn *hwfn, in qed_iov_chk_ucast() argument
3254 vf = qed_iov_get_public_vf_info(hwfn, vfid, true); in qed_iov_chk_ucast()
3264 qed_iov_bulletin_set_mac(hwfn, vf->mac, vfid); in qed_iov_chk_ucast()
3267 qed_schedule_iov(hwfn, QED_IOV_WQ_BULLETIN_UPDATE_FLAG); in qed_iov_chk_ucast()
4421 void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag) in qed_schedule_iov() argument
4424 set_bit(flag, &hwfn->iov_task_flags); in qed_schedule_iov()
4426 DP_VERBOSE(hwfn, QED_MSG_IOV, "Scheduling iov task [Flag: %d]\n", flag); in qed_schedule_iov()
4427 queue_delayed_work(hwfn->iov_wq, &hwfn->iov_task, 0); in qed_schedule_iov()
[all …]
Dqed_sriov.h391 void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag);
394 void qed_inform_vf_link_state(struct qed_hwfn *hwfn);
452 static inline void qed_schedule_iov(struct qed_hwfn *hwfn, in qed_schedule_iov() argument
466 static inline void qed_inform_vf_link_state(struct qed_hwfn *hwfn) in qed_inform_vf_link_state() argument
Dqed_vf.c1637 static bool qed_vf_bulletin_get_forced_mac(struct qed_hwfn *hwfn, in qed_vf_bulletin_get_forced_mac() argument
1642 bulletin = &hwfn->vf_iov_info->bulletin_shadow; in qed_vf_bulletin_get_forced_mac()
1685 static void qed_handle_bulletin_change(struct qed_hwfn *hwfn) in qed_handle_bulletin_change() argument
1687 struct qed_eth_cb_ops *ops = hwfn->cdev->protocol_ops.eth; in qed_handle_bulletin_change()
1689 void *cookie = hwfn->cdev->ops_cookie; in qed_handle_bulletin_change()
1692 qed_vf_bulletin_get_udp_ports(hwfn, &vxlan_port, &geneve_port); in qed_handle_bulletin_change()
1693 is_mac_exist = qed_vf_bulletin_get_forced_mac(hwfn, mac, in qed_handle_bulletin_change()
1701 qed_link_update(hwfn, NULL); in qed_handle_bulletin_change()
1706 struct qed_hwfn *hwfn = container_of(work, struct qed_hwfn, in qed_iov_vf_task() local
1710 if (test_and_clear_bit(QED_IOV_WQ_STOP_WQ_FLAG, &hwfn->iov_task_flags)) in qed_iov_vf_task()
[all …]
Dqed.h975 void qed_link_update(struct qed_hwfn *hwfn, struct qed_ptt *ptt);
983 int qed_slowpath_irq_req(struct qed_hwfn *hwfn);
985 int qed_mfw_tlv_req(struct qed_hwfn *hwfn);
987 int qed_mfw_fill_tlv_data(struct qed_hwfn *hwfn,
Dqed_fcoe.c748 struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); in qed_fill_fcoe_dev_info() local
755 qed_fcoe_get_primary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_fcoe_dev_info()
757 qed_fcoe_get_secondary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_fcoe_dev_info()
759 info->wwpn = hwfn->mcp_info->func_info.wwn_port; in qed_fill_fcoe_dev_info()
760 info->wwnn = hwfn->mcp_info->func_info.wwn_node; in qed_fill_fcoe_dev_info()
762 info->num_cqs = FEAT_NUM(hwfn, QED_FCOE_CQ); in qed_fill_fcoe_dev_info()
Dqed_l2.c2221 struct qed_hwfn *hwfn = &cdev->hwfns[i]; in qed_fill_eth_dev_info() local
2222 u16 l2_queues = (u16)FEAT_NUM(hwfn, in qed_fill_eth_dev_info()
2226 cids = hwfn->pf_params.eth_pf_params.num_cons; in qed_fill_eth_dev_info()
2637 struct qed_hwfn *hwfn = &cdev->hwfns[i]; in qed_tunn_configure() local
2641 tun = &hwfn->cdev->tunnel; in qed_tunn_configure()
2643 p_ptt = qed_ptt_acquire(hwfn); in qed_tunn_configure()
2650 rc = qed_sp_pf_update_tunn_cfg(hwfn, p_ptt, &tunn_info, in qed_tunn_configure()
2654 qed_ptt_release(hwfn, p_ptt); in qed_tunn_configure()
2658 if (IS_PF_SRIOV(hwfn)) { in qed_tunn_configure()
2665 qed_for_each_vf(hwfn, j) { in qed_tunn_configure()
[all …]
Dqed_iscsi.c1085 struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); in qed_fill_iscsi_dev_info() local
1093 qed_iscsi_get_primary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_iscsi_dev_info()
1095 qed_iscsi_get_secondary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_iscsi_dev_info()
1097 info->num_cqs = FEAT_NUM(hwfn, QED_ISCSI_CQ); in qed_fill_iscsi_dev_info()