/drivers/net/ethernet/qlogic/qed/ |
D | qed_dcbx.c | 879 static void qed_dcbx_aen(struct qed_hwfn *hwfn, u32 mib_type) in qed_dcbx_aen() argument 881 struct qed_common_cb_ops *op = hwfn->cdev->protocol_ops.common; in qed_dcbx_aen() 882 void *cookie = hwfn->cdev->ops_cookie; in qed_dcbx_aen() 885 op->dcbx_aen(cookie, &hwfn->p_dcbx_info->get, mib_type); in qed_dcbx_aen() 1308 static struct qed_dcbx_get *qed_dcbnl_get_dcbx(struct qed_hwfn *hwfn, in qed_dcbnl_get_dcbx() argument 1317 if (qed_dcbx_query_params(hwfn, dcbx_info, type)) { in qed_dcbnl_get_dcbx() 1324 DP_INFO(hwfn, "DCBX is not enabled/operational\n"); in qed_dcbnl_get_dcbx() 1334 struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev); in qed_dcbnl_getstate() local 1338 dcbx_info = qed_dcbnl_get_dcbx(hwfn, QED_DCBX_OPERATIONAL_MIB); in qed_dcbnl_getstate() 1343 DP_VERBOSE(hwfn, QED_MSG_DCB, "DCB state = %d\n", enabled); in qed_dcbnl_getstate() [all …]
|
D | qed_main.c | 640 struct qed_hwfn *hwfn = &cdev->hwfns[index % cdev->num_hwfns]; in qed_simd_handler_config() local 643 hwfn->simd_proto_handler[relative_idx].func = handler; in qed_simd_handler_config() 644 hwfn->simd_proto_handler[relative_idx].token = token; in qed_simd_handler_config() 649 struct qed_hwfn *hwfn = &cdev->hwfns[index % cdev->num_hwfns]; in qed_simd_handler_clean() local 652 memset(&hwfn->simd_proto_handler[relative_idx], 0, in qed_simd_handler_clean() 665 struct qed_hwfn *hwfn; in qed_single_int() local 676 hwfn = &cdev->hwfns[i]; in qed_single_int() 680 tasklet_schedule(hwfn->sp_dpc); in qed_single_int() 689 &hwfn->simd_proto_handler[j]; in qed_single_int() 694 DP_NOTICE(hwfn, in qed_single_int() [all …]
|
D | qed_init_ops.h | 103 #define STORE_RT_REG(hwfn, offset, val) \ argument 104 qed_init_store_rt_reg(hwfn, offset, val) 106 #define OVERWRITE_RT_REG(hwfn, offset, val) \ argument 107 qed_init_store_rt_reg(hwfn, offset, val) 123 #define STORE_RT_REG_AGG(hwfn, offset, val) \ argument 124 qed_init_store_rt_agg(hwfn, offset, (u32 *)&val, sizeof(val))
|
D | qed_sriov.c | 2997 static int qed_iov_pre_update_vport(struct qed_hwfn *hwfn, in qed_iov_pre_update_vport() argument 3013 vf_info = qed_iov_get_public_vf_info(hwfn, vfid, true); in qed_iov_pre_update_vport() 3249 static int qed_iov_chk_ucast(struct qed_hwfn *hwfn, in qed_iov_chk_ucast() argument 3254 vf = qed_iov_get_public_vf_info(hwfn, vfid, true); in qed_iov_chk_ucast() 3264 qed_iov_bulletin_set_mac(hwfn, vf->mac, vfid); in qed_iov_chk_ucast() 3267 qed_schedule_iov(hwfn, QED_IOV_WQ_BULLETIN_UPDATE_FLAG); in qed_iov_chk_ucast() 4421 void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag) in qed_schedule_iov() argument 4424 set_bit(flag, &hwfn->iov_task_flags); in qed_schedule_iov() 4426 DP_VERBOSE(hwfn, QED_MSG_IOV, "Scheduling iov task [Flag: %d]\n", flag); in qed_schedule_iov() 4427 queue_delayed_work(hwfn->iov_wq, &hwfn->iov_task, 0); in qed_schedule_iov() [all …]
|
D | qed_sriov.h | 391 void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag); 394 void qed_inform_vf_link_state(struct qed_hwfn *hwfn); 452 static inline void qed_schedule_iov(struct qed_hwfn *hwfn, in qed_schedule_iov() argument 466 static inline void qed_inform_vf_link_state(struct qed_hwfn *hwfn) in qed_inform_vf_link_state() argument
|
D | qed_vf.c | 1637 static bool qed_vf_bulletin_get_forced_mac(struct qed_hwfn *hwfn, in qed_vf_bulletin_get_forced_mac() argument 1642 bulletin = &hwfn->vf_iov_info->bulletin_shadow; in qed_vf_bulletin_get_forced_mac() 1685 static void qed_handle_bulletin_change(struct qed_hwfn *hwfn) in qed_handle_bulletin_change() argument 1687 struct qed_eth_cb_ops *ops = hwfn->cdev->protocol_ops.eth; in qed_handle_bulletin_change() 1689 void *cookie = hwfn->cdev->ops_cookie; in qed_handle_bulletin_change() 1692 qed_vf_bulletin_get_udp_ports(hwfn, &vxlan_port, &geneve_port); in qed_handle_bulletin_change() 1693 is_mac_exist = qed_vf_bulletin_get_forced_mac(hwfn, mac, in qed_handle_bulletin_change() 1701 qed_link_update(hwfn, NULL); in qed_handle_bulletin_change() 1706 struct qed_hwfn *hwfn = container_of(work, struct qed_hwfn, in qed_iov_vf_task() local 1710 if (test_and_clear_bit(QED_IOV_WQ_STOP_WQ_FLAG, &hwfn->iov_task_flags)) in qed_iov_vf_task() [all …]
|
D | qed.h | 975 void qed_link_update(struct qed_hwfn *hwfn, struct qed_ptt *ptt); 983 int qed_slowpath_irq_req(struct qed_hwfn *hwfn); 985 int qed_mfw_tlv_req(struct qed_hwfn *hwfn); 987 int qed_mfw_fill_tlv_data(struct qed_hwfn *hwfn,
|
D | qed_fcoe.c | 748 struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); in qed_fill_fcoe_dev_info() local 755 qed_fcoe_get_primary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_fcoe_dev_info() 757 qed_fcoe_get_secondary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_fcoe_dev_info() 759 info->wwpn = hwfn->mcp_info->func_info.wwn_port; in qed_fill_fcoe_dev_info() 760 info->wwnn = hwfn->mcp_info->func_info.wwn_node; in qed_fill_fcoe_dev_info() 762 info->num_cqs = FEAT_NUM(hwfn, QED_FCOE_CQ); in qed_fill_fcoe_dev_info()
|
D | qed_l2.c | 2221 struct qed_hwfn *hwfn = &cdev->hwfns[i]; in qed_fill_eth_dev_info() local 2222 u16 l2_queues = (u16)FEAT_NUM(hwfn, in qed_fill_eth_dev_info() 2226 cids = hwfn->pf_params.eth_pf_params.num_cons; in qed_fill_eth_dev_info() 2637 struct qed_hwfn *hwfn = &cdev->hwfns[i]; in qed_tunn_configure() local 2641 tun = &hwfn->cdev->tunnel; in qed_tunn_configure() 2643 p_ptt = qed_ptt_acquire(hwfn); in qed_tunn_configure() 2650 rc = qed_sp_pf_update_tunn_cfg(hwfn, p_ptt, &tunn_info, in qed_tunn_configure() 2654 qed_ptt_release(hwfn, p_ptt); in qed_tunn_configure() 2658 if (IS_PF_SRIOV(hwfn)) { in qed_tunn_configure() 2665 qed_for_each_vf(hwfn, j) { in qed_tunn_configure() [all …]
|
D | qed_iscsi.c | 1085 struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); in qed_fill_iscsi_dev_info() local 1093 qed_iscsi_get_primary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_iscsi_dev_info() 1095 qed_iscsi_get_secondary_bdq_prod(hwfn, BDQ_ID_RQ); in qed_fill_iscsi_dev_info() 1097 info->num_cqs = FEAT_NUM(hwfn, QED_ISCSI_CQ); in qed_fill_iscsi_dev_info()
|