/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_mbx.c | 149 static u16 fm10k_mbx_index_len(struct fm10k_mbx_info *mbx, u16 head, u16 tail) in fm10k_mbx_index_len() argument 157 return len & ((mbx->mbmem_len << 1) - 1); in fm10k_mbx_index_len() 168 static u16 fm10k_mbx_tail_add(struct fm10k_mbx_info *mbx, u16 offset) in fm10k_mbx_tail_add() argument 170 u16 tail = (mbx->tail + offset + 1) & ((mbx->mbmem_len << 1) - 1); in fm10k_mbx_tail_add() 173 return (tail > mbx->tail) ? --tail : ++tail; in fm10k_mbx_tail_add() 184 static u16 fm10k_mbx_tail_sub(struct fm10k_mbx_info *mbx, u16 offset) in fm10k_mbx_tail_sub() argument 186 u16 tail = (mbx->tail - offset - 1) & ((mbx->mbmem_len << 1) - 1); in fm10k_mbx_tail_sub() 189 return (tail < mbx->tail) ? ++tail : --tail; in fm10k_mbx_tail_sub() 200 static u16 fm10k_mbx_head_add(struct fm10k_mbx_info *mbx, u16 offset) in fm10k_mbx_head_add() argument 202 u16 head = (mbx->head + offset + 1) & ((mbx->mbmem_len << 1) - 1); in fm10k_mbx_head_add() [all …]
|
D | fm10k_vf.c | 186 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_update_vlan_vf() local 206 return mbx->ops.enqueue_tx(hw, mbx, msg); in fm10k_update_vlan_vf() 218 struct fm10k_mbx_info *mbx) in fm10k_msg_mac_vlan_vf() argument 290 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_update_uc_addr_vf() local 315 return mbx->ops.enqueue_tx(hw, mbx, msg); in fm10k_update_uc_addr_vf() 332 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_update_mc_addr_vf() local 353 return mbx->ops.enqueue_tx(hw, mbx, msg); in fm10k_update_mc_addr_vf() 365 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_update_int_moderator_vf() local 372 mbx->ops.enqueue_tx(hw, mbx, msg); in fm10k_update_int_moderator_vf() 393 struct fm10k_mbx_info *mbx) in fm10k_msg_lport_state_vf() argument [all …]
|
D | fm10k_iov.c | 26 struct fm10k_mbx_info *mbx) in fm10k_iov_msg_error() argument 28 struct fm10k_vf_info *vf_info = (struct fm10k_vf_info *)mbx; in fm10k_iov_msg_error() 35 return fm10k_tlv_msg_error(hw, results, mbx); in fm10k_iov_msg_error() 85 vf_info->mbx.ops.connect(hw, &vf_info->mbx); in fm10k_iov_event() 126 struct fm10k_mbx_info *mbx = &vf_info->mbx; in fm10k_iov_mbx() local 130 hw->mbx.ops.process(hw, &hw->mbx); in fm10k_iov_mbx() 137 if (!mbx->timeout) { in fm10k_iov_mbx() 139 mbx->ops.connect(hw, mbx); in fm10k_iov_mbx() 143 if (!hw->mbx.ops.tx_ready(&hw->mbx, FM10K_VFMBX_MSG_MTU)) { in fm10k_iov_mbx() 150 mbx->ops.process(hw, mbx); in fm10k_iov_mbx() [all …]
|
D | fm10k_pf.c | 328 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_update_xc_addr_pf() local 357 return mbx->ops.enqueue_tx(hw, mbx, msg); in fm10k_update_xc_addr_pf() 415 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_update_xcast_mode_pf() local 436 return mbx->ops.enqueue_tx(hw, mbx, msg); in fm10k_update_xcast_mode_pf() 483 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_update_lport_state_pf() local 507 return mbx->ops.enqueue_tx(hw, mbx, msg); in fm10k_update_lport_state_pf() 898 if (vf_info->mbx.ops.enqueue_tx) { in fm10k_iov_assign_default_mac_vlan_pf() 899 err = vf_info->mbx.ops.enqueue_tx(hw, &vf_info->mbx, msg); in fm10k_iov_assign_default_mac_vlan_pf() 981 vf_info->mbx.timeout = 0; in fm10k_iov_reset_resources_pf() 982 if (vf_info->mbx.ops.disconnect) in fm10k_iov_reset_resources_pf() [all …]
|
D | fm10k_common.c | 492 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_get_host_state_generic() local 498 mbx->ops.process(hw, mbx); in fm10k_get_host_state_generic() 515 if (!mbx->timeout) { in fm10k_get_host_state_generic() 521 if (!mbx->ops.tx_ready(mbx, FM10K_VFMBX_MSG_MTU)) in fm10k_get_host_state_generic()
|
D | fm10k_pci.c | 950 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_msix_mbx_vf() local 959 mbx->ops.process(hw, mbx); in fm10k_msix_mbx_vf() 1073 vf_info->mbx.ops.connect(hw, &vf_info->mbx); in fm10k_handle_fault() 1145 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_msix_mbx_pf() local 1163 err = mbx->ops.process(hw, mbx); in fm10k_msix_mbx_pf() 1209 hw->mbx.ops.disconnect(hw, &hw->mbx); in fm10k_mbx_free_irq() 1233 struct fm10k_mbx_info *mbx) in fm10k_mbx_mac_addr() argument 1240 err = fm10k_msg_mac_vlan_vf(hw, results, mbx); in fm10k_mbx_mac_addr() 1261 struct fm10k_mbx_info __always_unused *mbx) in fm10k_mbx_error() argument 1293 err = hw->mbx.ops.register_handlers(&hw->mbx, vf_mbx_data); in fm10k_mbx_request_irq_vf() [all …]
|
D | fm10k_tlv.c | 561 struct fm10k_mbx_info *mbx, in fm10k_tlv_msg_parse() argument 593 return data->func(hw, results, mbx); in fm10k_tlv_msg_parse() 607 struct fm10k_mbx_info *mbx) in fm10k_tlv_msg_error() argument 723 struct fm10k_mbx_info *mbx) in fm10k_tlv_msg_test() argument 744 &mbx->test_result); in fm10k_tlv_msg_test() 864 return mbx->ops.enqueue_tx(hw, mbx, reply); in fm10k_tlv_msg_test()
|
D | fm10k_ethtool.c | 293 fm10k_add_ethtool_stats(&data, &interface->hw.mbx, in fm10k_get_ethtool_stats() 889 struct fm10k_mbx_info *mbx = &hw->mbx; in fm10k_mbx_test() local 906 mbx->test_result = FM10K_NOT_IMPLEMENTED; in fm10k_mbx_test() 907 err = mbx->ops.enqueue_tx(hw, mbx, test_msg); in fm10k_mbx_test() 919 mbx->ops.process(hw, mbx); in fm10k_mbx_test() 922 err = mbx->test_result; in fm10k_mbx_test()
|
/drivers/net/ethernet/intel/igb/ |
D | e1000_mbx.c | 37 struct e1000_mbx_info *mbx = &hw->mbx; in igb_read_mbx() local 41 if (size > mbx->size) in igb_read_mbx() 42 size = mbx->size; in igb_read_mbx() 44 if (mbx->ops.read) in igb_read_mbx() 45 ret_val = mbx->ops.read(hw, msg, size, mbx_id); in igb_read_mbx() 61 struct e1000_mbx_info *mbx = &hw->mbx; in igb_write_mbx() local 64 if (size > mbx->size) in igb_write_mbx() 67 else if (mbx->ops.write) in igb_write_mbx() 68 ret_val = mbx->ops.write(hw, msg, size, mbx_id); in igb_write_mbx() 82 struct e1000_mbx_info *mbx = &hw->mbx; in igb_check_for_msg() local [all …]
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_mbx.c | 45 struct ixgbe_mbx_info *mbx = &hw->mbx; in ixgbe_read_mbx() local 48 if (size > mbx->size) in ixgbe_read_mbx() 49 size = mbx->size; in ixgbe_read_mbx() 51 if (!mbx->ops) in ixgbe_read_mbx() 54 return mbx->ops->read(hw, msg, size, mbx_id); in ixgbe_read_mbx() 68 struct ixgbe_mbx_info *mbx = &hw->mbx; in ixgbe_write_mbx() local 70 if (size > mbx->size) in ixgbe_write_mbx() 73 if (!mbx->ops) in ixgbe_write_mbx() 76 return mbx->ops->write(hw, msg, size, mbx_id); in ixgbe_write_mbx() 88 struct ixgbe_mbx_info *mbx = &hw->mbx; in ixgbe_check_for_msg() local [all …]
|
/drivers/net/ethernet/intel/igbvf/ |
D | mbx.c | 37 struct e1000_mbx_info *mbx = &hw->mbx; in e1000_poll_for_msg() local 38 int countdown = mbx->timeout; in e1000_poll_for_msg() 40 if (!mbx->ops.check_for_msg) in e1000_poll_for_msg() 43 while (countdown && mbx->ops.check_for_msg(hw)) { in e1000_poll_for_msg() 45 udelay(mbx->usec_delay); in e1000_poll_for_msg() 50 mbx->timeout = 0; in e1000_poll_for_msg() 63 struct e1000_mbx_info *mbx = &hw->mbx; in e1000_poll_for_ack() local 64 int countdown = mbx->timeout; in e1000_poll_for_ack() 66 if (!mbx->ops.check_for_ack) in e1000_poll_for_ack() 69 while (countdown && mbx->ops.check_for_ack(hw)) { in e1000_poll_for_ack() [all …]
|
D | vf.c | 82 hw->mbx.ops.init_params = e1000_init_mbx_params_vf; in e1000_init_function_pointers_vf() 124 struct e1000_mbx_info *mbx = &hw->mbx; in e1000_reset_hw_vf() local 136 while (!mbx->ops.check_for_rst(hw) && timeout) { in e1000_reset_hw_vf() 143 mbx->timeout = E1000_VF_MBX_INIT_TIMEOUT; in e1000_reset_hw_vf() 147 mbx->ops.write_posted(hw, msgbuf, 1); in e1000_reset_hw_vf() 152 ret_val = mbx->ops.read_posted(hw, msgbuf, 3); in e1000_reset_hw_vf() 225 struct e1000_mbx_info *mbx = &hw->mbx; in e1000_update_mc_addr_list_vf() local 250 mbx->ops.write_posted(hw, msgbuf, E1000_VFMAILBOX_SIZE); in e1000_update_mc_addr_list_vf() 261 struct e1000_mbx_info *mbx = &hw->mbx; in e1000_set_vfta_vf() local 271 mbx->ops.write_posted(hw, msgbuf, 2); in e1000_set_vfta_vf() [all …]
|
/drivers/net/ethernet/intel/ixgbevf/ |
D | mbx.c | 38 struct ixgbe_mbx_info *mbx = &hw->mbx; in ixgbevf_poll_for_msg() local 39 int countdown = mbx->timeout; in ixgbevf_poll_for_msg() 41 while (countdown && mbx->ops.check_for_msg(hw)) { in ixgbevf_poll_for_msg() 43 udelay(mbx->udelay); in ixgbevf_poll_for_msg() 48 mbx->timeout = 0; in ixgbevf_poll_for_msg() 61 struct ixgbe_mbx_info *mbx = &hw->mbx; in ixgbevf_poll_for_ack() local 62 int countdown = mbx->timeout; in ixgbevf_poll_for_ack() 64 while (countdown && mbx->ops.check_for_ack(hw)) { in ixgbevf_poll_for_ack() 66 udelay(mbx->udelay); in ixgbevf_poll_for_ack() 71 mbx->timeout = 0; in ixgbevf_poll_for_ack() [all …]
|
D | vf.c | 39 struct ixgbe_mbx_info *mbx = &hw->mbx; in ixgbevf_write_msg_read_ack() local 40 s32 retval = mbx->ops.write_posted(hw, msg, size); in ixgbevf_write_msg_read_ack() 45 return mbx->ops.read_posted(hw, retmsg, size); in ixgbevf_write_msg_read_ack() 90 struct ixgbe_mbx_info *mbx = &hw->mbx; in ixgbevf_reset_hw_vf() local 106 while (!mbx->ops.check_for_rst(hw) && timeout) { in ixgbevf_reset_hw_vf() 115 mbx->timeout = IXGBE_VF_MBX_INIT_TIMEOUT; in ixgbevf_reset_hw_vf() 118 mbx->ops.write_posted(hw, msgbuf, 1); in ixgbevf_reset_hw_vf() 126 ret_val = mbx->ops.read_posted(hw, msgbuf, IXGBE_VF_PERMADDR_MSG_LEN); in ixgbevf_reset_hw_vf() 339 err = hw->mbx.ops.write_posted(hw, msgbuf, 1); in ixgbevf_get_reta_locked() 344 err = hw->mbx.ops.read_posted(hw, msgbuf, dwords + 1); in ixgbevf_get_reta_locked() [all …]
|
/drivers/net/ethernet/cavium/thunder/ |
D | nic_main.c | 142 static void nic_send_msg_to_vf(struct nicpf *nic, int vf, union nic_mbx *mbx) in nic_send_msg_to_vf() argument 145 u64 *msg = (u64 *)mbx; in nic_send_msg_to_vf() 169 union nic_mbx mbx = {}; in nic_mbx_send_ready() local 173 mbx.nic_cfg.msg = NIC_MBOX_MSG_READY; in nic_mbx_send_ready() 174 mbx.nic_cfg.vf_id = vf; in nic_mbx_send_ready() 176 mbx.nic_cfg.tns_mode = NIC_TNS_BYPASS_MODE; in nic_mbx_send_ready() 184 ether_addr_copy((u8 *)&mbx.nic_cfg.mac_addr, mac); in nic_mbx_send_ready() 186 mbx.nic_cfg.sqs_mode = (vf >= nic->num_vf_en) ? true : false; in nic_mbx_send_ready() 187 mbx.nic_cfg.node_id = nic->node; in nic_mbx_send_ready() 189 mbx.nic_cfg.loopback_supported = vf < nic->num_vf_en; in nic_mbx_send_ready() [all …]
|
D | nicvf_main.c | 108 static void nicvf_write_to_mbx(struct nicvf *nic, union nic_mbx *mbx) in nicvf_write_to_mbx() argument 110 u64 *msg = (u64 *)mbx; in nicvf_write_to_mbx() 116 int nicvf_send_msg_to_pf(struct nicvf *nic, union nic_mbx *mbx) in nicvf_send_msg_to_pf() argument 124 nicvf_write_to_mbx(nic, mbx); in nicvf_send_msg_to_pf() 131 (mbx->msg.msg & 0xFF), nic->vf_id); in nicvf_send_msg_to_pf() 141 (mbx->msg.msg & 0xFF), nic->vf_id); in nicvf_send_msg_to_pf() 153 union nic_mbx mbx = {}; in nicvf_check_pf_ready() local 155 mbx.msg.msg = NIC_MBOX_MSG_READY; in nicvf_check_pf_ready() 156 if (nicvf_send_msg_to_pf(nic, &mbx)) { in nicvf_check_pf_ready() 175 union nic_mbx mbx = {}; in nicvf_handle_mbx_intr() local [all …]
|
D | nicvf_queues.c | 406 union nic_mbx mbx = {}; in nicvf_reclaim_rcv_queue() local 409 mbx.msg.msg = NIC_MBOX_MSG_RQ_SW_SYNC; in nicvf_reclaim_rcv_queue() 410 nicvf_send_msg_to_pf(nic, &mbx); in nicvf_reclaim_rcv_queue() 497 union nic_mbx mbx = {}; in nicvf_reset_rcv_queue_stats() local 500 mbx.reset_stat.msg = NIC_MBOX_MSG_RESET_STAT_COUNTER; in nicvf_reset_rcv_queue_stats() 501 mbx.reset_stat.rx_stat_mask = 0x3FFF; in nicvf_reset_rcv_queue_stats() 502 mbx.reset_stat.tx_stat_mask = 0x1F; in nicvf_reset_rcv_queue_stats() 503 mbx.reset_stat.rq_stat_mask = 0xFFFF; in nicvf_reset_rcv_queue_stats() 504 mbx.reset_stat.sq_stat_mask = 0xFFFF; in nicvf_reset_rcv_queue_stats() 505 nicvf_send_msg_to_pf(nic, &mbx); in nicvf_reset_rcv_queue_stats() [all …]
|
/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_vfpf.c | 1126 struct bnx2x_vf_mbx *mbx = BP_VF_MBX(bp, vf->index); in bnx2x_vf_mbx_resp_single_tlv() local 1130 type = mbx->first_tlv.tl.type; in bnx2x_vf_mbx_resp_single_tlv() 1134 bnx2x_add_tlv(bp, &mbx->msg->resp, 0, type, length); in bnx2x_vf_mbx_resp_single_tlv() 1135 bnx2x_add_tlv(bp, &mbx->msg->resp, length, CHANNEL_TLV_LIST_END, in bnx2x_vf_mbx_resp_single_tlv() 1143 struct bnx2x_vf_mbx *mbx = BP_VF_MBX(bp, vf->index); in bnx2x_vf_mbx_resp_send_msg() local 1144 struct pfvf_general_resp_tlv *resp = &mbx->msg->resp.general_resp; in bnx2x_vf_mbx_resp_send_msg() 1151 mbx->vf_addr_hi, mbx->vf_addr_lo, mbx->first_tlv.resp_msg_offset); in bnx2x_vf_mbx_resp_send_msg() 1156 vf_addr = HILO_U64(mbx->vf_addr_hi, mbx->vf_addr_lo) + in bnx2x_vf_mbx_resp_send_msg() 1157 mbx->first_tlv.resp_msg_offset; in bnx2x_vf_mbx_resp_send_msg() 1158 pf_addr = mbx->msg_mapping + in bnx2x_vf_mbx_resp_send_msg() [all …]
|
/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_83xx_hw.c | 492 static inline void qlcnic_83xx_notify_mbx_response(struct qlcnic_mailbox *mbx) in qlcnic_83xx_notify_mbx_response() argument 494 mbx->rsp_status = QLC_83XX_MBX_RESPONSE_ARRIVED; in qlcnic_83xx_notify_mbx_response() 495 complete(&mbx->completion); in qlcnic_83xx_notify_mbx_response() 501 struct qlcnic_mailbox *mbx = adapter->ahw->mailbox; in qlcnic_83xx_poll_process_aen() local 504 spin_lock_irqsave(&mbx->aen_lock, flags); in qlcnic_83xx_poll_process_aen() 513 if (mbx->rsp_status != rsp_status) in qlcnic_83xx_poll_process_aen() 514 qlcnic_83xx_notify_mbx_response(mbx); in qlcnic_83xx_poll_process_aen() 518 spin_unlock_irqrestore(&mbx->aen_lock, flags); in qlcnic_83xx_poll_process_aen() 850 struct qlcnic_mailbox *mbx = adapter->ahw->mailbox; in qlcnic_83xx_issue_cmd() local 855 if (!mbx) in qlcnic_83xx_issue_cmd() [all …]
|
D | qlcnic_sriov_common.c | 312 struct qlcnic_mailbox *mbx = ahw->mailbox; in qlcnic_sriov_post_bc_msg() local 325 err = mbx->ops->enqueue_cmd(adapter, &cmd, &timeout); in qlcnic_sriov_post_bc_msg() 339 flush_workqueue(mbx->work_q); in qlcnic_sriov_post_bc_msg() 714 static int qlcnic_sriov_alloc_bc_mbx_args(struct qlcnic_cmd_args *mbx, u32 type) in qlcnic_sriov_alloc_bc_mbx_args() argument 724 mbx->op_type = QLC_BC_CMD; in qlcnic_sriov_alloc_bc_mbx_args() 725 mbx->req.num = mbx_tbl[i].in_args; in qlcnic_sriov_alloc_bc_mbx_args() 726 mbx->rsp.num = mbx_tbl[i].out_args; in qlcnic_sriov_alloc_bc_mbx_args() 727 mbx->req.arg = kcalloc(mbx->req.num, sizeof(u32), in qlcnic_sriov_alloc_bc_mbx_args() 729 if (!mbx->req.arg) in qlcnic_sriov_alloc_bc_mbx_args() 731 mbx->rsp.arg = kcalloc(mbx->rsp.num, sizeof(u32), in qlcnic_sriov_alloc_bc_mbx_args() [all …]
|
D | qlcnic_dcb.c | 495 struct qlcnic_dcb_mbx_params *mbx; in qlcnic_82xx_dcb_get_cee_cfg() local 498 mbx = dcb->param; in qlcnic_82xx_dcb_get_cee_cfg() 499 if (!mbx) in qlcnic_82xx_dcb_get_cee_cfg() 502 err = qlcnic_dcb_query_cee_param(dcb, (char *)&mbx->type[0], in qlcnic_82xx_dcb_get_cee_cfg() 507 err = qlcnic_dcb_query_cee_param(dcb, (char *)&mbx->type[1], in qlcnic_82xx_dcb_get_cee_cfg() 512 err = qlcnic_dcb_query_cee_param(dcb, (char *)&mbx->type[2], in qlcnic_82xx_dcb_get_cee_cfg() 517 mbx->prio_tc_map = QLC_82XX_DCB_PRIO_TC_MAP; in qlcnic_82xx_dcb_get_cee_cfg() 656 static void qlcnic_dcb_fill_cee_tc_params(struct qlcnic_dcb_mbx_params *mbx, in qlcnic_dcb_fill_cee_tc_params() argument 664 tc = QLC_DCB_GET_TC_PRIO(mbx->prio_tc_map, i); in qlcnic_dcb_fill_cee_tc_params() 745 struct qlcnic_dcb_mbx_params *mbx = adapter->dcb->param; in qlcnic_dcb_map_cee_params() local [all …]
|
/drivers/net/ethernet/qlogic/qed/ |
D | qed_sriov.c | 1019 struct qed_iov_vf_mbx *mbx = &p_vf->vf_mbx; in qed_iov_send_response() local 1023 mbx->reply_virt->default_resp.hdr.status = status; in qed_iov_send_response() 1025 qed_dp_tlv_list(p_hwfn, mbx->reply_virt); in qed_iov_send_response() 1033 qed_dmae_host2host(p_hwfn, p_ptt, mbx->reply_phys + sizeof(u64), in qed_iov_send_response() 1034 mbx->req_virt->first_tlv.reply_address + in qed_iov_send_response() 1039 qed_dmae_host2host(p_hwfn, p_ptt, mbx->reply_phys, in qed_iov_send_response() 1040 mbx->req_virt->first_tlv.reply_address, in qed_iov_send_response() 1122 struct qed_iov_vf_mbx *mbx = &vf_info->vf_mbx; in qed_iov_prepare_resp() local 1124 mbx->offset = (u8 *)mbx->reply_virt; in qed_iov_prepare_resp() 1126 qed_add_tlv(p_hwfn, &mbx->offset, type, length); in qed_iov_prepare_resp() [all …]
|
/drivers/atm/ |
D | zatm.c | 359 static void poll_rx(struct atm_dev *dev,int mbx) in poll_rx() argument 368 pos = (zatm_dev->mbx_start[mbx] & ~0xffffUL) | zin(MTA(mbx)); in poll_rx() 369 while (x = zin(MWA(mbx)), (pos & 0xffff) != x) { in poll_rx() 377 if (((pos += 16) & 0xffff) == zatm_dev->mbx_end[mbx]) in poll_rx() 378 pos = zatm_dev->mbx_start[mbx]; in poll_rx() 474 zout(pos & 0xffff,MTA(mbx)); in poll_rx() 742 static void poll_tx(struct atm_dev *dev,int mbx) in poll_tx() argument 750 pos = (zatm_dev->mbx_start[mbx] & ~0xffffUL) | zin(MTA(mbx)); in poll_tx() 751 while (x = zin(MWA(mbx)), (pos & 0xffff) != x) { in poll_tx() 766 chan = (zatm_dev->mbx_start[mbx][pos >> 2] & uPD98401_TXI_CONN) in poll_tx() [all …]
|
/drivers/scsi/qla2xxx/ |
D | qla_isr.c | 587 uint16_t cnt, mbx; in qla2x00_async_event() local 671 mbx = (IS_QLA81XX(ha) || IS_QLA83XX(ha) || IS_QLA27XX(ha)) ? in qla2x00_async_event() 675 "mbx7=%xh.\n", mb[1], mb[2], mb[3], mbx); in qla2x00_async_event() 688 if ((mbx & MBX_3) && (ha->port_no == 0)) in qla2x00_async_event() 767 mbx = (IS_QLA81XX(ha) || IS_QLA8031(ha)) in qla2x00_async_event() 769 mbx = (IS_P3P_TYPE(ha)) ? RD_REG_WORD(®82->mailbox_out[4]) in qla2x00_async_event() 770 : mbx; in qla2x00_async_event() 773 mb[1], mb[2], mb[3], mbx); in qla2x00_async_event() 1268 struct mbx_entry *mbx) in qla2x00_mbx_iocb_entry() argument 1278 sp = qla2x00_get_sp_from_handle(vha, func, req, mbx); in qla2x00_mbx_iocb_entry() [all …]
|
D | qla_iocb.c | 1894 qla2x00_login_iocb(srb_t *sp, struct mbx_entry *mbx) in qla2x00_login_iocb() argument 1900 mbx->entry_type = MBX_IOCB_TYPE; in qla2x00_login_iocb() 1901 SET_TARGET_ID(ha, mbx->loop_id, sp->fcport->loop_id); in qla2x00_login_iocb() 1902 mbx->mb0 = cpu_to_le16(MBC_LOGIN_FABRIC_PORT); in qla2x00_login_iocb() 1906 mbx->mb1 = cpu_to_le16(sp->fcport->loop_id); in qla2x00_login_iocb() 1907 mbx->mb10 = cpu_to_le16(opts); in qla2x00_login_iocb() 1909 mbx->mb1 = cpu_to_le16((sp->fcport->loop_id << 8) | opts); in qla2x00_login_iocb() 1911 mbx->mb2 = cpu_to_le16(sp->fcport->d_id.b.domain); in qla2x00_login_iocb() 1912 mbx->mb3 = cpu_to_le16(sp->fcport->d_id.b.area << 8 | in qla2x00_login_iocb() 1914 mbx->mb9 = cpu_to_le16(sp->fcport->vha->vp_idx); in qla2x00_login_iocb() [all …]
|