Home
last modified time | relevance | path

Searched full:hba (Results 1 – 25 of 622) sorted by relevance

12345678910>>...25

/kernel/linux/linux-6.6/drivers/ufs/core/
Dufshcd-priv.h9 static inline bool ufshcd_is_user_access_allowed(struct ufs_hba *hba) in ufshcd_is_user_access_allowed() argument
11 return !hba->shutting_down; in ufshcd_is_user_access_allowed()
14 void ufshcd_schedule_eh_work(struct ufs_hba *hba);
17 struct ufs_hba *hba) in ufshcd_keep_autobkops_enabled_except_suspend() argument
19 return hba->caps & UFSHCD_CAP_KEEP_AUTO_BKOPS_ENABLED_EXCEPT_SUSPEND; in ufshcd_keep_autobkops_enabled_except_suspend()
22 static inline u8 ufshcd_wb_get_query_index(struct ufs_hba *hba) in ufshcd_wb_get_query_index() argument
24 if (hba->dev_info.wb_buffer_type == WB_BUF_MODE_LU_DEDICATED) in ufshcd_wb_get_query_index()
25 return hba->dev_info.wb_dedicated_lu; in ufshcd_wb_get_query_index()
29 static inline bool ufshcd_is_wb_buf_flush_allowed(struct ufs_hba *hba) in ufshcd_is_wb_buf_flush_allowed() argument
31 return ufshcd_is_wb_allowed(hba) && in ufshcd_is_wb_buf_flush_allowed()
[all …]
Dufshcd.c104 static bool is_mcq_supported(struct ufs_hba *hba) in is_mcq_supported() argument
106 return hba->mcq_sup && use_mcq_mode; in is_mcq_supported()
129 int ufshcd_dump_regs(struct ufs_hba *hba, size_t offset, size_t len, in ufshcd_dump_regs() argument
147 regs[pos / 4] = ufshcd_readl(hba, offset + pos); in ufshcd_dump_regs()
265 static irqreturn_t ufshcd_tmc_handler(struct ufs_hba *hba);
267 static int ufshcd_reset_and_restore(struct ufs_hba *hba);
269 static int ufshcd_clear_tm_cmd(struct ufs_hba *hba, int tag);
270 static void ufshcd_hba_exit(struct ufs_hba *hba);
271 static int ufshcd_probe_hba(struct ufs_hba *hba, bool init_dev_params);
272 static int ufshcd_setup_clocks(struct ufs_hba *hba, bool on);
[all …]
Dufs-mcq.c85 * @hba: per adapter instance
91 void ufshcd_mcq_config_mac(struct ufs_hba *hba, u32 max_active_cmds) in ufshcd_mcq_config_mac() argument
95 val = ufshcd_readl(hba, REG_UFS_MCQ_CFG); in ufshcd_mcq_config_mac()
98 ufshcd_writel(hba, val, REG_UFS_MCQ_CFG); in ufshcd_mcq_config_mac()
105 * @hba: per adapter instance
111 struct ufs_hw_queue *ufshcd_mcq_req_to_hwq(struct ufs_hba *hba, in ufshcd_mcq_req_to_hwq() argument
117 return &hba->uhq[hwq]; in ufshcd_mcq_req_to_hwq()
122 * @hba: per adapter instance
132 int ufshcd_mcq_decide_queue_depth(struct ufs_hba *hba) in ufshcd_mcq_decide_queue_depth() argument
137 mac = ufshcd_mcq_vops_get_hba_mac(hba); in ufshcd_mcq_decide_queue_depth()
[all …]
Dufshcd-crypto.c20 static int ufshcd_program_key(struct ufs_hba *hba, in ufshcd_program_key() argument
24 u32 slot_offset = hba->crypto_cfg_register + slot * sizeof(*cfg); in ufshcd_program_key()
27 ufshcd_hold(hba); in ufshcd_program_key()
29 if (hba->vops && hba->vops->program_key) { in ufshcd_program_key()
30 err = hba->vops->program_key(hba, cfg, slot); in ufshcd_program_key()
35 ufshcd_writel(hba, 0, slot_offset + 16 * sizeof(cfg->reg_val[0])); in ufshcd_program_key()
37 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[i]), in ufshcd_program_key()
41 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[17]), in ufshcd_program_key()
44 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[16]), in ufshcd_program_key()
47 ufshcd_release(hba); in ufshcd_program_key()
[all …]
Dufs-debugfs.c18 /* @file corresponds to a debugfs attribute in directory hba->debugfs_root. */
36 struct ufs_hba *hba = hba_from_file(s->file); in ufs_debugfs_stats_show() local
37 struct ufs_event_hist *e = hba->ufs_stats.event; in ufs_debugfs_stats_show()
62 struct ufs_hba *hba = data; in ee_usr_mask_get() local
64 *val = hba->ee_usr_mask; in ee_usr_mask_get()
68 static int ufs_debugfs_get_user_access(struct ufs_hba *hba) in ufs_debugfs_get_user_access() argument
69 __acquires(&hba->host_sem) in ufs_debugfs_get_user_access()
71 down(&hba->host_sem); in ufs_debugfs_get_user_access()
72 if (!ufshcd_is_user_access_allowed(hba)) { in ufs_debugfs_get_user_access()
73 up(&hba->host_sem); in ufs_debugfs_get_user_access()
[all …]
Dufs-hwmon.c14 struct ufs_hba *hba; member
18 static int ufs_read_temp_enable(struct ufs_hba *hba, u8 mask, long *val) in ufs_read_temp_enable() argument
23 err = ufshcd_query_attr(hba, UPIU_QUERY_OPCODE_READ_ATTR, QUERY_ATTR_IDN_EE_CONTROL, 0, 0, in ufs_read_temp_enable()
33 static int ufs_get_temp(struct ufs_hba *hba, enum attr_idn idn, long *val) in ufs_get_temp() argument
38 err = ufshcd_query_attr(hba, UPIU_QUERY_OPCODE_READ_ATTR, idn, 0, 0, &value); in ufs_get_temp()
54 struct ufs_hba *hba = data->hba; in ufs_hwmon_read() local
57 down(&hba->host_sem); in ufs_hwmon_read()
59 if (!ufshcd_is_user_access_allowed(hba)) { in ufs_hwmon_read()
60 up(&hba->host_sem); in ufs_hwmon_read()
64 ufshcd_rpm_get_sync(hba); in ufs_hwmon_read()
[all …]
/kernel/linux/linux-5.10/drivers/scsi/ufs/
Dufshcd.c97 int ufshcd_dump_regs(struct ufs_hba *hba, size_t offset, size_t len, in ufshcd_dump_regs() argument
115 regs[pos / 4] = ufshcd_readl(hba, offset + pos); in ufshcd_dump_regs()
222 static irqreturn_t ufshcd_tmc_handler(struct ufs_hba *hba);
224 static int ufshcd_reset_and_restore(struct ufs_hba *hba);
226 static int ufshcd_clear_tm_cmd(struct ufs_hba *hba, int tag);
227 static void ufshcd_hba_exit(struct ufs_hba *hba);
228 static int ufshcd_probe_hba(struct ufs_hba *hba, bool async);
229 static int ufshcd_setup_clocks(struct ufs_hba *hba, bool on);
230 static int ufshcd_uic_hibern8_enter(struct ufs_hba *hba);
231 static inline void ufshcd_add_delay_before_dme_cmd(struct ufs_hba *hba);
[all …]
Dufs-mediatek.c63 static bool ufs_mtk_is_boost_crypt_enabled(struct ufs_hba *hba) in ufs_mtk_is_boost_crypt_enabled() argument
65 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_boost_crypt_enabled()
70 static void ufs_mtk_cfg_unipro_cg(struct ufs_hba *hba, bool enable) in ufs_mtk_cfg_unipro_cg() argument
75 ufshcd_dme_get(hba, in ufs_mtk_cfg_unipro_cg()
81 ufshcd_dme_set(hba, in ufs_mtk_cfg_unipro_cg()
84 ufshcd_dme_get(hba, in ufs_mtk_cfg_unipro_cg()
87 ufshcd_dme_set(hba, in ufs_mtk_cfg_unipro_cg()
90 ufshcd_dme_get(hba, in ufs_mtk_cfg_unipro_cg()
95 ufshcd_dme_set(hba, in ufs_mtk_cfg_unipro_cg()
98 ufshcd_dme_get(hba, in ufs_mtk_cfg_unipro_cg()
[all …]
Dufs-hisi.c23 static int ufs_hisi_check_hibern8(struct ufs_hba *hba) in ufs_hisi_check_hibern8() argument
31 err = ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 0), in ufs_hisi_check_hibern8()
33 err |= ufshcd_dme_get(hba, in ufs_hisi_check_hibern8()
48 err = ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 0), in ufs_hisi_check_hibern8()
50 err |= ufshcd_dme_get(hba, in ufs_hisi_check_hibern8()
55 dev_err(hba->dev, "%s: unable to get TX_FSM_STATE, err %d\n", in ufs_hisi_check_hibern8()
60 dev_err(hba->dev, "%s: invalid TX_FSM_STATE, lane0 = %d, lane1 = %d\n", in ufs_hisi_check_hibern8()
67 static void ufs_hisi_clk_init(struct ufs_hba *hba) in ufs_hisi_clk_init() argument
69 struct ufs_hisi_host *host = ufshcd_get_variant(hba); in ufs_hisi_clk_init()
81 static void ufs_hisi_soc_init(struct ufs_hba *hba) in ufs_hisi_soc_init() argument
[all …]
Dufshcd.h98 #define ufshcd_is_link_off(hba) ((hba)->uic_link_state == UIC_LINK_OFF_STATE) argument
99 #define ufshcd_is_link_active(hba) ((hba)->uic_link_state == \ argument
101 #define ufshcd_is_link_hibern8(hba) ((hba)->uic_link_state == \ argument
103 #define ufshcd_is_link_broken(hba) ((hba)->uic_link_state == \ argument
105 #define ufshcd_set_link_off(hba) ((hba)->uic_link_state = UIC_LINK_OFF_STATE) argument
106 #define ufshcd_set_link_active(hba) ((hba)->uic_link_state = \ argument
108 #define ufshcd_set_link_hibern8(hba) ((hba)->uic_link_state = \ argument
110 #define ufshcd_set_link_broken(hba) ((hba)->uic_link_state = \ argument
226 * @list: list headed by hba->clk_list_head
318 int (*apply_dev_quirks)(struct ufs_hba *hba);
[all …]
Dufshcd-crypto.c20 static int ufshcd_program_key(struct ufs_hba *hba, in ufshcd_program_key() argument
24 u32 slot_offset = hba->crypto_cfg_register + slot * sizeof(*cfg); in ufshcd_program_key()
27 ufshcd_hold(hba, false); in ufshcd_program_key()
29 if (hba->vops && hba->vops->program_key) { in ufshcd_program_key()
30 err = hba->vops->program_key(hba, cfg, slot); in ufshcd_program_key()
35 ufshcd_writel(hba, 0, slot_offset + 16 * sizeof(cfg->reg_val[0])); in ufshcd_program_key()
37 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[i]), in ufshcd_program_key()
41 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[17]), in ufshcd_program_key()
44 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[16]), in ufshcd_program_key()
47 ufshcd_release(hba); in ufshcd_program_key()
[all …]
Dcdns-pltfrm.c33 * @hba: per adapter instance
36 static void cdns_ufs_get_l4_attr(struct ufs_hba *hba) in cdns_ufs_get_l4_attr() argument
38 struct cdns_ufs_host *host = ufshcd_get_variant(hba); in cdns_ufs_get_l4_attr()
40 ufshcd_dme_get(hba, UIC_ARG_MIB(T_PEERDEVICEID), in cdns_ufs_get_l4_attr()
42 ufshcd_dme_get(hba, UIC_ARG_MIB(T_PEERCPORTID), in cdns_ufs_get_l4_attr()
44 ufshcd_dme_get(hba, UIC_ARG_MIB(T_TRAFFICCLASS), in cdns_ufs_get_l4_attr()
46 ufshcd_dme_get(hba, UIC_ARG_MIB(T_PROTOCOLID), in cdns_ufs_get_l4_attr()
48 ufshcd_dme_get(hba, UIC_ARG_MIB(T_CPORTFLAGS), in cdns_ufs_get_l4_attr()
50 ufshcd_dme_get(hba, UIC_ARG_MIB(T_TXTOKENVALUE), in cdns_ufs_get_l4_attr()
52 ufshcd_dme_get(hba, UIC_ARG_MIB(T_RXTOKENVALUE), in cdns_ufs_get_l4_attr()
[all …]
/kernel/linux/linux-6.6/drivers/ufs/host/
Dufs-mediatek.c31 static int ufs_mtk_config_mcq(struct ufs_hba *hba, bool irq);
94 static bool ufs_mtk_is_boost_crypt_enabled(struct ufs_hba *hba) in ufs_mtk_is_boost_crypt_enabled() argument
96 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_boost_crypt_enabled()
101 static bool ufs_mtk_is_va09_supported(struct ufs_hba *hba) in ufs_mtk_is_va09_supported() argument
103 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_va09_supported()
108 static bool ufs_mtk_is_broken_vcc(struct ufs_hba *hba) in ufs_mtk_is_broken_vcc() argument
110 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_broken_vcc()
115 static bool ufs_mtk_is_pmc_via_fastauto(struct ufs_hba *hba) in ufs_mtk_is_pmc_via_fastauto() argument
117 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_pmc_via_fastauto()
122 static void ufs_mtk_cfg_unipro_cg(struct ufs_hba *hba, bool enable) in ufs_mtk_cfg_unipro_cg() argument
[all …]
Dufs-hisi.c25 static int ufs_hisi_check_hibern8(struct ufs_hba *hba) in ufs_hisi_check_hibern8() argument
33 err = ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 0), in ufs_hisi_check_hibern8()
35 err |= ufshcd_dme_get(hba, in ufs_hisi_check_hibern8()
50 err = ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 0), in ufs_hisi_check_hibern8()
52 err |= ufshcd_dme_get(hba, in ufs_hisi_check_hibern8()
57 dev_err(hba->dev, "%s: unable to get TX_FSM_STATE, err %d\n", in ufs_hisi_check_hibern8()
62 dev_err(hba->dev, "%s: invalid TX_FSM_STATE, lane0 = %d, lane1 = %d\n", in ufs_hisi_check_hibern8()
69 static void ufs_hisi_clk_init(struct ufs_hba *hba) in ufs_hisi_clk_init() argument
71 struct ufs_hisi_host *host = ufshcd_get_variant(hba); in ufs_hisi_clk_init()
83 static void ufs_hisi_soc_init(struct ufs_hba *hba) in ufs_hisi_soc_init() argument
[all …]
Dufs-exynos.c214 struct ufs_hba *hba = ufs->hba; in exynosauto_ufs_post_hce_enable() local
217 ufshcd_rmwl(hba, MHCTRL_EN_VH_MASK, MHCTRL_EN_VH(1), MHCTRL); in exynosauto_ufs_post_hce_enable()
228 struct ufs_hba *hba = ufs->hba; in exynosauto_ufs_pre_link() local
235 ufshcd_dme_set(hba, UIC_ARG_MIB(0x200), 0x40); in exynosauto_ufs_pre_link()
237 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_CLK_PRD, i), in exynosauto_ufs_pre_link()
239 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_CLK_PRD_EN, i), 0x0); in exynosauto_ufs_pre_link()
241 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_LINERESET_VALUE2, i), in exynosauto_ufs_pre_link()
243 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_LINERESET_VALUE1, i), in exynosauto_ufs_pre_link()
245 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_LINERESET_VALUE0, i), in exynosauto_ufs_pre_link()
248 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x2f, i), 0x79); in exynosauto_ufs_pre_link()
[all …]
Dufshcd-pci.c24 void (*late_init)(struct ufs_hba *hba);
97 static int ufs_intel_hce_enable_notify(struct ufs_hba *hba, in ufs_intel_hce_enable_notify() argument
101 if (status == POST_CHANGE && hba->caps & UFSHCD_CAP_CRYPTO) { in ufs_intel_hce_enable_notify()
102 u32 hce = ufshcd_readl(hba, REG_CONTROLLER_ENABLE); in ufs_intel_hce_enable_notify()
105 ufshcd_writel(hba, hce, REG_CONTROLLER_ENABLE); in ufs_intel_hce_enable_notify()
111 static int ufs_intel_disable_lcc(struct ufs_hba *hba) in ufs_intel_disable_lcc() argument
116 ufshcd_dme_get(hba, attr, &lcc_enable); in ufs_intel_disable_lcc()
118 ufshcd_disable_host_tx_lcc(hba); in ufs_intel_disable_lcc()
123 static int ufs_intel_link_startup_notify(struct ufs_hba *hba, in ufs_intel_link_startup_notify() argument
130 err = ufs_intel_disable_lcc(hba); in ufs_intel_link_startup_notify()
[all …]
Dufs-qcom.c96 static int ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div(struct ufs_hba *hba,
108 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_enable()
114 struct ufs_hba *hba = host->hba; in ufs_qcom_ice_init() local
115 struct device *dev = hba->dev; in ufs_qcom_ice_init()
128 hba->caps |= UFSHCD_CAP_CRYPTO; in ufs_qcom_ice_init()
135 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_resume()
143 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_suspend()
149 static int ufs_qcom_ice_program_key(struct ufs_hba *hba, in ufs_qcom_ice_program_key() argument
153 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_ice_program_key()
159 cap = hba->crypto_cap_array[cfg->crypto_cap_idx]; in ufs_qcom_ice_program_key()
[all …]
/kernel/linux/linux-6.6/drivers/scsi/
Dhptiop.c39 static int iop_send_sync_msg(struct hptiop_hba *hba, u32 msg, u32 millisec);
40 static void hptiop_finish_scsi_req(struct hptiop_hba *hba, u32 tag,
42 static void hptiop_host_request_callback_itl(struct hptiop_hba *hba, u32 tag);
43 static void hptiop_iop_request_callback_itl(struct hptiop_hba *hba, u32 tag);
44 static void hptiop_message_callback(struct hptiop_hba *hba, u32 msg);
46 static int iop_wait_ready_itl(struct hptiop_hba *hba, u32 millisec) in iop_wait_ready_itl() argument
52 req = readl(&hba->u.itl.iop->inbound_queue); in iop_wait_ready_itl()
59 writel(req, &hba->u.itl.iop->outbound_queue); in iop_wait_ready_itl()
60 readl(&hba->u.itl.iop->outbound_intstatus); in iop_wait_ready_itl()
67 static int iop_wait_ready_mv(struct hptiop_hba *hba, u32 millisec) in iop_wait_ready_mv() argument
[all …]
Dstex.c392 static struct status_msg *stex_get_status(struct st_hba *hba) in stex_get_status() argument
394 struct status_msg *status = hba->status_buffer + hba->status_tail; in stex_get_status()
396 ++hba->status_tail; in stex_get_status()
397 hba->status_tail %= hba->sts_count+1; in stex_get_status()
410 static struct req_msg *stex_alloc_req(struct st_hba *hba) in stex_alloc_req() argument
412 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req()
414 ++hba->req_head; in stex_alloc_req()
415 hba->req_head %= hba->rq_count+1; in stex_alloc_req()
420 static struct req_msg *stex_ss_alloc_req(struct st_hba *hba) in stex_ss_alloc_req() argument
422 return (struct req_msg *)(hba->dma_mem + in stex_ss_alloc_req()
[all …]
/kernel/linux/linux-5.10/drivers/scsi/
Dhptiop.c39 static int iop_send_sync_msg(struct hptiop_hba *hba, u32 msg, u32 millisec);
40 static void hptiop_finish_scsi_req(struct hptiop_hba *hba, u32 tag,
42 static void hptiop_host_request_callback_itl(struct hptiop_hba *hba, u32 tag);
43 static void hptiop_iop_request_callback_itl(struct hptiop_hba *hba, u32 tag);
44 static void hptiop_message_callback(struct hptiop_hba *hba, u32 msg);
46 static int iop_wait_ready_itl(struct hptiop_hba *hba, u32 millisec) in iop_wait_ready_itl() argument
52 req = readl(&hba->u.itl.iop->inbound_queue); in iop_wait_ready_itl()
59 writel(req, &hba->u.itl.iop->outbound_queue); in iop_wait_ready_itl()
60 readl(&hba->u.itl.iop->outbound_intstatus); in iop_wait_ready_itl()
67 static int iop_wait_ready_mv(struct hptiop_hba *hba, u32 millisec) in iop_wait_ready_mv() argument
[all …]
Dstex.c392 static struct status_msg *stex_get_status(struct st_hba *hba) in stex_get_status() argument
394 struct status_msg *status = hba->status_buffer + hba->status_tail; in stex_get_status()
396 ++hba->status_tail; in stex_get_status()
397 hba->status_tail %= hba->sts_count+1; in stex_get_status()
413 static struct req_msg *stex_alloc_req(struct st_hba *hba) in stex_alloc_req() argument
415 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req()
417 ++hba->req_head; in stex_alloc_req()
418 hba->req_head %= hba->rq_count+1; in stex_alloc_req()
423 static struct req_msg *stex_ss_alloc_req(struct st_hba *hba) in stex_ss_alloc_req() argument
425 return (struct req_msg *)(hba->dma_mem + in stex_ss_alloc_req()
[all …]
/kernel/linux/linux-5.10/drivers/scsi/bnx2i/
Dbnx2i_iscsi.c34 static int bnx2i_adapter_ready(struct bnx2i_hba *hba) in bnx2i_adapter_ready() argument
38 if (!hba || !test_bit(ADAPTER_STATE_UP, &hba->adapter_state) || in bnx2i_adapter_ready()
39 test_bit(ADAPTER_STATE_GOING_DOWN, &hba->adapter_state) || in bnx2i_adapter_ready()
40 test_bit(ADAPTER_STATE_LINK_DOWN, &hba->adapter_state)) in bnx2i_adapter_ready()
138 * @hba: adapter instance
143 static int bnx2i_map_scsi_sg(struct bnx2i_hba *hba, struct bnx2i_cmd *cmd) in bnx2i_map_scsi_sg() argument
190 bd_count = bnx2i_map_scsi_sg(cmd->conn->hba, cmd); in bnx2i_iscsi_map_sg_list()
230 * @hba: pointer to adapter instance
238 static int bnx2i_bind_conn_to_iscsi_cid(struct bnx2i_hba *hba, in bnx2i_bind_conn_to_iscsi_cid() argument
242 if (hba && hba->cid_que.conn_cid_tbl[iscsi_cid]) { in bnx2i_bind_conn_to_iscsi_cid()
[all …]
Dbnx2i_init.c75 * @hba: Adapter structure pointer
82 void bnx2i_identify_device(struct bnx2i_hba *hba, struct cnic_dev *dev) in bnx2i_identify_device() argument
84 hba->cnic_dev_type = 0; in bnx2i_identify_device()
86 if (hba->pci_did == PCI_DEVICE_ID_NX2_5706 || in bnx2i_identify_device()
87 hba->pci_did == PCI_DEVICE_ID_NX2_5706S) { in bnx2i_identify_device()
88 set_bit(BNX2I_NX2_DEV_5706, &hba->cnic_dev_type); in bnx2i_identify_device()
89 } else if (hba->pci_did == PCI_DEVICE_ID_NX2_5708 || in bnx2i_identify_device()
90 hba->pci_did == PCI_DEVICE_ID_NX2_5708S) { in bnx2i_identify_device()
91 set_bit(BNX2I_NX2_DEV_5708, &hba->cnic_dev_type); in bnx2i_identify_device()
92 } else if (hba->pci_did == PCI_DEVICE_ID_NX2_5709 || in bnx2i_identify_device()
[all …]
/kernel/linux/linux-6.6/drivers/scsi/bnx2i/
Dbnx2i_iscsi.c34 static int bnx2i_adapter_ready(struct bnx2i_hba *hba) in bnx2i_adapter_ready() argument
38 if (!hba || !test_bit(ADAPTER_STATE_UP, &hba->adapter_state) || in bnx2i_adapter_ready()
39 test_bit(ADAPTER_STATE_GOING_DOWN, &hba->adapter_state) || in bnx2i_adapter_ready()
40 test_bit(ADAPTER_STATE_LINK_DOWN, &hba->adapter_state)) in bnx2i_adapter_ready()
138 * @hba: adapter instance
143 static int bnx2i_map_scsi_sg(struct bnx2i_hba *hba, struct bnx2i_cmd *cmd) in bnx2i_map_scsi_sg() argument
190 bd_count = bnx2i_map_scsi_sg(cmd->conn->hba, cmd); in bnx2i_iscsi_map_sg_list()
230 * @hba: pointer to adapter instance
238 static int bnx2i_bind_conn_to_iscsi_cid(struct bnx2i_hba *hba, in bnx2i_bind_conn_to_iscsi_cid() argument
242 if (hba && hba->cid_que.conn_cid_tbl[iscsi_cid]) { in bnx2i_bind_conn_to_iscsi_cid()
[all …]
Dbnx2i_init.c75 * @hba: Adapter structure pointer
82 void bnx2i_identify_device(struct bnx2i_hba *hba, struct cnic_dev *dev) in bnx2i_identify_device() argument
84 hba->cnic_dev_type = 0; in bnx2i_identify_device()
86 if (hba->pci_did == PCI_DEVICE_ID_NX2_5706 || in bnx2i_identify_device()
87 hba->pci_did == PCI_DEVICE_ID_NX2_5706S) { in bnx2i_identify_device()
88 set_bit(BNX2I_NX2_DEV_5706, &hba->cnic_dev_type); in bnx2i_identify_device()
89 } else if (hba->pci_did == PCI_DEVICE_ID_NX2_5708 || in bnx2i_identify_device()
90 hba->pci_did == PCI_DEVICE_ID_NX2_5708S) { in bnx2i_identify_device()
91 set_bit(BNX2I_NX2_DEV_5708, &hba->cnic_dev_type); in bnx2i_identify_device()
92 } else if (hba->pci_did == PCI_DEVICE_ID_NX2_5709 || in bnx2i_identify_device()
[all …]

12345678910>>...25