/drivers/crypto/qat/qat_common/ |
D | adf_init.c | 103 int adf_dev_init(struct adf_accel_dev *accel_dev) in adf_dev_init() argument 107 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_dev_init() 110 dev_err(&GET_DEV(accel_dev), in adf_dev_init() 115 if (!test_bit(ADF_STATUS_CONFIGURED, &accel_dev->status)) { in adf_dev_init() 116 dev_err(&GET_DEV(accel_dev), "Device not configured\n"); in adf_dev_init() 120 if (adf_init_etr_data(accel_dev)) { in adf_dev_init() 121 dev_err(&GET_DEV(accel_dev), "Failed initialize etr\n"); in adf_dev_init() 125 if (hw_data->init_admin_comms && hw_data->init_admin_comms(accel_dev)) { in adf_dev_init() 126 dev_err(&GET_DEV(accel_dev), "Failed initialize admin comms\n"); in adf_dev_init() 130 if (hw_data->init_arb && hw_data->init_arb(accel_dev)) { in adf_dev_init() [all …]
|
D | adf_vf_isr.c | 71 struct adf_accel_dev *accel_dev; member 75 static int adf_enable_msi(struct adf_accel_dev *accel_dev) in adf_enable_msi() argument 77 struct adf_accel_pci *pci_dev_info = &accel_dev->accel_pci_dev; in adf_enable_msi() 81 dev_err(&GET_DEV(accel_dev), in adf_enable_msi() 86 accel_dev->vf.irq_name = kzalloc(ADF_MAX_MSIX_VECTOR_NAME, GFP_KERNEL); in adf_enable_msi() 87 if (!accel_dev->vf.irq_name) in adf_enable_msi() 93 static void adf_disable_msi(struct adf_accel_dev *accel_dev) in adf_disable_msi() argument 95 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_disable_msi() 97 kfree(accel_dev->vf.irq_name); in adf_disable_msi() 105 struct adf_accel_dev *accel_dev = stop_data->accel_dev; in adf_dev_stop_async() local [all …]
|
D | adf_isr.c | 62 static int adf_enable_msix(struct adf_accel_dev *accel_dev) in adf_enable_msix() argument 64 struct adf_accel_pci *pci_dev_info = &accel_dev->accel_pci_dev; in adf_enable_msix() 65 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_msix() 69 if (!accel_dev->pf.vf_info) { in adf_enable_msix() 83 dev_err(&GET_DEV(accel_dev), "Failed to enable MSI-X IRQ(s)\n"); in adf_enable_msix() 105 struct adf_accel_dev *accel_dev = dev_ptr; in adf_msix_isr_ae() local 109 if (accel_dev->pf.vf_info) { in adf_msix_isr_ae() 110 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_msix_isr_ae() 112 &GET_BARS(accel_dev)[hw_data->get_misc_bar_id(hw_data)]; in adf_msix_isr_ae() 128 adf_disable_vf2pf_interrupts(accel_dev, vf_mask); in adf_msix_isr_ae() [all …]
|
D | adf_common_drv.h | 88 int (*event_hld)(struct adf_accel_dev *accel_dev, 104 int adf_dev_init(struct adf_accel_dev *accel_dev); 105 int adf_dev_start(struct adf_accel_dev *accel_dev); 106 void adf_dev_stop(struct adf_accel_dev *accel_dev); 107 void adf_dev_shutdown(struct adf_accel_dev *accel_dev); 109 int adf_iov_putmsg(struct adf_accel_dev *accel_dev, u32 msg, u8 vf_nr); 110 void adf_pf2vf_notify_restarting(struct adf_accel_dev *accel_dev); 111 int adf_enable_vf2pf_comms(struct adf_accel_dev *accel_dev); 121 int adf_devmgr_add_dev(struct adf_accel_dev *accel_dev, 123 void adf_devmgr_rm_dev(struct adf_accel_dev *accel_dev, [all …]
|
D | adf_accel_engine.c | 54 int adf_ae_fw_load(struct adf_accel_dev *accel_dev) in adf_ae_fw_load() argument 56 struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; in adf_ae_fw_load() 57 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_ae_fw_load() 65 &accel_dev->accel_pci_dev.pci_dev->dev)) { in adf_ae_fw_load() 66 dev_err(&GET_DEV(accel_dev), "Failed to load MMP firmware %s\n", in adf_ae_fw_load() 71 &accel_dev->accel_pci_dev.pci_dev->dev)) { in adf_ae_fw_load() 72 dev_err(&GET_DEV(accel_dev), "Failed to load UOF firmware %s\n", in adf_ae_fw_load() 82 dev_err(&GET_DEV(accel_dev), "Failed to load MMP\n"); in adf_ae_fw_load() 86 dev_err(&GET_DEV(accel_dev), "Failed to map FW\n"); in adf_ae_fw_load() 90 dev_err(&GET_DEV(accel_dev), "Failed to load UOF\n"); in adf_ae_fw_load() [all …]
|
D | adf_pf2vf_msg.c | 59 void adf_enable_pf2vf_interrupts(struct adf_accel_dev *accel_dev) in adf_enable_pf2vf_interrupts() argument 61 struct adf_accel_pci *pci_info = &accel_dev->accel_pci_dev; in adf_enable_pf2vf_interrupts() 62 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_pf2vf_interrupts() 69 void adf_disable_pf2vf_interrupts(struct adf_accel_dev *accel_dev) in adf_disable_pf2vf_interrupts() argument 71 struct adf_accel_pci *pci_info = &accel_dev->accel_pci_dev; in adf_disable_pf2vf_interrupts() 72 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_disable_pf2vf_interrupts() 79 void adf_enable_vf2pf_interrupts(struct adf_accel_dev *accel_dev, in adf_enable_vf2pf_interrupts() argument 82 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_vf2pf_interrupts() 84 &GET_BARS(accel_dev)[hw_data->get_misc_bar_id(hw_data)]; in adf_enable_vf2pf_interrupts() 103 void adf_disable_vf2pf_interrupts(struct adf_accel_dev *accel_dev, u32 vf_mask) in adf_disable_vf2pf_interrupts() argument [all …]
|
D | adf_aer.c | 61 struct adf_accel_dev *accel_dev = adf_devmgr_pci_to_accel_dev(pdev); in adf_error_detected() local 64 if (!accel_dev) { in adf_error_detected() 80 struct adf_accel_dev *accel_dev; member 85 void adf_reset_sbr(struct adf_accel_dev *accel_dev) in adf_reset_sbr() argument 87 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_reset_sbr() 95 dev_info(&GET_DEV(accel_dev), in adf_reset_sbr() 98 dev_info(&GET_DEV(accel_dev), "Secondary bus reset\n"); in adf_reset_sbr() 110 void adf_reset_flr(struct adf_accel_dev *accel_dev) in adf_reset_flr() argument 112 pcie_flr(accel_to_pci_dev(accel_dev)); in adf_reset_flr() 116 void adf_dev_restore(struct adf_accel_dev *accel_dev) in adf_dev_restore() argument [all …]
|
D | adf_sriov.c | 110 static int adf_enable_sriov(struct adf_accel_dev *accel_dev) in adf_enable_sriov() argument 112 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_enable_sriov() 114 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_sriov() 116 &GET_BARS(accel_dev)[hw_data->get_misc_bar_id(hw_data)]; in adf_enable_sriov() 122 for (i = 0, vf_info = accel_dev->pf.vf_info; i < totalvfs; in adf_enable_sriov() 125 vf_info->accel_dev = accel_dev; in adf_enable_sriov() 152 adf_enable_vf2pf_interrupts(accel_dev, GENMASK_ULL(totalvfs - 1, 0)); in adf_enable_sriov() 171 void adf_disable_sriov(struct adf_accel_dev *accel_dev) in adf_disable_sriov() argument 173 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_disable_sriov() 175 &GET_BARS(accel_dev)[hw_data->get_misc_bar_id(hw_data)]; in adf_disable_sriov() [all …]
|
D | qat_crypto.c | 65 adf_dev_put(inst->accel_dev); in qat_crypto_put_instance() 68 static int qat_crypto_free_instances(struct adf_accel_dev *accel_dev) in qat_crypto_free_instances() argument 73 list_for_each_entry_safe(inst, tmp, &accel_dev->crypto_list, list) { in qat_crypto_free_instances() 97 struct adf_accel_dev *accel_dev = NULL, *tmp_dev; in qat_crypto_get_instance_node() local 110 accel_dev = tmp_dev; in qat_crypto_get_instance_node() 116 if (!accel_dev) { in qat_crypto_get_instance_node() 122 accel_dev = tmp_dev; in qat_crypto_get_instance_node() 128 if (!accel_dev) in qat_crypto_get_instance_node() 132 list_for_each_entry(tmp_inst, &accel_dev->crypto_list, list) { in qat_crypto_get_instance_node() 142 if (adf_dev_get(accel_dev)) { in qat_crypto_get_instance_node() [all …]
|
D | adf_ctl_drv.c | 151 static int adf_add_key_value_data(struct adf_accel_dev *accel_dev, in adf_add_key_value_data() argument 159 if (adf_cfg_add_key_value_param(accel_dev, section, in adf_add_key_value_data() 162 dev_err(&GET_DEV(accel_dev), in adf_add_key_value_data() 167 if (adf_cfg_add_key_value_param(accel_dev, section, in adf_add_key_value_data() 170 dev_err(&GET_DEV(accel_dev), in adf_add_key_value_data() 178 static int adf_copy_key_value_data(struct adf_accel_dev *accel_dev, in adf_copy_key_value_data() argument 190 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data() 195 if (adf_cfg_section_add(accel_dev, section.name)) { in adf_copy_key_value_data() 196 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data() 206 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data() [all …]
|
D | adf_dev_mgr.c | 183 int adf_devmgr_add_dev(struct adf_accel_dev *accel_dev, in adf_devmgr_add_dev() argument 190 dev_err(&GET_DEV(accel_dev), "Only support up to %d devices\n", in adf_devmgr_add_dev() 196 atomic_set(&accel_dev->ref_count, 0); in adf_devmgr_add_dev() 199 if (!accel_dev->is_vf || (accel_dev->is_vf && !pf)) { in adf_devmgr_add_dev() 206 if (ptr == accel_dev) { in adf_devmgr_add_dev() 212 list_add_tail(&accel_dev->list, &accel_table); in adf_devmgr_add_dev() 213 accel_dev->accel_id = adf_find_free_id(); in adf_devmgr_add_dev() 214 if (accel_dev->accel_id > ADF_MAX_DEVICES) { in adf_devmgr_add_dev() 225 map->id = accel_dev->accel_id; in adf_devmgr_add_dev() 229 } else if (accel_dev->is_vf && pf) { in adf_devmgr_add_dev() [all …]
|
D | adf_admin.c | 154 static int adf_put_admin_msg_sync(struct adf_accel_dev *accel_dev, u32 ae, in adf_put_admin_msg_sync() argument 157 struct adf_admin_comms *admin = accel_dev->admin; in adf_put_admin_msg_sync() 184 dev_err(&GET_DEV(accel_dev), in adf_put_admin_msg_sync() 191 static int adf_send_admin_cmd(struct adf_accel_dev *accel_dev, int cmd) in adf_send_admin_cmd() argument 193 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_send_admin_cmd() 203 req.init_cfg_ptr = accel_dev->admin->const_tbl_addr; in adf_send_admin_cmd() 207 if (adf_put_admin_msg_sync(accel_dev, i, &req, &resp) || in adf_send_admin_cmd() 222 int adf_send_admin_init(struct adf_accel_dev *accel_dev) in adf_send_admin_init() argument 224 int ret = adf_send_admin_cmd(accel_dev, ICP_QAT_FW_INIT_ME); in adf_send_admin_init() 228 return adf_send_admin_cmd(accel_dev, ICP_QAT_FW_CONSTANTS_CFG); in adf_send_admin_init() [all …]
|
D | adf_accel_devices.h | 167 int (*alloc_irq)(struct adf_accel_dev *accel_dev); 168 void (*free_irq)(struct adf_accel_dev *accel_dev); 169 void (*enable_error_correction)(struct adf_accel_dev *accel_dev); 170 int (*init_admin_comms)(struct adf_accel_dev *accel_dev); 171 void (*exit_admin_comms)(struct adf_accel_dev *accel_dev); 172 int (*send_admin_init)(struct adf_accel_dev *accel_dev); 173 int (*init_arb)(struct adf_accel_dev *accel_dev); 174 void (*exit_arb)(struct adf_accel_dev *accel_dev); 175 void (*get_arb_mapping)(struct adf_accel_dev *accel_dev, 177 void (*disable_iov)(struct adf_accel_dev *accel_dev); [all …]
|
D | adf_transport.c | 181 struct adf_accel_dev *accel_dev = bank->accel_dev; in adf_init_ring() local 182 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_ring() 188 ring->base_addr = dma_alloc_coherent(&GET_DEV(accel_dev), in adf_init_ring() 197 dev_err(&GET_DEV(accel_dev), "Ring address not aligned\n"); in adf_init_ring() 198 dma_free_coherent(&GET_DEV(accel_dev), ring_size_bytes, in adf_init_ring() 224 dma_free_coherent(&GET_DEV(ring->bank->accel_dev), in adf_cleanup_ring() 230 int adf_create_ring(struct adf_accel_dev *accel_dev, const char *section, in adf_create_ring() argument 236 struct adf_etr_data *transport_data = accel_dev->transport; in adf_create_ring() 243 if (bank_num >= GET_MAX_BANKS(accel_dev)) { in adf_create_ring() 244 dev_err(&GET_DEV(accel_dev), "Invalid bank number\n"); in adf_create_ring() [all …]
|
D | adf_cfg.c | 128 int adf_cfg_dev_add(struct adf_accel_dev *accel_dev) in adf_cfg_dev_add() argument 137 accel_dev->cfg = dev_cfg_data; in adf_cfg_dev_add() 141 accel_dev->debugfs_dir, in adf_cfg_dev_add() 145 dev_err(&GET_DEV(accel_dev), in adf_cfg_dev_add() 148 accel_dev->cfg = NULL; in adf_cfg_dev_add() 157 void adf_cfg_del_all(struct adf_accel_dev *accel_dev) in adf_cfg_del_all() argument 159 struct adf_cfg_device_data *dev_cfg_data = accel_dev->cfg; in adf_cfg_del_all() 164 clear_bit(ADF_STATUS_CONFIGURED, &accel_dev->status); in adf_cfg_del_all() 177 void adf_cfg_dev_remove(struct adf_accel_dev *accel_dev) in adf_cfg_dev_remove() argument 179 struct adf_cfg_device_data *dev_cfg_data = accel_dev->cfg; in adf_cfg_dev_remove() [all …]
|
D | adf_vf2pf_msg.c | 59 int adf_vf2pf_init(struct adf_accel_dev *accel_dev) in adf_vf2pf_init() argument 64 if (adf_iov_putmsg(accel_dev, msg, 0)) { in adf_vf2pf_init() 65 dev_err(&GET_DEV(accel_dev), in adf_vf2pf_init() 69 set_bit(ADF_STATUS_PF_RUNNING, &accel_dev->status); in adf_vf2pf_init() 82 void adf_vf2pf_shutdown(struct adf_accel_dev *accel_dev) in adf_vf2pf_shutdown() argument 87 if (test_bit(ADF_STATUS_PF_RUNNING, &accel_dev->status)) in adf_vf2pf_shutdown() 88 if (adf_iov_putmsg(accel_dev, msg, 0)) in adf_vf2pf_shutdown() 89 dev_err(&GET_DEV(accel_dev), in adf_vf2pf_shutdown()
|
D | adf_hw_arbiter.c | 79 int adf_init_arb(struct adf_accel_dev *accel_dev) in adf_init_arb() argument 81 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_arb() 82 void __iomem *csr = accel_dev->transport->banks[0].csr_addr; in adf_init_arb() 97 hw_data->get_arb_mapping(accel_dev, &thd_2_arb_cfg); in adf_init_arb() 116 void adf_exit_arb(struct adf_accel_dev *accel_dev) in adf_exit_arb() argument 118 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_exit_arb() 122 if (!accel_dev->transport) in adf_exit_arb() 125 csr = accel_dev->transport->banks[0].csr_addr; in adf_exit_arb() 140 for (i = 0; i < GET_MAX_BANKS(accel_dev); i++) in adf_exit_arb()
|
D | adf_cfg.h | 76 int adf_cfg_dev_add(struct adf_accel_dev *accel_dev); 77 void adf_cfg_dev_remove(struct adf_accel_dev *accel_dev); 78 int adf_cfg_section_add(struct adf_accel_dev *accel_dev, const char *name); 79 void adf_cfg_del_all(struct adf_accel_dev *accel_dev); 80 int adf_cfg_add_key_value_param(struct adf_accel_dev *accel_dev, 84 int adf_cfg_get_param_value(struct adf_accel_dev *accel_dev,
|
/drivers/crypto/qat/qat_dh895xccvf/ |
D | adf_drv.c | 84 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 86 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 87 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 90 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 92 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 103 if (accel_dev->hw_device) { in adf_cleanup_accel() 106 adf_clean_hw_data_dh895xcciov(accel_dev->hw_device); in adf_cleanup_accel() 111 kfree(accel_dev->hw_device); in adf_cleanup_accel() 112 accel_dev->hw_device = NULL; in adf_cleanup_accel() 114 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
/drivers/crypto/qat/qat_c62xvf/ |
D | adf_drv.c | 84 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 86 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 87 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 90 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 92 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 103 if (accel_dev->hw_device) { in adf_cleanup_accel() 106 adf_clean_hw_data_c62xiov(accel_dev->hw_device); in adf_cleanup_accel() 111 kfree(accel_dev->hw_device); in adf_cleanup_accel() 112 accel_dev->hw_device = NULL; in adf_cleanup_accel() 114 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
/drivers/crypto/qat/qat_c3xxxvf/ |
D | adf_drv.c | 84 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 86 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 87 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 90 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 92 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 103 if (accel_dev->hw_device) { in adf_cleanup_accel() 106 adf_clean_hw_data_c3xxxiov(accel_dev->hw_device); in adf_cleanup_accel() 111 kfree(accel_dev->hw_device); in adf_cleanup_accel() 112 accel_dev->hw_device = NULL; in adf_cleanup_accel() 114 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
/drivers/crypto/qat/qat_c3xxx/ |
D | adf_drv.c | 85 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 87 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 88 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 91 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 93 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 103 if (accel_dev->hw_device) { in adf_cleanup_accel() 106 adf_clean_hw_data_c3xxx(accel_dev->hw_device); in adf_cleanup_accel() 111 kfree(accel_dev->hw_device); in adf_cleanup_accel() 112 accel_dev->hw_device = NULL; in adf_cleanup_accel() 114 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
/drivers/crypto/qat/qat_dh895xcc/ |
D | adf_drv.c | 85 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 87 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 88 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 91 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 93 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 103 if (accel_dev->hw_device) { in adf_cleanup_accel() 106 adf_clean_hw_data_dh895xcc(accel_dev->hw_device); in adf_cleanup_accel() 111 kfree(accel_dev->hw_device); in adf_cleanup_accel() 112 accel_dev->hw_device = NULL; in adf_cleanup_accel() 114 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
D | adf_dh895xcc_hw_data.c | 145 static void adf_get_arbiter_mapping(struct adf_accel_dev *accel_dev, in adf_get_arbiter_mapping() argument 148 switch (accel_dev->accel_pci_dev.sku) { in adf_get_arbiter_mapping() 158 dev_err(&GET_DEV(accel_dev), in adf_get_arbiter_mapping() 174 static void adf_enable_error_correction(struct adf_accel_dev *accel_dev) in adf_enable_error_correction() argument 176 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_enable_error_correction() 177 struct adf_bar *misc_bar = &GET_BARS(accel_dev)[ADF_DH895XCC_PMISC_BAR]; in adf_enable_error_correction() 202 static void adf_enable_ints(struct adf_accel_dev *accel_dev) in adf_enable_ints() argument 206 addr = (&GET_BARS(accel_dev)[ADF_DH895XCC_PMISC_BAR])->virt_addr; in adf_enable_ints() 210 accel_dev->pf.vf_info ? 0 : in adf_enable_ints() 211 GENMASK_ULL(GET_MAX_BANKS(accel_dev) - 1, 0)); in adf_enable_ints() [all …]
|
/drivers/crypto/qat/qat_c62x/ |
D | adf_drv.c | 85 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 87 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 88 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 91 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 93 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 103 if (accel_dev->hw_device) { in adf_cleanup_accel() 106 adf_clean_hw_data_c62x(accel_dev->hw_device); in adf_cleanup_accel() 111 kfree(accel_dev->hw_device); in adf_cleanup_accel() 112 accel_dev->hw_device = NULL; in adf_cleanup_accel() 114 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|