Home
last modified time | relevance | path

Searched refs:accel_dev (Results 1 – 25 of 37) sorted by relevance

12

/drivers/crypto/qat/qat_common/
Dadf_init.c59 int adf_dev_init(struct adf_accel_dev *accel_dev) in adf_dev_init() argument
63 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_dev_init()
67 dev_err(&GET_DEV(accel_dev), in adf_dev_init()
72 if (!test_bit(ADF_STATUS_CONFIGURED, &accel_dev->status)) { in adf_dev_init()
73 dev_err(&GET_DEV(accel_dev), "Device not configured\n"); in adf_dev_init()
77 if (adf_init_etr_data(accel_dev)) { in adf_dev_init()
78 dev_err(&GET_DEV(accel_dev), "Failed initialize etr\n"); in adf_dev_init()
82 if (hw_data->init_admin_comms && hw_data->init_admin_comms(accel_dev)) { in adf_dev_init()
83 dev_err(&GET_DEV(accel_dev), "Failed initialize admin comms\n"); in adf_dev_init()
87 if (hw_data->init_arb && hw_data->init_arb(accel_dev)) { in adf_dev_init()
[all …]
Dadf_vf_isr.c27 struct adf_accel_dev *accel_dev; member
31 static int adf_enable_msi(struct adf_accel_dev *accel_dev) in adf_enable_msi() argument
33 struct adf_accel_pci *pci_dev_info = &accel_dev->accel_pci_dev; in adf_enable_msi()
37 dev_err(&GET_DEV(accel_dev), in adf_enable_msi()
42 accel_dev->vf.irq_name = kzalloc(ADF_MAX_MSIX_VECTOR_NAME, GFP_KERNEL); in adf_enable_msi()
43 if (!accel_dev->vf.irq_name) in adf_enable_msi()
49 static void adf_disable_msi(struct adf_accel_dev *accel_dev) in adf_disable_msi() argument
51 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_disable_msi()
53 kfree(accel_dev->vf.irq_name); in adf_disable_msi()
61 struct adf_accel_dev *accel_dev = stop_data->accel_dev; in adf_dev_stop_async() local
[all …]
Dadf_common_drv.h44 int (*event_hld)(struct adf_accel_dev *accel_dev,
60 int adf_dev_init(struct adf_accel_dev *accel_dev);
61 int adf_dev_start(struct adf_accel_dev *accel_dev);
62 void adf_dev_stop(struct adf_accel_dev *accel_dev);
63 void adf_dev_shutdown(struct adf_accel_dev *accel_dev);
65 int adf_iov_putmsg(struct adf_accel_dev *accel_dev, u32 msg, u8 vf_nr);
66 void adf_pf2vf_notify_restarting(struct adf_accel_dev *accel_dev);
67 int adf_enable_vf2pf_comms(struct adf_accel_dev *accel_dev);
77 int adf_devmgr_add_dev(struct adf_accel_dev *accel_dev,
79 void adf_devmgr_rm_dev(struct adf_accel_dev *accel_dev,
[all …]
Dadf_isr.c20 static int adf_enable_msix(struct adf_accel_dev *accel_dev) in adf_enable_msix() argument
22 struct adf_accel_pci *pci_dev_info = &accel_dev->accel_pci_dev; in adf_enable_msix()
23 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_msix()
27 if (!accel_dev->pf.vf_info) { in adf_enable_msix()
41 dev_err(&GET_DEV(accel_dev), "Failed to enable MSI-X IRQ(s)\n"); in adf_enable_msix()
63 struct adf_accel_dev *accel_dev = dev_ptr; in adf_msix_isr_ae() local
67 if (accel_dev->pf.vf_info) { in adf_msix_isr_ae()
68 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_msix_isr_ae()
70 &GET_BARS(accel_dev)[hw_data->get_misc_bar_id(hw_data)]; in adf_msix_isr_ae()
86 adf_disable_vf2pf_interrupts(accel_dev, vf_mask); in adf_msix_isr_ae()
[all …]
Dadf_accel_engine.c10 int adf_ae_fw_load(struct adf_accel_dev *accel_dev) in adf_ae_fw_load() argument
12 struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; in adf_ae_fw_load()
13 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_ae_fw_load()
21 &accel_dev->accel_pci_dev.pci_dev->dev)) { in adf_ae_fw_load()
22 dev_err(&GET_DEV(accel_dev), "Failed to load MMP firmware %s\n", in adf_ae_fw_load()
27 &accel_dev->accel_pci_dev.pci_dev->dev)) { in adf_ae_fw_load()
28 dev_err(&GET_DEV(accel_dev), "Failed to load UOF firmware %s\n", in adf_ae_fw_load()
38 dev_err(&GET_DEV(accel_dev), "Failed to load MMP\n"); in adf_ae_fw_load()
42 dev_err(&GET_DEV(accel_dev), "Failed to map FW\n"); in adf_ae_fw_load()
46 dev_err(&GET_DEV(accel_dev), "Failed to load UOF\n"); in adf_ae_fw_load()
[all …]
Dadf_pf2vf_msg.c14 void adf_enable_pf2vf_interrupts(struct adf_accel_dev *accel_dev) in adf_enable_pf2vf_interrupts() argument
16 struct adf_accel_pci *pci_info = &accel_dev->accel_pci_dev; in adf_enable_pf2vf_interrupts()
17 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_pf2vf_interrupts()
24 void adf_disable_pf2vf_interrupts(struct adf_accel_dev *accel_dev) in adf_disable_pf2vf_interrupts() argument
26 struct adf_accel_pci *pci_info = &accel_dev->accel_pci_dev; in adf_disable_pf2vf_interrupts()
27 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_disable_pf2vf_interrupts()
34 void adf_enable_vf2pf_interrupts(struct adf_accel_dev *accel_dev, in adf_enable_vf2pf_interrupts() argument
37 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_vf2pf_interrupts()
39 &GET_BARS(accel_dev)[hw_data->get_misc_bar_id(hw_data)]; in adf_enable_vf2pf_interrupts()
58 void adf_disable_vf2pf_interrupts(struct adf_accel_dev *accel_dev, u32 vf_mask) in adf_disable_vf2pf_interrupts() argument
[all …]
Dadf_aer.c17 struct adf_accel_dev *accel_dev = adf_devmgr_pci_to_accel_dev(pdev); in adf_error_detected() local
20 if (!accel_dev) { in adf_error_detected()
36 struct adf_accel_dev *accel_dev; member
41 void adf_reset_sbr(struct adf_accel_dev *accel_dev) in adf_reset_sbr() argument
43 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_reset_sbr()
51 dev_info(&GET_DEV(accel_dev), in adf_reset_sbr()
54 dev_info(&GET_DEV(accel_dev), "Secondary bus reset\n"); in adf_reset_sbr()
66 void adf_reset_flr(struct adf_accel_dev *accel_dev) in adf_reset_flr() argument
68 pcie_flr(accel_to_pci_dev(accel_dev)); in adf_reset_flr()
72 void adf_dev_restore(struct adf_accel_dev *accel_dev) in adf_dev_restore() argument
[all …]
Dadf_sriov.c66 static int adf_enable_sriov(struct adf_accel_dev *accel_dev) in adf_enable_sriov() argument
68 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_enable_sriov()
70 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_sriov()
72 &GET_BARS(accel_dev)[hw_data->get_misc_bar_id(hw_data)]; in adf_enable_sriov()
78 for (i = 0, vf_info = accel_dev->pf.vf_info; i < totalvfs; in adf_enable_sriov()
81 vf_info->accel_dev = accel_dev; in adf_enable_sriov()
108 adf_enable_vf2pf_interrupts(accel_dev, GENMASK_ULL(totalvfs - 1, 0)); in adf_enable_sriov()
127 void adf_disable_sriov(struct adf_accel_dev *accel_dev) in adf_disable_sriov() argument
129 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_disable_sriov()
131 &GET_BARS(accel_dev)[hw_data->get_misc_bar_id(hw_data)]; in adf_disable_sriov()
[all …]
Dqat_crypto.c21 adf_dev_put(inst->accel_dev); in qat_crypto_put_instance()
24 static int qat_crypto_free_instances(struct adf_accel_dev *accel_dev) in qat_crypto_free_instances() argument
29 list_for_each_entry_safe(inst, tmp, &accel_dev->crypto_list, list) { in qat_crypto_free_instances()
53 struct adf_accel_dev *accel_dev = NULL, *tmp_dev; in qat_crypto_get_instance_node() local
66 accel_dev = tmp_dev; in qat_crypto_get_instance_node()
72 if (!accel_dev) { in qat_crypto_get_instance_node()
78 accel_dev = tmp_dev; in qat_crypto_get_instance_node()
84 if (!accel_dev) in qat_crypto_get_instance_node()
88 list_for_each_entry(tmp_inst, &accel_dev->crypto_list, list) { in qat_crypto_get_instance_node()
98 if (adf_dev_get(accel_dev)) { in qat_crypto_get_instance_node()
[all …]
Dadf_ctl_drv.c107 static int adf_add_key_value_data(struct adf_accel_dev *accel_dev, in adf_add_key_value_data() argument
115 if (adf_cfg_add_key_value_param(accel_dev, section, in adf_add_key_value_data()
118 dev_err(&GET_DEV(accel_dev), in adf_add_key_value_data()
123 if (adf_cfg_add_key_value_param(accel_dev, section, in adf_add_key_value_data()
126 dev_err(&GET_DEV(accel_dev), in adf_add_key_value_data()
134 static int adf_copy_key_value_data(struct adf_accel_dev *accel_dev, in adf_copy_key_value_data() argument
146 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data()
151 if (adf_cfg_section_add(accel_dev, section.name)) { in adf_copy_key_value_data()
152 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data()
162 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data()
[all …]
Dadf_admin.c113 static int adf_put_admin_msg_sync(struct adf_accel_dev *accel_dev, u32 ae, in adf_put_admin_msg_sync() argument
118 struct adf_admin_comms *admin = accel_dev->admin; in adf_put_admin_msg_sync()
140 dev_err(&GET_DEV(accel_dev), in adf_put_admin_msg_sync()
155 static int adf_send_admin(struct adf_accel_dev *accel_dev, in adf_send_admin() argument
163 if (adf_put_admin_msg_sync(accel_dev, ae, req, resp) || in adf_send_admin()
170 static int adf_init_me(struct adf_accel_dev *accel_dev) in adf_init_me() argument
174 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_init_me()
181 return adf_send_admin(accel_dev, &req, &resp, ae_mask); in adf_init_me()
184 static int adf_set_fw_constants(struct adf_accel_dev *accel_dev) in adf_set_fw_constants() argument
188 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_set_fw_constants()
[all …]
Dadf_dev_mgr.c139 int adf_devmgr_add_dev(struct adf_accel_dev *accel_dev, in adf_devmgr_add_dev() argument
146 dev_err(&GET_DEV(accel_dev), "Only support up to %d devices\n", in adf_devmgr_add_dev()
152 atomic_set(&accel_dev->ref_count, 0); in adf_devmgr_add_dev()
155 if (!accel_dev->is_vf || (accel_dev->is_vf && !pf)) { in adf_devmgr_add_dev()
162 if (ptr == accel_dev) { in adf_devmgr_add_dev()
168 list_add_tail(&accel_dev->list, &accel_table); in adf_devmgr_add_dev()
169 accel_dev->accel_id = adf_find_free_id(); in adf_devmgr_add_dev()
170 if (accel_dev->accel_id > ADF_MAX_DEVICES) { in adf_devmgr_add_dev()
181 map->id = accel_dev->accel_id; in adf_devmgr_add_dev()
185 } else if (accel_dev->is_vf && pf) { in adf_devmgr_add_dev()
[all …]
Dadf_accel_devices.h117 int (*alloc_irq)(struct adf_accel_dev *accel_dev);
118 void (*free_irq)(struct adf_accel_dev *accel_dev);
119 void (*enable_error_correction)(struct adf_accel_dev *accel_dev);
120 int (*init_admin_comms)(struct adf_accel_dev *accel_dev);
121 void (*exit_admin_comms)(struct adf_accel_dev *accel_dev);
122 int (*send_admin_init)(struct adf_accel_dev *accel_dev);
123 int (*init_arb)(struct adf_accel_dev *accel_dev);
124 void (*exit_arb)(struct adf_accel_dev *accel_dev);
125 void (*get_arb_mapping)(struct adf_accel_dev *accel_dev,
127 void (*disable_iov)(struct adf_accel_dev *accel_dev);
[all …]
Dadf_transport.c137 struct adf_accel_dev *accel_dev = bank->accel_dev; in adf_init_ring() local
138 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_ring()
144 ring->base_addr = dma_alloc_coherent(&GET_DEV(accel_dev), in adf_init_ring()
153 dev_err(&GET_DEV(accel_dev), "Ring address not aligned\n"); in adf_init_ring()
154 dma_free_coherent(&GET_DEV(accel_dev), ring_size_bytes, in adf_init_ring()
181 dma_free_coherent(&GET_DEV(ring->bank->accel_dev), in adf_cleanup_ring()
187 int adf_create_ring(struct adf_accel_dev *accel_dev, const char *section, in adf_create_ring() argument
193 struct adf_etr_data *transport_data = accel_dev->transport; in adf_create_ring()
200 if (bank_num >= GET_MAX_BANKS(accel_dev)) { in adf_create_ring()
201 dev_err(&GET_DEV(accel_dev), "Invalid bank number\n"); in adf_create_ring()
[all …]
Dadf_cfg.c67 int adf_cfg_dev_add(struct adf_accel_dev *accel_dev) in adf_cfg_dev_add() argument
76 accel_dev->cfg = dev_cfg_data; in adf_cfg_dev_add()
80 accel_dev->debugfs_dir, in adf_cfg_dev_add()
89 void adf_cfg_del_all(struct adf_accel_dev *accel_dev) in adf_cfg_del_all() argument
91 struct adf_cfg_device_data *dev_cfg_data = accel_dev->cfg; in adf_cfg_del_all()
96 clear_bit(ADF_STATUS_CONFIGURED, &accel_dev->status); in adf_cfg_del_all()
109 void adf_cfg_dev_remove(struct adf_accel_dev *accel_dev) in adf_cfg_dev_remove() argument
111 struct adf_cfg_device_data *dev_cfg_data = accel_dev->cfg; in adf_cfg_dev_remove()
121 accel_dev->cfg = NULL; in adf_cfg_dev_remove()
170 static struct adf_cfg_section *adf_cfg_sec_find(struct adf_accel_dev *accel_dev, in adf_cfg_sec_find() argument
[all …]
Dadf_vf2pf_msg.c15 int adf_vf2pf_notify_init(struct adf_accel_dev *accel_dev) in adf_vf2pf_notify_init() argument
20 if (adf_iov_putmsg(accel_dev, msg, 0)) { in adf_vf2pf_notify_init()
21 dev_err(&GET_DEV(accel_dev), in adf_vf2pf_notify_init()
25 set_bit(ADF_STATUS_PF_RUNNING, &accel_dev->status); in adf_vf2pf_notify_init()
38 void adf_vf2pf_notify_shutdown(struct adf_accel_dev *accel_dev) in adf_vf2pf_notify_shutdown() argument
43 if (test_bit(ADF_STATUS_PF_RUNNING, &accel_dev->status)) in adf_vf2pf_notify_shutdown()
44 if (adf_iov_putmsg(accel_dev, msg, 0)) in adf_vf2pf_notify_shutdown()
45 dev_err(&GET_DEV(accel_dev), in adf_vf2pf_notify_shutdown()
Dadf_hw_arbiter.c35 int adf_init_arb(struct adf_accel_dev *accel_dev) in adf_init_arb() argument
37 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_arb()
38 void __iomem *csr = accel_dev->transport->banks[0].csr_addr; in adf_init_arb()
53 hw_data->get_arb_mapping(accel_dev, &thd_2_arb_cfg); in adf_init_arb()
72 void adf_exit_arb(struct adf_accel_dev *accel_dev) in adf_exit_arb() argument
74 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_exit_arb()
78 if (!accel_dev->transport) in adf_exit_arb()
81 csr = accel_dev->transport->banks[0].csr_addr; in adf_exit_arb()
96 for (i = 0; i < GET_MAX_BANKS(accel_dev); i++) in adf_exit_arb()
Dadf_cfg.h32 int adf_cfg_dev_add(struct adf_accel_dev *accel_dev);
33 void adf_cfg_dev_remove(struct adf_accel_dev *accel_dev);
34 int adf_cfg_section_add(struct adf_accel_dev *accel_dev, const char *name);
35 void adf_cfg_del_all(struct adf_accel_dev *accel_dev);
36 int adf_cfg_add_key_value_param(struct adf_accel_dev *accel_dev,
40 int adf_cfg_get_param_value(struct adf_accel_dev *accel_dev,
/drivers/crypto/qat/qat_c62xvf/
Dadf_drv.c37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_c62xiov(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/drivers/crypto/qat/qat_dh895xccvf/
Dadf_drv.c37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_dh895xcciov(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/drivers/crypto/qat/qat_c3xxxvf/
Dadf_drv.c37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_c3xxxiov(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/drivers/crypto/qat/qat_dh895xcc/
Dadf_drv.c38 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
40 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
41 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
44 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
46 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_dh895xcc(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
Dadf_dh895xcc_hw_data.c105 static void adf_get_arbiter_mapping(struct adf_accel_dev *accel_dev, in adf_get_arbiter_mapping() argument
108 switch (accel_dev->accel_pci_dev.sku) { in adf_get_arbiter_mapping()
118 dev_err(&GET_DEV(accel_dev), in adf_get_arbiter_mapping()
134 static void adf_enable_error_correction(struct adf_accel_dev *accel_dev) in adf_enable_error_correction() argument
136 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_enable_error_correction()
137 struct adf_bar *misc_bar = &GET_BARS(accel_dev)[ADF_DH895XCC_PMISC_BAR]; in adf_enable_error_correction()
144 for_each_set_bit(i, &ae_mask, GET_MAX_ACCELENGINES(accel_dev)) { in adf_enable_error_correction()
164 static void adf_enable_ints(struct adf_accel_dev *accel_dev) in adf_enable_ints() argument
168 addr = (&GET_BARS(accel_dev)[ADF_DH895XCC_PMISC_BAR])->virt_addr; in adf_enable_ints()
172 accel_dev->pf.vf_info ? 0 : in adf_enable_ints()
[all …]
/drivers/crypto/qat/qat_c3xxx/
Dadf_drv.c38 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
40 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
41 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
44 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
46 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_c3xxx(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/drivers/crypto/qat/qat_c62x/
Dadf_drv.c38 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
40 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
41 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
44 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
46 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_c62x(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]

12