Home
last modified time | relevance | path

Searched refs:accel_dev (Results 1 – 25 of 57) sorted by relevance

123

/Linux-v6.1/drivers/crypto/qat/qat_common/
Dadf_init.c59 int adf_dev_init(struct adf_accel_dev *accel_dev) in adf_dev_init() argument
63 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_dev_init()
67 dev_err(&GET_DEV(accel_dev), in adf_dev_init()
72 if (!test_bit(ADF_STATUS_CONFIGURED, &accel_dev->status) && in adf_dev_init()
73 !accel_dev->is_vf) { in adf_dev_init()
74 dev_err(&GET_DEV(accel_dev), "Device not configured\n"); in adf_dev_init()
78 if (adf_init_etr_data(accel_dev)) { in adf_dev_init()
79 dev_err(&GET_DEV(accel_dev), "Failed initialize etr\n"); in adf_dev_init()
83 if (hw_data->init_device && hw_data->init_device(accel_dev)) { in adf_dev_init()
84 dev_err(&GET_DEV(accel_dev), "Failed to initialize device\n"); in adf_dev_init()
[all …]
Dadf_vf_isr.c27 struct adf_accel_dev *accel_dev; member
31 void adf_enable_pf2vf_interrupts(struct adf_accel_dev *accel_dev) in adf_enable_pf2vf_interrupts() argument
33 void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev); in adf_enable_pf2vf_interrupts()
38 void adf_disable_pf2vf_interrupts(struct adf_accel_dev *accel_dev) in adf_disable_pf2vf_interrupts() argument
40 void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev); in adf_disable_pf2vf_interrupts()
46 static int adf_enable_msi(struct adf_accel_dev *accel_dev) in adf_enable_msi() argument
48 struct adf_accel_pci *pci_dev_info = &accel_dev->accel_pci_dev; in adf_enable_msi()
52 dev_err(&GET_DEV(accel_dev), in adf_enable_msi()
60 static void adf_disable_msi(struct adf_accel_dev *accel_dev) in adf_disable_msi() argument
62 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_disable_msi()
[all …]
Dadf_isr.c21 static int adf_enable_msix(struct adf_accel_dev *accel_dev) in adf_enable_msix() argument
23 struct adf_accel_pci *pci_dev_info = &accel_dev->accel_pci_dev; in adf_enable_msix()
24 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_msix()
29 hw_data->set_msix_rttable(accel_dev); in adf_enable_msix()
34 dev_err(&GET_DEV(accel_dev), in adf_enable_msix()
50 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(bank->accel_dev); in adf_msix_isr_bundle()
59 void adf_enable_vf2pf_interrupts(struct adf_accel_dev *accel_dev, u32 vf_mask) in adf_enable_vf2pf_interrupts() argument
61 void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev); in adf_enable_vf2pf_interrupts()
64 spin_lock_irqsave(&accel_dev->pf.vf2pf_ints_lock, flags); in adf_enable_vf2pf_interrupts()
65 GET_PFVF_OPS(accel_dev)->enable_vf2pf_interrupts(pmisc_addr, vf_mask); in adf_enable_vf2pf_interrupts()
[all …]
Dadf_accel_engine.c10 static int adf_ae_fw_load_images(struct adf_accel_dev *accel_dev, void *fw_addr, in adf_ae_fw_load_images() argument
13 struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; in adf_ae_fw_load_images()
14 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_ae_fw_load_images()
25 obj_name = hw_device->uof_get_name(accel_dev, i); in adf_ae_fw_load_images()
26 ae_mask = hw_device->uof_get_ae_mask(accel_dev, i); in adf_ae_fw_load_images()
28 dev_err(&GET_DEV(accel_dev), "Invalid UOF image\n"); in adf_ae_fw_load_images()
33 dev_err(&GET_DEV(accel_dev), in adf_ae_fw_load_images()
38 dev_err(&GET_DEV(accel_dev), in adf_ae_fw_load_images()
43 dev_err(&GET_DEV(accel_dev), in adf_ae_fw_load_images()
53 adf_ae_fw_release(accel_dev); in adf_ae_fw_load_images()
[all …]
Dadf_sriov.c25 struct adf_accel_dev *accel_dev = vf_info->accel_dev; in adf_iov_send_resp() local
29 ret = adf_recv_and_handle_vf2pf_msg(accel_dev, vf_nr); in adf_iov_send_resp()
32 adf_enable_vf2pf_interrupts(accel_dev, 1 << vf_nr); in adf_iov_send_resp()
50 static int adf_enable_sriov(struct adf_accel_dev *accel_dev) in adf_enable_sriov() argument
52 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_enable_sriov()
54 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_sriov()
58 for (i = 0, vf_info = accel_dev->pf.vf_info; i < totalvfs; in adf_enable_sriov()
61 vf_info->accel_dev = accel_dev; in adf_enable_sriov()
73 hw_data->configure_iov_threads(accel_dev, true); in adf_enable_sriov()
76 adf_enable_vf2pf_interrupts(accel_dev, BIT_ULL(totalvfs) - 1); in adf_enable_sriov()
[all …]
Dadf_common_drv.h44 int (*event_hld)(struct adf_accel_dev *accel_dev,
55 int adf_dev_init(struct adf_accel_dev *accel_dev);
56 int adf_dev_start(struct adf_accel_dev *accel_dev);
57 void adf_dev_stop(struct adf_accel_dev *accel_dev);
58 void adf_dev_shutdown(struct adf_accel_dev *accel_dev);
59 int adf_dev_shutdown_cache_cfg(struct adf_accel_dev *accel_dev);
69 int adf_devmgr_add_dev(struct adf_accel_dev *accel_dev,
71 void adf_devmgr_rm_dev(struct adf_accel_dev *accel_dev,
79 int adf_devmgr_in_reset(struct adf_accel_dev *accel_dev);
80 int adf_dev_started(struct adf_accel_dev *accel_dev);
[all …]
Dadf_aer.c17 struct adf_accel_dev *accel_dev = adf_devmgr_pci_to_accel_dev(pdev); in adf_error_detected() local
20 if (!accel_dev) { in adf_error_detected()
36 struct adf_accel_dev *accel_dev; member
41 void adf_reset_sbr(struct adf_accel_dev *accel_dev) in adf_reset_sbr() argument
43 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_reset_sbr()
51 dev_info(&GET_DEV(accel_dev), in adf_reset_sbr()
54 dev_info(&GET_DEV(accel_dev), "Secondary bus reset\n"); in adf_reset_sbr()
66 void adf_reset_flr(struct adf_accel_dev *accel_dev) in adf_reset_flr() argument
68 pcie_flr(accel_to_pci_dev(accel_dev)); in adf_reset_flr()
72 void adf_dev_restore(struct adf_accel_dev *accel_dev) in adf_dev_restore() argument
[all …]
Dadf_pfvf_vf_proto.c33 int adf_send_vf2pf_msg(struct adf_accel_dev *accel_dev, struct pfvf_message msg) in adf_send_vf2pf_msg() argument
35 struct adf_pfvf_ops *pfvf_ops = GET_PFVF_OPS(accel_dev); in adf_send_vf2pf_msg()
38 return pfvf_ops->send_msg(accel_dev, msg, pfvf_offset, in adf_send_vf2pf_msg()
39 &accel_dev->vf.vf2pf_lock); in adf_send_vf2pf_msg()
50 static struct pfvf_message adf_recv_pf2vf_msg(struct adf_accel_dev *accel_dev) in adf_recv_pf2vf_msg() argument
52 struct adf_pfvf_ops *pfvf_ops = GET_PFVF_OPS(accel_dev); in adf_recv_pf2vf_msg()
55 return pfvf_ops->recv_msg(accel_dev, pfvf_offset, accel_dev->vf.pf_compat_ver); in adf_recv_pf2vf_msg()
69 int adf_send_vf2pf_req(struct adf_accel_dev *accel_dev, struct pfvf_message msg, in adf_send_vf2pf_req() argument
76 reinit_completion(&accel_dev->vf.msg_received); in adf_send_vf2pf_req()
80 ret = adf_send_vf2pf_msg(accel_dev, msg); in adf_send_vf2pf_req()
[all …]
Dadf_sysfs.c18 struct adf_accel_dev *accel_dev; in state_show() local
21 accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev)); in state_show()
22 if (!accel_dev) in state_show()
25 state = adf_dev_started(accel_dev) ? "up" : "down"; in state_show()
32 struct adf_accel_dev *accel_dev; in state_store() local
36 accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev)); in state_store()
37 if (!accel_dev) in state_store()
40 accel_id = accel_dev->accel_id; in state_store()
42 if (adf_devmgr_in_reset(accel_dev) || adf_dev_in_use(accel_dev)) { in state_store()
53 if (!adf_dev_started(accel_dev)) { in state_store()
[all …]
Dadf_pfvf_vf_msg.c18 int adf_vf2pf_notify_init(struct adf_accel_dev *accel_dev) in adf_vf2pf_notify_init() argument
22 if (adf_send_vf2pf_msg(accel_dev, msg)) { in adf_vf2pf_notify_init()
23 dev_err(&GET_DEV(accel_dev), in adf_vf2pf_notify_init()
27 set_bit(ADF_STATUS_PF_RUNNING, &accel_dev->status); in adf_vf2pf_notify_init()
40 void adf_vf2pf_notify_shutdown(struct adf_accel_dev *accel_dev) in adf_vf2pf_notify_shutdown() argument
44 if (test_bit(ADF_STATUS_PF_RUNNING, &accel_dev->status)) in adf_vf2pf_notify_shutdown()
45 if (adf_send_vf2pf_msg(accel_dev, msg)) in adf_vf2pf_notify_shutdown()
46 dev_err(&GET_DEV(accel_dev), in adf_vf2pf_notify_shutdown()
51 int adf_vf2pf_request_version(struct adf_accel_dev *accel_dev) in adf_vf2pf_request_version() argument
64 ret = adf_send_vf2pf_req(accel_dev, msg, &resp); in adf_vf2pf_request_version()
[all …]
Dqat_crypto.c22 adf_dev_put(inst->accel_dev); in qat_crypto_put_instance()
25 static int qat_crypto_free_instances(struct adf_accel_dev *accel_dev) in qat_crypto_free_instances() argument
30 list_for_each_entry_safe(inst, tmp, &accel_dev->crypto_list, list) { in qat_crypto_free_instances()
54 struct adf_accel_dev *accel_dev = NULL, *tmp_dev; in qat_crypto_get_instance_node() local
67 accel_dev = tmp_dev; in qat_crypto_get_instance_node()
73 if (!accel_dev) { in qat_crypto_get_instance_node()
79 accel_dev = tmp_dev; in qat_crypto_get_instance_node()
85 if (!accel_dev) in qat_crypto_get_instance_node()
89 list_for_each_entry(tmp_inst, &accel_dev->crypto_list, list) { in qat_crypto_get_instance_node()
99 if (adf_dev_get(accel_dev)) { in qat_crypto_get_instance_node()
[all …]
Dadf_admin.c109 static int adf_put_admin_msg_sync(struct adf_accel_dev *accel_dev, u32 ae, in adf_put_admin_msg_sync() argument
114 struct adf_admin_comms *admin = accel_dev->admin; in adf_put_admin_msg_sync()
136 dev_err(&GET_DEV(accel_dev), in adf_put_admin_msg_sync()
151 static int adf_send_admin(struct adf_accel_dev *accel_dev, in adf_send_admin() argument
159 if (adf_put_admin_msg_sync(accel_dev, ae, req, resp) || in adf_send_admin()
166 static int adf_init_ae(struct adf_accel_dev *accel_dev) in adf_init_ae() argument
170 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_init_ae()
177 return adf_send_admin(accel_dev, &req, &resp, ae_mask); in adf_init_ae()
180 static int adf_set_fw_constants(struct adf_accel_dev *accel_dev) in adf_set_fw_constants() argument
184 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_set_fw_constants()
[all …]
Dadf_ctl_drv.c110 static int adf_add_key_value_data(struct adf_accel_dev *accel_dev, in adf_add_key_value_data() argument
118 if (adf_cfg_add_key_value_param(accel_dev, section, in adf_add_key_value_data()
121 dev_err(&GET_DEV(accel_dev), in adf_add_key_value_data()
126 if (adf_cfg_add_key_value_param(accel_dev, section, in adf_add_key_value_data()
129 dev_err(&GET_DEV(accel_dev), in adf_add_key_value_data()
137 static int adf_copy_key_value_data(struct adf_accel_dev *accel_dev, in adf_copy_key_value_data() argument
150 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data()
155 if (adf_cfg_section_add(accel_dev, section.name)) { in adf_copy_key_value_data()
156 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data()
166 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data()
[all …]
Dadf_accel_devices.h154 int (*enable_comms)(struct adf_accel_dev *accel_dev);
160 int (*send_msg)(struct adf_accel_dev *accel_dev, struct pfvf_message msg,
162 struct pfvf_message (*recv_msg)(struct adf_accel_dev *accel_dev,
170 u32 (*get_accel_cap)(struct adf_accel_dev *accel_dev);
179 int (*alloc_irq)(struct adf_accel_dev *accel_dev);
180 void (*free_irq)(struct adf_accel_dev *accel_dev);
181 void (*enable_error_correction)(struct adf_accel_dev *accel_dev);
182 int (*init_admin_comms)(struct adf_accel_dev *accel_dev);
183 void (*exit_admin_comms)(struct adf_accel_dev *accel_dev);
184 int (*send_admin_init)(struct adf_accel_dev *accel_dev);
[all …]
Dadf_transport.c61 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(bank->accel_dev); in adf_enable_ring_irq()
74 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(bank->accel_dev); in adf_disable_ring_irq()
90 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(ring->bank->accel_dev); in adf_send_message()
114 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(ring->bank->accel_dev); in adf_handle_response()
138 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(ring->bank->accel_dev); in adf_configure_tx_ring()
149 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(ring->bank->accel_dev); in adf_configure_rx_ring()
163 struct adf_accel_dev *accel_dev = bank->accel_dev; in adf_init_ring() local
164 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_ring()
165 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(accel_dev); in adf_init_ring()
171 ring->base_addr = dma_alloc_coherent(&GET_DEV(accel_dev), in adf_init_ring()
[all …]
Dadf_pfvf_pf_proto.c32 int adf_send_pf2vf_msg(struct adf_accel_dev *accel_dev, u8 vf_nr, struct pfvf_message msg) in adf_send_pf2vf_msg() argument
34 struct adf_pfvf_ops *pfvf_ops = GET_PFVF_OPS(accel_dev); in adf_send_pf2vf_msg()
37 return pfvf_ops->send_msg(accel_dev, msg, pfvf_offset, in adf_send_pf2vf_msg()
38 &accel_dev->pf.vf_info[vf_nr].pf2vf_lock); in adf_send_pf2vf_msg()
50 static struct pfvf_message adf_recv_vf2pf_msg(struct adf_accel_dev *accel_dev, u8 vf_nr) in adf_recv_vf2pf_msg() argument
52 struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr]; in adf_recv_vf2pf_msg()
53 struct adf_pfvf_ops *pfvf_ops = GET_PFVF_OPS(accel_dev); in adf_recv_vf2pf_msg()
56 return pfvf_ops->recv_msg(accel_dev, pfvf_offset, vf_info->vf_compat_ver); in adf_recv_vf2pf_msg()
85 struct adf_accel_dev *accel_dev = vf_info->accel_dev; in adf_pf2vf_blkmsg_get_data() local
97 if (unlikely((*provider)(accel_dev, blkmsg, vf_info->vf_compat_ver))) { in adf_pf2vf_blkmsg_get_data()
[all …]
Dadf_dev_mgr.c139 int adf_devmgr_add_dev(struct adf_accel_dev *accel_dev, in adf_devmgr_add_dev() argument
146 dev_err(&GET_DEV(accel_dev), "Only support up to %d devices\n", in adf_devmgr_add_dev()
152 atomic_set(&accel_dev->ref_count, 0); in adf_devmgr_add_dev()
155 if (!accel_dev->is_vf || !pf) { in adf_devmgr_add_dev()
162 if (ptr == accel_dev) { in adf_devmgr_add_dev()
168 list_add_tail(&accel_dev->list, &accel_table); in adf_devmgr_add_dev()
169 accel_dev->accel_id = adf_find_free_id(); in adf_devmgr_add_dev()
170 if (accel_dev->accel_id > ADF_MAX_DEVICES) { in adf_devmgr_add_dev()
181 map->id = accel_dev->accel_id; in adf_devmgr_add_dev()
185 } else if (accel_dev->is_vf && pf) { in adf_devmgr_add_dev()
[all …]
Dadf_gen4_pm.c20 struct adf_accel_dev *accel_dev; member
24 static int send_host_msg(struct adf_accel_dev *accel_dev) in send_host_msg() argument
26 void __iomem *pmisc = adf_get_pmisc_base(accel_dev); in send_host_msg()
50 struct adf_accel_dev *accel_dev = pm_data->accel_dev; in pm_bh_handler() local
51 void __iomem *pmisc = adf_get_pmisc_base(accel_dev); in pm_bh_handler()
58 if (send_host_msg(accel_dev)) in pm_bh_handler()
59 dev_warn_ratelimited(&GET_DEV(accel_dev), in pm_bh_handler()
74 bool adf_gen4_handle_pm_interrupt(struct adf_accel_dev *accel_dev) in adf_gen4_handle_pm_interrupt() argument
76 void __iomem *pmisc = adf_get_pmisc_base(accel_dev); in adf_gen4_handle_pm_interrupt()
103 pm_data->accel_dev = accel_dev; in adf_gen4_handle_pm_interrupt()
[all …]
/Linux-v6.1/drivers/crypto/qat/qat_4xxx/
Dadf_drv.c22 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
24 if (accel_dev->hw_device) { in adf_cleanup_accel()
25 adf_clean_hw_data_4xxx(accel_dev->hw_device); in adf_cleanup_accel()
26 accel_dev->hw_device = NULL; in adf_cleanup_accel()
28 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
29 debugfs_remove(accel_dev->debugfs_dir); in adf_cleanup_accel()
30 adf_devmgr_rm_dev(accel_dev, NULL); in adf_cleanup_accel()
33 static int adf_cfg_dev_init(struct adf_accel_dev *accel_dev) in adf_cfg_dev_init() argument
38 config = accel_dev->accel_id % 2 ? ADF_CFG_DC : ADF_CFG_CY; in adf_cfg_dev_init()
40 ret = adf_cfg_section_add(accel_dev, ADF_GENERAL_SEC); in adf_cfg_dev_init()
[all …]
/Linux-v6.1/drivers/crypto/qat/qat_c3xxxvf/
Dadf_drv.c37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_c3xxxiov(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/Linux-v6.1/drivers/crypto/qat/qat_c62xvf/
Dadf_drv.c37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_c62xiov(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/Linux-v6.1/drivers/crypto/qat/qat_dh895xccvf/
Dadf_drv.c37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
56 if (accel_dev->hw_device) { in adf_cleanup_accel()
59 adf_clean_hw_data_dh895xcciov(accel_dev->hw_device); in adf_cleanup_accel()
64 kfree(accel_dev->hw_device); in adf_cleanup_accel()
65 accel_dev->hw_device = NULL; in adf_cleanup_accel()
67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/Linux-v6.1/drivers/crypto/qat/qat_c3xxx/
Dadf_drv.c39 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
41 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
42 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
45 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
47 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
57 if (accel_dev->hw_device) { in adf_cleanup_accel()
60 adf_clean_hw_data_c3xxx(accel_dev->hw_device); in adf_cleanup_accel()
65 kfree(accel_dev->hw_device); in adf_cleanup_accel()
66 accel_dev->hw_device = NULL; in adf_cleanup_accel()
68 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/Linux-v6.1/drivers/crypto/qat/qat_c62x/
Dadf_drv.c39 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
41 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
42 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
45 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
47 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
57 if (accel_dev->hw_device) { in adf_cleanup_accel()
60 adf_clean_hw_data_c62x(accel_dev->hw_device); in adf_cleanup_accel()
65 kfree(accel_dev->hw_device); in adf_cleanup_accel()
66 accel_dev->hw_device = NULL; in adf_cleanup_accel()
68 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]
/Linux-v6.1/drivers/crypto/qat/qat_dh895xcc/
Dadf_drv.c39 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument
41 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
42 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev()
45 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument
47 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel()
57 if (accel_dev->hw_device) { in adf_cleanup_accel()
60 adf_clean_hw_data_dh895xcc(accel_dev->hw_device); in adf_cleanup_accel()
65 kfree(accel_dev->hw_device); in adf_cleanup_accel()
66 accel_dev->hw_device = NULL; in adf_cleanup_accel()
68 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel()
[all …]

123