| /Linux-v6.1/drivers/crypto/qat/qat_common/ |
| D | adf_init.c | 59 int adf_dev_init(struct adf_accel_dev *accel_dev) in adf_dev_init() argument 63 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_dev_init() 67 dev_err(&GET_DEV(accel_dev), in adf_dev_init() 72 if (!test_bit(ADF_STATUS_CONFIGURED, &accel_dev->status) && in adf_dev_init() 73 !accel_dev->is_vf) { in adf_dev_init() 74 dev_err(&GET_DEV(accel_dev), "Device not configured\n"); in adf_dev_init() 78 if (adf_init_etr_data(accel_dev)) { in adf_dev_init() 79 dev_err(&GET_DEV(accel_dev), "Failed initialize etr\n"); in adf_dev_init() 83 if (hw_data->init_device && hw_data->init_device(accel_dev)) { in adf_dev_init() 84 dev_err(&GET_DEV(accel_dev), "Failed to initialize device\n"); in adf_dev_init() [all …]
|
| D | adf_vf_isr.c | 27 struct adf_accel_dev *accel_dev; member 31 void adf_enable_pf2vf_interrupts(struct adf_accel_dev *accel_dev) in adf_enable_pf2vf_interrupts() argument 33 void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev); in adf_enable_pf2vf_interrupts() 38 void adf_disable_pf2vf_interrupts(struct adf_accel_dev *accel_dev) in adf_disable_pf2vf_interrupts() argument 40 void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev); in adf_disable_pf2vf_interrupts() 46 static int adf_enable_msi(struct adf_accel_dev *accel_dev) in adf_enable_msi() argument 48 struct adf_accel_pci *pci_dev_info = &accel_dev->accel_pci_dev; in adf_enable_msi() 52 dev_err(&GET_DEV(accel_dev), in adf_enable_msi() 60 static void adf_disable_msi(struct adf_accel_dev *accel_dev) in adf_disable_msi() argument 62 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_disable_msi() [all …]
|
| D | adf_isr.c | 21 static int adf_enable_msix(struct adf_accel_dev *accel_dev) in adf_enable_msix() argument 23 struct adf_accel_pci *pci_dev_info = &accel_dev->accel_pci_dev; in adf_enable_msix() 24 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_msix() 29 hw_data->set_msix_rttable(accel_dev); in adf_enable_msix() 34 dev_err(&GET_DEV(accel_dev), in adf_enable_msix() 50 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(bank->accel_dev); in adf_msix_isr_bundle() 59 void adf_enable_vf2pf_interrupts(struct adf_accel_dev *accel_dev, u32 vf_mask) in adf_enable_vf2pf_interrupts() argument 61 void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev); in adf_enable_vf2pf_interrupts() 64 spin_lock_irqsave(&accel_dev->pf.vf2pf_ints_lock, flags); in adf_enable_vf2pf_interrupts() 65 GET_PFVF_OPS(accel_dev)->enable_vf2pf_interrupts(pmisc_addr, vf_mask); in adf_enable_vf2pf_interrupts() [all …]
|
| D | adf_accel_engine.c | 10 static int adf_ae_fw_load_images(struct adf_accel_dev *accel_dev, void *fw_addr, in adf_ae_fw_load_images() argument 13 struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; in adf_ae_fw_load_images() 14 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_ae_fw_load_images() 25 obj_name = hw_device->uof_get_name(accel_dev, i); in adf_ae_fw_load_images() 26 ae_mask = hw_device->uof_get_ae_mask(accel_dev, i); in adf_ae_fw_load_images() 28 dev_err(&GET_DEV(accel_dev), "Invalid UOF image\n"); in adf_ae_fw_load_images() 33 dev_err(&GET_DEV(accel_dev), in adf_ae_fw_load_images() 38 dev_err(&GET_DEV(accel_dev), in adf_ae_fw_load_images() 43 dev_err(&GET_DEV(accel_dev), in adf_ae_fw_load_images() 53 adf_ae_fw_release(accel_dev); in adf_ae_fw_load_images() [all …]
|
| D | adf_sriov.c | 25 struct adf_accel_dev *accel_dev = vf_info->accel_dev; in adf_iov_send_resp() local 29 ret = adf_recv_and_handle_vf2pf_msg(accel_dev, vf_nr); in adf_iov_send_resp() 32 adf_enable_vf2pf_interrupts(accel_dev, 1 << vf_nr); in adf_iov_send_resp() 50 static int adf_enable_sriov(struct adf_accel_dev *accel_dev) in adf_enable_sriov() argument 52 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_enable_sriov() 54 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_sriov() 58 for (i = 0, vf_info = accel_dev->pf.vf_info; i < totalvfs; in adf_enable_sriov() 61 vf_info->accel_dev = accel_dev; in adf_enable_sriov() 73 hw_data->configure_iov_threads(accel_dev, true); in adf_enable_sriov() 76 adf_enable_vf2pf_interrupts(accel_dev, BIT_ULL(totalvfs) - 1); in adf_enable_sriov() [all …]
|
| D | adf_common_drv.h | 44 int (*event_hld)(struct adf_accel_dev *accel_dev, 55 int adf_dev_init(struct adf_accel_dev *accel_dev); 56 int adf_dev_start(struct adf_accel_dev *accel_dev); 57 void adf_dev_stop(struct adf_accel_dev *accel_dev); 58 void adf_dev_shutdown(struct adf_accel_dev *accel_dev); 59 int adf_dev_shutdown_cache_cfg(struct adf_accel_dev *accel_dev); 69 int adf_devmgr_add_dev(struct adf_accel_dev *accel_dev, 71 void adf_devmgr_rm_dev(struct adf_accel_dev *accel_dev, 79 int adf_devmgr_in_reset(struct adf_accel_dev *accel_dev); 80 int adf_dev_started(struct adf_accel_dev *accel_dev); [all …]
|
| D | adf_aer.c | 17 struct adf_accel_dev *accel_dev = adf_devmgr_pci_to_accel_dev(pdev); in adf_error_detected() local 20 if (!accel_dev) { in adf_error_detected() 36 struct adf_accel_dev *accel_dev; member 41 void adf_reset_sbr(struct adf_accel_dev *accel_dev) in adf_reset_sbr() argument 43 struct pci_dev *pdev = accel_to_pci_dev(accel_dev); in adf_reset_sbr() 51 dev_info(&GET_DEV(accel_dev), in adf_reset_sbr() 54 dev_info(&GET_DEV(accel_dev), "Secondary bus reset\n"); in adf_reset_sbr() 66 void adf_reset_flr(struct adf_accel_dev *accel_dev) in adf_reset_flr() argument 68 pcie_flr(accel_to_pci_dev(accel_dev)); in adf_reset_flr() 72 void adf_dev_restore(struct adf_accel_dev *accel_dev) in adf_dev_restore() argument [all …]
|
| D | adf_pfvf_vf_proto.c | 33 int adf_send_vf2pf_msg(struct adf_accel_dev *accel_dev, struct pfvf_message msg) in adf_send_vf2pf_msg() argument 35 struct adf_pfvf_ops *pfvf_ops = GET_PFVF_OPS(accel_dev); in adf_send_vf2pf_msg() 38 return pfvf_ops->send_msg(accel_dev, msg, pfvf_offset, in adf_send_vf2pf_msg() 39 &accel_dev->vf.vf2pf_lock); in adf_send_vf2pf_msg() 50 static struct pfvf_message adf_recv_pf2vf_msg(struct adf_accel_dev *accel_dev) in adf_recv_pf2vf_msg() argument 52 struct adf_pfvf_ops *pfvf_ops = GET_PFVF_OPS(accel_dev); in adf_recv_pf2vf_msg() 55 return pfvf_ops->recv_msg(accel_dev, pfvf_offset, accel_dev->vf.pf_compat_ver); in adf_recv_pf2vf_msg() 69 int adf_send_vf2pf_req(struct adf_accel_dev *accel_dev, struct pfvf_message msg, in adf_send_vf2pf_req() argument 76 reinit_completion(&accel_dev->vf.msg_received); in adf_send_vf2pf_req() 80 ret = adf_send_vf2pf_msg(accel_dev, msg); in adf_send_vf2pf_req() [all …]
|
| D | adf_sysfs.c | 18 struct adf_accel_dev *accel_dev; in state_show() local 21 accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev)); in state_show() 22 if (!accel_dev) in state_show() 25 state = adf_dev_started(accel_dev) ? "up" : "down"; in state_show() 32 struct adf_accel_dev *accel_dev; in state_store() local 36 accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev)); in state_store() 37 if (!accel_dev) in state_store() 40 accel_id = accel_dev->accel_id; in state_store() 42 if (adf_devmgr_in_reset(accel_dev) || adf_dev_in_use(accel_dev)) { in state_store() 53 if (!adf_dev_started(accel_dev)) { in state_store() [all …]
|
| D | adf_pfvf_vf_msg.c | 18 int adf_vf2pf_notify_init(struct adf_accel_dev *accel_dev) in adf_vf2pf_notify_init() argument 22 if (adf_send_vf2pf_msg(accel_dev, msg)) { in adf_vf2pf_notify_init() 23 dev_err(&GET_DEV(accel_dev), in adf_vf2pf_notify_init() 27 set_bit(ADF_STATUS_PF_RUNNING, &accel_dev->status); in adf_vf2pf_notify_init() 40 void adf_vf2pf_notify_shutdown(struct adf_accel_dev *accel_dev) in adf_vf2pf_notify_shutdown() argument 44 if (test_bit(ADF_STATUS_PF_RUNNING, &accel_dev->status)) in adf_vf2pf_notify_shutdown() 45 if (adf_send_vf2pf_msg(accel_dev, msg)) in adf_vf2pf_notify_shutdown() 46 dev_err(&GET_DEV(accel_dev), in adf_vf2pf_notify_shutdown() 51 int adf_vf2pf_request_version(struct adf_accel_dev *accel_dev) in adf_vf2pf_request_version() argument 64 ret = adf_send_vf2pf_req(accel_dev, msg, &resp); in adf_vf2pf_request_version() [all …]
|
| D | qat_crypto.c | 22 adf_dev_put(inst->accel_dev); in qat_crypto_put_instance() 25 static int qat_crypto_free_instances(struct adf_accel_dev *accel_dev) in qat_crypto_free_instances() argument 30 list_for_each_entry_safe(inst, tmp, &accel_dev->crypto_list, list) { in qat_crypto_free_instances() 54 struct adf_accel_dev *accel_dev = NULL, *tmp_dev; in qat_crypto_get_instance_node() local 67 accel_dev = tmp_dev; in qat_crypto_get_instance_node() 73 if (!accel_dev) { in qat_crypto_get_instance_node() 79 accel_dev = tmp_dev; in qat_crypto_get_instance_node() 85 if (!accel_dev) in qat_crypto_get_instance_node() 89 list_for_each_entry(tmp_inst, &accel_dev->crypto_list, list) { in qat_crypto_get_instance_node() 99 if (adf_dev_get(accel_dev)) { in qat_crypto_get_instance_node() [all …]
|
| D | adf_admin.c | 109 static int adf_put_admin_msg_sync(struct adf_accel_dev *accel_dev, u32 ae, in adf_put_admin_msg_sync() argument 114 struct adf_admin_comms *admin = accel_dev->admin; in adf_put_admin_msg_sync() 136 dev_err(&GET_DEV(accel_dev), in adf_put_admin_msg_sync() 151 static int adf_send_admin(struct adf_accel_dev *accel_dev, in adf_send_admin() argument 159 if (adf_put_admin_msg_sync(accel_dev, ae, req, resp) || in adf_send_admin() 166 static int adf_init_ae(struct adf_accel_dev *accel_dev) in adf_init_ae() argument 170 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_init_ae() 177 return adf_send_admin(accel_dev, &req, &resp, ae_mask); in adf_init_ae() 180 static int adf_set_fw_constants(struct adf_accel_dev *accel_dev) in adf_set_fw_constants() argument 184 struct adf_hw_device_data *hw_device = accel_dev->hw_device; in adf_set_fw_constants() [all …]
|
| D | adf_ctl_drv.c | 110 static int adf_add_key_value_data(struct adf_accel_dev *accel_dev, in adf_add_key_value_data() argument 118 if (adf_cfg_add_key_value_param(accel_dev, section, in adf_add_key_value_data() 121 dev_err(&GET_DEV(accel_dev), in adf_add_key_value_data() 126 if (adf_cfg_add_key_value_param(accel_dev, section, in adf_add_key_value_data() 129 dev_err(&GET_DEV(accel_dev), in adf_add_key_value_data() 137 static int adf_copy_key_value_data(struct adf_accel_dev *accel_dev, in adf_copy_key_value_data() argument 150 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data() 155 if (adf_cfg_section_add(accel_dev, section.name)) { in adf_copy_key_value_data() 156 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data() 166 dev_err(&GET_DEV(accel_dev), in adf_copy_key_value_data() [all …]
|
| D | adf_accel_devices.h | 154 int (*enable_comms)(struct adf_accel_dev *accel_dev); 160 int (*send_msg)(struct adf_accel_dev *accel_dev, struct pfvf_message msg, 162 struct pfvf_message (*recv_msg)(struct adf_accel_dev *accel_dev, 170 u32 (*get_accel_cap)(struct adf_accel_dev *accel_dev); 179 int (*alloc_irq)(struct adf_accel_dev *accel_dev); 180 void (*free_irq)(struct adf_accel_dev *accel_dev); 181 void (*enable_error_correction)(struct adf_accel_dev *accel_dev); 182 int (*init_admin_comms)(struct adf_accel_dev *accel_dev); 183 void (*exit_admin_comms)(struct adf_accel_dev *accel_dev); 184 int (*send_admin_init)(struct adf_accel_dev *accel_dev); [all …]
|
| D | adf_transport.c | 61 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(bank->accel_dev); in adf_enable_ring_irq() 74 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(bank->accel_dev); in adf_disable_ring_irq() 90 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(ring->bank->accel_dev); in adf_send_message() 114 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(ring->bank->accel_dev); in adf_handle_response() 138 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(ring->bank->accel_dev); in adf_configure_tx_ring() 149 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(ring->bank->accel_dev); in adf_configure_rx_ring() 163 struct adf_accel_dev *accel_dev = bank->accel_dev; in adf_init_ring() local 164 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_ring() 165 struct adf_hw_csr_ops *csr_ops = GET_CSR_OPS(accel_dev); in adf_init_ring() 171 ring->base_addr = dma_alloc_coherent(&GET_DEV(accel_dev), in adf_init_ring() [all …]
|
| D | adf_pfvf_pf_proto.c | 32 int adf_send_pf2vf_msg(struct adf_accel_dev *accel_dev, u8 vf_nr, struct pfvf_message msg) in adf_send_pf2vf_msg() argument 34 struct adf_pfvf_ops *pfvf_ops = GET_PFVF_OPS(accel_dev); in adf_send_pf2vf_msg() 37 return pfvf_ops->send_msg(accel_dev, msg, pfvf_offset, in adf_send_pf2vf_msg() 38 &accel_dev->pf.vf_info[vf_nr].pf2vf_lock); in adf_send_pf2vf_msg() 50 static struct pfvf_message adf_recv_vf2pf_msg(struct adf_accel_dev *accel_dev, u8 vf_nr) in adf_recv_vf2pf_msg() argument 52 struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr]; in adf_recv_vf2pf_msg() 53 struct adf_pfvf_ops *pfvf_ops = GET_PFVF_OPS(accel_dev); in adf_recv_vf2pf_msg() 56 return pfvf_ops->recv_msg(accel_dev, pfvf_offset, vf_info->vf_compat_ver); in adf_recv_vf2pf_msg() 85 struct adf_accel_dev *accel_dev = vf_info->accel_dev; in adf_pf2vf_blkmsg_get_data() local 97 if (unlikely((*provider)(accel_dev, blkmsg, vf_info->vf_compat_ver))) { in adf_pf2vf_blkmsg_get_data() [all …]
|
| D | adf_dev_mgr.c | 139 int adf_devmgr_add_dev(struct adf_accel_dev *accel_dev, in adf_devmgr_add_dev() argument 146 dev_err(&GET_DEV(accel_dev), "Only support up to %d devices\n", in adf_devmgr_add_dev() 152 atomic_set(&accel_dev->ref_count, 0); in adf_devmgr_add_dev() 155 if (!accel_dev->is_vf || !pf) { in adf_devmgr_add_dev() 162 if (ptr == accel_dev) { in adf_devmgr_add_dev() 168 list_add_tail(&accel_dev->list, &accel_table); in adf_devmgr_add_dev() 169 accel_dev->accel_id = adf_find_free_id(); in adf_devmgr_add_dev() 170 if (accel_dev->accel_id > ADF_MAX_DEVICES) { in adf_devmgr_add_dev() 181 map->id = accel_dev->accel_id; in adf_devmgr_add_dev() 185 } else if (accel_dev->is_vf && pf) { in adf_devmgr_add_dev() [all …]
|
| D | adf_gen4_pm.c | 20 struct adf_accel_dev *accel_dev; member 24 static int send_host_msg(struct adf_accel_dev *accel_dev) in send_host_msg() argument 26 void __iomem *pmisc = adf_get_pmisc_base(accel_dev); in send_host_msg() 50 struct adf_accel_dev *accel_dev = pm_data->accel_dev; in pm_bh_handler() local 51 void __iomem *pmisc = adf_get_pmisc_base(accel_dev); in pm_bh_handler() 58 if (send_host_msg(accel_dev)) in pm_bh_handler() 59 dev_warn_ratelimited(&GET_DEV(accel_dev), in pm_bh_handler() 74 bool adf_gen4_handle_pm_interrupt(struct adf_accel_dev *accel_dev) in adf_gen4_handle_pm_interrupt() argument 76 void __iomem *pmisc = adf_get_pmisc_base(accel_dev); in adf_gen4_handle_pm_interrupt() 103 pm_data->accel_dev = accel_dev; in adf_gen4_handle_pm_interrupt() [all …]
|
| /Linux-v6.1/drivers/crypto/qat/qat_4xxx/ |
| D | adf_drv.c | 22 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 24 if (accel_dev->hw_device) { in adf_cleanup_accel() 25 adf_clean_hw_data_4xxx(accel_dev->hw_device); in adf_cleanup_accel() 26 accel_dev->hw_device = NULL; in adf_cleanup_accel() 28 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() 29 debugfs_remove(accel_dev->debugfs_dir); in adf_cleanup_accel() 30 adf_devmgr_rm_dev(accel_dev, NULL); in adf_cleanup_accel() 33 static int adf_cfg_dev_init(struct adf_accel_dev *accel_dev) in adf_cfg_dev_init() argument 38 config = accel_dev->accel_id % 2 ? ADF_CFG_DC : ADF_CFG_CY; in adf_cfg_dev_init() 40 ret = adf_cfg_section_add(accel_dev, ADF_GENERAL_SEC); in adf_cfg_dev_init() [all …]
|
| /Linux-v6.1/drivers/crypto/qat/qat_c3xxxvf/ |
| D | adf_drv.c | 37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 56 if (accel_dev->hw_device) { in adf_cleanup_accel() 59 adf_clean_hw_data_c3xxxiov(accel_dev->hw_device); in adf_cleanup_accel() 64 kfree(accel_dev->hw_device); in adf_cleanup_accel() 65 accel_dev->hw_device = NULL; in adf_cleanup_accel() 67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
| /Linux-v6.1/drivers/crypto/qat/qat_c62xvf/ |
| D | adf_drv.c | 37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 56 if (accel_dev->hw_device) { in adf_cleanup_accel() 59 adf_clean_hw_data_c62xiov(accel_dev->hw_device); in adf_cleanup_accel() 64 kfree(accel_dev->hw_device); in adf_cleanup_accel() 65 accel_dev->hw_device = NULL; in adf_cleanup_accel() 67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
| /Linux-v6.1/drivers/crypto/qat/qat_dh895xccvf/ |
| D | adf_drv.c | 37 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 39 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 40 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 43 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 45 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 56 if (accel_dev->hw_device) { in adf_cleanup_accel() 59 adf_clean_hw_data_dh895xcciov(accel_dev->hw_device); in adf_cleanup_accel() 64 kfree(accel_dev->hw_device); in adf_cleanup_accel() 65 accel_dev->hw_device = NULL; in adf_cleanup_accel() 67 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
| /Linux-v6.1/drivers/crypto/qat/qat_c3xxx/ |
| D | adf_drv.c | 39 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 41 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 42 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 45 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 47 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 57 if (accel_dev->hw_device) { in adf_cleanup_accel() 60 adf_clean_hw_data_c3xxx(accel_dev->hw_device); in adf_cleanup_accel() 65 kfree(accel_dev->hw_device); in adf_cleanup_accel() 66 accel_dev->hw_device = NULL; in adf_cleanup_accel() 68 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
| /Linux-v6.1/drivers/crypto/qat/qat_c62x/ |
| D | adf_drv.c | 39 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 41 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 42 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 45 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 47 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 57 if (accel_dev->hw_device) { in adf_cleanup_accel() 60 adf_clean_hw_data_c62x(accel_dev->hw_device); in adf_cleanup_accel() 65 kfree(accel_dev->hw_device); in adf_cleanup_accel() 66 accel_dev->hw_device = NULL; in adf_cleanup_accel() 68 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|
| /Linux-v6.1/drivers/crypto/qat/qat_dh895xcc/ |
| D | adf_drv.c | 39 static void adf_cleanup_pci_dev(struct adf_accel_dev *accel_dev) in adf_cleanup_pci_dev() argument 41 pci_release_regions(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 42 pci_disable_device(accel_dev->accel_pci_dev.pci_dev); in adf_cleanup_pci_dev() 45 static void adf_cleanup_accel(struct adf_accel_dev *accel_dev) in adf_cleanup_accel() argument 47 struct adf_accel_pci *accel_pci_dev = &accel_dev->accel_pci_dev; in adf_cleanup_accel() 57 if (accel_dev->hw_device) { in adf_cleanup_accel() 60 adf_clean_hw_data_dh895xcc(accel_dev->hw_device); in adf_cleanup_accel() 65 kfree(accel_dev->hw_device); in adf_cleanup_accel() 66 accel_dev->hw_device = NULL; in adf_cleanup_accel() 68 adf_cfg_dev_remove(accel_dev); in adf_cleanup_accel() [all …]
|