Lines Matching refs:virt

114 	struct amdgpu_virt *virt = &adev->virt;  in amdgpu_virt_request_full_gpu()  local
117 if (virt->ops && virt->ops->req_full_gpu) { in amdgpu_virt_request_full_gpu()
118 r = virt->ops->req_full_gpu(adev, init); in amdgpu_virt_request_full_gpu()
122 adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_request_full_gpu()
137 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_release_full_gpu() local
140 if (virt->ops && virt->ops->rel_full_gpu) { in amdgpu_virt_release_full_gpu()
141 r = virt->ops->rel_full_gpu(adev, init); in amdgpu_virt_release_full_gpu()
145 adev->virt.caps |= AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_release_full_gpu()
158 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_reset_gpu() local
161 if (virt->ops && virt->ops->reset_gpu) { in amdgpu_virt_reset_gpu()
162 r = virt->ops->reset_gpu(adev); in amdgpu_virt_reset_gpu()
166 adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_reset_gpu()
174 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_request_init_data() local
176 if (virt->ops && virt->ops->req_init_data) in amdgpu_virt_request_init_data()
177 virt->ops->req_init_data(adev); in amdgpu_virt_request_init_data()
179 if (adev->virt.req_init_data_ver > 0) in amdgpu_virt_request_init_data()
193 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_wait_reset() local
195 if (!virt->ops || !virt->ops->wait_reset) in amdgpu_virt_wait_reset()
198 return virt->ops->wait_reset(adev); in amdgpu_virt_wait_reset()
211 if (!amdgpu_sriov_vf(adev) || adev->virt.mm_table.gpu_addr) in amdgpu_virt_alloc_mm_table()
216 &adev->virt.mm_table.bo, in amdgpu_virt_alloc_mm_table()
217 &adev->virt.mm_table.gpu_addr, in amdgpu_virt_alloc_mm_table()
218 (void *)&adev->virt.mm_table.cpu_addr); in amdgpu_virt_alloc_mm_table()
224 memset((void *)adev->virt.mm_table.cpu_addr, 0, PAGE_SIZE); in amdgpu_virt_alloc_mm_table()
226 adev->virt.mm_table.gpu_addr, in amdgpu_virt_alloc_mm_table()
227 adev->virt.mm_table.cpu_addr); in amdgpu_virt_alloc_mm_table()
238 if (!amdgpu_sriov_vf(adev) || !adev->virt.mm_table.gpu_addr) in amdgpu_virt_free_mm_table()
241 amdgpu_bo_free_kernel(&adev->virt.mm_table.bo, in amdgpu_virt_free_mm_table()
242 &adev->virt.mm_table.gpu_addr, in amdgpu_virt_free_mm_table()
243 (void *)&adev->virt.mm_table.cpu_addr); in amdgpu_virt_free_mm_table()
244 adev->virt.mm_table.gpu_addr = 0; in amdgpu_virt_free_mm_table()
270 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_init_ras_err_handler_data() local
271 struct amdgpu_virt_ras_err_handler_data **data = &virt->virt_eh_data; in amdgpu_virt_init_ras_err_handler_data()
298 virt->ras_init_done = true; in amdgpu_virt_init_ras_err_handler_data()
305 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_ras_release_bp() local
306 struct amdgpu_virt_ras_err_handler_data *data = virt->virt_eh_data; in amdgpu_virt_ras_release_bp()
323 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_release_ras_err_handler_data() local
324 struct amdgpu_virt_ras_err_handler_data *data = virt->virt_eh_data; in amdgpu_virt_release_ras_err_handler_data()
326 virt->ras_init_done = false; in amdgpu_virt_release_ras_err_handler_data()
336 virt->virt_eh_data = NULL; in amdgpu_virt_release_ras_err_handler_data()
342 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_ras_add_bps() local
343 struct amdgpu_virt_ras_err_handler_data *data = virt->virt_eh_data; in amdgpu_virt_ras_add_bps()
354 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_ras_reserve_bps() local
355 struct amdgpu_virt_ras_err_handler_data *data = virt->virt_eh_data; in amdgpu_virt_ras_reserve_bps()
386 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_ras_check_bad_page() local
387 struct amdgpu_virt_ras_err_handler_data *data = virt->virt_eh_data; in amdgpu_virt_ras_check_bad_page()
426 struct amd_sriov_msg_pf2vf_info_header *pf2vf_info = adev->virt.fw_reserve.p_pf2vf; in amdgpu_virt_read_pf2vf_data()
430 if (adev->virt.fw_reserve.p_pf2vf == NULL) in amdgpu_virt_read_pf2vf_data()
442 adev->virt.fw_reserve.p_pf2vf, pf2vf_info->size, in amdgpu_virt_read_pf2vf_data()
443 adev->virt.fw_reserve.checksum_key, checksum); in amdgpu_virt_read_pf2vf_data()
449 adev->virt.gim_feature = in amdgpu_virt_read_pf2vf_data()
456 adev->virt.fw_reserve.p_pf2vf, pf2vf_info->size, in amdgpu_virt_read_pf2vf_data()
463 adev->virt.vf2pf_update_interval_ms = in amdgpu_virt_read_pf2vf_data()
465 adev->virt.gim_feature = in amdgpu_virt_read_pf2vf_data()
475 if (adev->virt.vf2pf_update_interval_ms < 200 || adev->virt.vf2pf_update_interval_ms > 10000) in amdgpu_virt_read_pf2vf_data()
476 adev->virt.vf2pf_update_interval_ms = 2000; in amdgpu_virt_read_pf2vf_data()
484 vf2pf_info = (struct amd_sriov_msg_vf2pf_info *) adev->virt.fw_reserve.p_vf2pf; in amdgpu_virt_populate_vf2pf_ucode_info()
486 if (adev->virt.fw_reserve.p_vf2pf == NULL) in amdgpu_virt_populate_vf2pf_ucode_info()
517 vf2pf_info = (struct amd_sriov_msg_vf2pf_info *) adev->virt.fw_reserve.p_vf2pf; in amdgpu_virt_write_vf2pf_data()
519 if (adev->virt.fw_reserve.p_vf2pf == NULL) in amdgpu_virt_write_vf2pf_data()
560 struct amdgpu_device *adev = container_of(work, struct amdgpu_device, virt.vf2pf_work.work); in amdgpu_virt_update_vf2pf_work_item()
565 schedule_delayed_work(&(adev->virt.vf2pf_work), adev->virt.vf2pf_update_interval_ms); in amdgpu_virt_update_vf2pf_work_item()
570 if (adev->virt.vf2pf_update_interval_ms != 0) { in amdgpu_virt_fini_data_exchange()
572 flush_delayed_work(&adev->virt.vf2pf_work); in amdgpu_virt_fini_data_exchange()
573 cancel_delayed_work_sync(&adev->virt.vf2pf_work); in amdgpu_virt_fini_data_exchange()
583 adev->virt.fw_reserve.p_pf2vf = NULL; in amdgpu_virt_init_data_exchange()
584 adev->virt.fw_reserve.p_vf2pf = NULL; in amdgpu_virt_init_data_exchange()
585 adev->virt.vf2pf_update_interval_ms = 0; in amdgpu_virt_init_data_exchange()
588 adev->virt.vf2pf_update_interval_ms = 2000; in amdgpu_virt_init_data_exchange()
590 adev->virt.fw_reserve.p_pf2vf = in amdgpu_virt_init_data_exchange()
593 adev->virt.fw_reserve.p_vf2pf = in amdgpu_virt_init_data_exchange()
601 if (adev->virt.fw_reserve.p_pf2vf->version == 2) { in amdgpu_virt_init_data_exchange()
602 pf2vf_v2 = (struct amd_sriov_msg_pf2vf_info *)adev->virt.fw_reserve.p_pf2vf; in amdgpu_virt_init_data_exchange()
608 if (bp_block_size && !adev->virt.ras_init_done) in amdgpu_virt_init_data_exchange()
611 if (adev->virt.ras_init_done) in amdgpu_virt_init_data_exchange()
616 if (adev->virt.vf2pf_update_interval_ms != 0) { in amdgpu_virt_init_data_exchange()
617 INIT_DELAYED_WORK(&adev->virt.vf2pf_work, amdgpu_virt_update_vf2pf_work_item); in amdgpu_virt_init_data_exchange()
618 schedule_delayed_work(&(adev->virt.vf2pf_work), adev->virt.vf2pf_update_interval_ms); in amdgpu_virt_init_data_exchange()
645 adev->virt.caps |= AMDGPU_SRIOV_CAPS_IS_VF; in amdgpu_detect_virtualization()
648 adev->virt.caps |= AMDGPU_SRIOV_CAPS_ENABLE_IOV; in amdgpu_detect_virtualization()
652 adev->virt.caps |= AMDGPU_PASSTHROUGH_MODE; in amdgpu_detect_virtualization()
698 adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_enable_access_debugfs()
708 adev->virt.caps |= AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_disable_access_debugfs()