Lines Matching refs:pdd
259 struct kfd_process_device *pdd; in kfd_ioctl_create_queue() local
279 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_create_queue()
280 if (IS_ERR(pdd)) { in kfd_ioctl_create_queue()
472 struct kfd_process_device *pdd; in kfd_ioctl_set_memory_policy() local
491 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_set_memory_policy()
492 if (IS_ERR(pdd)) { in kfd_ioctl_set_memory_policy()
505 &pdd->qpd, in kfd_ioctl_set_memory_policy()
524 struct kfd_process_device *pdd; in kfd_ioctl_set_trap_handler() local
532 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_set_trap_handler()
533 if (IS_ERR(pdd)) { in kfd_ioctl_set_trap_handler()
539 &pdd->qpd, in kfd_ioctl_set_trap_handler()
556 struct kfd_process_device *pdd; in kfd_ioctl_dbg_register() local
576 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_dbg_register()
577 if (IS_ERR(pdd)) { in kfd_ioctl_dbg_register()
578 status = PTR_ERR(pdd); in kfd_ioctl_dbg_register()
856 struct kfd_process_device *pdd; in kfd_ioctl_get_process_apertures() local
867 pdd = kfd_get_first_process_device_data(p); in kfd_ioctl_get_process_apertures()
871 pAperture->gpu_id = pdd->dev->id; in kfd_ioctl_get_process_apertures()
872 pAperture->lds_base = pdd->lds_base; in kfd_ioctl_get_process_apertures()
873 pAperture->lds_limit = pdd->lds_limit; in kfd_ioctl_get_process_apertures()
874 pAperture->gpuvm_base = pdd->gpuvm_base; in kfd_ioctl_get_process_apertures()
875 pAperture->gpuvm_limit = pdd->gpuvm_limit; in kfd_ioctl_get_process_apertures()
876 pAperture->scratch_base = pdd->scratch_base; in kfd_ioctl_get_process_apertures()
877 pAperture->scratch_limit = pdd->scratch_limit; in kfd_ioctl_get_process_apertures()
882 "gpu id %u\n", pdd->dev->id); in kfd_ioctl_get_process_apertures()
884 "lds_base %llX\n", pdd->lds_base); in kfd_ioctl_get_process_apertures()
886 "lds_limit %llX\n", pdd->lds_limit); in kfd_ioctl_get_process_apertures()
888 "gpuvm_base %llX\n", pdd->gpuvm_base); in kfd_ioctl_get_process_apertures()
890 "gpuvm_limit %llX\n", pdd->gpuvm_limit); in kfd_ioctl_get_process_apertures()
892 "scratch_base %llX\n", pdd->scratch_base); in kfd_ioctl_get_process_apertures()
894 "scratch_limit %llX\n", pdd->scratch_limit); in kfd_ioctl_get_process_apertures()
898 pdd = kfd_get_next_process_device_data(p, pdd); in kfd_ioctl_get_process_apertures()
899 } while (pdd && (args->num_of_nodes < NUM_OF_SUPPORTED_GPUS)); in kfd_ioctl_get_process_apertures()
912 struct kfd_process_device *pdd; in kfd_ioctl_get_process_apertures_new() local
928 pdd = kfd_get_first_process_device_data(p); in kfd_ioctl_get_process_apertures_new()
931 pdd = kfd_get_next_process_device_data(p, pdd); in kfd_ioctl_get_process_apertures_new()
932 } while (pdd); in kfd_ioctl_get_process_apertures_new()
955 pdd = kfd_get_first_process_device_data(p); in kfd_ioctl_get_process_apertures_new()
957 pa[nodes].gpu_id = pdd->dev->id; in kfd_ioctl_get_process_apertures_new()
958 pa[nodes].lds_base = pdd->lds_base; in kfd_ioctl_get_process_apertures_new()
959 pa[nodes].lds_limit = pdd->lds_limit; in kfd_ioctl_get_process_apertures_new()
960 pa[nodes].gpuvm_base = pdd->gpuvm_base; in kfd_ioctl_get_process_apertures_new()
961 pa[nodes].gpuvm_limit = pdd->gpuvm_limit; in kfd_ioctl_get_process_apertures_new()
962 pa[nodes].scratch_base = pdd->scratch_base; in kfd_ioctl_get_process_apertures_new()
963 pa[nodes].scratch_limit = pdd->scratch_limit; in kfd_ioctl_get_process_apertures_new()
966 "gpu id %u\n", pdd->dev->id); in kfd_ioctl_get_process_apertures_new()
968 "lds_base %llX\n", pdd->lds_base); in kfd_ioctl_get_process_apertures_new()
970 "lds_limit %llX\n", pdd->lds_limit); in kfd_ioctl_get_process_apertures_new()
972 "gpuvm_base %llX\n", pdd->gpuvm_base); in kfd_ioctl_get_process_apertures_new()
974 "gpuvm_limit %llX\n", pdd->gpuvm_limit); in kfd_ioctl_get_process_apertures_new()
976 "scratch_base %llX\n", pdd->scratch_base); in kfd_ioctl_get_process_apertures_new()
978 "scratch_limit %llX\n", pdd->scratch_limit); in kfd_ioctl_get_process_apertures_new()
981 pdd = kfd_get_next_process_device_data(p, pdd); in kfd_ioctl_get_process_apertures_new()
982 } while (pdd && (nodes < args->num_of_nodes)); in kfd_ioctl_get_process_apertures_new()
1010 struct kfd_process_device *pdd; in kfd_ioctl_create_event() local
1026 pdd = kfd_bind_process_to_device(kfd, p); in kfd_ioctl_create_event()
1027 if (IS_ERR(pdd)) { in kfd_ioctl_create_event()
1028 err = PTR_ERR(pdd); in kfd_ioctl_create_event()
1032 mem = kfd_process_device_translate_handle(pdd, in kfd_ioctl_create_event()
1110 struct kfd_process_device *pdd; in kfd_ioctl_set_scratch_backing_va() local
1120 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_set_scratch_backing_va()
1121 if (IS_ERR(pdd)) { in kfd_ioctl_set_scratch_backing_va()
1122 err = PTR_ERR(pdd); in kfd_ioctl_set_scratch_backing_va()
1126 pdd->qpd.sh_hidden_private_base = args->va_addr; in kfd_ioctl_set_scratch_backing_va()
1131 pdd->qpd.vmid != 0) in kfd_ioctl_set_scratch_backing_va()
1133 dev->kgd, args->va_addr, pdd->qpd.vmid); in kfd_ioctl_set_scratch_backing_va()
1188 struct kfd_process_device *pdd; in kfd_ioctl_acquire_vm() local
1203 pdd = kfd_get_process_device_data(dev, p); in kfd_ioctl_acquire_vm()
1204 if (!pdd) { in kfd_ioctl_acquire_vm()
1209 if (pdd->drm_file) { in kfd_ioctl_acquire_vm()
1210 ret = pdd->drm_file == drm_file ? 0 : -EBUSY; in kfd_ioctl_acquire_vm()
1214 ret = kfd_process_device_init_vm(pdd, drm_file); in kfd_ioctl_acquire_vm()
1251 struct kfd_process_device *pdd; in kfd_ioctl_alloc_memory_of_gpu() local
1287 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_alloc_memory_of_gpu()
1288 if (IS_ERR(pdd)) { in kfd_ioctl_alloc_memory_of_gpu()
1289 err = PTR_ERR(pdd); in kfd_ioctl_alloc_memory_of_gpu()
1295 pdd->vm, (struct kgd_mem **) &mem, &offset, in kfd_ioctl_alloc_memory_of_gpu()
1301 idr_handle = kfd_process_device_create_obj_handle(pdd, mem); in kfd_ioctl_alloc_memory_of_gpu()
1333 struct kfd_process_device *pdd; in kfd_ioctl_free_memory_of_gpu() local
1344 pdd = kfd_get_process_device_data(dev, p); in kfd_ioctl_free_memory_of_gpu()
1345 if (!pdd) { in kfd_ioctl_free_memory_of_gpu()
1352 pdd, GET_IDR_HANDLE(args->handle)); in kfd_ioctl_free_memory_of_gpu()
1366 pdd, GET_IDR_HANDLE(args->handle)); in kfd_ioctl_free_memory_of_gpu()
1377 struct kfd_process_device *pdd, *peer_pdd; in kfd_ioctl_map_memory_to_gpu() local
1412 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_map_memory_to_gpu()
1413 if (IS_ERR(pdd)) { in kfd_ioctl_map_memory_to_gpu()
1414 err = PTR_ERR(pdd); in kfd_ioctl_map_memory_to_gpu()
1418 mem = kfd_process_device_translate_handle(pdd, in kfd_ioctl_map_memory_to_gpu()
1487 struct kfd_process_device *pdd, *peer_pdd; in kfd_ioctl_unmap_memory_from_gpu() local
1521 pdd = kfd_get_process_device_data(dev, p); in kfd_ioctl_unmap_memory_from_gpu()
1522 if (!pdd) { in kfd_ioctl_unmap_memory_from_gpu()
1527 mem = kfd_process_device_translate_handle(pdd, in kfd_ioctl_unmap_memory_from_gpu()
1629 struct kfd_process_device *pdd; in kfd_ioctl_import_dmabuf() local
1647 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_import_dmabuf()
1648 if (IS_ERR(pdd)) { in kfd_ioctl_import_dmabuf()
1649 r = PTR_ERR(pdd); in kfd_ioctl_import_dmabuf()
1654 args->va_addr, pdd->vm, in kfd_ioctl_import_dmabuf()
1660 idr_handle = kfd_process_device_create_obj_handle(pdd, mem); in kfd_ioctl_import_dmabuf()