Lines Matching refs:pdd
260 struct kfd_process_device *pdd; in kfd_ioctl_create_queue() local
280 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_create_queue()
281 if (IS_ERR(pdd)) { in kfd_ioctl_create_queue()
456 struct kfd_process_device *pdd; in kfd_ioctl_set_memory_policy() local
475 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_set_memory_policy()
476 if (IS_ERR(pdd)) { in kfd_ioctl_set_memory_policy()
489 &pdd->qpd, in kfd_ioctl_set_memory_policy()
508 struct kfd_process_device *pdd; in kfd_ioctl_set_trap_handler() local
516 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_set_trap_handler()
517 if (IS_ERR(pdd)) { in kfd_ioctl_set_trap_handler()
523 &pdd->qpd, in kfd_ioctl_set_trap_handler()
540 struct kfd_process_device *pdd; in kfd_ioctl_dbg_register() local
560 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_dbg_register()
561 if (IS_ERR(pdd)) { in kfd_ioctl_dbg_register()
562 status = PTR_ERR(pdd); in kfd_ioctl_dbg_register()
841 struct kfd_process_device *pdd; in kfd_ioctl_get_process_apertures() local
852 pdd = kfd_get_first_process_device_data(p); in kfd_ioctl_get_process_apertures()
856 pAperture->gpu_id = pdd->dev->id; in kfd_ioctl_get_process_apertures()
857 pAperture->lds_base = pdd->lds_base; in kfd_ioctl_get_process_apertures()
858 pAperture->lds_limit = pdd->lds_limit; in kfd_ioctl_get_process_apertures()
859 pAperture->gpuvm_base = pdd->gpuvm_base; in kfd_ioctl_get_process_apertures()
860 pAperture->gpuvm_limit = pdd->gpuvm_limit; in kfd_ioctl_get_process_apertures()
861 pAperture->scratch_base = pdd->scratch_base; in kfd_ioctl_get_process_apertures()
862 pAperture->scratch_limit = pdd->scratch_limit; in kfd_ioctl_get_process_apertures()
867 "gpu id %u\n", pdd->dev->id); in kfd_ioctl_get_process_apertures()
869 "lds_base %llX\n", pdd->lds_base); in kfd_ioctl_get_process_apertures()
871 "lds_limit %llX\n", pdd->lds_limit); in kfd_ioctl_get_process_apertures()
873 "gpuvm_base %llX\n", pdd->gpuvm_base); in kfd_ioctl_get_process_apertures()
875 "gpuvm_limit %llX\n", pdd->gpuvm_limit); in kfd_ioctl_get_process_apertures()
877 "scratch_base %llX\n", pdd->scratch_base); in kfd_ioctl_get_process_apertures()
879 "scratch_limit %llX\n", pdd->scratch_limit); in kfd_ioctl_get_process_apertures()
883 pdd = kfd_get_next_process_device_data(p, pdd); in kfd_ioctl_get_process_apertures()
884 } while (pdd && (args->num_of_nodes < NUM_OF_SUPPORTED_GPUS)); in kfd_ioctl_get_process_apertures()
897 struct kfd_process_device *pdd; in kfd_ioctl_get_process_apertures_new() local
913 pdd = kfd_get_first_process_device_data(p); in kfd_ioctl_get_process_apertures_new()
916 pdd = kfd_get_next_process_device_data(p, pdd); in kfd_ioctl_get_process_apertures_new()
917 } while (pdd); in kfd_ioctl_get_process_apertures_new()
940 pdd = kfd_get_first_process_device_data(p); in kfd_ioctl_get_process_apertures_new()
942 pa[nodes].gpu_id = pdd->dev->id; in kfd_ioctl_get_process_apertures_new()
943 pa[nodes].lds_base = pdd->lds_base; in kfd_ioctl_get_process_apertures_new()
944 pa[nodes].lds_limit = pdd->lds_limit; in kfd_ioctl_get_process_apertures_new()
945 pa[nodes].gpuvm_base = pdd->gpuvm_base; in kfd_ioctl_get_process_apertures_new()
946 pa[nodes].gpuvm_limit = pdd->gpuvm_limit; in kfd_ioctl_get_process_apertures_new()
947 pa[nodes].scratch_base = pdd->scratch_base; in kfd_ioctl_get_process_apertures_new()
948 pa[nodes].scratch_limit = pdd->scratch_limit; in kfd_ioctl_get_process_apertures_new()
951 "gpu id %u\n", pdd->dev->id); in kfd_ioctl_get_process_apertures_new()
953 "lds_base %llX\n", pdd->lds_base); in kfd_ioctl_get_process_apertures_new()
955 "lds_limit %llX\n", pdd->lds_limit); in kfd_ioctl_get_process_apertures_new()
957 "gpuvm_base %llX\n", pdd->gpuvm_base); in kfd_ioctl_get_process_apertures_new()
959 "gpuvm_limit %llX\n", pdd->gpuvm_limit); in kfd_ioctl_get_process_apertures_new()
961 "scratch_base %llX\n", pdd->scratch_base); in kfd_ioctl_get_process_apertures_new()
963 "scratch_limit %llX\n", pdd->scratch_limit); in kfd_ioctl_get_process_apertures_new()
966 pdd = kfd_get_next_process_device_data(p, pdd); in kfd_ioctl_get_process_apertures_new()
967 } while (pdd && (nodes < args->num_of_nodes)); in kfd_ioctl_get_process_apertures_new()
995 struct kfd_process_device *pdd; in kfd_ioctl_create_event() local
1011 pdd = kfd_bind_process_to_device(kfd, p); in kfd_ioctl_create_event()
1012 if (IS_ERR(pdd)) { in kfd_ioctl_create_event()
1013 err = PTR_ERR(pdd); in kfd_ioctl_create_event()
1017 mem = kfd_process_device_translate_handle(pdd, in kfd_ioctl_create_event()
1095 struct kfd_process_device *pdd; in kfd_ioctl_set_scratch_backing_va() local
1105 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_set_scratch_backing_va()
1106 if (IS_ERR(pdd)) { in kfd_ioctl_set_scratch_backing_va()
1107 err = PTR_ERR(pdd); in kfd_ioctl_set_scratch_backing_va()
1111 pdd->qpd.sh_hidden_private_base = args->va_addr; in kfd_ioctl_set_scratch_backing_va()
1116 pdd->qpd.vmid != 0) in kfd_ioctl_set_scratch_backing_va()
1118 dev->kgd, args->va_addr, pdd->qpd.vmid); in kfd_ioctl_set_scratch_backing_va()
1173 struct kfd_process_device *pdd; in kfd_ioctl_acquire_vm() local
1188 pdd = kfd_get_process_device_data(dev, p); in kfd_ioctl_acquire_vm()
1189 if (!pdd) { in kfd_ioctl_acquire_vm()
1194 if (pdd->drm_file) { in kfd_ioctl_acquire_vm()
1195 ret = pdd->drm_file == drm_file ? 0 : -EBUSY; in kfd_ioctl_acquire_vm()
1199 ret = kfd_process_device_init_vm(pdd, drm_file); in kfd_ioctl_acquire_vm()
1236 struct kfd_process_device *pdd; in kfd_ioctl_alloc_memory_of_gpu() local
1260 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_alloc_memory_of_gpu()
1261 if (IS_ERR(pdd)) { in kfd_ioctl_alloc_memory_of_gpu()
1262 err = PTR_ERR(pdd); in kfd_ioctl_alloc_memory_of_gpu()
1268 pdd->vm, (struct kgd_mem **) &mem, &offset, in kfd_ioctl_alloc_memory_of_gpu()
1274 idr_handle = kfd_process_device_create_obj_handle(pdd, mem); in kfd_ioctl_alloc_memory_of_gpu()
1298 struct kfd_process_device *pdd; in kfd_ioctl_free_memory_of_gpu() local
1309 pdd = kfd_get_process_device_data(dev, p); in kfd_ioctl_free_memory_of_gpu()
1310 if (!pdd) { in kfd_ioctl_free_memory_of_gpu()
1317 pdd, GET_IDR_HANDLE(args->handle)); in kfd_ioctl_free_memory_of_gpu()
1330 pdd, GET_IDR_HANDLE(args->handle)); in kfd_ioctl_free_memory_of_gpu()
1341 struct kfd_process_device *pdd, *peer_pdd; in kfd_ioctl_map_memory_to_gpu() local
1376 pdd = kfd_bind_process_to_device(dev, p); in kfd_ioctl_map_memory_to_gpu()
1377 if (IS_ERR(pdd)) { in kfd_ioctl_map_memory_to_gpu()
1378 err = PTR_ERR(pdd); in kfd_ioctl_map_memory_to_gpu()
1382 mem = kfd_process_device_translate_handle(pdd, in kfd_ioctl_map_memory_to_gpu()
1451 struct kfd_process_device *pdd, *peer_pdd; in kfd_ioctl_unmap_memory_from_gpu() local
1485 pdd = kfd_get_process_device_data(dev, p); in kfd_ioctl_unmap_memory_from_gpu()
1486 if (!pdd) { in kfd_ioctl_unmap_memory_from_gpu()
1491 mem = kfd_process_device_translate_handle(pdd, in kfd_ioctl_unmap_memory_from_gpu()