Lines Matching refs:kdev

1335 static int kfd_fill_gpu_cache_info_from_gfx_config(struct kfd_dev *kdev,  in kfd_fill_gpu_cache_info_from_gfx_config()  argument
1338 struct amdgpu_device *adev = kdev->adev; in kfd_fill_gpu_cache_info_from_gfx_config()
1407 int kfd_get_gpu_cache_info(struct kfd_node *kdev, struct kfd_gpu_cache_info **pcache_info) in kfd_get_gpu_cache_info() argument
1411 switch (kdev->adev->asic_type) { in kfd_get_gpu_cache_info()
1449 switch (KFD_GC_VERSION(kdev)) { in kfd_get_gpu_cache_info()
1526 kfd_fill_gpu_cache_info_from_gfx_config(kdev->kfd, *pcache_info); in kfd_get_gpu_cache_info()
1790 struct kfd_node *kdev, uint8_t type, uint64_t size, in kfd_fill_gpu_memory_affinity() argument
1819 static void kfd_find_numa_node_in_srat(struct kfd_node *kdev) in kfd_find_numa_node_in_srat() argument
1824 u32 pci_id = pci_domain_nr(kdev->adev->pdev->bus) << 16 | in kfd_find_numa_node_in_srat()
1825 pci_dev_id(kdev->adev->pdev); in kfd_find_numa_node_in_srat()
1900 set_dev_node(&kdev->adev->pdev->dev, numa_node); in kfd_find_numa_node_in_srat()
1917 struct kfd_node *kdev, in kfd_fill_gpu_direct_io_link_to_cpu() argument
1931 if (kfd_dev_is_large_bar(kdev)) in kfd_fill_gpu_direct_io_link_to_cpu()
1937 if (kdev->adev->gmc.xgmi.connected_to_cpu || in kfd_fill_gpu_direct_io_link_to_cpu()
1938 (KFD_GC_VERSION(kdev) == IP_VERSION(9, 4, 3) && in kfd_fill_gpu_direct_io_link_to_cpu()
1939 kdev->adev->smuio.funcs->get_pkg_type(kdev->adev) == in kfd_fill_gpu_direct_io_link_to_cpu()
1941 bool ext_cpu = KFD_GC_VERSION(kdev) != IP_VERSION(9, 4, 3); in kfd_fill_gpu_direct_io_link_to_cpu()
1945 kdev->adev, NULL, true) : mem_bw; in kfd_fill_gpu_direct_io_link_to_cpu()
1960 amdgpu_amdkfd_get_pcie_bandwidth_mbytes(kdev->adev, true); in kfd_fill_gpu_direct_io_link_to_cpu()
1962 amdgpu_amdkfd_get_pcie_bandwidth_mbytes(kdev->adev, false); in kfd_fill_gpu_direct_io_link_to_cpu()
1968 if (kdev->adev->pdev->dev.numa_node == NUMA_NO_NODE && in kfd_fill_gpu_direct_io_link_to_cpu()
1970 kfd_find_numa_node_in_srat(kdev); in kfd_fill_gpu_direct_io_link_to_cpu()
1973 if (kdev->adev->pdev->dev.numa_node == NUMA_NO_NODE) in kfd_fill_gpu_direct_io_link_to_cpu()
1976 sub_type_hdr->proximity_domain_to = kdev->adev->pdev->dev.numa_node; in kfd_fill_gpu_direct_io_link_to_cpu()
1984 struct kfd_node *kdev, in kfd_fill_gpu_xgmi_link_to_gpu() argument
1990 bool use_ta_info = kdev->kfd->num_nodes == 1; in kfd_fill_gpu_xgmi_link_to_gpu()
2009 amdgpu_amdkfd_get_xgmi_hops_count(kdev->adev, peer_kdev->adev); in kfd_fill_gpu_xgmi_link_to_gpu()
2011 amdgpu_amdkfd_get_xgmi_bandwidth_mbytes(kdev->adev, in kfd_fill_gpu_xgmi_link_to_gpu()
2014 amdgpu_amdkfd_get_xgmi_bandwidth_mbytes(kdev->adev, NULL, true) : 0; in kfd_fill_gpu_xgmi_link_to_gpu()
2016 bool is_single_hop = kdev->kfd == peer_kdev->kfd; in kfd_fill_gpu_xgmi_link_to_gpu()
2036 size_t *size, struct kfd_node *kdev, in kfd_create_vcrat_image_gpu() argument
2088 amdgpu_amdkfd_get_cu_info(kdev->adev, &cu_info); in kfd_create_vcrat_image_gpu()
2091 (cu_info.cu_active_number / kdev->kfd->num_nodes); in kfd_create_vcrat_image_gpu()
2114 local_mem_info = kdev->local_mem_info; in kfd_create_vcrat_image_gpu()
2123 kdev, HSA_MEM_HEAP_TYPE_FB_PUBLIC, in kfd_create_vcrat_image_gpu()
2130 kdev, HSA_MEM_HEAP_TYPE_FB_PRIVATE, in kfd_create_vcrat_image_gpu()
2148 ret = kfd_fill_gpu_direct_io_link_to_cpu(&avail_size, kdev, in kfd_create_vcrat_image_gpu()
2166 if (kdev->kfd->hive_id) { in kfd_create_vcrat_image_gpu()
2171 if (peer_dev->gpu->kfd->hive_id != kdev->kfd->hive_id) in kfd_create_vcrat_image_gpu()
2177 &avail_size, kdev, peer_dev->gpu, in kfd_create_vcrat_image_gpu()
2210 int flags, struct kfd_node *kdev, in kfd_create_crat_image_virtual() argument
2242 if (!kdev) in kfd_create_crat_image_virtual()
2248 ret = kfd_create_vcrat_image_gpu(pcrat_image, size, kdev, in kfd_create_crat_image_virtual()