Home
last modified time | relevance | path

Searched refs:xgmi (Results 1 – 12 of 12) sorted by relevance

/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/
Damdgpu_xgmi.c130 return snprintf(buf, PAGE_SIZE, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id()
231 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
238 if (tmp->hive_id == adev->gmc.xgmi.hive_id) { in amdgpu_get_xgmi_hive()
259 tmp->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive()
290 adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate()
291 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate()
307 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology()
308 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology()
321 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count()
335 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_add_device()
[all …]
Dgfxhub_v1_1.c39 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info()
40 if (adev->gmc.xgmi.num_physical_nodes > 4) in gfxhub_v1_1_get_xgmi_info()
43 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info()
45 if (adev->gmc.xgmi.physical_node_id > 3) in gfxhub_v1_1_get_xgmi_info()
47 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v1_1_get_xgmi_info()
Damdgpu_xgmi.h50 adev->gmc.xgmi.hive_id && in amdgpu_xgmi_same_hive()
51 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id); in amdgpu_xgmi_same_hive()
Damdgpu_device.c1842 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_init()
2057 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_fini()
3612 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset()
3614 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset()
3630 gmc.xgmi.head) { in amdgpu_do_asic_reset()
3631 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset()
3640 gmc.xgmi.head) { in amdgpu_do_asic_reset()
3647 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset()
3695 if (hive && tmp_adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_do_asic_reset()
3816 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_gpu_recover()
[all …]
Dgmc_v9_0.c950 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location()
959 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location()
1202 if (adev->gmc.xgmi.supported) { in gmc_v9_0_sw_init()
Damdgpu_amdkfd.c569 return adev->gmc.xgmi.hive_id; in amdgpu_amdkfd_get_hive_id()
579 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_hops_count()
580 peer_adev->gmc.xgmi.physical_node_id, ret); in amdgpu_amdkfd_get_xgmi_hops_count()
Damdgpu_gmc.h178 struct amdgpu_xgmi xgmi; member
Damdgpu_gmc.c154 if (mc->xgmi.num_physical_nodes == 0) { in amdgpu_gmc_vram_location()
Damdgpu_ib.c353 } else if (adev->gmc.xgmi.hive_id) { in amdgpu_ib_ring_tests()
Damdgpu_psp.c833 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_hw_start()
1209 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_hw_fini()
1241 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_suspend()
Damdgpu_amdkfd_gfx_v9.c679 if (adev->gmc.xgmi.num_physical_nodes && in kgd_gfx_v9_invalidate_tlbs()
Dsoc15.c668 adev->gmc.xgmi.supported = true; in soc15_set_ip_blocks()