Searched refs:xgmi (Results 1 – 12 of 12) sorted by relevance
| /Linux-v5.4/drivers/gpu/drm/amd/amdgpu/ |
| D | amdgpu_xgmi.c | 130 return snprintf(buf, PAGE_SIZE, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id() 231 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive() 238 if (tmp->hive_id == adev->gmc.xgmi.hive_id) { in amdgpu_get_xgmi_hive() 259 tmp->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive() 290 adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate() 291 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate() 307 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology() 308 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology() 321 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count() 335 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_add_device() [all …]
|
| D | gfxhub_v1_1.c | 39 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info() 40 if (adev->gmc.xgmi.num_physical_nodes > 4) in gfxhub_v1_1_get_xgmi_info() 43 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info() 45 if (adev->gmc.xgmi.physical_node_id > 3) in gfxhub_v1_1_get_xgmi_info() 47 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v1_1_get_xgmi_info()
|
| D | amdgpu_xgmi.h | 50 adev->gmc.xgmi.hive_id && in amdgpu_xgmi_same_hive() 51 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id); in amdgpu_xgmi_same_hive()
|
| D | amdgpu_device.c | 1842 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_init() 2057 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_fini() 3612 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset() 3614 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset() 3630 gmc.xgmi.head) { in amdgpu_do_asic_reset() 3631 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset() 3640 gmc.xgmi.head) { in amdgpu_do_asic_reset() 3647 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset() 3695 if (hive && tmp_adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_do_asic_reset() 3816 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_gpu_recover() [all …]
|
| D | gmc_v9_0.c | 950 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location() 959 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location() 1202 if (adev->gmc.xgmi.supported) { in gmc_v9_0_sw_init()
|
| D | amdgpu_amdkfd.c | 569 return adev->gmc.xgmi.hive_id; in amdgpu_amdkfd_get_hive_id() 579 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_hops_count() 580 peer_adev->gmc.xgmi.physical_node_id, ret); in amdgpu_amdkfd_get_xgmi_hops_count()
|
| D | amdgpu_gmc.h | 178 struct amdgpu_xgmi xgmi; member
|
| D | amdgpu_gmc.c | 154 if (mc->xgmi.num_physical_nodes == 0) { in amdgpu_gmc_vram_location()
|
| D | amdgpu_ib.c | 353 } else if (adev->gmc.xgmi.hive_id) { in amdgpu_ib_ring_tests()
|
| D | amdgpu_psp.c | 833 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_hw_start() 1209 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_hw_fini() 1241 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_suspend()
|
| D | amdgpu_amdkfd_gfx_v9.c | 679 if (adev->gmc.xgmi.num_physical_nodes && in kgd_gfx_v9_invalidate_tlbs()
|
| D | soc15.c | 668 adev->gmc.xgmi.supported = true; in soc15_set_ip_blocks()
|