Home
last modified time | relevance | path

Searched refs:xgmi (Results 1 – 15 of 15) sorted by relevance

/Linux-v5.10/drivers/gpu/drm/amd/amdgpu/
Damdgpu_xgmi.c220 return snprintf(buf, PAGE_SIZE, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id()
330 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
342 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
366 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive()
432 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate()
433 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate()
461 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology()
462 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology()
475 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count()
489 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_add_device()
[all …]
Dgfxhub_v1_1.c55 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info()
56 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v1_1_get_xgmi_info()
59 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info()
61 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v1_1_get_xgmi_info()
63 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v1_1_get_xgmi_info()
Damdgpu_xgmi.h73 adev->gmc.xgmi.hive_id && in amdgpu_xgmi_same_hive()
74 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id); in amdgpu_xgmi_same_hive()
Dgfxhub_v2_1.c518 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v2_1_get_xgmi_info()
519 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v2_1_get_xgmi_info()
522 adev->gmc.xgmi.physical_node_id = in gfxhub_v2_1_get_xgmi_info()
524 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v2_1_get_xgmi_info()
527 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v2_1_get_xgmi_info()
Damdgpu_device.c2255 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_init()
2467 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_late_init()
2483 if (mgpu_info.num_dgpu == adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_ip_late_init()
2524 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_fini()
4257 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset()
4259 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset()
4275 gmc.xgmi.head) { in amdgpu_do_asic_reset()
4276 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset()
4287 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset()
4296 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset()
[all …]
Dgmc_v9_0.c729 if (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb()
855 bool vega20_xgmi_wa = (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb_pasid()
1241 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location()
1250 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location()
1464 if (adev->gmc.xgmi.supported) { in gmc_v9_0_sw_init()
Dgmc_v10_0.c697 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location()
707 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location()
857 if (adev->gmc.xgmi.supported) { in gmc_v10_0_sw_init()
Damdgpu_amdkfd.c538 return adev->gmc.xgmi.hive_id; in amdgpu_amdkfd_get_hive_id()
556 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_hops_count()
557 peer_adev->gmc.xgmi.physical_node_id, ret); in amdgpu_amdkfd_get_xgmi_hops_count()
Damdgpu_gmc.h240 struct amdgpu_xgmi xgmi; member
Damdgpu_ib.c391 } else if (adev->gmc.xgmi.hive_id) { in amdgpu_ib_ring_tests()
Damdgpu_ras.c871 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_error_inject()
1576 if (hive && adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_do_recovery()
1580 list_add_tail(&adev->gmc.xgmi.head, &device_list); in amdgpu_ras_do_recovery()
1585 device_list_handle, gmc.xgmi.head) { in amdgpu_ras_do_recovery()
Damdgpu_gmc.c157 if (mc->xgmi.num_physical_nodes == 0) { in amdgpu_gmc_vram_location()
Dnv.c475 adev->gmc.xgmi.supported = true; in nv_set_ip_blocks()
Damdgpu_psp.c2170 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_suspend()
2253 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_resume()
Dsoc15.c701 adev->gmc.xgmi.supported = true; in soc15_set_ip_blocks()