/Linux-v6.1/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_xgmi.c | 244 return sysfs_emit(buf, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id() 359 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive() 370 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive() 423 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive() 494 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate() 495 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate() 526 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology() 527 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology() 547 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count() 559 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_num_links() [all …]
|
D | gfxhub_v1_1.c | 88 if (max_region || adev->gmc.xgmi.connected_to_cpu) { in gfxhub_v1_1_get_xgmi_info() 89 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info() 91 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v1_1_get_xgmi_info() 95 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info() 99 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info() 104 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v1_1_get_xgmi_info() 107 adev->gmc.xgmi.node_segment_size = seg_size; in gfxhub_v1_1_get_xgmi_info()
|
D | amdgpu_gmc.c | 51 u64 vram_size = adev->gmc.xgmi.node_segment_size * adev->gmc.xgmi.num_physical_nodes; in amdgpu_gmc_pdb0_alloc() 211 if (mc->xgmi.num_physical_nodes == 0) { in amdgpu_gmc_vram_location() 239 u64 hive_vram_end = mc->xgmi.node_segment_size * mc->xgmi.num_physical_nodes - 1; in amdgpu_gmc_sysvm_location() 240 mc->vram_start = mc->xgmi.node_segment_size * mc->xgmi.physical_node_id; in amdgpu_gmc_sysvm_location() 241 mc->vram_end = mc->vram_start + mc->xgmi.node_segment_size - 1; in amdgpu_gmc_sysvm_location() 444 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_gmc_ras_early_init() 445 adev->gmc.xgmi.ras = &xgmi_ras; in amdgpu_gmc_ras_early_init() 446 amdgpu_ras_register_ras_block(adev, &adev->gmc.xgmi.ras->ras_block); in amdgpu_gmc_ras_early_init() 447 adev->gmc.xgmi.ras_if = &adev->gmc.xgmi.ras->ras_block.ras_comm; in amdgpu_gmc_ras_early_init() 709 u64 vram_size = adev->gmc.xgmi.node_segment_size * adev->gmc.xgmi.num_physical_nodes; in amdgpu_gmc_init_pdb0() [all …]
|
D | amdgpu_xgmi.h | 73 adev->gmc.xgmi.hive_id && in amdgpu_xgmi_same_hive() 74 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id); in amdgpu_xgmi_same_hive()
|
D | gmc_v9_0.c | 687 !adev->gmc.xgmi.connected_to_cpu) { in gmc_v9_0_set_irq_funcs() 772 if (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb() 913 bool vega20_xgmi_wa = (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb_pasid() 1227 if (!adev->gmc.xgmi.connected_to_cpu) in gmc_v9_0_set_umc_funcs() 1315 if (!adev->gmc.xgmi.connected_to_cpu) in gmc_v9_0_set_mca_funcs() 1331 adev->gmc.xgmi.supported = true; in gmc_v9_0_early_init() 1334 adev->gmc.xgmi.supported = true; in gmc_v9_0_early_init() 1335 adev->gmc.xgmi.connected_to_cpu = in gmc_v9_0_early_init() 1407 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location() 1408 if (adev->gmc.xgmi.connected_to_cpu) { in gmc_v9_0_vram_gtt_location() [all …]
|
D | aldebaran.c | 39 adev->gmc.xgmi.connected_to_cpu)) in aldebaran_is_mode2_default() 176 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in aldebaran_mode2_perform_reset() 193 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in aldebaran_mode2_perform_reset() 366 tmp_adev->gmc.xgmi.num_physical_nodes > 1) in aldebaran_mode2_restore_hwcontext()
|
D | amdgpu_device.c | 1307 if (adev->gmc.xgmi.pending_reset) in amdgpu_device_need_post() 2460 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_init() 2485 if (!adev->gmc.xgmi.pending_reset) in amdgpu_device_ip_init() 2711 …if (amdgpu_passthrough(adev) && ((adev->asic_type == CHIP_ARCTURUS && adev->gmc.xgmi.num_physical_… in amdgpu_device_ip_late_init() 2715 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_late_init() 2731 if (mgpu_info.num_dgpu == adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_ip_late_init() 2845 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_fini() 2994 if (adev->gmc.xgmi.pending_reset && in amdgpu_device_ip_suspend_phase2() 3679 if (adev->gmc.xgmi.supported) { in amdgpu_device_init() 3716 if (adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_init() [all …]
|
D | amdgpu_amdkfd.c | 559 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_hops_count() 560 peer_adev->gmc.xgmi.physical_node_id, ret); in amdgpu_amdkfd_get_xgmi_hops_count() 583 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_bandwidth_mbytes() 584 peer_adev->gmc.xgmi.physical_node_id, num_links); in amdgpu_amdkfd_get_xgmi_bandwidth_mbytes()
|
D | gfxhub_v2_1.c | 521 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v2_1_get_xgmi_info() 522 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v2_1_get_xgmi_info() 525 adev->gmc.xgmi.physical_node_id = in gfxhub_v2_1_get_xgmi_info() 527 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v2_1_get_xgmi_info() 530 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v2_1_get_xgmi_info()
|
D | amdgpu_ras.c | 1100 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_error_inject() 1586 if (!adev->gmc.xgmi.connected_to_cpu) in amdgpu_ras_interrupt_poison_consumption_handler() 1929 if (hive && adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_do_recovery() 1933 list_add_tail(&adev->gmc.xgmi.head, &device_list); in amdgpu_ras_do_recovery() 1938 device_list_handle, gmc.xgmi.head) { in amdgpu_ras_do_recovery() 2198 if (adev->gmc.xgmi.pending_reset) in amdgpu_ras_recovery_init() 2223 (adev->gmc.xgmi.connected_to_cpu)) in amdgpu_ras_recovery_init() 2332 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_ras_check_supported() 2457 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_ras_init() 2483 if (adev->gmc.xgmi.connected_to_cpu) { in amdgpu_ras_init() [all …]
|
D | gmc_v10_0.c | 788 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location() 799 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location()
|
D | amdgpu_psp.c | 387 !adev->gmc.xgmi.connected_to_cpu && in psp_sw_init() 1179 adev->gmc.xgmi.connected_to_cpu)) in psp_xgmi_terminate() 1291 uint64_t src_node_id = psp->adev->gmc.xgmi.node_id; in psp_xgmi_reflect_topology_info() 1297 list_for_each_entry(mirror_adev, &hive->device_list, gmc.xgmi.head) { in psp_xgmi_reflect_topology_info() 1301 if (mirror_adev->gmc.xgmi.node_id != dst_node_id) in psp_xgmi_reflect_topology_info() 1628 if (!adev->gmc.xgmi.connected_to_cpu) in psp_ras_initialize() 2571 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_load_fw() 2664 if (adev->gmc.xgmi.num_physical_nodes > 1) in psp_hw_fini() 2684 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_suspend() 2780 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_resume() [all …]
|
D | psp_v11_0.c | 157 le32_to_cpu(ta_hdr->xgmi.fw_version); in psp_v11_0_init_microcode() 159 le32_to_cpu(ta_hdr->xgmi.size_bytes); in psp_v11_0_init_microcode()
|
D | amdgpu_gmc.h | 253 struct amdgpu_xgmi xgmi; member
|
D | amdgpu_ib.c | 379 } else if (adev->gmc.xgmi.hive_id) { in amdgpu_ib_ring_tests()
|
D | amdgpu_ucode.h | 140 struct psp_fw_legacy_bin_desc xgmi; member
|
D | gfxhub_v1_0.c | 206 if (adev->gmc.xgmi.connected_to_cpu) { in gfxhub_v1_0_init_cache_regs()
|
D | amdgpu_object.c | 1046 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_bo_init() 1083 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_bo_fini()
|
D | soc15.c | 529 if (adev->gmc.xgmi.supported && adev->gmc.xgmi.connected_to_cpu) in soc15_asic_reset_method()
|
D | amdgpu_vram_mgr.c | 548 if (adev->gmc.xgmi.connected_to_cpu) in amdgpu_vram_mgr_new()
|
D | amdgpu_drv.c | 2208 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_pci_remove() 2309 adev->gmc.xgmi.pending_reset = false; in amdgpu_drv_delayed_reset_work_handler()
|
D | mmhub_v1_7.c | 212 if (adev->gmc.xgmi.connected_to_cpu) { in mmhub_v1_7_init_cache_regs()
|
/Linux-v6.1/Documentation/gpu/amdgpu/ |
D | index.rst | 13 xgmi
|
/Linux-v6.1/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_device.c | 600 kfd->hive_id = kfd->adev->gmc.xgmi.hive_id; in kgd2kfd_device_init() 1145 if (!kdev->adev->gmc.xgmi.supported) in kfd_get_num_sdma_engines()
|
D | kfd_migrate.c | 687 if (adev->gmc.xgmi.connected_to_cpu) in svm_migrate_vma_to_ram() 1010 if (adev->gmc.xgmi.connected_to_cpu) { in svm_migrate_init()
|