Home
last modified time | relevance | path

Searched refs:gart (Results 1 – 25 of 49) sorted by relevance

12

/Linux-v6.1/drivers/iommu/
Dtegra-gart.c58 #define FLUSH_GART_REGS(gart) readl_relaxed((gart)->regs + GART_CONFIG) argument
60 #define for_each_gart_pte(gart, iova) \ argument
61 for (iova = gart->iovmm_base; \
62 iova < gart->iovmm_end; \
65 static inline void gart_set_pte(struct gart_device *gart, in gart_set_pte() argument
68 writel_relaxed(iova, gart->regs + GART_ENTRY_ADDR); in gart_set_pte()
69 writel_relaxed(pte, gart->regs + GART_ENTRY_DATA); in gart_set_pte()
72 static inline unsigned long gart_read_pte(struct gart_device *gart, in gart_read_pte() argument
77 writel_relaxed(iova, gart->regs + GART_ENTRY_ADDR); in gart_read_pte()
78 pte = readl_relaxed(gart->regs + GART_ENTRY_DATA); in gart_read_pte()
[all …]
DMakefile24 obj-$(CONFIG_TEGRA_IOMMU_GART) += tegra-gart.o
/Linux-v6.1/drivers/gpu/drm/radeon/
Dradeon_gart.c75 ptr = dma_alloc_coherent(&rdev->pdev->dev, rdev->gart.table_size, in radeon_gart_table_ram_alloc()
76 &rdev->gart.table_addr, GFP_KERNEL); in radeon_gart_table_ram_alloc()
84 rdev->gart.table_size >> PAGE_SHIFT); in radeon_gart_table_ram_alloc()
87 rdev->gart.ptr = ptr; in radeon_gart_table_ram_alloc()
102 if (rdev->gart.ptr == NULL) { in radeon_gart_table_ram_free()
108 set_memory_wb((unsigned long)rdev->gart.ptr, in radeon_gart_table_ram_free()
109 rdev->gart.table_size >> PAGE_SHIFT); in radeon_gart_table_ram_free()
112 dma_free_coherent(&rdev->pdev->dev, rdev->gart.table_size, in radeon_gart_table_ram_free()
113 (void *)rdev->gart.ptr, rdev->gart.table_addr); in radeon_gart_table_ram_free()
114 rdev->gart.ptr = NULL; in radeon_gart_table_ram_free()
[all …]
Drs400.c84 if (rdev->gart.ptr) { in rs400_gart_init()
106 rdev->gart.table_size = rdev->gart.num_gpu_pages * 4; in rs400_gart_init()
164 tmp = (u32)rdev->gart.table_addr & 0xfffff000; in rs400_gart_enable()
165 tmp |= (upper_32_bits(rdev->gart.table_addr) & 0xff) << 4; in rs400_gart_enable()
192 (unsigned long long)rdev->gart.table_addr); in rs400_gart_enable()
193 rdev->gart.ready = true; in rs400_gart_enable()
236 u32 *gtt = rdev->gart.ptr; in rs400_gart_set_page()
Dradeon_asic.c167 rdev->asic->gart.tlb_flush = &rv370_pcie_gart_tlb_flush; in radeon_agp_disable()
168 rdev->asic->gart.get_page_entry = &rv370_pcie_gart_get_page_entry; in radeon_agp_disable()
169 rdev->asic->gart.set_page = &rv370_pcie_gart_set_page; in radeon_agp_disable()
173 rdev->asic->gart.tlb_flush = &r100_pci_gart_tlb_flush; in radeon_agp_disable()
174 rdev->asic->gart.get_page_entry = &r100_pci_gart_get_page_entry; in radeon_agp_disable()
175 rdev->asic->gart.set_page = &r100_pci_gart_set_page; in radeon_agp_disable()
209 .gart = {
277 .gart = {
373 .gart = {
441 .gart = {
[all …]
Dr300.c122 void __iomem *ptr = rdev->gart.ptr; in rv370_pcie_gart_set_page()
134 if (rdev->gart.robj) { in rv370_pcie_gart_init()
144 rdev->gart.table_size = rdev->gart.num_gpu_pages * 4; in rv370_pcie_gart_init()
145 rdev->asic->gart.tlb_flush = &rv370_pcie_gart_tlb_flush; in rv370_pcie_gart_init()
146 rdev->asic->gart.get_page_entry = &rv370_pcie_gart_get_page_entry; in rv370_pcie_gart_init()
147 rdev->asic->gart.set_page = &rv370_pcie_gart_set_page; in rv370_pcie_gart_init()
157 if (rdev->gart.robj == NULL) { in rv370_pcie_gart_enable()
172 table_addr = rdev->gart.table_addr; in rv370_pcie_gart_enable()
187 rdev->gart.ready = true; in rv370_pcie_gart_enable()
Drs600.c550 if (rdev->gart.robj) { in rs600_gart_init()
559 rdev->gart.table_size = rdev->gart.num_gpu_pages * 8; in rs600_gart_init()
568 if (rdev->gart.robj == NULL) { in rs600_gart_enable()
605 rdev->gart.table_addr); in rs600_gart_enable()
622 (unsigned long long)rdev->gart.table_addr); in rs600_gart_enable()
623 rdev->gart.ready = true; in rs600_gart_enable()
663 void __iomem *ptr = (void *)rdev->gart.ptr; in rs600_gart_set_page()
Dradeon_ttm.c867 if (p >= rdev->gart.num_cpu_pages) in radeon_ttm_gtt_read()
870 page = rdev->gart.pages[p]; in radeon_ttm_gtt_read()
876 kunmap(rdev->gart.pages[p]); in radeon_ttm_gtt_read()
Drv770.c910 if (rdev->gart.robj == NULL) { in rv770_pcie_gart_enable()
939 WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR, rdev->gart.table_addr >> 12); in rv770_pcie_gart_enable()
950 (unsigned long long)rdev->gart.table_addr); in rv770_pcie_gart_enable()
951 rdev->gart.ready = true; in rv770_pcie_gart_enable()
Dr100.c655 if (rdev->gart.ptr) { in r100_pci_gart_init()
663 rdev->gart.table_size = rdev->gart.num_gpu_pages * 4; in r100_pci_gart_init()
664 rdev->asic->gart.tlb_flush = &r100_pci_gart_tlb_flush; in r100_pci_gart_init()
665 rdev->asic->gart.get_page_entry = &r100_pci_gart_get_page_entry; in r100_pci_gart_init()
666 rdev->asic->gart.set_page = &r100_pci_gart_set_page; in r100_pci_gart_init()
681 WREG32(RADEON_AIC_PT_BASE, rdev->gart.table_addr); in r100_pci_gart_enable()
687 (unsigned long long)rdev->gart.table_addr); in r100_pci_gart_enable()
688 rdev->gart.ready = true; in r100_pci_gart_enable()
711 u32 *gtt = rdev->gart.ptr; in r100_pci_gart_set_page()
Dni.c1265 if (rdev->gart.robj == NULL) { in cayman_pcie_gart_enable()
1294 WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR, rdev->gart.table_addr >> 12); in cayman_pcie_gart_enable()
1340 (unsigned long long)rdev->gart.table_addr); in cayman_pcie_gart_enable()
1341 rdev->gart.ready = true; in cayman_pcie_gart_enable()
/Linux-v6.1/drivers/gpu/drm/amd/amdgpu/
Damdgpu_gart.c117 if (adev->gart.bo != NULL) in amdgpu_gart_table_vram_alloc()
120 return amdgpu_bo_create_kernel(adev, adev->gart.table_size, PAGE_SIZE, in amdgpu_gart_table_vram_alloc()
121 AMDGPU_GEM_DOMAIN_VRAM, &adev->gart.bo, in amdgpu_gart_table_vram_alloc()
122 NULL, (void *)&adev->gart.ptr); in amdgpu_gart_table_vram_alloc()
136 amdgpu_bo_free_kernel(&adev->gart.bo, NULL, (void *)&adev->gart.ptr); in amdgpu_gart_table_vram_free()
164 if (!adev->gart.ptr) in amdgpu_gart_unbind()
174 if (!adev->gart.ptr) in amdgpu_gart_unbind()
178 amdgpu_gmc_set_pte_pde(adev, adev->gart.ptr, in amdgpu_gart_unbind()
244 if (!adev->gart.ptr) in amdgpu_gart_bind()
247 amdgpu_gart_map(adev, offset, pages, dma_addr, flags, adev->gart.ptr); in amdgpu_gart_bind()
[all …]
Dgmc_v11_0.c714 if (adev->gart.bo) { in gmc_v11_0_gart_init()
724 adev->gart.table_size = adev->gart.num_gpu_pages * 8; in gmc_v11_0_gart_init()
725 adev->gart.gart_pte_flags = AMDGPU_PTE_MTYPE_NV10(MTYPE_UC) | in gmc_v11_0_gart_init()
866 if (adev->gart.bo == NULL) { in gmc_v11_0_gart_enable()
888 (unsigned long long)amdgpu_bo_gpu_offset(adev->gart.bo)); in gmc_v11_0_gart_enable()
Dgmc_v6_0.c475 if (adev->gart.bo == NULL) { in gmc_v6_0_gart_enable()
481 table_addr = amdgpu_bo_gpu_offset(adev->gart.bo); in gmc_v6_0_gart_enable()
566 if (adev->gart.bo) { in gmc_v6_0_gart_init()
573 adev->gart.table_size = adev->gart.num_gpu_pages * 8; in gmc_v6_0_gart_init()
574 adev->gart.gart_pte_flags = 0; in gmc_v6_0_gart_init()
Dgmc_v10_0.c379 job->vm_pd_addr = amdgpu_gmc_pd_addr(adev->gart.bo); in gmc_v10_0_flush_gpu_tlb()
866 if (adev->gart.bo) { in gmc_v10_0_gart_init()
876 adev->gart.table_size = adev->gart.num_gpu_pages * 8; in gmc_v10_0_gart_init()
877 adev->gart.gart_pte_flags = AMDGPU_PTE_MTYPE_NV10(MTYPE_UC) | in gmc_v10_0_gart_init()
1055 if (adev->gart.bo == NULL) { in gmc_v10_0_gart_enable()
1084 (unsigned long long)amdgpu_bo_gpu_offset(adev->gart.bo)); in gmc_v10_0_gart_enable()
Dgmc_v7_0.c620 if (adev->gart.bo == NULL) { in gmc_v7_0_gart_enable()
625 table_addr = amdgpu_bo_gpu_offset(adev->gart.bo); in gmc_v7_0_gart_enable()
720 if (adev->gart.bo) { in gmc_v7_0_gart_init()
728 adev->gart.table_size = adev->gart.num_gpu_pages * 8; in gmc_v7_0_gart_init()
729 adev->gart.gart_pte_flags = 0; in gmc_v7_0_gart_init()
Dgmc_v8_0.c843 if (adev->gart.bo == NULL) { in gmc_v8_0_gart_enable()
848 table_addr = amdgpu_bo_gpu_offset(adev->gart.bo); in gmc_v8_0_gart_enable()
960 if (adev->gart.bo) { in gmc_v8_0_gart_init()
968 adev->gart.table_size = adev->gart.num_gpu_pages * 8; in gmc_v8_0_gart_init()
969 adev->gart.gart_pte_flags = AMDGPU_PTE_EXECUTABLE; in gmc_v8_0_gart_init()
Dgmc_v9_0.c1511 if (adev->gart.bo) { in gmc_v9_0_gart_init()
1528 adev->gart.table_size = adev->gart.num_gpu_pages * 8; in gmc_v9_0_gart_init()
1529 adev->gart.gart_pte_flags = AMDGPU_PTE_MTYPE_VG10(MTYPE_UC) | in gmc_v9_0_gart_init()
1794 if (adev->gart.bo == NULL) { in gmc_v9_0_gart_enable()
1814 (unsigned long long)amdgpu_bo_gpu_offset(adev->gart.bo)); in gmc_v9_0_gart_enable()
Damdgpu_gmc.c706 uint64_t flags = adev->gart.gart_pte_flags; //TODO it is UC. explore NC/RW? in amdgpu_gmc_init_pdb0()
714 u64 gart_ptb_gpu_pa = amdgpu_gmc_vram_pa(adev, adev->gart.bo); in amdgpu_gmc_init_pdb0()
/Linux-v6.1/include/soc/tegra/
Dmc.h118 int tegra_gart_suspend(struct gart_device *gart);
119 int tegra_gart_resume(struct gart_device *gart);
127 static inline int tegra_gart_suspend(struct gart_device *gart) in tegra_gart_suspend() argument
132 static inline int tegra_gart_resume(struct gart_device *gart) in tegra_gart_resume() argument
217 struct gart_device *gart; member
/Linux-v6.1/drivers/gpu/drm/nouveau/
Dnouveau_chan.h24 struct nvif_object gart; member
Dnouveau_chan.c97 nvif_object_dtor(&chan->gart); in nouveau_channel_del()
381 nouveau_channel_init(struct nouveau_channel *chan, u32 vram, u32 gart) in nouveau_channel_init() argument
446 ret = nvif_object_ctor(&chan->user, "abi16ChanGartCtxDma", gart, in nouveau_channel_init()
448 &chan->gart); in nouveau_channel_init()
/Linux-v6.1/drivers/memory/tegra/
Dtegra20.c694 if (IS_ENABLED(CONFIG_TEGRA_IOMMU_GART) && mc->gart) { in tegra20_mc_suspend()
695 err = tegra_gart_suspend(mc->gart); in tegra20_mc_suspend()
707 if (IS_ENABLED(CONFIG_TEGRA_IOMMU_GART) && mc->gart) { in tegra20_mc_resume()
708 err = tegra_gart_resume(mc->gart); in tegra20_mc_resume()
Dmc.c899 mc->gart = tegra_gart_probe(&pdev->dev, mc); in tegra_mc_probe()
900 if (IS_ERR(mc->gart)) { in tegra_mc_probe()
902 PTR_ERR(mc->gart)); in tegra_mc_probe()
903 mc->gart = NULL; in tegra_mc_probe()
/Linux-v6.1/drivers/gpu/drm/amd/amdkfd/
Dkfd_migrate.c76 dst_addr = amdgpu_bo_gpu_offset(adev->gart.bo); in svm_migrate_gart_map()
87 pte_flags |= adev->gart.gart_pte_flags; in svm_migrate_gart_map()

12