Searched refs:VTD_PAGE_SIZE (Results 1 – 3 of 3) sorted by relevance
| /Linux-v6.1/drivers/iommu/intel/ |
| D | iommu.h | 33 #define VTD_PAGE_SIZE (1UL << VTD_PAGE_SHIFT) macro 35 #define VTD_PAGE_ALIGN(addr) (((addr) + VTD_PAGE_SIZE - 1) & VTD_PAGE_MASK) 698 return IS_ALIGNED((unsigned long)pte, VTD_PAGE_SIZE); in first_pte_in_page() 704 (struct dma_pte *)ALIGN((unsigned long)pte, VTD_PAGE_SIZE) - pte; in nr_pte_to_next_page()
|
| D | dmar.c | 891 addr = ioremap(drhd->address, VTD_PAGE_SIZE); in dmar_validate_one_drhd() 893 addr = early_ioremap(drhd->address, VTD_PAGE_SIZE); in dmar_validate_one_drhd() 905 early_iounmap(addr, VTD_PAGE_SIZE); in dmar_validate_one_drhd() 969 iommu->reg_size = VTD_PAGE_SIZE; in map_iommu() 1575 if (!IS_ALIGNED(addr, VTD_PAGE_SIZE << size_order)) in qi_flush_dev_iotlb_pasid()
|
| D | iommu.c | 34 #define ROOT_SIZE VTD_PAGE_SIZE 35 #define CONTEXT_SIZE VTD_PAGE_SIZE 140 #define ROOT_ENTRY_NR (VTD_PAGE_SIZE/sizeof(struct root_entry)) 960 domain_flush_cache(domain, tmp_page, VTD_PAGE_SIZE); in pfn_to_dma_pte() 2287 pteval += lvl_pages * VTD_PAGE_SIZE; in __domain_mapping() 2696 VTD_PAGE_SIZE); in copy_context_table() 2748 __iommu_flush_cache(iommu, new_ce, VTD_PAGE_SIZE); in copy_context_table() 4350 if (size < VTD_PAGE_SIZE << level_to_offset_bits(level)) in intel_iommu_unmap() 4351 size = VTD_PAGE_SIZE << level_to_offset_bits(level); in intel_iommu_unmap()
|