Lines Matching refs:phys_addr
176 __ioremap_caller(resource_size_t phys_addr, unsigned long size, in __ioremap_caller() argument
181 const resource_size_t unaligned_phys_addr = phys_addr; in __ioremap_caller()
191 last_addr = phys_addr + size - 1; in __ioremap_caller()
192 if (!size || last_addr < phys_addr) in __ioremap_caller()
195 if (!phys_addr_valid(phys_addr)) { in __ioremap_caller()
197 (unsigned long long)phys_addr); in __ioremap_caller()
202 __ioremap_check_mem(phys_addr, size, &io_desc); in __ioremap_caller()
209 &phys_addr, &last_addr); in __ioremap_caller()
216 offset = phys_addr & ~PAGE_MASK; in __ioremap_caller()
217 phys_addr &= PHYSICAL_PAGE_MASK; in __ioremap_caller()
218 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_caller()
220 retval = memtype_reserve(phys_addr, (u64)phys_addr + size, in __ioremap_caller()
228 if (!is_new_memtype_allowed(phys_addr, size, pcm, new_pcm)) { in __ioremap_caller()
231 (unsigned long long)phys_addr, in __ioremap_caller()
232 (unsigned long long)(phys_addr + size), in __ioremap_caller()
276 area->phys_addr = phys_addr; in __ioremap_caller()
279 if (memtype_kernel_map_sync(phys_addr, size, pcm)) in __ioremap_caller()
282 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) in __ioremap_caller()
299 memtype_free(phys_addr, phys_addr + size); in __ioremap_caller()
324 void __iomem *ioremap(resource_size_t phys_addr, unsigned long size) in ioremap() argument
336 return __ioremap_caller(phys_addr, size, pcm, in ioremap()
365 void __iomem *ioremap_uc(resource_size_t phys_addr, unsigned long size) in ioremap_uc() argument
369 return __ioremap_caller(phys_addr, size, pcm, in ioremap_uc()
384 void __iomem *ioremap_wc(resource_size_t phys_addr, unsigned long size) in ioremap_wc() argument
386 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WC, in ioremap_wc()
401 void __iomem *ioremap_wt(resource_size_t phys_addr, unsigned long size) in ioremap_wt() argument
403 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WT, in ioremap_wt()
408 void __iomem *ioremap_encrypted(resource_size_t phys_addr, unsigned long size) in ioremap_encrypted() argument
410 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WB, in ioremap_encrypted()
415 void __iomem *ioremap_cache(resource_size_t phys_addr, unsigned long size) in ioremap_cache() argument
417 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WB, in ioremap_cache()
422 void __iomem *ioremap_prot(resource_size_t phys_addr, unsigned long size, in ioremap_prot() argument
425 return __ioremap_caller(phys_addr, size, in ioremap_prot()
475 memtype_free(p->phys_addr, p->phys_addr + get_vm_area_size(p)); in iounmap()
539 static bool memremap_should_map_decrypted(resource_size_t phys_addr, in memremap_should_map_decrypted() argument
548 is_pmem = region_intersects(phys_addr, size, IORESOURCE_MEM, in memremap_should_map_decrypted()
558 switch (efi_mem_type(phys_addr)) { in memremap_should_map_decrypted()
560 if (efi_mem_attributes(phys_addr) & EFI_MEMORY_NV) in memremap_should_map_decrypted()
569 switch (e820__get_entry_type(phys_addr, phys_addr + size - 1)) { in memremap_should_map_decrypted()
592 static bool memremap_is_efi_data(resource_size_t phys_addr, in memremap_is_efi_data() argument
604 if (phys_addr == paddr) in memremap_is_efi_data()
610 if (phys_addr == paddr) in memremap_is_efi_data()
613 if (efi_is_table_address(phys_addr)) in memremap_is_efi_data()
616 switch (efi_mem_type(phys_addr)) { in memremap_is_efi_data()
631 static bool memremap_is_setup_data(resource_size_t phys_addr, in memremap_is_setup_data() argument
641 if (phys_addr == paddr) in memremap_is_setup_data()
650 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) { in memremap_is_setup_data()
663 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) in memremap_is_setup_data()
676 static bool __init early_memremap_is_setup_data(resource_size_t phys_addr, in early_memremap_is_setup_data() argument
686 if (phys_addr == paddr) in early_memremap_is_setup_data()
696 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) in early_memremap_is_setup_data()
710 bool arch_memremap_can_ram_remap(resource_size_t phys_addr, unsigned long size, in arch_memremap_can_ram_remap() argument
723 if (memremap_is_setup_data(phys_addr, size) || in arch_memremap_can_ram_remap()
724 memremap_is_efi_data(phys_addr, size)) in arch_memremap_can_ram_remap()
728 return !memremap_should_map_decrypted(phys_addr, size); in arch_memremap_can_ram_remap()
737 pgprot_t __init early_memremap_pgprot_adjust(resource_size_t phys_addr, in early_memremap_pgprot_adjust() argument
749 if (early_memremap_is_setup_data(phys_addr, size) || in early_memremap_pgprot_adjust()
750 memremap_is_efi_data(phys_addr, size)) in early_memremap_pgprot_adjust()
754 if (encrypted_prot && memremap_should_map_decrypted(phys_addr, size)) in early_memremap_pgprot_adjust()
761 bool phys_mem_access_encrypted(unsigned long phys_addr, unsigned long size) in phys_mem_access_encrypted() argument
763 return arch_memremap_can_ram_remap(phys_addr, size, 0); in phys_mem_access_encrypted()
768 void __init *early_memremap_encrypted(resource_size_t phys_addr, in early_memremap_encrypted() argument
771 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_ENC); in early_memremap_encrypted()
778 void __init *early_memremap_encrypted_wp(resource_size_t phys_addr, in early_memremap_encrypted_wp() argument
783 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_ENC_WP); in early_memremap_encrypted_wp()
787 void __init *early_memremap_decrypted(resource_size_t phys_addr, in early_memremap_decrypted() argument
790 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_NOENC); in early_memremap_decrypted()
797 void __init *early_memremap_decrypted_wp(resource_size_t phys_addr, in early_memremap_decrypted_wp() argument
802 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_NOENC_WP); in early_memremap_decrypted_wp()