Lines Matching refs:phys_addr
155 __ioremap_caller(resource_size_t phys_addr, unsigned long size, in __ioremap_caller() argument
160 const resource_size_t unaligned_phys_addr = phys_addr; in __ioremap_caller()
170 last_addr = phys_addr + size - 1; in __ioremap_caller()
171 if (!size || last_addr < phys_addr) in __ioremap_caller()
174 if (!phys_addr_valid(phys_addr)) { in __ioremap_caller()
176 (unsigned long long)phys_addr); in __ioremap_caller()
181 __ioremap_check_mem(phys_addr, size, &io_desc); in __ioremap_caller()
188 &phys_addr, &last_addr); in __ioremap_caller()
195 offset = phys_addr & ~PAGE_MASK; in __ioremap_caller()
196 phys_addr &= PHYSICAL_PAGE_MASK; in __ioremap_caller()
197 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_caller()
199 retval = reserve_memtype(phys_addr, (u64)phys_addr + size, in __ioremap_caller()
207 if (!is_new_memtype_allowed(phys_addr, size, pcm, new_pcm)) { in __ioremap_caller()
210 (unsigned long long)phys_addr, in __ioremap_caller()
211 (unsigned long long)(phys_addr + size), in __ioremap_caller()
255 area->phys_addr = phys_addr; in __ioremap_caller()
258 if (kernel_map_sync_memtype(phys_addr, size, pcm)) in __ioremap_caller()
261 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) in __ioremap_caller()
278 free_memtype(phys_addr, phys_addr + size); in __ioremap_caller()
303 void __iomem *ioremap_nocache(resource_size_t phys_addr, unsigned long size) in ioremap_nocache() argument
315 return __ioremap_caller(phys_addr, size, pcm, in ioremap_nocache()
344 void __iomem *ioremap_uc(resource_size_t phys_addr, unsigned long size) in ioremap_uc() argument
348 return __ioremap_caller(phys_addr, size, pcm, in ioremap_uc()
363 void __iomem *ioremap_wc(resource_size_t phys_addr, unsigned long size) in ioremap_wc() argument
365 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WC, in ioremap_wc()
380 void __iomem *ioremap_wt(resource_size_t phys_addr, unsigned long size) in ioremap_wt() argument
382 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WT, in ioremap_wt()
387 void __iomem *ioremap_encrypted(resource_size_t phys_addr, unsigned long size) in ioremap_encrypted() argument
389 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WB, in ioremap_encrypted()
394 void __iomem *ioremap_cache(resource_size_t phys_addr, unsigned long size) in ioremap_cache() argument
396 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WB, in ioremap_cache()
401 void __iomem *ioremap_prot(resource_size_t phys_addr, unsigned long size, in ioremap_prot() argument
404 return __ioremap_caller(phys_addr, size, in ioremap_prot()
454 free_memtype(p->phys_addr, p->phys_addr + get_vm_area_size(p)); in iounmap()
518 static bool memremap_should_map_decrypted(resource_size_t phys_addr, in memremap_should_map_decrypted() argument
527 is_pmem = region_intersects(phys_addr, size, IORESOURCE_MEM, in memremap_should_map_decrypted()
537 switch (efi_mem_type(phys_addr)) { in memremap_should_map_decrypted()
539 if (efi_mem_attributes(phys_addr) & EFI_MEMORY_NV) in memremap_should_map_decrypted()
548 switch (e820__get_entry_type(phys_addr, phys_addr + size - 1)) { in memremap_should_map_decrypted()
571 static bool memremap_is_efi_data(resource_size_t phys_addr, in memremap_is_efi_data() argument
583 if (phys_addr == paddr) in memremap_is_efi_data()
589 if (phys_addr == paddr) in memremap_is_efi_data()
592 if (efi_is_table_address(phys_addr)) in memremap_is_efi_data()
595 switch (efi_mem_type(phys_addr)) { in memremap_is_efi_data()
610 static bool memremap_is_setup_data(resource_size_t phys_addr, in memremap_is_setup_data() argument
620 if (phys_addr == paddr) in memremap_is_setup_data()
631 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) in memremap_is_setup_data()
644 static bool __init early_memremap_is_setup_data(resource_size_t phys_addr, in early_memremap_is_setup_data() argument
654 if (phys_addr == paddr) in early_memremap_is_setup_data()
664 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) in early_memremap_is_setup_data()
678 bool arch_memremap_can_ram_remap(resource_size_t phys_addr, unsigned long size, in arch_memremap_can_ram_remap() argument
691 if (memremap_is_setup_data(phys_addr, size) || in arch_memremap_can_ram_remap()
692 memremap_is_efi_data(phys_addr, size)) in arch_memremap_can_ram_remap()
696 return !memremap_should_map_decrypted(phys_addr, size); in arch_memremap_can_ram_remap()
705 pgprot_t __init early_memremap_pgprot_adjust(resource_size_t phys_addr, in early_memremap_pgprot_adjust() argument
717 if (early_memremap_is_setup_data(phys_addr, size) || in early_memremap_pgprot_adjust()
718 memremap_is_efi_data(phys_addr, size)) in early_memremap_pgprot_adjust()
722 if (encrypted_prot && memremap_should_map_decrypted(phys_addr, size)) in early_memremap_pgprot_adjust()
729 bool phys_mem_access_encrypted(unsigned long phys_addr, unsigned long size) in phys_mem_access_encrypted() argument
731 return arch_memremap_can_ram_remap(phys_addr, size, 0); in phys_mem_access_encrypted()
736 void __init *early_memremap_encrypted(resource_size_t phys_addr, in early_memremap_encrypted() argument
739 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_ENC); in early_memremap_encrypted()
746 void __init *early_memremap_encrypted_wp(resource_size_t phys_addr, in early_memremap_encrypted_wp() argument
753 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_ENC_WP); in early_memremap_encrypted_wp()
757 void __init *early_memremap_decrypted(resource_size_t phys_addr, in early_memremap_decrypted() argument
760 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_NOENC); in early_memremap_decrypted()
767 void __init *early_memremap_decrypted_wp(resource_size_t phys_addr, in early_memremap_decrypted_wp() argument
774 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_NOENC_WP); in early_memremap_decrypted_wp()