Home
last modified time | relevance | path

Searched refs:__pa (Results 1 – 25 of 421) sorted by relevance

12345678910>>...17

/Linux-v5.4/arch/xtensa/kernel/
Dsetup.c335 !mem_reserve(__pa(initrd_start), __pa(initrd_end))) in setup_arch()
341 mem_reserve(__pa(_stext), __pa(_end)); in setup_arch()
344 mem_reserve(__pa(&_WindowVectors_text_start), in setup_arch()
345 __pa(&_WindowVectors_text_end)); in setup_arch()
347 mem_reserve(__pa(&_DebugInterruptVector_text_start), in setup_arch()
348 __pa(&_DebugInterruptVector_text_end)); in setup_arch()
350 mem_reserve(__pa(&_KernelExceptionVector_text_start), in setup_arch()
351 __pa(&_KernelExceptionVector_text_end)); in setup_arch()
353 mem_reserve(__pa(&_UserExceptionVector_text_start), in setup_arch()
354 __pa(&_UserExceptionVector_text_end)); in setup_arch()
[all …]
/Linux-v5.4/arch/parisc/kernel/
Dfirmware.c161 __pa(pdc_result), 0); in set_firmware_width_unlocked()
242 retval = mem_pdc_call(PDC_INSTR, 0UL, __pa(pdc_result)); in pdc_instr()
267 __pa(pdc_result), __pa(pdc_result2), len); in pdc_chassis_info()
291 retval = mem_pdc_call(PDC_PAT_CHASSIS_LOG, PDC_PAT_CHASSIS_WRITE_LOG, __pa(&state), __pa(&data)); in pdc_pat_chassis_send_log()
337 retval = mem_pdc_call(PDC_CHASSIS, PDC_CHASSIS_WARN, __pa(pdc_result)); in pdc_chassis_warn()
348 ret = mem_pdc_call(PDC_COPROC, PDC_COPROC_CFG, __pa(pdc_result)); in pdc_coproc_cfg_unlocked()
395 retval = mem_pdc_call(PDC_IODC, PDC_IODC_READ, __pa(pdc_result), hpa, in pdc_iodc_read()
396 index, __pa(pdc_result2), iodc_data_size); in pdc_iodc_read()
422 retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_MODULE, __pa(pdc_result), in pdc_system_map_find_mods()
423 __pa(pdc_result2), mod_index); in pdc_system_map_find_mods()
[all …]
/Linux-v5.4/arch/x86/include/asm/
Dpgalloc.h66 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel()
67 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel()
73 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel_safe()
74 set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel_safe()
126 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate()
127 set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate()
132 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate_safe()
133 set_pud_safe(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate_safe()
140 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate()
141 set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); in p4d_populate()
[all …]
Dpage.h41 #ifndef __pa
42 #define __pa(x) __phys_addr((unsigned long)(x)) macro
63 #define __boot_pa(x) __pa(x)
69 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
/Linux-v5.4/arch/x86/kernel/
Dhead32.c74 #ifdef __pa in mk_early_pgtbl_32()
75 #undef __pa in mk_early_pgtbl_32()
77 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) in mk_early_pgtbl_32() macro
82 const unsigned long limit = __pa(_end) + in mk_early_pgtbl_32()
85 pmd_t pl2, *pl2p = (pmd_t *)__pa(initial_pg_pmd); in mk_early_pgtbl_32()
88 pgd_t pl2, *pl2p = (pgd_t *)__pa(initial_page_table); in mk_early_pgtbl_32()
92 ptep = (pte_t *)__pa(__brk_base); in mk_early_pgtbl_32()
112 ptr = (unsigned long *)__pa(&max_pfn_mapped); in mk_early_pgtbl_32()
116 ptr = (unsigned long *)__pa(&_brk_end); in mk_early_pgtbl_32()
Dmachine_kexec_32.c101 set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT)); in machine_kexec_page_table_set_one()
107 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in machine_kexec_page_table_set_one()
123 (unsigned long)control_page, __pa(control_page)); in machine_kexec_prepare_page_tables()
129 __pa(control_page), __pa(control_page)); in machine_kexec_prepare_page_tables()
213 page_list[PA_CONTROL_PAGE] = __pa(control_page); in machine_kexec()
215 page_list[PA_PGD] = __pa(image->arch.pgd); in machine_kexec()
Despfix_64.c170 pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap()
171 paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT); in init_espfix_ap()
182 pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap()
183 paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT); in init_espfix_ap()
194 pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); in init_espfix_ap()
/Linux-v5.4/arch/arm/mach-omap2/
Domap-secure.c50 outer_clean_range(__pa(param), __pa(param + 5)); in omap_secure_dispatcher()
51 ret = omap_smc2(idx, flag, __pa(param)); in omap_secure_dispatcher()
82 param[1] = __pa(addr); /* Physical address for saving */ in omap3_save_secure_ram()
87 ret = save_secure_ram_context(__pa(param)); in omap3_save_secure_ram()
125 outer_clean_range(__pa(param), __pa(param + 5)); in rx51_secure_dispatcher()
126 ret = omap_smc3(idx, process, flag, __pa(param)); in rx51_secure_dispatcher()
/Linux-v5.4/arch/riscv/mm/
Dinit.c69 if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { in setup_initrd()
75 memblock_reserve(__pa(initrd_start), size); in setup_initrd()
94 phys_addr_t vmlinux_end = __pa(&_end); in setup_bootmem()
95 phys_addr_t vmlinux_start = __pa(&_start); in setup_bootmem()
273 #define get_pgd_next_virt(__pa) get_pmd_virt(__pa) argument
274 #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \ argument
275 create_pmd_mapping(__nextp, __va, __pa, __sz, __prot)
281 #define get_pgd_next_virt(__pa) get_pte_virt(__pa) argument
282 #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \ argument
283 create_pte_mapping(__nextp, __va, __pa, __sz, __prot)
[all …]
/Linux-v5.4/arch/parisc/mm/
Dinit.c294 memblock_reserve(__pa(KERNEL_BINARY_TEXT_START), in setup_bootmem()
310 if (__pa(initrd_start) < mem_max) { in setup_bootmem()
313 if (__pa(initrd_end) > mem_max) { in setup_bootmem()
314 initrd_reserve = mem_max - __pa(initrd_start); in setup_bootmem()
319 …printk(KERN_INFO "initrd: reserving %08lx-%08lx (mem_max %08lx)\n", __pa(initrd_start), __pa(initr… in setup_bootmem()
321 memblock_reserve(__pa(initrd_start), initrd_reserve); in setup_bootmem()
368 ro_start = __pa((unsigned long)_text); in map_pages()
369 ro_end = __pa((unsigned long)&data_start); in map_pages()
370 kernel_start = __pa((unsigned long)&__init_begin); in map_pages()
371 kernel_end = __pa((unsigned long)&_end); in map_pages()
[all …]
/Linux-v5.4/arch/powerpc/mm/nohash/
D8xx.c36 return __pa(va); in v_block_mapped()
121 _ALIGN(__pa(_einittext), 8 << 20)); in mmu_mapin_ram()
148 mmu_patch_cmp_limit(&patch__itlbmiss_linmem_top, __pa(_etext)); in mmu_mark_initmem_nx()
156 -__pa(((unsigned long)_sinittext) & in mmu_mark_rodata_ro()
158 mmu_patch_addis(&patch__dtlbmiss_romem_top, -__pa(_sinittext)); in mmu_mark_rodata_ro()
184 s16 offset = (s16)(__pa(swapper_pg_dir)); in set_context()
197 mtspr(SPRN_M_TWB, __pa(pgd) - offset); in set_context()
/Linux-v5.4/arch/um/kernel/
Dphysmem.c107 os_seek_file(physmem_fd, __pa(__syscall_stub_start)); in setup_physmem()
111 memblock_add(__pa(start), len + highmem); in setup_physmem()
112 memblock_reserve(__pa(start), reserve); in setup_physmem()
114 min_low_pfn = PFN_UP(__pa(reserve_end)); in setup_physmem()
126 else if (phys < __pa(end_iomem)) { in phys_mapping()
139 else if (phys < __pa(end_iomem) + highmem) { in phys_mapping()
212 region->phys = __pa(region->virt); in setup_iomem()
Dmem.c49 map_memory(brk_end, __pa(brk_end), uml_reserved - brk_end, 1, 1, 0); in mem_init()
50 memblock_free(__pa(brk_end), uml_reserved - brk_end); in mem_init()
75 (unsigned long) __pa(pte))); in one_page_table_init()
89 set_pud(pud, __pud(_KERNPG_TABLE + (unsigned long) __pa(pmd_table))); in one_md_table_init()
143 p = __pa(v); in fixaddr_user_init()
/Linux-v5.4/arch/arm/mm/
Dpmsa-v8.c254 subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
255 subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
259 subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
260 subtract_range(io, ARRAY_SIZE(io), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
276 err |= pmsav8_setup_fixed(PMSAv8_XIP_REGION, CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
279 err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
/Linux-v5.4/arch/x86/power/
Dhibernate_64.c71 __pud(__pa(pmd) | pgprot_val(pgtable_prot))); in set_up_temporary_text_mapping()
73 p4d_t new_p4d = __p4d(__pa(pud) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping()
74 pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping()
80 pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping()
123 temp_pgt = __pa(pgd); in set_up_temporary_mappings()
/Linux-v5.4/arch/x86/xen/
Dp2m.c346 paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list()
348 paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list()
351 pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
353 pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
390 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); in xen_rebuild_p2m_list()
400 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
410 set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE)); in xen_rebuild_p2m_list()
459 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) in get_phys_to_machine()
497 paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT); in alloc_p2m_pmd()
509 __pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE)); in alloc_p2m_pmd()
[all …]
Dmmu_pv.c807 xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd))); in __xen_pgd_pin()
812 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_pin()
821 xen_do_pin(MMUEXT_PIN_L3_TABLE, PFN_DOWN(__pa(pgd))); in __xen_pgd_pin()
926 xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd))); in __xen_pgd_unpin()
934 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_unpin()
1002 if (this_cpu_read(xen_current_cr3) == __pa(mm->pgd)) in drop_mm_ref_this_cpu()
1021 if (per_cpu(xen_current_cr3, cpu) != __pa(mm->pgd)) in xen_drop_mm_ref()
1037 if (per_cpu(xen_current_cr3, cpu) == __pa(mm->pgd)) in xen_drop_mm_ref()
1129 unsigned long pa = __pa(pgtbl) & PHYSICAL_PAGE_MASK; in xen_cleanmfnmap_free_pgtbl()
1250 memblock_free(__pa(addr), size); in xen_pagetable_p2m_free()
[all …]
/Linux-v5.4/arch/x86/mm/
Dkasan_init_64.c30 __pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid); in early_alloc()
34 (void *)_RET_IP_, nid, __pa(MAX_DMA_ADDRESS)); in early_alloc()
51 if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL)) in kasan_populate_pmd()
54 memblock_free(__pa(p), PMD_SIZE); in kasan_populate_pmd()
70 entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL); in kasan_populate_pmd()
88 if (p && pud_set_huge(pud, __pa(p), PAGE_KERNEL)) in kasan_populate_pud()
91 memblock_free(__pa(p), PUD_SIZE); in kasan_populate_pud()
326 __pgd(__pa(tmp_p4d_table) | _KERNPG_TABLE)); in kasan_init()
367 early_pfn_to_nid(__pa(_stext))); in kasan_init()
388 pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot)); in kasan_init()
/Linux-v5.4/arch/hexagon/include/asm/
Dpage.h86 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + PHYS_OFFSET) macro
93 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr)))
100 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
130 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
/Linux-v5.4/arch/s390/mm/
Dkasan_init.c168 pmd_val(*pm_dir) = __pa(page) | sgt_prot; in kasan_early_vmemmap_populate()
187 pte_val(*pt_dir) = __pa(page) | pgt_prot; in kasan_early_vmemmap_populate()
192 pte_val(*pt_dir) = __pa(page) | pgt_prot; in kasan_early_vmemmap_populate()
196 pte_val(*pt_dir) = __pa(page) | pgt_prot_zero; in kasan_early_vmemmap_populate()
209 S390_lowcore.kernel_asce = (__pa(pgd) & PAGE_MASK) | asce_bits; in kasan_set_pgd()
249 pmd_t pmd_z = __pmd(__pa(kasan_early_shadow_pte) | _SEGMENT_ENTRY); in kasan_early_init()
250 pud_t pud_z = __pud(__pa(kasan_early_shadow_pmd) | _REGION3_ENTRY); in kasan_early_init()
251 p4d_t p4d_z = __p4d(__pa(kasan_early_shadow_pud) | _REGION2_ENTRY); in kasan_early_init()
256 pte_z = __pte(__pa(kasan_early_shadow_page) | pgt_prot); in kasan_early_init()
/Linux-v5.4/arch/parisc/include/asm/
Dpgalloc.h40 + (__u32)(__pa((unsigned long)pgd) >> PxD_VALUE_SHIFT)); in pgd_alloc()
65 (__u32)(__pa((unsigned long)pmd) >> PxD_VALUE_SHIFT)); in pgd_populate()
116 + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)); in pmd_populate_kernel()
120 + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)); in pmd_populate_kernel()
/Linux-v5.4/arch/powerpc/platforms/pseries/
Dsvm.c42 uv_unshare_page(PHYS_PFN(__pa(addr)), numpages); in set_memory_encrypted()
52 uv_share_page(PHYS_PFN(__pa(addr)), numpages); in set_memory_decrypted()
76 unsigned long pfn = PHYS_PFN(__pa(addr)); in dtl_cache_ctor()
/Linux-v5.4/arch/s390/boot/
Dstartup.c28 unsigned long __bootdata_preserved(__sdma) = __pa(&_sdma);
29 unsigned long __bootdata_preserved(__edma) = __pa(&_edma);
30 unsigned long __bootdata_preserved(__stext_dma) = __pa(&_stext_dma);
31 unsigned long __bootdata_preserved(__etext_dma) = __pa(&_etext_dma);
52 unsigned long __bootdata_preserved(__swsusp_reset_dma) = __pa(_swsusp_reset_dma);
/Linux-v5.4/arch/ia64/mm/
Dinit.c437 map_start = vmem_map + (__pa(start) >> PAGE_SHIFT); in create_mem_map_page_table()
438 map_end = vmem_map + (__pa(end) >> PAGE_SHIFT); in create_mem_map_page_table()
442 node = paddr_to_nid(__pa(start)); in create_mem_map_page_table()
475 set_pte(pte, pfn_pte(__pa(page) >> PAGE_SHIFT, in create_mem_map_page_table()
501 map_start = vmem_map + (__pa(start) >> PAGE_SHIFT); in virtual_memmap_init()
502 map_end = vmem_map + (__pa(end) >> PAGE_SHIFT); in virtual_memmap_init()
586 memblock_add_node(__pa(start), end - start, nid); in register_active_ranges()
595 pfn_start = (PAGE_ALIGN(__pa(start))) >> PAGE_SHIFT; in find_max_min_low_pfn()
596 pfn_end = (PAGE_ALIGN(__pa(end - 1))) >> PAGE_SHIFT; in find_max_min_low_pfn()
598 pfn_start = GRANULEROUNDDOWN(__pa(start)) >> PAGE_SHIFT; in find_max_min_low_pfn()
[all …]
/Linux-v5.4/drivers/i2c/busses/
Di2c-opal.c90 req.buffer_ra = cpu_to_be64(__pa(msgs[0].buf)); in i2c_opal_master_xfer()
101 req.buffer_ra = cpu_to_be64(__pa(msgs[1].buf)); in i2c_opal_master_xfer()
126 req.buffer_ra = cpu_to_be64(__pa(&data->byte)); in i2c_opal_smbus_xfer()
134 req.buffer_ra = cpu_to_be64(__pa(&data->byte)); in i2c_opal_smbus_xfer()
146 req.buffer_ra = cpu_to_be64(__pa(local)); in i2c_opal_smbus_xfer()
154 req.buffer_ra = cpu_to_be64(__pa(&data->block[1])); in i2c_opal_smbus_xfer()

12345678910>>...17