Searched refs:kernel_map (Results 1 – 9 of 9) sorted by relevance
36 #define PAGE_OFFSET kernel_map.page_offset116 extern struct kernel_mapping kernel_map;120 ((x) >= kernel_map.virt_addr && (x) < (kernel_map.virt_addr + kernel_map.size))125 #define linear_mapping_pa_to_va(x) ((void *)((unsigned long)(x) + kernel_map.va_pa_offset))129 (void *)((unsigned long)(_y) + kernel_map.va_kernel_xip_pa_offset) : \130 (void *)((unsigned long)(_y) + kernel_map.va_kernel_pa_offset + XIP_OFFSET); \134 #define linear_mapping_va_to_pa(x) ((unsigned long)(x) - kernel_map.va_pa_offset)137 (IS_ENABLED(CONFIG_XIP_KERNEL) && _y < kernel_map.virt_addr + XIP_OFFSET) ? \138 ((unsigned long)(_y) - kernel_map.va_kernel_xip_pa_offset) : \139 ((unsigned long)(_y) - kernel_map.va_kernel_pa_offset - XIP_OFFSET); \
34 struct kernel_mapping kernel_map __ro_after_init;35 EXPORT_SYMBOL(kernel_map);37 #define kernel_map (*(struct kernel_mapping *)XIP_FIXUP(&kernel_map)) macro433 BUG_ON((va - kernel_map.virt_addr) >> PUD_SHIFT); in alloc_pmd_early()499 BUG_ON((va - kernel_map.virt_addr) >> PGDIR_SHIFT); in alloc_pud_early()537 BUG_ON((va - kernel_map.virt_addr) >> PGDIR_SHIFT); in alloc_p4d_early()737 kernel_map.page_offset = PAGE_OFFSET_L4; in disable_pgtable_l5()744 kernel_map.page_offset = PAGE_OFFSET_L3; in disable_pgtable_l4()829 end_va = kernel_map.virt_addr + kernel_map.xiprom_sz; in create_kernel_page_table()830 for (va = kernel_map.virt_addr; va < end_va; va += PMD_SIZE) in create_kernel_page_table()[all …]
24 unsigned long kernel_start = kernel_map.virt_addr; in __phys_addr_symbol()
382 address_markers[KERNEL_MAPPING_NR].start_address = kernel_map.virt_addr; in ptdump_init()
481 struct map *kernel_map = machine__kernel_map(kmap->kmaps->machine); in map__rip_2objdump() local483 if (kernel_map) in map__rip_2objdump()484 map = kernel_map; in map__rip_2objdump()
595 struct map *kernel_map = machine__kernel_map(&rep->session->machines.host); in report__warn_kptr_restrict() local596 struct kmap *kernel_kmap = kernel_map ? map__kmap(kernel_map) : NULL; in report__warn_kptr_restrict()601 if (kernel_map == NULL || in report__warn_kptr_restrict()602 (kernel_map->dso->hit && in report__warn_kptr_restrict()609 if (kernel_map && map__has_symbols(kernel_map)) { in report__warn_kptr_restrict()
333 struct map *kernel_map; in build_alloc_func_list() local350 kernel_map = machine__kernel_map(machine); in build_alloc_func_list()351 if (map__load(kernel_map) < 0) { in build_alloc_func_list()356 map__for_each_symbol(kernel_map, sym, node) { in build_alloc_func_list()
237 this_hart_id, kernel_map.va_pa_offset); in machine_kexec()
76 la a1, kernel_map