Lines Matching refs:PAGE_SIZE

117 		vma->vm_end = vma->vm_start + PAGE_SIZE;  in ia64_init_addr_space()
134 vma->vm_end = PAGE_SIZE; in ia64_init_addr_space()
196 for (; start < end; start += PAGE_SIZE) { in free_initrd_mem()
252 page = virt_to_page(ia64_imva(__start_gate_section + PAGE_SIZE)); in setup_gate()
253 put_kernel_page(page, GATE_ADDR + PAGE_SIZE, PAGE_GATE); in setup_gate()
260 for (addr = GATE_ADDR + PAGE_SIZE; in setup_gate()
262 addr += PAGE_SIZE) in setup_gate()
420 end_address += PAGE_SIZE; in vmemmap_find_next_valid_pfn()
455 for (address = start_page; address < end_page; address += PAGE_SIZE) { in create_mem_map_page_table()
458 p4d = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node); in create_mem_map_page_table()
466 pud = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node); in create_mem_map_page_table()
474 pmd = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node); in create_mem_map_page_table()
482 pte = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node); in create_mem_map_page_table()
490 void *page = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, in create_mem_map_page_table()
502 __func__, PAGE_SIZE, PAGE_SIZE, node); in create_mem_map_page_table()
533 map_start -= ((unsigned long) map_start & (PAGE_SIZE - 1)) / sizeof(struct page); in virtual_memmap_init()
649 BUG_ON(PTRS_PER_PGD * sizeof(pgd_t) != PAGE_SIZE); in mem_init()
650 BUG_ON(PTRS_PER_PMD * sizeof(pmd_t) != PAGE_SIZE); in mem_init()
651 BUG_ON(PTRS_PER_PTE * sizeof(pte_t) != PAGE_SIZE); in mem_init()
671 high_memory = __va(max_low_pfn * PAGE_SIZE); in mem_init()