Lines Matching refs:start
91 unsigned long long start; member
120 if (one->start + one->size <= two->start) in mem_overlaps()
123 if (one->start >= two->start + two->size) in mem_overlaps()
138 parse_memmap(char *p, unsigned long long *start, unsigned long long *size) in parse_memmap() argument
158 *start = memparse(p + 1, &p); in parse_memmap()
170 *start = 0; in parse_memmap()
186 unsigned long long start, size; in mem_avoid_memmap() local
192 rc = parse_memmap(str, &start, &size); in mem_avoid_memmap()
197 if (start == 0) { in mem_avoid_memmap()
205 mem_avoid[MEM_AVOID_MEMMAP_BEGIN + i].start = start; in mem_avoid_memmap()
384 mem_avoid[MEM_AVOID_ZO_RANGE].start = input; in mem_avoid_init()
386 add_identity_map(mem_avoid[MEM_AVOID_ZO_RANGE].start, in mem_avoid_init()
394 mem_avoid[MEM_AVOID_INITRD].start = initrd_start; in mem_avoid_init()
405 mem_avoid[MEM_AVOID_CMDLINE].start = cmd_line; in mem_avoid_init()
407 add_identity_map(mem_avoid[MEM_AVOID_CMDLINE].start, in mem_avoid_init()
411 mem_avoid[MEM_AVOID_BOOTPARAMS].start = (unsigned long)boot_params; in mem_avoid_init()
413 add_identity_map(mem_avoid[MEM_AVOID_BOOTPARAMS].start, in mem_avoid_init()
436 unsigned long earliest = img->start + img->size; in mem_avoid_overlap()
441 mem_avoid[i].start < earliest) { in mem_avoid_overlap()
443 earliest = overlap->start; in mem_avoid_overlap()
453 avoid.start = (unsigned long)ptr; in mem_avoid_overlap()
456 if (mem_overlaps(img, &avoid) && (avoid.start < earliest)) { in mem_avoid_overlap()
458 earliest = overlap->start; in mem_avoid_overlap()
488 slot_area.addr = region->start; in store_slot_info()
514 addr = ALIGN(region->start, PUD_SIZE); in process_gb_huge_pages()
516 if (addr < region->start + region->size) in process_gb_huge_pages()
517 size = region->size - (addr - region->start); in process_gb_huge_pages()
538 if (addr >= region->start + image_size) { in process_gb_huge_pages()
539 tmp.start = region->start; in process_gb_huge_pages()
540 tmp.size = addr - region->start; in process_gb_huge_pages()
544 size = region->size - (addr - region->start) - i * PUD_SIZE; in process_gb_huge_pages()
546 tmp.start = addr + i * PUD_SIZE; in process_gb_huge_pages()
586 if (IS_ENABLED(CONFIG_X86_32) && entry->start >= KERNEL_IMAGE_SIZE) in process_mem_region()
590 if (entry->start + entry->size < minimum) in process_mem_region()
594 end = min(entry->size + entry->start, mem_limit); in process_mem_region()
595 if (entry->start >= end) in process_mem_region()
597 cur_entry.start = entry->start; in process_mem_region()
598 cur_entry.size = end - entry->start; in process_mem_region()
600 region.start = cur_entry.start; in process_mem_region()
605 start_orig = region.start; in process_mem_region()
608 if (region.start < minimum) in process_mem_region()
609 region.start = minimum; in process_mem_region()
612 region.start = ALIGN(region.start, CONFIG_PHYSICAL_ALIGN); in process_mem_region()
615 if (region.start > cur_entry.start + cur_entry.size) in process_mem_region()
619 region.size -= region.start - start_orig; in process_mem_region()
623 region.start + region.size > KERNEL_IMAGE_SIZE) in process_mem_region()
624 region.size = KERNEL_IMAGE_SIZE - region.start; in process_mem_region()
637 if (overlap.start > region.start + image_size) { in process_mem_region()
640 beginning.start = region.start; in process_mem_region()
641 beginning.size = overlap.start - region.start; in process_mem_region()
646 if (overlap.start + overlap.size >= region.start + region.size) in process_mem_region()
650 region.size -= overlap.start - region.start + overlap.size; in process_mem_region()
651 region.start = overlap.start + overlap.size; in process_mem_region()
718 region.start = md->phys_addr; in process_efi_entries()
749 region.start = entry->addr; in process_e820_entries()