Lines Matching refs:image

143 static struct page *kimage_alloc_page(struct kimage *image,
147 int sanity_check_segment_list(struct kimage *image) in sanity_check_segment_list() argument
150 unsigned long nr_segments = image->nr_segments; in sanity_check_segment_list()
170 mstart = image->segment[i].mem; in sanity_check_segment_list()
171 mend = mstart + image->segment[i].memsz; in sanity_check_segment_list()
189 mstart = image->segment[i].mem; in sanity_check_segment_list()
190 mend = mstart + image->segment[i].memsz; in sanity_check_segment_list()
194 pstart = image->segment[j].mem; in sanity_check_segment_list()
195 pend = pstart + image->segment[j].memsz; in sanity_check_segment_list()
208 if (image->segment[i].bufsz > image->segment[i].memsz) in sanity_check_segment_list()
218 if (PAGE_COUNT(image->segment[i].memsz) > nr_pages / 2) in sanity_check_segment_list()
221 total_pages += PAGE_COUNT(image->segment[i].memsz); in sanity_check_segment_list()
237 if (image->type == KEXEC_TYPE_CRASH) { in sanity_check_segment_list()
241 mstart = image->segment[i].mem; in sanity_check_segment_list()
242 mend = mstart + image->segment[i].memsz - 1; in sanity_check_segment_list()
255 struct kimage *image; in do_kimage_alloc_init() local
258 image = kzalloc(sizeof(*image), GFP_KERNEL); in do_kimage_alloc_init()
259 if (!image) in do_kimage_alloc_init()
262 image->head = 0; in do_kimage_alloc_init()
263 image->entry = &image->head; in do_kimage_alloc_init()
264 image->last_entry = &image->head; in do_kimage_alloc_init()
265 image->control_page = ~0; /* By default this does not apply */ in do_kimage_alloc_init()
266 image->type = KEXEC_TYPE_DEFAULT; in do_kimage_alloc_init()
269 INIT_LIST_HEAD(&image->control_pages); in do_kimage_alloc_init()
272 INIT_LIST_HEAD(&image->dest_pages); in do_kimage_alloc_init()
275 INIT_LIST_HEAD(&image->unusable_pages); in do_kimage_alloc_init()
278 image->hp_action = KEXEC_CRASH_HP_NONE; in do_kimage_alloc_init()
279 image->elfcorehdr_index = -1; in do_kimage_alloc_init()
280 image->elfcorehdr_updated = false; in do_kimage_alloc_init()
283 return image; in do_kimage_alloc_init()
286 int kimage_is_destination_range(struct kimage *image, in kimage_is_destination_range() argument
292 for (i = 0; i < image->nr_segments; i++) { in kimage_is_destination_range()
295 mstart = image->segment[i].mem; in kimage_is_destination_range()
296 mend = mstart + image->segment[i].memsz; in kimage_is_destination_range()
355 static struct page *kimage_alloc_normal_control_pages(struct kimage *image, in kimage_alloc_normal_control_pages() argument
392 kimage_is_destination_range(image, addr, eaddr)) { in kimage_alloc_normal_control_pages()
400 list_add(&pages->lru, &image->control_pages); in kimage_alloc_normal_control_pages()
421 static struct page *kimage_alloc_crash_control_pages(struct kimage *image, in kimage_alloc_crash_control_pages() argument
450 hole_start = (image->control_page + (size - 1)) & ~(size - 1); in kimage_alloc_crash_control_pages()
460 for (i = 0; i < image->nr_segments; i++) { in kimage_alloc_crash_control_pages()
463 mstart = image->segment[i].mem; in kimage_alloc_crash_control_pages()
464 mend = mstart + image->segment[i].memsz - 1; in kimage_alloc_crash_control_pages()
473 if (i == image->nr_segments) { in kimage_alloc_crash_control_pages()
475 image->control_page = hole_end; in kimage_alloc_crash_control_pages()
488 struct page *kimage_alloc_control_pages(struct kimage *image, in kimage_alloc_control_pages() argument
493 switch (image->type) { in kimage_alloc_control_pages()
495 pages = kimage_alloc_normal_control_pages(image, order); in kimage_alloc_control_pages()
498 pages = kimage_alloc_crash_control_pages(image, order); in kimage_alloc_control_pages()
505 int kimage_crash_copy_vmcoreinfo(struct kimage *image) in kimage_crash_copy_vmcoreinfo() argument
510 if (image->type != KEXEC_TYPE_CRASH) in kimage_crash_copy_vmcoreinfo()
522 vmcoreinfo_page = kimage_alloc_control_pages(image, 0); in kimage_crash_copy_vmcoreinfo()
533 image->vmcoreinfo_data_copy = safecopy; in kimage_crash_copy_vmcoreinfo()
539 static int kimage_add_entry(struct kimage *image, kimage_entry_t entry) in kimage_add_entry() argument
541 if (*image->entry != 0) in kimage_add_entry()
542 image->entry++; in kimage_add_entry()
544 if (image->entry == image->last_entry) { in kimage_add_entry()
548 page = kimage_alloc_page(image, GFP_KERNEL, KIMAGE_NO_DEST); in kimage_add_entry()
553 *image->entry = virt_to_boot_phys(ind_page) | IND_INDIRECTION; in kimage_add_entry()
554 image->entry = ind_page; in kimage_add_entry()
555 image->last_entry = ind_page + in kimage_add_entry()
558 *image->entry = entry; in kimage_add_entry()
559 image->entry++; in kimage_add_entry()
560 *image->entry = 0; in kimage_add_entry()
565 static int kimage_set_destination(struct kimage *image, in kimage_set_destination() argument
570 return kimage_add_entry(image, destination | IND_DESTINATION); in kimage_set_destination()
574 static int kimage_add_page(struct kimage *image, unsigned long page) in kimage_add_page() argument
578 return kimage_add_entry(image, page | IND_SOURCE); in kimage_add_page()
582 static void kimage_free_extra_pages(struct kimage *image) in kimage_free_extra_pages() argument
585 kimage_free_page_list(&image->dest_pages); in kimage_free_extra_pages()
588 kimage_free_page_list(&image->unusable_pages); in kimage_free_extra_pages()
592 void kimage_terminate(struct kimage *image) in kimage_terminate() argument
594 if (*image->entry != 0) in kimage_terminate()
595 image->entry++; in kimage_terminate()
597 *image->entry = IND_DONE; in kimage_terminate()
600 #define for_each_kimage_entry(image, ptr, entry) \ argument
601 for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); \
613 void kimage_free(struct kimage *image) in kimage_free() argument
618 if (!image) in kimage_free()
621 if (image->vmcoreinfo_data_copy) { in kimage_free()
623 vunmap(image->vmcoreinfo_data_copy); in kimage_free()
626 kimage_free_extra_pages(image); in kimage_free()
627 for_each_kimage_entry(image, ptr, entry) { in kimage_free()
644 machine_kexec_cleanup(image); in kimage_free()
647 kimage_free_page_list(&image->control_pages); in kimage_free()
653 if (image->file_mode) in kimage_free()
654 kimage_file_post_load_cleanup(image); in kimage_free()
656 kfree(image); in kimage_free()
659 static kimage_entry_t *kimage_dst_used(struct kimage *image, in kimage_dst_used() argument
665 for_each_kimage_entry(image, ptr, entry) { in kimage_dst_used()
678 static struct page *kimage_alloc_page(struct kimage *image, in kimage_alloc_page() argument
707 list_for_each_entry(page, &image->dest_pages, lru) { in kimage_alloc_page()
725 list_add(&page->lru, &image->unusable_pages); in kimage_alloc_page()
735 if (!kimage_is_destination_range(image, addr, in kimage_alloc_page()
744 old = kimage_dst_used(image, addr); in kimage_alloc_page()
768 list_add(&page->lru, &image->dest_pages); in kimage_alloc_page()
774 static int kimage_load_normal_segment(struct kimage *image, in kimage_load_normal_segment() argument
783 if (image->file_mode) in kimage_load_normal_segment()
791 result = kimage_set_destination(image, maddr); in kimage_load_normal_segment()
800 page = kimage_alloc_page(image, GFP_HIGHUSER, maddr); in kimage_load_normal_segment()
805 result = kimage_add_page(image, page_to_boot_pfn(page) in kimage_load_normal_segment()
819 if (image->file_mode) in kimage_load_normal_segment()
830 if (image->file_mode) in kimage_load_normal_segment()
842 static int kimage_load_crash_segment(struct kimage *image, in kimage_load_crash_segment() argument
856 if (image->file_mode) in kimage_load_crash_segment()
885 if (image->file_mode) in kimage_load_crash_segment()
898 if (image->file_mode) in kimage_load_crash_segment()
910 int kimage_load_segment(struct kimage *image, in kimage_load_segment() argument
915 switch (image->type) { in kimage_load_segment()
917 result = kimage_load_normal_segment(image, segment); in kimage_load_segment()
920 result = kimage_load_crash_segment(image, segment); in kimage_load_segment()