Lines Matching refs:pg

462 	struct ftrace_profile_page *pg;  in function_stat_next()  local
464 pg = (struct ftrace_profile_page *)((unsigned long)rec & PAGE_MASK); in function_stat_next()
470 if ((void *)rec >= (void *)&pg->records[pg->index]) { in function_stat_next()
471 pg = pg->next; in function_stat_next()
472 if (!pg) in function_stat_next()
474 rec = &pg->records[0]; in function_stat_next()
603 struct ftrace_profile_page *pg; in ftrace_profile_reset() local
605 pg = stat->pages = stat->start; in ftrace_profile_reset()
607 while (pg) { in ftrace_profile_reset()
608 memset(pg->records, 0, PROFILE_RECORDS_SIZE); in ftrace_profile_reset()
609 pg->index = 0; in ftrace_profile_reset()
610 pg = pg->next; in ftrace_profile_reset()
619 struct ftrace_profile_page *pg; in ftrace_profile_pages_init() local
645 pg = stat->start = stat->pages; in ftrace_profile_pages_init()
650 pg->next = (void *)get_zeroed_page(GFP_KERNEL); in ftrace_profile_pages_init()
651 if (!pg->next) in ftrace_profile_pages_init()
653 pg = pg->next; in ftrace_profile_pages_init()
659 pg = stat->start; in ftrace_profile_pages_init()
660 while (pg) { in ftrace_profile_pages_init()
661 unsigned long tmp = (unsigned long)pg; in ftrace_profile_pages_init()
663 pg = pg->next; in ftrace_profile_pages_init()
1537 #define do_for_each_ftrace_rec(pg, rec) \ argument
1538 for (pg = ftrace_pages_start; pg; pg = pg->next) { \
1540 for (_____i = 0; _____i < pg->index; _____i++) { \
1541 rec = &pg->records[_____i];
1574 struct ftrace_page *pg; in ftrace_location_range() local
1581 for (pg = ftrace_pages_start; pg; pg = pg->next) { in ftrace_location_range()
1582 if (end < pg->records[0].ip || in ftrace_location_range()
1583 start >= (pg->records[pg->index - 1].ip + MCOUNT_INSN_SIZE)) in ftrace_location_range()
1585 rec = bsearch(&key, pg->records, pg->index, in ftrace_location_range()
1655 struct ftrace_page *pg; in __ftrace_hash_rec_update() local
1693 do_for_each_ftrace_rec(pg, rec) { in __ftrace_hash_rec_update()
1865 struct ftrace_page *pg; in __ftrace_hash_update_ipmodify() local
1884 do_for_each_ftrace_rec(pg, rec) { in __ftrace_hash_update_ipmodify()
1910 do_for_each_ftrace_rec(pg, rec) { in __ftrace_hash_update_ipmodify()
2421 struct ftrace_page *pg; in ftrace_replace_code() local
2427 do_for_each_ftrace_rec(pg, rec) { in ftrace_replace_code()
2442 struct ftrace_page *pg; member
2464 iter->pg = ftrace_pages_start; in ftrace_rec_iter_start()
2468 while (iter->pg && !iter->pg->index) in ftrace_rec_iter_start()
2469 iter->pg = iter->pg->next; in ftrace_rec_iter_start()
2471 if (!iter->pg) in ftrace_rec_iter_start()
2487 if (iter->index >= iter->pg->index) { in ftrace_rec_iter_next()
2488 iter->pg = iter->pg->next; in ftrace_rec_iter_next()
2492 while (iter->pg && !iter->pg->index) in ftrace_rec_iter_next()
2493 iter->pg = iter->pg->next; in ftrace_rec_iter_next()
2496 if (!iter->pg) in ftrace_rec_iter_next()
2510 return &iter->pg->records[iter->index]; in ftrace_rec_iter_record()
2791 struct ftrace_page *pg; in ftrace_shutdown() local
2794 do_for_each_ftrace_rec(pg, rec) { in ftrace_shutdown()
2920 struct ftrace_page *pg; in ftrace_update_code() local
2943 for (pg = new_pgs; pg; pg = pg->next) { in ftrace_update_code()
2945 for (i = 0; i < pg->index; i++) { in ftrace_update_code()
2951 p = &pg->records[i]; in ftrace_update_code()
2974 static int ftrace_allocate_records(struct ftrace_page *pg, int count) in ftrace_allocate_records() argument
2992 pg->records = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in ftrace_allocate_records()
2994 if (!pg->records) { in ftrace_allocate_records()
3003 pg->size = cnt; in ftrace_allocate_records()
3015 struct ftrace_page *pg; in ftrace_allocate_pages() local
3022 start_pg = pg = kzalloc(sizeof(*pg), GFP_KERNEL); in ftrace_allocate_pages()
3023 if (!pg) in ftrace_allocate_pages()
3032 cnt = ftrace_allocate_records(pg, num_to_init); in ftrace_allocate_pages()
3040 pg->next = kzalloc(sizeof(*pg), GFP_KERNEL); in ftrace_allocate_pages()
3041 if (!pg->next) in ftrace_allocate_pages()
3044 pg = pg->next; in ftrace_allocate_pages()
3050 pg = start_pg; in ftrace_allocate_pages()
3051 while (pg) { in ftrace_allocate_pages()
3052 order = get_count_order(pg->size / ENTRIES_PER_PAGE); in ftrace_allocate_pages()
3053 free_pages((unsigned long)pg->records, order); in ftrace_allocate_pages()
3054 start_pg = pg->next; in ftrace_allocate_pages()
3055 kfree(pg); in ftrace_allocate_pages()
3056 pg = start_pg; in ftrace_allocate_pages()
3068 struct ftrace_page *pg; member
3290 if (iter->idx >= iter->pg->index) { in t_func_next()
3291 if (iter->pg->next) { in t_func_next()
3292 iter->pg = iter->pg->next; in t_func_next()
3297 rec = &iter->pg->records[iter->idx++]; in t_func_next()
3396 iter->pg = ftrace_pages_start; in t_start()
3504 iter->pg = ftrace_pages_start; in ftrace_avail_open()
3519 iter->pg = ftrace_pages_start; in ftrace_enabled_open()
3602 iter->pg = ftrace_pages_start; in ftrace_regex_open()
3758 struct ftrace_page *pg; in match_records() local
3785 do_for_each_ftrace_rec(pg, rec) { in match_records()
5334 struct ftrace_page *pg; in ftrace_graph_set_hash() local
5352 do_for_each_ftrace_rec(pg, rec) { in ftrace_graph_set_hash()
5506 struct ftrace_page *pg; in ftrace_process_locs() local
5551 pg = start_pg; in ftrace_process_locs()
5563 if (pg->index == pg->size) { in ftrace_process_locs()
5565 if (WARN_ON(!pg->next)) in ftrace_process_locs()
5567 pg = pg->next; in ftrace_process_locs()
5570 rec = &pg->records[pg->index++]; in ftrace_process_locs()
5575 WARN_ON(pg->next); in ftrace_process_locs()
5578 ftrace_pages = pg; in ftrace_process_locs()
5637 clear_mod_from_hash(struct ftrace_page *pg, struct ftrace_hash *hash) in clear_mod_from_hash() argument
5646 for (i = 0; i < pg->index; i++) { in clear_mod_from_hash()
5647 rec = &pg->records[i]; in clear_mod_from_hash()
5660 static void clear_mod_from_hashes(struct ftrace_page *pg) in clear_mod_from_hashes() argument
5669 clear_mod_from_hash(pg, tr->ops->func_hash->filter_hash); in clear_mod_from_hashes()
5670 clear_mod_from_hash(pg, tr->ops->func_hash->notrace_hash); in clear_mod_from_hashes()
5699 struct ftrace_page *pg; in ftrace_release_mod() local
5720 for (pg = ftrace_pages_start; pg; pg = *last_pg) { in ftrace_release_mod()
5721 rec = &pg->records[0]; in ftrace_release_mod()
5728 if (WARN_ON(pg == ftrace_pages_start)) in ftrace_release_mod()
5732 if (pg == ftrace_pages) in ftrace_release_mod()
5735 ftrace_update_tot_cnt -= pg->index; in ftrace_release_mod()
5736 *last_pg = pg->next; in ftrace_release_mod()
5738 pg->next = tmp_page; in ftrace_release_mod()
5739 tmp_page = pg; in ftrace_release_mod()
5741 last_pg = &pg->next; in ftrace_release_mod()
5746 for (pg = tmp_page; pg; pg = tmp_page) { in ftrace_release_mod()
5749 clear_mod_from_hashes(pg); in ftrace_release_mod()
5751 order = get_count_order(pg->size / ENTRIES_PER_PAGE); in ftrace_release_mod()
5752 free_pages((unsigned long)pg->records, order); in ftrace_release_mod()
5753 tmp_page = pg->next; in ftrace_release_mod()
5754 kfree(pg); in ftrace_release_mod()
5761 struct ftrace_page *pg; in ftrace_module_enable() local
5784 do_for_each_ftrace_rec(pg, rec) { in ftrace_module_enable()
6054 struct ftrace_page *pg; in ftrace_free_mem() local
6077 for (pg = ftrace_pages_start; pg; last_pg = &pg->next, pg = *last_pg) { in ftrace_free_mem()
6078 if (end < pg->records[0].ip || in ftrace_free_mem()
6079 start >= (pg->records[pg->index - 1].ip + MCOUNT_INSN_SIZE)) in ftrace_free_mem()
6082 rec = bsearch(&key, pg->records, pg->index, in ftrace_free_mem()
6094 pg->index--; in ftrace_free_mem()
6096 if (!pg->index) { in ftrace_free_mem()
6097 *last_pg = pg->next; in ftrace_free_mem()
6098 order = get_count_order(pg->size / ENTRIES_PER_PAGE); in ftrace_free_mem()
6099 free_pages((unsigned long)pg->records, order); in ftrace_free_mem()
6100 kfree(pg); in ftrace_free_mem()
6101 pg = container_of(last_pg, struct ftrace_page, next); in ftrace_free_mem()
6103 ftrace_pages = pg; in ftrace_free_mem()
6107 (pg->index - (rec - pg->records)) * sizeof(*rec)); in ftrace_free_mem()