Lines Matching +full:0 +full:x2c000

65 	for (i = 0; i < THREADS__TABLE_SIZE; i++) {  in machine__threads_init()
69 threads->nr = 0; in machine__threads_init()
98 machine->pid) < 0) in machine__set_mmap_name()
101 return machine->mmap_name ? 0 : -ENOMEM; in machine__set_mmap_name()
109 thread__set_comm(thread, comm, 0); in thread__set_guest_comm()
116 memset(machine, 0, sizeof(*machine)); in machine__init()
131 machine->id_hdr_size = 0; in machine__init()
134 machine->kernel_start = 0; in machine__init()
156 err = 0; in machine__init()
164 return 0; in machine__init()
174 if (machine__create_kernel_maps(machine) < 0) in machine__new_host()
193 if (machine && machine__load_kallsyms(machine, "/proc/kallsyms") <= 0) { in machine__new_kallsyms()
228 for (i = 0; i < THREADS__TABLE_SIZE; i++) { in machine__delete_threads()
259 for (i = 0; i < THREADS__TABLE_SIZE; i++) { in machine__exit()
297 if (machine__init(machine, root_dir, pid) != 0) { in machines__add()
721 * Threads are identified by pid and tid, and the idle task has pid == tid == 0.
729 struct thread *thread = machine__findnew_thread(machine, 0, 0); in machine__idle_thread()
731 if (!thread || thread__set_comm(thread, "swapper", 0) || in machine__idle_thread()
732 thread__set_namespaces(thread, 0, NULL)) in machine__idle_thread()
754 int err = 0; in machine__process_comm_event()
780 int err = 0; in machine__process_namespaces_event()
817 return 0; in machine__process_cgroup_event()
825 return 0; in machine__process_lost_event()
833 return 0; in machine__process_lost_samples_event()
866 return 0; in machine__process_aux_event()
874 return 0; in machine__process_itrace_start_event()
882 return 0; in machine__process_aux_output_hw_id_event()
890 return 0; in machine__process_switch_event()
901 int err = 0; in machine__process_ksymbol_register()
911 map = map__new2(0, dso); in machine__process_ksymbol_register()
949 0, 0, event->ksymbol.name); in machine__process_ksymbol_register()
970 return 0; in machine__process_ksymbol_unregister()
982 return 0; in machine__process_ksymbol_unregister()
1009 return 0; in machine__process_text_poke()
1013 return 0; in machine__process_text_poke()
1037 return 0; in machine__process_text_poke()
1107 size_t printed = 0; in machine__fprintf_vmlinux_path()
1114 printed += fprintf(fp, "[0] %s\n", filename); in machine__fprintf_vmlinux_path()
1117 for (i = 0; i < vmlinux_path__nr_entries; ++i) in machine__fprintf_vmlinux_path()
1130 for (i = 0; i < THREADS__TABLE_SIZE; i++) { in machine__fprintf()
1197 u64 addr = 0; in machine__get_running_kernel_start()
1202 return 0; in machine__get_running_kernel_start()
1204 for (i = 0; (name = ref_reloc_sym_names[i]) != NULL; i++) { in machine__get_running_kernel_start()
1224 return 0; in machine__get_running_kernel_start()
1272 for (i = 0; i < ARRAY_SIZE(syms); i++) { in find_entry_trampoline()
1278 return 0; in find_entry_trampoline()
1285 #define X86_64_CPU_ENTRY_AREA_PER_CPU 0xfffffe0000000000ULL
1286 #define X86_64_CPU_ENTRY_AREA_SIZE 0x2c000
1287 #define X86_64_ENTRY_TRAMPOLINE 0x6000
1316 return 0; in machine__map_x86_64_entry_trampolines()
1320 return 0; in machine__map_x86_64_entry_trampolines()
1325 for (cpu = 0; cpu < nr_cpus_avail; cpu++) { in machine__map_x86_64_entry_trampolines()
1337 if (machine__create_extra_kernel_map(machine, kernel, &xm) < 0) in machine__map_x86_64_entry_trampolines()
1343 return 0; in machine__map_x86_64_entry_trampolines()
1349 return 0; in machine__create_extra_kernel_maps()
1359 machine->vmlinux_map = map__new2(0, kernel); in __machine__create_kernel_maps()
1388 int ret = 0; in machines__create_guest_kernel_maps()
1390 int i, items = 0; in machines__create_guest_kernel_maps()
1403 if (items <= 0) in machines__create_guest_kernel_maps()
1405 for (i = 0; i < items; i++) { in machines__create_guest_kernel_maps()
1406 if (!isdigit(namelist[i]->d_name[0])) { in machines__create_guest_kernel_maps()
1411 if ((*endp != '\0') || in machines__create_guest_kernel_maps()
1466 if (ret > 0) { in machine__load_kallsyms()
1485 if (ret > 0) in machine__load_vmlinux_path()
1514 *tmp = '\0'; in get_kernel_version()
1532 return 0; in maps__set_module_path()
1551 return 0; in maps__set_module_path()
1558 int ret = 0; in maps__set_modules_path_dir()
1580 if (depth == 0) { in maps__set_modules_path_dir()
1587 if (ret < 0) in maps__set_modules_path_dir()
1624 return maps__set_modules_path_dir(machine__kernel_maps(machine), modules_path, 0); in machine__set_modules_path()
1630 return 0; in arch__fix_module_text_start()
1639 if (arch__fix_module_text_start(&start, &size, name) < 0) in machine__create_module()
1649 return 0; in machine__create_module()
1671 return 0; in machine__create_modules()
1675 return 0; in machine__create_modules()
1687 if (start == 0 && end == 0) in machine__set_kernel_mmap()
1688 map__set_end(machine->vmlinux_map, ~0ULL); in machine__set_kernel_mmap()
1713 u64 start = 0, end = ~0ULL; in machine__create_kernel_maps()
1720 if (ret < 0) in machine__create_kernel_maps()
1723 if (symbol_conf.use_modules && machine__create_modules(machine) < 0) { in machine__create_kernel_maps()
1745 if (ret < 0) in machine__create_kernel_maps()
1752 if (end == ~0ULL) { in machine__create_kernel_maps()
1807 return 0; in machine__process_kernel_mmap_event()
1814 is_kernel_mmap = memcmp(xm->name, mmap_name, strlen(mmap_name) - 1) == 0; in machine__process_kernel_mmap_event()
1822 is_kernel_mmap = memcmp(xm->name, mmap_name, strlen(mmap_name) - 1) == 0; in machine__process_kernel_mmap_event()
1824 if (xm->name[0] == '/' || in machine__process_kernel_mmap_event()
1825 (!is_kernel_mmap && xm->name[0] == '[')) { in machine__process_kernel_mmap_event()
1884 if (__machine__create_kernel_maps(machine, kernel) < 0) { in machine__process_kernel_mmap_event()
1892 if (machine__update_kernel_mmap(machine, xm->start, xm->end) < 0) { in machine__process_kernel_mmap_event()
1905 if (xm->pgoff != 0) { in machine__process_kernel_mmap_event()
1921 return 0; in machine__process_kernel_mmap_event()
1939 int ret = 0; in machine__process_mmap2_event()
1959 if (ret < 0) in machine__process_mmap2_event()
1961 return 0; in machine__process_mmap2_event()
1984 return 0; in machine__process_mmap2_event()
1992 return 0; in machine__process_mmap2_event()
2000 u32 prot = 0; in machine__process_mmap_event()
2001 int ret = 0; in machine__process_mmap_event()
2016 if (ret < 0) in machine__process_mmap_event()
2018 return 0; in machine__process_mmap_event()
2031 NULL, prot, 0, NULL, event->mmap.filename, thread); in machine__process_mmap_event()
2042 return 0; in machine__process_mmap_event()
2050 return 0; in machine__process_mmap_event()
2067 BUG_ON(refcount_read(thread__refcnt(th)) == 0); in __machine__remove_thread()
2095 int err = 0; in machine__process_fork_event()
2141 thread__fork(thread, parent, sample->time, do_maps_clone) < 0) { in machine__process_fork_event()
2164 return 0; in machine__process_exit_event()
2216 if (!regexec(regex, sym->name, 0, NULL, 0)) in symbol__match_regex()
2243 ams->phys_addr = 0; in ip__resolve_ams()
2244 ams->data_page_size = 0; in ip__resolve_ams()
2327 int nr_loop_iter = 0, err = 0; in add_callchain_ip()
2328 u64 iter_cycles = 0; in add_callchain_ip()
2332 al.filtered = 0; in add_callchain_ip()
2411 for (i = 0; i < bs->nr; i++) { in sample__resolve_bstack()
2425 iter->cycles = 0; in save_iterations()
2427 for (i = 0; i < nr; i++) in save_iterations()
2433 #define NO_ENTRY 0xff
2448 for (i = 0; i < nr; i++) { in remove_loops()
2457 off = 0; in remove_loops()
2465 if (j > 0) { in remove_loops()
2496 for (i = 0; i < end + 1; i++) { in lbr_callchain_add_kernel_ip()
2503 return 0; in lbr_callchain_add_kernel_ip()
2506 for (i = end; i >= 0; i--) { in lbr_callchain_add_kernel_ip()
2514 return 0; in lbr_callchain_add_kernel_ip()
2568 for (i = 0; i < (int)(cursor->nr - 1); i++) in lbr_callchain_add_lbr_ip()
2575 ip = entries[0].to; in lbr_callchain_add_lbr_ip()
2576 flags = &entries[0].flags; in lbr_callchain_add_lbr_ip()
2577 *branch_from = entries[0].from; in lbr_callchain_add_lbr_ip()
2588 * But does not need to save current cursor node for entry 0. in lbr_callchain_add_lbr_ip()
2600 for (i = 0; i < lbr_nr; i++) { in lbr_callchain_add_lbr_ip()
2611 return 0; in lbr_callchain_add_lbr_ip()
2615 for (i = lbr_nr - 1; i >= 0; i--) { in lbr_callchain_add_lbr_ip()
2628 ip = entries[0].to; in lbr_callchain_add_lbr_ip()
2629 flags = &entries[0].flags; in lbr_callchain_add_lbr_ip()
2630 *branch_from = entries[0].from; in lbr_callchain_add_lbr_ip()
2638 return 0; in lbr_callchain_add_lbr_ip()
2663 return 0; in lbr_callchain_add_stitched_lbr_ip()
2693 int i, j, nr_identical_branches = 0; in has_stitched_lbr()
2717 for (i = distance, j = cur_stack->nr - 1; (i >= 0) && (j >= 0); i--, j--) { in has_stitched_lbr()
2786 * 0 no available LBR callchain information, should try fp
2802 u64 branch_from = 0; in resolve_lbr_callchain_sample()
2805 for (i = 0; i < chain_nr; i++) { in resolve_lbr_callchain_sample()
2812 return 0; in resolve_lbr_callchain_sample()
2815 (max_lbr > 0) && alloc_lbr_stitch(thread, max_lbr)) { in resolve_lbr_callchain_sample()
2869 return (err < 0) ? err : 0; in resolve_lbr_callchain_sample()
2878 int err = 0; in find_prev_cpumode()
2880 while (--ent >= 0) { in find_prev_cpumode()
2886 false, NULL, NULL, 0); in find_prev_cpumode()
2899 return 0; in get_leaf_frame_caller()
2913 int chain_nr = 0; in thread__resolve_callchain_sample()
2917 int first_call = 0; in thread__resolve_callchain_sample()
2928 !env ? 0 : env->max_branches); in thread__resolve_callchain_sample()
2930 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
2961 for (i = 0; i < nr; i++) { in thread__resolve_callchain_sample()
2985 memset(iter, 0, sizeof(struct iterations) * nr); in thread__resolve_callchain_sample()
2988 for (i = 0; i < nr; i++) { in thread__resolve_callchain_sample()
2999 &iter[i], 0); in thread__resolve_callchain_sample()
3006 if (chain_nr == 0) in thread__resolve_callchain_sample()
3007 return 0; in thread__resolve_callchain_sample()
3017 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
3019 for (i = first_call, nr_entries = 0; in thread__resolve_callchain_sample()
3039 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
3052 if (usr_idx >= 0 && chain->ips[usr_idx] == PERF_CONTEXT_USER) { in thread__resolve_callchain_sample()
3065 false, NULL, NULL, 0); in thread__resolve_callchain_sample()
3067 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
3073 false, NULL, NULL, 0); in thread__resolve_callchain_sample()
3076 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
3079 return 0; in thread__resolve_callchain_sample()
3115 NULL, 0, 0, 0, ilist->srcline); in append_inlines()
3117 if (ret != 0) in append_inlines()
3133 return 0; in unwind_entry()
3135 if (append_inlines(cursor, &entry->ms, entry->ip) == 0) in unwind_entry()
3136 return 0; in unwind_entry()
3147 false, NULL, 0, 0, 0, srcline); in unwind_entry()
3159 return 0; in thread__resolve_callchain_unwind()
3164 return 0; in thread__resolve_callchain_unwind()
3178 int ret = 0; in thread__resolve_callchain()
3216 int rc = 0; in machine__for_each_thread()
3219 for (i = 0; i < THREADS__TABLE_SIZE; i++) { in machine__for_each_thread()
3226 if (rc != 0) in machine__for_each_thread()
3238 int rc = 0; in machines__for_each_thread()
3241 if (rc != 0) in machines__for_each_thread()
3248 if (rc != 0) in machines__for_each_thread()
3256 if (cpu < 0 || (size_t)cpu >= machine->current_tid_sz) in machine__get_current_tid()
3268 if (cpu < 0) in machine__set_current_tid()
3286 return 0; in machine__set_current_tid()
3305 return machine ? perf_env__nr_cpus_avail(machine->env) : 0; in machine__nr_cpus_avail()
3311 int err = 0; in machine__get_kernel_start()
3389 int err = 0; in machine__for_each_dso()
3402 int err = 0; in machine__for_each_kernel_map()
3406 if (err != 0) { in machine__for_each_kernel_map()