Lines Matching +full:0 +full:x2c000

62 	for (i = 0; i < THREADS__TABLE_SIZE; i++) {  in machine__threads_init()
66 threads->nr = 0; in machine__threads_init()
79 machine->pid) < 0) in machine__set_mmap_name()
82 return machine->mmap_name ? 0 : -ENOMEM; in machine__set_mmap_name()
89 memset(machine, 0, sizeof(*machine)); in machine__init()
101 machine->id_hdr_size = 0; in machine__init()
104 machine->kernel_start = 0; in machine__init()
123 thread__set_comm(thread, comm, 0); in machine__init()
128 err = 0; in machine__init()
135 return 0; in machine__init()
145 if (machine__create_kernel_maps(machine) < 0) in machine__new_host()
164 if (machine && machine__load_kallsyms(machine, "/proc/kallsyms") <= 0) { in machine__new_kallsyms()
199 for (i = 0; i < THREADS__TABLE_SIZE; i++) { in machine__delete_threads()
228 for (i = 0; i < THREADS__TABLE_SIZE; i++) { in machine__exit()
277 if (machine__init(machine, root_dir, pid) != 0) { in machines__add()
602 * Threads are identified by pid and tid, and the idle task has pid == tid == 0.
610 struct thread *thread = machine__findnew_thread(machine, 0, 0); in machine__idle_thread()
612 if (!thread || thread__set_comm(thread, "swapper", 0) || in machine__idle_thread()
613 thread__set_namespaces(thread, 0, NULL)) in machine__idle_thread()
635 int err = 0; in machine__process_comm_event()
661 int err = 0; in machine__process_namespaces_event()
698 return 0; in machine__process_cgroup_event()
706 return 0; in machine__process_lost_event()
714 return 0; in machine__process_lost_samples_event()
747 return 0; in machine__process_aux_event()
755 return 0; in machine__process_itrace_start_event()
763 return 0; in machine__process_switch_event()
778 map = map__new2(0, dso); in machine__process_ksymbol_register()
806 0, 0, event->ksymbol.name); in machine__process_ksymbol_register()
810 return 0; in machine__process_ksymbol_register()
822 return 0; in machine__process_ksymbol_unregister()
832 return 0; in machine__process_ksymbol_unregister()
858 return 0; in machine__process_text_poke()
862 return 0; in machine__process_text_poke()
886 return 0; in machine__process_text_poke()
953 size_t printed = 0; in machine__fprintf_vmlinux_path()
960 printed += fprintf(fp, "[0] %s\n", filename); in machine__fprintf_vmlinux_path()
963 for (i = 0; i < vmlinux_path__nr_entries; ++i) in machine__fprintf_vmlinux_path()
976 for (i = 0; i < THREADS__TABLE_SIZE; i++) { in machine__fprintf()
1047 u64 addr = 0; in machine__get_running_kernel_start()
1052 return 0; in machine__get_running_kernel_start()
1054 for (i = 0; (name = ref_reloc_sym_names[i]) != NULL; i++) { in machine__get_running_kernel_start()
1072 return 0; in machine__get_running_kernel_start()
1100 return 0; in machine__create_extra_kernel_map()
1117 for (i = 0; i < ARRAY_SIZE(syms); i++) { in find_entry_trampoline()
1123 return 0; in find_entry_trampoline()
1130 #define X86_64_CPU_ENTRY_AREA_PER_CPU 0xfffffe0000000000ULL
1131 #define X86_64_CPU_ENTRY_AREA_SIZE 0x2c000
1132 #define X86_64_ENTRY_TRAMPOLINE 0x6000
1161 return 0; in machine__map_x86_64_entry_trampolines()
1165 return 0; in machine__map_x86_64_entry_trampolines()
1170 for (cpu = 0; cpu < nr_cpus_avail; cpu++) { in machine__map_x86_64_entry_trampolines()
1182 if (machine__create_extra_kernel_map(machine, kernel, &xm) < 0) in machine__map_x86_64_entry_trampolines()
1188 return 0; in machine__map_x86_64_entry_trampolines()
1194 return 0; in machine__create_extra_kernel_maps()
1203 machine->vmlinux_map = map__new2(0, kernel); in __machine__create_kernel_maps()
1209 return 0; in __machine__create_kernel_maps()
1232 int ret = 0; in machines__create_guest_kernel_maps()
1234 int i, items = 0; in machines__create_guest_kernel_maps()
1247 if (items <= 0) in machines__create_guest_kernel_maps()
1249 for (i = 0; i < items; i++) { in machines__create_guest_kernel_maps()
1250 if (!isdigit(namelist[i]->d_name[0])) { in machines__create_guest_kernel_maps()
1255 if ((*endp != '\0') || in machines__create_guest_kernel_maps()
1309 if (ret > 0) { in machine__load_kallsyms()
1327 if (ret > 0) in machine__load_vmlinux_path()
1356 *tmp = '\0'; in get_kernel_version()
1373 return 0; in maps__set_module_path()
1391 return 0; in maps__set_module_path()
1398 int ret = 0; in maps__set_modules_path_dir()
1420 if (depth == 0) { in maps__set_modules_path_dir()
1427 if (ret < 0) in maps__set_modules_path_dir()
1464 return maps__set_modules_path_dir(&machine->kmaps, modules_path, 0); in machine__set_modules_path()
1470 return 0; in arch__fix_module_text_start()
1479 if (arch__fix_module_text_start(&start, &size, name) < 0) in machine__create_module()
1489 return 0; in machine__create_module()
1511 return 0; in machine__create_modules()
1515 return 0; in machine__create_modules()
1527 if (start == 0 && end == 0) in machine__set_kernel_mmap()
1528 machine->vmlinux_map->end = ~0ULL; in machine__set_kernel_mmap()
1550 u64 start = 0, end = ~0ULL; in machine__create_kernel_maps()
1557 if (ret < 0) in machine__create_kernel_maps()
1560 if (symbol_conf.use_modules && machine__create_modules(machine) < 0) { in machine__create_kernel_maps()
1587 if (end == ~0ULL) { in machine__create_kernel_maps()
1639 return 0; in machine__process_kernel_mmap_event()
1647 strlen(machine->mmap_name) - 1) == 0; in machine__process_kernel_mmap_event()
1648 if (xm->name[0] == '/' || in machine__process_kernel_mmap_event()
1649 (!is_kernel_mmap && xm->name[0] == '[')) { in machine__process_kernel_mmap_event()
1707 if (__machine__create_kernel_maps(machine, kernel) < 0) { in machine__process_kernel_mmap_event()
1725 if (xm->pgoff != 0) { in machine__process_kernel_mmap_event()
1740 return 0; in machine__process_kernel_mmap_event()
1758 int ret = 0; in machine__process_mmap2_event()
1778 if (ret < 0) in machine__process_mmap2_event()
1780 return 0; in machine__process_mmap2_event()
1803 return 0; in machine__process_mmap2_event()
1811 return 0; in machine__process_mmap2_event()
1819 u32 prot = 0; in machine__process_mmap_event()
1820 int ret = 0; in machine__process_mmap_event()
1835 if (ret < 0) in machine__process_mmap_event()
1837 return 0; in machine__process_mmap_event()
1850 NULL, prot, 0, NULL, event->mmap.filename, thread); in machine__process_mmap_event()
1861 return 0; in machine__process_mmap_event()
1869 return 0; in machine__process_mmap_event()
1882 BUG_ON(refcount_read(&th->refcnt) == 0); in __machine__remove_thread()
1920 int err = 0; in machine__process_fork_event()
1966 thread__fork(thread, parent, sample->time, do_maps_clone) < 0) { in machine__process_fork_event()
1991 return 0; in machine__process_exit_event()
2041 if (!regexec(regex, sym->name, 0, NULL, 0)) in symbol__match_regex()
2052 memset(&al, 0, sizeof(al)); in ip__resolve_ams()
2067 ams->phys_addr = 0; in ip__resolve_ams()
2068 ams->data_page_size = 0; in ip__resolve_ams()
2077 memset(&al, 0, sizeof(al)); in ip__resolve_data()
2146 int nr_loop_iter = 0; in add_callchain_ip()
2147 u64 iter_cycles = 0; in add_callchain_ip()
2150 al.filtered = 0; in add_callchain_ip()
2177 return 0; in add_callchain_ip()
2196 return 0; in add_callchain_ip()
2223 for (i = 0; i < bs->nr; i++) { in sample__resolve_bstack()
2237 iter->cycles = 0; in save_iterations()
2239 for (i = 0; i < nr; i++) in save_iterations()
2245 #define NO_ENTRY 0xff
2260 for (i = 0; i < nr; i++) { in remove_loops()
2269 off = 0; in remove_loops()
2277 if (j > 0) { in remove_loops()
2308 for (i = 0; i < end + 1; i++) { in lbr_callchain_add_kernel_ip()
2315 return 0; in lbr_callchain_add_kernel_ip()
2318 for (i = end; i >= 0; i--) { in lbr_callchain_add_kernel_ip()
2326 return 0; in lbr_callchain_add_kernel_ip()
2380 for (i = 0; i < (int)(cursor->nr - 1); i++) in lbr_callchain_add_lbr_ip()
2387 ip = entries[0].to; in lbr_callchain_add_lbr_ip()
2388 flags = &entries[0].flags; in lbr_callchain_add_lbr_ip()
2389 *branch_from = entries[0].from; in lbr_callchain_add_lbr_ip()
2400 * But does not need to save current cursor node for entry 0. in lbr_callchain_add_lbr_ip()
2412 for (i = 0; i < lbr_nr; i++) { in lbr_callchain_add_lbr_ip()
2423 return 0; in lbr_callchain_add_lbr_ip()
2427 for (i = lbr_nr - 1; i >= 0; i--) { in lbr_callchain_add_lbr_ip()
2440 ip = entries[0].to; in lbr_callchain_add_lbr_ip()
2441 flags = &entries[0].flags; in lbr_callchain_add_lbr_ip()
2442 *branch_from = entries[0].from; in lbr_callchain_add_lbr_ip()
2450 return 0; in lbr_callchain_add_lbr_ip()
2475 return 0; in lbr_callchain_add_stitched_lbr_ip()
2505 int i, j, nr_identical_branches = 0; in has_stitched_lbr()
2529 for (i = distance, j = cur_stack->nr - 1; (i >= 0) && (j >= 0); i--, j--) { in has_stitched_lbr()
2596 * 0 no available LBR callchain information, should try fp
2612 u64 branch_from = 0; in resolve_lbr_callchain_sample()
2615 for (i = 0; i < chain_nr; i++) { in resolve_lbr_callchain_sample()
2622 return 0; in resolve_lbr_callchain_sample()
2625 (max_lbr > 0) && alloc_lbr_stitch(thread, max_lbr)) { in resolve_lbr_callchain_sample()
2679 return (err < 0) ? err : 0; in resolve_lbr_callchain_sample()
2688 int err = 0; in find_prev_cpumode()
2690 while (--ent >= 0) { in find_prev_cpumode()
2696 false, NULL, NULL, 0); in find_prev_cpumode()
2714 int chain_nr = 0; in thread__resolve_callchain_sample()
2718 int first_call = 0; in thread__resolve_callchain_sample()
2728 !env ? 0 : env->max_branches); in thread__resolve_callchain_sample()
2730 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
2761 for (i = 0; i < nr; i++) { in thread__resolve_callchain_sample()
2785 memset(iter, 0, sizeof(struct iterations) * nr); in thread__resolve_callchain_sample()
2788 for (i = 0; i < nr; i++) { in thread__resolve_callchain_sample()
2799 &iter[i], 0); in thread__resolve_callchain_sample()
2806 if (chain_nr == 0) in thread__resolve_callchain_sample()
2807 return 0; in thread__resolve_callchain_sample()
2817 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
2819 for (i = first_call, nr_entries = 0; in thread__resolve_callchain_sample()
2839 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
2845 false, NULL, NULL, 0); in thread__resolve_callchain_sample()
2848 return (err < 0) ? err : 0; in thread__resolve_callchain_sample()
2851 return 0; in thread__resolve_callchain_sample()
2884 NULL, 0, 0, 0, ilist->srcline); in append_inlines()
2886 if (ret != 0) in append_inlines()
2900 return 0; in unwind_entry()
2902 if (append_inlines(cursor, &entry->ms, entry->ip) == 0) in unwind_entry()
2903 return 0; in unwind_entry()
2914 false, NULL, 0, 0, 0, srcline); in unwind_entry()
2926 return 0; in thread__resolve_callchain_unwind()
2931 return 0; in thread__resolve_callchain_unwind()
2945 int ret = 0; in thread__resolve_callchain()
2981 int rc = 0; in machine__for_each_thread()
2984 for (i = 0; i < THREADS__TABLE_SIZE; i++) { in machine__for_each_thread()
2990 if (rc != 0) in machine__for_each_thread()
2996 if (rc != 0) in machine__for_each_thread()
3008 int rc = 0; in machines__for_each_thread()
3011 if (rc != 0) in machines__for_each_thread()
3018 if (rc != 0) in machines__for_each_thread()
3028 if (cpu < 0 || cpu >= nr_cpus || !machine->current_tid) in machine__get_current_tid()
3040 if (cpu < 0) in machine__set_current_tid()
3049 for (i = 0; i < nr_cpus; i++) in machine__set_current_tid()
3068 return 0; in machine__set_current_tid()
3082 return machine ? perf_env__nr_cpus_avail(machine->env) : 0; in machine__nr_cpus_avail()
3088 int err = 0; in machine__get_kernel_start()
3166 int err = 0; in machine__for_each_dso()