Lines Matching refs:env

16 void perf_env__insert_bpf_prog_info(struct perf_env *env,  in perf_env__insert_bpf_prog_info()  argument
24 down_write(&env->bpf_progs.lock); in perf_env__insert_bpf_prog_info()
25 p = &env->bpf_progs.infos.rb_node; in perf_env__insert_bpf_prog_info()
41 rb_insert_color(&info_node->rb_node, &env->bpf_progs.infos); in perf_env__insert_bpf_prog_info()
42 env->bpf_progs.infos_cnt++; in perf_env__insert_bpf_prog_info()
44 up_write(&env->bpf_progs.lock); in perf_env__insert_bpf_prog_info()
47 struct bpf_prog_info_node *perf_env__find_bpf_prog_info(struct perf_env *env, in perf_env__find_bpf_prog_info() argument
53 down_read(&env->bpf_progs.lock); in perf_env__find_bpf_prog_info()
54 n = env->bpf_progs.infos.rb_node; in perf_env__find_bpf_prog_info()
68 up_read(&env->bpf_progs.lock); in perf_env__find_bpf_prog_info()
72 void perf_env__insert_btf(struct perf_env *env, struct btf_node *btf_node) in perf_env__insert_btf() argument
79 down_write(&env->bpf_progs.lock); in perf_env__insert_btf()
80 p = &env->bpf_progs.btfs.rb_node; in perf_env__insert_btf()
96 rb_insert_color(&btf_node->rb_node, &env->bpf_progs.btfs); in perf_env__insert_btf()
97 env->bpf_progs.btfs_cnt++; in perf_env__insert_btf()
99 up_write(&env->bpf_progs.lock); in perf_env__insert_btf()
102 struct btf_node *perf_env__find_btf(struct perf_env *env, __u32 btf_id) in perf_env__find_btf() argument
107 down_read(&env->bpf_progs.lock); in perf_env__find_btf()
108 n = env->bpf_progs.btfs.rb_node; in perf_env__find_btf()
122 up_read(&env->bpf_progs.lock); in perf_env__find_btf()
127 static void perf_env__purge_bpf(struct perf_env *env) in perf_env__purge_bpf() argument
132 down_write(&env->bpf_progs.lock); in perf_env__purge_bpf()
134 root = &env->bpf_progs.infos; in perf_env__purge_bpf()
146 env->bpf_progs.infos_cnt = 0; in perf_env__purge_bpf()
148 root = &env->bpf_progs.btfs; in perf_env__purge_bpf()
160 env->bpf_progs.btfs_cnt = 0; in perf_env__purge_bpf()
162 up_write(&env->bpf_progs.lock); in perf_env__purge_bpf()
165 void perf_env__exit(struct perf_env *env) in perf_env__exit() argument
169 perf_env__purge_bpf(env); in perf_env__exit()
170 zfree(&env->hostname); in perf_env__exit()
171 zfree(&env->os_release); in perf_env__exit()
172 zfree(&env->version); in perf_env__exit()
173 zfree(&env->arch); in perf_env__exit()
174 zfree(&env->cpu_desc); in perf_env__exit()
175 zfree(&env->cpuid); in perf_env__exit()
176 zfree(&env->cmdline); in perf_env__exit()
177 zfree(&env->cmdline_argv); in perf_env__exit()
178 zfree(&env->sibling_cores); in perf_env__exit()
179 zfree(&env->sibling_threads); in perf_env__exit()
180 zfree(&env->pmu_mappings); in perf_env__exit()
181 zfree(&env->cpu); in perf_env__exit()
183 for (i = 0; i < env->nr_numa_nodes; i++) in perf_env__exit()
184 perf_cpu_map__put(env->numa_nodes[i].map); in perf_env__exit()
185 zfree(&env->numa_nodes); in perf_env__exit()
187 for (i = 0; i < env->caches_cnt; i++) in perf_env__exit()
188 cpu_cache_level__free(&env->caches[i]); in perf_env__exit()
189 zfree(&env->caches); in perf_env__exit()
191 for (i = 0; i < env->nr_memory_nodes; i++) in perf_env__exit()
192 zfree(&env->memory_nodes[i].set); in perf_env__exit()
193 zfree(&env->memory_nodes); in perf_env__exit()
196 void perf_env__init(struct perf_env *env) in perf_env__init() argument
198 env->bpf_progs.infos = RB_ROOT; in perf_env__init()
199 env->bpf_progs.btfs = RB_ROOT; in perf_env__init()
200 init_rwsem(&env->bpf_progs.lock); in perf_env__init()
203 int perf_env__set_cmdline(struct perf_env *env, int argc, const char *argv[]) in perf_env__set_cmdline() argument
208 env->cmdline_argv = calloc(argc, sizeof(char *)); in perf_env__set_cmdline()
209 if (env->cmdline_argv == NULL) in perf_env__set_cmdline()
217 env->cmdline_argv[i] = argv[i]; in perf_env__set_cmdline()
218 if (env->cmdline_argv[i] == NULL) in perf_env__set_cmdline()
222 env->nr_cmdline = argc; in perf_env__set_cmdline()
226 zfree(&env->cmdline_argv); in perf_env__set_cmdline()
231 int perf_env__read_cpu_topology_map(struct perf_env *env) in perf_env__read_cpu_topology_map() argument
235 if (env->cpu != NULL) in perf_env__read_cpu_topology_map()
238 if (env->nr_cpus_avail == 0) in perf_env__read_cpu_topology_map()
239 env->nr_cpus_avail = cpu__max_present_cpu(); in perf_env__read_cpu_topology_map()
241 nr_cpus = env->nr_cpus_avail; in perf_env__read_cpu_topology_map()
245 env->cpu = calloc(nr_cpus, sizeof(env->cpu[0])); in perf_env__read_cpu_topology_map()
246 if (env->cpu == NULL) in perf_env__read_cpu_topology_map()
250 env->cpu[cpu].core_id = cpu_map__get_core_id(cpu); in perf_env__read_cpu_topology_map()
251 env->cpu[cpu].socket_id = cpu_map__get_socket_id(cpu); in perf_env__read_cpu_topology_map()
252 env->cpu[cpu].die_id = cpu_map__get_die_id(cpu); in perf_env__read_cpu_topology_map()
255 env->nr_cpus_avail = nr_cpus; in perf_env__read_cpu_topology_map()
259 static int perf_env__read_arch(struct perf_env *env) in perf_env__read_arch() argument
263 if (env->arch) in perf_env__read_arch()
267 env->arch = strdup(uts.machine); in perf_env__read_arch()
269 return env->arch ? 0 : -ENOMEM; in perf_env__read_arch()
272 static int perf_env__read_nr_cpus_avail(struct perf_env *env) in perf_env__read_nr_cpus_avail() argument
274 if (env->nr_cpus_avail == 0) in perf_env__read_nr_cpus_avail()
275 env->nr_cpus_avail = cpu__max_present_cpu(); in perf_env__read_nr_cpus_avail()
277 return env->nr_cpus_avail ? 0 : -ENOENT; in perf_env__read_nr_cpus_avail()
280 const char *perf_env__raw_arch(struct perf_env *env) in perf_env__raw_arch() argument
282 return env && !perf_env__read_arch(env) ? env->arch : "unknown"; in perf_env__raw_arch()
285 int perf_env__nr_cpus_avail(struct perf_env *env) in perf_env__nr_cpus_avail() argument
287 return env && !perf_env__read_nr_cpus_avail(env) ? env->nr_cpus_avail : 0; in perf_env__nr_cpus_avail()
327 const char *perf_env__arch(struct perf_env *env) in perf_env__arch() argument
332 if (!env || !env->arch) { /* Assume local operation */ in perf_env__arch()
337 arch_name = env->arch; in perf_env__arch()