| /Linux-v5.4/kernel/sched/ | 
| D | topology.c | 276 static void perf_domain_debug(const struct cpumask *cpu_map,  in perf_domain_debug()  argument282 	printk(KERN_DEBUG "root_domain %*pbl:", cpumask_pr_args(cpu_map));  in perf_domain_debug()
 342 static bool build_perf_domains(const struct cpumask *cpu_map)  in build_perf_domains()  argument
 344 	int i, nr_pd = 0, nr_cs = 0, nr_cpus = cpumask_weight(cpu_map);  in build_perf_domains()
 346 	int cpu = cpumask_first(cpu_map);  in build_perf_domains()
 358 					cpumask_pr_args(cpu_map));  in build_perf_domains()
 363 	for_each_cpu(i, cpu_map) {  in build_perf_domains()
 377 						cpumask_pr_args(cpu_map));  in build_perf_domains()
 399 						cpumask_pr_args(cpu_map));  in build_perf_domains()
 403 	perf_domain_debug(cpu_map, pd);  in build_perf_domains()
 [all …]
 
 | 
| D | sched.h | 788 extern int sched_init_domains(const struct cpumask *cpu_map);
 | 
| /Linux-v5.4/tools/power/x86/intel-speed-select/ | 
| D | isst-config.c | 59 struct _cpu_map *cpu_map;  variable355 	cpu_map = malloc(sizeof(*cpu_map) * topo_max_cpus);  in create_cpu_map()
 356 	if (!cpu_map)  in create_cpu_map()
 368 		map.cpu_map[0].logical_cpu = i;  in create_cpu_map()
 371 			     map.cpu_map[0].logical_cpu);  in create_cpu_map()
 375 				map.cpu_map[0].logical_cpu);  in create_cpu_map()
 378 		cpu_map[i].core_id = get_physical_core_id(i);  in create_cpu_map()
 379 		cpu_map[i].pkg_id = get_physical_package_id(i);  in create_cpu_map()
 380 		cpu_map[i].die_id = get_physical_die_id(i);  in create_cpu_map()
 381 		cpu_map[i].punit_cpu = map.cpu_map[0].physical_cpu;  in create_cpu_map()
 [all …]
 
 | 
| /Linux-v5.4/arch/ia64/mm/ | 
| D | discontig.c | 184 	unsigned int *cpu_map;  in setup_per_cpu_areas()  local194 	cpu_map = ai->groups[0].cpu_map;  in setup_per_cpu_areas()
 208 				cpu_map[unit++] = cpu;  in setup_per_cpu_areas()
 233 		cpu = cpu_map[unit];  in setup_per_cpu_areas()
 245 		gi->cpu_map		= &cpu_map[unit];  in setup_per_cpu_areas()
 
 | 
| D | contig.c | 119 		gi->cpu_map[gi->nr_units++] = cpu;  in setup_per_cpu_areas()
 | 
| /Linux-v5.4/drivers/platform/x86/intel_speed_select_if/ | 
| D | isst_if_common.c | 374 	struct isst_if_cpu_map *cpu_map;  in isst_if_proc_phyid_req()  local376 	cpu_map = (struct isst_if_cpu_map *)cmd_ptr;  in isst_if_proc_phyid_req()
 377 	if (cpu_map->logical_cpu >= nr_cpu_ids ||  in isst_if_proc_phyid_req()
 378 	    cpu_map->logical_cpu >= num_possible_cpus())  in isst_if_proc_phyid_req()
 382 	cpu_map->physical_cpu = isst_cpu_info[cpu_map->logical_cpu].punit_cpu_id;  in isst_if_proc_phyid_req()
 505 		cmd_cb.offset = offsetof(struct isst_if_cpu_maps, cpu_map);  in isst_if_def_ioctl()
 
 | 
| /Linux-v5.4/kernel/bpf/ | 
| D | cpumap.c | 74 	struct bpf_cpu_map_entry **cpu_map;  member126 	cmap->cpu_map = bpf_map_area_alloc(cmap->map.max_entries *  in cpu_map_alloc()
 129 	if (!cmap->cpu_map)  in cpu_map_alloc()
 447 	old_rcpu = xchg(&cmap->cpu_map[key_cpu], rcpu);  in __cpu_map_entry_replace()
 543 		rcpu = READ_ONCE(cmap->cpu_map[i]);  in cpu_map_free()
 551 	bpf_map_area_free(cmap->cpu_map);  in cpu_map_free()
 563 	rcpu = READ_ONCE(cmap->cpu_map[key]);  in __cpu_map_lookup_elem()
 
 | 
| /Linux-v5.4/samples/bpf/ | 
| D | xdp_redirect_cpu_kern.c | 21 struct bpf_map_def SEC("maps") cpu_map = {227 	return bpf_redirect_map(&cpu_map, cpu_dest, 0);  in xdp_prognum0_no_touch()
 270 	return bpf_redirect_map(&cpu_map, cpu_dest, 0);  in xdp_prognum1_touch_data()
 318 	return bpf_redirect_map(&cpu_map, cpu_dest, 0);  in xdp_prognum2_round_robin()
 386 	return bpf_redirect_map(&cpu_map, cpu_dest, 0);  in xdp_prognum3_proto_separate()
 462 	return bpf_redirect_map(&cpu_map, cpu_dest, 0);  in xdp_prognum4_ddos_filter_pktgen()
 565 	return bpf_redirect_map(&cpu_map, cpu_dest, 0);  in xdp_prognum5_lb_hash_ip_pairs()
 
 | 
| /Linux-v5.4/tools/perf/tests/ | 
| D | cpumap.c | 19 	struct perf_record_cpu_map *map_event = &event->cpu_map;  in process_event_mask()53 	struct perf_record_cpu_map *map_event = &event->cpu_map;  in process_event_cpus()
 
 | 
| /Linux-v5.4/mm/ | 
| D | percpu.c | 2129 			  __alignof__(ai->groups[0].cpu_map[0]));  in pcpu_alloc_alloc_info()2130 	ai_size = base_size + nr_units * sizeof(ai->groups[0].cpu_map[0]);  in pcpu_alloc_alloc_info()
 2138 	ai->groups[0].cpu_map = ptr;  in pcpu_alloc_alloc_info()
 2141 		ai->groups[0].cpu_map[unit] = NR_CPUS;  in pcpu_alloc_alloc_info()
 2207 				if (gi->cpu_map[unit] != NR_CPUS)  in pcpu_dump_alloc_info()
 2209 						cpu_width, gi->cpu_map[unit]);  in pcpu_dump_alloc_info()
 2354 			cpu = gi->cpu_map[i];  in pcpu_setup_first_chunk()
 2543 	unsigned int *cpu_map;  in pcpu_build_alloc_info()  local
 2629 	cpu_map = ai->groups[0].cpu_map;  in pcpu_build_alloc_info()
 2632 		ai->groups[group].cpu_map = cpu_map;  in pcpu_build_alloc_info()
 [all …]
 
 | 
| /Linux-v5.4/tools/perf/util/ | 
| D | mmap.c | 333 	const struct perf_cpu_map *cpu_map = NULL;  in build_node_mask()  local335 	cpu_map = cpu_map__online();  in build_node_mask()
 336 	if (!cpu_map)  in build_node_mask()
 339 	nr_cpus = perf_cpu_map__nr(cpu_map);  in build_node_mask()
 341 		cpu = cpu_map->map[c]; /* map c index to online cpu index */  in build_node_mask()
 
 | 
| D | tool.h | 71 			cpu_map,  member
 | 
| D | session.c | 518 	if (tool->cpu_map == NULL)  in perf_tool__fill_defaults()519 		tool->cpu_map = process_event_cpu_map_stub;  in perf_tool__fill_defaults()
 848 	struct perf_record_cpu_map_data *data = &event->cpu_map.data;  in perf_event__cpu_map_swap()
 1551 		return tool->cpu_map(session, event);  in perf_session__process_user_event()
 
 | 
| D | event.c | 297 	struct perf_cpu_map *cpus = cpu_map__new_data(&event->cpu_map.data);  in perf_event__fprintf_cpu_map()
 | 
| /Linux-v5.4/tools/perf/arch/nds32/util/ | 
| D | header.c | 15 	struct cpu_map *cpus;  in get_cpuid_str()
 | 
| /Linux-v5.4/tools/perf/arch/arm/util/ | 
| D | cs-etm.c | 642 	struct perf_cpu_map *cpu_map;  in cs_etm_info_fill()  local657 		cpu_map = online_cpus;  in cs_etm_info_fill()
 666 		cpu_map = event_cpus;  in cs_etm_info_fill()
 669 	nr_cpu = perf_cpu_map__nr(cpu_map);  in cs_etm_info_fill()
 683 		if (cpu_map__has(cpu_map, i))  in cs_etm_info_fill()
 
 | 
| /Linux-v5.4/tools/perf/python/ | 
| D | twatch.py | 12 	cpus = perf.cpu_map()
 | 
| D | tracepoint.py | 19     cpus    = perf.cpu_map()
 | 
| /Linux-v5.4/include/uapi/linux/ | 
| D | isst_if.h | 63 	struct isst_if_cpu_map cpu_map[1];  member
 | 
| /Linux-v5.4/include/linux/ | 
| D | percpu.h | 71 	unsigned int		*cpu_map;	/* unit->cpu map, empty  member
 | 
| /Linux-v5.4/tools/perf/lib/include/perf/ | 
| D | event.h | 376 	struct perf_record_cpu_map		cpu_map;  member
 | 
| /Linux-v5.4/drivers/scsi/lpfc/ | 
| D | lpfc_init.c | 6875 	phba->sli4_hba.cpu_map = kcalloc(phba->sli4_hba.num_possible_cpu,  in lpfc_sli4_driver_resource_setup()6878 	if (!phba->sli4_hba.cpu_map) {  in lpfc_sli4_driver_resource_setup()
 6913 	kfree(phba->sli4_hba.cpu_map);  in lpfc_sli4_driver_resource_setup()
 6952 	kfree(phba->sli4_hba.cpu_map);  in lpfc_sli4_driver_resource_unset()
 8810 		cpup = &phba->sli4_hba.cpu_map[cpu];  in lpfc_sli4_queue_create()
 8843 		cpup = &phba->sli4_hba.cpu_map[cpu];  in lpfc_sli4_queue_create()
 8856 		eqcpup = &phba->sli4_hba.cpu_map[eqcpu];  in lpfc_sli4_queue_create()
 9416 			cpup = &phba->sli4_hba.cpu_map[cpu];  in lpfc_sli4_queue_setup()
 9451 		cpup = &phba->sli4_hba.cpu_map[cpu];  in lpfc_sli4_queue_setup()
 10576 		cpup = &phba->sli4_hba.cpu_map[cpu];  in lpfc_find_cpu_handle()
 [all …]
 
 | 
| D | lpfc_sli4.h | 891 	struct lpfc_vector_map_info *cpu_map;  member
 | 
| /Linux-v5.4/tools/perf/ | 
| D | builtin-stat.c | 1553 	cpus = cpu_map__new_data(&event->cpu_map.data);  in process_cpu_map_event()1600 		.cpu_map	= process_cpu_map_event,
 
 | 
| D | builtin-script.c | 3326 	script->cpus = cpu_map__new_data(&event->cpu_map.data);  in process_cpu_map_event()3439 			.cpu_map	 = process_cpu_map_event,  in cmd_script()
 
 |