/Linux-v6.6/arch/riscv/kernel/vdso/ |
D | hwprobe.c | 11 size_t cpu_count, unsigned long *cpus, 16 size_t cpu_count, unsigned long *cpus, 20 size_t cpu_count, unsigned long *cpus, in __vdso_riscv_hwprobe() argument 25 bool all_cpus = !cpu_count && !cpus; in __vdso_riscv_hwprobe() 36 return riscv_hwprobe(pairs, pair_count, cpu_count, cpus, flags); in __vdso_riscv_hwprobe()
|
/Linux-v6.6/arch/riscv/kernel/ |
D | smpboot.c | 78 static unsigned int cpu_count = 1; variable 108 early_map_cpu_to_node(0, acpi_numa_get_nid(cpu_count)); in acpi_parse_rintc() 112 if (cpu_count >= NR_CPUS) { in acpi_parse_rintc() 117 cpuid_to_hartid_map(cpu_count) = hart; in acpi_parse_rintc() 118 early_map_cpu_to_node(cpu_count, acpi_numa_get_nid(cpu_count)); in acpi_parse_rintc() 119 cpu_count++; in acpi_parse_rintc()
|
D | sys_riscv.c | 231 size_t pair_count, size_t cpu_count, in do_riscv_hwprobe() argument 249 if (!cpu_count && !cpus_user) { in do_riscv_hwprobe() 252 if (cpu_count > cpumask_size()) in do_riscv_hwprobe() 253 cpu_count = cpumask_size(); in do_riscv_hwprobe() 255 ret = copy_from_user(&cpus, cpus_user, cpu_count); in do_riscv_hwprobe() 332 size_t, pair_count, size_t, cpu_count, unsigned long __user *, in SYSCALL_DEFINE5() argument 335 return do_riscv_hwprobe(pairs, pair_count, cpu_count, in SYSCALL_DEFINE5()
|
D | patch.c | 22 atomic_t cpu_count; member 223 if (atomic_inc_return(&patch->cpu_count) == num_online_cpus()) { in patch_text_cb() 229 atomic_inc(&patch->cpu_count); in patch_text_cb() 231 while (atomic_read(&patch->cpu_count) <= num_online_cpus()) in patch_text_cb() 247 .cpu_count = ATOMIC_INIT(0), in patch_text()
|
/Linux-v6.6/tools/power/cpupower/utils/idle_monitor/ |
D | mperf_monitor.c | 231 for (cpu = 0; cpu < cpu_count; cpu++) { in mperf_start() 243 for (cpu = 0; cpu < cpu_count; cpu++) { in mperf_stop() 345 is_valid = calloc(cpu_count, sizeof(int)); in mperf_register() 346 mperf_previous_count = calloc(cpu_count, sizeof(unsigned long long)); in mperf_register() 347 aperf_previous_count = calloc(cpu_count, sizeof(unsigned long long)); in mperf_register() 348 mperf_current_count = calloc(cpu_count, sizeof(unsigned long long)); in mperf_register() 349 aperf_current_count = calloc(cpu_count, sizeof(unsigned long long)); in mperf_register() 350 tsc_at_measure_start = calloc(cpu_count, sizeof(unsigned long long)); in mperf_register() 351 tsc_at_measure_end = calloc(cpu_count, sizeof(unsigned long long)); in mperf_register()
|
D | cpuidle_sysfs.c | 48 for (cpu = 0; cpu < cpu_count; cpu++) { in cpuidle_start() 67 for (cpu = 0; cpu < cpu_count; cpu++) { in cpuidle_stop() 178 previous_count = malloc(sizeof(long long *) * cpu_count); in cpuidle_register() 179 current_count = malloc(sizeof(long long *) * cpu_count); in cpuidle_register() 180 for (num = 0; num < cpu_count; num++) { in cpuidle_register() 195 for (num = 0; num < cpu_count; num++) { in cpuidle_unregister()
|
D | hsw_ext_idle.c | 119 for (cpu = 0; cpu < cpu_count; cpu++) { in hsw_ext_start() 136 for (cpu = 0; cpu < cpu_count; cpu++) { in hsw_ext_stop() 161 is_valid = calloc(cpu_count, sizeof(int)); in hsw_ext_register() 163 previous_count[num] = calloc(cpu_count, in hsw_ext_register() 165 current_count[num] = calloc(cpu_count, in hsw_ext_register()
|
D | snb_idle.c | 117 for (cpu = 0; cpu < cpu_count; cpu++) { in snb_start() 134 for (cpu = 0; cpu < cpu_count; cpu++) { in snb_stop() 166 is_valid = calloc(cpu_count, sizeof(int)); in snb_register() 168 previous_count[num] = calloc(cpu_count, in snb_register() 170 current_count[num] = calloc(cpu_count, in snb_register()
|
D | nhm_idle.c | 134 for (cpu = 0; cpu < cpu_count; cpu++) { in nhm_start() 153 for (cpu = 0; cpu < cpu_count; cpu++) { in nhm_stop() 180 is_valid = calloc(cpu_count, sizeof(int)); in intel_nhm_register() 182 previous_count[num] = calloc(cpu_count, in intel_nhm_register() 184 current_count[num] = calloc(cpu_count, in intel_nhm_register()
|
D | cpupower-monitor.c | 30 int cpu_count; variable 330 for (cpu = 0; cpu < cpu_count; cpu++) in do_interval_measure() 342 for (cpu = 0; cpu < cpu_count; cpu++) in do_interval_measure() 394 cpu_count = get_cpu_topology(&cpu_top); in cmd_monitor() 395 if (cpu_count < 0) { in cmd_monitor() 407 dprint("System has up to %d CPU cores\n", cpu_count); in cmd_monitor() 438 cpu_top.pkgs, cpu_top.cores, cpu_count); in cmd_monitor() 455 for (cpu = 0; cpu < cpu_count; cpu++) { in cmd_monitor()
|
D | amd_fam14h_idle.c | 233 for (cpu = 0; cpu < cpu_count; cpu++) in amd_fam14h_start() 253 for (cpu = 0; cpu < cpu_count; cpu++) in amd_fam14h_stop() 294 previous_count[num] = calloc(cpu_count, in amd_fam14h_register() 296 current_count[num] = calloc(cpu_count, in amd_fam14h_register()
|
/Linux-v6.6/arch/xtensa/kernel/ |
D | jump_label.c | 27 atomic_t cpu_count; member 43 if (atomic_inc_return(&patch->cpu_count) == num_online_cpus()) { in patch_text_stop_machine() 45 atomic_inc(&patch->cpu_count); in patch_text_stop_machine() 47 while (atomic_read(&patch->cpu_count) <= num_online_cpus()) in patch_text_stop_machine() 58 .cpu_count = ATOMIC_INIT(0), in patch_text()
|
/Linux-v6.6/drivers/thermal/intel/ |
D | intel_hfi.c | 200 int i = 0, cpu_count; in update_capabilities() local 205 cpu_count = cpumask_weight(hfi_instance->cpus); in update_capabilities() 208 if (!cpu_count) in update_capabilities() 211 cpu_caps = kcalloc(cpu_count, sizeof(*cpu_caps), GFP_KERNEL); in update_capabilities() 217 if (cpu_count < HFI_MAX_THERM_NOTIFY_COUNT) in update_capabilities() 222 (i + HFI_MAX_THERM_NOTIFY_COUNT) <= cpu_count; in update_capabilities() 227 cpu_count = cpu_count - i; in update_capabilities() 231 if (cpu_count) in update_capabilities() 232 thermal_genl_cpu_capability_event(cpu_count, &cpu_caps[i]); in update_capabilities()
|
/Linux-v6.6/arch/arm64/kernel/ |
D | smp.c | 501 static unsigned int cpu_count = 1; variable 533 if (is_mpidr_duplicate(cpu_count, hwid)) { in acpi_map_gic_cpu_interface() 550 if (cpu_count >= NR_CPUS) in acpi_map_gic_cpu_interface() 554 set_cpu_logical_map(cpu_count, hwid); in acpi_map_gic_cpu_interface() 556 cpu_madt_gicc[cpu_count] = *processor; in acpi_map_gic_cpu_interface() 567 acpi_set_mailbox_entry(cpu_count, processor); in acpi_map_gic_cpu_interface() 569 cpu_count++; in acpi_map_gic_cpu_interface() 633 if (is_mpidr_duplicate(cpu_count, hwid)) { in of_parse_and_init_cpus() 664 if (cpu_count >= NR_CPUS) in of_parse_and_init_cpus() 668 set_cpu_logical_map(cpu_count, hwid); in of_parse_and_init_cpus() [all …]
|
D | patching.c | 129 atomic_t cpu_count; member 138 if (atomic_inc_return(&pp->cpu_count) == num_online_cpus()) { in aarch64_insn_patch_text_cb() 143 atomic_inc(&pp->cpu_count); in aarch64_insn_patch_text_cb() 145 while (atomic_read(&pp->cpu_count) <= num_online_cpus()) in aarch64_insn_patch_text_cb() 159 .cpu_count = ATOMIC_INIT(0), in aarch64_insn_patch_text()
|
/Linux-v6.6/arch/s390/hypfs/ |
D | hypfs_diag0c.c | 33 unsigned int cpu_count, cpu, i; in diag0c_store() local 37 cpu_count = num_online_cpus(); in diag0c_store() 43 diag0c_data = kzalloc(struct_size(diag0c_data, entry, cpu_count), in diag0c_store() 55 *count = cpu_count; in diag0c_store()
|
/Linux-v6.6/tools/testing/selftests/rcutorture/bin/ |
D | kvm-test-1-run.sh | 144 cpu_count=`configNR_CPUS.sh $resdir/ConfigFragment` 145 cpu_count=`configfrag_boot_cpus "$boot_args_in" "$config_template" "$cpu_count"` 146 if test "$cpu_count" -gt "$TORTURE_ALLOTED_CPUS" 148 echo CPU count limited from $cpu_count to $TORTURE_ALLOTED_CPUS | tee -a $resdir/Warnings 149 cpu_count=$TORTURE_ALLOTED_CPUS 151 qemu_args="`specify_qemu_cpus "$QEMU" "$qemu_args" "$cpu_count"`" 207 echo "# TORTURE_CPU_COUNT=$cpu_count" >> $resdir/qemu-cmd
|
D | kvm.sh | 321 cpu_count=`configNR_CPUS.sh $T/KCONFIG_ARG` 323 cpu_count=`configNR_CPUS.sh $CONFIGFRAG/$CF1` 325 cpu_count=`configfrag_boot_cpus "$TORTURE_BOOTARGS" "$CONFIGFRAG/$CF1" "$cpu_count"` 326 cpu_count=`configfrag_boot_maxcpus "$TORTURE_BOOTARGS" "$CONFIGFRAG/$CF1" "$cpu_count"` 327 echo 'scenariocpu["'"$CF1"'"] = '"$cpu_count"';' >> $T/cfgcpu.awk
|
D | kvm-test-1-run-batch.sh | 78 affinity_export="`awk -f $T/cpubatches.awk -v cpu_count="$cpu_count" -v scenario=$i < /dev/null`"
|
/Linux-v6.6/arch/arm/mach-axxia/ |
D | platsmp.c | 56 int cpu_count = 0; in axxia_smp_prepare_cpus() local 73 if (cpu_count < max_cpus) { in axxia_smp_prepare_cpus() 75 cpu_count++; in axxia_smp_prepare_cpus()
|
/Linux-v6.6/arch/csky/kernel/ |
D | ftrace.c | 205 atomic_t cpu_count; member 212 if (atomic_inc_return(¶m->cpu_count) == 1) { in __ftrace_modify_code() 214 atomic_inc(¶m->cpu_count); in __ftrace_modify_code() 216 while (atomic_read(¶m->cpu_count) <= num_online_cpus()) in __ftrace_modify_code()
|
/Linux-v6.6/scripts/ |
D | checkkconfigsymbols.py | 18 from multiprocessing import Pool, cpu_count 276 pool = Pool(cpu_count(), init_worker) 283 for part in partition(kfiles, cpu_count()): 313 pool = Pool(cpu_count(), init_worker) 340 arglist = partition(source_files, cpu_count()) 346 for part in partition(kconfig_files, cpu_count()):
|
/Linux-v6.6/tools/power/x86/intel-speed-select/ |
D | isst-core-tpmi.c | 277 int ret, cpu_count; in tpmi_get_coremask_info() local 290 ctdp_level->core_cpumask, &cpu_count); in tpmi_get_coremask_info() 291 ctdp_level->cpu_count = cpu_count; in tpmi_get_coremask_info() 294 id->cpu, config_index, ctdp_level->cpu_count); in tpmi_get_coremask_info() 390 int ret, cpu_count; in _pbf_get_coremask_info() local 403 pbf_info->core_cpumask, &cpu_count); in _pbf_get_coremask_info() 406 id->cpu, config_index, cpu_count); in _pbf_get_coremask_info()
|
/Linux-v6.6/arch/csky/kernel/probes/ |
D | kprobes.c | 25 atomic_t cpu_count; member 33 if (atomic_inc_return(¶m->cpu_count) == num_online_cpus()) { in patch_text_cb() 36 atomic_inc(¶m->cpu_count); in patch_text_cb() 38 while (atomic_read(¶m->cpu_count) <= num_online_cpus()) in patch_text_cb()
|
/Linux-v6.6/arch/parisc/kernel/ |
D | processor.c | 94 if (boot_cpu_data.cpu_count > 0) { in processor_probe() 103 cpuid = boot_cpu_data.cpu_count; in processor_probe() 152 boot_cpu_data.cpu_count--; in processor_probe() 162 boot_cpu_data.cpu_count++; in processor_probe()
|