/Linux-v5.4/tools/perf/util/ |
D | counts.c | 10 struct perf_counts *counts = zalloc(sizeof(*counts)); in perf_counts__new() local 12 if (counts) { in perf_counts__new() 17 free(counts); in perf_counts__new() 21 counts->values = values; in perf_counts__new() 25 xyarray__delete(counts->values); in perf_counts__new() 26 free(counts); in perf_counts__new() 30 counts->loaded = values; in perf_counts__new() 33 return counts; in perf_counts__new() 36 void perf_counts__delete(struct perf_counts *counts) in perf_counts__delete() argument 38 if (counts) { in perf_counts__delete() [all …]
|
D | counts.h | 21 perf_counts(struct perf_counts *counts, int cpu, int thread) in perf_counts() argument 23 return xyarray__entry(counts->values, cpu, thread); in perf_counts() 27 perf_counts__is_loaded(struct perf_counts *counts, int cpu, int thread) in perf_counts__is_loaded() argument 29 return *((bool *) xyarray__entry(counts->loaded, cpu, thread)); in perf_counts__is_loaded() 33 perf_counts__set_loaded(struct perf_counts *counts, int cpu, int thread, bool loaded) in perf_counts__set_loaded() argument 35 *((bool *) xyarray__entry(counts->loaded, cpu, thread)) = loaded; in perf_counts__set_loaded() 39 void perf_counts__delete(struct perf_counts *counts);
|
D | branch.c | 24 st->counts[flags->type]++; in branch_type_count() 67 total += st->counts[i]; in branch_type_stat_display() 101 if (st->counts[i] > 0) in branch_type_stat_display() 105 (double)st->counts[i] / (double)total); in branch_type_stat_display() 120 total += st->counts[i]; in branch_type_str() 135 if (st->counts[i] > 0) in branch_type_str()
|
D | stat.c | 150 struct perf_counts *counts; in perf_evsel__alloc_prev_raw_counts() local 152 counts = perf_counts__new(ncpus, nthreads); in perf_evsel__alloc_prev_raw_counts() 153 if (counts) in perf_evsel__alloc_prev_raw_counts() 154 evsel->prev_raw_counts = counts; in perf_evsel__alloc_prev_raw_counts() 156 return counts ? 0 : -ENOMEM; in perf_evsel__alloc_prev_raw_counts() 285 struct perf_counts_values *aggr = &evsel->counts->aggr; in process_counter_values() 345 perf_counts(counter->counts, cpu, thread))) in process_counter_maps() 356 struct perf_counts_values *aggr = &counter->counts->aggr; in perf_stat_process_counter() 358 u64 *count = counter->counts->aggr.values; in perf_stat_process_counter() 385 perf_counts_values__scale(aggr, config->scale, &counter->counts->scaled); in perf_stat_process_counter() [all …]
|
D | stat-display.c | 433 if (run == 0 || ena == 0 || counter->counts->scaled == -1) { in printout() 507 val += perf_counts(counter->counts, cpu, 0)->val; in aggr_update_shadow() 595 struct perf_counts_values *counts; in aggr_cb() local 602 counts = perf_counts(counter->counts, cpu, 0); in aggr_cb() 607 if (counts->ena == 0 || counts->run == 0 || in aggr_cb() 608 counter->counts->scaled == -1) { in aggr_cb() 613 ad->val += counts->val; in aggr_cb() 614 ad->ena += counts->ena; in aggr_cb() 615 ad->run += counts->run; in aggr_cb() 712 val += perf_counts(counter->counts, cpu, thread)->val; in sort_aggr_thread() [all …]
|
/Linux-v5.4/arch/powerpc/platforms/pseries/ |
D | msi.c | 230 struct msi_counts *counts = data; in count_non_bridge_devices() local 240 counts->num_devices++; in count_non_bridge_devices() 247 struct msi_counts *counts = data; in count_spare_msis() local 251 if (dn == counts->requestor) in count_spare_msis() 252 req = counts->request; in count_spare_msis() 266 if (req < counts->quota) in count_spare_msis() 267 counts->spare += counts->quota - req; in count_spare_msis() 268 else if (req > counts->quota) in count_spare_msis() 269 counts->over_quota++; in count_spare_msis() 277 struct msi_counts counts; in msi_quota_for_device() local [all …]
|
/Linux-v5.4/tools/perf/lib/tests/ |
D | test-evsel.c | 36 struct perf_counts_values counts = { .val = 0 }; in test_stat_cpu() local 38 perf_evsel__read(evsel, cpu, 0, &counts); in test_stat_cpu() 39 __T("failed to read value for evsel", counts.val != 0); in test_stat_cpu() 51 struct perf_counts_values counts = { .val = 0 }; in test_stat_thread() local 71 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread() 72 __T("failed to read value for evsel", counts.val != 0); in test_stat_thread() 83 struct perf_counts_values counts = { .val = 0 }; in test_stat_thread_enable() local 104 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread_enable() 105 __T("failed to read value for evsel", counts.val == 0); in test_stat_thread_enable() 110 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread_enable() [all …]
|
D | test-evlist.c | 57 struct perf_counts_values counts = { .val = 0 }; in test_stat_cpu() local 59 perf_evsel__read(evsel, cpu, 0, &counts); in test_stat_cpu() 60 __T("failed to read value for evsel", counts.val != 0); in test_stat_cpu() 73 struct perf_counts_values counts = { .val = 0 }; in test_stat_thread() local 111 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread() 112 __T("failed to read value for evsel", counts.val != 0); in test_stat_thread() 124 struct perf_counts_values counts = { .val = 0 }; in test_stat_thread_enable() local 164 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread_enable() 165 __T("failed to read value for evsel", counts.val == 0); in test_stat_thread_enable() 171 perf_evsel__read(evsel, 0, 0, &counts); in test_stat_thread_enable() [all …]
|
/Linux-v5.4/samples/bpf/ |
D | sampleip_user.c | 78 struct ipcount counts[MAX_IPS]; variable 98 counts[i].ip = next_key; in print_ip_map() 99 counts[i++].count = value; in print_ip_map() 105 qsort(counts, max, sizeof(struct ipcount), count_cmp); in print_ip_map() 107 if (counts[i].ip > PAGE_OFFSET) { in print_ip_map() 108 sym = ksym_search(counts[i].ip); in print_ip_map() 114 printf("0x%-17llx %-32s %u\n", counts[i].ip, sym->name, in print_ip_map() 115 counts[i].count); in print_ip_map() 117 printf("0x%-17llx %-32s %u\n", counts[i].ip, "(user)", in print_ip_map() 118 counts[i].count); in print_ip_map()
|
D | trace_event_kern.c | 20 struct bpf_map_def SEC("maps") counts = { 71 val = bpf_map_lookup_elem(&counts, &key); in bpf_prog1() 75 bpf_map_update_elem(&counts, &key, &one, BPF_NOEXIST); in bpf_prog1()
|
D | offwaketime_kern.c | 25 struct bpf_map_def SEC("maps") counts = { 94 val = bpf_map_lookup_elem(&counts, &key); in update_counts() 96 bpf_map_update_elem(&counts, &key, &zero, BPF_NOEXIST); in update_counts() 97 val = bpf_map_lookup_elem(&counts, &key); in update_counts()
|
D | test_lru_dist.c | 175 unsigned int counts = 0; in read_keys() local 191 counts++; in read_keys() 193 counts++; /* in case the last line has no \n */ in read_keys() 195 retkeys = malloc(counts * sizeof(unsigned long long)); in read_keys() 198 counts = 0; in read_keys() 200 retkeys[counts++] = strtoull(l, NULL, 10); in read_keys() 205 return counts; in read_keys()
|
/Linux-v5.4/tools/testing/selftests/bpf/progs/ |
D | test_btf_newkv.c | 38 struct ipv_counts *counts; in test_long_fname_2() local 44 counts = bpf_map_lookup_elem(&btf_map, &key); in test_long_fname_2() 45 if (!counts) in test_long_fname_2() 48 counts->v6++; in test_long_fname_2() 51 counts = bpf_map_lookup_elem(&btf_map_legacy, &key); in test_long_fname_2() 52 if (!counts) in test_long_fname_2()
|
D | test_btf_nokv.c | 28 struct ipv_counts *counts; in test_long_fname_2() local 34 counts = bpf_map_lookup_elem(&btf_map, &key); in test_long_fname_2() 35 if (!counts) in test_long_fname_2() 38 counts->v6++; in test_long_fname_2()
|
D | test_btf_haskv.c | 30 struct ipv_counts *counts; in test_long_fname_2() local 36 counts = bpf_map_lookup_elem(&btf_map, &key); in test_long_fname_2() 37 if (!counts) in test_long_fname_2() 40 counts->v6++; in test_long_fname_2()
|
/Linux-v5.4/drivers/md/ |
D | md-bitmap.c | 932 unsigned long chunk = block >> bitmap->counts.chunkshift; in md_bitmap_file_set_bit() 961 unsigned long chunk = block >> bitmap->counts.chunkshift; in md_bitmap_file_clear_bit() 989 unsigned long chunk = block >> bitmap->counts.chunkshift; in md_bitmap_file_test_bit() 1071 chunks = bitmap->counts.chunks; in md_bitmap_init_from_disk() 1080 int needed = ((sector_t)(i+1) << (bitmap->counts.chunkshift) in md_bitmap_init_from_disk() 1083 (sector_t)i << bitmap->counts.chunkshift, in md_bitmap_init_from_disk() 1161 int needed = ((sector_t)(i+1) << bitmap->counts.chunkshift in md_bitmap_init_from_disk() 1164 (sector_t)i << bitmap->counts.chunkshift, in md_bitmap_init_from_disk() 1236 struct bitmap_counts *counts; in md_bitmap_daemon_work() local 1290 counts = &bitmap->counts; in md_bitmap_daemon_work() [all …]
|
/Linux-v5.4/kernel/gcov/ |
D | gcc_3_4.c | 81 struct gcov_ctr_info counts[0]; member 190 memset(info->counts[i].values, 0, in gcov_info_reset() 191 info->counts[i].num * sizeof(gcov_type)); in gcov_info_reset() 220 for (j = 0; j < dest->counts[i].num; j++) { in gcov_info_add() 221 dest->counts[i].values[j] += in gcov_info_add() 222 source->counts[i].values[j]; in gcov_info_add() 260 dup = kzalloc(struct_size(dup, counts, active), GFP_KERNEL); in gcov_info_dup() 278 struct gcov_ctr_info *ctr = &info->counts[i]; in gcov_info_dup() 281 dup->counts[i].num = ctr->num; in gcov_info_dup() 282 dup->counts[i].merge = ctr->merge; in gcov_info_dup() [all …]
|
/Linux-v5.4/drivers/scsi/ |
D | ch.c | 116 u_int counts[CH_TYPES]; member 229 ch->counts[i]) in ch_elem_to_typecode() 317 ch->counts[CHET_MT] = in ch_readconfig() 321 ch->counts[CHET_ST] = in ch_readconfig() 325 ch->counts[CHET_IE] = in ch_readconfig() 329 ch->counts[CHET_DT] = in ch_readconfig() 333 ch->counts[CHET_MT]); in ch_readconfig() 336 ch->counts[CHET_ST]); in ch_readconfig() 339 ch->counts[CHET_IE]); in ch_readconfig() 342 ch->counts[CHET_DT]); in ch_readconfig() [all …]
|
/Linux-v5.4/arch/x86/kernel/cpu/resctrl/ |
D | pseudo_lock.c | 934 struct residency_counts *counts) in measure_residency_fn() argument 1040 counts->miss_before = miss_before; in measure_residency_fn() 1041 counts->hits_before = hits_before; in measure_residency_fn() 1042 counts->miss_after = miss_after; in measure_residency_fn() 1043 counts->hits_after = hits_after; in measure_residency_fn() 1050 struct residency_counts counts = {0}; in measure_l2_residency() local 1072 measure_residency_fn(&perf_miss_attr, &perf_hit_attr, plr, &counts); in measure_l2_residency() 1077 trace_pseudo_lock_l2(counts.hits_after - counts.hits_before, in measure_l2_residency() 1078 counts.miss_after - counts.miss_before); in measure_l2_residency() 1088 struct residency_counts counts = {0}; in measure_l3_residency() local [all …]
|
/Linux-v5.4/drivers/gpu/drm/msm/disp/dpu1/ |
D | dpu_core_irq.c | 115 int i, ret = 0, counts; in dpu_core_irq_enable() local 122 counts = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idxs[0]]); in dpu_core_irq_enable() 123 if (counts) in dpu_core_irq_enable() 124 DRM_ERROR("irq_idx=%d enable_count=%d\n", irq_idxs[0], counts); in dpu_core_irq_enable() 170 int i, ret = 0, counts; in dpu_core_irq_disable() local 177 counts = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idxs[0]]); in dpu_core_irq_disable() 178 if (counts == 2) in dpu_core_irq_disable() 179 DRM_ERROR("irq_idx=%d enable_count=%d\n", irq_idxs[0], counts); in dpu_core_irq_disable()
|
/Linux-v5.4/drivers/watchdog/ |
D | loongson1_wdt.c | 41 unsigned int counts; in ls1x_wdt_set_timeout() local 45 counts = drvdata->clk_rate * min(timeout, max_hw_heartbeat); in ls1x_wdt_set_timeout() 46 writel(counts, drvdata->base + WDT_TIMER); in ls1x_wdt_set_timeout()
|
/Linux-v5.4/tools/perf/tests/ |
D | openat-syscall.c | 54 if (perf_counts(evsel->counts, 0, 0)->val != nr_openat_calls) { in test__openat_syscall_event() 56 nr_openat_calls, perf_counts(evsel->counts, 0, 0)->val); in test__openat_syscall_event()
|
/Linux-v5.4/arch/ia64/kernel/ |
D | smp.c | 258 unsigned short *counts = __ia64_per_cpu_var(shadow_flush_counts); in smp_flush_tlb_cpumask() local 266 counts[cpu] = local_tlb_flush_counts[cpu].count & 0xffff; in smp_flush_tlb_cpumask() 280 while(counts[cpu] == (local_tlb_flush_counts[cpu].count & 0xffff)) in smp_flush_tlb_cpumask()
|
/Linux-v5.4/tools/perf/Documentation/ |
D | perf-stat.txt | 46 'percore' is a event qualifier that sums up the event counts for both 107 Do not aggregate counts across all monitored CPUs. 119 print counts using a CSV-style output to make it easy to import directly into 197 Aggregate counts per processor socket for system-wide mode measurements. This 204 Aggregate counts per processor die for system-wide mode measurements. This 211 Aggregate counts per physical processor for system-wide mode measurements. This 217 Aggregate counts per monitored threads, when monitoring threads (-t option) 247 Aggregate counts per processor socket for system-wide mode measurements. 250 Aggregate counts per processor die for system-wide mode measurements. 253 Aggregate counts per physical processor for system-wide mode measurements. [all …]
|
/Linux-v5.4/tools/memory-model/scripts/ |
D | cmplitmushist.sh | 36 echo Matching Observation result and counts: $2 71 echo Matching Observation result and counts: $obsline 1>&2
|