| /Linux-v5.4/arch/sh/kernel/ |
| D | perf_event.c | 28 struct cpu_hw_events { struct 34 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 219 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_stop() 237 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_start() 254 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_del() 264 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_add() 354 struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu); in sh_pmu_prepare_cpu() 356 memset(cpuhw, 0, sizeof(struct cpu_hw_events)); in sh_pmu_prepare_cpu()
|
| /Linux-v5.4/arch/x86/events/amd/ |
| D | core.c | 335 static inline int amd_has_nb(struct cpu_hw_events *cpuc) in amd_has_nb() 363 static void __amd_put_nb_event_constraints(struct cpu_hw_events *cpuc, in __amd_put_nb_event_constraints() 420 __amd_get_nb_event_constraints(struct cpu_hw_events *cpuc, struct perf_event *event, in __amd_get_nb_event_constraints() 496 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_prepare() 512 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_starting() 526 nb = per_cpu(cpu_hw_events, i).amd_nb; in amd_pmu_cpu_starting() 543 struct cpu_hw_events *cpuhw; in amd_pmu_cpu_dead() 548 cpuhw = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_dead() 591 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_disable_all() 655 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_handle_irq() [all …]
|
| /Linux-v5.4/arch/x86/events/intel/ |
| D | ds.c | 265 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in init_debug_store_on_cpu() 277 if (!per_cpu(cpu_hw_events, cpu).ds) in fini_debug_store_on_cpu() 336 struct cpu_hw_events *hwev = per_cpu_ptr(&cpu_hw_events, cpu); in alloc_pebs_buffer() 374 struct cpu_hw_events *hwev = per_cpu_ptr(&cpu_hw_events, cpu); in release_pebs_buffer() 392 struct cpu_hw_events *hwev = per_cpu_ptr(&cpu_hw_events, cpu); in alloc_bts_buffer() 421 struct cpu_hw_events *hwev = per_cpu_ptr(&cpu_hw_events, cpu); in release_bts_buffer() 439 per_cpu(cpu_hw_events, cpu).ds = ds; in alloc_ds_buffer() 445 per_cpu(cpu_hw_events, cpu).ds = NULL; in release_ds_buffer() 565 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_disable_bts() 582 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_drain_bts_buffer() [all …]
|
| D | lbr.c | 146 static void intel_pmu_lbr_filter(struct cpu_hw_events *cpuc); 155 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_lbr_enable() 219 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_reset() 342 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_lbr_restore() 390 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_lbr_save() 422 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_sched_task() 459 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_add() 500 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_del() 522 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_enable_all() 530 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_disable_all() [all …]
|
| D | core.c | 1951 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_disable_all() 1969 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_enable_all() 2008 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_nhm_workaround() 2073 static void intel_set_tfa(struct cpu_hw_events *cpuc, bool on) in intel_set_tfa() 2083 static void intel_tfa_commit_scheduling(struct cpu_hw_events *cpuc, int idx, int cntr) in intel_tfa_commit_scheduling() 2094 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_tfa_pmu_enable_all() 2152 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_disable_event() 2234 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_enable_event() 2237 if (!__this_cpu_read(cpu_hw_events.enabled)) in intel_pmu_enable_event() 2294 struct debug_store *ds = __this_cpu_read(cpu_hw_events.ds); in intel_pmu_reset() [all …]
|
| D | bts.c | 141 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in bts_config_buffer() 180 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in bts_update() 250 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_start() 295 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_stop() 440 struct debug_store *ds = this_cpu_ptr(&cpu_hw_events)->ds; in intel_bts_interrupt() 510 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_add()
|
| D | knc.c | 216 struct cpu_hw_events *cpuc; in knc_pmu_handle_irq() 221 cpuc = this_cpu_ptr(&cpu_hw_events); in knc_pmu_handle_irq()
|
| D | p4.c | 918 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_disable_all() 987 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_enable_all() 1001 struct cpu_hw_events *cpuc; in p4_pmu_handle_irq() 1007 cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_handle_irq() 1206 static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) in p4_pmu_schedule_events()
|
| /Linux-v5.4/arch/powerpc/perf/ |
| D | core-fsl-emb.c | 19 struct cpu_hw_events { struct 25 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 205 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_disable() 209 cpuhw = this_cpu_ptr(&cpu_hw_events); in fsl_emb_pmu_disable() 244 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_enable() 248 cpuhw = this_cpu_ptr(&cpu_hw_events); in fsl_emb_pmu_enable() 291 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_add() 298 cpuhw = &get_cpu_var(cpu_hw_events); in fsl_emb_pmu_add() 344 put_cpu_var(cpu_hw_events); in fsl_emb_pmu_add() 352 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_del() [all …]
|
| D | core-book3s.c | 30 struct cpu_hw_events { struct 64 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 122 static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) in ebb_switch_in() 130 static inline void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) {} in power_pmu_bhrb_read() 369 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_bhrb_enable() 385 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_bhrb_disable() 447 static void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) in power_pmu_bhrb_read() 596 static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) in ebb_switch_in() 866 static int power_check_constraints(struct cpu_hw_events *cpuhw, in power_check_constraints() 1113 static void freeze_limited_counters(struct cpu_hw_events *cpuhw, in freeze_limited_counters() [all …]
|
| /Linux-v5.4/arch/alpha/kernel/ |
| D | perf_event.c | 34 struct cpu_hw_events { struct 53 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 391 static void maybe_change_configuration(struct cpu_hw_events *cpuc) in maybe_change_configuration() 435 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_add() 487 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_del() 535 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_stop() 555 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_start() 722 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_enable() 748 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_disable() 807 struct cpu_hw_events *cpuc; in alpha_perf_event_irq_handler() [all …]
|
| /Linux-v5.4/arch/x86/events/ |
| D | core.c | 48 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { 611 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_disable_all() 642 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_disable() 659 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_enable_all() 866 int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) in x86_schedule_events() 999 static int collect_events(struct cpu_hw_events *cpuc, struct perf_event *leader, bool dogrp) in collect_events() 1054 struct cpu_hw_events *cpuc, int i) in x86_assign_hw_event() 1098 struct cpu_hw_events *cpuc, in match_prev_assignment() 1110 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_enable() 1251 if (__this_cpu_read(cpu_hw_events.enabled)) in x86_pmu_enable_event() [all …]
|
| D | perf_event.h | 186 struct cpu_hw_events { struct 580 int (*schedule_events)(struct cpu_hw_events *cpuc, int n, int *assign); 599 (*get_event_constraints)(struct cpu_hw_events *cpuc, 603 void (*put_event_constraints)(struct cpu_hw_events *cpuc, 606 void (*start_scheduling)(struct cpu_hw_events *cpuc); 608 void (*commit_scheduling)(struct cpu_hw_events *cpuc, int idx, int cntr); 610 void (*stop_scheduling)(struct cpu_hw_events *cpuc); 773 DECLARE_PER_CPU(struct cpu_hw_events, cpu_hw_events); 836 u64 disable_mask = __this_cpu_read(cpu_hw_events.perf_ctr_virt_mask); in __x86_pmu_enable_event() 847 int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign); [all …]
|
| /Linux-v5.4/arch/sparc/kernel/ |
| D | perf_event.c | 78 struct cpu_hw_events { struct 115 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { .enabled = 1, }; argument 827 static inline void sparc_pmu_enable_event(struct cpu_hw_events *cpuc, struct hw_perf_event *hwc, in… in sparc_pmu_enable_event() 845 static inline void sparc_pmu_disable_event(struct cpu_hw_events *cpuc, struct hw_perf_event *hwc, i… in sparc_pmu_disable_event() 923 static void read_in_all_counters(struct cpu_hw_events *cpuc) in read_in_all_counters() 947 static void calculate_single_pcr(struct cpu_hw_events *cpuc) in calculate_single_pcr() 983 static void calculate_multiple_pcrs(struct cpu_hw_events *cpuc) in calculate_multiple_pcrs() 1017 static void update_pcrs_for_enable(struct cpu_hw_events *cpuc) in update_pcrs_for_enable() 1031 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_enable() 1049 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_disable() [all …]
|
| /Linux-v5.4/arch/riscv/kernel/ |
| D | perf_event.c | 36 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); 299 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in riscv_pmu_add() 330 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in riscv_pmu_del()
|
| /Linux-v5.4/arch/mips/kernel/ |
| D | perf_event_mipsxx.c | 31 struct cpu_hw_events { struct 48 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { argument 285 static int mipsxx_pmu_alloc_counter(struct cpu_hw_events *cpuc, in mipsxx_pmu_alloc_counter() 318 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_enable_event() 358 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_disable_event() 458 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipspmu_add() 494 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipspmu_del() 714 struct cpu_hw_events fake_cpuc; in validate_group() 733 static void handle_associated_event(struct cpu_hw_events *cpuc, in handle_associated_event() 1356 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in pause_local_counters() [all …]
|
| /Linux-v5.4/arch/riscv/include/asm/ |
| D | perf_event.h | 50 struct cpu_hw_events { struct
|
| /Linux-v5.4/arch/nds32/kernel/ |
| D | perf_event_cpu.c | 33 static DEFINE_PER_CPU(struct pmu_hw_events, cpu_hw_events); 1068 return this_cpu_ptr(&cpu_hw_events); in cpu_pmu_get_cpu_events() 1109 struct pmu_hw_events *events = &per_cpu(cpu_hw_events, cpu); in cpu_pmu_init()
|