/Linux-v4.19/arch/x86/kvm/ |
D | pmu_intel.c | 38 static void reprogram_fixed_counters(struct kvm_pmu *pmu, u64 data) in reprogram_fixed_counters() argument 42 for (i = 0; i < pmu->nr_arch_fixed_counters; i++) { in reprogram_fixed_counters() 44 u8 old_ctrl = fixed_ctrl_field(pmu->fixed_ctr_ctrl, i); in reprogram_fixed_counters() 47 pmc = get_fixed_pmc(pmu, MSR_CORE_PERF_FIXED_CTR0 + i); in reprogram_fixed_counters() 55 pmu->fixed_ctr_ctrl = data; in reprogram_fixed_counters() 59 static void global_ctrl_changed(struct kvm_pmu *pmu, u64 data) in global_ctrl_changed() argument 62 u64 diff = pmu->global_ctrl ^ data; in global_ctrl_changed() 64 pmu->global_ctrl = data; in global_ctrl_changed() 67 reprogram_counter(pmu, bit); in global_ctrl_changed() 70 static unsigned intel_find_arch_event(struct kvm_pmu *pmu, in intel_find_arch_event() argument [all …]
|
D | pmu_amd.c | 49 static unsigned int get_msr_base(struct kvm_pmu *pmu, enum pmu_type type) in get_msr_base() argument 51 struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); in get_msr_base() 100 static inline struct kvm_pmc *get_gp_pmc_amd(struct kvm_pmu *pmu, u32 msr, in get_gp_pmc_amd() argument 128 return &pmu->gp_counters[msr_to_index(msr)]; in get_gp_pmc_amd() 131 static unsigned amd_find_arch_event(struct kvm_pmu *pmu, in amd_find_arch_event() argument 162 static struct kvm_pmc *amd_pmc_idx_to_pmc(struct kvm_pmu *pmu, int pmc_idx) in amd_pmc_idx_to_pmc() argument 164 unsigned int base = get_msr_base(pmu, PMU_TYPE_COUNTER); in amd_pmc_idx_to_pmc() 165 struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); in amd_pmc_idx_to_pmc() 175 return get_gp_pmc_amd(pmu, base + pmc_idx, PMU_TYPE_COUNTER); in amd_pmc_idx_to_pmc() 181 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_is_valid_msr_idx() local [all …]
|
D | pmu.h | 5 #define vcpu_to_pmu(vcpu) (&(vcpu)->arch.pmu) 6 #define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu)) argument 7 #define pmc_to_pmu(pmc) (&(pmc)->vcpu->arch.pmu) 23 unsigned (*find_arch_event)(struct kvm_pmu *pmu, u8 event_select, 27 struct kvm_pmc *(*pmc_idx_to_pmc)(struct kvm_pmu *pmu, int pmc_idx); 40 struct kvm_pmu *pmu = pmc_to_pmu(pmc); in pmc_bitmask() local 42 return pmu->counter_bitmask[pmc->type]; in pmc_bitmask() 85 static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr, in get_gp_pmc() argument 88 if (msr >= base && msr < base + pmu->nr_arch_gp_counters) in get_gp_pmc() 89 return &pmu->gp_counters[msr - base]; in get_gp_pmc() [all …]
|
/Linux-v4.19/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
D | base.c | 30 nvkm_pmu_pgob(struct nvkm_pmu *pmu, bool enable) in nvkm_pmu_pgob() argument 32 if (pmu && pmu->func->pgob) in nvkm_pmu_pgob() 33 pmu->func->pgob(pmu, enable); in nvkm_pmu_pgob() 39 struct nvkm_pmu *pmu = container_of(work, typeof(*pmu), recv.work); in nvkm_pmu_recv() local 40 return pmu->func->recv(pmu); in nvkm_pmu_recv() 44 nvkm_pmu_send(struct nvkm_pmu *pmu, u32 reply[2], in nvkm_pmu_send() argument 47 if (!pmu || !pmu->func->send) in nvkm_pmu_send() 49 return pmu->func->send(pmu, reply, process, message, data0, data1); in nvkm_pmu_send() 55 struct nvkm_pmu *pmu = nvkm_pmu(subdev); in nvkm_pmu_intr() local 56 if (!pmu->func->intr) in nvkm_pmu_intr() [all …]
|
D | gk20a.c | 51 gk20a_pmu_dvfs_target(struct gk20a_pmu *pmu, int *state) in gk20a_pmu_dvfs_target() argument 53 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_target() 59 gk20a_pmu_dvfs_get_cur_state(struct gk20a_pmu *pmu, int *state) in gk20a_pmu_dvfs_get_cur_state() argument 61 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_get_cur_state() 67 gk20a_pmu_dvfs_get_target_state(struct gk20a_pmu *pmu, in gk20a_pmu_dvfs_get_target_state() argument 70 struct gk20a_pmu_dvfs_data *data = pmu->data; in gk20a_pmu_dvfs_get_target_state() 71 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_get_target_state() 86 nvkm_trace(&pmu->base.subdev, "cur level = %d, new level = %d\n", in gk20a_pmu_dvfs_get_target_state() 95 gk20a_pmu_dvfs_get_dev_status(struct gk20a_pmu *pmu, in gk20a_pmu_dvfs_get_dev_status() argument 98 struct nvkm_falcon *falcon = pmu->base.falcon; in gk20a_pmu_dvfs_get_dev_status() [all …]
|
D | gt215.c | 30 gt215_pmu_send(struct nvkm_pmu *pmu, u32 reply[2], in gt215_pmu_send() argument 33 struct nvkm_subdev *subdev = &pmu->subdev; in gt215_pmu_send() 54 pmu->recv.message = message; in gt215_pmu_send() 55 pmu->recv.process = process; in gt215_pmu_send() 65 pmu->send.base)); in gt215_pmu_send() 77 wait_event(pmu->recv.wait, (pmu->recv.process == 0)); in gt215_pmu_send() 78 reply[0] = pmu->recv.data[0]; in gt215_pmu_send() 79 reply[1] = pmu->recv.data[1]; in gt215_pmu_send() 87 gt215_pmu_recv(struct nvkm_pmu *pmu) in gt215_pmu_recv() argument 89 struct nvkm_subdev *subdev = &pmu->subdev; in gt215_pmu_recv() [all …]
|
D | Kbuild | 1 nvkm-y += nvkm/subdev/pmu/base.o 2 nvkm-y += nvkm/subdev/pmu/memx.o 3 nvkm-y += nvkm/subdev/pmu/gt215.o 4 nvkm-y += nvkm/subdev/pmu/gf100.o 5 nvkm-y += nvkm/subdev/pmu/gf119.o 6 nvkm-y += nvkm/subdev/pmu/gk104.o 7 nvkm-y += nvkm/subdev/pmu/gk110.o 8 nvkm-y += nvkm/subdev/pmu/gk208.o 9 nvkm-y += nvkm/subdev/pmu/gk20a.o 10 nvkm-y += nvkm/subdev/pmu/gm107.o [all …]
|
D | memx.c | 7 struct nvkm_pmu *pmu; member 20 struct nvkm_device *device = memx->pmu->subdev.device; in memx_out() 44 nvkm_memx_init(struct nvkm_pmu *pmu, struct nvkm_memx **pmemx) in nvkm_memx_init() argument 46 struct nvkm_device *device = pmu->subdev.device; in nvkm_memx_init() 51 ret = nvkm_pmu_send(pmu, reply, PROC_MEMX, MEMX_MSG_INFO, in nvkm_memx_init() 59 memx->pmu = pmu; in nvkm_memx_init() 75 struct nvkm_pmu *pmu = memx->pmu; in nvkm_memx_fini() local 76 struct nvkm_subdev *subdev = &pmu->subdev; in nvkm_memx_fini() 89 nvkm_pmu_send(pmu, reply, PROC_MEMX, MEMX_MSG_EXEC, in nvkm_memx_fini() 102 nvkm_debug(&memx->pmu->subdev, "R[%06x] = %08x\n", addr, data); in nvkm_memx_wr32() [all …]
|
/Linux-v4.19/drivers/soc/dove/ |
D | pmu.c | 50 struct pmu_data *pmu = rcdev_to_pmu(rc); in pmu_reset_reset() local 54 spin_lock_irqsave(&pmu->lock, flags); in pmu_reset_reset() 55 val = readl_relaxed(pmu->pmc_base + PMC_SW_RST); in pmu_reset_reset() 56 writel_relaxed(val & ~BIT(id), pmu->pmc_base + PMC_SW_RST); in pmu_reset_reset() 57 writel_relaxed(val | BIT(id), pmu->pmc_base + PMC_SW_RST); in pmu_reset_reset() 58 spin_unlock_irqrestore(&pmu->lock, flags); in pmu_reset_reset() 65 struct pmu_data *pmu = rcdev_to_pmu(rc); in pmu_reset_assert() local 69 spin_lock_irqsave(&pmu->lock, flags); in pmu_reset_assert() 70 val &= readl_relaxed(pmu->pmc_base + PMC_SW_RST); in pmu_reset_assert() 71 writel_relaxed(val, pmu->pmc_base + PMC_SW_RST); in pmu_reset_assert() [all …]
|
/Linux-v4.19/drivers/perf/ |
D | arm_pmu_acpi.c | 127 struct arm_pmu *pmu; in arm_pmu_acpi_find_alloc_pmu() local 131 pmu = per_cpu(probed_pmus, cpu); in arm_pmu_acpi_find_alloc_pmu() 132 if (!pmu || pmu->acpi_cpuid != cpuid) in arm_pmu_acpi_find_alloc_pmu() 135 return pmu; in arm_pmu_acpi_find_alloc_pmu() 138 pmu = armpmu_alloc_atomic(); in arm_pmu_acpi_find_alloc_pmu() 139 if (!pmu) { in arm_pmu_acpi_find_alloc_pmu() 145 pmu->acpi_cpuid = cpuid; in arm_pmu_acpi_find_alloc_pmu() 147 return pmu; in arm_pmu_acpi_find_alloc_pmu() 154 static bool pmu_irq_matches(struct arm_pmu *pmu, int irq) in pmu_irq_matches() argument 156 struct pmu_hw_events __percpu *hw_events = pmu->hw_events; in pmu_irq_matches() [all …]
|
D | arm_pmu_platform.c | 25 static int probe_current_pmu(struct arm_pmu *pmu, in probe_current_pmu() argument 37 ret = info->init(pmu); in probe_current_pmu() 45 static int pmu_parse_percpu_irq(struct arm_pmu *pmu, int irq) in pmu_parse_percpu_irq() argument 48 struct pmu_hw_events __percpu *hw_events = pmu->hw_events; in pmu_parse_percpu_irq() 50 ret = irq_get_percpu_devid_partition(irq, &pmu->supported_cpus); in pmu_parse_percpu_irq() 54 for_each_cpu(cpu, &pmu->supported_cpus) in pmu_parse_percpu_irq() 96 static int pmu_parse_irqs(struct arm_pmu *pmu) in pmu_parse_irqs() argument 99 struct platform_device *pdev = pmu->plat_device; in pmu_parse_irqs() 100 struct pmu_hw_events __percpu *hw_events = pmu->hw_events; in pmu_parse_irqs() 114 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_parse_irqs() [all …]
|
D | arm_pmu.c | 104 if (type == event->pmu->type) in armpmu_map_event() 121 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period() 163 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update() 193 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop() 209 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start() 234 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del() 250 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add() 283 validate_event(struct pmu *pmu, struct pmu_hw_events *hw_events, in validate_event() argument 296 if (event->pmu != pmu) in validate_event() 305 armpmu = to_arm_pmu(event->pmu); in validate_event() [all …]
|
/Linux-v4.19/drivers/gpu/drm/i915/ |
D | i915_pmu.c | 82 enable = i915->pmu.enable; in pmu_needs_timer() 115 if (!i915->pmu.base.event_init) in i915_pmu_gt_parked() 118 spin_lock_irq(&i915->pmu.lock); in i915_pmu_gt_parked() 123 i915->pmu.timer_enabled = pmu_needs_timer(i915, false); in i915_pmu_gt_parked() 124 spin_unlock_irq(&i915->pmu.lock); in i915_pmu_gt_parked() 129 if (!i915->pmu.timer_enabled && pmu_needs_timer(i915, true)) { in __i915_pmu_maybe_start_timer() 130 i915->pmu.timer_enabled = true; in __i915_pmu_maybe_start_timer() 131 i915->pmu.timer_last = ktime_get(); in __i915_pmu_maybe_start_timer() 132 hrtimer_start_range_ns(&i915->pmu.timer, in __i915_pmu_maybe_start_timer() 140 if (!i915->pmu.base.event_init) in i915_pmu_gt_unparked() [all …]
|
/Linux-v4.19/arch/x86/events/intel/ |
D | rapl.c | 144 struct pmu *pmu; member 150 struct pmu pmu; member 230 static void rapl_start_hrtimer(struct rapl_pmu *pmu) in rapl_start_hrtimer() argument 232 hrtimer_start(&pmu->hrtimer, pmu->timer_interval, in rapl_start_hrtimer() 238 struct rapl_pmu *pmu = container_of(hrtimer, struct rapl_pmu, hrtimer); in rapl_hrtimer_handle() local 242 if (!pmu->n_active) in rapl_hrtimer_handle() 245 raw_spin_lock_irqsave(&pmu->lock, flags); in rapl_hrtimer_handle() 247 list_for_each_entry(event, &pmu->active_list, active_entry) in rapl_hrtimer_handle() 250 raw_spin_unlock_irqrestore(&pmu->lock, flags); in rapl_hrtimer_handle() 252 hrtimer_forward_now(hrtimer, pmu->timer_interval); in rapl_hrtimer_handle() [all …]
|
D | uncore.h | 71 struct pmu *pmu; /* for custom pmu ops */ member 93 struct pmu pmu; member 123 struct intel_uncore_pmu *pmu; member 188 return box->pmu->type->box_ctl; in uncore_pci_box_ctl() 193 return box->pmu->type->fixed_ctl; in uncore_pci_fixed_ctl() 198 return box->pmu->type->fixed_ctr; in uncore_pci_fixed_ctr() 205 return idx * 8 + box->pmu->type->event_ctl; in uncore_pci_event_ctl() 207 return idx * 4 + box->pmu->type->event_ctl; in uncore_pci_event_ctl() 213 return idx * 8 + box->pmu->type->perf_ctr; in uncore_pci_perf_ctr() 218 struct intel_uncore_pmu *pmu = box->pmu; in uncore_msr_box_offset() local [all …]
|
D | uncore.c | 101 struct intel_uncore_box *uncore_pmu_to_box(struct intel_uncore_pmu *pmu, int cpu) in uncore_pmu_to_box() argument 109 return pkgid < max_packages ? pmu->boxes[pkgid] : NULL; in uncore_pmu_to_box() 332 return &box->pmu->pmu == event->pmu; in is_box_event() 342 max_count = box->pmu->type->num_counters; in uncore_collect_events() 343 if (box->pmu->type->fixed_ctl) in uncore_collect_events() 376 struct intel_uncore_type *type = box->pmu->type; in uncore_get_event_constraint() 401 if (box->pmu->type->ops->put_constraint) in uncore_put_event_constraint() 402 box->pmu->type->ops->put_constraint(box, event); in uncore_put_event_constraint() 644 static int uncore_validate_group(struct intel_uncore_pmu *pmu, in uncore_validate_group() argument 655 fake_box = uncore_alloc_box(pmu->type, NUMA_NO_NODE); in uncore_validate_group() [all …]
|
/Linux-v4.19/Documentation/devicetree/bindings/pinctrl/ |
D | marvell,dove-pinctrl.txt | 14 Note: pmu* also allows for Power Management functions listed below 18 mpp0 0 gpio, pmu, uart2(rts), sdio0(cd), lcd0(pwm), pmu* 19 mpp1 1 gpio, pmu, uart2(cts), sdio0(wp), lcd1(pwm), pmu* 20 mpp2 2 gpio, pmu, uart2(txd), sdio0(buspwr), sata(prsnt), 21 uart1(rts), pmu* 22 mpp3 3 gpio, pmu, uart2(rxd), sdio0(ledctrl), sata(act), 23 uart1(cts), lcd-spi(cs1), pmu* 24 mpp4 4 gpio, pmu, uart3(rts), sdio1(cd), spi1(miso), pmu* 25 mpp5 5 gpio, pmu, uart3(cts), sdio1(wp), spi1(cs), pmu* 26 mpp6 6 gpio, pmu, uart3(txd), sdio1(buspwr), spi1(mosi), pmu* [all …]
|
/Linux-v4.19/drivers/soc/rockchip/ |
D | pm_domains.c | 70 struct rockchip_pmu *pmu; member 137 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_idle() local 141 regmap_read(pmu->regmap, pmu->info->idle_offset, &val); in rockchip_pmu_domain_is_idle() 145 static unsigned int rockchip_pmu_read_ack(struct rockchip_pmu *pmu) in rockchip_pmu_read_ack() argument 149 regmap_read(pmu->regmap, pmu->info->ack_offset, &val); in rockchip_pmu_read_ack() 158 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_set_idle_request() local 167 regmap_write(pmu->regmap, pmu->info->req_offset, in rockchip_pmu_set_idle_request() 171 regmap_update_bits(pmu->regmap, pmu->info->req_offset, in rockchip_pmu_set_idle_request() 178 ret = readx_poll_timeout_atomic(rockchip_pmu_read_ack, pmu, val, in rockchip_pmu_set_idle_request() 182 dev_err(pmu->dev, in rockchip_pmu_set_idle_request() [all …]
|
/Linux-v4.19/arch/powerpc/perf/ |
D | imc-pmu.c | 48 return container_of(event->pmu, struct imc_pmu, pmu); in imc_event_to_pmu() 73 struct pmu *pmu = dev_get_drvdata(dev); in imc_pmu_cpumask_get_attr() local 74 struct imc_pmu *imc_pmu = container_of(pmu, struct imc_pmu, pmu); in imc_pmu_cpumask_get_attr() 189 static int update_events_in_group(struct device_node *node, struct imc_pmu *pmu) in update_events_in_group() argument 226 pmu->events = kcalloc(ct, sizeof(struct imc_events), GFP_KERNEL); in update_events_in_group() 227 if (!pmu->events) in update_events_in_group() 233 ret = imc_parse_event(np, g_scale, g_unit, prefix, base_reg, &pmu->events[ct]); in update_events_in_group() 241 imc_free_events(pmu->events, ct); in update_events_in_group() 256 imc_free_events(pmu->events, ct); in update_events_in_group() 263 ev_val_str = kasprintf(GFP_KERNEL, "event=0x%x", pmu->events[i].value); in update_events_in_group() [all …]
|
D | Makefile | 7 obj64-$(CONFIG_PPC_PERF_CTRS) += ppc970-pmu.o power5-pmu.o \ 8 power5+-pmu.o power6-pmu.o power7-pmu.o \ 9 isa207-common.o power8-pmu.o power9-pmu.o 10 obj32-$(CONFIG_PPC_PERF_CTRS) += mpc7450-pmu.o 12 obj-$(CONFIG_PPC_POWERNV) += imc-pmu.o 14 obj-$(CONFIG_FSL_EMB_PERF_EVENT_E500) += e500-pmu.o e6500-pmu.o 18 obj-$(CONFIG_PPC_8xx) += 8xx-pmu.o
|
/Linux-v4.19/virt/kvm/arm/ |
D | pmu.c | 35 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_get_counter_value() local 36 struct kvm_pmc *pmc = &pmu->pmc[select_idx]; in kvm_pmu_get_counter_value() 96 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_reset() local 99 kvm_pmu_stop_counter(vcpu, &pmu->pmc[i]); in kvm_pmu_vcpu_reset() 100 pmu->pmc[i].idx = i; in kvm_pmu_vcpu_reset() 101 pmu->pmc[i].bitmask = 0xffffffffUL; in kvm_pmu_vcpu_reset() 113 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_destroy() local 116 struct kvm_pmc *pmc = &pmu->pmc[i]; in kvm_pmu_vcpu_destroy() 147 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_enable_counter() local 157 pmc = &pmu->pmc[i]; in kvm_pmu_enable_counter() [all …]
|
/Linux-v4.19/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
D | gm200.c | 32 pmu_code(struct nv50_devinit *init, u32 pmu, u32 img, u32 len, bool sec) in pmu_code() argument 38 nvkm_wr32(device, 0x10a180, 0x01000000 | (sec ? 0x10000000 : 0) | pmu); in pmu_code() 41 nvkm_wr32(device, 0x10a188, (pmu + i) >> 8); in pmu_code() 52 pmu_data(struct nv50_devinit *init, u32 pmu, u32 img, u32 len) in pmu_data() argument 58 nvkm_wr32(device, 0x10a1c0, 0x01000000 | pmu); in pmu_data() 87 struct nvbios_pmuR pmu; in pmu_load() local 89 if (!nvbios_pmuRm(bios, type, &pmu)) in pmu_load() 95 pmu_code(init, pmu.boot_addr_pmu, pmu.boot_addr, pmu.boot_size, false); in pmu_load() 96 pmu_code(init, pmu.code_addr_pmu, pmu.code_addr, pmu.code_size, true); in pmu_load() 97 pmu_data(init, pmu.data_addr_pmu, pmu.data_addr, pmu.data_size); in pmu_load() [all …]
|
/Linux-v4.19/Documentation/devicetree/bindings/arm/ |
D | pmu.txt | 10 "apm,potenza-pmu" 12 "arm,cortex-a73-pmu" 13 "arm,cortex-a72-pmu" 14 "arm,cortex-a57-pmu" 15 "arm,cortex-a53-pmu" 16 "arm,cortex-a35-pmu" 17 "arm,cortex-a17-pmu" 18 "arm,cortex-a15-pmu" 19 "arm,cortex-a12-pmu" 20 "arm,cortex-a9-pmu" [all …]
|
/Linux-v4.19/tools/perf/util/ |
D | pmu.c | 664 char * __weak get_cpuid_str(struct perf_pmu *pmu __maybe_unused) in get_cpuid_str() 697 static char *perf_pmu__getcpuid(struct perf_pmu *pmu) in perf_pmu__getcpuid() argument 706 cpuid = get_cpuid_str(pmu); in perf_pmu__getcpuid() 717 struct pmu_events_map *perf_pmu__find_map(struct perf_pmu *pmu) in perf_pmu__find_map() argument 720 char *cpuid = perf_pmu__getcpuid(pmu); in perf_pmu__find_map() 749 static void pmu_add_cpu_aliases(struct list_head *head, struct perf_pmu *pmu) in pmu_add_cpu_aliases() argument 754 const char *name = pmu->name; in pmu_add_cpu_aliases() 757 map = perf_pmu__find_map(pmu); in pmu_add_cpu_aliases() 775 pname = pe->pmu ? pe->pmu : "cpu"; in pmu_add_cpu_aliases() 791 perf_pmu__get_default_config(struct perf_pmu *pmu __maybe_unused) in perf_pmu__get_default_config() [all …]
|
/Linux-v4.19/tools/perf/pmu-events/ |
D | Build | 5 pmu-events-y += pmu-events.o 6 JDIR = pmu-events/arch/$(SRCARCH) 11 # Locate/process JSON files in pmu-events/arch/ 12 # directory and create tables in pmu-events.c. 14 $(OUTPUT)pmu-events/pmu-events.c: $(JSON) $(JEVENTS) 15 $(Q)$(call echo-cmd,gen)$(JEVENTS) $(SRCARCH) pmu-events/arch $(OUTPUT)pmu-events/pmu-events.c $(V)
|