/Linux-v4.19/arch/x86/oprofile/ |
D | op_model_ppro.c | 33 static void ppro_shutdown(struct op_msrs const * const msrs) in ppro_shutdown() argument 38 if (!msrs->counters[i].addr) in ppro_shutdown() 45 static int ppro_fill_in_addresses(struct op_msrs * const msrs) in ppro_fill_in_addresses() argument 57 msrs->counters[i].addr = MSR_P6_PERFCTR0 + i; in ppro_fill_in_addresses() 58 msrs->controls[i].addr = MSR_P6_EVNTSEL0 + i; in ppro_fill_in_addresses() 64 ppro_shutdown(msrs); in ppro_fill_in_addresses() 73 struct op_msrs const * const msrs) in ppro_setup_ctrs() argument 97 if (!msrs->controls[i].addr) in ppro_setup_ctrs() 99 rdmsrl(msrs->controls[i].addr, val); in ppro_setup_ctrs() 103 wrmsrl(msrs->controls[i].addr, val); in ppro_setup_ctrs() [all …]
|
D | op_model_amd.c | 133 struct op_msrs const * const msrs) in op_amd_handle_ibs() argument 269 struct op_msrs const * const msrs) in op_mux_switch_ctrl() argument 279 rdmsrl(msrs->controls[i].addr, val); in op_mux_switch_ctrl() 282 wrmsrl(msrs->controls[i].addr, val); in op_mux_switch_ctrl() 290 static void op_amd_shutdown(struct op_msrs const * const msrs) in op_amd_shutdown() argument 295 if (!msrs->counters[i].addr) in op_amd_shutdown() 302 static int op_amd_fill_in_addresses(struct op_msrs * const msrs) in op_amd_fill_in_addresses() argument 315 msrs->counters[i].addr = MSR_F15H_PERF_CTR + (i << 1); in op_amd_fill_in_addresses() 316 msrs->controls[i].addr = MSR_F15H_PERF_CTL + (i << 1); in op_amd_fill_in_addresses() 318 msrs->controls[i].addr = MSR_K7_EVNTSEL0 + i; in op_amd_fill_in_addresses() [all …]
|
D | nmi_int.c | 75 static void nmi_cpu_save_registers(struct op_msrs *msrs) in nmi_cpu_save_registers() argument 77 struct op_msr *counters = msrs->counters; in nmi_cpu_save_registers() 78 struct op_msr *controls = msrs->controls; in nmi_cpu_save_registers() 94 struct op_msrs const *msrs = this_cpu_ptr(&cpu_msrs); in nmi_cpu_start() local 95 if (!msrs->controls) in nmi_cpu_start() 98 model->start(msrs); in nmi_cpu_start() 114 struct op_msrs const *msrs = this_cpu_ptr(&cpu_msrs); in nmi_cpu_stop() local 115 if (!msrs->controls) in nmi_cpu_stop() 118 model->stop(msrs); in nmi_cpu_stop() 181 static void nmi_cpu_setup_mux(int cpu, struct op_msrs const * const msrs) in nmi_cpu_setup_mux() argument [all …]
|
D | op_model_p4.c | 388 static void p4_shutdown(struct op_msrs const * const msrs) in p4_shutdown() argument 393 if (msrs->counters[i].addr) in p4_shutdown() 394 release_perfctr_nmi(msrs->counters[i].addr); in p4_shutdown() 402 if (msrs->controls[i].addr) in p4_shutdown() 403 release_evntsel_nmi(msrs->controls[i].addr); in p4_shutdown() 407 static int p4_fill_in_addresses(struct op_msrs * const msrs) in p4_fill_in_addresses() argument 420 msrs->counters[i].addr = addr; in p4_fill_in_addresses() 421 msrs->controls[i].addr = cccraddr; in p4_fill_in_addresses() 429 msrs->controls[i].addr = addr; in p4_fill_in_addresses() 438 msrs->controls[i].addr = addr; in p4_fill_in_addresses() [all …]
|
D | op_x86_model.h | 43 int (*fill_in_addresses)(struct op_msrs * const msrs); 45 struct op_msrs const * const msrs); 47 struct op_msrs const * const msrs); 48 void (*start)(struct op_msrs const * const msrs); 49 void (*stop)(struct op_msrs const * const msrs); 50 void (*shutdown)(struct op_msrs const * const msrs); 53 struct op_msrs const * const msrs);
|
/Linux-v4.19/arch/x86/lib/ |
D | msr.c | 11 struct msr *msrs = NULL; in msrs_alloc() local 13 msrs = alloc_percpu(struct msr); in msrs_alloc() 14 if (!msrs) { in msrs_alloc() 19 return msrs; in msrs_alloc() 23 void msrs_free(struct msr *msrs) in msrs_free() argument 25 free_percpu(msrs); in msrs_free()
|
D | msr-smp.c | 14 if (rv->msrs) in __rdmsr_on_cpu() 15 reg = per_cpu_ptr(rv->msrs, this_cpu); in __rdmsr_on_cpu() 28 if (rv->msrs) in __wrmsr_on_cpu() 29 reg = per_cpu_ptr(rv->msrs, this_cpu); in __wrmsr_on_cpu() 100 struct msr *msrs, in __rwmsr_on_cpus() argument 108 rv.msrs = msrs; in __rwmsr_on_cpus() 127 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) in rdmsr_on_cpus() argument 129 __rwmsr_on_cpus(mask, msr_no, msrs, __rdmsr_on_cpu); in rdmsr_on_cpus() 141 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) in wrmsr_on_cpus() argument 143 __rwmsr_on_cpus(mask, msr_no, msrs, __wrmsr_on_cpu); in wrmsr_on_cpus()
|
/Linux-v4.19/Documentation/trace/postprocess/ |
D | decode_msr.py | 7 msrs = dict() variable 13 msrs[int(m.group(2), 16)] = m.group(1) 26 if num in msrs: 27 r = msrs[num]
|
/Linux-v4.19/arch/x86/include/asm/ |
D | msr.h | 27 struct msr *msrs; member 323 void msrs_free(struct msr *msrs); 332 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs); 333 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs); 362 struct msr *msrs) in rdmsr_on_cpus() argument 364 rdmsr_on_cpu(0, msr_no, &(msrs[0].l), &(msrs[0].h)); in rdmsr_on_cpus() 367 struct msr *msrs) in wrmsr_on_cpus() argument 369 wrmsr_on_cpu(0, msr_no, msrs[0].l, msrs[0].h); in wrmsr_on_cpus()
|
/Linux-v4.19/drivers/powercap/ |
D | intel_rapl.c | 169 int msrs[RAPL_DOMAIN_MSR_MAX]; member 652 rd->msrs[0] = MSR_PKG_POWER_LIMIT; in rapl_init_domains() 653 rd->msrs[1] = MSR_PKG_ENERGY_STATUS; in rapl_init_domains() 654 rd->msrs[2] = MSR_PKG_PERF_STATUS; in rapl_init_domains() 655 rd->msrs[3] = 0; in rapl_init_domains() 656 rd->msrs[4] = MSR_PKG_POWER_INFO; in rapl_init_domains() 665 rd->msrs[0] = MSR_PP0_POWER_LIMIT; in rapl_init_domains() 666 rd->msrs[1] = MSR_PP0_ENERGY_STATUS; in rapl_init_domains() 667 rd->msrs[2] = 0; in rapl_init_domains() 668 rd->msrs[3] = MSR_PP0_POLICY; in rapl_init_domains() [all …]
|
/Linux-v4.19/arch/x86/kvm/ |
D | vmx.c | 864 struct nested_vmx_msrs msrs; member 1941 return vmx_misc_cr3_count(to_vmx(vcpu)->nested.msrs.misc_low); in nested_cpu_vmx_misc_cr3_count() 1951 return to_vmx(vcpu)->nested.msrs.misc_low & in nested_cpu_has_vmwrite_any_field() 1957 return to_vmx(vcpu)->nested.msrs.misc_low & VMX_MISC_ZERO_LEN_INS; in nested_cpu_has_zero_length_injection() 1962 return to_vmx(vcpu)->nested.msrs.procbased_ctls_high & in nested_cpu_supports_monitor_trap_flag() 1968 return to_vmx(vcpu)->nested.msrs.secondary_ctls_high & in nested_cpu_has_vmx_shadow_vmcs() 3479 static void nested_vmx_setup_ctls_msrs(struct nested_vmx_msrs *msrs, bool apicv) in nested_vmx_setup_ctls_msrs() argument 3482 memset(msrs, 0, sizeof(*msrs)); in nested_vmx_setup_ctls_msrs() 3503 msrs->pinbased_ctls_low, in nested_vmx_setup_ctls_msrs() 3504 msrs->pinbased_ctls_high); in nested_vmx_setup_ctls_msrs() [all …]
|
D | x86.c | 157 u32 msrs[KVM_NR_SHARED_MSRS]; member 246 wrmsrl(shared_msrs_global.msrs[slot], values->host); in kvm_on_user_return() 272 shared_msrs_global.msrs[slot] = msr; in kvm_define_shared_msr() 283 shared_msr_update(i, shared_msrs_global.msrs[i]); in kvm_shared_msr_cpu_online() 295 err = wrmsrl_safe(shared_msrs_global.msrs[slot], value); in kvm_set_shared_msr() 2814 static int __msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs, in __msr_io() argument 2821 for (i = 0; i < msrs->nmsrs; ++i) in __msr_io() 2838 struct kvm_msrs msrs; in msr_io() local 2844 if (copy_from_user(&msrs, user_msrs, sizeof msrs)) in msr_io() 2848 if (msrs.nmsrs >= MAX_IO_MSRS) in msr_io() [all …]
|
/Linux-v4.19/tools/testing/selftests/kvm/lib/ |
D | x86.c | 757 struct kvm_msrs msrs; member 795 state = malloc(sizeof(*state) + nmsrs * sizeof(state->msrs.entries[0])); in vcpu_save_state() 831 state->msrs.nmsrs = nmsrs; in vcpu_save_state() 833 state->msrs.entries[i].index = list->indices[i]; in vcpu_save_state() 834 r = ioctl(vcpu->fd, KVM_GET_MSRS, &state->msrs); in vcpu_save_state() 869 r = ioctl(vcpu->fd, KVM_SET_MSRS, &state->msrs); in vcpu_load_state() 870 … TEST_ASSERT(r == state->msrs.nmsrs, "Unexpected result from KVM_SET_MSRS, r: %i (failed at %x)", in vcpu_load_state() 871 r, r == state->msrs.nmsrs ? -1 : state->msrs.entries[r].index); in vcpu_load_state()
|
/Linux-v4.19/Documentation/virtual/kvm/ |
D | cpuid.txt | 32 KVM_FEATURE_CLOCKSOURCE || 0 || kvmclock available at msrs 40 KVM_FEATURE_CLOCKSOURCE2 || 3 || kvmclock available at msrs
|
D | api.txt | 139 __u32 nmsrs; /* number of msrs in entries */ 144 kvm adjusts nmsrs to reflect the actual number of msrs and fills in the 147 KVM_GET_MSR_INDEX_LIST returns the guest msrs that are supported. The list 490 Returns: number of msrs successfully returned; 504 __u32 nmsrs; /* number of msrs in entries */
|
/Linux-v4.19/drivers/edac/ |
D | amd64_edac.c | 16 static struct msr __percpu *msrs; variable 2887 rdmsr_on_cpus(mask, MSR_IA32_MCG_CTL, msrs); in nb_mce_bank_enabled_on_node() 2890 struct msr *reg = per_cpu_ptr(msrs, cpu); in nb_mce_bank_enabled_on_node() 2919 rdmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs); in toggle_ecc_err_reporting() 2923 struct msr *reg = per_cpu_ptr(msrs, cpu); in toggle_ecc_err_reporting() 2938 wrmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs); in toggle_ecc_err_reporting() 3458 msrs = msrs_alloc(); in amd64_edac_init() 3459 if (!msrs) in amd64_edac_init() 3498 msrs_free(msrs); in amd64_edac_init() 3499 msrs = NULL; in amd64_edac_init() [all …]
|
/Linux-v4.19/arch/x86/kernel/cpu/mcheck/ |
D | mce.c | 1620 u32 msrs[] = { in __mcheck_cpu_apply_quirks() local 1634 for (i = 0; i < ARRAY_SIZE(msrs); i++) in __mcheck_cpu_apply_quirks() 1635 msr_clear_bit(msrs[i], 62); in __mcheck_cpu_apply_quirks()
|