Lines Matching refs:kvm_vcpu

97 bool intel_pmu_lbr_is_compatible(struct kvm_vcpu *vcpu);
98 bool intel_pmu_lbr_is_enabled(struct kvm_vcpu *vcpu);
100 int intel_pmu_create_guest_lbr_event(struct kvm_vcpu *vcpu);
101 void vmx_passthrough_lbr_msrs(struct kvm_vcpu *vcpu);
228 struct kvm_vcpu vcpu;
348 bool nested_vmx_allowed(struct kvm_vcpu *vcpu);
349 void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu,
354 void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu);
357 int vmx_get_cpl(struct kvm_vcpu *vcpu);
358 bool vmx_emulation_required(struct kvm_vcpu *vcpu);
359 unsigned long vmx_get_rflags(struct kvm_vcpu *vcpu);
360 void vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
361 u32 vmx_get_interrupt_shadow(struct kvm_vcpu *vcpu);
362 void vmx_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask);
363 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer);
364 void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
365 void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
367 void ept_save_pdptrs(struct kvm_vcpu *vcpu);
368 void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
369 void __vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
370 u64 construct_eptp(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level);
372 bool vmx_guest_inject_ac(struct kvm_vcpu *vcpu);
373 void vmx_update_exception_bitmap(struct kvm_vcpu *vcpu);
374 bool vmx_nmi_blocked(struct kvm_vcpu *vcpu);
375 bool vmx_interrupt_blocked(struct kvm_vcpu *vcpu);
376 bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu);
377 void vmx_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked);
378 void vmx_set_virtual_apic_mode(struct kvm_vcpu *vcpu);
380 void pt_update_intercept_for_msr(struct kvm_vcpu *vcpu);
384 void vmx_ept_load_pdptrs(struct kvm_vcpu *vcpu);
386 void vmx_disable_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, int type);
387 void vmx_enable_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, int type);
389 u64 vmx_get_l2_tsc_offset(struct kvm_vcpu *vcpu);
390 u64 vmx_get_l2_tsc_multiplier(struct kvm_vcpu *vcpu);
392 static inline void vmx_set_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, in vmx_set_intercept_for_msr()
401 void vmx_update_cpu_dirty_logging(struct kvm_vcpu *vcpu);
438 static inline void vmx_register_cache_reset(struct kvm_vcpu *vcpu) in BUILD_CONTROLS_SHADOW()
457 static inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu) in to_vmx()
462 static inline unsigned long vmx_get_exit_qual(struct kvm_vcpu *vcpu) in vmx_get_exit_qual()
473 static inline u32 vmx_get_intr_info(struct kvm_vcpu *vcpu) in vmx_get_intr_info()
502 static inline bool vmx_need_pf_intercept(struct kvm_vcpu *vcpu) in vmx_need_pf_intercept()
510 static inline bool is_unrestricted_guest(struct kvm_vcpu *vcpu) in is_unrestricted_guest()
517 bool __vmx_guest_state_valid(struct kvm_vcpu *vcpu);
518 static inline bool vmx_guest_state_valid(struct kvm_vcpu *vcpu) in vmx_guest_state_valid()
523 void dump_vmcs(struct kvm_vcpu *vcpu);