Lines Matching refs:vcpu_vmx
953 struct vcpu_vmx { struct
1062 static inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu) in to_vmx() argument
1064 return container_of(vcpu, struct vcpu_vmx, vcpu); in to_vmx()
1280 static void copy_shadow_to_vmcs12(struct vcpu_vmx *vmx);
2070 static int __find_msr_index(struct vcpu_vmx *vmx, u32 msr) in __find_msr_index()
2108 static struct shared_msr_entry *find_msr_entry(struct vcpu_vmx *vmx, u32 msr) in find_msr_entry()
2444 static inline void vm_entry_controls_reset_shadow(struct vcpu_vmx *vmx) in vm_entry_controls_reset_shadow()
2449 static inline void vm_entry_controls_init(struct vcpu_vmx *vmx, u32 val) in vm_entry_controls_init()
2455 static inline void vm_entry_controls_set(struct vcpu_vmx *vmx, u32 val) in vm_entry_controls_set()
2461 static inline u32 vm_entry_controls_get(struct vcpu_vmx *vmx) in vm_entry_controls_get()
2467 static inline void vm_entry_controls_setbit(struct vcpu_vmx *vmx, u32 val) in vm_entry_controls_setbit()
2472 static inline void vm_entry_controls_clearbit(struct vcpu_vmx *vmx, u32 val) in vm_entry_controls_clearbit()
2477 static inline void vm_exit_controls_reset_shadow(struct vcpu_vmx *vmx) in vm_exit_controls_reset_shadow()
2482 static inline void vm_exit_controls_init(struct vcpu_vmx *vmx, u32 val) in vm_exit_controls_init()
2488 static inline void vm_exit_controls_set(struct vcpu_vmx *vmx, u32 val) in vm_exit_controls_set()
2494 static inline u32 vm_exit_controls_get(struct vcpu_vmx *vmx) in vm_exit_controls_get()
2500 static inline void vm_exit_controls_setbit(struct vcpu_vmx *vmx, u32 val) in vm_exit_controls_setbit()
2505 static inline void vm_exit_controls_clearbit(struct vcpu_vmx *vmx, u32 val) in vm_exit_controls_clearbit()
2510 static void vmx_segment_cache_clear(struct vcpu_vmx *vmx) in vmx_segment_cache_clear()
2515 static bool vmx_segment_cache_test_set(struct vcpu_vmx *vmx, unsigned seg, in vmx_segment_cache_test_set()
2530 static u16 vmx_read_guest_seg_selector(struct vcpu_vmx *vmx, unsigned seg) in vmx_read_guest_seg_selector()
2539 static ulong vmx_read_guest_seg_base(struct vcpu_vmx *vmx, unsigned seg) in vmx_read_guest_seg_base()
2548 static u32 vmx_read_guest_seg_limit(struct vcpu_vmx *vmx, unsigned seg) in vmx_read_guest_seg_limit()
2557 static u32 vmx_read_guest_seg_ar(struct vcpu_vmx *vmx, unsigned seg) in vmx_read_guest_seg_ar()
2646 static void clear_atomic_switch_msr_special(struct vcpu_vmx *vmx, in clear_atomic_switch_msr_special()
2664 static void clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr) in clear_atomic_switch_msr()
2704 static void add_atomic_switch_msr_special(struct vcpu_vmx *vmx, in add_atomic_switch_msr_special()
2715 static void add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr, in add_atomic_switch_msr()
2780 static bool update_transition_efer(struct vcpu_vmx *vmx, int efer_offset) in update_transition_efer()
2865 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_prepare_switch_to_guest()
2943 static void vmx_prepare_switch_to_host(struct vcpu_vmx *vmx) in vmx_prepare_switch_to_host()
2983 static u64 vmx_read_guest_kernel_gs_base(struct vcpu_vmx *vmx) in vmx_read_guest_kernel_gs_base()
2992 static void vmx_write_guest_kernel_gs_base(struct vcpu_vmx *vmx, u64 data) in vmx_write_guest_kernel_gs_base()
3048 static void decache_tsc_multiplier(struct vcpu_vmx *vmx) in decache_tsc_multiplier()
3060 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_load()
3323 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_queue_exception()
3370 static void move_msr_up(struct vcpu_vmx *vmx, int from, int to) in move_msr_up()
3384 static void setup_msrs(struct vcpu_vmx *vmx) in setup_msrs()
3733 static int vmx_restore_vmx_basic(struct vcpu_vmx *vmx, u64 data) in vmx_restore_vmx_basic()
3764 vmx_restore_control_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data) in vmx_restore_control_msr()
3809 static int vmx_restore_vmx_misc(struct vcpu_vmx *vmx, u64 data) in vmx_restore_vmx_misc()
3854 static int vmx_restore_vmx_ept_vpid_cap(struct vcpu_vmx *vmx, u64 data) in vmx_restore_vmx_ept_vpid_cap()
3870 static int vmx_restore_fixed0_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data) in vmx_restore_fixed0_msr()
3903 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_vmx_msr()
4067 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_msr()
4159 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_msr()
4945 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_pmode()
5018 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_rmode()
5066 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_efer()
5264 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_cr0()
5429 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_segment()
5476 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_cpl()
5509 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_segment()
6096 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_update_msr_bitmap()
6139 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_complete_nested_posted_interrupt()
6176 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_guest_apic_has_interrupt()
6237 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_deliver_nested_posted_interrupt()
6263 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_deliver_posted_interrupt()
6287 static void vmx_set_constant_host_state(struct vcpu_vmx *vmx) in vmx_set_constant_host_state()
6344 static void set_cr4_guest_host_mask(struct vcpu_vmx *vmx) in set_cr4_guest_host_mask()
6355 static u32 vmx_pin_based_exec_ctrl(struct vcpu_vmx *vmx) in vmx_pin_based_exec_ctrl()
6372 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_refresh_apicv_exec_ctrl()
6390 static u32 vmx_exec_control(struct vcpu_vmx *vmx) in vmx_exec_control()
6428 static void vmx_compute_secondary_exec_control(struct vcpu_vmx *vmx) in vmx_compute_secondary_exec_control()
6567 static void vmx_vcpu_setup(struct vcpu_vmx *vmx) in vmx_vcpu_setup()
6680 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_reset()
6819 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_inject_irq()
6848 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_inject_nmi()
6880 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_nmi_mask()
6894 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_nmi_mask()
7041 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_exception()
7584 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_task_switch()
7736 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invalid_guest_state()
7799 struct vcpu_vmx *vmx = to_vmx(vcpu); in grow_ple_window()
7814 struct vcpu_vmx *vmx = to_vmx(vcpu); in shrink_ple_window()
8119 struct vcpu_vmx *vmx = in vmx_preemption_timer_fn()
8120 container_of(timer, struct vcpu_vmx, nested.preemption_timer); in vmx_preemption_timer_fn()
8250 struct vcpu_vmx *vmx = to_vmx(vcpu); in alloc_shadow_vmcs()
8271 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_vmx_operation()
8324 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmon()
8420 static void vmx_disable_shadow_vmcs(struct vcpu_vmx *vmx) in vmx_disable_shadow_vmcs()
8426 static inline void nested_release_vmcs12(struct vcpu_vmx *vmx) in nested_release_vmcs12()
8452 static void free_nested(struct vcpu_vmx *vmx) in free_nested()
8502 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmclear()
8619 static void copy_shadow_to_vmcs12(struct vcpu_vmx *vmx) in copy_shadow_to_vmcs12()
8657 static void copy_vmcs12_to_shadow(struct vcpu_vmx *vmx) in copy_vmcs12_to_shadow()
8692 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_check_vmcs12()
8762 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmwrite()
8853 static void set_current_vmptr(struct vcpu_vmx *vmx, gpa_t vmptr) in set_current_vmptr()
8869 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmptrld()
8950 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invept()
9012 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invvpid()
9230 struct vcpu_vmx *vmx = to_vmx(vcpu); in valid_ept_address()
9311 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmfunc()
9617 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_exit_reflected()
9822 static void vmx_destroy_pml_buffer(struct vcpu_vmx *vmx) in vmx_destroy_pml_buffer()
9832 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_flush_pml_buffer()
10043 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_handle_exit()
10325 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_sync_pir_to_irr()
10382 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_apicv_post_state_restore()
10388 static void vmx_complete_atomic_exit(struct vcpu_vmx *vmx) in vmx_complete_atomic_exit()
10427 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_handle_external_intr()
10488 static void vmx_recover_nmi_blocking(struct vcpu_vmx *vmx) in vmx_recover_nmi_blocking()
10585 static void vmx_complete_interrupts(struct vcpu_vmx *vmx) in vmx_complete_interrupts()
10602 static void atomic_switch_perf_msrs(struct vcpu_vmx *vmx) in atomic_switch_perf_msrs()
10620 static void vmx_arm_hv_timer(struct vcpu_vmx *vmx, u32 val) in vmx_arm_hv_timer()
10631 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_update_hv_timer()
10661 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_run()
10826 [launched]"i"(offsetof(struct vcpu_vmx, __launched)), in vmx_vcpu_run()
10827 [fail]"i"(offsetof(struct vcpu_vmx, fail)), in vmx_vcpu_run()
10828 [host_rsp]"i"(offsetof(struct vcpu_vmx, host_rsp)), in vmx_vcpu_run()
10829 [rax]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RAX])), in vmx_vcpu_run()
10830 [rbx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RBX])), in vmx_vcpu_run()
10831 [rcx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RCX])), in vmx_vcpu_run()
10832 [rdx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RDX])), in vmx_vcpu_run()
10833 [rsi]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RSI])), in vmx_vcpu_run()
10834 [rdi]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RDI])), in vmx_vcpu_run()
10835 [rbp]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RBP])), in vmx_vcpu_run()
10837 [r8]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R8])), in vmx_vcpu_run()
10838 [r9]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R9])), in vmx_vcpu_run()
10839 [r10]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R10])), in vmx_vcpu_run()
10840 [r11]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R11])), in vmx_vcpu_run()
10841 [r12]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R12])), in vmx_vcpu_run()
10842 [r13]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R13])), in vmx_vcpu_run()
10843 [r14]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R14])), in vmx_vcpu_run()
10844 [r15]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R15])), in vmx_vcpu_run()
10846 [cr2]"i"(offsetof(struct vcpu_vmx, vcpu.arch.cr2)), in vmx_vcpu_run()
10950 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_switch_vmcs()
10969 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_free_vcpu_nested()
10979 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_free_vcpu()
10995 struct vcpu_vmx *vmx = kmem_cache_zalloc(kvm_vcpu_cache, GFP_KERNEL); in vmx_create_vcpu()
11221 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_cr_fixed1_bits_update()
11260 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_entry_exit_ctls_update()
11277 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_cpuid_update()
11307 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_ept_inject_page_fault()
11398 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_vmcs12_pages()
11488 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_start_preemption_timer()
11662 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_flush_cached_shadow_vmcs12()
11965 struct vcpu_vmx *vmx = to_vmx(vcpu); in prepare_vmcs02_full()
12105 struct vcpu_vmx *vmx = to_vmx(vcpu); in prepare_vmcs02()
12385 struct vcpu_vmx *vmx = to_vmx(vcpu); in check_vmentry_prereqs()
12609 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_vmx_non_root_mode()
12699 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_run()
12895 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_check_nested_events()
13287 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_vmexit()
13507 struct vcpu_vmx *vmx; in vmx_set_hv_timer()
13578 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_write_pml_buffer()
13863 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_pre_enter_smm()
13877 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_pre_leave_smm()
13906 struct vcpu_vmx *vmx; in vmx_get_nested_state()
13988 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_nested_state()
14325 r = kvm_init(&vmx_x86_ops, sizeof(struct vcpu_vmx), in vmx_init()
14326 __alignof__(struct vcpu_vmx), THIS_MODULE); in vmx_init()