Lines Matching refs:kvm_x86_ops
136 struct kvm_x86_ops kvm_x86_ops __read_mostly;
140 *(((struct kvm_x86_ops *)0)->func));
643 kvm_x86_ops.nested_ops->leave_nested(vcpu); in kvm_leave_nested()
663 kvm_x86_ops.nested_ops->is_exception_vmexit(vcpu, nr, error_code)) { in kvm_multiple_exception()
4597 r = kvm_x86_ops.nested_ops->get_state ? in kvm_vm_ioctl_check_extension()
4598 kvm_x86_ops.nested_ops->get_state(NULL, NULL, 0) : 0; in kvm_vm_ioctl_check_extension()
4601 r = kvm_x86_ops.enable_l2_tlb_flush != NULL; in kvm_vm_ioctl_check_extension()
4604 r = kvm_x86_ops.nested_ops->enable_evmcs != NULL; in kvm_vm_ioctl_check_extension()
5613 if (!kvm_x86_ops.nested_ops->enable_evmcs) in kvm_vcpu_ioctl_enable_cap()
5615 r = kvm_x86_ops.nested_ops->enable_evmcs(vcpu, &vmcs_version); in kvm_vcpu_ioctl_enable_cap()
5624 if (!kvm_x86_ops.enable_l2_tlb_flush) in kvm_vcpu_ioctl_enable_cap()
5966 if (!kvm_x86_ops.nested_ops->get_state) in kvm_arch_vcpu_ioctl()
5974 r = kvm_x86_ops.nested_ops->get_state(vcpu, user_kvm_nested_state, in kvm_arch_vcpu_ioctl()
5996 if (!kvm_x86_ops.nested_ops->set_state) in kvm_arch_vcpu_ioctl()
6019 r = kvm_x86_ops.nested_ops->set_state(vcpu, user_kvm_nested_state, &kvm_state); in kvm_arch_vcpu_ioctl()
6410 if (!kvm_x86_ops.vm_copy_enc_context_from) in kvm_vm_ioctl_enable_cap()
6417 if (!kvm_x86_ops.vm_move_enc_context_from) in kvm_vm_ioctl_enable_cap()
7046 if (!kvm_x86_ops.mem_enc_ioctl) in kvm_arch_vm_ioctl()
7060 if (!kvm_x86_ops.mem_enc_register_region) in kvm_arch_vm_ioctl()
7074 if (!kvm_x86_ops.mem_enc_unregister_region) in kvm_arch_vm_ioctl()
9408 memcpy(&kvm_x86_ops, ops->runtime_ops, sizeof(kvm_x86_ops)); in kvm_ops_update()
9411 static_call_update(kvm_x86_##func, kvm_x86_ops.func); in kvm_ops_update()
9413 WARN_ON(!kvm_x86_ops.func); __KVM_X86_OP(func) in kvm_ops_update()
9416 static_call_update(kvm_x86_##func, (void *)kvm_x86_ops.func ? : \ in kvm_ops_update()
9454 if (kvm_x86_ops.hardware_enable) { in __kvm_x86_vendor_init()
9455 pr_err("already loaded vendor module '%s'\n", kvm_x86_ops.name); in __kvm_x86_vendor_init()
9573 kvm_x86_ops.hardware_enable = NULL; in __kvm_x86_vendor_init()
9625 kvm_x86_ops.hardware_enable = NULL; in kvm_x86_vendor_exit()
9972 if (!kvm_x86_ops.update_cr8_intercept) in update_cr8_intercept()
9998 kvm_x86_ops.nested_ops->triple_fault(vcpu); in kvm_check_nested_events()
10002 return kvm_x86_ops.nested_ops->check_events(vcpu); in kvm_check_nested_events()
10240 kvm_x86_ops.nested_ops->has_events && in kvm_check_and_inject_events()
10241 kvm_x86_ops.nested_ops->has_events(vcpu)) in kvm_check_and_inject_events()
10377 kvm_x86_ops.allow_apicv_in_x2apic_without_x2apic_virtualization) in kvm_vcpu_update_apicv()
10390 if (!(kvm_x86_ops.required_apicv_inhibits & BIT(reason))) in __kvm_set_or_clear_apicv_inhibit()
10523 if (unlikely(!kvm_x86_ops.nested_ops->get_nested_state_pages(vcpu))) { in vcpu_enter_guest()
10573 kvm_x86_ops.nested_ops->triple_fault(vcpu); in vcpu_enter_guest()
11172 kvm_x86_ops.nested_ops->is_exception_vmexit(vcpu, ex->vector, in kvm_arch_vcpu_ioctl_run()
12671 if (!kvm_x86_ops.cpu_dirty_log_size) in kvm_mmu_update_cpu_dirty_logging()
12747 if (kvm_x86_ops.cpu_dirty_log_size) { in kvm_mmu_slot_apply_flags()
12869 kvm_x86_ops.nested_ops->has_events && in kvm_vcpu_has_events()
12870 kvm_x86_ops.nested_ops->has_events(vcpu)) in kvm_vcpu_has_events()