Lines Matching refs:static_call
842 if (static_call(kvm_x86_get_cpl)(vcpu) <= required_cpl) in kvm_require_cpl()
926 return static_call(kvm_x86_is_valid_cr0)(vcpu, cr0); in kvm_is_valid_cr0()
990 static_call(kvm_x86_get_cs_db_l_bits)(vcpu, &cs_db, &cs_l); in kvm_set_cr0()
1004 static_call(kvm_x86_set_cr0)(vcpu, cr0); in kvm_set_cr0()
1122 if (static_call(kvm_x86_get_cpl)(vcpu) != 0 || in kvm_emulate_xsetbv()
1147 static_call(kvm_x86_is_valid_cr4)(vcpu, cr4); in kvm_is_valid_cr4()
1215 static_call(kvm_x86_set_cr4)(vcpu, cr4); in kvm_set_cr4()
1354 static_call(kvm_x86_set_dr7)(vcpu, dr7); in kvm_update_dr7()
1694 return static_call(kvm_x86_get_msr_feature)(msr); in kvm_get_msr_feature()
1773 r = static_call(kvm_x86_set_efer)(vcpu, efer); in set_efer()
1899 return static_call(kvm_x86_set_msr)(vcpu, &msr); in __kvm_set_msr()
1941 ret = static_call(kvm_x86_get_msr)(vcpu, &msr); in __kvm_get_msr()
2009 return static_call(kvm_x86_complete_emulated_msr)(vcpu, vcpu->run->msr.error); in complete_fast_msr_access()
2073 return static_call(kvm_x86_complete_emulated_msr)(vcpu, r); in kvm_emulate_rdmsr()
2098 return static_call(kvm_x86_complete_emulated_msr)(vcpu, r); in kvm_emulate_wrmsr()
2628 static_call(kvm_x86_get_l2_tsc_offset)(vcpu), in kvm_vcpu_write_tsc_offset()
2629 static_call(kvm_x86_get_l2_tsc_multiplier)(vcpu)); in kvm_vcpu_write_tsc_offset()
2633 static_call(kvm_x86_write_tsc_offset)(vcpu); in kvm_vcpu_write_tsc_offset()
2644 static_call(kvm_x86_get_l2_tsc_multiplier)(vcpu)); in kvm_vcpu_write_tsc_multiplier()
2649 static_call(kvm_x86_write_tsc_multiplier)(vcpu); in kvm_vcpu_write_tsc_multiplier()
3466 static_call(kvm_x86_flush_tlb_all)(vcpu); in kvm_vcpu_flush_tlb_all()
3487 static_call(kvm_x86_flush_tlb_guest)(vcpu); in kvm_vcpu_flush_tlb_guest()
3500 static_call(kvm_x86_flush_tlb_current)(vcpu); in kvm_vcpu_flush_tlb_current()
4569 r = static_call(kvm_x86_has_emulated_msr)(kvm, MSR_IA32_SMBASE); in kvm_vm_ioctl_check_extension()
4808 if (static_call(kvm_x86_has_wbinvd_exit)()) in kvm_arch_vcpu_load()
4815 static_call(kvm_x86_vcpu_load)(vcpu, cpu); in kvm_arch_vcpu_load()
4910 vcpu->arch.preempted_in_kernel = !static_call(kvm_x86_get_cpl)(vcpu); in kvm_arch_vcpu_put()
4924 static_call(kvm_x86_vcpu_put)(vcpu); in kvm_arch_vcpu_put()
5048 static_call(kvm_x86_setup_mce)(vcpu); in kvm_vcpu_ioctl_x86_setup_mce()
5208 events->interrupt.shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); in kvm_vcpu_ioctl_x86_get_vcpu_events()
5212 events->nmi.masked = static_call(kvm_x86_get_nmi_mask)(vcpu); in kvm_vcpu_ioctl_x86_get_vcpu_events()
5294 static_call(kvm_x86_set_interrupt_shadow)(vcpu, in kvm_vcpu_ioctl_x86_set_vcpu_events()
5303 static_call(kvm_x86_set_nmi_mask)(vcpu, events->nmi.masked); in kvm_vcpu_ioctl_x86_set_vcpu_events()
5627 return static_call(kvm_x86_enable_l2_tlb_flush)(vcpu); in kvm_vcpu_ioctl_enable_cap()
6096 ret = static_call(kvm_x86_set_tss_addr)(kvm, addr); in kvm_vm_ioctl_set_tss_addr()
6103 return static_call(kvm_x86_set_identity_map_addr)(kvm, ident_addr); in kvm_vm_ioctl_set_identity_map_addr()
6413 r = static_call(kvm_x86_vm_copy_enc_context_from)(kvm, cap->args[0]); in kvm_vm_ioctl_enable_cap()
6420 r = static_call(kvm_x86_vm_move_enc_context_from)(kvm, cap->args[0]); in kvm_vm_ioctl_enable_cap()
7049 r = static_call(kvm_x86_mem_enc_ioctl)(kvm, argp); in kvm_arch_vm_ioctl()
7063 r = static_call(kvm_x86_mem_enc_register_region)(kvm, ®ion); in kvm_arch_vm_ioctl()
7077 r = static_call(kvm_x86_mem_enc_unregister_region)(kvm, ®ion); in kvm_arch_vm_ioctl()
7226 if (!static_call(kvm_x86_has_emulated_msr)(NULL, emulated_msrs_all[i])) in kvm_init_msr_lists()
7285 static_call(kvm_x86_set_segment)(vcpu, var, seg); in kvm_set_segment()
7291 static_call(kvm_x86_get_segment)(vcpu, var, seg); in kvm_get_segment()
7314 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_mmu_gva_to_gpa_read()
7324 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_mmu_gva_to_gpa_write()
7377 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_fetch_guest_virt()
7402 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_read_guest_virt()
7425 else if (static_call(kvm_x86_get_cpl)(vcpu) == 3) in emulator_read_std()
7470 else if (static_call(kvm_x86_get_cpl)(vcpu) == 3) in emulator_write_std()
7491 return static_call(kvm_x86_can_emulate_instruction)(vcpu, emul_type, in kvm_can_emulate_insn()
7540 u64 access = ((static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0) in vcpu_mmio_gva_to_gpa()
7957 return static_call(kvm_x86_get_segment_base)(vcpu, seg); in get_segment_base()
7970 if (static_call(kvm_x86_has_wbinvd_exit)()) { in kvm_emulate_wbinvd_noskip()
8075 return static_call(kvm_x86_get_cpl)(emul_to_vcpu(ctxt)); in emulator_get_cpl()
8080 static_call(kvm_x86_get_gdt)(emul_to_vcpu(ctxt), dt); in emulator_get_gdt()
8085 static_call(kvm_x86_get_idt)(emul_to_vcpu(ctxt), dt); in emulator_get_idt()
8090 static_call(kvm_x86_set_gdt)(emul_to_vcpu(ctxt), dt); in emulator_set_gdt()
8095 static_call(kvm_x86_set_idt)(emul_to_vcpu(ctxt), dt); in emulator_set_idt()
8245 return static_call(kvm_x86_check_intercept)(emul_to_vcpu(ctxt), info, stage, in emulator_intercept()
8283 static_call(kvm_x86_set_nmi_mask)(emul_to_vcpu(ctxt), masked); in emulator_set_nmi_mask()
8370 u32 int_shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); in toggle_interruptibility()
8381 static_call(kvm_x86_set_interrupt_shadow)(vcpu, mask); in toggle_interruptibility()
8422 static_call(kvm_x86_get_cs_db_l_bits)(vcpu, &cs_db, &cs_l); in init_emulate_ctxt()
8478 static_call(kvm_x86_get_exit_info)(vcpu, (u32 *)&info[0], &info[1], in prepare_emulation_failure_exit()
8557 if (!is_guest_mode(vcpu) && static_call(kvm_x86_get_cpl)(vcpu) == 0) { in handle_emulation_failure()
8722 unsigned long rflags = static_call(kvm_x86_get_rflags)(vcpu); in kvm_skip_emulated_instruction()
8725 r = static_call(kvm_x86_skip_emulated_instruction)(vcpu); in kvm_skip_emulated_instruction()
8757 shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); in kvm_is_code_breakpoint_inhibited()
9025 unsigned long rflags = static_call(kvm_x86_get_rflags)(vcpu); in x86_emulate_instruction()
9441 return static_call(kvm_x86_check_processor_compatibility)(); in kvm_x86_check_processor_compatibility()
9574 static_call(kvm_x86_hardware_unsetup)(); in __kvm_x86_vendor_init()
9616 static_call(kvm_x86_hardware_unsetup)(); in kvm_x86_vendor_exit()
9742 ulong vcpu_reasons = static_call(kvm_x86_vcpu_get_apicv_inhibit_reasons)(vcpu); in kvm_vcpu_apicv_activated()
9845 if (static_call(kvm_x86_get_cpl)(vcpu) != 0) { in kvm_emulate_hypercall()
9939 static_call(kvm_x86_patch_hypercall)(vcpu, instruction); in emulator_fix_hypercall()
9956 kvm_run->if_flag = static_call(kvm_x86_get_if_flag)(vcpu); in post_kvm_run_save()
9991 static_call(kvm_x86_update_cr8_intercept)(vcpu, tpr, max_irr); in update_cr8_intercept()
10021 static_call(kvm_x86_inject_exception)(vcpu); in kvm_inject_exception()
10107 static_call(kvm_x86_inject_nmi)(vcpu); in kvm_check_and_inject_events()
10109 static_call(kvm_x86_inject_irq)(vcpu, true); in kvm_check_and_inject_events()
10194 r = can_inject ? static_call(kvm_x86_smi_allowed)(vcpu, true) : -EBUSY; in kvm_check_and_inject_events()
10203 static_call(kvm_x86_enable_smi_window)(vcpu); in kvm_check_and_inject_events()
10208 r = can_inject ? static_call(kvm_x86_nmi_allowed)(vcpu, true) : -EBUSY; in kvm_check_and_inject_events()
10214 static_call(kvm_x86_inject_nmi)(vcpu); in kvm_check_and_inject_events()
10216 WARN_ON(static_call(kvm_x86_nmi_allowed)(vcpu, true) < 0); in kvm_check_and_inject_events()
10219 static_call(kvm_x86_enable_nmi_window)(vcpu); in kvm_check_and_inject_events()
10223 r = can_inject ? static_call(kvm_x86_interrupt_allowed)(vcpu, true) : -EBUSY; in kvm_check_and_inject_events()
10231 static_call(kvm_x86_inject_irq)(vcpu, false); in kvm_check_and_inject_events()
10232 WARN_ON(static_call(kvm_x86_interrupt_allowed)(vcpu, true) < 0); in kvm_check_and_inject_events()
10236 static_call(kvm_x86_enable_irq_window)(vcpu); in kvm_check_and_inject_events()
10282 if (static_call(kvm_x86_get_nmi_mask)(vcpu) || vcpu->arch.nmi_injected) in process_nmi()
10291 if (static_call(kvm_x86_is_vnmi_pending)(vcpu)) in process_nmi()
10298 (static_call(kvm_x86_set_vnmi_pending)(vcpu))) in process_nmi()
10309 static_call(kvm_x86_is_vnmi_pending)(vcpu); in kvm_get_nr_pending_nmis()
10343 static_call(kvm_x86_refresh_apicv_exec_ctrl)(vcpu); in __kvm_vcpu_update_apicv()
10652 static_call(kvm_x86_msr_filter_changed)(vcpu); in vcpu_enter_guest()
10655 static_call(kvm_x86_update_cpu_dirty_logging)(vcpu); in vcpu_enter_guest()
10677 static_call(kvm_x86_enable_irq_window)(vcpu); in vcpu_enter_guest()
10692 static_call(kvm_x86_prepare_switch_to_guest)(vcpu); in vcpu_enter_guest()
10742 static_call(kvm_x86_request_immediate_exit)(vcpu); in vcpu_enter_guest()
10774 exit_fastpath = static_call(kvm_x86_vcpu_run)(vcpu); in vcpu_enter_guest()
10798 static_call(kvm_x86_sync_dirty_debug_regs)(vcpu); in vcpu_enter_guest()
10827 static_call(kvm_x86_handle_exit_irqoff)(vcpu); in vcpu_enter_guest()
10873 r = static_call(kvm_x86_handle_exit)(vcpu, exit_fastpath); in vcpu_enter_guest()
10879 static_call(kvm_x86_cancel_injection)(vcpu); in vcpu_enter_guest()
11198 r = static_call(kvm_x86_vcpu_pre_run)(vcpu); in kvm_arch_vcpu_ioctl_run()
11318 static_call(kvm_x86_get_idt)(vcpu, &dt); in __get_sregs_common()
11321 static_call(kvm_x86_get_gdt)(vcpu, &dt); in __get_sregs_common()
11517 static_call(kvm_x86_set_idt)(vcpu, &dt); in __set_sregs_common()
11520 static_call(kvm_x86_set_gdt)(vcpu, &dt); in __set_sregs_common()
11531 static_call(kvm_x86_set_efer)(vcpu, sregs->efer); in __set_sregs_common()
11534 static_call(kvm_x86_set_cr0)(vcpu, sregs->cr0); in __set_sregs_common()
11538 static_call(kvm_x86_set_cr4)(vcpu, sregs->cr4); in __set_sregs_common()
11708 static_call(kvm_x86_update_exception_bitmap)(vcpu); in kvm_arch_vcpu_ioctl_set_guest_debug()
11845 return static_call(kvm_x86_vcpu_precreate)(kvm); in kvm_arch_vcpu_precreate()
11934 r = static_call(kvm_x86_vcpu_create)(vcpu); in kvm_arch_vcpu_create()
11992 static_call(kvm_x86_vcpu_free)(vcpu); in kvm_arch_vcpu_destroy()
12113 static_call(kvm_x86_vcpu_reset)(vcpu, init_event); in kvm_vcpu_reset()
12132 static_call(kvm_x86_set_cr0)(vcpu, new_cr0); in kvm_vcpu_reset()
12133 static_call(kvm_x86_set_cr4)(vcpu, 0); in kvm_vcpu_reset()
12134 static_call(kvm_x86_set_efer)(vcpu, 0); in kvm_vcpu_reset()
12135 static_call(kvm_x86_update_exception_bitmap)(vcpu); in kvm_vcpu_reset()
12192 ret = static_call(kvm_x86_hardware_enable)(); in kvm_arch_hardware_enable()
12274 static_call(kvm_x86_hardware_disable)(); in kvm_arch_hardware_disable()
12300 static_call(kvm_x86_sched_in)(vcpu, cpu); in kvm_arch_sched_in()
12324 ret = static_call(kvm_x86_vm_init)(kvm); in kvm_arch_init_vm()
12827 static_call(kvm_x86_guest_apic_has_interrupt)(vcpu)); in kvm_guest_apic_has_interrupt()
12847 static_call(kvm_x86_nmi_allowed)(vcpu, false))) in kvm_vcpu_has_events()
12853 static_call(kvm_x86_smi_allowed)(vcpu, false))) in kvm_vcpu_has_events()
12887 static_call(kvm_x86_dy_apicv_has_pending_interrupt)(vcpu)) in kvm_arch_dy_has_pending_interrupt()
12928 return static_call(kvm_x86_interrupt_allowed)(vcpu, false); in kvm_arch_interrupt_allowed()
12954 rflags = static_call(kvm_x86_get_rflags)(vcpu); in kvm_get_rflags()
12966 static_call(kvm_x86_set_rflags)(vcpu, rflags); in __kvm_set_rflags()
13078 static_call(kvm_x86_get_cpl)(vcpu) == 0) in kvm_can_deliver_async_pf()
13237 ret = static_call(kvm_x86_pi_update_irte)(irqfd->kvm, in kvm_arch_irq_bypass_add_producer()
13262 ret = static_call(kvm_x86_pi_update_irte)(irqfd->kvm, prod->irq, irqfd->gsi, 0); in kvm_arch_irq_bypass_del_producer()
13273 return static_call(kvm_x86_pi_update_irte)(kvm, host_irq, guest_irq, set); in kvm_arch_update_irqfd_routing()