Lines Matching refs:kvm_x86_get_cpl
835 if (static_call(kvm_x86_get_cpl)(vcpu) <= required_cpl)
1093 if (static_call(kvm_x86_get_cpl)(vcpu) != 0 ||
4796 vcpu->arch.preempted_in_kernel = !static_call(kvm_x86_get_cpl)(vcpu);
7160 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0;
7170 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0;
7180 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0;
7233 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0;
7258 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0;
7281 else if (static_call(kvm_x86_get_cpl)(vcpu) == 3)
7335 else if (static_call(kvm_x86_get_cpl)(vcpu) == 3)
7405 u64 access = ((static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0)
7940 return static_call(kvm_x86_get_cpl)(emul_to_vcpu(ctxt));
8456 if (!is_guest_mode(vcpu) && static_call(kvm_x86_get_cpl)(vcpu) == 0) {
9666 if (static_call(kvm_x86_get_cpl)(vcpu) != 0) {
13133 static_call(kvm_x86_get_cpl)(vcpu) == 0)