Lines Matching refs:kvm_ops

421 	struct kvmppc_ops *kvm_ops = NULL;  in kvm_arch_init_vm()  local
427 kvm_ops = kvmppc_hv_ops; in kvm_arch_init_vm()
429 kvm_ops = kvmppc_pr_ops; in kvm_arch_init_vm()
430 if (!kvm_ops) in kvm_arch_init_vm()
435 kvm_ops = kvmppc_hv_ops; in kvm_arch_init_vm()
439 kvm_ops = kvmppc_pr_ops; in kvm_arch_init_vm()
443 if (kvm_ops->owner && !try_module_get(kvm_ops->owner)) in kvm_arch_init_vm()
446 kvm->arch.kvm_ops = kvm_ops; in kvm_arch_init_vm()
491 module_put(kvm->arch.kvm_ops->owner); in kvm_arch_destroy_vm()
823 if (kvm->arch.kvm_ops->irq_bypass_add_producer) in kvm_arch_irq_bypass_add_producer()
824 return kvm->arch.kvm_ops->irq_bypass_add_producer(cons, prod); in kvm_arch_irq_bypass_add_producer()
836 if (kvm->arch.kvm_ops->irq_bypass_del_producer) in kvm_arch_irq_bypass_del_producer()
837 kvm->arch.kvm_ops->irq_bypass_del_producer(cons, prod); in kvm_arch_irq_bypass_del_producer()
1142 if (vcpu->kvm->arch.kvm_ops->giveup_ext) in kvmppc_complete_mmio_load()
1143 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_FP); in kvmppc_complete_mmio_load()
1158 if (vcpu->kvm->arch.kvm_ops->giveup_ext) in kvmppc_complete_mmio_load()
1159 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VSX); in kvmppc_complete_mmio_load()
1175 if (vcpu->kvm->arch.kvm_ops->giveup_ext) in kvmppc_complete_mmio_load()
1176 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VEC); in kvmppc_complete_mmio_load()
2113 if (kvm->arch.kvm_ops->set_smt_mode) in kvm_vm_ioctl_enable_cap()
2114 r = kvm->arch.kvm_ops->set_smt_mode(kvm, mode, flags); in kvm_vm_ioctl_enable_cap()
2310 r = kvm->arch.kvm_ops->get_smmu_info(kvm, &info); in kvm_arch_vm_ioctl()
2326 if (!kvm->arch.kvm_ops->configure_mmu) in kvm_arch_vm_ioctl()
2331 r = kvm->arch.kvm_ops->configure_mmu(kvm, &cfg); in kvm_arch_vm_ioctl()
2339 if (!kvm->arch.kvm_ops->get_rmmu_info) in kvm_arch_vm_ioctl()
2341 r = kvm->arch.kvm_ops->get_rmmu_info(kvm, &info); in kvm_arch_vm_ioctl()
2356 r = kvm->arch.kvm_ops->arch_vm_ioctl(filp, ioctl, arg); in kvm_arch_vm_ioctl()