Lines Matching refs:root_role

2251 	iterator->level = vcpu->arch.mmu->root_role.level;  in shadow_walk_init_using_root()
2255 !vcpu->arch.mmu->root_role.direct) in shadow_walk_init_using_root()
2653 if (vcpu->arch.mmu->root_role.direct) in kvm_mmu_unprotect_page_virt()
3501 WARN_ON_ONCE(mmu->root_role.guest_mode); in kvm_mmu_free_guest_mode_roots()
3533 union kvm_mmu_page_role role = vcpu->arch.mmu->root_role; in mmu_alloc_root()
3551 u8 shadow_root_level = mmu->root_role.level; in mmu_alloc_direct_roots()
3700 mmu->root_role.level); in mmu_alloc_shadow_roots()
3716 if (mmu->root_role.level >= PT64_ROOT_4LEVEL) { in mmu_alloc_shadow_roots()
3725 if (mmu->root_role.level == PT64_ROOT_5LEVEL) { in mmu_alloc_shadow_roots()
3757 if (mmu->root_role.level == PT64_ROOT_5LEVEL) in mmu_alloc_shadow_roots()
3759 else if (mmu->root_role.level == PT64_ROOT_4LEVEL) in mmu_alloc_shadow_roots()
3775 bool need_pml5 = mmu->root_role.level > PT64_ROOT_4LEVEL; in mmu_alloc_special_roots()
3786 if (mmu->root_role.direct || in mmu_alloc_special_roots()
3788 mmu->root_role.level < PT64_ROOT_4LEVEL) in mmu_alloc_special_roots()
3884 if (vcpu->arch.mmu->root_role.direct) in kvm_mmu_sync_roots()
4114 arch.direct_map = vcpu->arch.mmu->root_role.direct; in kvm_arch_setup_async_pf()
4125 if ((vcpu->arch.mmu->root_role.direct != work->arch.direct_map) || in kvm_arch_async_page_ready()
4133 if (!vcpu->arch.mmu->root_role.direct && in kvm_arch_async_page_ready()
4453 union kvm_mmu_page_role new_role = mmu->root_role; in kvm_mmu_new_pgd()
4707 WARN_ON_ONCE(context->root_role.level < PT32E_ROOT_LEVEL); in reset_shadow_zero_bits_mask()
4711 context->root_role.level, in reset_shadow_zero_bits_mask()
4712 context->root_role.efer_nx, in reset_shadow_zero_bits_mask()
4718 for (i = context->root_role.level; --i >= 0;) { in reset_shadow_zero_bits_mask()
4753 context->root_role.level, true, in reset_tdp_shadow_zero_bits_mask()
4764 for (i = context->root_role.level; --i >= 0;) { in reset_tdp_shadow_zero_bits_mask()
5035 union kvm_mmu_page_role root_role = kvm_calc_tdp_mmu_root_page_role(vcpu, cpu_role); in init_kvm_tdp_mmu() local
5038 root_role.word == context->root_role.word) in init_kvm_tdp_mmu()
5042 context->root_role.word = root_role.word; in init_kvm_tdp_mmu()
5063 union kvm_mmu_page_role root_role) in shadow_mmu_init_context() argument
5066 root_role.word == context->root_role.word) in shadow_mmu_init_context()
5070 context->root_role.word = root_role.word; in shadow_mmu_init_context()
5087 union kvm_mmu_page_role root_role; in kvm_init_shadow_mmu() local
5089 root_role = cpu_role.base; in kvm_init_shadow_mmu()
5092 root_role.level = max_t(u32, root_role.level, PT32E_ROOT_LEVEL); in kvm_init_shadow_mmu()
5103 root_role.efer_nx = true; in kvm_init_shadow_mmu()
5105 shadow_mmu_init_context(vcpu, context, cpu_role, root_role); in kvm_init_shadow_mmu()
5118 union kvm_mmu_page_role root_role; in kvm_init_shadow_npt_mmu() local
5123 root_role = cpu_role.base; in kvm_init_shadow_npt_mmu()
5124 root_role.level = kvm_mmu_get_tdp_level(vcpu); in kvm_init_shadow_npt_mmu()
5125 if (root_role.level == PT64_ROOT_5LEVEL && in kvm_init_shadow_npt_mmu()
5127 root_role.passthrough = 1; in kvm_init_shadow_npt_mmu()
5129 shadow_mmu_init_context(vcpu, context, cpu_role, root_role); in kvm_init_shadow_npt_mmu()
5172 context->root_role.word = new_mode.base.word; in kvm_init_shadow_ept_mmu()
5268 vcpu->arch.root_mmu.root_role.word = 0; in kvm_mmu_after_set_cpuid()
5269 vcpu->arch.guest_mmu.root_role.word = 0; in kvm_mmu_after_set_cpuid()
5270 vcpu->arch.nested_mmu.root_role.word = 0; in kvm_mmu_after_set_cpuid()
5294 r = mmu_topup_memory_caches(vcpu, !vcpu->arch.mmu->root_role.direct); in kvm_mmu_load()
5300 if (vcpu->arch.mmu->root_role.direct) in kvm_mmu_load()
5537 bool direct = vcpu->arch.mmu->root_role.direct; in kvm_mmu_page_fault()
5568 if (vcpu->arch.mmu->root_role.direct && in kvm_mmu_page_fault()