Lines Matching refs:mmu_lock
2341 if (need_resched() || spin_needbreak(&vcpu->kvm->mmu_lock)) { in mmu_sync_children()
2343 cond_resched_lock(&vcpu->kvm->mmu_lock); in mmu_sync_children()
2711 spin_lock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()
2725 spin_unlock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()
2736 spin_lock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()
2744 spin_unlock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()
3448 spin_lock(&vcpu->kvm->mmu_lock); in nonpaging_map()
3456 spin_unlock(&vcpu->kvm->mmu_lock); in nonpaging_map()
3461 spin_unlock(&vcpu->kvm->mmu_lock); in nonpaging_map()
3503 spin_lock(&vcpu->kvm->mmu_lock); in kvm_mmu_free_roots()
3526 spin_unlock(&vcpu->kvm->mmu_lock); in kvm_mmu_free_roots()
3548 spin_lock(&vcpu->kvm->mmu_lock); in mmu_alloc_direct_roots()
3550 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_direct_roots()
3556 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_direct_roots()
3563 spin_lock(&vcpu->kvm->mmu_lock); in mmu_alloc_direct_roots()
3565 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_direct_roots()
3572 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_direct_roots()
3603 spin_lock(&vcpu->kvm->mmu_lock); in mmu_alloc_shadow_roots()
3605 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_shadow_roots()
3612 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_shadow_roots()
3640 spin_lock(&vcpu->kvm->mmu_lock); in mmu_alloc_shadow_roots()
3642 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_shadow_roots()
3649 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_shadow_roots()
3723 spin_lock(&vcpu->kvm->mmu_lock); in kvm_mmu_sync_roots()
3729 spin_unlock(&vcpu->kvm->mmu_lock); in kvm_mmu_sync_roots()
3733 spin_lock(&vcpu->kvm->mmu_lock); in kvm_mmu_sync_roots()
3747 spin_unlock(&vcpu->kvm->mmu_lock); in kvm_mmu_sync_roots()
4085 spin_lock(&vcpu->kvm->mmu_lock); in tdp_page_fault()
4093 spin_unlock(&vcpu->kvm->mmu_lock); in tdp_page_fault()
4098 spin_unlock(&vcpu->kvm->mmu_lock); in tdp_page_fault()
5158 spin_lock(&vcpu->kvm->mmu_lock); in kvm_mmu_pte_write()
5189 spin_unlock(&vcpu->kvm->mmu_lock); in kvm_mmu_pte_write()
5484 if (need_resched() || spin_needbreak(&kvm->mmu_lock)) { in slot_handle_level_range()
5489 cond_resched_lock(&kvm->mmu_lock); in slot_handle_level_range()
5542 spin_lock(&kvm->mmu_lock); in kvm_zap_gfn_range()
5559 spin_unlock(&kvm->mmu_lock); in kvm_zap_gfn_range()
5573 spin_lock(&kvm->mmu_lock); in kvm_mmu_slot_remove_write_access()
5576 spin_unlock(&kvm->mmu_lock); in kvm_mmu_slot_remove_write_access()
5637 spin_lock(&kvm->mmu_lock); in kvm_mmu_zap_collapsible_sptes()
5640 spin_unlock(&kvm->mmu_lock); in kvm_mmu_zap_collapsible_sptes()
5648 spin_lock(&kvm->mmu_lock); in kvm_mmu_slot_leaf_clear_dirty()
5650 spin_unlock(&kvm->mmu_lock); in kvm_mmu_slot_leaf_clear_dirty()
5670 spin_lock(&kvm->mmu_lock); in kvm_mmu_slot_largepage_remove_write_access()
5673 spin_unlock(&kvm->mmu_lock); in kvm_mmu_slot_largepage_remove_write_access()
5688 spin_lock(&kvm->mmu_lock); in kvm_mmu_slot_set_dirty()
5690 spin_unlock(&kvm->mmu_lock); in kvm_mmu_slot_set_dirty()
5731 cond_resched_lock(&kvm->mmu_lock)) { in kvm_zap_obsolete_pages()
5762 spin_lock(&kvm->mmu_lock); in kvm_mmu_invalidate_zap_all_pages()
5778 spin_unlock(&kvm->mmu_lock); in kvm_mmu_invalidate_zap_all_pages()
5830 spin_lock(&kvm->mmu_lock); in mmu_shrink_scan()
5843 spin_unlock(&kvm->mmu_lock); in mmu_shrink_scan()