Lines Matching refs:memslot
1826 struct kvm_memory_slot *memslot; in kvm_handle_hva_range() local
1833 kvm_for_each_memslot(memslot, slots) { in kvm_handle_hva_range()
1837 hva_start = max(start, memslot->userspace_addr); in kvm_handle_hva_range()
1838 hva_end = min(end, memslot->userspace_addr + in kvm_handle_hva_range()
1839 (memslot->npages << PAGE_SHIFT)); in kvm_handle_hva_range()
1846 gfn_start = hva_to_gfn_memslot(hva_start, memslot); in kvm_handle_hva_range()
1847 gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); in kvm_handle_hva_range()
1849 for_each_slot_rmap_range(memslot, PT_PAGE_TABLE_LEVEL, in kvm_handle_hva_range()
1853 ret |= handler(kvm, iterator.rmap, memslot, in kvm_handle_hva_range()
5472 slot_handle_level_range(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_level_range() argument
5479 for_each_slot_rmap_range(memslot, start_level, end_level, start_gfn, in slot_handle_level_range()
5502 slot_handle_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_level() argument
5506 return slot_handle_level_range(kvm, memslot, fn, start_level, in slot_handle_level()
5507 end_level, memslot->base_gfn, in slot_handle_level()
5508 memslot->base_gfn + memslot->npages - 1, in slot_handle_level()
5513 slot_handle_all_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_all_level() argument
5516 return slot_handle_level(kvm, memslot, fn, PT_PAGE_TABLE_LEVEL, in slot_handle_all_level()
5521 slot_handle_large_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_large_level() argument
5524 return slot_handle_level(kvm, memslot, fn, PT_PAGE_TABLE_LEVEL + 1, in slot_handle_large_level()
5529 slot_handle_leaf(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_leaf() argument
5532 return slot_handle_level(kvm, memslot, fn, PT_PAGE_TABLE_LEVEL, in slot_handle_leaf()
5539 struct kvm_memory_slot *memslot; in kvm_zap_gfn_range() local
5545 kvm_for_each_memslot(memslot, slots) { in kvm_zap_gfn_range()
5548 start = max(gfn_start, memslot->base_gfn); in kvm_zap_gfn_range()
5549 end = min(gfn_end, memslot->base_gfn + memslot->npages); in kvm_zap_gfn_range()
5553 slot_handle_level_range(kvm, memslot, kvm_zap_rmapp, in kvm_zap_gfn_range()
5569 struct kvm_memory_slot *memslot) in kvm_mmu_slot_remove_write_access() argument
5574 flush = slot_handle_all_level(kvm, memslot, slot_rmap_write_protect, in kvm_mmu_slot_remove_write_access()
5634 const struct kvm_memory_slot *memslot) in kvm_mmu_zap_collapsible_sptes() argument
5638 slot_handle_leaf(kvm, (struct kvm_memory_slot *)memslot, in kvm_mmu_zap_collapsible_sptes()
5644 struct kvm_memory_slot *memslot) in kvm_mmu_slot_leaf_clear_dirty() argument
5649 flush = slot_handle_leaf(kvm, memslot, __rmap_clear_dirty, false); in kvm_mmu_slot_leaf_clear_dirty()
5666 struct kvm_memory_slot *memslot) in kvm_mmu_slot_largepage_remove_write_access() argument
5671 flush = slot_handle_large_level(kvm, memslot, slot_rmap_write_protect, in kvm_mmu_slot_largepage_remove_write_access()
5684 struct kvm_memory_slot *memslot) in kvm_mmu_slot_set_dirty() argument
5689 flush = slot_handle_all_level(kvm, memslot, __rmap_set_dirty, false); in kvm_mmu_slot_set_dirty()
5917 struct kvm_memory_slot *memslot; in kvm_mmu_calculate_mmu_pages() local
5923 kvm_for_each_memslot(memslot, slots) in kvm_mmu_calculate_mmu_pages()
5924 nr_pages += memslot->npages; in kvm_mmu_calculate_mmu_pages()