Lines Matching refs:memslot
54 static bool memslot_is_logging(struct kvm_memory_slot *memslot) in memslot_is_logging() argument
56 return memslot->dirty_bitmap && !(memslot->flags & KVM_MEM_READONLY); in memslot_is_logging()
405 struct kvm_memory_slot *memslot) in stage2_flush_memslot() argument
407 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_flush_memslot()
408 phys_addr_t end = addr + PAGE_SIZE * memslot->npages; in stage2_flush_memslot()
429 struct kvm_memory_slot *memslot; in stage2_flush_vm() local
436 kvm_for_each_memslot(memslot, slots) in stage2_flush_vm()
437 stage2_flush_memslot(kvm, memslot); in stage2_flush_vm()
909 struct kvm_memory_slot *memslot) in stage2_unmap_memslot() argument
911 hva_t hva = memslot->userspace_addr; in stage2_unmap_memslot()
912 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_unmap_memslot()
913 phys_addr_t size = PAGE_SIZE * memslot->npages; in stage2_unmap_memslot()
942 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); in stage2_unmap_memslot()
959 struct kvm_memory_slot *memslot; in stage2_unmap_vm() local
967 kvm_for_each_memslot(memslot, slots) in stage2_unmap_vm()
968 stage2_unmap_memslot(kvm, memslot); in stage2_unmap_vm()
1394 struct kvm_memory_slot *memslot = id_to_memslot(slots, slot); in kvm_mmu_wp_memory_region() local
1395 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1396 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1470 struct kvm_memory_slot *memslot, unsigned long hva, in user_mem_abort() argument
1482 bool logging_active = memslot_is_logging(memslot); in user_mem_abort()
1516 if ((memslot->userspace_addr & ~PMD_MASK) != in user_mem_abort()
1517 ((memslot->base_gfn << PAGE_SHIFT) & ~PMD_MASK)) in user_mem_abort()
1685 struct kvm_memory_slot *memslot; in kvm_handle_guest_abort() local
1727 memslot = gfn_to_memslot(vcpu->kvm, gfn); in kvm_handle_guest_abort()
1728 hva = gfn_to_hva_memslot_prot(memslot, gfn, &writable); in kvm_handle_guest_abort()
1774 ret = user_mem_abort(vcpu, fault_ipa, memslot, hva, fault_status); in kvm_handle_guest_abort()
1791 struct kvm_memory_slot *memslot; in handle_hva_to_gpa() local
1797 kvm_for_each_memslot(memslot, slots) { in handle_hva_to_gpa()
1801 hva_start = max(start, memslot->userspace_addr); in handle_hva_to_gpa()
1802 hva_end = min(end, memslot->userspace_addr + in handle_hva_to_gpa()
1803 (memslot->npages << PAGE_SHIFT)); in handle_hva_to_gpa()
1807 gpa = hva_to_gfn_memslot(hva_start, memslot) << PAGE_SHIFT; in handle_hva_to_gpa()
2047 struct kvm_memory_slot *memslot, in kvm_arch_prepare_memory_region() argument
2064 if (memslot->base_gfn + memslot->npages >= in kvm_arch_prepare_memory_region()
2112 if (memslot->flags & KVM_MEM_LOG_DIRTY_PAGES) { in kvm_arch_prepare_memory_region()
2133 stage2_flush_memslot(kvm, memslot); in kvm_arch_prepare_memory_region()