Lines Matching refs:phys_addr_t
44 static phys_addr_t hyp_idmap_vector;
70 static void kvm_tlb_flush_vmid_ipa(struct kvm *kvm, phys_addr_t ipa) in kvm_tlb_flush_vmid_ipa()
109 static void stage2_dissolve_pmd(struct kvm *kvm, phys_addr_t addr, pmd_t *pmd) in stage2_dissolve_pmd()
151 static void clear_stage2_pgd_entry(struct kvm *kvm, pgd_t *pgd, phys_addr_t addr) in clear_stage2_pgd_entry()
160 static void clear_stage2_pud_entry(struct kvm *kvm, pud_t *pud, phys_addr_t addr) in clear_stage2_pud_entry()
170 static void clear_stage2_pmd_entry(struct kvm *kvm, pmd_t *pmd, phys_addr_t addr) in clear_stage2_pmd_entry()
234 phys_addr_t addr, phys_addr_t end) in unmap_stage2_ptes()
236 phys_addr_t start_addr = addr; in unmap_stage2_ptes()
260 phys_addr_t addr, phys_addr_t end) in unmap_stage2_pmds()
262 phys_addr_t next, start_addr = addr; in unmap_stage2_pmds()
289 phys_addr_t addr, phys_addr_t end) in unmap_stage2_puds()
291 phys_addr_t next, start_addr = addr; in unmap_stage2_puds()
326 static void unmap_stage2_range(struct kvm *kvm, phys_addr_t start, u64 size) in unmap_stage2_range()
329 phys_addr_t addr = start, end = start + size; in unmap_stage2_range()
330 phys_addr_t next; in unmap_stage2_range()
357 phys_addr_t addr, phys_addr_t end) in stage2_flush_ptes()
369 phys_addr_t addr, phys_addr_t end) in stage2_flush_pmds()
372 phys_addr_t next; in stage2_flush_pmds()
387 phys_addr_t addr, phys_addr_t end) in stage2_flush_puds()
390 phys_addr_t next; in stage2_flush_puds()
407 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_flush_memslot()
408 phys_addr_t end = addr + PAGE_SIZE * memslot->npages; in stage2_flush_memslot()
409 phys_addr_t next; in stage2_flush_memslot()
469 static void unmap_hyp_ptes(pmd_t *pmd, phys_addr_t addr, phys_addr_t end) in unmap_hyp_ptes()
485 static void unmap_hyp_pmds(pud_t *pud, phys_addr_t addr, phys_addr_t end) in unmap_hyp_pmds()
487 phys_addr_t next; in unmap_hyp_pmds()
502 static void unmap_hyp_puds(pgd_t *pgd, phys_addr_t addr, phys_addr_t end) in unmap_hyp_puds()
504 phys_addr_t next; in unmap_hyp_puds()
525 phys_addr_t start, u64 size) in __unmap_hyp_range()
528 phys_addr_t addr = start, end = start + size; in __unmap_hyp_range()
529 phys_addr_t next; in __unmap_hyp_range()
543 static void unmap_hyp_range(pgd_t *pgdp, phys_addr_t start, u64 size) in unmap_hyp_range()
548 static void unmap_hyp_idmap_range(pgd_t *pgdp, phys_addr_t start, u64 size) in unmap_hyp_idmap_range()
719 static phys_addr_t kvm_kaddr_to_phys(void *kaddr) in kvm_kaddr_to_phys()
742 phys_addr_t phys_addr; in create_hyp_mappings()
768 static int __create_hyp_private_mapping(phys_addr_t phys_addr, size_t size, in __create_hyp_private_mapping()
825 int create_hyp_io_mappings(phys_addr_t phys_addr, size_t size, in create_hyp_io_mappings()
860 int create_hyp_exec_mappings(phys_addr_t phys_addr, size_t size, in create_hyp_exec_mappings()
912 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_unmap_memslot()
913 phys_addr_t size = PAGE_SIZE * memslot->npages; in stage2_unmap_memslot()
1001 phys_addr_t addr) in stage2_get_pud()
1019 phys_addr_t addr) in stage2_get_pmd()
1040 *cache, phys_addr_t addr, const pmd_t *new_pmd) in stage2_set_pmd_huge()
1086 static bool stage2_is_exec(struct kvm *kvm, phys_addr_t addr) in stage2_is_exec()
1106 phys_addr_t addr, const pte_t *new_pte, in stage2_set_pte()
1193 int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa, in kvm_phys_addr_ioremap()
1194 phys_addr_t pa, unsigned long size, bool writable) in kvm_phys_addr_ioremap()
1196 phys_addr_t addr, end; in kvm_phys_addr_ioremap()
1229 static bool transparent_hugepage_adjust(kvm_pfn_t *pfnp, phys_addr_t *ipap) in transparent_hugepage_adjust()
1284 static void stage2_wp_ptes(pmd_t *pmd, phys_addr_t addr, phys_addr_t end) in stage2_wp_ptes()
1303 static void stage2_wp_pmds(pud_t *pud, phys_addr_t addr, phys_addr_t end) in stage2_wp_pmds()
1306 phys_addr_t next; in stage2_wp_pmds()
1331 static void stage2_wp_puds(pgd_t *pgd, phys_addr_t addr, phys_addr_t end) in stage2_wp_puds()
1334 phys_addr_t next; in stage2_wp_puds()
1353 static void stage2_wp_range(struct kvm *kvm, phys_addr_t addr, phys_addr_t end) in stage2_wp_range()
1356 phys_addr_t next; in stage2_wp_range()
1395 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1396 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1419 phys_addr_t base_gfn = slot->base_gfn + gfn_offset; in kvm_mmu_write_protect_pt_masked()
1420 phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; in kvm_mmu_write_protect_pt_masked()
1421 phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; in kvm_mmu_write_protect_pt_masked()
1469 static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa, in user_mem_abort()
1634 static void handle_access_fault(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa) in handle_access_fault()
1684 phys_addr_t fault_ipa; in kvm_handle_guest_abort()
1929 phys_addr_t kvm_mmu_get_httbr(void) in kvm_mmu_get_httbr()
1937 phys_addr_t kvm_get_idmap_vector(void) in kvm_get_idmap_vector()
2106 phys_addr_t pa; in kvm_arch_prepare_memory_region()
2108 pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT; in kvm_arch_prepare_memory_region()
2164 phys_addr_t size = slot->npages << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot()