Lines Matching refs:pte_index

83 			     unsigned long *rmap, long pte_index, int realmode)  in kvmppc_add_revmap_chain()  argument
98 tail->forw = pte_index; in kvmppc_add_revmap_chain()
99 head->back = pte_index; in kvmppc_add_revmap_chain()
101 rev->forw = rev->back = pte_index; in kvmppc_add_revmap_chain()
103 pte_index | KVMPPC_RMAP_PRESENT; in kvmppc_add_revmap_chain()
161 static void remove_revmap_chain(struct kvm *kvm, long pte_index, in remove_revmap_chain() argument
184 if (head == pte_index) { in remove_revmap_chain()
186 if (head == pte_index) in remove_revmap_chain()
199 long pte_index, unsigned long pteh, unsigned long ptel, in kvmppc_do_h_enter() argument
318 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) in kvmppc_do_h_enter()
321 pte_index &= ~7UL; in kvmppc_do_h_enter()
322 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); in kvmppc_do_h_enter()
351 pte_index += i; in kvmppc_do_h_enter()
353 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); in kvmppc_do_h_enter()
370 rev = &kvm->arch.hpt.rev[pte_index]; in kvmppc_do_h_enter()
391 kvmppc_add_revmap_chain(kvm, rev, rmap, pte_index, in kvmppc_do_h_enter()
411 *pte_idx_ret = pte_index; in kvmppc_do_h_enter()
417 long pte_index, unsigned long pteh, unsigned long ptel) in kvmppc_h_enter() argument
419 return kvmppc_do_h_enter(vcpu->kvm, flags, pte_index, pteh, ptel, in kvmppc_h_enter()
478 unsigned long pte_index, unsigned long avpn, in kvmppc_do_h_remove() argument
488 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) in kvmppc_do_h_remove()
490 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); in kvmppc_do_h_remove()
506 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); in kvmppc_do_h_remove()
510 rb = compute_tlbie_rb(v, pte_r, pte_index); in kvmppc_do_h_remove()
521 remove_revmap_chain(kvm, pte_index, rev, v, in kvmppc_do_h_remove()
540 unsigned long pte_index, unsigned long avpn) in kvmppc_h_remove() argument
542 return kvmppc_do_h_remove(vcpu->kvm, flags, pte_index, avpn, in kvmppc_h_remove()
553 unsigned long flags, req, pte_index, rcbits; in kvmppc_h_bulk_remove() local
566 pte_index = args[j]; in kvmppc_h_bulk_remove()
567 flags = pte_index >> 56; in kvmppc_h_bulk_remove()
568 pte_index &= ((1ul << 56) - 1); in kvmppc_h_bulk_remove()
576 pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) { in kvmppc_h_bulk_remove()
578 args[j] = ((0xa0 | flags) << 56) + pte_index; in kvmppc_h_bulk_remove()
582 hp = (__be64 *) (kvm->arch.hpt.virt + (pte_index << 4)); in kvmppc_h_bulk_remove()
614 args[j] = ((0x90 | flags) << 56) + pte_index; in kvmppc_h_bulk_remove()
618 args[j] = ((0x80 | flags) << 56) + pte_index; in kvmppc_h_bulk_remove()
619 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); in kvmppc_h_bulk_remove()
634 tlbrb[n] = compute_tlbie_rb(hp0, hp1, pte_index); in kvmppc_h_bulk_remove()
650 pte_index = args[j] & ((1ul << 56) - 1); in kvmppc_h_bulk_remove()
653 remove_revmap_chain(kvm, pte_index, rev, in kvmppc_h_bulk_remove()
665 unsigned long pte_index, unsigned long avpn, in kvmppc_h_protect() argument
676 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) in kvmppc_h_protect()
679 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); in kvmppc_h_protect()
699 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); in kvmppc_h_protect()
718 rb = compute_tlbie_rb(v, r, pte_index); in kvmppc_h_protect()
736 unsigned long pte_index) in kvmppc_h_read() argument
746 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) in kvmppc_h_read()
749 pte_index &= ~3; in kvmppc_h_read()
752 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); in kvmppc_h_read()
753 for (i = 0; i < n; ++i, ++pte_index) { in kvmppc_h_read()
754 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); in kvmppc_h_read()
776 unsigned long pte_index) in kvmppc_h_clear_ref() argument
787 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) in kvmppc_h_clear_ref()
790 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); in kvmppc_h_clear_ref()
791 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); in kvmppc_h_clear_ref()
807 kvmppc_clear_ref_hpte(kvm, hpte, pte_index); in kvmppc_h_clear_ref()
824 unsigned long pte_index) in kvmppc_h_clear_mod() argument
834 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) in kvmppc_h_clear_mod()
837 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); in kvmppc_h_clear_mod()
838 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); in kvmppc_h_clear_mod()
854 kvmppc_invalidate_hpte(kvm, hpte, pte_index); in kvmppc_h_clear_mod()
871 unsigned long pte_index) in kvmppc_invalidate_hpte() argument
883 rb = compute_tlbie_rb(hp0, hp1, pte_index); in kvmppc_invalidate_hpte()
889 unsigned long pte_index) in kvmppc_clear_ref_hpte() argument
901 rb = compute_tlbie_rb(hp0, hp1, pte_index); in kvmppc_clear_ref_hpte()
1065 index = cache_entry->pte_index; in kvmppc_hpte_hv_fault()
1141 cache_entry->pte_index = index; in kvmppc_hpte_hv_fault()