Lines Matching refs:kvm_vcpu
89 struct kvm_vcpu *runnable_threads[MAX_SMT_THREADS];
96 struct kvm_vcpu *runner;
156 extern void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong ea, ulong ea_mask);
157 extern void kvmppc_mmu_pte_vflush(struct kvm_vcpu *vcpu, u64 vp, u64 vp_mask);
158 extern void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end);
159 extern void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 new_msr);
160 extern void kvmppc_mmu_book3s_64_init(struct kvm_vcpu *vcpu);
161 extern void kvmppc_mmu_book3s_32_init(struct kvm_vcpu *vcpu);
162 extern void kvmppc_mmu_book3s_hv_init(struct kvm_vcpu *vcpu);
163 extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte,
165 extern void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte);
166 extern int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr);
167 extern void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong eaddr, ulong seg_size);
168 extern void kvmppc_mmu_flush_segments(struct kvm_vcpu *vcpu);
170 struct kvm_vcpu *vcpu, unsigned long addr,
174 extern int kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu,
177 extern void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte);
178 extern struct hpte_cache *kvmppc_mmu_hpte_cache_next(struct kvm_vcpu *vcpu);
180 extern void kvmppc_mmu_hpte_destroy(struct kvm_vcpu *vcpu);
181 extern int kvmppc_mmu_hpte_init(struct kvm_vcpu *vcpu);
182 extern void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte);
189 struct kvm_vcpu *vcpu,
191 extern int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
208 extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr, bool data);
209 extern void kvmppc_book3s_queue_irqprio(struct kvm_vcpu *vcpu, unsigned int vec);
210 extern void kvmppc_book3s_dequeue_irqprio(struct kvm_vcpu *vcpu,
212 extern void kvmppc_inject_interrupt(struct kvm_vcpu *vcpu, int vec, u64 flags);
213 extern void kvmppc_trigger_fac_interrupt(struct kvm_vcpu *vcpu, ulong fac);
214 extern void kvmppc_set_bat(struct kvm_vcpu *vcpu, struct kvmppc_bat *bat,
216 extern void kvmppc_giveup_ext(struct kvm_vcpu *vcpu, ulong msr);
217 extern int kvmppc_emulate_paired_single(struct kvm_run *run, struct kvm_vcpu *vcpu);
218 extern kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa,
245 extern void kvmppc_set_fscr(struct kvm_vcpu *vcpu, u64 fscr);
247 extern int kvmhv_p9_tm_emulation_early(struct kvm_vcpu *vcpu);
248 extern int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu);
249 extern void kvmhv_emulate_tm_rollback(struct kvm_vcpu *vcpu);
253 extern u32 kvmppc_alignment_dsisr(struct kvm_vcpu *vcpu, unsigned int inst);
254 extern ulong kvmppc_alignment_dar(struct kvm_vcpu *vcpu, unsigned int inst);
255 extern int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd);
259 extern void kvmppc_copy_to_svcpu(struct kvm_vcpu *vcpu);
260 extern void kvmppc_copy_from_svcpu(struct kvm_vcpu *vcpu);
263 void kvmppc_save_tm_pr(struct kvm_vcpu *vcpu);
264 void kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu);
265 void kvmppc_save_tm_sprs(struct kvm_vcpu *vcpu);
266 void kvmppc_restore_tm_sprs(struct kvm_vcpu *vcpu);
268 static inline void kvmppc_save_tm_pr(struct kvm_vcpu *vcpu) {} in kvmppc_save_tm_pr()
269 static inline void kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu) {} in kvmppc_restore_tm_pr()
270 static inline void kvmppc_save_tm_sprs(struct kvm_vcpu *vcpu) {} in kvmppc_save_tm_sprs()
271 static inline void kvmppc_restore_tm_sprs(struct kvm_vcpu *vcpu) {} in kvmppc_restore_tm_sprs()
274 void kvmppc_giveup_fac(struct kvm_vcpu *vcpu, ulong fac);
278 static inline struct kvmppc_vcpu_book3s *to_book3s(struct kvm_vcpu *vcpu) in to_book3s()
292 static inline void kvmppc_set_gpr(struct kvm_vcpu *vcpu, int num, ulong val) in kvmppc_set_gpr()
297 static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num) in kvmppc_get_gpr()
302 static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val) in kvmppc_set_cr()
307 static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu) in kvmppc_get_cr()
312 static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_xer()
317 static inline ulong kvmppc_get_xer(struct kvm_vcpu *vcpu) in kvmppc_get_xer()
322 static inline void kvmppc_set_ctr(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_ctr()
327 static inline ulong kvmppc_get_ctr(struct kvm_vcpu *vcpu) in kvmppc_get_ctr()
332 static inline void kvmppc_set_lr(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_lr()
337 static inline ulong kvmppc_get_lr(struct kvm_vcpu *vcpu) in kvmppc_get_lr()
342 static inline void kvmppc_set_pc(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_pc()
347 static inline ulong kvmppc_get_pc(struct kvm_vcpu *vcpu) in kvmppc_get_pc()
352 static inline u64 kvmppc_get_msr(struct kvm_vcpu *vcpu);
353 static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu) in kvmppc_need_byteswap()
358 static inline ulong kvmppc_get_fault_dar(struct kvm_vcpu *vcpu) in kvmppc_get_fault_dar()
369 static inline bool kvmppc_supports_magic_page(struct kvm_vcpu *vcpu) in kvmppc_supports_magic_page()
375 extern int kvmppc_h_logical_ci_load(struct kvm_vcpu *vcpu);
376 extern int kvmppc_h_logical_ci_store(struct kvm_vcpu *vcpu);