Lines Matching refs:svcpu

271 	struct kvmppc_book3s_shadow_vcpu *svcpu = svcpu_get(vcpu);  in kvmppc_mmu_next_segment()  local
278 for (i = 0; i < svcpu->slb_max; i++) { in kvmppc_mmu_next_segment()
279 if (!(svcpu->slb[i].esid & SLB_ESID_V)) in kvmppc_mmu_next_segment()
281 else if ((svcpu->slb[i].esid & ESID_MASK) == esid) { in kvmppc_mmu_next_segment()
299 if ((svcpu->slb_max) == max_slb_size) in kvmppc_mmu_next_segment()
302 r = svcpu->slb_max; in kvmppc_mmu_next_segment()
303 svcpu->slb_max++; in kvmppc_mmu_next_segment()
306 svcpu_put(svcpu); in kvmppc_mmu_next_segment()
312 struct kvmppc_book3s_shadow_vcpu *svcpu = svcpu_get(vcpu); in kvmppc_mmu_map_segment() local
325 svcpu->slb[slb_index].esid = 0; in kvmppc_mmu_map_segment()
346 svcpu->slb[slb_index].esid = slb_esid; in kvmppc_mmu_map_segment()
347 svcpu->slb[slb_index].vsid = slb_vsid; in kvmppc_mmu_map_segment()
352 svcpu_put(svcpu); in kvmppc_mmu_map_segment()
358 struct kvmppc_book3s_shadow_vcpu *svcpu = svcpu_get(vcpu); in kvmppc_mmu_flush_segment() local
362 for (i = 0; i < svcpu->slb_max; i++) { in kvmppc_mmu_flush_segment()
363 if ((svcpu->slb[i].esid & SLB_ESID_V) && in kvmppc_mmu_flush_segment()
364 (svcpu->slb[i].esid & seg_mask) == ea) { in kvmppc_mmu_flush_segment()
366 svcpu->slb[i].esid = 0; in kvmppc_mmu_flush_segment()
370 svcpu_put(svcpu); in kvmppc_mmu_flush_segment()
375 struct kvmppc_book3s_shadow_vcpu *svcpu = svcpu_get(vcpu); in kvmppc_mmu_flush_segments() local
376 svcpu->slb_max = 0; in kvmppc_mmu_flush_segments()
377 svcpu->slb[0].esid = 0; in kvmppc_mmu_flush_segments()
378 svcpu_put(svcpu); in kvmppc_mmu_flush_segments()