Lines Matching refs:gprs
66 arch->gprs[insn.r_format.rd] = epc + 8; in kvm_compute_return_epc()
69 nextpc = arch->gprs[insn.r_format.rs]; in kvm_compute_return_epc()
85 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc()
94 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc()
103 arch->gprs[31] = epc + 8; in kvm_compute_return_epc()
104 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc()
113 arch->gprs[31] = epc + 8; in kvm_compute_return_epc()
114 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc()
142 arch->gprs[31] = instpc + 8; in kvm_compute_return_epc()
155 if (arch->gprs[insn.i_format.rs] == in kvm_compute_return_epc()
156 arch->gprs[insn.i_format.rt]) in kvm_compute_return_epc()
165 if (arch->gprs[insn.i_format.rs] != in kvm_compute_return_epc()
166 arch->gprs[insn.i_format.rt]) in kvm_compute_return_epc()
179 if ((long)arch->gprs[insn.i_format.rs] <= 0) in kvm_compute_return_epc()
192 if ((long)arch->gprs[insn.i_format.rs] > 0) in kvm_compute_return_epc()
1321 vcpu->arch.gprs[rt] = in kvm_mips_emulate_CP0()
1324 vcpu->arch.gprs[rt] = 0x0; in kvm_mips_emulate_CP0()
1329 vcpu->arch.gprs[rt] = (s32)cop0->reg[rd][sel]; in kvm_mips_emulate_CP0()
1338 vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1342 vcpu->arch.gprs[rt] = cop0->reg[rd][sel]; in kvm_mips_emulate_CP0()
1346 vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1355 vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1358 && (vcpu->arch.gprs[rt] >= in kvm_mips_emulate_CP0()
1361 vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1371 vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1374 vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1378 kvm_mips_write_count(vcpu, vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1384 vcpu->arch.gprs[rt], in kvm_mips_emulate_CP0()
1390 val = vcpu->arch.gprs[rt]; in kvm_mips_emulate_CP0()
1462 val = vcpu->arch.gprs[rt]; in kvm_mips_emulate_CP0()
1499 new_cause = vcpu->arch.gprs[rt]; in kvm_mips_emulate_CP0()
1519 cop0->reg[rd][sel] = vcpu->arch.gprs[rt] & mask; in kvm_mips_emulate_CP0()
1521 cop0->reg[rd][sel] = vcpu->arch.gprs[rt]; in kvm_mips_emulate_CP0()
1533 vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1542 vcpu->arch.gprs[rt] = in kvm_mips_emulate_CP0()
1571 vcpu->arch.gprs[rt]); in kvm_mips_emulate_CP0()
1572 vcpu->arch.gprs[rd] = vcpu->arch.gprs[rt]; in kvm_mips_emulate_CP0()
1628 *(u64 *)data = vcpu->arch.gprs[rt]; in kvm_mips_emulate_store()
1632 vcpu->arch.gprs[rt], *(u64 *)data); in kvm_mips_emulate_store()
1638 *(u32 *)data = vcpu->arch.gprs[rt]; in kvm_mips_emulate_store()
1642 vcpu->arch.gprs[rt], *(u32 *)data); in kvm_mips_emulate_store()
1647 *(u16 *)data = vcpu->arch.gprs[rt]; in kvm_mips_emulate_store()
1651 vcpu->arch.gprs[rt], *(u16 *)data); in kvm_mips_emulate_store()
1656 *(u8 *)data = vcpu->arch.gprs[rt]; in kvm_mips_emulate_store()
1660 vcpu->arch.gprs[rt], *(u8 *)data); in kvm_mips_emulate_store()
1827 va = arch->gprs[base] + offset; in kvm_mips_emulate_cache()
1830 cache, op, base, arch->gprs[base], offset); in kvm_mips_emulate_cache()
1839 vcpu->arch.pc, vcpu->arch.gprs[31], cache, op, base, in kvm_mips_emulate_cache()
1840 arch->gprs[base], offset); in kvm_mips_emulate_cache()
1916 cache, op, base, arch->gprs[base], offset); in kvm_mips_emulate_cache()
2530 arch->gprs[rt] = vcpu->vcpu_id; in kvm_mips_handle_ri()
2533 arch->gprs[rt] = min(current_cpu_data.dcache.linesz, in kvm_mips_handle_ri()
2537 arch->gprs[rt] = (s32)kvm_mips_read_count(vcpu); in kvm_mips_handle_ri()
2543 arch->gprs[rt] = 1; in kvm_mips_handle_ri()
2546 arch->gprs[rt] = 2; in kvm_mips_handle_ri()
2550 arch->gprs[rt] = kvm_read_c0_guest_userlocal(cop0); in kvm_mips_handle_ri()
2559 vcpu->arch.gprs[rt]); in kvm_mips_handle_ri()
2580 unsigned long *gpr = &vcpu->arch.gprs[vcpu->arch.io_gpr]; in kvm_mips_complete_mmio_load()