| /Linux-v4.19/virt/kvm/arm/vgic/ |
| D | vgic.c | 107 return &vcpu->arch.vgic_cpu.private_irqs[intid]; in vgic_get_irq() 274 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_sort_ap_list() local 276 DEBUG_SPINLOCK_BUG_ON(!spin_is_locked(&vgic_cpu->ap_list_lock)); in vgic_sort_ap_list() 278 list_sort(NULL, &vgic_cpu->ap_list_head, vgic_irq_cmp); in vgic_sort_ap_list() 354 spin_lock_irqsave(&vcpu->arch.vgic_cpu.ap_list_lock, flags); in vgic_queue_irq_unlock() 371 spin_unlock_irqrestore(&vcpu->arch.vgic_cpu.ap_list_lock, flags); in vgic_queue_irq_unlock() 382 list_add_tail(&irq->ap_list, &vcpu->arch.vgic_cpu.ap_list_head); in vgic_queue_irq_unlock() 386 spin_unlock_irqrestore(&vcpu->arch.vgic_cpu.ap_list_lock, flags); in vgic_queue_irq_unlock() 594 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_prune_ap_list() local 600 spin_lock(&vgic_cpu->ap_list_lock); in vgic_prune_ap_list() [all …]
|
| D | vgic-mmio-v3.c | 185 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_mmio_read_v3r_ctlr() local 187 return vgic_cpu->lpis_enabled ? GICR_CTLR_ENABLE_LPIS : 0; in vgic_mmio_read_v3r_ctlr() 195 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_mmio_write_v3r_ctlr() local 196 bool was_enabled = vgic_cpu->lpis_enabled; in vgic_mmio_write_v3r_ctlr() 201 vgic_cpu->lpis_enabled = val & GICR_CTLR_ENABLE_LPIS; in vgic_mmio_write_v3r_ctlr() 203 if (!was_enabled && vgic_cpu->lpis_enabled) in vgic_mmio_write_v3r_ctlr() 211 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_mmio_read_v3r_typer() local 212 struct vgic_redist_region *rdreg = vgic_cpu->rdreg; in vgic_mmio_read_v3r_typer() 402 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_mmio_write_propbase() local 406 if (vgic_cpu->lpis_enabled) in vgic_mmio_write_propbase() [all …]
|
| D | vgic-v2.c | 42 struct vgic_v2_cpu_if *cpuif = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_set_underflow() 62 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_v2_fold_lr_state() local 63 struct vgic_v2_cpu_if *cpuif = &vgic_cpu->vgic_v2; in vgic_v2_fold_lr_state() 70 for (lr = 0; lr < vgic_cpu->used_lrs; lr++) { in vgic_v2_fold_lr_state() 134 vgic_cpu->used_lrs = 0; in vgic_v2_fold_lr_state() 220 vcpu->arch.vgic_cpu.vgic_v2.vgic_lr[lr] = val; in vgic_v2_populate_lr() 225 vcpu->arch.vgic_cpu.vgic_v2.vgic_lr[lr] = 0; in vgic_v2_clear_lr() 230 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_set_vmcr() 257 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_get_vmcr() 290 vcpu->arch.vgic_cpu.vgic_v2.vgic_vmcr = 0; in vgic_v2_enable() [all …]
|
| D | vgic-v3.c | 32 struct vgic_v3_cpu_if *cpuif = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_set_underflow() 45 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in vgic_v3_fold_lr_state() local 46 struct vgic_v3_cpu_if *cpuif = &vgic_cpu->vgic_v3; in vgic_v3_fold_lr_state() 54 for (lr = 0; lr < vgic_cpu->used_lrs; lr++) { in vgic_v3_fold_lr_state() 126 vgic_cpu->used_lrs = 0; in vgic_v3_fold_lr_state() 206 vcpu->arch.vgic_cpu.vgic_v3.vgic_lr[lr] = val; in vgic_v3_populate_lr() 211 vcpu->arch.vgic_cpu.vgic_v3.vgic_lr[lr] = 0; in vgic_v3_clear_lr() 216 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_set_vmcr() 246 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_get_vmcr() 282 struct vgic_v3_cpu_if *vgic_v3 = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_enable() [all …]
|
| D | vgic-init.c | 200 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in kvm_vgic_vcpu_init() local 205 vgic_cpu->rd_iodev.base_addr = VGIC_ADDR_UNDEF; in kvm_vgic_vcpu_init() 206 vgic_cpu->sgi_iodev.base_addr = VGIC_ADDR_UNDEF; in kvm_vgic_vcpu_init() 208 INIT_LIST_HEAD(&vgic_cpu->ap_list_head); in kvm_vgic_vcpu_init() 209 spin_lock_init(&vgic_cpu->ap_list_lock); in kvm_vgic_vcpu_init() 216 struct vgic_irq *irq = &vgic_cpu->private_irqs[i]; in kvm_vgic_vcpu_init() 349 struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; in kvm_vgic_vcpu_destroy() local 351 INIT_LIST_HEAD(&vgic_cpu->ap_list_head); in kvm_vgic_vcpu_destroy()
|
| D | vgic-v4.c | 99 vcpu->arch.vgic_cpu.vgic_v3.its_vpe.pending_last = true; in vgic_v4_doorbell_handler() 137 dist->its_vm.vpes[i] = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe; in vgic_v4_init() 211 return its_schedule_vpe(&vcpu->arch.vgic_cpu.vgic_v3.its_vpe, false); in vgic_v4_sync_hwstate() 216 int irq = vcpu->arch.vgic_cpu.vgic_v3.its_vpe.irq; in vgic_v4_flush_hwstate() 232 err = its_schedule_vpe(&vcpu->arch.vgic_cpu.vgic_v3.its_vpe, true); in vgic_v4_flush_hwstate() 294 .vpe = &irq->target_vcpu->arch.vgic_cpu.vgic_v3.its_vpe, in kvm_vgic_v4_set_forwarding() 353 int irq = vcpu->arch.vgic_cpu.vgic_v3.its_vpe.irq; in kvm_vgic_v4_enable_doorbell() 362 int irq = vcpu->arch.vgic_cpu.vgic_v3.its_vpe.irq; in kvm_vgic_v4_disable_doorbell()
|
| D | vgic-mmio-v2.c | 370 return vcpu->arch.vgic_cpu.vgic_v2.vgic_apr; in vgic_mmio_read_apr() 372 struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_mmio_read_apr() 396 vcpu->arch.vgic_cpu.vgic_v2.vgic_apr = val; in vgic_mmio_write_apr() 398 struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_mmio_write_apr()
|
| D | vgic.h | 267 struct vgic_cpu *cpu_if = &vcpu->arch.vgic_cpu; in vgic_v3_max_apr_idx()
|
| D | vgic-its.c | 373 map.vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe; in update_affinity() 430 gpa_t pendbase = GICR_PENDBASER_ADDRESS(vcpu->arch.vgic_cpu.pendbaser); in its_sync_lpi_pending_table() 565 if (!vcpu->arch.vgic_cpu.lpis_enabled) in vgic_its_resolve_lpi() 1178 if (vcpu->arch.vgic_cpu.vgic_v3.its_vpe.its_vm) in vgic_its_cmd_handle_invall() 1179 its_invall_vpe(&vcpu->arch.vgic_cpu.vgic_v3.its_vpe); in vgic_its_cmd_handle_invall() 1648 if (!(vcpu->arch.vgic_cpu.pendbaser & GICR_PENDBASER_PTZ)) in vgic_enable_lpis()
|
| /Linux-v4.19/virt/kvm/arm/hyp/ |
| D | vgic-v3-sr.c | 210 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in __vgic_v3_save_state() 211 u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; in __vgic_v3_save_state() 246 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in __vgic_v3_restore_state() 247 u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; in __vgic_v3_restore_state() 273 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in __vgic_v3_activate_traps() 322 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in __vgic_v3_deactivate_traps() 354 cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in __vgic_v3_save_aprs() 387 cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in __vgic_v3_restore_aprs() 459 unsigned int used_lrs = vcpu->arch.vgic_cpu.used_lrs; in __vgic_v3_highest_priority_lr() 498 unsigned int used_lrs = vcpu->arch.vgic_cpu.used_lrs; in __vgic_v3_find_active_lr()
|
| /Linux-v4.19/arch/arm64/kvm/ |
| D | vgic-sys-reg-v3.c | 25 struct vgic_cpu *vgic_v3_cpu = &vcpu->arch.vgic_cpu; in access_gic_ctlr() 194 struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_access_apr_reg() 241 struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3; in access_gic_sre()
|
| D | sys_regs.c | 303 p->regval = vcpu->arch.vgic_cpu.vgic_v3.vgic_sre; in access_gic_sre()
|
| /Linux-v4.19/arch/arm/include/asm/ |
| D | kvm_host.h | 169 struct vgic_cpu vgic_cpu; member
|
| /Linux-v4.19/arch/arm64/include/asm/ |
| D | kvm_host.h | 256 struct vgic_cpu vgic_cpu; member
|
| /Linux-v4.19/include/kvm/ |
| D | arm_vgic.h | 300 struct vgic_cpu { struct
|
| /Linux-v4.19/arch/arm/kvm/ |
| D | coproc.c | 287 *vcpu_reg(vcpu, p->Rt1) = vcpu->arch.vgic_cpu.vgic_v3.vgic_sre; in access_gic_sre()
|