Lines Matching full:ics

37 /* -- ICS routines -- */
39 struct kvmppc_ics *ics, struct kvmppc_icp *icp) in ics_rm_check_resend() argument
44 struct ics_irq_state *state = &ics->irq_state[i]; in ics_rm_check_resend()
229 struct kvmppc_ics *ics = xics->ics[icsid]; in icp_rm_check_resend() local
233 if (!ics) in icp_rm_check_resend()
235 ics_rm_check_resend(xics, ics, icp); in icp_rm_check_resend()
281 struct kvmppc_ics *ics; in icp_rm_deliver_irq() local
301 /* Get the ICS state and lock it */ in icp_rm_deliver_irq()
302 ics = kvmppc_xics_find_ics(xics, new_irq, &src); in icp_rm_deliver_irq()
303 if (!ics) { in icp_rm_deliver_irq()
308 state = &ics->irq_state[src]; in icp_rm_deliver_irq()
310 /* Get a lock on the ICS */ in icp_rm_deliver_irq()
311 arch_spin_lock(&ics->lock); in icp_rm_deliver_irq()
357 * ics spin lock. in icp_rm_deliver_irq()
371 arch_spin_unlock(&ics->lock); in icp_rm_deliver_irq()
380 * resend map bit and mark the ICS state as needing a resend in icp_rm_deliver_irq()
389 set_bit(ics->icsid, icp->resend_map); in icp_rm_deliver_irq()
400 arch_spin_unlock(&ics->lock); in icp_rm_deliver_irq()
406 arch_spin_unlock(&ics->lock); in icp_rm_deliver_irq()
426 * and notify ICS of a potential resend. This is done in icp_rm_down_cppr()
679 struct kvmppc_ics *ics; in ics_rm_eoi() local
685 * ICS EOI handling: For LSI, if P bit is still set, we need to in ics_rm_eoi()
692 ics = kvmppc_xics_find_ics(xics, irq, &src); in ics_rm_eoi()
693 if (!ics) in ics_rm_eoi()
696 state = &ics->irq_state[src]; in ics_rm_eoi()
751 * The sending of an EOI to the ICS is handled after the in xics_rm_h_eoi()
852 struct kvmppc_ics *ics; in kvmppc_deliver_irq_passthru() local
864 ics = kvmppc_xics_find_ics(xics, irq, &src); in kvmppc_deliver_irq_passthru()
865 if (!ics) in kvmppc_deliver_irq_passthru()
868 state = &ics->irq_state[src]; in kvmppc_deliver_irq_passthru()