Lines Matching full:ics
37 /* -- ICS routines -- */
39 struct kvmppc_ics *ics, struct kvmppc_icp *icp) in ics_rm_check_resend() argument
44 struct ics_irq_state *state = &ics->irq_state[i]; in ics_rm_check_resend()
236 struct kvmppc_ics *ics = xics->ics[icsid]; in icp_rm_check_resend() local
240 if (!ics) in icp_rm_check_resend()
242 ics_rm_check_resend(xics, ics, icp); in icp_rm_check_resend()
288 struct kvmppc_ics *ics; in icp_rm_deliver_irq() local
308 /* Get the ICS state and lock it */ in icp_rm_deliver_irq()
309 ics = kvmppc_xics_find_ics(xics, new_irq, &src); in icp_rm_deliver_irq()
310 if (!ics) { in icp_rm_deliver_irq()
315 state = &ics->irq_state[src]; in icp_rm_deliver_irq()
317 /* Get a lock on the ICS */ in icp_rm_deliver_irq()
318 arch_spin_lock(&ics->lock); in icp_rm_deliver_irq()
364 * ics spin lock. in icp_rm_deliver_irq()
378 arch_spin_unlock(&ics->lock); in icp_rm_deliver_irq()
387 * resend map bit and mark the ICS state as needing a resend in icp_rm_deliver_irq()
396 set_bit(ics->icsid, icp->resend_map); in icp_rm_deliver_irq()
407 arch_spin_unlock(&ics->lock); in icp_rm_deliver_irq()
413 arch_spin_unlock(&ics->lock); in icp_rm_deliver_irq()
433 * and notify ICS of a potential resend. This is done in icp_rm_down_cppr()
681 struct kvmppc_ics *ics; in ics_rm_eoi() local
687 * ICS EOI handling: For LSI, if P bit is still set, we need to in ics_rm_eoi()
694 ics = kvmppc_xics_find_ics(xics, irq, &src); in ics_rm_eoi()
695 if (!ics) in ics_rm_eoi()
698 state = &ics->irq_state[src]; in ics_rm_eoi()
752 * The sending of an EOI to the ICS is handled after the in xics_rm_h_eoi()
861 struct kvmppc_ics *ics; in kvmppc_deliver_irq_passthru() local
873 ics = kvmppc_xics_find_ics(xics, irq, &src); in kvmppc_deliver_irq_passthru()
874 if (!ics) in kvmppc_deliver_irq_passthru()
877 state = &ics->irq_state[src]; in kvmppc_deliver_irq_passthru()