Lines Matching full:irq

9  * chip.  When an event is received, it is mapped to an irq and sent
29 #include <linux/irq.h>
45 #include <asm/irq.h>
84 * Packed IRQ information:
86 * event channel - irq->event channel mapping
89 * PIRQ - vector, with MSB being "needs EIO", or physical IRQ of the HVM
90 * guest, or GSI (real passthrough IRQ) of the device.
107 unsigned irq; member
149 * Adding of event channels is no issue as the associated IRQ becomes active
150 * only after everything is setup (before request_[threaded_]irq() the handler
161 * IRQ-desc lock
168 /* IRQ <-> VIRQ mapping. */
171 /* IRQ <-> IPI mapping */
181 static bool (*pirq_needs_eoi)(unsigned irq);
220 static int set_evtchn_to_irq(evtchn_port_t evtchn, unsigned int irq) in set_evtchn_to_irq() argument
233 /* Unallocated irq entries return -1 anyway */ in set_evtchn_to_irq()
234 if (irq == -1) in set_evtchn_to_irq()
251 WRITE_ONCE(evtchn_to_irq[row][col], irq); in set_evtchn_to_irq()
264 /* Get info for IRQ */
265 static struct irq_info *info_for_irq(unsigned irq) in info_for_irq() argument
267 if (irq < nr_legacy_irqs()) in info_for_irq()
268 return legacy_info_ptrs[irq]; in info_for_irq()
270 return irq_get_chip_data(irq); in info_for_irq()
273 static void set_info_for_irq(unsigned int irq, struct irq_info *info) in set_info_for_irq() argument
275 if (irq < nr_legacy_irqs()) in set_info_for_irq()
276 legacy_info_ptrs[irq] = info; in set_info_for_irq()
278 irq_set_chip_data(irq, info); in set_info_for_irq()
307 /* Constructors for packed IRQ information. */
309 unsigned irq, in xen_irq_info_common_setup() argument
319 info->irq = irq; in xen_irq_info_common_setup()
325 ret = set_evtchn_to_irq(evtchn, irq); in xen_irq_info_common_setup()
329 irq_clear_status_flags(irq, IRQ_NOREQUEST|IRQ_NOAUTOEN); in xen_irq_info_common_setup()
334 static int xen_irq_info_evtchn_setup(unsigned irq, in xen_irq_info_evtchn_setup() argument
338 struct irq_info *info = info_for_irq(irq); in xen_irq_info_evtchn_setup()
341 ret = xen_irq_info_common_setup(info, irq, IRQT_EVTCHN, evtchn, 0); in xen_irq_info_evtchn_setup()
350 unsigned irq, in xen_irq_info_ipi_setup() argument
354 struct irq_info *info = info_for_irq(irq); in xen_irq_info_ipi_setup()
358 per_cpu(ipi_to_irq, cpu)[ipi] = irq; in xen_irq_info_ipi_setup()
360 return xen_irq_info_common_setup(info, irq, IRQT_IPI, evtchn, 0); in xen_irq_info_ipi_setup()
364 unsigned irq, in xen_irq_info_virq_setup() argument
368 struct irq_info *info = info_for_irq(irq); in xen_irq_info_virq_setup()
372 per_cpu(virq_to_irq, cpu)[virq] = irq; in xen_irq_info_virq_setup()
374 return xen_irq_info_common_setup(info, irq, IRQT_VIRQ, evtchn, 0); in xen_irq_info_virq_setup()
377 static int xen_irq_info_pirq_setup(unsigned irq, in xen_irq_info_pirq_setup() argument
384 struct irq_info *info = info_for_irq(irq); in xen_irq_info_pirq_setup()
391 return xen_irq_info_common_setup(info, irq, IRQT_PIRQ, evtchn, 0); in xen_irq_info_pirq_setup()
403 * Accessors for packed IRQ information.
405 evtchn_port_t evtchn_from_irq(unsigned irq) in evtchn_from_irq() argument
409 if (likely(irq < nr_irqs)) in evtchn_from_irq()
410 info = info_for_irq(irq); in evtchn_from_irq()
428 static enum ipi_vector ipi_from_irq(unsigned irq) in ipi_from_irq() argument
430 struct irq_info *info = info_for_irq(irq); in ipi_from_irq()
438 static unsigned virq_from_irq(unsigned irq) in virq_from_irq() argument
440 struct irq_info *info = info_for_irq(irq); in virq_from_irq()
448 static unsigned pirq_from_irq(unsigned irq) in pirq_from_irq() argument
450 struct irq_info *info = info_for_irq(irq); in pirq_from_irq()
458 static enum xen_irq_type type_from_irq(unsigned irq) in type_from_irq() argument
460 return info_for_irq(irq)->type; in type_from_irq()
463 static unsigned cpu_from_irq(unsigned irq) in cpu_from_irq() argument
465 return info_for_irq(irq)->cpu; in cpu_from_irq()
470 int irq = get_evtchn_to_irq(evtchn); in cpu_from_evtchn() local
473 if (irq != -1) in cpu_from_evtchn()
474 ret = cpu_from_irq(irq); in cpu_from_evtchn()
508 static bool pirq_check_eoi_map(unsigned irq) in pirq_check_eoi_map() argument
510 return test_bit(pirq_from_irq(irq), pirq_eoi_map); in pirq_check_eoi_map()
514 static bool pirq_needs_eoi_flag(unsigned irq) in pirq_needs_eoi_flag() argument
516 struct irq_info *info = info_for_irq(irq); in pirq_needs_eoi_flag()
525 int irq = get_evtchn_to_irq(evtchn); in bind_evtchn_to_cpu() local
526 struct irq_info *info = info_for_irq(irq); in bind_evtchn_to_cpu()
528 BUG_ON(irq == -1); in bind_evtchn_to_cpu()
531 cpumask_copy(irq_get_affinity_mask(irq), cpumask_of(cpu)); in bind_evtchn_to_cpu()
532 cpumask_copy(irq_get_effective_affinity_mask(irq), in bind_evtchn_to_cpu()
544 * notify_remote_via_irq - send event to remote end of event channel via irq
545 * @irq: irq of event channel to send event to
551 void notify_remote_via_irq(int irq) in notify_remote_via_irq() argument
553 evtchn_port_t evtchn = evtchn_from_irq(irq); in notify_remote_via_irq()
706 void xen_irq_lateeoi(unsigned int irq, unsigned int eoi_flags) in xen_irq_lateeoi() argument
713 info = info_for_irq(irq); in xen_irq_lateeoi()
722 static void xen_irq_init(unsigned irq) in xen_irq_init() argument
728 panic("Unable to allocate metadata for IRQ%d\n", irq); in xen_irq_init()
733 set_info_for_irq(irq, info); in xen_irq_init()
738 irq_set_status_flags(irq, IRQ_MOVE_PCNTXT); in xen_irq_init()
746 int i, irq = irq_alloc_descs(-1, 0, nvec, -1); in xen_allocate_irqs_dynamic() local
748 if (irq >= 0) { in xen_allocate_irqs_dynamic()
750 xen_irq_init(irq + i); in xen_allocate_irqs_dynamic()
753 return irq; in xen_allocate_irqs_dynamic()
764 int irq; in xen_allocate_irq_gsi() local
769 * all IRQs are dynamically allocated from the entire IRQ in xen_allocate_irq_gsi()
775 /* Legacy IRQ descriptors are already allocated by the arch. */ in xen_allocate_irq_gsi()
777 irq = gsi; in xen_allocate_irq_gsi()
779 irq = irq_alloc_desc_at(gsi, -1); in xen_allocate_irq_gsi()
781 xen_irq_init(irq); in xen_allocate_irq_gsi()
783 return irq; in xen_allocate_irq_gsi()
786 static void xen_free_irq(unsigned irq) in xen_free_irq() argument
788 struct irq_info *info = info_for_irq(irq); in xen_free_irq()
801 set_info_for_irq(irq, NULL); in xen_free_irq()
809 /* Legacy IRQ descriptors are managed by the arch. */ in xen_free_irq()
810 if (irq < nr_legacy_irqs()) in xen_free_irq()
813 irq_free_desc(irq); in xen_free_irq()
832 static void pirq_query_unmask(int irq) in pirq_query_unmask() argument
835 struct irq_info *info = info_for_irq(irq); in pirq_query_unmask()
839 irq_status.irq = pirq_from_irq(irq); in pirq_query_unmask()
850 struct irq_info *info = info_for_irq(data->irq); in eoi_pirq()
852 struct physdev_eoi eoi = { .irq = pirq_from_irq(data->irq) }; in eoi_pirq()
860 if (pirq_needs_eoi(data->irq)) { in eoi_pirq()
872 static unsigned int __startup_pirq(unsigned int irq) in __startup_pirq() argument
875 struct irq_info *info = info_for_irq(irq); in __startup_pirq()
876 evtchn_port_t evtchn = evtchn_from_irq(irq); in __startup_pirq()
884 bind_pirq.pirq = pirq_from_irq(irq); in __startup_pirq()
890 pr_warn("Failed to obtain physical IRQ %d\n", irq); in __startup_pirq()
895 pirq_query_unmask(irq); in __startup_pirq()
897 rc = set_evtchn_to_irq(evtchn, irq); in __startup_pirq()
911 eoi_pirq(irq_get_irq_data(irq)); in __startup_pirq()
916 pr_err("irq%d: Failed to set port to irq mapping (%d)\n", irq, rc); in __startup_pirq()
923 return __startup_pirq(data->irq); in startup_pirq()
928 unsigned int irq = data->irq; in shutdown_pirq() local
929 struct irq_info *info = info_for_irq(irq); in shutdown_pirq()
930 evtchn_port_t evtchn = evtchn_from_irq(irq); in shutdown_pirq()
961 return info->irq; in xen_irq_from_gsi()
968 static void __unbind_from_irq(unsigned int irq) in __unbind_from_irq() argument
970 evtchn_port_t evtchn = evtchn_from_irq(irq); in __unbind_from_irq()
971 struct irq_info *info = info_for_irq(irq); in __unbind_from_irq()
980 unsigned int cpu = cpu_from_irq(irq); in __unbind_from_irq()
985 switch (type_from_irq(irq)) { in __unbind_from_irq()
987 per_cpu(virq_to_irq, cpu)[virq_from_irq(irq)] = -1; in __unbind_from_irq()
990 per_cpu(ipi_to_irq, cpu)[ipi_from_irq(irq)] = -1; in __unbind_from_irq()
1004 xen_free_irq(irq); in __unbind_from_irq()
1009 * IRQ number returned here and the Xen pirq argument.
1011 * Note: We don't assign an event channel until the irq actually started
1012 * up. Return an existing irq if we've already got one for the gsi.
1020 int irq; in xen_bind_pirq_gsi_to_irq() local
1026 irq = xen_irq_from_gsi(gsi); in xen_bind_pirq_gsi_to_irq()
1027 if (irq != -1) { in xen_bind_pirq_gsi_to_irq()
1028 pr_info("%s: returning irq %d for gsi %u\n", in xen_bind_pirq_gsi_to_irq()
1029 __func__, irq, gsi); in xen_bind_pirq_gsi_to_irq()
1033 irq = xen_allocate_irq_gsi(gsi); in xen_bind_pirq_gsi_to_irq()
1034 if (irq < 0) in xen_bind_pirq_gsi_to_irq()
1037 irq_op.irq = irq; in xen_bind_pirq_gsi_to_irq()
1045 xen_free_irq(irq); in xen_bind_pirq_gsi_to_irq()
1046 irq = -ENOSPC; in xen_bind_pirq_gsi_to_irq()
1050 ret = xen_irq_info_pirq_setup(irq, 0, pirq, gsi, DOMID_SELF, in xen_bind_pirq_gsi_to_irq()
1053 __unbind_from_irq(irq); in xen_bind_pirq_gsi_to_irq()
1054 irq = ret; in xen_bind_pirq_gsi_to_irq()
1058 pirq_query_unmask(irq); in xen_bind_pirq_gsi_to_irq()
1075 irq_set_chip_and_handler_name(irq, &xen_pirq_chip, in xen_bind_pirq_gsi_to_irq()
1078 irq_set_chip_and_handler_name(irq, &xen_pirq_chip, in xen_bind_pirq_gsi_to_irq()
1084 return irq; in xen_bind_pirq_gsi_to_irq()
1105 int i, irq, ret; in xen_bind_pirq_msi_to_irq() local
1109 irq = xen_allocate_irqs_dynamic(nvec); in xen_bind_pirq_msi_to_irq()
1110 if (irq < 0) in xen_bind_pirq_msi_to_irq()
1114 irq_set_chip_and_handler_name(irq + i, &xen_pirq_chip, handle_edge_irq, name); in xen_bind_pirq_msi_to_irq()
1116 ret = xen_irq_info_pirq_setup(irq + i, 0, pirq + i, 0, domid, in xen_bind_pirq_msi_to_irq()
1122 ret = irq_set_msi_desc(irq, msidesc); in xen_bind_pirq_msi_to_irq()
1127 return irq; in xen_bind_pirq_msi_to_irq()
1130 __unbind_from_irq(irq + nvec); in xen_bind_pirq_msi_to_irq()
1136 int xen_destroy_irq(int irq) in xen_destroy_irq() argument
1139 struct irq_info *info = info_for_irq(irq); in xen_destroy_irq()
1161 pr_warn("unmap irq failed %d\n", rc); in xen_destroy_irq()
1166 xen_free_irq(irq); in xen_destroy_irq()
1175 int irq; in xen_irq_from_pirq() local
1184 irq = info->irq; in xen_irq_from_pirq()
1188 irq = -1; in xen_irq_from_pirq()
1192 return irq; in xen_irq_from_pirq()
1196 int xen_pirq_from_irq(unsigned irq) in xen_pirq_from_irq() argument
1198 return pirq_from_irq(irq); in xen_pirq_from_irq()
1205 int irq; in bind_evtchn_to_irq_chip() local
1213 irq = get_evtchn_to_irq(evtchn); in bind_evtchn_to_irq_chip()
1215 if (irq == -1) { in bind_evtchn_to_irq_chip()
1216 irq = xen_allocate_irq_dynamic(); in bind_evtchn_to_irq_chip()
1217 if (irq < 0) in bind_evtchn_to_irq_chip()
1220 irq_set_chip_and_handler_name(irq, chip, in bind_evtchn_to_irq_chip()
1223 ret = xen_irq_info_evtchn_setup(irq, evtchn, dev); in bind_evtchn_to_irq_chip()
1225 __unbind_from_irq(irq); in bind_evtchn_to_irq_chip()
1226 irq = ret; in bind_evtchn_to_irq_chip()
1238 struct irq_info *info = info_for_irq(irq); in bind_evtchn_to_irq_chip()
1245 return irq; in bind_evtchn_to_irq_chip()
1258 int ret, irq; in bind_ipi_to_irq() local
1262 irq = per_cpu(ipi_to_irq, cpu)[ipi]; in bind_ipi_to_irq()
1264 if (irq == -1) { in bind_ipi_to_irq()
1265 irq = xen_allocate_irq_dynamic(); in bind_ipi_to_irq()
1266 if (irq < 0) in bind_ipi_to_irq()
1269 irq_set_chip_and_handler_name(irq, &xen_percpu_chip, in bind_ipi_to_irq()
1278 ret = xen_irq_info_ipi_setup(cpu, irq, evtchn, ipi); in bind_ipi_to_irq()
1280 __unbind_from_irq(irq); in bind_ipi_to_irq()
1281 irq = ret; in bind_ipi_to_irq()
1290 struct irq_info *info = info_for_irq(irq); in bind_ipi_to_irq()
1296 return irq; in bind_ipi_to_irq()
1364 int irq, ret; in bind_virq_to_irq() local
1368 irq = per_cpu(virq_to_irq, cpu)[virq]; in bind_virq_to_irq()
1370 if (irq == -1) { in bind_virq_to_irq()
1371 irq = xen_allocate_irq_dynamic(); in bind_virq_to_irq()
1372 if (irq < 0) in bind_virq_to_irq()
1376 irq_set_chip_and_handler_name(irq, &xen_percpu_chip, in bind_virq_to_irq()
1379 irq_set_chip_and_handler_name(irq, &xen_dynamic_chip, in bind_virq_to_irq()
1394 ret = xen_irq_info_virq_setup(cpu, irq, evtchn, virq); in bind_virq_to_irq()
1396 __unbind_from_irq(irq); in bind_virq_to_irq()
1397 irq = ret; in bind_virq_to_irq()
1407 struct irq_info *info = info_for_irq(irq); in bind_virq_to_irq()
1414 return irq; in bind_virq_to_irq()
1417 static void unbind_from_irq(unsigned int irq) in unbind_from_irq() argument
1420 __unbind_from_irq(irq); in unbind_from_irq()
1430 int irq, retval; in bind_evtchn_to_irqhandler_chip() local
1432 irq = bind_evtchn_to_irq_chip(evtchn, chip, NULL); in bind_evtchn_to_irqhandler_chip()
1433 if (irq < 0) in bind_evtchn_to_irqhandler_chip()
1434 return irq; in bind_evtchn_to_irqhandler_chip()
1435 retval = request_irq(irq, handler, irqflags, devname, dev_id); in bind_evtchn_to_irqhandler_chip()
1437 unbind_from_irq(irq); in bind_evtchn_to_irqhandler_chip()
1441 return irq; in bind_evtchn_to_irqhandler_chip()
1471 int irq, retval; in bind_interdomain_evtchn_to_irqhandler_chip() local
1473 irq = bind_interdomain_evtchn_to_irq_chip(dev, remote_port, chip); in bind_interdomain_evtchn_to_irqhandler_chip()
1474 if (irq < 0) in bind_interdomain_evtchn_to_irqhandler_chip()
1475 return irq; in bind_interdomain_evtchn_to_irqhandler_chip()
1477 retval = request_irq(irq, handler, irqflags, devname, dev_id); in bind_interdomain_evtchn_to_irqhandler_chip()
1479 unbind_from_irq(irq); in bind_interdomain_evtchn_to_irqhandler_chip()
1483 return irq; in bind_interdomain_evtchn_to_irqhandler_chip()
1503 int irq, retval; in bind_virq_to_irqhandler() local
1505 irq = bind_virq_to_irq(virq, cpu, irqflags & IRQF_PERCPU); in bind_virq_to_irqhandler()
1506 if (irq < 0) in bind_virq_to_irqhandler()
1507 return irq; in bind_virq_to_irqhandler()
1508 retval = request_irq(irq, handler, irqflags, devname, dev_id); in bind_virq_to_irqhandler()
1510 unbind_from_irq(irq); in bind_virq_to_irqhandler()
1514 return irq; in bind_virq_to_irqhandler()
1525 int irq, retval; in bind_ipi_to_irqhandler() local
1527 irq = bind_ipi_to_irq(ipi, cpu); in bind_ipi_to_irqhandler()
1528 if (irq < 0) in bind_ipi_to_irqhandler()
1529 return irq; in bind_ipi_to_irqhandler()
1532 retval = request_irq(irq, handler, irqflags, devname, dev_id); in bind_ipi_to_irqhandler()
1534 unbind_from_irq(irq); in bind_ipi_to_irqhandler()
1538 return irq; in bind_ipi_to_irqhandler()
1541 void unbind_from_irqhandler(unsigned int irq, void *dev_id) in unbind_from_irqhandler() argument
1543 struct irq_info *info = info_for_irq(irq); in unbind_from_irqhandler()
1547 free_irq(irq, dev_id); in unbind_from_irqhandler()
1548 unbind_from_irq(irq); in unbind_from_irqhandler()
1554 * @irq:irq bound to an event channel.
1557 int xen_set_irq_priority(unsigned irq, unsigned priority) in xen_set_irq_priority() argument
1561 set_priority.port = evtchn_from_irq(irq); in xen_set_irq_priority()
1571 int irq = get_evtchn_to_irq(evtchn); in evtchn_make_refcounted() local
1574 if (irq == -1) in evtchn_make_refcounted()
1577 info = info_for_irq(irq); in evtchn_make_refcounted()
1592 int irq; in evtchn_get() local
1601 irq = get_evtchn_to_irq(evtchn); in evtchn_get()
1602 if (irq == -1) in evtchn_get()
1605 info = info_for_irq(irq); in evtchn_get()
1625 int irq = get_evtchn_to_irq(evtchn); in evtchn_put() local
1626 if (WARN_ON(irq == -1)) in evtchn_put()
1628 unbind_from_irq(irq); in evtchn_put()
1634 int irq; in xen_send_IPI_one() local
1645 irq = per_cpu(ipi_to_irq, cpu)[vector]; in xen_send_IPI_one()
1646 BUG_ON(irq < 0); in xen_send_IPI_one()
1647 notify_remote_via_irq(irq); in xen_send_IPI_one()
1658 int irq; in handle_irq_for_port() local
1662 irq = get_evtchn_to_irq(port); in handle_irq_for_port()
1663 if (irq == -1) in handle_irq_for_port()
1688 info = info_for_irq(irq); in handle_irq_for_port()
1702 generic_handle_irq(irq); in handle_irq_for_port()
1752 /* Rebind a new event channel to an existing irq. */
1753 void rebind_evtchn_irq(evtchn_port_t evtchn, int irq) in rebind_evtchn_irq() argument
1755 struct irq_info *info = info_for_irq(irq); in rebind_evtchn_irq()
1760 /* Make sure the irq is masked, since the new event channel in rebind_evtchn_irq()
1762 disable_irq(irq); in rebind_evtchn_irq()
1766 /* After resume the irq<->evtchn mappings are all cleared out */ in rebind_evtchn_irq()
1768 /* Expect irq to have been bound before, in rebind_evtchn_irq()
1772 (void)xen_irq_info_evtchn_setup(irq, evtchn, NULL); in rebind_evtchn_irq()
1779 enable_irq(irq); in rebind_evtchn_irq()
1851 ret = xen_rebind_evtchn_to_cpu(info_for_irq(data->irq), tcpu); in set_affinity_irq()
1860 struct irq_info *info = info_for_irq(data->irq); in enable_dynirq()
1869 struct irq_info *info = info_for_irq(data->irq); in disable_dynirq()
1878 struct irq_info *info = info_for_irq(data->irq); in ack_dynirq()
1893 struct irq_info *info = info_for_irq(data->irq); in lateeoi_ack_dynirq()
1909 struct irq_info *info = info_for_irq(data->irq); in lateeoi_mask_ack_dynirq()
1920 struct irq_info *info = info_for_irq(data->irq); in retrigger_dynirq()
1935 int pirq, rc, irq, gsi; in restore_pirqs() local
1945 irq = info->irq; in restore_pirqs()
1959 pr_warn("xen map irq failed gsi=%d irq=%d pirq=%d rc=%d\n", in restore_pirqs()
1960 gsi, irq, pirq, rc); in restore_pirqs()
1961 xen_free_irq(irq); in restore_pirqs()
1965 printk(KERN_DEBUG "xen: --> irq=%d, pirq=%d\n", irq, map_irq.pirq); in restore_pirqs()
1967 __startup_pirq(irq); in restore_pirqs()
1975 int virq, irq; in restore_cpu_virqs() local
1978 if ((irq = per_cpu(virq_to_irq, cpu)[virq]) == -1) in restore_cpu_virqs()
1981 BUG_ON(virq_from_irq(irq) != virq); in restore_cpu_virqs()
1992 (void)xen_irq_info_virq_setup(cpu, irq, evtchn, virq); in restore_cpu_virqs()
2002 int ipi, irq; in restore_cpu_ipis() local
2005 if ((irq = per_cpu(ipi_to_irq, cpu)[ipi]) == -1) in restore_cpu_ipis()
2008 BUG_ON(ipi_from_irq(irq) != ipi); in restore_cpu_ipis()
2018 (void)xen_irq_info_ipi_setup(cpu, irq, evtchn, ipi); in restore_cpu_ipis()
2024 /* Clear an irq's pending state, in preparation for polling on it */
2025 void xen_clear_irq_pending(int irq) in xen_clear_irq_pending() argument
2027 struct irq_info *info = info_for_irq(irq); in xen_clear_irq_pending()
2034 void xen_set_irq_pending(int irq) in xen_set_irq_pending() argument
2036 evtchn_port_t evtchn = evtchn_from_irq(irq); in xen_set_irq_pending()
2042 bool xen_test_irq_pending(int irq) in xen_test_irq_pending() argument
2044 evtchn_port_t evtchn = evtchn_from_irq(irq); in xen_test_irq_pending()
2053 /* Poll waiting for an irq to become pending with timeout. In the usual case,
2054 * the irq will be disabled so it won't deliver an interrupt. */
2055 void xen_poll_irq_timeout(int irq, u64 timeout) in xen_poll_irq_timeout() argument
2057 evtchn_port_t evtchn = evtchn_from_irq(irq); in xen_poll_irq_timeout()
2071 /* Poll waiting for an irq to become pending. In the usual case, the
2072 * irq will be disabled so it won't deliver an interrupt. */
2073 void xen_poll_irq(int irq) in xen_poll_irq() argument
2075 xen_poll_irq_timeout(irq, 0 /* no timeout */); in xen_poll_irq()
2078 /* Check whether the IRQ line is shared with other guests. */
2079 int xen_test_irq_shared(int irq) in xen_test_irq_shared() argument
2081 struct irq_info *info = info_for_irq(irq); in xen_test_irq_shared()
2087 irq_status.irq = info->u.pirq.pirq; in xen_test_irq_shared()
2103 /* No IRQ <-> event-channel mappings. */ in xen_irq_resume()