Lines Matching refs:desc
44 struct irq_desc *desc = irq_get_desc_lock(irq, &flags, 0); in irq_set_chip() local
46 if (!desc) in irq_set_chip()
52 desc->irq_data.chip = chip; in irq_set_chip()
53 irq_put_desc_unlock(desc, flags); in irq_set_chip()
71 struct irq_desc *desc = irq_get_desc_buslock(irq, &flags, IRQ_GET_DESC_CHECK_GLOBAL); in irq_set_irq_type() local
74 if (!desc) in irq_set_irq_type()
77 ret = __irq_set_trigger(desc, type); in irq_set_irq_type()
78 irq_put_desc_busunlock(desc, flags); in irq_set_irq_type()
93 struct irq_desc *desc = irq_get_desc_lock(irq, &flags, 0); in irq_set_handler_data() local
95 if (!desc) in irq_set_handler_data()
97 desc->irq_common_data.handler_data = data; in irq_set_handler_data()
98 irq_put_desc_unlock(desc, flags); in irq_set_handler_data()
115 …struct irq_desc *desc = irq_get_desc_lock(irq_base + irq_offset, &flags, IRQ_GET_DESC_CHECK_GLOBAL… in irq_set_msi_desc_off() local
117 if (!desc) in irq_set_msi_desc_off()
119 desc->irq_common_data.msi_desc = entry; in irq_set_msi_desc_off()
122 irq_put_desc_unlock(desc, flags); in irq_set_msi_desc_off()
148 struct irq_desc *desc = irq_get_desc_lock(irq, &flags, 0); in irq_set_chip_data() local
150 if (!desc) in irq_set_chip_data()
152 desc->irq_data.chip_data = data; in irq_set_chip_data()
153 irq_put_desc_unlock(desc, flags); in irq_set_chip_data()
160 struct irq_desc *desc = irq_to_desc(irq); in irq_get_irq_data() local
162 return desc ? &desc->irq_data : NULL; in irq_get_irq_data()
166 static void irq_state_clr_disabled(struct irq_desc *desc) in irq_state_clr_disabled() argument
168 irqd_clear(&desc->irq_data, IRQD_IRQ_DISABLED); in irq_state_clr_disabled()
171 static void irq_state_clr_masked(struct irq_desc *desc) in irq_state_clr_masked() argument
173 irqd_clear(&desc->irq_data, IRQD_IRQ_MASKED); in irq_state_clr_masked()
176 static void irq_state_clr_started(struct irq_desc *desc) in irq_state_clr_started() argument
178 irqd_clear(&desc->irq_data, IRQD_IRQ_STARTED); in irq_state_clr_started()
181 static void irq_state_set_started(struct irq_desc *desc) in irq_state_set_started() argument
183 irqd_set(&desc->irq_data, IRQD_IRQ_STARTED); in irq_state_set_started()
194 __irq_startup_managed(struct irq_desc *desc, struct cpumask *aff, bool force) in __irq_startup_managed() argument
196 struct irq_data *d = irq_desc_get_irq_data(desc); in __irq_startup_managed()
230 __irq_startup_managed(struct irq_desc *desc, struct cpumask *aff, bool force) in __irq_startup_managed() argument
236 static int __irq_startup(struct irq_desc *desc) in __irq_startup() argument
238 struct irq_data *d = irq_desc_get_irq_data(desc); in __irq_startup()
246 irq_state_clr_disabled(desc); in __irq_startup()
247 irq_state_clr_masked(desc); in __irq_startup()
249 irq_enable(desc); in __irq_startup()
251 irq_state_set_started(desc); in __irq_startup()
255 int irq_startup(struct irq_desc *desc, bool resend, bool force) in irq_startup() argument
257 struct irq_data *d = irq_desc_get_irq_data(desc); in irq_startup()
261 desc->depth = 0; in irq_startup()
264 irq_enable(desc); in irq_startup()
266 switch (__irq_startup_managed(desc, aff, force)) { in irq_startup()
268 ret = __irq_startup(desc); in irq_startup()
269 irq_setup_affinity(desc); in irq_startup()
273 ret = __irq_startup(desc); in irq_startup()
281 check_irq_resend(desc); in irq_startup()
286 int irq_activate(struct irq_desc *desc) in irq_activate() argument
288 struct irq_data *d = irq_desc_get_irq_data(desc); in irq_activate()
295 int irq_activate_and_startup(struct irq_desc *desc, bool resend) in irq_activate_and_startup() argument
297 if (WARN_ON(irq_activate(desc))) in irq_activate_and_startup()
299 return irq_startup(desc, resend, IRQ_START_FORCE); in irq_activate_and_startup()
302 static void __irq_disable(struct irq_desc *desc, bool mask);
304 void irq_shutdown(struct irq_desc *desc) in irq_shutdown() argument
306 if (irqd_is_started(&desc->irq_data)) { in irq_shutdown()
307 desc->depth = 1; in irq_shutdown()
308 if (desc->irq_data.chip->irq_shutdown) { in irq_shutdown()
309 desc->irq_data.chip->irq_shutdown(&desc->irq_data); in irq_shutdown()
310 irq_state_set_disabled(desc); in irq_shutdown()
311 irq_state_set_masked(desc); in irq_shutdown()
313 __irq_disable(desc, true); in irq_shutdown()
315 irq_state_clr_started(desc); in irq_shutdown()
323 irq_domain_deactivate_irq(&desc->irq_data); in irq_shutdown()
326 void irq_enable(struct irq_desc *desc) in irq_enable() argument
328 if (!irqd_irq_disabled(&desc->irq_data)) { in irq_enable()
329 unmask_irq(desc); in irq_enable()
331 irq_state_clr_disabled(desc); in irq_enable()
332 if (desc->irq_data.chip->irq_enable) { in irq_enable()
333 desc->irq_data.chip->irq_enable(&desc->irq_data); in irq_enable()
334 irq_state_clr_masked(desc); in irq_enable()
336 unmask_irq(desc); in irq_enable()
341 static void __irq_disable(struct irq_desc *desc, bool mask) in __irq_disable() argument
343 if (irqd_irq_disabled(&desc->irq_data)) { in __irq_disable()
345 mask_irq(desc); in __irq_disable()
347 irq_state_set_disabled(desc); in __irq_disable()
348 if (desc->irq_data.chip->irq_disable) { in __irq_disable()
349 desc->irq_data.chip->irq_disable(&desc->irq_data); in __irq_disable()
350 irq_state_set_masked(desc); in __irq_disable()
352 mask_irq(desc); in __irq_disable()
377 void irq_disable(struct irq_desc *desc) in irq_disable() argument
379 __irq_disable(desc, irq_settings_disable_unlazy(desc)); in irq_disable()
382 void irq_percpu_enable(struct irq_desc *desc, unsigned int cpu) in irq_percpu_enable() argument
384 if (desc->irq_data.chip->irq_enable) in irq_percpu_enable()
385 desc->irq_data.chip->irq_enable(&desc->irq_data); in irq_percpu_enable()
387 desc->irq_data.chip->irq_unmask(&desc->irq_data); in irq_percpu_enable()
388 cpumask_set_cpu(cpu, desc->percpu_enabled); in irq_percpu_enable()
391 void irq_percpu_disable(struct irq_desc *desc, unsigned int cpu) in irq_percpu_disable() argument
393 if (desc->irq_data.chip->irq_disable) in irq_percpu_disable()
394 desc->irq_data.chip->irq_disable(&desc->irq_data); in irq_percpu_disable()
396 desc->irq_data.chip->irq_mask(&desc->irq_data); in irq_percpu_disable()
397 cpumask_clear_cpu(cpu, desc->percpu_enabled); in irq_percpu_disable()
400 static inline void mask_ack_irq(struct irq_desc *desc) in mask_ack_irq() argument
402 if (desc->irq_data.chip->irq_mask_ack) { in mask_ack_irq()
403 desc->irq_data.chip->irq_mask_ack(&desc->irq_data); in mask_ack_irq()
404 irq_state_set_masked(desc); in mask_ack_irq()
406 mask_irq(desc); in mask_ack_irq()
407 if (desc->irq_data.chip->irq_ack) in mask_ack_irq()
408 desc->irq_data.chip->irq_ack(&desc->irq_data); in mask_ack_irq()
412 void mask_irq(struct irq_desc *desc) in mask_irq() argument
414 if (irqd_irq_masked(&desc->irq_data)) in mask_irq()
417 if (desc->irq_data.chip->irq_mask) { in mask_irq()
418 desc->irq_data.chip->irq_mask(&desc->irq_data); in mask_irq()
419 irq_state_set_masked(desc); in mask_irq()
423 void unmask_irq(struct irq_desc *desc) in unmask_irq() argument
425 if (!irqd_irq_masked(&desc->irq_data)) in unmask_irq()
428 if (desc->irq_data.chip->irq_unmask) { in unmask_irq()
429 desc->irq_data.chip->irq_unmask(&desc->irq_data); in unmask_irq()
430 irq_state_clr_masked(desc); in unmask_irq()
434 void unmask_threaded_irq(struct irq_desc *desc) in unmask_threaded_irq() argument
436 struct irq_chip *chip = desc->irq_data.chip; in unmask_threaded_irq()
439 chip->irq_eoi(&desc->irq_data); in unmask_threaded_irq()
441 unmask_irq(desc); in unmask_threaded_irq()
454 struct irq_desc *desc = irq_to_desc(irq); in handle_nested_irq() local
460 raw_spin_lock_irq(&desc->lock); in handle_nested_irq()
462 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_nested_irq()
464 action = desc->action; in handle_nested_irq()
465 if (unlikely(!action || irqd_irq_disabled(&desc->irq_data))) { in handle_nested_irq()
466 desc->istate |= IRQS_PENDING; in handle_nested_irq()
470 kstat_incr_irqs_this_cpu(desc); in handle_nested_irq()
471 irqd_set(&desc->irq_data, IRQD_IRQ_INPROGRESS); in handle_nested_irq()
472 raw_spin_unlock_irq(&desc->lock); in handle_nested_irq()
475 for_each_action_of_desc(desc, action) in handle_nested_irq()
479 note_interrupt(desc, action_ret); in handle_nested_irq()
481 raw_spin_lock_irq(&desc->lock); in handle_nested_irq()
482 irqd_clear(&desc->irq_data, IRQD_IRQ_INPROGRESS); in handle_nested_irq()
485 raw_spin_unlock_irq(&desc->lock); in handle_nested_irq()
489 static bool irq_check_poll(struct irq_desc *desc) in irq_check_poll() argument
491 if (!(desc->istate & IRQS_POLL_INPROGRESS)) in irq_check_poll()
493 return irq_wait_for_poll(desc); in irq_check_poll()
496 static bool irq_may_run(struct irq_desc *desc) in irq_may_run() argument
504 if (!irqd_has_set(&desc->irq_data, mask)) in irq_may_run()
512 if (irq_pm_check_wakeup(desc)) in irq_may_run()
518 return irq_check_poll(desc); in irq_may_run()
532 void handle_simple_irq(struct irq_desc *desc) in handle_simple_irq() argument
534 raw_spin_lock(&desc->lock); in handle_simple_irq()
536 if (!irq_may_run(desc)) in handle_simple_irq()
539 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_simple_irq()
541 if (unlikely(!desc->action || irqd_irq_disabled(&desc->irq_data))) { in handle_simple_irq()
542 desc->istate |= IRQS_PENDING; in handle_simple_irq()
546 kstat_incr_irqs_this_cpu(desc); in handle_simple_irq()
547 handle_irq_event(desc); in handle_simple_irq()
550 raw_spin_unlock(&desc->lock); in handle_simple_irq()
567 void handle_untracked_irq(struct irq_desc *desc) in handle_untracked_irq() argument
571 raw_spin_lock(&desc->lock); in handle_untracked_irq()
573 if (!irq_may_run(desc)) in handle_untracked_irq()
576 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_untracked_irq()
578 if (unlikely(!desc->action || irqd_irq_disabled(&desc->irq_data))) { in handle_untracked_irq()
579 desc->istate |= IRQS_PENDING; in handle_untracked_irq()
583 desc->istate &= ~IRQS_PENDING; in handle_untracked_irq()
584 irqd_set(&desc->irq_data, IRQD_IRQ_INPROGRESS); in handle_untracked_irq()
585 raw_spin_unlock(&desc->lock); in handle_untracked_irq()
587 __handle_irq_event_percpu(desc, &flags); in handle_untracked_irq()
589 raw_spin_lock(&desc->lock); in handle_untracked_irq()
590 irqd_clear(&desc->irq_data, IRQD_IRQ_INPROGRESS); in handle_untracked_irq()
593 raw_spin_unlock(&desc->lock); in handle_untracked_irq()
601 static void cond_unmask_irq(struct irq_desc *desc) in cond_unmask_irq() argument
610 if (!irqd_irq_disabled(&desc->irq_data) && in cond_unmask_irq()
611 irqd_irq_masked(&desc->irq_data) && !desc->threads_oneshot) in cond_unmask_irq()
612 unmask_irq(desc); in cond_unmask_irq()
624 void handle_level_irq(struct irq_desc *desc) in handle_level_irq() argument
626 raw_spin_lock(&desc->lock); in handle_level_irq()
627 mask_ack_irq(desc); in handle_level_irq()
629 if (!irq_may_run(desc)) in handle_level_irq()
632 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_level_irq()
638 if (unlikely(!desc->action || irqd_irq_disabled(&desc->irq_data))) { in handle_level_irq()
639 desc->istate |= IRQS_PENDING; in handle_level_irq()
643 kstat_incr_irqs_this_cpu(desc); in handle_level_irq()
644 handle_irq_event(desc); in handle_level_irq()
646 cond_unmask_irq(desc); in handle_level_irq()
649 raw_spin_unlock(&desc->lock); in handle_level_irq()
654 static inline void preflow_handler(struct irq_desc *desc) in preflow_handler() argument
656 if (desc->preflow_handler) in preflow_handler()
657 desc->preflow_handler(&desc->irq_data); in preflow_handler()
660 static inline void preflow_handler(struct irq_desc *desc) { } in preflow_handler() argument
663 static void cond_unmask_eoi_irq(struct irq_desc *desc, struct irq_chip *chip) in cond_unmask_eoi_irq() argument
665 if (!(desc->istate & IRQS_ONESHOT)) { in cond_unmask_eoi_irq()
666 chip->irq_eoi(&desc->irq_data); in cond_unmask_eoi_irq()
675 if (!irqd_irq_disabled(&desc->irq_data) && in cond_unmask_eoi_irq()
676 irqd_irq_masked(&desc->irq_data) && !desc->threads_oneshot) { in cond_unmask_eoi_irq()
677 chip->irq_eoi(&desc->irq_data); in cond_unmask_eoi_irq()
678 unmask_irq(desc); in cond_unmask_eoi_irq()
680 chip->irq_eoi(&desc->irq_data); in cond_unmask_eoi_irq()
693 void handle_fasteoi_irq(struct irq_desc *desc) in handle_fasteoi_irq() argument
695 struct irq_chip *chip = desc->irq_data.chip; in handle_fasteoi_irq()
697 raw_spin_lock(&desc->lock); in handle_fasteoi_irq()
699 if (!irq_may_run(desc)) in handle_fasteoi_irq()
702 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_fasteoi_irq()
708 if (unlikely(!desc->action || irqd_irq_disabled(&desc->irq_data))) { in handle_fasteoi_irq()
709 desc->istate |= IRQS_PENDING; in handle_fasteoi_irq()
710 mask_irq(desc); in handle_fasteoi_irq()
714 kstat_incr_irqs_this_cpu(desc); in handle_fasteoi_irq()
715 if (desc->istate & IRQS_ONESHOT) in handle_fasteoi_irq()
716 mask_irq(desc); in handle_fasteoi_irq()
718 preflow_handler(desc); in handle_fasteoi_irq()
719 handle_irq_event(desc); in handle_fasteoi_irq()
721 cond_unmask_eoi_irq(desc, chip); in handle_fasteoi_irq()
723 raw_spin_unlock(&desc->lock); in handle_fasteoi_irq()
727 chip->irq_eoi(&desc->irq_data); in handle_fasteoi_irq()
728 raw_spin_unlock(&desc->lock); in handle_fasteoi_irq()
747 void handle_edge_irq(struct irq_desc *desc) in handle_edge_irq() argument
749 raw_spin_lock(&desc->lock); in handle_edge_irq()
751 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_edge_irq()
753 if (!irq_may_run(desc)) { in handle_edge_irq()
754 desc->istate |= IRQS_PENDING; in handle_edge_irq()
755 mask_ack_irq(desc); in handle_edge_irq()
763 if (irqd_irq_disabled(&desc->irq_data) || !desc->action) { in handle_edge_irq()
764 desc->istate |= IRQS_PENDING; in handle_edge_irq()
765 mask_ack_irq(desc); in handle_edge_irq()
769 kstat_incr_irqs_this_cpu(desc); in handle_edge_irq()
772 desc->irq_data.chip->irq_ack(&desc->irq_data); in handle_edge_irq()
775 if (unlikely(!desc->action)) { in handle_edge_irq()
776 mask_irq(desc); in handle_edge_irq()
785 if (unlikely(desc->istate & IRQS_PENDING)) { in handle_edge_irq()
786 if (!irqd_irq_disabled(&desc->irq_data) && in handle_edge_irq()
787 irqd_irq_masked(&desc->irq_data)) in handle_edge_irq()
788 unmask_irq(desc); in handle_edge_irq()
791 handle_irq_event(desc); in handle_edge_irq()
793 } while ((desc->istate & IRQS_PENDING) && in handle_edge_irq()
794 !irqd_irq_disabled(&desc->irq_data)); in handle_edge_irq()
797 raw_spin_unlock(&desc->lock); in handle_edge_irq()
809 void handle_edge_eoi_irq(struct irq_desc *desc) in handle_edge_eoi_irq() argument
811 struct irq_chip *chip = irq_desc_get_chip(desc); in handle_edge_eoi_irq()
813 raw_spin_lock(&desc->lock); in handle_edge_eoi_irq()
815 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_edge_eoi_irq()
817 if (!irq_may_run(desc)) { in handle_edge_eoi_irq()
818 desc->istate |= IRQS_PENDING; in handle_edge_eoi_irq()
826 if (irqd_irq_disabled(&desc->irq_data) || !desc->action) { in handle_edge_eoi_irq()
827 desc->istate |= IRQS_PENDING; in handle_edge_eoi_irq()
831 kstat_incr_irqs_this_cpu(desc); in handle_edge_eoi_irq()
834 if (unlikely(!desc->action)) in handle_edge_eoi_irq()
837 handle_irq_event(desc); in handle_edge_eoi_irq()
839 } while ((desc->istate & IRQS_PENDING) && in handle_edge_eoi_irq()
840 !irqd_irq_disabled(&desc->irq_data)); in handle_edge_eoi_irq()
843 chip->irq_eoi(&desc->irq_data); in handle_edge_eoi_irq()
844 raw_spin_unlock(&desc->lock); in handle_edge_eoi_irq()
854 void handle_percpu_irq(struct irq_desc *desc) in handle_percpu_irq() argument
856 struct irq_chip *chip = irq_desc_get_chip(desc); in handle_percpu_irq()
858 kstat_incr_irqs_this_cpu(desc); in handle_percpu_irq()
861 chip->irq_ack(&desc->irq_data); in handle_percpu_irq()
863 handle_irq_event_percpu(desc); in handle_percpu_irq()
866 chip->irq_eoi(&desc->irq_data); in handle_percpu_irq()
880 void handle_percpu_devid_irq(struct irq_desc *desc) in handle_percpu_devid_irq() argument
882 struct irq_chip *chip = irq_desc_get_chip(desc); in handle_percpu_devid_irq()
883 struct irqaction *action = desc->action; in handle_percpu_devid_irq()
884 unsigned int irq = irq_desc_get_irq(desc); in handle_percpu_devid_irq()
887 kstat_incr_irqs_this_cpu(desc); in handle_percpu_devid_irq()
890 chip->irq_ack(&desc->irq_data); in handle_percpu_devid_irq()
898 bool enabled = cpumask_test_cpu(cpu, desc->percpu_enabled); in handle_percpu_devid_irq()
901 irq_percpu_disable(desc, cpu); in handle_percpu_devid_irq()
908 chip->irq_eoi(&desc->irq_data); in handle_percpu_devid_irq()
912 __irq_do_set_handler(struct irq_desc *desc, irq_flow_handler_t handle, in __irq_do_set_handler() argument
918 struct irq_data *irq_data = &desc->irq_data; in __irq_do_set_handler()
947 if (desc->irq_data.chip != &no_irq_chip) in __irq_do_set_handler()
948 mask_ack_irq(desc); in __irq_do_set_handler()
949 irq_state_set_disabled(desc); in __irq_do_set_handler()
951 desc->action = NULL; in __irq_do_set_handler()
952 desc->depth = 1; in __irq_do_set_handler()
954 desc->handle_irq = handle; in __irq_do_set_handler()
955 desc->name = name; in __irq_do_set_handler()
958 unsigned int type = irqd_get_trigger_type(&desc->irq_data); in __irq_do_set_handler()
969 __irq_set_trigger(desc, type); in __irq_do_set_handler()
970 desc->handle_irq = handle; in __irq_do_set_handler()
973 irq_settings_set_noprobe(desc); in __irq_do_set_handler()
974 irq_settings_set_norequest(desc); in __irq_do_set_handler()
975 irq_settings_set_nothread(desc); in __irq_do_set_handler()
976 desc->action = &chained_action; in __irq_do_set_handler()
977 irq_activate_and_startup(desc, IRQ_RESEND); in __irq_do_set_handler()
986 struct irq_desc *desc = irq_get_desc_buslock(irq, &flags, 0); in __irq_set_handler() local
988 if (!desc) in __irq_set_handler()
991 __irq_do_set_handler(desc, handle, is_chained, name); in __irq_set_handler()
992 irq_put_desc_busunlock(desc, flags); in __irq_set_handler()
1001 struct irq_desc *desc = irq_get_desc_buslock(irq, &flags, 0); in irq_set_chained_handler_and_data() local
1003 if (!desc) in irq_set_chained_handler_and_data()
1006 desc->irq_common_data.handler_data = data; in irq_set_chained_handler_and_data()
1007 __irq_do_set_handler(desc, handle, 1, NULL); in irq_set_chained_handler_and_data()
1009 irq_put_desc_busunlock(desc, flags); in irq_set_chained_handler_and_data()
1025 struct irq_desc *desc = irq_get_desc_lock(irq, &flags, 0); in irq_modify_status() local
1027 if (!desc) in irq_modify_status()
1034 WARN_ON_ONCE(!desc->depth && (set & _IRQ_NOAUTOEN)); in irq_modify_status()
1036 irq_settings_clr_and_set(desc, clr, set); in irq_modify_status()
1038 trigger = irqd_get_trigger_type(&desc->irq_data); in irq_modify_status()
1040 irqd_clear(&desc->irq_data, IRQD_NO_BALANCING | IRQD_PER_CPU | in irq_modify_status()
1042 if (irq_settings_has_no_balance_set(desc)) in irq_modify_status()
1043 irqd_set(&desc->irq_data, IRQD_NO_BALANCING); in irq_modify_status()
1044 if (irq_settings_is_per_cpu(desc)) in irq_modify_status()
1045 irqd_set(&desc->irq_data, IRQD_PER_CPU); in irq_modify_status()
1046 if (irq_settings_can_move_pcntxt(desc)) in irq_modify_status()
1047 irqd_set(&desc->irq_data, IRQD_MOVE_PCNTXT); in irq_modify_status()
1048 if (irq_settings_is_level(desc)) in irq_modify_status()
1049 irqd_set(&desc->irq_data, IRQD_LEVEL); in irq_modify_status()
1051 tmp = irq_settings_get_trigger_mask(desc); in irq_modify_status()
1055 irqd_set(&desc->irq_data, trigger); in irq_modify_status()
1057 irq_put_desc_unlock(desc, flags); in irq_modify_status()
1069 struct irq_desc *desc; in irq_cpu_online() local
1075 desc = irq_to_desc(irq); in irq_cpu_online()
1076 if (!desc) in irq_cpu_online()
1079 raw_spin_lock_irqsave(&desc->lock, flags); in irq_cpu_online()
1081 chip = irq_data_get_irq_chip(&desc->irq_data); in irq_cpu_online()
1084 !irqd_irq_disabled(&desc->irq_data))) in irq_cpu_online()
1085 chip->irq_cpu_online(&desc->irq_data); in irq_cpu_online()
1087 raw_spin_unlock_irqrestore(&desc->lock, flags); in irq_cpu_online()
1099 struct irq_desc *desc; in irq_cpu_offline() local
1105 desc = irq_to_desc(irq); in irq_cpu_offline()
1106 if (!desc) in irq_cpu_offline()
1109 raw_spin_lock_irqsave(&desc->lock, flags); in irq_cpu_offline()
1111 chip = irq_data_get_irq_chip(&desc->irq_data); in irq_cpu_offline()
1114 !irqd_irq_disabled(&desc->irq_data))) in irq_cpu_offline()
1115 chip->irq_cpu_offline(&desc->irq_data); in irq_cpu_offline()
1117 raw_spin_unlock_irqrestore(&desc->lock, flags); in irq_cpu_offline()
1134 void handle_fasteoi_ack_irq(struct irq_desc *desc) in handle_fasteoi_ack_irq() argument
1136 struct irq_chip *chip = desc->irq_data.chip; in handle_fasteoi_ack_irq()
1138 raw_spin_lock(&desc->lock); in handle_fasteoi_ack_irq()
1140 if (!irq_may_run(desc)) in handle_fasteoi_ack_irq()
1143 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_fasteoi_ack_irq()
1149 if (unlikely(!desc->action || irqd_irq_disabled(&desc->irq_data))) { in handle_fasteoi_ack_irq()
1150 desc->istate |= IRQS_PENDING; in handle_fasteoi_ack_irq()
1151 mask_irq(desc); in handle_fasteoi_ack_irq()
1155 kstat_incr_irqs_this_cpu(desc); in handle_fasteoi_ack_irq()
1156 if (desc->istate & IRQS_ONESHOT) in handle_fasteoi_ack_irq()
1157 mask_irq(desc); in handle_fasteoi_ack_irq()
1160 desc->irq_data.chip->irq_ack(&desc->irq_data); in handle_fasteoi_ack_irq()
1162 preflow_handler(desc); in handle_fasteoi_ack_irq()
1163 handle_irq_event(desc); in handle_fasteoi_ack_irq()
1165 cond_unmask_eoi_irq(desc, chip); in handle_fasteoi_ack_irq()
1167 raw_spin_unlock(&desc->lock); in handle_fasteoi_ack_irq()
1171 chip->irq_eoi(&desc->irq_data); in handle_fasteoi_ack_irq()
1172 raw_spin_unlock(&desc->lock); in handle_fasteoi_ack_irq()
1186 void handle_fasteoi_mask_irq(struct irq_desc *desc) in handle_fasteoi_mask_irq() argument
1188 struct irq_chip *chip = desc->irq_data.chip; in handle_fasteoi_mask_irq()
1190 raw_spin_lock(&desc->lock); in handle_fasteoi_mask_irq()
1191 mask_ack_irq(desc); in handle_fasteoi_mask_irq()
1193 if (!irq_may_run(desc)) in handle_fasteoi_mask_irq()
1196 desc->istate &= ~(IRQS_REPLAY | IRQS_WAITING); in handle_fasteoi_mask_irq()
1202 if (unlikely(!desc->action || irqd_irq_disabled(&desc->irq_data))) { in handle_fasteoi_mask_irq()
1203 desc->istate |= IRQS_PENDING; in handle_fasteoi_mask_irq()
1204 mask_irq(desc); in handle_fasteoi_mask_irq()
1208 kstat_incr_irqs_this_cpu(desc); in handle_fasteoi_mask_irq()
1209 if (desc->istate & IRQS_ONESHOT) in handle_fasteoi_mask_irq()
1210 mask_irq(desc); in handle_fasteoi_mask_irq()
1212 preflow_handler(desc); in handle_fasteoi_mask_irq()
1213 handle_irq_event(desc); in handle_fasteoi_mask_irq()
1215 cond_unmask_eoi_irq(desc, chip); in handle_fasteoi_mask_irq()
1217 raw_spin_unlock(&desc->lock); in handle_fasteoi_mask_irq()
1221 chip->irq_eoi(&desc->irq_data); in handle_fasteoi_mask_irq()
1222 raw_spin_unlock(&desc->lock); in handle_fasteoi_mask_irq()