Lines Matching refs:trig

52 	struct iio_trigger *trig = to_iio_trigger(dev);  in name_show()  local
54 return sysfs_emit(buf, "%s\n", trig->name); in name_show()
115 int iio_trigger_set_immutable(struct iio_dev *indio_dev, struct iio_trigger *trig) in iio_trigger_set_immutable() argument
119 if (!indio_dev || !trig) in iio_trigger_set_immutable()
126 indio_dev->trig = iio_trigger_get(trig); in iio_trigger_set_immutable()
148 struct iio_trigger *trig = NULL, *iter; in iio_trigger_acquire_by_name() local
153 trig = iter; in iio_trigger_acquire_by_name()
154 iio_trigger_get(trig); in iio_trigger_acquire_by_name()
159 return trig; in iio_trigger_acquire_by_name()
164 struct iio_trigger *trig = container_of(work, struct iio_trigger, in iio_reenable_work_fn() local
171 trig->ops->reenable(trig); in iio_reenable_work_fn()
188 static void iio_trigger_notify_done_atomic(struct iio_trigger *trig) in iio_trigger_notify_done_atomic() argument
190 if (atomic_dec_and_test(&trig->use_count) && trig->ops && in iio_trigger_notify_done_atomic()
191 trig->ops->reenable) in iio_trigger_notify_done_atomic()
192 schedule_work(&trig->reenable_work); in iio_trigger_notify_done_atomic()
201 void iio_trigger_poll(struct iio_trigger *trig) in iio_trigger_poll() argument
205 if (!atomic_read(&trig->use_count)) { in iio_trigger_poll()
206 atomic_set(&trig->use_count, CONFIG_IIO_CONSUMERS_PER_TRIGGER); in iio_trigger_poll()
209 if (trig->subirqs[i].enabled) in iio_trigger_poll()
210 generic_handle_irq(trig->subirq_base + i); in iio_trigger_poll()
212 iio_trigger_notify_done_atomic(trig); in iio_trigger_poll()
232 void iio_trigger_poll_nested(struct iio_trigger *trig) in iio_trigger_poll_nested() argument
236 if (!atomic_read(&trig->use_count)) { in iio_trigger_poll_nested()
237 atomic_set(&trig->use_count, CONFIG_IIO_CONSUMERS_PER_TRIGGER); in iio_trigger_poll_nested()
240 if (trig->subirqs[i].enabled) in iio_trigger_poll_nested()
241 handle_nested_irq(trig->subirq_base + i); in iio_trigger_poll_nested()
243 iio_trigger_notify_done(trig); in iio_trigger_poll_nested()
249 void iio_trigger_notify_done(struct iio_trigger *trig) in iio_trigger_notify_done() argument
251 if (atomic_dec_and_test(&trig->use_count) && trig->ops && in iio_trigger_notify_done()
252 trig->ops->reenable) in iio_trigger_notify_done()
253 trig->ops->reenable(trig); in iio_trigger_notify_done()
258 static int iio_trigger_get_irq(struct iio_trigger *trig) in iio_trigger_get_irq() argument
262 mutex_lock(&trig->pool_lock); in iio_trigger_get_irq()
263 ret = bitmap_find_free_region(trig->pool, in iio_trigger_get_irq()
266 mutex_unlock(&trig->pool_lock); in iio_trigger_get_irq()
268 ret += trig->subirq_base; in iio_trigger_get_irq()
273 static void iio_trigger_put_irq(struct iio_trigger *trig, int irq) in iio_trigger_put_irq() argument
275 mutex_lock(&trig->pool_lock); in iio_trigger_put_irq()
276 clear_bit(irq - trig->subirq_base, trig->pool); in iio_trigger_put_irq()
277 mutex_unlock(&trig->pool_lock); in iio_trigger_put_irq()
287 int iio_trigger_attach_poll_func(struct iio_trigger *trig, in iio_trigger_attach_poll_func() argument
292 bitmap_empty(trig->pool, CONFIG_IIO_CONSUMERS_PER_TRIGGER); in iio_trigger_attach_poll_func()
299 pf->irq = iio_trigger_get_irq(trig); in iio_trigger_attach_poll_func()
302 trig->name, CONFIG_IIO_CONSUMERS_PER_TRIGGER); in iio_trigger_attach_poll_func()
314 if (trig->ops && trig->ops->set_trigger_state && notinuse) { in iio_trigger_attach_poll_func()
315 ret = trig->ops->set_trigger_state(trig, true); in iio_trigger_attach_poll_func()
325 if (iio_validate_own_trigger(pf->indio_dev, trig)) in iio_trigger_attach_poll_func()
326 trig->attached_own_device = true; in iio_trigger_attach_poll_func()
333 iio_trigger_put_irq(trig, pf->irq); in iio_trigger_attach_poll_func()
339 int iio_trigger_detach_poll_func(struct iio_trigger *trig, in iio_trigger_detach_poll_func() argument
344 bitmap_weight(trig->pool, CONFIG_IIO_CONSUMERS_PER_TRIGGER) == 1; in iio_trigger_detach_poll_func()
347 if (trig->ops && trig->ops->set_trigger_state && no_other_users) { in iio_trigger_detach_poll_func()
348 ret = trig->ops->set_trigger_state(trig, false); in iio_trigger_detach_poll_func()
352 if (pf->indio_dev->dev.parent == trig->dev.parent) in iio_trigger_detach_poll_func()
353 trig->attached_own_device = false; in iio_trigger_detach_poll_func()
354 iio_trigger_put_irq(trig, pf->irq); in iio_trigger_detach_poll_func()
425 if (indio_dev->trig) in current_trigger_show()
426 return sysfs_emit(buf, "%s\n", indio_dev->trig->name); in current_trigger_show()
450 struct iio_trigger *oldtrig = indio_dev->trig; in current_trigger_store()
451 struct iio_trigger *trig; in current_trigger_store() local
465 trig = iio_trigger_acquire_by_name(buf); in current_trigger_store()
466 if (oldtrig == trig) { in current_trigger_store()
471 if (trig && indio_dev->info->validate_trigger) { in current_trigger_store()
472 ret = indio_dev->info->validate_trigger(indio_dev, trig); in current_trigger_store()
477 if (trig && trig->ops && trig->ops->validate_device) { in current_trigger_store()
478 ret = trig->ops->validate_device(trig, indio_dev); in current_trigger_store()
483 indio_dev->trig = trig; in current_trigger_store()
491 if (indio_dev->trig) { in current_trigger_store()
493 iio_trigger_attach_poll_func(indio_dev->trig, in current_trigger_store()
500 if (trig) in current_trigger_store()
501 iio_trigger_put(trig); in current_trigger_store()
519 struct iio_trigger *trig = to_iio_trigger(device); in iio_trig_release() local
522 if (trig->subirq_base) { in iio_trig_release()
524 irq_modify_status(trig->subirq_base + i, in iio_trig_release()
527 irq_set_chip(trig->subirq_base + i, in iio_trig_release()
529 irq_set_handler(trig->subirq_base + i, in iio_trig_release()
533 irq_free_descs(trig->subirq_base, in iio_trig_release()
536 kfree(trig->name); in iio_trig_release()
537 kfree(trig); in iio_trig_release()
548 struct iio_trigger *trig = container_of(chip, struct iio_trigger, subirq_chip); in iio_trig_subirqmask() local
550 trig->subirqs[d->irq - trig->subirq_base].enabled = false; in iio_trig_subirqmask()
556 struct iio_trigger *trig = container_of(chip, struct iio_trigger, subirq_chip); in iio_trig_subirqunmask() local
558 trig->subirqs[d->irq - trig->subirq_base].enabled = true; in iio_trig_subirqunmask()
567 struct iio_trigger *trig; in viio_trigger_alloc() local
570 trig = kzalloc(sizeof(*trig), GFP_KERNEL); in viio_trigger_alloc()
571 if (!trig) in viio_trigger_alloc()
574 trig->dev.parent = parent; in viio_trigger_alloc()
575 trig->dev.type = &iio_trig_type; in viio_trigger_alloc()
576 trig->dev.bus = &iio_bus_type; in viio_trigger_alloc()
577 device_initialize(&trig->dev); in viio_trigger_alloc()
578 INIT_WORK(&trig->reenable_work, iio_reenable_work_fn); in viio_trigger_alloc()
580 mutex_init(&trig->pool_lock); in viio_trigger_alloc()
581 trig->subirq_base = irq_alloc_descs(-1, 0, in viio_trigger_alloc()
584 if (trig->subirq_base < 0) in viio_trigger_alloc()
587 trig->name = kvasprintf(GFP_KERNEL, fmt, vargs); in viio_trigger_alloc()
588 if (trig->name == NULL) in viio_trigger_alloc()
591 INIT_LIST_HEAD(&trig->list); in viio_trigger_alloc()
593 trig->owner = this_mod; in viio_trigger_alloc()
595 trig->subirq_chip.name = trig->name; in viio_trigger_alloc()
596 trig->subirq_chip.irq_mask = &iio_trig_subirqmask; in viio_trigger_alloc()
597 trig->subirq_chip.irq_unmask = &iio_trig_subirqunmask; in viio_trigger_alloc()
599 irq_set_chip(trig->subirq_base + i, &trig->subirq_chip); in viio_trigger_alloc()
600 irq_set_handler(trig->subirq_base + i, &handle_simple_irq); in viio_trigger_alloc()
601 irq_modify_status(trig->subirq_base + i, in viio_trigger_alloc()
605 return trig; in viio_trigger_alloc()
608 irq_free_descs(trig->subirq_base, CONFIG_IIO_CONSUMERS_PER_TRIGGER); in viio_trigger_alloc()
610 kfree(trig); in viio_trigger_alloc()
629 struct iio_trigger *trig; in __iio_trigger_alloc() local
633 trig = viio_trigger_alloc(parent, this_mod, fmt, vargs); in __iio_trigger_alloc()
636 return trig; in __iio_trigger_alloc()
640 void iio_trigger_free(struct iio_trigger *trig) in iio_trigger_free() argument
642 if (trig) in iio_trigger_free()
643 put_device(&trig->dev); in iio_trigger_free()
671 struct iio_trigger **ptr, *trig; in __devm_iio_trigger_alloc() local
681 trig = viio_trigger_alloc(parent, this_mod, fmt, vargs); in __devm_iio_trigger_alloc()
683 if (trig) { in __devm_iio_trigger_alloc()
684 *ptr = trig; in __devm_iio_trigger_alloc()
690 return trig; in __devm_iio_trigger_alloc()
727 return indio_dev->trig->attached_own_device; in iio_trigger_using_own()
743 int iio_validate_own_trigger(struct iio_dev *idev, struct iio_trigger *trig) in iio_validate_own_trigger() argument
745 if (idev->dev.parent != trig->dev.parent) in iio_validate_own_trigger()
763 int iio_trigger_validate_own_device(struct iio_trigger *trig, in iio_trigger_validate_own_device() argument
766 if (indio_dev->dev.parent != trig->dev.parent) in iio_trigger_validate_own_device()
781 if (indio_dev->trig) in iio_device_unregister_trigger_consumer()
782 iio_trigger_put(indio_dev->trig); in iio_device_unregister_trigger_consumer()