Lines Matching refs:phy
30 struct phy *phy = *(struct phy **)res; in devm_phy_release() local
32 phy_put(dev, phy); in devm_phy_release()
44 struct phy *phy = *(struct phy **)res; in devm_phy_consume() local
46 phy_destroy(phy); in devm_phy_consume()
51 struct phy **phy = res; in devm_phy_match() local
53 return *phy == match_data; in devm_phy_match()
64 int phy_create_lookup(struct phy *phy, const char *con_id, const char *dev_id) in phy_create_lookup() argument
68 if (!phy || !dev_id || !con_id) in phy_create_lookup()
77 pl->phy = phy; in phy_create_lookup()
96 void phy_remove_lookup(struct phy *phy, const char *con_id, const char *dev_id) in phy_remove_lookup() argument
100 if (!phy || !dev_id || !con_id) in phy_remove_lookup()
105 if (pl->phy == phy && !strcmp(pl->dev_id, dev_id) && in phy_remove_lookup()
115 static struct phy *phy_find(struct device *dev, const char *con_id) in phy_find()
128 return pl ? pl->phy : ERR_PTR(-ENODEV); in phy_find()
148 int phy_pm_runtime_get(struct phy *phy) in phy_pm_runtime_get() argument
152 if (!phy) in phy_pm_runtime_get()
155 if (!pm_runtime_enabled(&phy->dev)) in phy_pm_runtime_get()
158 ret = pm_runtime_get(&phy->dev); in phy_pm_runtime_get()
160 pm_runtime_put_noidle(&phy->dev); in phy_pm_runtime_get()
166 int phy_pm_runtime_get_sync(struct phy *phy) in phy_pm_runtime_get_sync() argument
170 if (!phy) in phy_pm_runtime_get_sync()
173 if (!pm_runtime_enabled(&phy->dev)) in phy_pm_runtime_get_sync()
176 ret = pm_runtime_get_sync(&phy->dev); in phy_pm_runtime_get_sync()
178 pm_runtime_put_sync(&phy->dev); in phy_pm_runtime_get_sync()
184 int phy_pm_runtime_put(struct phy *phy) in phy_pm_runtime_put() argument
186 if (!phy) in phy_pm_runtime_put()
189 if (!pm_runtime_enabled(&phy->dev)) in phy_pm_runtime_put()
192 return pm_runtime_put(&phy->dev); in phy_pm_runtime_put()
196 int phy_pm_runtime_put_sync(struct phy *phy) in phy_pm_runtime_put_sync() argument
198 if (!phy) in phy_pm_runtime_put_sync()
201 if (!pm_runtime_enabled(&phy->dev)) in phy_pm_runtime_put_sync()
204 return pm_runtime_put_sync(&phy->dev); in phy_pm_runtime_put_sync()
208 void phy_pm_runtime_allow(struct phy *phy) in phy_pm_runtime_allow() argument
210 if (!phy) in phy_pm_runtime_allow()
213 if (!pm_runtime_enabled(&phy->dev)) in phy_pm_runtime_allow()
216 pm_runtime_allow(&phy->dev); in phy_pm_runtime_allow()
220 void phy_pm_runtime_forbid(struct phy *phy) in phy_pm_runtime_forbid() argument
222 if (!phy) in phy_pm_runtime_forbid()
225 if (!pm_runtime_enabled(&phy->dev)) in phy_pm_runtime_forbid()
228 pm_runtime_forbid(&phy->dev); in phy_pm_runtime_forbid()
243 int phy_init(struct phy *phy) in phy_init() argument
247 if (!phy) in phy_init()
250 ret = phy_pm_runtime_get_sync(phy); in phy_init()
255 mutex_lock(&phy->mutex); in phy_init()
256 if (phy->power_count > phy->init_count) in phy_init()
257 dev_warn(&phy->dev, "phy_power_on was called before phy_init\n"); in phy_init()
259 if (phy->init_count == 0 && phy->ops->init) { in phy_init()
260 ret = phy->ops->init(phy); in phy_init()
262 dev_err(&phy->dev, "phy init failed --> %d\n", ret); in phy_init()
266 ++phy->init_count; in phy_init()
269 mutex_unlock(&phy->mutex); in phy_init()
270 phy_pm_runtime_put(phy); in phy_init()
283 int phy_exit(struct phy *phy) in phy_exit() argument
287 if (!phy) in phy_exit()
290 ret = phy_pm_runtime_get_sync(phy); in phy_exit()
295 mutex_lock(&phy->mutex); in phy_exit()
296 if (phy->init_count == 1 && phy->ops->exit) { in phy_exit()
297 ret = phy->ops->exit(phy); in phy_exit()
299 dev_err(&phy->dev, "phy exit failed --> %d\n", ret); in phy_exit()
303 --phy->init_count; in phy_exit()
306 mutex_unlock(&phy->mutex); in phy_exit()
307 phy_pm_runtime_put(phy); in phy_exit()
320 int phy_power_on(struct phy *phy) in phy_power_on() argument
324 if (!phy) in phy_power_on()
327 if (phy->pwr) { in phy_power_on()
328 ret = regulator_enable(phy->pwr); in phy_power_on()
333 ret = phy_pm_runtime_get_sync(phy); in phy_power_on()
339 mutex_lock(&phy->mutex); in phy_power_on()
340 if (phy->power_count == 0 && phy->ops->power_on) { in phy_power_on()
341 ret = phy->ops->power_on(phy); in phy_power_on()
343 dev_err(&phy->dev, "phy poweron failed --> %d\n", ret); in phy_power_on()
347 ++phy->power_count; in phy_power_on()
348 mutex_unlock(&phy->mutex); in phy_power_on()
352 mutex_unlock(&phy->mutex); in phy_power_on()
353 phy_pm_runtime_put_sync(phy); in phy_power_on()
355 if (phy->pwr) in phy_power_on()
356 regulator_disable(phy->pwr); in phy_power_on()
370 int phy_power_off(struct phy *phy) in phy_power_off() argument
374 if (!phy) in phy_power_off()
377 mutex_lock(&phy->mutex); in phy_power_off()
378 if (phy->power_count == 1 && phy->ops->power_off) { in phy_power_off()
379 ret = phy->ops->power_off(phy); in phy_power_off()
381 dev_err(&phy->dev, "phy poweroff failed --> %d\n", ret); in phy_power_off()
382 mutex_unlock(&phy->mutex); in phy_power_off()
386 --phy->power_count; in phy_power_off()
387 mutex_unlock(&phy->mutex); in phy_power_off()
388 phy_pm_runtime_put(phy); in phy_power_off()
390 if (phy->pwr) in phy_power_off()
391 regulator_disable(phy->pwr); in phy_power_off()
397 int phy_set_mode_ext(struct phy *phy, enum phy_mode mode, int submode) in phy_set_mode_ext() argument
401 if (!phy || !phy->ops->set_mode) in phy_set_mode_ext()
404 mutex_lock(&phy->mutex); in phy_set_mode_ext()
405 ret = phy->ops->set_mode(phy, mode, submode); in phy_set_mode_ext()
407 phy->attrs.mode = mode; in phy_set_mode_ext()
408 mutex_unlock(&phy->mutex); in phy_set_mode_ext()
414 int phy_set_media(struct phy *phy, enum phy_media media) in phy_set_media() argument
418 if (!phy || !phy->ops->set_media) in phy_set_media()
421 mutex_lock(&phy->mutex); in phy_set_media()
422 ret = phy->ops->set_media(phy, media); in phy_set_media()
423 mutex_unlock(&phy->mutex); in phy_set_media()
429 int phy_set_speed(struct phy *phy, int speed) in phy_set_speed() argument
433 if (!phy || !phy->ops->set_speed) in phy_set_speed()
436 mutex_lock(&phy->mutex); in phy_set_speed()
437 ret = phy->ops->set_speed(phy, speed); in phy_set_speed()
438 mutex_unlock(&phy->mutex); in phy_set_speed()
444 int phy_reset(struct phy *phy) in phy_reset() argument
448 if (!phy || !phy->ops->reset) in phy_reset()
451 ret = phy_pm_runtime_get_sync(phy); in phy_reset()
455 mutex_lock(&phy->mutex); in phy_reset()
456 ret = phy->ops->reset(phy); in phy_reset()
457 mutex_unlock(&phy->mutex); in phy_reset()
459 phy_pm_runtime_put(phy); in phy_reset()
475 int phy_calibrate(struct phy *phy) in phy_calibrate() argument
479 if (!phy || !phy->ops->calibrate) in phy_calibrate()
482 mutex_lock(&phy->mutex); in phy_calibrate()
483 ret = phy->ops->calibrate(phy); in phy_calibrate()
484 mutex_unlock(&phy->mutex); in phy_calibrate()
501 int phy_configure(struct phy *phy, union phy_configure_opts *opts) in phy_configure() argument
505 if (!phy) in phy_configure()
508 if (!phy->ops->configure) in phy_configure()
511 mutex_lock(&phy->mutex); in phy_configure()
512 ret = phy->ops->configure(phy, opts); in phy_configure()
513 mutex_unlock(&phy->mutex); in phy_configure()
535 int phy_validate(struct phy *phy, enum phy_mode mode, int submode, in phy_validate() argument
540 if (!phy) in phy_validate()
543 if (!phy->ops->validate) in phy_validate()
546 mutex_lock(&phy->mutex); in phy_validate()
547 ret = phy->ops->validate(phy, mode, submode, opts); in phy_validate()
548 mutex_unlock(&phy->mutex); in phy_validate()
565 static struct phy *_of_phy_get(struct device_node *np, int index) in _of_phy_get()
569 struct phy *phy = NULL; in _of_phy_get() local
584 phy = ERR_PTR(-EPROBE_DEFER); in _of_phy_get()
590 phy = ERR_PTR(-ENODEV); in _of_phy_get()
594 phy = phy_provider->of_xlate(phy_provider->dev, &args); in _of_phy_get()
603 return phy; in _of_phy_get()
615 struct phy *of_phy_get(struct device_node *np, const char *con_id) in of_phy_get()
617 struct phy *phy = NULL; in of_phy_get() local
623 phy = _of_phy_get(np, index); in of_phy_get()
624 if (IS_ERR(phy)) in of_phy_get()
625 return phy; in of_phy_get()
627 if (!try_module_get(phy->ops->owner)) in of_phy_get()
630 get_device(&phy->dev); in of_phy_get()
632 return phy; in of_phy_get()
642 void of_phy_put(struct phy *phy) in of_phy_put() argument
644 if (!phy || IS_ERR(phy)) in of_phy_put()
647 mutex_lock(&phy->mutex); in of_phy_put()
648 if (phy->ops->release) in of_phy_put()
649 phy->ops->release(phy); in of_phy_put()
650 mutex_unlock(&phy->mutex); in of_phy_put()
652 module_put(phy->ops->owner); in of_phy_put()
653 put_device(&phy->dev); in of_phy_put()
664 void phy_put(struct device *dev, struct phy *phy) in phy_put() argument
666 device_link_remove(dev, &phy->dev); in phy_put()
667 of_phy_put(phy); in phy_put()
679 void devm_phy_put(struct device *dev, struct phy *phy) in devm_phy_put() argument
683 if (!phy) in devm_phy_put()
686 r = devres_destroy(dev, devm_phy_release, devm_phy_match, phy); in devm_phy_put()
701 struct phy *of_phy_simple_xlate(struct device *dev, struct of_phandle_args in of_phy_simple_xlate()
704 struct phy *phy; in of_phy_simple_xlate() local
709 phy = to_phy(dev); in of_phy_simple_xlate()
710 if (args->np != phy->dev.of_node) in of_phy_simple_xlate()
714 return phy; in of_phy_simple_xlate()
732 struct phy *phy_get(struct device *dev, const char *string) in phy_get()
735 struct phy *phy; in phy_get() local
744 phy = _of_phy_get(dev->of_node, index); in phy_get()
750 phy = phy_find(dev, string); in phy_get()
752 if (IS_ERR(phy)) in phy_get()
753 return phy; in phy_get()
755 if (!try_module_get(phy->ops->owner)) in phy_get()
758 get_device(&phy->dev); in phy_get()
760 link = device_link_add(dev, &phy->dev, DL_FLAG_STATELESS); in phy_get()
763 dev_name(phy->dev.parent)); in phy_get()
765 return phy; in phy_get()
779 struct phy *phy_optional_get(struct device *dev, const char *string) in phy_optional_get()
781 struct phy *phy = phy_get(dev, string); in phy_optional_get() local
783 if (PTR_ERR(phy) == -ENODEV) in phy_optional_get()
784 phy = NULL; in phy_optional_get()
786 return phy; in phy_optional_get()
800 struct phy *devm_phy_get(struct device *dev, const char *string) in devm_phy_get()
802 struct phy **ptr, *phy; in devm_phy_get() local
808 phy = phy_get(dev, string); in devm_phy_get()
809 if (!IS_ERR(phy)) { in devm_phy_get()
810 *ptr = phy; in devm_phy_get()
816 return phy; in devm_phy_get()
833 struct phy *devm_phy_optional_get(struct device *dev, const char *string) in devm_phy_optional_get()
835 struct phy *phy = devm_phy_get(dev, string); in devm_phy_optional_get() local
837 if (PTR_ERR(phy) == -ENODEV) in devm_phy_optional_get()
838 phy = NULL; in devm_phy_optional_get()
840 return phy; in devm_phy_optional_get()
854 struct phy *devm_of_phy_get(struct device *dev, struct device_node *np, in devm_of_phy_get()
857 struct phy **ptr, *phy; in devm_of_phy_get() local
864 phy = of_phy_get(np, con_id); in devm_of_phy_get()
865 if (!IS_ERR(phy)) { in devm_of_phy_get()
866 *ptr = phy; in devm_of_phy_get()
870 return phy; in devm_of_phy_get()
873 link = device_link_add(dev, &phy->dev, DL_FLAG_STATELESS); in devm_of_phy_get()
876 dev_name(phy->dev.parent)); in devm_of_phy_get()
878 return phy; in devm_of_phy_get()
894 struct phy *devm_of_phy_get_by_index(struct device *dev, struct device_node *np, in devm_of_phy_get_by_index()
897 struct phy **ptr, *phy; in devm_of_phy_get_by_index() local
904 phy = _of_phy_get(np, index); in devm_of_phy_get_by_index()
905 if (IS_ERR(phy)) { in devm_of_phy_get_by_index()
907 return phy; in devm_of_phy_get_by_index()
910 if (!try_module_get(phy->ops->owner)) { in devm_of_phy_get_by_index()
915 get_device(&phy->dev); in devm_of_phy_get_by_index()
917 *ptr = phy; in devm_of_phy_get_by_index()
920 link = device_link_add(dev, &phy->dev, DL_FLAG_STATELESS); in devm_of_phy_get_by_index()
923 dev_name(phy->dev.parent)); in devm_of_phy_get_by_index()
925 return phy; in devm_of_phy_get_by_index()
937 struct phy *phy_create(struct device *dev, struct device_node *node, in phy_create()
942 struct phy *phy; in phy_create() local
947 phy = kzalloc(sizeof(*phy), GFP_KERNEL); in phy_create()
948 if (!phy) in phy_create()
958 device_initialize(&phy->dev); in phy_create()
959 mutex_init(&phy->mutex); in phy_create()
961 phy->dev.class = phy_class; in phy_create()
962 phy->dev.parent = dev; in phy_create()
963 phy->dev.of_node = node ?: dev->of_node; in phy_create()
964 phy->id = id; in phy_create()
965 phy->ops = ops; in phy_create()
967 ret = dev_set_name(&phy->dev, "phy-%s.%d", dev_name(dev), id); in phy_create()
972 phy->pwr = regulator_get_optional(&phy->dev, "phy"); in phy_create()
973 if (IS_ERR(phy->pwr)) { in phy_create()
974 ret = PTR_ERR(phy->pwr); in phy_create()
978 phy->pwr = NULL; in phy_create()
981 ret = device_add(&phy->dev); in phy_create()
986 pm_runtime_enable(&phy->dev); in phy_create()
987 pm_runtime_no_callbacks(&phy->dev); in phy_create()
990 return phy; in phy_create()
993 put_device(&phy->dev); /* calls phy_release() which frees resources */ in phy_create()
997 kfree(phy); in phy_create()
1013 struct phy *devm_phy_create(struct device *dev, struct device_node *node, in devm_phy_create()
1016 struct phy **ptr, *phy; in devm_phy_create() local
1022 phy = phy_create(dev, node, ops); in devm_phy_create()
1023 if (!IS_ERR(phy)) { in devm_phy_create()
1024 *ptr = phy; in devm_phy_create()
1030 return phy; in devm_phy_create()
1040 void phy_destroy(struct phy *phy) in phy_destroy() argument
1042 pm_runtime_disable(&phy->dev); in phy_destroy()
1043 device_unregister(&phy->dev); in phy_destroy()
1055 void devm_phy_destroy(struct device *dev, struct phy *phy) in devm_phy_destroy() argument
1059 r = devres_destroy(dev, devm_phy_consume, devm_phy_match, phy); in devm_phy_destroy()
1084 struct phy * (*of_xlate)(struct device *dev, in __of_phy_provider_register()
1147 struct phy * (*of_xlate)(struct device *dev, in __devm_of_phy_provider_register()
1216 struct phy *phy; in phy_release() local
1218 phy = to_phy(dev); in phy_release()
1220 regulator_put(phy->pwr); in phy_release()
1221 ida_simple_remove(&phy_ida, phy->id); in phy_release()
1222 kfree(phy); in phy_release()