Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
at v6.5 1261 lines 31 kB view raw
1// SPDX-License-Identifier: GPL-2.0-or-later 2/* 3 * phy-core.c -- Generic Phy framework. 4 * 5 * Copyright (C) 2013 Texas Instruments Incorporated - http://www.ti.com 6 * 7 * Author: Kishon Vijay Abraham I <kishon@ti.com> 8 */ 9 10#include <linux/kernel.h> 11#include <linux/export.h> 12#include <linux/module.h> 13#include <linux/err.h> 14#include <linux/debugfs.h> 15#include <linux/device.h> 16#include <linux/slab.h> 17#include <linux/of.h> 18#include <linux/phy/phy.h> 19#include <linux/idr.h> 20#include <linux/pm_runtime.h> 21#include <linux/regulator/consumer.h> 22 23static struct class *phy_class; 24static struct dentry *phy_debugfs_root; 25static DEFINE_MUTEX(phy_provider_mutex); 26static LIST_HEAD(phy_provider_list); 27static LIST_HEAD(phys); 28static DEFINE_IDA(phy_ida); 29 30static void devm_phy_release(struct device *dev, void *res) 31{ 32 struct phy *phy = *(struct phy **)res; 33 34 phy_put(dev, phy); 35} 36 37static void devm_phy_provider_release(struct device *dev, void *res) 38{ 39 struct phy_provider *phy_provider = *(struct phy_provider **)res; 40 41 of_phy_provider_unregister(phy_provider); 42} 43 44static void devm_phy_consume(struct device *dev, void *res) 45{ 46 struct phy *phy = *(struct phy **)res; 47 48 phy_destroy(phy); 49} 50 51static int devm_phy_match(struct device *dev, void *res, void *match_data) 52{ 53 struct phy **phy = res; 54 55 return *phy == match_data; 56} 57 58/** 59 * phy_create_lookup() - allocate and register PHY/device association 60 * @phy: the phy of the association 61 * @con_id: connection ID string on device 62 * @dev_id: the device of the association 63 * 64 * Creates and registers phy_lookup entry. 65 */ 66int phy_create_lookup(struct phy *phy, const char *con_id, const char *dev_id) 67{ 68 struct phy_lookup *pl; 69 70 if (!phy || !dev_id || !con_id) 71 return -EINVAL; 72 73 pl = kzalloc(sizeof(*pl), GFP_KERNEL); 74 if (!pl) 75 return -ENOMEM; 76 77 pl->dev_id = dev_id; 78 pl->con_id = con_id; 79 pl->phy = phy; 80 81 mutex_lock(&phy_provider_mutex); 82 list_add_tail(&pl->node, &phys); 83 mutex_unlock(&phy_provider_mutex); 84 85 return 0; 86} 87EXPORT_SYMBOL_GPL(phy_create_lookup); 88 89/** 90 * phy_remove_lookup() - find and remove PHY/device association 91 * @phy: the phy of the association 92 * @con_id: connection ID string on device 93 * @dev_id: the device of the association 94 * 95 * Finds and unregisters phy_lookup entry that was created with 96 * phy_create_lookup(). 97 */ 98void phy_remove_lookup(struct phy *phy, const char *con_id, const char *dev_id) 99{ 100 struct phy_lookup *pl; 101 102 if (!phy || !dev_id || !con_id) 103 return; 104 105 mutex_lock(&phy_provider_mutex); 106 list_for_each_entry(pl, &phys, node) 107 if (pl->phy == phy && !strcmp(pl->dev_id, dev_id) && 108 !strcmp(pl->con_id, con_id)) { 109 list_del(&pl->node); 110 kfree(pl); 111 break; 112 } 113 mutex_unlock(&phy_provider_mutex); 114} 115EXPORT_SYMBOL_GPL(phy_remove_lookup); 116 117static struct phy *phy_find(struct device *dev, const char *con_id) 118{ 119 const char *dev_id = dev_name(dev); 120 struct phy_lookup *p, *pl = NULL; 121 122 mutex_lock(&phy_provider_mutex); 123 list_for_each_entry(p, &phys, node) 124 if (!strcmp(p->dev_id, dev_id) && !strcmp(p->con_id, con_id)) { 125 pl = p; 126 break; 127 } 128 mutex_unlock(&phy_provider_mutex); 129 130 return pl ? pl->phy : ERR_PTR(-ENODEV); 131} 132 133static struct phy_provider *of_phy_provider_lookup(struct device_node *node) 134{ 135 struct phy_provider *phy_provider; 136 struct device_node *child; 137 138 list_for_each_entry(phy_provider, &phy_provider_list, list) { 139 if (phy_provider->dev->of_node == node) 140 return phy_provider; 141 142 for_each_child_of_node(phy_provider->children, child) 143 if (child == node) 144 return phy_provider; 145 } 146 147 return ERR_PTR(-EPROBE_DEFER); 148} 149 150int phy_pm_runtime_get(struct phy *phy) 151{ 152 int ret; 153 154 if (!phy) 155 return 0; 156 157 if (!pm_runtime_enabled(&phy->dev)) 158 return -ENOTSUPP; 159 160 ret = pm_runtime_get(&phy->dev); 161 if (ret < 0 && ret != -EINPROGRESS) 162 pm_runtime_put_noidle(&phy->dev); 163 164 return ret; 165} 166EXPORT_SYMBOL_GPL(phy_pm_runtime_get); 167 168int phy_pm_runtime_get_sync(struct phy *phy) 169{ 170 int ret; 171 172 if (!phy) 173 return 0; 174 175 if (!pm_runtime_enabled(&phy->dev)) 176 return -ENOTSUPP; 177 178 ret = pm_runtime_get_sync(&phy->dev); 179 if (ret < 0) 180 pm_runtime_put_sync(&phy->dev); 181 182 return ret; 183} 184EXPORT_SYMBOL_GPL(phy_pm_runtime_get_sync); 185 186int phy_pm_runtime_put(struct phy *phy) 187{ 188 if (!phy) 189 return 0; 190 191 if (!pm_runtime_enabled(&phy->dev)) 192 return -ENOTSUPP; 193 194 return pm_runtime_put(&phy->dev); 195} 196EXPORT_SYMBOL_GPL(phy_pm_runtime_put); 197 198int phy_pm_runtime_put_sync(struct phy *phy) 199{ 200 if (!phy) 201 return 0; 202 203 if (!pm_runtime_enabled(&phy->dev)) 204 return -ENOTSUPP; 205 206 return pm_runtime_put_sync(&phy->dev); 207} 208EXPORT_SYMBOL_GPL(phy_pm_runtime_put_sync); 209 210void phy_pm_runtime_allow(struct phy *phy) 211{ 212 if (!phy) 213 return; 214 215 if (!pm_runtime_enabled(&phy->dev)) 216 return; 217 218 pm_runtime_allow(&phy->dev); 219} 220EXPORT_SYMBOL_GPL(phy_pm_runtime_allow); 221 222void phy_pm_runtime_forbid(struct phy *phy) 223{ 224 if (!phy) 225 return; 226 227 if (!pm_runtime_enabled(&phy->dev)) 228 return; 229 230 pm_runtime_forbid(&phy->dev); 231} 232EXPORT_SYMBOL_GPL(phy_pm_runtime_forbid); 233 234/** 235 * phy_init - phy internal initialization before phy operation 236 * @phy: the phy returned by phy_get() 237 * 238 * Used to allow phy's driver to perform phy internal initialization, 239 * such as PLL block powering, clock initialization or anything that's 240 * is required by the phy to perform the start of operation. 241 * Must be called before phy_power_on(). 242 * 243 * Return: %0 if successful, a negative error code otherwise 244 */ 245int phy_init(struct phy *phy) 246{ 247 int ret; 248 249 if (!phy) 250 return 0; 251 252 ret = phy_pm_runtime_get_sync(phy); 253 if (ret < 0 && ret != -ENOTSUPP) 254 return ret; 255 ret = 0; /* Override possible ret == -ENOTSUPP */ 256 257 mutex_lock(&phy->mutex); 258 if (phy->power_count > phy->init_count) 259 dev_warn(&phy->dev, "phy_power_on was called before phy_init\n"); 260 261 if (phy->init_count == 0 && phy->ops->init) { 262 ret = phy->ops->init(phy); 263 if (ret < 0) { 264 dev_err(&phy->dev, "phy init failed --> %d\n", ret); 265 goto out; 266 } 267 } 268 ++phy->init_count; 269 270out: 271 mutex_unlock(&phy->mutex); 272 phy_pm_runtime_put(phy); 273 return ret; 274} 275EXPORT_SYMBOL_GPL(phy_init); 276 277/** 278 * phy_exit - Phy internal un-initialization 279 * @phy: the phy returned by phy_get() 280 * 281 * Must be called after phy_power_off(). 282 * 283 * Return: %0 if successful, a negative error code otherwise 284 */ 285int phy_exit(struct phy *phy) 286{ 287 int ret; 288 289 if (!phy) 290 return 0; 291 292 ret = phy_pm_runtime_get_sync(phy); 293 if (ret < 0 && ret != -ENOTSUPP) 294 return ret; 295 ret = 0; /* Override possible ret == -ENOTSUPP */ 296 297 mutex_lock(&phy->mutex); 298 if (phy->init_count == 1 && phy->ops->exit) { 299 ret = phy->ops->exit(phy); 300 if (ret < 0) { 301 dev_err(&phy->dev, "phy exit failed --> %d\n", ret); 302 goto out; 303 } 304 } 305 --phy->init_count; 306 307out: 308 mutex_unlock(&phy->mutex); 309 phy_pm_runtime_put(phy); 310 return ret; 311} 312EXPORT_SYMBOL_GPL(phy_exit); 313 314/** 315 * phy_power_on - Enable the phy and enter proper operation 316 * @phy: the phy returned by phy_get() 317 * 318 * Must be called after phy_init(). 319 * 320 * Return: %0 if successful, a negative error code otherwise 321 */ 322int phy_power_on(struct phy *phy) 323{ 324 int ret = 0; 325 326 if (!phy) 327 goto out; 328 329 if (phy->pwr) { 330 ret = regulator_enable(phy->pwr); 331 if (ret) 332 goto out; 333 } 334 335 ret = phy_pm_runtime_get_sync(phy); 336 if (ret < 0 && ret != -ENOTSUPP) 337 goto err_pm_sync; 338 339 ret = 0; /* Override possible ret == -ENOTSUPP */ 340 341 mutex_lock(&phy->mutex); 342 if (phy->power_count == 0 && phy->ops->power_on) { 343 ret = phy->ops->power_on(phy); 344 if (ret < 0) { 345 dev_err(&phy->dev, "phy poweron failed --> %d\n", ret); 346 goto err_pwr_on; 347 } 348 } 349 ++phy->power_count; 350 mutex_unlock(&phy->mutex); 351 return 0; 352 353err_pwr_on: 354 mutex_unlock(&phy->mutex); 355 phy_pm_runtime_put_sync(phy); 356err_pm_sync: 357 if (phy->pwr) 358 regulator_disable(phy->pwr); 359out: 360 return ret; 361} 362EXPORT_SYMBOL_GPL(phy_power_on); 363 364/** 365 * phy_power_off - Disable the phy. 366 * @phy: the phy returned by phy_get() 367 * 368 * Must be called before phy_exit(). 369 * 370 * Return: %0 if successful, a negative error code otherwise 371 */ 372int phy_power_off(struct phy *phy) 373{ 374 int ret; 375 376 if (!phy) 377 return 0; 378 379 mutex_lock(&phy->mutex); 380 if (phy->power_count == 1 && phy->ops->power_off) { 381 ret = phy->ops->power_off(phy); 382 if (ret < 0) { 383 dev_err(&phy->dev, "phy poweroff failed --> %d\n", ret); 384 mutex_unlock(&phy->mutex); 385 return ret; 386 } 387 } 388 --phy->power_count; 389 mutex_unlock(&phy->mutex); 390 phy_pm_runtime_put(phy); 391 392 if (phy->pwr) 393 regulator_disable(phy->pwr); 394 395 return 0; 396} 397EXPORT_SYMBOL_GPL(phy_power_off); 398 399int phy_set_mode_ext(struct phy *phy, enum phy_mode mode, int submode) 400{ 401 int ret; 402 403 if (!phy || !phy->ops->set_mode) 404 return 0; 405 406 mutex_lock(&phy->mutex); 407 ret = phy->ops->set_mode(phy, mode, submode); 408 if (!ret) 409 phy->attrs.mode = mode; 410 mutex_unlock(&phy->mutex); 411 412 return ret; 413} 414EXPORT_SYMBOL_GPL(phy_set_mode_ext); 415 416int phy_set_media(struct phy *phy, enum phy_media media) 417{ 418 int ret; 419 420 if (!phy || !phy->ops->set_media) 421 return 0; 422 423 mutex_lock(&phy->mutex); 424 ret = phy->ops->set_media(phy, media); 425 mutex_unlock(&phy->mutex); 426 427 return ret; 428} 429EXPORT_SYMBOL_GPL(phy_set_media); 430 431int phy_set_speed(struct phy *phy, int speed) 432{ 433 int ret; 434 435 if (!phy || !phy->ops->set_speed) 436 return 0; 437 438 mutex_lock(&phy->mutex); 439 ret = phy->ops->set_speed(phy, speed); 440 mutex_unlock(&phy->mutex); 441 442 return ret; 443} 444EXPORT_SYMBOL_GPL(phy_set_speed); 445 446int phy_reset(struct phy *phy) 447{ 448 int ret; 449 450 if (!phy || !phy->ops->reset) 451 return 0; 452 453 ret = phy_pm_runtime_get_sync(phy); 454 if (ret < 0 && ret != -ENOTSUPP) 455 return ret; 456 457 mutex_lock(&phy->mutex); 458 ret = phy->ops->reset(phy); 459 mutex_unlock(&phy->mutex); 460 461 phy_pm_runtime_put(phy); 462 463 return ret; 464} 465EXPORT_SYMBOL_GPL(phy_reset); 466 467/** 468 * phy_calibrate() - Tunes the phy hw parameters for current configuration 469 * @phy: the phy returned by phy_get() 470 * 471 * Used to calibrate phy hardware, typically by adjusting some parameters in 472 * runtime, which are otherwise lost after host controller reset and cannot 473 * be applied in phy_init() or phy_power_on(). 474 * 475 * Return: %0 if successful, a negative error code otherwise 476 */ 477int phy_calibrate(struct phy *phy) 478{ 479 int ret; 480 481 if (!phy || !phy->ops->calibrate) 482 return 0; 483 484 mutex_lock(&phy->mutex); 485 ret = phy->ops->calibrate(phy); 486 mutex_unlock(&phy->mutex); 487 488 return ret; 489} 490EXPORT_SYMBOL_GPL(phy_calibrate); 491 492/** 493 * phy_configure() - Changes the phy parameters 494 * @phy: the phy returned by phy_get() 495 * @opts: New configuration to apply 496 * 497 * Used to change the PHY parameters. phy_init() must have been called 498 * on the phy. The configuration will be applied on the current phy 499 * mode, that can be changed using phy_set_mode(). 500 * 501 * Return: %0 if successful, a negative error code otherwise 502 */ 503int phy_configure(struct phy *phy, union phy_configure_opts *opts) 504{ 505 int ret; 506 507 if (!phy) 508 return -EINVAL; 509 510 if (!phy->ops->configure) 511 return -EOPNOTSUPP; 512 513 mutex_lock(&phy->mutex); 514 ret = phy->ops->configure(phy, opts); 515 mutex_unlock(&phy->mutex); 516 517 return ret; 518} 519EXPORT_SYMBOL_GPL(phy_configure); 520 521/** 522 * phy_validate() - Checks the phy parameters 523 * @phy: the phy returned by phy_get() 524 * @mode: phy_mode the configuration is applicable to. 525 * @submode: PHY submode the configuration is applicable to. 526 * @opts: Configuration to check 527 * 528 * Used to check that the current set of parameters can be handled by 529 * the phy. Implementations are free to tune the parameters passed as 530 * arguments if needed by some implementation detail or 531 * constraints. It will not change any actual configuration of the 532 * PHY, so calling it as many times as deemed fit will have no side 533 * effect. 534 * 535 * Return: %0 if successful, a negative error code otherwise 536 */ 537int phy_validate(struct phy *phy, enum phy_mode mode, int submode, 538 union phy_configure_opts *opts) 539{ 540 int ret; 541 542 if (!phy) 543 return -EINVAL; 544 545 if (!phy->ops->validate) 546 return -EOPNOTSUPP; 547 548 mutex_lock(&phy->mutex); 549 ret = phy->ops->validate(phy, mode, submode, opts); 550 mutex_unlock(&phy->mutex); 551 552 return ret; 553} 554EXPORT_SYMBOL_GPL(phy_validate); 555 556/** 557 * _of_phy_get() - lookup and obtain a reference to a phy by phandle 558 * @np: device_node for which to get the phy 559 * @index: the index of the phy 560 * 561 * Returns the phy associated with the given phandle value, 562 * after getting a refcount to it or -ENODEV if there is no such phy or 563 * -EPROBE_DEFER if there is a phandle to the phy, but the device is 564 * not yet loaded. This function uses of_xlate call back function provided 565 * while registering the phy_provider to find the phy instance. 566 */ 567static struct phy *_of_phy_get(struct device_node *np, int index) 568{ 569 int ret; 570 struct phy_provider *phy_provider; 571 struct phy *phy = NULL; 572 struct of_phandle_args args; 573 574 ret = of_parse_phandle_with_args(np, "phys", "#phy-cells", 575 index, &args); 576 if (ret) 577 return ERR_PTR(-ENODEV); 578 579 /* This phy type handled by the usb-phy subsystem for now */ 580 if (of_device_is_compatible(args.np, "usb-nop-xceiv")) 581 return ERR_PTR(-ENODEV); 582 583 mutex_lock(&phy_provider_mutex); 584 phy_provider = of_phy_provider_lookup(args.np); 585 if (IS_ERR(phy_provider) || !try_module_get(phy_provider->owner)) { 586 phy = ERR_PTR(-EPROBE_DEFER); 587 goto out_unlock; 588 } 589 590 if (!of_device_is_available(args.np)) { 591 dev_warn(phy_provider->dev, "Requested PHY is disabled\n"); 592 phy = ERR_PTR(-ENODEV); 593 goto out_put_module; 594 } 595 596 phy = phy_provider->of_xlate(phy_provider->dev, &args); 597 598out_put_module: 599 module_put(phy_provider->owner); 600 601out_unlock: 602 mutex_unlock(&phy_provider_mutex); 603 of_node_put(args.np); 604 605 return phy; 606} 607 608/** 609 * of_phy_get() - lookup and obtain a reference to a phy using a device_node. 610 * @np: device_node for which to get the phy 611 * @con_id: name of the phy from device's point of view 612 * 613 * Returns the phy driver, after getting a refcount to it; or 614 * -ENODEV if there is no such phy. The caller is responsible for 615 * calling phy_put() to release that count. 616 */ 617struct phy *of_phy_get(struct device_node *np, const char *con_id) 618{ 619 struct phy *phy = NULL; 620 int index = 0; 621 622 if (con_id) 623 index = of_property_match_string(np, "phy-names", con_id); 624 625 phy = _of_phy_get(np, index); 626 if (IS_ERR(phy)) 627 return phy; 628 629 if (!try_module_get(phy->ops->owner)) 630 return ERR_PTR(-EPROBE_DEFER); 631 632 get_device(&phy->dev); 633 634 return phy; 635} 636EXPORT_SYMBOL_GPL(of_phy_get); 637 638/** 639 * of_phy_put() - release the PHY 640 * @phy: the phy returned by of_phy_get() 641 * 642 * Releases a refcount the caller received from of_phy_get(). 643 */ 644void of_phy_put(struct phy *phy) 645{ 646 if (!phy || IS_ERR(phy)) 647 return; 648 649 mutex_lock(&phy->mutex); 650 if (phy->ops->release) 651 phy->ops->release(phy); 652 mutex_unlock(&phy->mutex); 653 654 module_put(phy->ops->owner); 655 put_device(&phy->dev); 656} 657EXPORT_SYMBOL_GPL(of_phy_put); 658 659/** 660 * phy_put() - release the PHY 661 * @dev: device that wants to release this phy 662 * @phy: the phy returned by phy_get() 663 * 664 * Releases a refcount the caller received from phy_get(). 665 */ 666void phy_put(struct device *dev, struct phy *phy) 667{ 668 device_link_remove(dev, &phy->dev); 669 of_phy_put(phy); 670} 671EXPORT_SYMBOL_GPL(phy_put); 672 673/** 674 * devm_phy_put() - release the PHY 675 * @dev: device that wants to release this phy 676 * @phy: the phy returned by devm_phy_get() 677 * 678 * destroys the devres associated with this phy and invokes phy_put 679 * to release the phy. 680 */ 681void devm_phy_put(struct device *dev, struct phy *phy) 682{ 683 int r; 684 685 if (!phy) 686 return; 687 688 r = devres_destroy(dev, devm_phy_release, devm_phy_match, phy); 689 dev_WARN_ONCE(dev, r, "couldn't find PHY resource\n"); 690} 691EXPORT_SYMBOL_GPL(devm_phy_put); 692 693/** 694 * of_phy_simple_xlate() - returns the phy instance from phy provider 695 * @dev: the PHY provider device 696 * @args: of_phandle_args (not used here) 697 * 698 * Intended to be used by phy provider for the common case where #phy-cells is 699 * 0. For other cases where #phy-cells is greater than '0', the phy provider 700 * should provide a custom of_xlate function that reads the *args* and returns 701 * the appropriate phy. 702 */ 703struct phy *of_phy_simple_xlate(struct device *dev, struct of_phandle_args 704 *args) 705{ 706 struct phy *phy; 707 struct class_dev_iter iter; 708 709 class_dev_iter_init(&iter, phy_class, NULL, NULL); 710 while ((dev = class_dev_iter_next(&iter))) { 711 phy = to_phy(dev); 712 if (args->np != phy->dev.of_node) 713 continue; 714 715 class_dev_iter_exit(&iter); 716 return phy; 717 } 718 719 class_dev_iter_exit(&iter); 720 return ERR_PTR(-ENODEV); 721} 722EXPORT_SYMBOL_GPL(of_phy_simple_xlate); 723 724/** 725 * phy_get() - lookup and obtain a reference to a phy. 726 * @dev: device that requests this phy 727 * @string: the phy name as given in the dt data or the name of the controller 728 * port for non-dt case 729 * 730 * Returns the phy driver, after getting a refcount to it; or 731 * -ENODEV if there is no such phy. The caller is responsible for 732 * calling phy_put() to release that count. 733 */ 734struct phy *phy_get(struct device *dev, const char *string) 735{ 736 int index = 0; 737 struct phy *phy; 738 struct device_link *link; 739 740 if (dev->of_node) { 741 if (string) 742 index = of_property_match_string(dev->of_node, "phy-names", 743 string); 744 else 745 index = 0; 746 phy = _of_phy_get(dev->of_node, index); 747 } else { 748 if (string == NULL) { 749 dev_WARN(dev, "missing string\n"); 750 return ERR_PTR(-EINVAL); 751 } 752 phy = phy_find(dev, string); 753 } 754 if (IS_ERR(phy)) 755 return phy; 756 757 if (!try_module_get(phy->ops->owner)) 758 return ERR_PTR(-EPROBE_DEFER); 759 760 get_device(&phy->dev); 761 762 link = device_link_add(dev, &phy->dev, DL_FLAG_STATELESS); 763 if (!link) 764 dev_dbg(dev, "failed to create device link to %s\n", 765 dev_name(phy->dev.parent)); 766 767 return phy; 768} 769EXPORT_SYMBOL_GPL(phy_get); 770 771/** 772 * devm_phy_get() - lookup and obtain a reference to a phy. 773 * @dev: device that requests this phy 774 * @string: the phy name as given in the dt data or phy device name 775 * for non-dt case 776 * 777 * Gets the phy using phy_get(), and associates a device with it using 778 * devres. On driver detach, release function is invoked on the devres data, 779 * then, devres data is freed. 780 */ 781struct phy *devm_phy_get(struct device *dev, const char *string) 782{ 783 struct phy **ptr, *phy; 784 785 ptr = devres_alloc(devm_phy_release, sizeof(*ptr), GFP_KERNEL); 786 if (!ptr) 787 return ERR_PTR(-ENOMEM); 788 789 phy = phy_get(dev, string); 790 if (!IS_ERR(phy)) { 791 *ptr = phy; 792 devres_add(dev, ptr); 793 } else { 794 devres_free(ptr); 795 } 796 797 return phy; 798} 799EXPORT_SYMBOL_GPL(devm_phy_get); 800 801/** 802 * devm_phy_optional_get() - lookup and obtain a reference to an optional phy. 803 * @dev: device that requests this phy 804 * @string: the phy name as given in the dt data or phy device name 805 * for non-dt case 806 * 807 * Gets the phy using phy_get(), and associates a device with it using 808 * devres. On driver detach, release function is invoked on the devres 809 * data, then, devres data is freed. This differs to devm_phy_get() in 810 * that if the phy does not exist, it is not considered an error and 811 * -ENODEV will not be returned. Instead the NULL phy is returned, 812 * which can be passed to all other phy consumer calls. 813 */ 814struct phy *devm_phy_optional_get(struct device *dev, const char *string) 815{ 816 struct phy *phy = devm_phy_get(dev, string); 817 818 if (PTR_ERR(phy) == -ENODEV) 819 phy = NULL; 820 821 return phy; 822} 823EXPORT_SYMBOL_GPL(devm_phy_optional_get); 824 825/** 826 * devm_of_phy_get() - lookup and obtain a reference to a phy. 827 * @dev: device that requests this phy 828 * @np: node containing the phy 829 * @con_id: name of the phy from device's point of view 830 * 831 * Gets the phy using of_phy_get(), and associates a device with it using 832 * devres. On driver detach, release function is invoked on the devres data, 833 * then, devres data is freed. 834 */ 835struct phy *devm_of_phy_get(struct device *dev, struct device_node *np, 836 const char *con_id) 837{ 838 struct phy **ptr, *phy; 839 struct device_link *link; 840 841 ptr = devres_alloc(devm_phy_release, sizeof(*ptr), GFP_KERNEL); 842 if (!ptr) 843 return ERR_PTR(-ENOMEM); 844 845 phy = of_phy_get(np, con_id); 846 if (!IS_ERR(phy)) { 847 *ptr = phy; 848 devres_add(dev, ptr); 849 } else { 850 devres_free(ptr); 851 return phy; 852 } 853 854 link = device_link_add(dev, &phy->dev, DL_FLAG_STATELESS); 855 if (!link) 856 dev_dbg(dev, "failed to create device link to %s\n", 857 dev_name(phy->dev.parent)); 858 859 return phy; 860} 861EXPORT_SYMBOL_GPL(devm_of_phy_get); 862 863/** 864 * devm_of_phy_optional_get() - lookup and obtain a reference to an optional 865 * phy. 866 * @dev: device that requests this phy 867 * @np: node containing the phy 868 * @con_id: name of the phy from device's point of view 869 * 870 * Gets the phy using of_phy_get(), and associates a device with it using 871 * devres. On driver detach, release function is invoked on the devres data, 872 * then, devres data is freed. This differs to devm_of_phy_get() in 873 * that if the phy does not exist, it is not considered an error and 874 * -ENODEV will not be returned. Instead the NULL phy is returned, 875 * which can be passed to all other phy consumer calls. 876 */ 877struct phy *devm_of_phy_optional_get(struct device *dev, struct device_node *np, 878 const char *con_id) 879{ 880 struct phy *phy = devm_of_phy_get(dev, np, con_id); 881 882 if (PTR_ERR(phy) == -ENODEV) 883 phy = NULL; 884 885 if (IS_ERR(phy)) 886 dev_err_probe(dev, PTR_ERR(phy), "failed to get PHY %pOF:%s", 887 np, con_id); 888 889 return phy; 890} 891EXPORT_SYMBOL_GPL(devm_of_phy_optional_get); 892 893/** 894 * devm_of_phy_get_by_index() - lookup and obtain a reference to a phy by index. 895 * @dev: device that requests this phy 896 * @np: node containing the phy 897 * @index: index of the phy 898 * 899 * Gets the phy using _of_phy_get(), then gets a refcount to it, 900 * and associates a device with it using devres. On driver detach, 901 * release function is invoked on the devres data, 902 * then, devres data is freed. 903 * 904 */ 905struct phy *devm_of_phy_get_by_index(struct device *dev, struct device_node *np, 906 int index) 907{ 908 struct phy **ptr, *phy; 909 struct device_link *link; 910 911 ptr = devres_alloc(devm_phy_release, sizeof(*ptr), GFP_KERNEL); 912 if (!ptr) 913 return ERR_PTR(-ENOMEM); 914 915 phy = _of_phy_get(np, index); 916 if (IS_ERR(phy)) { 917 devres_free(ptr); 918 return phy; 919 } 920 921 if (!try_module_get(phy->ops->owner)) { 922 devres_free(ptr); 923 return ERR_PTR(-EPROBE_DEFER); 924 } 925 926 get_device(&phy->dev); 927 928 *ptr = phy; 929 devres_add(dev, ptr); 930 931 link = device_link_add(dev, &phy->dev, DL_FLAG_STATELESS); 932 if (!link) 933 dev_dbg(dev, "failed to create device link to %s\n", 934 dev_name(phy->dev.parent)); 935 936 return phy; 937} 938EXPORT_SYMBOL_GPL(devm_of_phy_get_by_index); 939 940/** 941 * phy_create() - create a new phy 942 * @dev: device that is creating the new phy 943 * @node: device node of the phy 944 * @ops: function pointers for performing phy operations 945 * 946 * Called to create a phy using phy framework. 947 */ 948struct phy *phy_create(struct device *dev, struct device_node *node, 949 const struct phy_ops *ops) 950{ 951 int ret; 952 int id; 953 struct phy *phy; 954 955 if (WARN_ON(!dev)) 956 return ERR_PTR(-EINVAL); 957 958 phy = kzalloc(sizeof(*phy), GFP_KERNEL); 959 if (!phy) 960 return ERR_PTR(-ENOMEM); 961 962 id = ida_simple_get(&phy_ida, 0, 0, GFP_KERNEL); 963 if (id < 0) { 964 dev_err(dev, "unable to get id\n"); 965 ret = id; 966 goto free_phy; 967 } 968 969 device_initialize(&phy->dev); 970 mutex_init(&phy->mutex); 971 972 phy->dev.class = phy_class; 973 phy->dev.parent = dev; 974 phy->dev.of_node = node ?: dev->of_node; 975 phy->id = id; 976 phy->ops = ops; 977 978 ret = dev_set_name(&phy->dev, "phy-%s.%d", dev_name(dev), id); 979 if (ret) 980 goto put_dev; 981 982 /* phy-supply */ 983 phy->pwr = regulator_get_optional(&phy->dev, "phy"); 984 if (IS_ERR(phy->pwr)) { 985 ret = PTR_ERR(phy->pwr); 986 if (ret == -EPROBE_DEFER) 987 goto put_dev; 988 989 phy->pwr = NULL; 990 } 991 992 ret = device_add(&phy->dev); 993 if (ret) 994 goto put_dev; 995 996 if (pm_runtime_enabled(dev)) { 997 pm_runtime_enable(&phy->dev); 998 pm_runtime_no_callbacks(&phy->dev); 999 } 1000 1001 phy->debugfs = debugfs_create_dir(dev_name(&phy->dev), phy_debugfs_root); 1002 1003 return phy; 1004 1005put_dev: 1006 put_device(&phy->dev); /* calls phy_release() which frees resources */ 1007 return ERR_PTR(ret); 1008 1009free_phy: 1010 kfree(phy); 1011 return ERR_PTR(ret); 1012} 1013EXPORT_SYMBOL_GPL(phy_create); 1014 1015/** 1016 * devm_phy_create() - create a new phy 1017 * @dev: device that is creating the new phy 1018 * @node: device node of the phy 1019 * @ops: function pointers for performing phy operations 1020 * 1021 * Creates a new PHY device adding it to the PHY class. 1022 * While at that, it also associates the device with the phy using devres. 1023 * On driver detach, release function is invoked on the devres data, 1024 * then, devres data is freed. 1025 */ 1026struct phy *devm_phy_create(struct device *dev, struct device_node *node, 1027 const struct phy_ops *ops) 1028{ 1029 struct phy **ptr, *phy; 1030 1031 ptr = devres_alloc(devm_phy_consume, sizeof(*ptr), GFP_KERNEL); 1032 if (!ptr) 1033 return ERR_PTR(-ENOMEM); 1034 1035 phy = phy_create(dev, node, ops); 1036 if (!IS_ERR(phy)) { 1037 *ptr = phy; 1038 devres_add(dev, ptr); 1039 } else { 1040 devres_free(ptr); 1041 } 1042 1043 return phy; 1044} 1045EXPORT_SYMBOL_GPL(devm_phy_create); 1046 1047/** 1048 * phy_destroy() - destroy the phy 1049 * @phy: the phy to be destroyed 1050 * 1051 * Called to destroy the phy. 1052 */ 1053void phy_destroy(struct phy *phy) 1054{ 1055 pm_runtime_disable(&phy->dev); 1056 device_unregister(&phy->dev); 1057} 1058EXPORT_SYMBOL_GPL(phy_destroy); 1059 1060/** 1061 * devm_phy_destroy() - destroy the PHY 1062 * @dev: device that wants to release this phy 1063 * @phy: the phy returned by devm_phy_get() 1064 * 1065 * destroys the devres associated with this phy and invokes phy_destroy 1066 * to destroy the phy. 1067 */ 1068void devm_phy_destroy(struct device *dev, struct phy *phy) 1069{ 1070 int r; 1071 1072 r = devres_destroy(dev, devm_phy_consume, devm_phy_match, phy); 1073 dev_WARN_ONCE(dev, r, "couldn't find PHY resource\n"); 1074} 1075EXPORT_SYMBOL_GPL(devm_phy_destroy); 1076 1077/** 1078 * __of_phy_provider_register() - create/register phy provider with the framework 1079 * @dev: struct device of the phy provider 1080 * @children: device node containing children (if different from dev->of_node) 1081 * @owner: the module owner containing of_xlate 1082 * @of_xlate: function pointer to obtain phy instance from phy provider 1083 * 1084 * Creates struct phy_provider from dev and of_xlate function pointer. 1085 * This is used in the case of dt boot for finding the phy instance from 1086 * phy provider. 1087 * 1088 * If the PHY provider doesn't nest children directly but uses a separate 1089 * child node to contain the individual children, the @children parameter 1090 * can be used to override the default. If NULL, the default (dev->of_node) 1091 * will be used. If non-NULL, the device node must be a child (or further 1092 * descendant) of dev->of_node. Otherwise an ERR_PTR()-encoded -EINVAL 1093 * error code is returned. 1094 */ 1095struct phy_provider *__of_phy_provider_register(struct device *dev, 1096 struct device_node *children, struct module *owner, 1097 struct phy * (*of_xlate)(struct device *dev, 1098 struct of_phandle_args *args)) 1099{ 1100 struct phy_provider *phy_provider; 1101 1102 /* 1103 * If specified, the device node containing the children must itself 1104 * be the provider's device node or a child (or further descendant) 1105 * thereof. 1106 */ 1107 if (children) { 1108 struct device_node *parent = of_node_get(children), *next; 1109 1110 while (parent) { 1111 if (parent == dev->of_node) 1112 break; 1113 1114 next = of_get_parent(parent); 1115 of_node_put(parent); 1116 parent = next; 1117 } 1118 1119 if (!parent) 1120 return ERR_PTR(-EINVAL); 1121 1122 of_node_put(parent); 1123 } else { 1124 children = dev->of_node; 1125 } 1126 1127 phy_provider = kzalloc(sizeof(*phy_provider), GFP_KERNEL); 1128 if (!phy_provider) 1129 return ERR_PTR(-ENOMEM); 1130 1131 phy_provider->dev = dev; 1132 phy_provider->children = of_node_get(children); 1133 phy_provider->owner = owner; 1134 phy_provider->of_xlate = of_xlate; 1135 1136 mutex_lock(&phy_provider_mutex); 1137 list_add_tail(&phy_provider->list, &phy_provider_list); 1138 mutex_unlock(&phy_provider_mutex); 1139 1140 return phy_provider; 1141} 1142EXPORT_SYMBOL_GPL(__of_phy_provider_register); 1143 1144/** 1145 * __devm_of_phy_provider_register() - create/register phy provider with the 1146 * framework 1147 * @dev: struct device of the phy provider 1148 * @children: device node containing children (if different from dev->of_node) 1149 * @owner: the module owner containing of_xlate 1150 * @of_xlate: function pointer to obtain phy instance from phy provider 1151 * 1152 * Creates struct phy_provider from dev and of_xlate function pointer. 1153 * This is used in the case of dt boot for finding the phy instance from 1154 * phy provider. While at that, it also associates the device with the 1155 * phy provider using devres. On driver detach, release function is invoked 1156 * on the devres data, then, devres data is freed. 1157 */ 1158struct phy_provider *__devm_of_phy_provider_register(struct device *dev, 1159 struct device_node *children, struct module *owner, 1160 struct phy * (*of_xlate)(struct device *dev, 1161 struct of_phandle_args *args)) 1162{ 1163 struct phy_provider **ptr, *phy_provider; 1164 1165 ptr = devres_alloc(devm_phy_provider_release, sizeof(*ptr), GFP_KERNEL); 1166 if (!ptr) 1167 return ERR_PTR(-ENOMEM); 1168 1169 phy_provider = __of_phy_provider_register(dev, children, owner, 1170 of_xlate); 1171 if (!IS_ERR(phy_provider)) { 1172 *ptr = phy_provider; 1173 devres_add(dev, ptr); 1174 } else { 1175 devres_free(ptr); 1176 } 1177 1178 return phy_provider; 1179} 1180EXPORT_SYMBOL_GPL(__devm_of_phy_provider_register); 1181 1182/** 1183 * of_phy_provider_unregister() - unregister phy provider from the framework 1184 * @phy_provider: phy provider returned by of_phy_provider_register() 1185 * 1186 * Removes the phy_provider created using of_phy_provider_register(). 1187 */ 1188void of_phy_provider_unregister(struct phy_provider *phy_provider) 1189{ 1190 if (IS_ERR(phy_provider)) 1191 return; 1192 1193 mutex_lock(&phy_provider_mutex); 1194 list_del(&phy_provider->list); 1195 of_node_put(phy_provider->children); 1196 kfree(phy_provider); 1197 mutex_unlock(&phy_provider_mutex); 1198} 1199EXPORT_SYMBOL_GPL(of_phy_provider_unregister); 1200 1201/** 1202 * devm_of_phy_provider_unregister() - remove phy provider from the framework 1203 * @dev: struct device of the phy provider 1204 * @phy_provider: phy provider returned by of_phy_provider_register() 1205 * 1206 * destroys the devres associated with this phy provider and invokes 1207 * of_phy_provider_unregister to unregister the phy provider. 1208 */ 1209void devm_of_phy_provider_unregister(struct device *dev, 1210 struct phy_provider *phy_provider) 1211{ 1212 int r; 1213 1214 r = devres_destroy(dev, devm_phy_provider_release, devm_phy_match, 1215 phy_provider); 1216 dev_WARN_ONCE(dev, r, "couldn't find PHY provider device resource\n"); 1217} 1218EXPORT_SYMBOL_GPL(devm_of_phy_provider_unregister); 1219 1220/** 1221 * phy_release() - release the phy 1222 * @dev: the dev member within phy 1223 * 1224 * When the last reference to the device is removed, it is called 1225 * from the embedded kobject as release method. 1226 */ 1227static void phy_release(struct device *dev) 1228{ 1229 struct phy *phy; 1230 1231 phy = to_phy(dev); 1232 dev_vdbg(dev, "releasing '%s'\n", dev_name(dev)); 1233 debugfs_remove_recursive(phy->debugfs); 1234 regulator_put(phy->pwr); 1235 ida_simple_remove(&phy_ida, phy->id); 1236 kfree(phy); 1237} 1238 1239static int __init phy_core_init(void) 1240{ 1241 phy_class = class_create("phy"); 1242 if (IS_ERR(phy_class)) { 1243 pr_err("failed to create phy class --> %ld\n", 1244 PTR_ERR(phy_class)); 1245 return PTR_ERR(phy_class); 1246 } 1247 1248 phy_class->dev_release = phy_release; 1249 1250 phy_debugfs_root = debugfs_create_dir("phy", NULL); 1251 1252 return 0; 1253} 1254device_initcall(phy_core_init); 1255 1256static void __exit phy_core_exit(void) 1257{ 1258 debugfs_remove_recursive(phy_debugfs_root); 1259 class_destroy(phy_class); 1260} 1261module_exit(phy_core_exit);