mirror of
				https://source.denx.de/u-boot/u-boot.git
				synced 2025-10-31 08:21:36 +01:00 
			
		
		
		
	In drivers usb/host/{ehci,ohci}-generic.c, {ehci,ohci}_setup_phy() and
{ehci,ohci}_shutdown_phy() shares 95% of common code.
Factorize this code in new generic_{setup,shudown}_phy() functions.
Signed-off-by: Patrice Chotard <patrice.chotard@foss.st.com>
Cc: Marek Vasut <marex@denx.de>
Cc: Simon Glass <sjg@chromium.org>
		
	
			
		
			
				
	
	
		
			507 lines
		
	
	
		
			10 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			507 lines
		
	
	
		
			10 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0+
 | |
| /*
 | |
|  * Copyright (C) 2017 Texas Instruments Incorporated - http://www.ti.com/
 | |
|  * Written by Jean-Jacques Hiblot  <jjhiblot@ti.com>
 | |
|  */
 | |
| 
 | |
| #define LOG_CATEGORY UCLASS_PHY
 | |
| 
 | |
| #include <common.h>
 | |
| #include <dm.h>
 | |
| #include <dm/device_compat.h>
 | |
| #include <dm/devres.h>
 | |
| #include <generic-phy.h>
 | |
| #include <linux/list.h>
 | |
| 
 | |
| /**
 | |
|  * struct phy_counts - Init and power-on counts of a single PHY port
 | |
|  *
 | |
|  * This structure is used to keep track of PHY initialization and power
 | |
|  * state change requests, so that we don't power off and deinitialize a
 | |
|  * PHY instance until all of its users want it done. Otherwise, multiple
 | |
|  * consumers using the same PHY port can cause problems (e.g. one might
 | |
|  * call power_off() after another's exit() and hang indefinitely).
 | |
|  *
 | |
|  * @id: The PHY ID within a PHY provider
 | |
|  * @power_on_count: Times generic_phy_power_on() was called for this ID
 | |
|  *                  without a matching generic_phy_power_off() afterwards
 | |
|  * @init_count: Times generic_phy_init() was called for this ID
 | |
|  *              without a matching generic_phy_exit() afterwards
 | |
|  * @list: Handle for a linked list of these structures corresponding to
 | |
|  *        ports of the same PHY provider
 | |
|  */
 | |
| struct phy_counts {
 | |
| 	unsigned long id;
 | |
| 	int power_on_count;
 | |
| 	int init_count;
 | |
| 	struct list_head list;
 | |
| };
 | |
| 
 | |
| static inline struct phy_ops *phy_dev_ops(struct udevice *dev)
 | |
| {
 | |
| 	return (struct phy_ops *)dev->driver->ops;
 | |
| }
 | |
| 
 | |
| static struct phy_counts *phy_get_counts(struct phy *phy)
 | |
| {
 | |
| 	struct list_head *uc_priv;
 | |
| 	struct phy_counts *counts;
 | |
| 
 | |
| 	if (!generic_phy_valid(phy))
 | |
| 		return NULL;
 | |
| 
 | |
| 	uc_priv = dev_get_uclass_priv(phy->dev);
 | |
| 	list_for_each_entry(counts, uc_priv, list)
 | |
| 		if (counts->id == phy->id)
 | |
| 			return counts;
 | |
| 
 | |
| 	return NULL;
 | |
| }
 | |
| 
 | |
| static int phy_alloc_counts(struct phy *phy)
 | |
| {
 | |
| 	struct list_head *uc_priv;
 | |
| 	struct phy_counts *counts;
 | |
| 
 | |
| 	if (!generic_phy_valid(phy))
 | |
| 		return 0;
 | |
| 	if (phy_get_counts(phy))
 | |
| 		return 0;
 | |
| 
 | |
| 	uc_priv = dev_get_uclass_priv(phy->dev);
 | |
| 	counts = kzalloc(sizeof(*counts), GFP_KERNEL);
 | |
| 	if (!counts)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	counts->id = phy->id;
 | |
| 	counts->power_on_count = 0;
 | |
| 	counts->init_count = 0;
 | |
| 	list_add(&counts->list, uc_priv);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int phy_uclass_pre_probe(struct udevice *dev)
 | |
| {
 | |
| 	struct list_head *uc_priv = dev_get_uclass_priv(dev);
 | |
| 
 | |
| 	INIT_LIST_HEAD(uc_priv);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int phy_uclass_pre_remove(struct udevice *dev)
 | |
| {
 | |
| 	struct list_head *uc_priv = dev_get_uclass_priv(dev);
 | |
| 	struct phy_counts *counts, *next;
 | |
| 
 | |
| 	list_for_each_entry_safe(counts, next, uc_priv, list)
 | |
| 		kfree(counts);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int generic_phy_xlate_offs_flags(struct phy *phy,
 | |
| 					struct ofnode_phandle_args *args)
 | |
| {
 | |
| 	debug("%s(phy=%p)\n", __func__, phy);
 | |
| 
 | |
| 	if (args->args_count > 1) {
 | |
| 		debug("Invalid args_count: %d\n", args->args_count);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	if (args->args_count)
 | |
| 		phy->id = args->args[0];
 | |
| 	else
 | |
| 		phy->id = 0;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int generic_phy_get_by_index_nodev(ofnode node, int index, struct phy *phy)
 | |
| {
 | |
| 	struct ofnode_phandle_args args;
 | |
| 	struct phy_ops *ops;
 | |
| 	struct udevice *phydev;
 | |
| 	int i, ret;
 | |
| 
 | |
| 	debug("%s(node=%s, index=%d, phy=%p)\n",
 | |
| 	      __func__, ofnode_get_name(node), index, phy);
 | |
| 
 | |
| 	assert(phy);
 | |
| 	phy->dev = NULL;
 | |
| 	ret = ofnode_parse_phandle_with_args(node, "phys", "#phy-cells", 0,
 | |
| 					     index, &args);
 | |
| 	if (ret) {
 | |
| 		debug("%s: dev_read_phandle_with_args failed: err=%d\n",
 | |
| 		      __func__, ret);
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	ret = uclass_get_device_by_ofnode(UCLASS_PHY, args.node, &phydev);
 | |
| 	if (ret) {
 | |
| 		debug("%s: uclass_get_device_by_ofnode failed: err=%d\n",
 | |
| 		      __func__, ret);
 | |
| 
 | |
| 		/* Check if args.node's parent is a PHY provider */
 | |
| 		ret = uclass_get_device_by_ofnode(UCLASS_PHY,
 | |
| 						  ofnode_get_parent(args.node),
 | |
| 						  &phydev);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 
 | |
| 		/* insert phy idx at first position into args array */
 | |
| 		for (i = args.args_count; i >= 1 ; i--)
 | |
| 			args.args[i] = args.args[i - 1];
 | |
| 
 | |
| 		args.args_count++;
 | |
| 		args.args[0] = ofnode_read_u32_default(args.node, "reg", -1);
 | |
| 	}
 | |
| 
 | |
| 	phy->dev = phydev;
 | |
| 
 | |
| 	ops = phy_dev_ops(phydev);
 | |
| 
 | |
| 	if (ops->of_xlate)
 | |
| 		ret = ops->of_xlate(phy, &args);
 | |
| 	else
 | |
| 		ret = generic_phy_xlate_offs_flags(phy, &args);
 | |
| 	if (ret) {
 | |
| 		debug("of_xlate() failed: %d\n", ret);
 | |
| 		goto err;
 | |
| 	}
 | |
| 
 | |
| 	ret = phy_alloc_counts(phy);
 | |
| 	if (ret) {
 | |
| 		debug("phy_alloc_counts() failed: %d\n", ret);
 | |
| 		goto err;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| err:
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_get_by_index(struct udevice *dev, int index,
 | |
| 			     struct phy *phy)
 | |
| {
 | |
| 	return generic_phy_get_by_index_nodev(dev_ofnode(dev), index, phy);
 | |
| }
 | |
| 
 | |
| int generic_phy_get_by_name(struct udevice *dev, const char *phy_name,
 | |
| 			    struct phy *phy)
 | |
| {
 | |
| 	int index;
 | |
| 
 | |
| 	debug("%s(dev=%p, name=%s, phy=%p)\n", __func__, dev, phy_name, phy);
 | |
| 
 | |
| 	index = dev_read_stringlist_search(dev, "phy-names", phy_name);
 | |
| 	if (index < 0) {
 | |
| 		debug("dev_read_stringlist_search() failed: %d\n", index);
 | |
| 		return index;
 | |
| 	}
 | |
| 
 | |
| 	return generic_phy_get_by_index(dev, index, phy);
 | |
| }
 | |
| 
 | |
| int generic_phy_init(struct phy *phy)
 | |
| {
 | |
| 	struct phy_counts *counts;
 | |
| 	struct phy_ops const *ops;
 | |
| 	int ret;
 | |
| 
 | |
| 	if (!generic_phy_valid(phy))
 | |
| 		return 0;
 | |
| 	ops = phy_dev_ops(phy->dev);
 | |
| 	if (!ops->init)
 | |
| 		return 0;
 | |
| 
 | |
| 	counts = phy_get_counts(phy);
 | |
| 	if (counts->init_count > 0) {
 | |
| 		counts->init_count++;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	ret = ops->init(phy);
 | |
| 	if (ret)
 | |
| 		dev_err(phy->dev, "PHY: Failed to init %s: %d.\n",
 | |
| 			phy->dev->name, ret);
 | |
| 	else
 | |
| 		counts->init_count = 1;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_reset(struct phy *phy)
 | |
| {
 | |
| 	struct phy_ops const *ops;
 | |
| 	int ret;
 | |
| 
 | |
| 	if (!generic_phy_valid(phy))
 | |
| 		return 0;
 | |
| 	ops = phy_dev_ops(phy->dev);
 | |
| 	if (!ops->reset)
 | |
| 		return 0;
 | |
| 	ret = ops->reset(phy);
 | |
| 	if (ret)
 | |
| 		dev_err(phy->dev, "PHY: Failed to reset %s: %d.\n",
 | |
| 			phy->dev->name, ret);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_exit(struct phy *phy)
 | |
| {
 | |
| 	struct phy_counts *counts;
 | |
| 	struct phy_ops const *ops;
 | |
| 	int ret;
 | |
| 
 | |
| 	if (!generic_phy_valid(phy))
 | |
| 		return 0;
 | |
| 	ops = phy_dev_ops(phy->dev);
 | |
| 	if (!ops->exit)
 | |
| 		return 0;
 | |
| 
 | |
| 	counts = phy_get_counts(phy);
 | |
| 	if (counts->init_count == 0)
 | |
| 		return 0;
 | |
| 	if (counts->init_count > 1) {
 | |
| 		counts->init_count--;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	ret = ops->exit(phy);
 | |
| 	if (ret)
 | |
| 		dev_err(phy->dev, "PHY: Failed to exit %s: %d.\n",
 | |
| 			phy->dev->name, ret);
 | |
| 	else
 | |
| 		counts->init_count = 0;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_power_on(struct phy *phy)
 | |
| {
 | |
| 	struct phy_counts *counts;
 | |
| 	struct phy_ops const *ops;
 | |
| 	int ret;
 | |
| 
 | |
| 	if (!generic_phy_valid(phy))
 | |
| 		return 0;
 | |
| 	ops = phy_dev_ops(phy->dev);
 | |
| 	if (!ops->power_on)
 | |
| 		return 0;
 | |
| 
 | |
| 	counts = phy_get_counts(phy);
 | |
| 	if (counts->power_on_count > 0) {
 | |
| 		counts->power_on_count++;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	ret = ops->power_on(phy);
 | |
| 	if (ret)
 | |
| 		dev_err(phy->dev, "PHY: Failed to power on %s: %d.\n",
 | |
| 			phy->dev->name, ret);
 | |
| 	else
 | |
| 		counts->power_on_count = 1;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_power_off(struct phy *phy)
 | |
| {
 | |
| 	struct phy_counts *counts;
 | |
| 	struct phy_ops const *ops;
 | |
| 	int ret;
 | |
| 
 | |
| 	if (!generic_phy_valid(phy))
 | |
| 		return 0;
 | |
| 	ops = phy_dev_ops(phy->dev);
 | |
| 	if (!ops->power_off)
 | |
| 		return 0;
 | |
| 
 | |
| 	counts = phy_get_counts(phy);
 | |
| 	if (counts->power_on_count == 0)
 | |
| 		return 0;
 | |
| 	if (counts->power_on_count > 1) {
 | |
| 		counts->power_on_count--;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	ret = ops->power_off(phy);
 | |
| 	if (ret)
 | |
| 		dev_err(phy->dev, "PHY: Failed to power off %s: %d.\n",
 | |
| 			phy->dev->name, ret);
 | |
| 	else
 | |
| 		counts->power_on_count = 0;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_configure(struct phy *phy, void *params)
 | |
| {
 | |
| 	struct phy_ops const *ops;
 | |
| 
 | |
| 	if (!generic_phy_valid(phy))
 | |
| 		return 0;
 | |
| 	ops = phy_dev_ops(phy->dev);
 | |
| 
 | |
| 	return ops->configure ? ops->configure(phy, params) : 0;
 | |
| }
 | |
| 
 | |
| int generic_phy_get_bulk(struct udevice *dev, struct phy_bulk *bulk)
 | |
| {
 | |
| 	int i, ret, count;
 | |
| 	struct udevice *phydev = dev;
 | |
| 
 | |
| 	bulk->count = 0;
 | |
| 
 | |
| 	/* Return if no phy declared */
 | |
| 	if (!dev_read_prop(dev, "phys", NULL)) {
 | |
| 		phydev = dev->parent;
 | |
| 		if (!dev_read_prop(phydev, "phys", NULL)) {
 | |
| 			pr_err("%s : no phys property\n", __func__);
 | |
| 			return 0;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	count = dev_count_phandle_with_args(phydev, "phys", "#phy-cells", 0);
 | |
| 	if (count < 1) {
 | |
| 		pr_err("%s : no phys found %d\n", __func__, count);
 | |
| 		return count;
 | |
| 	}
 | |
| 
 | |
| 	bulk->phys = devm_kcalloc(phydev, count, sizeof(struct phy), GFP_KERNEL);
 | |
| 	if (!bulk->phys)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	for (i = 0; i < count; i++) {
 | |
| 		ret = generic_phy_get_by_index(phydev, i, &bulk->phys[i]);
 | |
| 		if (ret) {
 | |
| 			pr_err("Failed to get PHY%d for %s\n", i, dev->name);
 | |
| 			return ret;
 | |
| 		}
 | |
| 		bulk->count++;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int generic_phy_init_bulk(struct phy_bulk *bulk)
 | |
| {
 | |
| 	struct phy *phys = bulk->phys;
 | |
| 	int i, ret;
 | |
| 
 | |
| 	for (i = 0; i < bulk->count; i++) {
 | |
| 		ret = generic_phy_init(&phys[i]);
 | |
| 		if (ret) {
 | |
| 			pr_err("Can't init PHY%d\n", i);
 | |
| 			goto phys_init_err;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| phys_init_err:
 | |
| 	for (; i > 0; i--)
 | |
| 		generic_phy_exit(&phys[i - 1]);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_exit_bulk(struct phy_bulk *bulk)
 | |
| {
 | |
| 	struct phy *phys = bulk->phys;
 | |
| 	int i, ret = 0;
 | |
| 
 | |
| 	for (i = 0; i < bulk->count; i++)
 | |
| 		ret |= generic_phy_exit(&phys[i]);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_power_on_bulk(struct phy_bulk *bulk)
 | |
| {
 | |
| 	struct phy *phys = bulk->phys;
 | |
| 	int i, ret;
 | |
| 
 | |
| 	for (i = 0; i < bulk->count; i++) {
 | |
| 		ret = generic_phy_power_on(&phys[i]);
 | |
| 		if (ret) {
 | |
| 			pr_err("Can't power on PHY%d\n", i);
 | |
| 			goto phys_poweron_err;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| phys_poweron_err:
 | |
| 	for (; i > 0; i--)
 | |
| 		generic_phy_power_off(&phys[i - 1]);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_phy_power_off_bulk(struct phy_bulk *bulk)
 | |
| {
 | |
| 	struct phy *phys = bulk->phys;
 | |
| 	int i, ret = 0;
 | |
| 
 | |
| 	for (i = 0; i < bulk->count; i++)
 | |
| 		ret |= generic_phy_power_off(&phys[i]);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_setup_phy(struct udevice *dev, struct phy *phy, int index)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	if (!phy)
 | |
| 		return 0;
 | |
| 
 | |
| 	ret = generic_phy_get_by_index(dev, index, phy);
 | |
| 	if (ret) {
 | |
| 		if (ret != -ENOENT)
 | |
| 			return ret;
 | |
| 	} else {
 | |
| 		ret = generic_phy_init(phy);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 
 | |
| 		ret = generic_phy_power_on(phy);
 | |
| 		if (ret)
 | |
| 			ret = generic_phy_exit(phy);
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int generic_shutdown_phy(struct phy *phy)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	if (!phy)
 | |
| 		return 0;
 | |
| 
 | |
| 	if (generic_phy_valid(phy)) {
 | |
| 		ret = generic_phy_power_off(phy);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 
 | |
| 		ret = generic_phy_exit(phy);
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| UCLASS_DRIVER(phy) = {
 | |
| 	.id		= UCLASS_PHY,
 | |
| 	.name		= "phy",
 | |
| 	.pre_probe	= phy_uclass_pre_probe,
 | |
| 	.pre_remove	= phy_uclass_pre_remove,
 | |
| 	.per_device_auto = sizeof(struct list_head),
 | |
| };
 |