linux-arm-kernel.lists.infradead.org archive mirror
 help / color / mirror / Atom feed
From: Tanmay Shah <tanmay.shah@amd.com>
To: Mathieu Poirier <mathieu.poirier@linaro.org>
Cc: andersson@kernel.org, robh+dt@kernel.org,
	krzysztof.kozlowski+dt@linaro.org, michal.simek@amd.com,
	radhey.shyam.pandey@amd.com, ben.levinsky@amd.com,
	linux-remoteproc@vger.kernel.org, devicetree@vger.kernel.org,
	linux-arm-kernel@lists.infradead.org
Subject: Re: [PATCH v5 3/4] remoteproc: zynqmp: add pm domains support
Date: Mon, 2 Oct 2023 15:39:14 -0500	[thread overview]
Message-ID: <4cd0d913-dabc-41a7-a71d-8d48ca68995d@amd.com> (raw)
In-Reply-To: <ZRr5xI1w4Pd3/8+S@p14s>

Hello Mathiew,

Thanks for reviews.

On 10/2/23 12:11 PM, Mathieu Poirier wrote:
> On Thu, Sep 28, 2023 at 08:58:59AM -0700, Tanmay Shah wrote:
> > Use TCM pm domains extracted from device-tree
> > to power on/off TCM using general pm domain framework
> > 
> > Signed-off-by: Tanmay Shah <tanmay.shah@amd.com>
> > ---
> >  drivers/remoteproc/xlnx_r5_remoteproc.c | 224 ++++++++++++++++++++++--
> >  1 file changed, 205 insertions(+), 19 deletions(-)
> > 
> > diff --git a/drivers/remoteproc/xlnx_r5_remoteproc.c b/drivers/remoteproc/xlnx_r5_remoteproc.c
> > index 4395edea9a64..27ed2c070ebb 100644
> > --- a/drivers/remoteproc/xlnx_r5_remoteproc.c
> > +++ b/drivers/remoteproc/xlnx_r5_remoteproc.c
> > @@ -16,6 +16,7 @@
> >  #include <linux/of_reserved_mem.h>
> >  #include <linux/platform_device.h>
> >  #include <linux/remoteproc.h>
> > +#include <linux/pm_domain.h>
> >  
> >  #include "remoteproc_internal.h"
> >  
> > @@ -102,6 +103,12 @@ static const struct mem_bank_data zynqmp_tcm_banks_lockstep[] = {
> >   * @rproc: rproc handle
> >   * @pm_domain_id: RPU CPU power domain id
> >   * @ipi: pointer to mailbox information
> > + * @num_pm_dev: number of tcm pm domain devices for this core
> > + * @pm_dev1: pm domain virtual devices for power domain framework
> > + * @pm_dev_link1: pm domain device links after registration
> > + * @pm_dev2: used only in lockstep mode. second core's pm domain virtual devices
> > + * @pm_dev_link2: used only in lockstep mode. second core's pm device links after
> > + * registration
> >   */
> >  struct zynqmp_r5_core {
> >  	struct device *dev;
> > @@ -111,6 +118,11 @@ struct zynqmp_r5_core {
> >  	struct rproc *rproc;
> >  	u32 pm_domain_id;
> >  	struct mbox_info *ipi;
> > +	int num_pm_dev;
> > +	struct device **pm_dev1;
> > +	struct device_link **pm_dev_link1;
> > +	struct device **pm_dev2;
> > +	struct device_link **pm_dev_link2;
> >  };
> >  
> >  /**
> > @@ -575,12 +587,21 @@ static int add_tcm_carveout_split_mode(struct rproc *rproc)
> >  		bank_size = r5_core->tcm_banks[i]->size;
> >  		pm_domain_id = r5_core->tcm_banks[i]->pm_domain_id;
> >  
> > -		ret = zynqmp_pm_request_node(pm_domain_id,
> > -					     ZYNQMP_PM_CAPABILITY_ACCESS, 0,
> > -					     ZYNQMP_PM_REQUEST_ACK_BLOCKING);
> > -		if (ret < 0) {
> > -			dev_err(dev, "failed to turn on TCM 0x%x", pm_domain_id);
> > -			goto release_tcm_split;
> > +		/*
> > +		 * If TCM information is available in device-tree then
> > +		 * in that case, pm domain framework will power on/off TCM.
> > +		 * In that case pm_domain_id is set to 0. If hardcode
> > +		 * bindings from driver is used, then only this driver will
> > +		 * use pm_domain_id.
> > +		 */
> > +		if (pm_domain_id) {
> > +			ret = zynqmp_pm_request_node(pm_domain_id,
> > +						     ZYNQMP_PM_CAPABILITY_ACCESS, 0,
> > +						     ZYNQMP_PM_REQUEST_ACK_BLOCKING);
> > +			if (ret < 0) {
> > +				dev_err(dev, "failed to turn on TCM 0x%x", pm_domain_id);
> > +				goto release_tcm_split;
> > +			}
> >  		}
> >  
> >  		dev_dbg(dev, "TCM carveout split mode %s addr=%llx, da=0x%x, size=0x%lx",
> > @@ -646,13 +667,16 @@ static int add_tcm_carveout_lockstep_mode(struct rproc *rproc)
> >  	for (i = 0; i < num_banks; i++) {
> >  		pm_domain_id = r5_core->tcm_banks[i]->pm_domain_id;
> >  
> > -		/* Turn on each TCM bank individually */
> > -		ret = zynqmp_pm_request_node(pm_domain_id,
> > -					     ZYNQMP_PM_CAPABILITY_ACCESS, 0,
> > -					     ZYNQMP_PM_REQUEST_ACK_BLOCKING);
> > -		if (ret < 0) {
> > -			dev_err(dev, "failed to turn on TCM 0x%x", pm_domain_id);
> > -			goto release_tcm_lockstep;
> > +		if (pm_domain_id) {
> > +			/* Turn on each TCM bank individually */
> > +			ret = zynqmp_pm_request_node(pm_domain_id,
> > +						     ZYNQMP_PM_CAPABILITY_ACCESS, 0,
> > +						     ZYNQMP_PM_REQUEST_ACK_BLOCKING);
> > +			if (ret < 0) {
> > +				dev_err(dev, "failed to turn on TCM 0x%x",
> > +					pm_domain_id);
> > +				goto release_tcm_lockstep;
> > +			}
> >  		}
> >  
> >  		bank_size = r5_core->tcm_banks[i]->size;
> > @@ -687,8 +711,10 @@ static int add_tcm_carveout_lockstep_mode(struct rproc *rproc)
> >  	/* If failed, Turn off all TCM banks turned on before */
> >  	for (i--; i >= 0; i--) {
> >  		pm_domain_id = r5_core->tcm_banks[i]->pm_domain_id;
> > -		zynqmp_pm_release_node(pm_domain_id);
> > +		if (pm_domain_id)
> > +			zynqmp_pm_release_node(pm_domain_id);
> >  	}
> > +
>
> Spurious change

Ack, will remove it.

>
> >  	return ret;
> >  }
> >  
> > @@ -758,6 +784,153 @@ static int zynqmp_r5_parse_fw(struct rproc *rproc, const struct firmware *fw)
> >  	return ret;
> >  }
> >  
> > +static void zynqmp_r5_remove_pm_domains(struct rproc *rproc)
> > +{
> > +	struct zynqmp_r5_core *r5_core = rproc->priv;
> > +	struct device *dev = r5_core->dev;
> > +	struct zynqmp_r5_cluster *cluster;
> > +	int i;
> > +
> > +	cluster = platform_get_drvdata(to_platform_device(dev->parent));
> > +
> > +	for (i = 0; i < r5_core->num_pm_dev; i++) {
> > +		if (r5_core->pm_dev_link1 && r5_core->pm_dev_link1[i])
> > +			device_link_del(r5_core->pm_dev_link1[i]);
> > +		if (r5_core->pm_dev1 && !IS_ERR_OR_NULL(r5_core->pm_dev1[i]))
> > +			dev_pm_domain_detach(r5_core->pm_dev1[i], false);
> > +	}
> > +
>
> A global function such as this one should not have to deal with error
> conditions.  Those should be dealt with in the allocation function where cleanup
> is done on error conditions.  See my comment below in
> zynqmp_r5_add_pm_domains().

Ack, I won't use IS_ERR_OR_NULL here. Instead will make sure that if this function is called,

all the pm domains are acquired successfully.

> > +	kfree(r5_core->pm_dev1);
> > +	r5_core->pm_dev1 = NULL;
> > +	kfree(r5_core->pm_dev_link1);
> > +	r5_core->pm_dev_link1 = NULL;
> > +
> > +	if (cluster->mode == SPLIT_MODE) {
> > +		r5_core->num_pm_dev = 0;
> > +		return;
> > +	}
> > +
> > +	for (i = 0; i < r5_core->num_pm_dev; i++) {
> > +		if (r5_core->pm_dev_link2 && r5_core->pm_dev_link2[i])
> > +			device_link_del(r5_core->pm_dev_link2[i]);
> > +		if (r5_core->pm_dev2 && !IS_ERR_OR_NULL(r5_core->pm_dev2[i]))
> > +			dev_pm_domain_detach(r5_core->pm_dev2[i], false);
> > +	}
> > +
> > +	kfree(r5_core->pm_dev2);
> > +	r5_core->pm_dev2 = NULL;
> > +	kfree(r5_core->pm_dev_link2);
> > +	r5_core->pm_dev_link2 = NULL;
> > +	r5_core->num_pm_dev = 0;
> > +}
> > +
> > +static int zynqmp_r5_add_pm_domains(struct rproc *rproc)
> > +{
> > +	struct zynqmp_r5_core *r5_core = rproc->priv;
> > +	struct device *dev = r5_core->dev, *dev2;
> > +	struct zynqmp_r5_cluster *cluster;
> > +	struct platform_device *pdev;
> > +	struct device_node *np;
> > +	int i, num_pm_dev, ret;
>
> I'm not sure 'ret' is needed - see below.

You are right. It's not needed. will fix it.


>
> > +
> > +	cluster = platform_get_drvdata(to_platform_device(dev->parent));
>
> Why not use dev_get_drvdata() as it is done elsewhere in this driver?
>
> > +
> > +	/* get number of power-domains */
> > +	num_pm_dev = of_count_phandle_with_args(r5_core->np, "power-domains",
> > +						"#power-domain-cells");
> > +
> > +	if (num_pm_dev <= 0)
> > +		return -EINVAL;
> > +
> > +	r5_core->pm_dev1 = kcalloc(num_pm_dev,
> > +				   sizeof(struct device *),
> > +				   GFP_KERNEL);
> > +	if (!r5_core->pm_dev1)
> > +		ret = -ENOMEM;
>
> What's the goal of the assignment?  Did you mean to return an error instead?

Ack, it should have been return. I implemented multiple ways to return errors, and missed to fix this before sending the patch.

>
> > +
> > +	r5_core->pm_dev_link1 = kcalloc(num_pm_dev,
> > +					sizeof(struct device_link *),
> > +					GFP_KERNEL);
> > +	if (!r5_core->pm_dev_link1)
> > +		return -ENOMEM;
>
> In case of error, always cleanup the work done in the current function.  That
> way cleanup functions such as zynqmp_r5_remove_pm_domains() are simple and easy
> to read.

Ack. I will make sure that if we fail here for any reason, all the memory allocated is freed

here. Also any acquired devices and pm domains should be detached and released in this function

only upon failure in between.

Thanks,

Tanmay

>
> > +
> > +	r5_core->num_pm_dev = num_pm_dev;
> > +
> > +	/* for zynqmp we only add TCM power domains and not core's power domain */
> > +	for (i = 1; i < r5_core->num_pm_dev; i++) {
> > +		r5_core->pm_dev1[i] = dev_pm_domain_attach_by_id(dev, i);
> > +		if (IS_ERR(r5_core->pm_dev1[i])) {
> > +			dev_dbg(dev, "failed to attach pm domain %d\n", i);
> > +			return PTR_ERR(r5_core->pm_dev1[i]);
> > +		}
> > +		if (!r5_core->pm_dev1[i]) {
> > +			dev_dbg(dev, "can't attach to pm domain %d\n", i);
> > +			return -EINVAL;
> > +		}
> > +
> > +		r5_core->pm_dev_link1[i] = device_link_add(dev, r5_core->pm_dev1[i],
> > +							   DL_FLAG_STATELESS |
> > +							   DL_FLAG_RPM_ACTIVE |
> > +							   DL_FLAG_PM_RUNTIME);
> > +		if (!r5_core->pm_dev_link1[i]) {
> > +			dev_pm_domain_detach(r5_core->pm_dev1[i], true);
> > +			r5_core->pm_dev1[i] = NULL;
> > +			return -EINVAL;
> > +		}
> > +	}
> > +
> > +	if (cluster->mode == SPLIT_MODE)
> > +		return 0;
> > +
> > +	r5_core->pm_dev2 = kcalloc(num_pm_dev,
> > +				   sizeof(struct device *),
> > +				   GFP_KERNEL);
> > +	if (!r5_core->pm_dev2)
> > +		return -ENOMEM;
> > +
> > +	r5_core->pm_dev_link2 = kcalloc(num_pm_dev,
> > +					sizeof(struct device_link *),
> > +					GFP_KERNEL);
> > +	if (!r5_core->pm_dev_link2)
> > +		return -ENOMEM;
> > +
> > +	/* get second core's device to detach its power-domains */
> > +	np = of_get_next_child(cluster->dev->of_node, of_node_get(dev->of_node));
> > +
> > +	pdev = of_find_device_by_node(np);
> > +	if (!pdev) {
> > +		dev_err(cluster->dev, "core1 platform device not available\n");
> > +		return -EINVAL;
> > +	}
> > +
> > +	dev2 = &pdev->dev;
> > +
> > +	/* for zynqmp we only add TCM power domains and not core's power domain */
> > +	for (i = 1; i < r5_core->num_pm_dev; i++) {
> > +		r5_core->pm_dev2[i] = dev_pm_domain_attach_by_id(dev2, i);
> > +		if (IS_ERR(r5_core->pm_dev2[i])) {
> > +			dev_dbg(dev, "can't attach to pm domain %d\n", i);
> > +			return PTR_ERR(r5_core->pm_dev2[i]);
> > +		}
> > +		if (!r5_core->pm_dev2[i]) {
> > +			dev_dbg(dev, "can't attach to pm domain %d\n", i);
> > +			return -EINVAL;
> > +		}
> > +
> > +		r5_core->pm_dev_link2[i] = device_link_add(dev, r5_core->pm_dev2[i],
> > +							   DL_FLAG_STATELESS |
> > +							   DL_FLAG_RPM_ACTIVE |
> > +							   DL_FLAG_PM_RUNTIME);
> > +		if (!r5_core->pm_dev_link2[i]) {
> > +			dev_pm_domain_detach(r5_core->pm_dev2[i], true);
> > +			r5_core->pm_dev2[i] = NULL;
> > +			return -ENODEV;
> > +		}
> > +	}
> > +
> > +	return 0;
> > +}
> > +
> >  /**
> >   * zynqmp_r5_rproc_prepare()
> >   * adds carveouts for TCM bank and reserved memory regions
> > @@ -770,19 +943,30 @@ static int zynqmp_r5_rproc_prepare(struct rproc *rproc)
> >  {
> >  	int ret;
> >  
> > +	ret = zynqmp_r5_add_pm_domains(rproc);
> > +	if (ret) {
> > +		dev_err(&rproc->dev, "failed to add pm domains\n");
> > +		goto fail_prepare;
> > +	}
> > +
> >  	ret = add_tcm_banks(rproc);
> >  	if (ret) {
> >  		dev_err(&rproc->dev, "failed to get TCM banks, err %d\n", ret);
> > -		return ret;
> > +		goto fail_prepare;
> >  	}
> >  
> >  	ret = add_mem_regions_carveout(rproc);
> >  	if (ret) {
> >  		dev_err(&rproc->dev, "failed to get reserve mem regions %d\n", ret);
> > -		return ret;
> > +		goto fail_prepare;
> >  	}
> >  
> >  	return 0;
> > +
> > +fail_prepare:
> > +	zynqmp_r5_remove_pm_domains(rproc);
> > +
> > +	return ret;
> >  }
> >  
> >  /**
> > @@ -801,11 +985,13 @@ static int zynqmp_r5_rproc_unprepare(struct rproc *rproc)
> >  
> >  	r5_core = rproc->priv;
> >  
> > +	zynqmp_r5_remove_pm_domains(rproc);
> > +
> >  	for (i = 0; i < r5_core->tcm_bank_count; i++) {
> >  		pm_domain_id = r5_core->tcm_banks[i]->pm_domain_id;
> > -		if (zynqmp_pm_release_node(pm_domain_id))
> > -			dev_warn(r5_core->dev,
> > -				 "can't turn off TCM bank 0x%x", pm_domain_id);
> > +		if (pm_domain_id && zynqmp_pm_release_node(pm_domain_id))
> > +			dev_dbg(r5_core->dev,
> > +				"can't turn off TCM bank 0x%x", pm_domain_id);
> >  	}
> >  
> >  	return 0;
> > -- 
> > 2.25.1
> > 

_______________________________________________
linux-arm-kernel mailing list
linux-arm-kernel@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-arm-kernel

  reply	other threads:[~2023-10-02 20:40 UTC|newest]

Thread overview: 15+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-09-28 15:58 [PATCH v5 0/4] add zynqmp TCM bindings Tanmay Shah
2023-09-28 15:58 ` [PATCH v5 1/4] dt-bindings: remoteproc: add Tightly Coupled Memory (TCM) bindings Tanmay Shah
2023-09-28 15:58 ` [PATCH v5 2/4] dts: zynqmp: add properties for TCM in remoteproc Tanmay Shah
2023-10-02 15:55   ` Mathieu Poirier
2023-10-02 16:25     ` Tanmay Shah
2023-10-02 17:12       ` Tanmay Shah
2023-10-02 20:17         ` Mathieu Poirier
2023-10-02 20:54           ` Tanmay Shah
2023-10-03 15:31             ` Mathieu Poirier
2023-10-03 16:32               ` Tanmay Shah
2023-09-28 15:58 ` [PATCH v5 3/4] remoteproc: zynqmp: add pm domains support Tanmay Shah
2023-10-02 17:11   ` Mathieu Poirier
2023-10-02 20:39     ` Tanmay Shah [this message]
2023-09-28 15:59 ` [PATCH v5 4/4] remoteproc: zynqmp: parse TCM from device tree Tanmay Shah
2023-10-02 17:39   ` Mathieu Poirier

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4cd0d913-dabc-41a7-a71d-8d48ca68995d@amd.com \
    --to=tanmay.shah@amd.com \
    --cc=andersson@kernel.org \
    --cc=ben.levinsky@amd.com \
    --cc=devicetree@vger.kernel.org \
    --cc=krzysztof.kozlowski+dt@linaro.org \
    --cc=linux-arm-kernel@lists.infradead.org \
    --cc=linux-remoteproc@vger.kernel.org \
    --cc=mathieu.poirier@linaro.org \
    --cc=michal.simek@amd.com \
    --cc=radhey.shyam.pandey@amd.com \
    --cc=robh+dt@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).