From: Heiko Stuebner <heiko@sntech.de>
To: Elaine Zhang <zhangqing@rock-chips.com>
Cc: khilman@baylibre.com, xf@rock-chips.com, wxt@rock-chips.com,
linux-arm-kernel@lists.infradead.org, huangtao@rock-chips.com,
zyw@rock-chips.com, xxx@rock-chips.com, jay.xu@rock-chips.com,
linux-rockchip@lists.infradead.org, linux-kernel@vger.kernel.org
Subject: Re: [PATCH v1 2/2] rockchip: power-domain: support qos save and restore
Date: Thu, 31 Mar 2016 18:31:46 +0200 [thread overview]
Message-ID: <6919893.LfaTZNRxZs@phil> (raw)
In-Reply-To: <1458285444-31129-3-git-send-email-zhangqing@rock-chips.com>
Hi Elaine,
Am Freitag, 18. März 2016, 15:17:24 schrieb Elaine Zhang:
> support qos save and restore when power domain on/off.
>
> Signed-off-by: Elaine Zhang <zhangqing@rock-chips.com>
overall looks nice already ... some implementation-specific comments below.
> ---
> drivers/soc/rockchip/pm_domains.c | 87
> +++++++++++++++++++++++++++++++++++++-- 1 file changed, 84 insertions(+),
> 3 deletions(-)
>
> diff --git a/drivers/soc/rockchip/pm_domains.c
> b/drivers/soc/rockchip/pm_domains.c index 18aee6b..c5f4be6 100644
> --- a/drivers/soc/rockchip/pm_domains.c
> +++ b/drivers/soc/rockchip/pm_domains.c
> @@ -45,10 +45,21 @@ struct rockchip_pmu_info {
> const struct rockchip_domain_info *domain_info;
> };
>
> +#define MAX_QOS_NODE_NUM 20
> +#define MAX_QOS_REGS_NUM 5
> +#define QOS_PRIORITY 0x08
> +#define QOS_MODE 0x0c
> +#define QOS_BANDWIDTH 0x10
> +#define QOS_SATURATION 0x14
> +#define QOS_EXTCONTROL 0x18
> +
> struct rockchip_pm_domain {
> struct generic_pm_domain genpd;
> const struct rockchip_domain_info *info;
> struct rockchip_pmu *pmu;
> + int num_qos;
> + struct regmap *qos_regmap[MAX_QOS_NODE_NUM];
> + u32 qos_save_regs[MAX_QOS_NODE_NUM][MAX_QOS_REGS_NUM];
struct regmap **qos_regmap;
u32 *qos_save_regs;
> int num_clks;
> struct clk *clks[];
> };
> @@ -111,6 +122,55 @@ static int rockchip_pmu_set_idle_request(struct
> rockchip_pm_domain *pd, return 0;
> }
>
> +static int rockchip_pmu_save_qos(struct rockchip_pm_domain *pd)
> +{
> + int i;
> +
> + for (i = 0; i < pd->num_qos; i++) {
> + regmap_read(pd->qos_regmap[i],
> + QOS_PRIORITY,
> + &pd->qos_save_regs[i][0]);
> + regmap_read(pd->qos_regmap[i],
> + QOS_MODE,
> + &pd->qos_save_regs[i][1]);
> + regmap_read(pd->qos_regmap[i],
> + QOS_BANDWIDTH,
> + &pd->qos_save_regs[i][2]);
> + regmap_read(pd->qos_regmap[i],
> + QOS_SATURATION,
> + &pd->qos_save_regs[i][3]);
> + regmap_read(pd->qos_regmap[i],
> + QOS_EXTCONTROL,
> + &pd->qos_save_regs[i][4]);
> + }
> + return 0;
> +}
> +
> +static int rockchip_pmu_restore_qos(struct rockchip_pm_domain *pd)
> +{
> + int i;
> +
> + for (i = 0; i < pd->num_qos; i++) {
> + regmap_write(pd->qos_regmap[i],
> + QOS_PRIORITY,
> + pd->qos_save_regs[i][0]);
> + regmap_write(pd->qos_regmap[i],
> + QOS_MODE,
> + pd->qos_save_regs[i][1]);
> + regmap_write(pd->qos_regmap[i],
> + QOS_BANDWIDTH,
> + pd->qos_save_regs[i][2]);
> + regmap_write(pd->qos_regmap[i],
> + QOS_SATURATION,
> + pd->qos_save_regs[i][3]);
> + regmap_write(pd->qos_regmap[i],
> + QOS_EXTCONTROL,
> + pd->qos_save_regs[i][4]);
> + }
> +
> + return 0;
> +}
> +
> static bool rockchip_pmu_domain_is_on(struct rockchip_pm_domain *pd)
> {
> struct rockchip_pmu *pmu = pd->pmu;
> @@ -147,7 +207,7 @@ static int rockchip_pd_power(struct rockchip_pm_domain
> *pd, bool power_on) clk_enable(pd->clks[i]);
>
> if (!power_on) {
> - /* FIXME: add code to save AXI_QOS */
> + rockchip_pmu_save_qos(pd);
>
> /* if powering down, idle request to NIU first */
> rockchip_pmu_set_idle_request(pd, true);
> @@ -159,7 +219,7 @@ static int rockchip_pd_power(struct rockchip_pm_domain
> *pd, bool power_on) /* if powering up, leave idle mode */
> rockchip_pmu_set_idle_request(pd, false);
>
> - /* FIXME: add code to restore AXI_QOS */
> + rockchip_pmu_restore_qos(pd);
> }
>
> for (i = pd->num_clks - 1; i >= 0; i--)
> @@ -227,9 +287,10 @@ static int rockchip_pm_add_one_domain(struct
> rockchip_pmu *pmu, {
> const struct rockchip_domain_info *pd_info;
> struct rockchip_pm_domain *pd;
> + struct device_node *qos_node;
> struct clk *clk;
> int clk_cnt;
> - int i;
> + int i, j;
> u32 id;
> int error;
>
> @@ -289,6 +350,26 @@ static int rockchip_pm_add_one_domain(struct
> rockchip_pmu *pmu, clk, node->name);
> }
>
> + pd->num_qos = of_count_phandle_with_args(node, "pm_qos",
> + NULL);
missing error handling here:
if (pd->num_qos < 0) {
error = pd->num_qos;
goto err_out;
}
Right now, you always allocate MAX_QOS_NODE_NUM entries for regmaps and
registers for each domain - a bit of a waste over all domains, so maybe
like:
pd->qos_regmap = kcalloc(pd->num_qos, sizeof(*pd->qos_regmap), GFP_KERNEL);
pd->qos_save_regs = kcalloc, pd->num_qos * MAX_QOS_REGS_NUM, sizeof(u32),
GFP_KERNEL);
+ of course error handling for both + cleanup in rockchip_remove_one_domain
> +
> + for (j = 0; j < pd->num_qos; j++) {
> + qos_node = of_parse_phandle(node, "pm_qos", j);
> + if (!qos_node) {
> + error = -ENODEV;
> + goto err_out;
> + }
> + pd->qos_regmap[j] = syscon_node_to_regmap(qos_node);
missing
if (IS_ERR(pd->qos_regmap[j])) { ...}
> + of_node_put(qos_node);
> + }
> +
> error = rockchip_pd_power(pd, true);
> if (error) {
> dev_err(pmu->dev,
next prev parent reply other threads:[~2016-03-31 16:32 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2016-03-18 7:17 [PATCH v1 0/2] rockchip: power-domain: support qos save and restore Elaine Zhang
2016-03-18 7:17 ` [PATCH v1 1/2] dt-bindings: modify document of Rockchip power domains Elaine Zhang
2016-03-18 16:18 ` Kevin Hilman
2016-03-18 22:16 ` Heiko Stuebner
2016-04-12 2:00 ` Heiko Stuebner
2016-03-18 7:17 ` [PATCH v1 2/2] rockchip: power-domain: support qos save and restore Elaine Zhang
2016-03-31 16:31 ` Heiko Stuebner [this message]
2016-04-01 2:33 ` Elaine Zhang
2016-04-01 16:19 ` Heiko Stuebner
2016-04-05 1:57 ` Elaine Zhang
2016-04-05 17:26 ` Heiko Stuebner
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=6919893.LfaTZNRxZs@phil \
--to=heiko@sntech.de \
--cc=huangtao@rock-chips.com \
--cc=jay.xu@rock-chips.com \
--cc=khilman@baylibre.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-rockchip@lists.infradead.org \
--cc=wxt@rock-chips.com \
--cc=xf@rock-chips.com \
--cc=xxx@rock-chips.com \
--cc=zhangqing@rock-chips.com \
--cc=zyw@rock-chips.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox