From: Konrad Dybcio <konrad.dybcio@linaro.org>
To: Robert Marko <robimarko@gmail.com>,
ilia.lin@kernel.org, agross@kernel.org, andersson@kernel.org,
rafael@kernel.org, viresh.kumar@linaro.org, robh+dt@kernel.org,
krzysztof.kozlowski+dt@linaro.org, conor+dt@kernel.org,
linux-pm@vger.kernel.org, linux-arm-msm@vger.kernel.org,
devicetree@vger.kernel.org, linux-kernel@vger.kernel.org
Cc: ansuelsmth@gmail.com
Subject: Re: [PATCH 2/2] cpufreq: qcom-nvmem: add support for IPQ8074
Date: Sat, 27 May 2023 17:56:06 +0200 [thread overview]
Message-ID: <b227820b-239d-32fa-4dd4-78cba2a148e0@linaro.org> (raw)
In-Reply-To: <20230527095229.12019-2-robimarko@gmail.com>
On 27.05.2023 11:52, Robert Marko wrote:
> IPQ8074 comes in 2 families:
> * IPQ8070A/IPQ8071A (Acorn) up to 1.4GHz
> * IPQ8072A/IPQ8074A/IPQ8076A/IPQ8078A (Hawkeye) up to 2.2GHz
>
> So, in order to be able to share one OPP table lets add support for IPQ8074
> family based of SMEM SoC ID-s as speedbin fuse is always 0 on IPQ8074.
>
> IPQ8074 compatible is blacklisted from DT platdev as the cpufreq device
> will get created by NVMEM CPUFreq driver.
>
> Signed-off-by: Robert Marko <robimarko@gmail.com>
> ---
> drivers/cpufreq/cpufreq-dt-platdev.c | 1 +
> drivers/cpufreq/qcom-cpufreq-nvmem.c | 40 ++++++++++++++++++++++++++++
> 2 files changed, 41 insertions(+)
>
> diff --git a/drivers/cpufreq/cpufreq-dt-platdev.c b/drivers/cpufreq/cpufreq-dt-platdev.c
> index 14aa8281c7f4..e4d6d128647d 100644
> --- a/drivers/cpufreq/cpufreq-dt-platdev.c
> +++ b/drivers/cpufreq/cpufreq-dt-platdev.c
> @@ -169,6 +169,7 @@ static const struct of_device_id blocklist[] __initconst = {
> { .compatible = "ti,am625", },
>
> { .compatible = "qcom,ipq8064", },
> + { .compatible = "qcom,ipq8074", },
> { .compatible = "qcom,apq8064", },
> { .compatible = "qcom,msm8974", },
> { .compatible = "qcom,msm8960", },
> diff --git a/drivers/cpufreq/qcom-cpufreq-nvmem.c b/drivers/cpufreq/qcom-cpufreq-nvmem.c
> index a88b6fe5db50..607fc0273e9c 100644
> --- a/drivers/cpufreq/qcom-cpufreq-nvmem.c
> +++ b/drivers/cpufreq/qcom-cpufreq-nvmem.c
> @@ -31,6 +31,9 @@
>
> #include <dt-bindings/arm/qcom,ids.h>
>
> +#define IPQ8074_HAWKEYE_VERSION BIT(0)
> +#define IPQ8074_ACORN_VERSION BIT(1)
> +
> struct qcom_cpufreq_drv;
>
> struct qcom_cpufreq_match_data {
> @@ -204,6 +207,38 @@ static int qcom_cpufreq_krait_name_version(struct device *cpu_dev,
> return ret;
> }
>
> +static int qcom_cpufreq_ipq8074_name_version(struct device *cpu_dev,
> + struct nvmem_cell *speedbin_nvmem,
> + char **pvs_name,
> + struct qcom_cpufreq_drv *drv)
> +{
> + u32 msm_id;
> + int ret;
> + *pvs_name = NULL;
> +
> + ret = qcom_smem_get_soc_id(&msm_id);
> + if (ret)
> + return ret;
> +
> + switch (msm_id) {
> + case QCOM_ID_IPQ8070A:
> + case QCOM_ID_IPQ8071A:
> + drv->versions = IPQ8074_ACORN_VERSION;
> + break;
> + case QCOM_ID_IPQ8072A:
> + case QCOM_ID_IPQ8074A:
> + case QCOM_ID_IPQ8076A:
> + case QCOM_ID_IPQ8078A:
> + drv->versions = IPQ8074_HAWKEYE_VERSION;
> + break;
> + default:
> + BUG();
I'd say pr_err, or at least WARN() + setting the slowest bin would be
more desirable here, cpufreq probes early so people without uart etc.
will be unlikely to find out why their kernel dies.
Konrad
> + break;
> + }
> +
> + return 0;
> +}
> +
> static const struct qcom_cpufreq_match_data match_data_kryo = {
> .get_version = qcom_cpufreq_kryo_name_version,
> };
> @@ -218,6 +253,10 @@ static const struct qcom_cpufreq_match_data match_data_qcs404 = {
> .genpd_names = qcs404_genpd_names,
> };
>
> +static const struct qcom_cpufreq_match_data match_data_ipq8074 = {
> + .get_version = qcom_cpufreq_ipq8074_name_version,
> +};
> +
> static int qcom_cpufreq_probe(struct platform_device *pdev)
> {
> struct qcom_cpufreq_drv *drv;
> @@ -363,6 +402,7 @@ static const struct of_device_id qcom_cpufreq_match_list[] __initconst = {
> { .compatible = "qcom,msm8996", .data = &match_data_kryo },
> { .compatible = "qcom,qcs404", .data = &match_data_qcs404 },
> { .compatible = "qcom,ipq8064", .data = &match_data_krait },
> + { .compatible = "qcom,ipq8074", .data = &match_data_ipq8074 },
> { .compatible = "qcom,apq8064", .data = &match_data_krait },
> { .compatible = "qcom,msm8974", .data = &match_data_krait },
> { .compatible = "qcom,msm8960", .data = &match_data_krait },
next prev parent reply other threads:[~2023-05-27 15:56 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-05-27 9:52 [PATCH 1/2] dt-bindings: cpufreq: qcom-cpufreq-nvmem: document IPQ8074 Robert Marko
2023-05-27 9:52 ` [PATCH 2/2] cpufreq: qcom-nvmem: add support for IPQ8074 Robert Marko
2023-05-27 15:56 ` Konrad Dybcio [this message]
2023-05-27 10:30 ` [PATCH 1/2] dt-bindings: cpufreq: qcom-cpufreq-nvmem: document IPQ8074 Conor Dooley
2023-05-29 5:07 ` Viresh Kumar
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=b227820b-239d-32fa-4dd4-78cba2a148e0@linaro.org \
--to=konrad.dybcio@linaro.org \
--cc=agross@kernel.org \
--cc=andersson@kernel.org \
--cc=ansuelsmth@gmail.com \
--cc=conor+dt@kernel.org \
--cc=devicetree@vger.kernel.org \
--cc=ilia.lin@kernel.org \
--cc=krzysztof.kozlowski+dt@linaro.org \
--cc=linux-arm-msm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-pm@vger.kernel.org \
--cc=rafael@kernel.org \
--cc=robh+dt@kernel.org \
--cc=robimarko@gmail.com \
--cc=viresh.kumar@linaro.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).