* [PATCH v2 01/11] opp: core: implement dev_pm_opp_get_bw
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-19 17:56 ` [PATCH v2 02/11] drm/msm: adreno: rename quirks that are features Neil Armstrong
` (9 subsequent siblings)
10 siblings, 0 replies; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
Add and implement the dev_pm_opp_get_bw() to retrieve
the OPP's bandwidth in the same way as the dev_pm_opp_get_voltage()
helper.
Retrieving bandwidth is required in the case of the Adreno GPU
where the GPU Management Unit can handle the Bandwidth scaling.
The helper can get the peak or average bandwidth for any of
the interconnect path.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/opp/core.c | 25 +++++++++++++++++++++++++
include/linux/pm_opp.h | 7 +++++++
2 files changed, 32 insertions(+)
diff --git a/drivers/opp/core.c b/drivers/opp/core.c
index 494f8860220d97fc690ebab5ed3b7f5f04f22d73..864b9b99b0129acaffaf45c584c5f34b8bababed 100644
--- a/drivers/opp/core.c
+++ b/drivers/opp/core.c
@@ -106,6 +106,31 @@ static bool assert_single_clk(struct opp_table *opp_table)
return !WARN_ON(opp_table->clk_count > 1);
}
+/**
+ * dev_pm_opp_get_bw() - Gets the bandwidth corresponding to an opp
+ * @opp: opp for which voltage has to be returned for
+ * @peak: select peak or average bandwidth
+ * @index: bandwidth index
+ *
+ * Return: bandwidth in kBps, else return 0
+ */
+unsigned long dev_pm_opp_get_bw(struct dev_pm_opp *opp, bool peak, int index)
+{
+ if (IS_ERR_OR_NULL(opp)) {
+ pr_err("%s: Invalid parameters\n", __func__);
+ return 0;
+ }
+
+ if (index > opp->opp_table->path_count)
+ return 0;
+
+ if (!opp->bandwidth)
+ return 0;
+
+ return peak ? opp->bandwidth[index].peak : opp->bandwidth[index].avg;
+}
+EXPORT_SYMBOL_GPL(dev_pm_opp_get_bw);
+
/**
* dev_pm_opp_get_voltage() - Gets the voltage corresponding to an opp
* @opp: opp for which voltage has to be returned for
diff --git a/include/linux/pm_opp.h b/include/linux/pm_opp.h
index 6424692c30b71fca471a1b7d63e018605dd9324b..cd9a257b8e7766d6c8631351a10a845c88414a74 100644
--- a/include/linux/pm_opp.h
+++ b/include/linux/pm_opp.h
@@ -106,6 +106,8 @@ struct dev_pm_opp_data {
struct opp_table *dev_pm_opp_get_opp_table(struct device *dev);
void dev_pm_opp_put_opp_table(struct opp_table *opp_table);
+unsigned long dev_pm_opp_get_bw(struct dev_pm_opp *opp, bool peak, int index);
+
unsigned long dev_pm_opp_get_voltage(struct dev_pm_opp *opp);
int dev_pm_opp_get_supplies(struct dev_pm_opp *opp, struct dev_pm_opp_supply *supplies);
@@ -209,6 +211,11 @@ static inline struct opp_table *dev_pm_opp_get_opp_table_indexed(struct device *
static inline void dev_pm_opp_put_opp_table(struct opp_table *opp_table) {}
+static inline unsigned long dev_pm_opp_get_bw(struct dev_pm_opp *opp, bool peak, int index)
+{
+ return 0;
+}
+
static inline unsigned long dev_pm_opp_get_voltage(struct dev_pm_opp *opp)
{
return 0;
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* [PATCH v2 02/11] drm/msm: adreno: rename quirks that are features
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
2024-11-19 17:56 ` [PATCH v2 01/11] opp: core: implement dev_pm_opp_get_bw Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-19 17:56 ` [PATCH v2 03/11] drm/msm: adreno: move features bits in a separate variable Neil Armstrong
` (8 subsequent siblings)
10 siblings, 0 replies; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
Half of the current "Quirks" are in fact features, so rename
the defines with FEAT instead of QUIRK.
They will be moved in a separate bitfield in a second time.
No functional changes.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/gpu/drm/msm/adreno/a6xx_catalog.c | 62 +++++++++++++++---------------
drivers/gpu/drm/msm/adreno/a6xx_gpu.c | 4 +-
drivers/gpu/drm/msm/adreno/adreno_device.c | 2 +-
drivers/gpu/drm/msm/adreno/adreno_gpu.h | 7 ++--
4 files changed, 38 insertions(+), 37 deletions(-)
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
index 0c560e84ad5a53bb4e8a49ba4e153ce9cf33f7ae..825c820def315968d508973c8ae40c7c7b646569 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
@@ -743,7 +743,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a615_zap.mbn",
.a6xx = &(const struct a6xx_info) {
@@ -769,7 +769,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.a6xx = &(const struct a6xx_info) {
.protect = &a630_protect,
@@ -839,7 +839,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a615_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -864,8 +864,8 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a620_zap.mbn",
.a6xx = &(const struct a6xx_info) {
@@ -892,7 +892,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a630_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -911,7 +911,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a640_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -934,8 +934,8 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M + SZ_128K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a650_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -961,8 +961,8 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M + SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a660_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -981,8 +981,8 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M + SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.a6xx = &(const struct a6xx_info) {
.hwcg = a690_hwcg,
@@ -1000,8 +1000,8 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a660_zap.mbn",
.a6xx = &(const struct a6xx_info) {
@@ -1028,7 +1028,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_2M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a640_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -1046,8 +1046,8 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_4M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a690_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -1331,7 +1331,7 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = SZ_128K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_HW_APRIV,
+ .quirks = ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a702_zap.mbn",
.a6xx = &(const struct a6xx_info) {
@@ -1355,9 +1355,9 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = SZ_2M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV |
- ADRENO_QUIRK_PREEMPTION,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV |
+ ADRENO_FEAT_PREEMPTION,
.init = a6xx_gpu_init,
.zapfw = "a730_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -1377,9 +1377,9 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = 3 * SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV |
- ADRENO_QUIRK_PREEMPTION,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV |
+ ADRENO_FEAT_PREEMPTION,
.init = a6xx_gpu_init,
.zapfw = "a740_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -1400,9 +1400,9 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = 3 * SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV |
- ADRENO_QUIRK_PREEMPTION,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV |
+ ADRENO_FEAT_PREEMPTION,
.init = a6xx_gpu_init,
.a6xx = &(const struct a6xx_info) {
.hwcg = a740_hwcg,
@@ -1422,9 +1422,9 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = 3 * SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_QUIRK_HAS_CACHED_COHERENT |
- ADRENO_QUIRK_HAS_HW_APRIV |
- ADRENO_QUIRK_PREEMPTION,
+ .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ ADRENO_FEAT_HAS_HW_APRIV |
+ ADRENO_FEAT_PREEMPTION,
.init = a6xx_gpu_init,
.zapfw = "gen70900_zap.mbn",
.a6xx = &(const struct a6xx_info) {
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
index 019610341df1506c89f44e86b8d1deeb27d61857..2ebd3fac212576a1507e0b6afe2560cd0408dd89 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
@@ -2478,7 +2478,7 @@ struct msm_gpu *a6xx_gpu_init(struct drm_device *dev)
adreno_gpu->gmu_is_wrapper = of_device_is_compatible(node, "qcom,adreno-gmu-wrapper");
adreno_gpu->base.hw_apriv =
- !!(config->info->quirks & ADRENO_QUIRK_HAS_HW_APRIV);
+ !!(config->info->quirks & ADRENO_FEAT_HAS_HW_APRIV);
/* gpu->info only gets assigned in adreno_gpu_init() */
is_a7xx = config->info->family == ADRENO_7XX_GEN1 ||
@@ -2495,7 +2495,7 @@ struct msm_gpu *a6xx_gpu_init(struct drm_device *dev)
}
if ((enable_preemption == 1) || (enable_preemption == -1 &&
- (config->info->quirks & ADRENO_QUIRK_PREEMPTION)))
+ (config->info->quirks & ADRENO_FEAT_PREEMPTION)))
ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_a7xx, 4);
else if (is_a7xx)
ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_a7xx, 1);
diff --git a/drivers/gpu/drm/msm/adreno/adreno_device.c b/drivers/gpu/drm/msm/adreno/adreno_device.c
index 9ffe91920fbfb4841b28aabec9fbde94539fdd83..09d4569f77528c2a20cabc814668c4c930dd07f1 100644
--- a/drivers/gpu/drm/msm/adreno/adreno_device.c
+++ b/drivers/gpu/drm/msm/adreno/adreno_device.c
@@ -207,7 +207,7 @@ static int adreno_bind(struct device *dev, struct device *master, void *data)
priv->is_a2xx = info->family < ADRENO_3XX;
priv->has_cached_coherent =
- !!(info->quirks & ADRENO_QUIRK_HAS_CACHED_COHERENT);
+ !!(info->quirks & ADRENO_FEAT_HAS_CACHED_COHERENT);
gpu = info->init(drm);
if (IS_ERR(gpu)) {
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
index e71f420f8b3a8e6cfc52dd1c4d5a63ef3704a07f..8782c25e8a393ec7d9dc23ad450908d039bd08c5 100644
--- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h
+++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
@@ -54,9 +54,10 @@ enum adreno_family {
#define ADRENO_QUIRK_TWO_PASS_USE_WFI BIT(0)
#define ADRENO_QUIRK_FAULT_DETECT_MASK BIT(1)
#define ADRENO_QUIRK_LMLOADKILL_DISABLE BIT(2)
-#define ADRENO_QUIRK_HAS_HW_APRIV BIT(3)
-#define ADRENO_QUIRK_HAS_CACHED_COHERENT BIT(4)
-#define ADRENO_QUIRK_PREEMPTION BIT(5)
+
+#define ADRENO_FEAT_HAS_HW_APRIV BIT(3)
+#define ADRENO_FEAT_HAS_CACHED_COHERENT BIT(4)
+#define ADRENO_FEAT_PREEMPTION BIT(5)
/* Helper for formating the chip_id in the way that userspace tools like
* crashdec expect.
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* [PATCH v2 03/11] drm/msm: adreno: move features bits in a separate variable
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
2024-11-19 17:56 ` [PATCH v2 01/11] opp: core: implement dev_pm_opp_get_bw Neil Armstrong
2024-11-19 17:56 ` [PATCH v2 02/11] drm/msm: adreno: rename quirks that are features Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-20 11:18 ` Dmitry Baryshkov
2024-11-19 17:56 ` [PATCH v2 04/11] drm/msm: adreno: add GMU_BW_VOTE feature flag Neil Armstrong
` (7 subsequent siblings)
10 siblings, 1 reply; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
Now the features defines have the right name, introduce a features
bitfield and move the features defines in it, fixing all code checking
for them.
No functional changes intended.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/gpu/drm/msm/adreno/a6xx_catalog.c | 34 +++++++++++++++---------------
drivers/gpu/drm/msm/adreno/a6xx_gpu.c | 4 ++--
drivers/gpu/drm/msm/adreno/adreno_device.c | 2 +-
drivers/gpu/drm/msm/adreno/adreno_gpu.h | 7 +++---
4 files changed, 24 insertions(+), 23 deletions(-)
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
index 825c820def315968d508973c8ae40c7c7b646569..93f0d4bf50ba773ecde93e6c29a2fcec24ebb7b3 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
@@ -743,7 +743,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a615_zap.mbn",
.a6xx = &(const struct a6xx_info) {
@@ -769,7 +769,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.a6xx = &(const struct a6xx_info) {
.protect = &a630_protect,
@@ -839,7 +839,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a615_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -864,7 +864,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a620_zap.mbn",
@@ -892,7 +892,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a630_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -911,7 +911,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a640_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -934,7 +934,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M + SZ_128K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a650_zap.mdt",
@@ -961,7 +961,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M + SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a660_zap.mdt",
@@ -981,7 +981,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_1M + SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.a6xx = &(const struct a6xx_info) {
@@ -1000,7 +1000,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_512K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a660_zap.mbn",
@@ -1028,7 +1028,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_2M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
.init = a6xx_gpu_init,
.zapfw = "a640_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -1046,7 +1046,7 @@ static const struct adreno_info a6xx_gpus[] = {
},
.gmem = SZ_4M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a690_zap.mdt",
@@ -1331,7 +1331,7 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = SZ_128K,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_HW_APRIV,
+ .features = ADRENO_FEAT_HAS_HW_APRIV,
.init = a6xx_gpu_init,
.zapfw = "a702_zap.mbn",
.a6xx = &(const struct a6xx_info) {
@@ -1355,7 +1355,7 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = SZ_2M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV |
ADRENO_FEAT_PREEMPTION,
.init = a6xx_gpu_init,
@@ -1377,7 +1377,7 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = 3 * SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV |
ADRENO_FEAT_PREEMPTION,
.init = a6xx_gpu_init,
@@ -1400,7 +1400,7 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = 3 * SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV |
ADRENO_FEAT_PREEMPTION,
.init = a6xx_gpu_init,
@@ -1422,7 +1422,7 @@ static const struct adreno_info a7xx_gpus[] = {
},
.gmem = 3 * SZ_1M,
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
- .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
+ .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV |
ADRENO_FEAT_PREEMPTION,
.init = a6xx_gpu_init,
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
index 2ebd3fac212576a1507e0b6afe2560cd0408dd89..654d0cff421f15901cd4bf33b41e70004634ec75 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
@@ -2478,7 +2478,7 @@ struct msm_gpu *a6xx_gpu_init(struct drm_device *dev)
adreno_gpu->gmu_is_wrapper = of_device_is_compatible(node, "qcom,adreno-gmu-wrapper");
adreno_gpu->base.hw_apriv =
- !!(config->info->quirks & ADRENO_FEAT_HAS_HW_APRIV);
+ !!(config->info->features & ADRENO_FEAT_HAS_HW_APRIV);
/* gpu->info only gets assigned in adreno_gpu_init() */
is_a7xx = config->info->family == ADRENO_7XX_GEN1 ||
@@ -2495,7 +2495,7 @@ struct msm_gpu *a6xx_gpu_init(struct drm_device *dev)
}
if ((enable_preemption == 1) || (enable_preemption == -1 &&
- (config->info->quirks & ADRENO_FEAT_PREEMPTION)))
+ (config->info->features & ADRENO_FEAT_PREEMPTION)))
ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_a7xx, 4);
else if (is_a7xx)
ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_a7xx, 1);
diff --git a/drivers/gpu/drm/msm/adreno/adreno_device.c b/drivers/gpu/drm/msm/adreno/adreno_device.c
index 09d4569f77528c2a20cabc814668c4c930dd07f1..11a228472fa0cef3b6e4e21a366470fbbc3ea8df 100644
--- a/drivers/gpu/drm/msm/adreno/adreno_device.c
+++ b/drivers/gpu/drm/msm/adreno/adreno_device.c
@@ -207,7 +207,7 @@ static int adreno_bind(struct device *dev, struct device *master, void *data)
priv->is_a2xx = info->family < ADRENO_3XX;
priv->has_cached_coherent =
- !!(info->quirks & ADRENO_FEAT_HAS_CACHED_COHERENT);
+ !!(info->features & ADRENO_FEAT_HAS_CACHED_COHERENT);
gpu = info->init(drm);
if (IS_ERR(gpu)) {
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
index 8782c25e8a393ec7d9dc23ad450908d039bd08c5..4702d4cfca3b58fb3cbb25cb6805f1c19be2ebcb 100644
--- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h
+++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
@@ -55,9 +55,9 @@ enum adreno_family {
#define ADRENO_QUIRK_FAULT_DETECT_MASK BIT(1)
#define ADRENO_QUIRK_LMLOADKILL_DISABLE BIT(2)
-#define ADRENO_FEAT_HAS_HW_APRIV BIT(3)
-#define ADRENO_FEAT_HAS_CACHED_COHERENT BIT(4)
-#define ADRENO_FEAT_PREEMPTION BIT(5)
+#define ADRENO_FEAT_HAS_HW_APRIV BIT(0)
+#define ADRENO_FEAT_HAS_CACHED_COHERENT BIT(1)
+#define ADRENO_FEAT_PREEMPTION BIT(2)
/* Helper for formating the chip_id in the way that userspace tools like
* crashdec expect.
@@ -98,6 +98,7 @@ struct adreno_info {
uint32_t revn;
const char *fw[ADRENO_FW_MAX];
uint32_t gmem;
+ u64 features;
u64 quirks;
struct msm_gpu *(*init)(struct drm_device *dev);
const char *zapfw;
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* Re: [PATCH v2 03/11] drm/msm: adreno: move features bits in a separate variable
2024-11-19 17:56 ` [PATCH v2 03/11] drm/msm: adreno: move features bits in a separate variable Neil Armstrong
@ 2024-11-20 11:18 ` Dmitry Baryshkov
0 siblings, 0 replies; 19+ messages in thread
From: Dmitry Baryshkov @ 2024-11-20 11:18 UTC (permalink / raw)
To: Neil Armstrong
Cc: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Marijn Suijten, David Airlie, Simona Vetter,
Bjorn Andersson, Rob Herring, Krzysztof Kozlowski, Conor Dooley,
Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree
On Tue, Nov 19, 2024 at 06:56:38PM +0100, Neil Armstrong wrote:
> Now the features defines have the right name, introduce a features
> bitfield and move the features defines in it, fixing all code checking
> for them.
>
> No functional changes intended.
I think it might be better to squahs this patch into the previous one,
it would simplify checking that we use .quirks for ADRENO_QUIRK_foo and
.features for ADRENO_FEAT_bar.
>
> Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
> ---
> drivers/gpu/drm/msm/adreno/a6xx_catalog.c | 34 +++++++++++++++---------------
> drivers/gpu/drm/msm/adreno/a6xx_gpu.c | 4 ++--
> drivers/gpu/drm/msm/adreno/adreno_device.c | 2 +-
> drivers/gpu/drm/msm/adreno/adreno_gpu.h | 7 +++---
> 4 files changed, 24 insertions(+), 23 deletions(-)
>
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
> index 825c820def315968d508973c8ae40c7c7b646569..93f0d4bf50ba773ecde93e6c29a2fcec24ebb7b3 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
> @@ -743,7 +743,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_512K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
> .init = a6xx_gpu_init,
> .zapfw = "a615_zap.mbn",
> .a6xx = &(const struct a6xx_info) {
> @@ -769,7 +769,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_512K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
> .init = a6xx_gpu_init,
> .a6xx = &(const struct a6xx_info) {
> .protect = &a630_protect,
> @@ -839,7 +839,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_512K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
> .init = a6xx_gpu_init,
> .zapfw = "a615_zap.mdt",
> .a6xx = &(const struct a6xx_info) {
> @@ -864,7 +864,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_512K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV,
> .init = a6xx_gpu_init,
> .zapfw = "a620_zap.mbn",
> @@ -892,7 +892,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_1M,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
> .init = a6xx_gpu_init,
> .zapfw = "a630_zap.mdt",
> .a6xx = &(const struct a6xx_info) {
> @@ -911,7 +911,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_1M,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
> .init = a6xx_gpu_init,
> .zapfw = "a640_zap.mdt",
> .a6xx = &(const struct a6xx_info) {
> @@ -934,7 +934,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_1M + SZ_128K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV,
> .init = a6xx_gpu_init,
> .zapfw = "a650_zap.mdt",
> @@ -961,7 +961,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_1M + SZ_512K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV,
> .init = a6xx_gpu_init,
> .zapfw = "a660_zap.mdt",
> @@ -981,7 +981,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_1M + SZ_512K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV,
> .init = a6xx_gpu_init,
> .a6xx = &(const struct a6xx_info) {
> @@ -1000,7 +1000,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_512K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV,
> .init = a6xx_gpu_init,
> .zapfw = "a660_zap.mbn",
> @@ -1028,7 +1028,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_2M,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT,
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT,
> .init = a6xx_gpu_init,
> .zapfw = "a640_zap.mdt",
> .a6xx = &(const struct a6xx_info) {
> @@ -1046,7 +1046,7 @@ static const struct adreno_info a6xx_gpus[] = {
> },
> .gmem = SZ_4M,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV,
> .init = a6xx_gpu_init,
> .zapfw = "a690_zap.mdt",
> @@ -1331,7 +1331,7 @@ static const struct adreno_info a7xx_gpus[] = {
> },
> .gmem = SZ_128K,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_HW_APRIV,
> + .features = ADRENO_FEAT_HAS_HW_APRIV,
> .init = a6xx_gpu_init,
> .zapfw = "a702_zap.mbn",
> .a6xx = &(const struct a6xx_info) {
> @@ -1355,7 +1355,7 @@ static const struct adreno_info a7xx_gpus[] = {
> },
> .gmem = SZ_2M,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV |
> ADRENO_FEAT_PREEMPTION,
> .init = a6xx_gpu_init,
> @@ -1377,7 +1377,7 @@ static const struct adreno_info a7xx_gpus[] = {
> },
> .gmem = 3 * SZ_1M,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV |
> ADRENO_FEAT_PREEMPTION,
> .init = a6xx_gpu_init,
> @@ -1400,7 +1400,7 @@ static const struct adreno_info a7xx_gpus[] = {
> },
> .gmem = 3 * SZ_1M,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV |
> ADRENO_FEAT_PREEMPTION,
> .init = a6xx_gpu_init,
> @@ -1422,7 +1422,7 @@ static const struct adreno_info a7xx_gpus[] = {
> },
> .gmem = 3 * SZ_1M,
> .inactive_period = DRM_MSM_INACTIVE_PERIOD,
> - .quirks = ADRENO_FEAT_HAS_CACHED_COHERENT |
> + .features = ADRENO_FEAT_HAS_CACHED_COHERENT |
> ADRENO_FEAT_HAS_HW_APRIV |
> ADRENO_FEAT_PREEMPTION,
> .init = a6xx_gpu_init,
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> index 2ebd3fac212576a1507e0b6afe2560cd0408dd89..654d0cff421f15901cd4bf33b41e70004634ec75 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> @@ -2478,7 +2478,7 @@ struct msm_gpu *a6xx_gpu_init(struct drm_device *dev)
> adreno_gpu->gmu_is_wrapper = of_device_is_compatible(node, "qcom,adreno-gmu-wrapper");
>
> adreno_gpu->base.hw_apriv =
> - !!(config->info->quirks & ADRENO_FEAT_HAS_HW_APRIV);
> + !!(config->info->features & ADRENO_FEAT_HAS_HW_APRIV);
>
> /* gpu->info only gets assigned in adreno_gpu_init() */
> is_a7xx = config->info->family == ADRENO_7XX_GEN1 ||
> @@ -2495,7 +2495,7 @@ struct msm_gpu *a6xx_gpu_init(struct drm_device *dev)
> }
>
> if ((enable_preemption == 1) || (enable_preemption == -1 &&
> - (config->info->quirks & ADRENO_FEAT_PREEMPTION)))
> + (config->info->features & ADRENO_FEAT_PREEMPTION)))
> ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_a7xx, 4);
> else if (is_a7xx)
> ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_a7xx, 1);
> diff --git a/drivers/gpu/drm/msm/adreno/adreno_device.c b/drivers/gpu/drm/msm/adreno/adreno_device.c
> index 09d4569f77528c2a20cabc814668c4c930dd07f1..11a228472fa0cef3b6e4e21a366470fbbc3ea8df 100644
> --- a/drivers/gpu/drm/msm/adreno/adreno_device.c
> +++ b/drivers/gpu/drm/msm/adreno/adreno_device.c
> @@ -207,7 +207,7 @@ static int adreno_bind(struct device *dev, struct device *master, void *data)
>
> priv->is_a2xx = info->family < ADRENO_3XX;
> priv->has_cached_coherent =
> - !!(info->quirks & ADRENO_FEAT_HAS_CACHED_COHERENT);
> + !!(info->features & ADRENO_FEAT_HAS_CACHED_COHERENT);
>
> gpu = info->init(drm);
> if (IS_ERR(gpu)) {
> diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> index 8782c25e8a393ec7d9dc23ad450908d039bd08c5..4702d4cfca3b58fb3cbb25cb6805f1c19be2ebcb 100644
> --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> @@ -55,9 +55,9 @@ enum adreno_family {
> #define ADRENO_QUIRK_FAULT_DETECT_MASK BIT(1)
> #define ADRENO_QUIRK_LMLOADKILL_DISABLE BIT(2)
>
> -#define ADRENO_FEAT_HAS_HW_APRIV BIT(3)
> -#define ADRENO_FEAT_HAS_CACHED_COHERENT BIT(4)
> -#define ADRENO_FEAT_PREEMPTION BIT(5)
> +#define ADRENO_FEAT_HAS_HW_APRIV BIT(0)
> +#define ADRENO_FEAT_HAS_CACHED_COHERENT BIT(1)
> +#define ADRENO_FEAT_PREEMPTION BIT(2)
>
> /* Helper for formating the chip_id in the way that userspace tools like
> * crashdec expect.
> @@ -98,6 +98,7 @@ struct adreno_info {
> uint32_t revn;
> const char *fw[ADRENO_FW_MAX];
> uint32_t gmem;
> + u64 features;
> u64 quirks;
> struct msm_gpu *(*init)(struct drm_device *dev);
> const char *zapfw;
>
> --
> 2.34.1
>
--
With best wishes
Dmitry
^ permalink raw reply [flat|nested] 19+ messages in thread
* [PATCH v2 04/11] drm/msm: adreno: add GMU_BW_VOTE feature flag
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
` (2 preceding siblings ...)
2024-11-19 17:56 ` [PATCH v2 03/11] drm/msm: adreno: move features bits in a separate variable Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-20 11:19 ` Dmitry Baryshkov
2024-11-19 17:56 ` [PATCH v2 05/11] drm/msm: adreno: add plumbing to generate bandwidth vote table for GMU Neil Armstrong
` (6 subsequent siblings)
10 siblings, 1 reply; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
The Adreno GMU Management Unit (GNU) can also scale the DDR Bandwidth
along the Frequency and Power Domain level, but by default we leave the
OPP core vote for the interconnect ddr path.
While scaling via the interconnect path was sufficient, newer GPUs
like the A750 requires specific vote paremeters and bandwidth to
achieve full functionality.
While the feature will require some data in a6xx_info, it's safer
to only enable tested platforms with this flag first.
Add a new feature enabling DDR Bandwidth vote via GMU.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/gpu/drm/msm/adreno/adreno_gpu.h | 1 +
1 file changed, 1 insertion(+)
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
index 4702d4cfca3b58fb3cbb25cb6805f1c19be2ebcb..394b96eb6c83354ae008b15b562bedb96cd391dd 100644
--- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h
+++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
@@ -58,6 +58,7 @@ enum adreno_family {
#define ADRENO_FEAT_HAS_HW_APRIV BIT(0)
#define ADRENO_FEAT_HAS_CACHED_COHERENT BIT(1)
#define ADRENO_FEAT_PREEMPTION BIT(2)
+#define ADRENO_FEAT_GMU_BW_VOTE BIT(3)
/* Helper for formating the chip_id in the way that userspace tools like
* crashdec expect.
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* Re: [PATCH v2 04/11] drm/msm: adreno: add GMU_BW_VOTE feature flag
2024-11-19 17:56 ` [PATCH v2 04/11] drm/msm: adreno: add GMU_BW_VOTE feature flag Neil Armstrong
@ 2024-11-20 11:19 ` Dmitry Baryshkov
0 siblings, 0 replies; 19+ messages in thread
From: Dmitry Baryshkov @ 2024-11-20 11:19 UTC (permalink / raw)
To: Neil Armstrong
Cc: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Marijn Suijten, David Airlie, Simona Vetter,
Bjorn Andersson, Rob Herring, Krzysztof Kozlowski, Conor Dooley,
Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree
On Tue, Nov 19, 2024 at 06:56:39PM +0100, Neil Armstrong wrote:
> The Adreno GMU Management Unit (GNU) can also scale the DDR Bandwidth
> along the Frequency and Power Domain level, but by default we leave the
> OPP core vote for the interconnect ddr path.
>
> While scaling via the interconnect path was sufficient, newer GPUs
> like the A750 requires specific vote paremeters and bandwidth to
> achieve full functionality.
>
> While the feature will require some data in a6xx_info, it's safer
> to only enable tested platforms with this flag first.
>
> Add a new feature enabling DDR Bandwidth vote via GMU.
Squash into the implementation patch.
>
> Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
> ---
> drivers/gpu/drm/msm/adreno/adreno_gpu.h | 1 +
> 1 file changed, 1 insertion(+)
>
> diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> index 4702d4cfca3b58fb3cbb25cb6805f1c19be2ebcb..394b96eb6c83354ae008b15b562bedb96cd391dd 100644
> --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> @@ -58,6 +58,7 @@ enum adreno_family {
> #define ADRENO_FEAT_HAS_HW_APRIV BIT(0)
> #define ADRENO_FEAT_HAS_CACHED_COHERENT BIT(1)
> #define ADRENO_FEAT_PREEMPTION BIT(2)
> +#define ADRENO_FEAT_GMU_BW_VOTE BIT(3)
>
> /* Helper for formating the chip_id in the way that userspace tools like
> * crashdec expect.
>
> --
> 2.34.1
>
--
With best wishes
Dmitry
^ permalink raw reply [flat|nested] 19+ messages in thread
* [PATCH v2 05/11] drm/msm: adreno: add plumbing to generate bandwidth vote table for GMU
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
` (3 preceding siblings ...)
2024-11-19 17:56 ` [PATCH v2 04/11] drm/msm: adreno: add GMU_BW_VOTE feature flag Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-20 11:42 ` Dmitry Baryshkov
2024-11-19 17:56 ` [PATCH v2 06/11] drm/msm: adreno: dynamically generate GMU bw table Neil Armstrong
` (5 subsequent siblings)
10 siblings, 1 reply; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
The Adreno GMU Management Unit (GMU) can also scale DDR Bandwidth along
the Frequency and Power Domain level, but by default we leave the
OPP core scale the interconnect ddr path.
In order to calculate vote values used by the GPU Management
Unit (GMU), we need to parse all the possible OPP Bandwidths and
create a vote value to be sent to the appropriate Bus Control
Modules (BCMs) declared in the GPU info struct.
The vote array will then be used to dynamically generate the GMU
bw_table sent during the GMU power-up.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/gpu/drm/msm/adreno/a6xx_gmu.c | 153 ++++++++++++++++++++++++++++++++++
drivers/gpu/drm/msm/adreno/a6xx_gmu.h | 14 ++++
drivers/gpu/drm/msm/adreno/a6xx_gpu.h | 1 +
3 files changed, 168 insertions(+)
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
index 14db7376c712d19446b38152e480bd5a1e0a5198..f6814d92a4edb29ba8a34a34aabb8b2324e9c6a4 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
@@ -9,6 +9,7 @@
#include <linux/pm_domain.h>
#include <linux/pm_opp.h>
#include <soc/qcom/cmd-db.h>
+#include <soc/qcom/tcs.h>
#include <drm/drm_gem.h>
#include "a6xx_gpu.h"
@@ -1287,6 +1288,109 @@ static int a6xx_gmu_memory_probe(struct a6xx_gmu *gmu)
return 0;
}
+/**
+ * struct bcm_db - Auxiliary data pertaining to each Bus Clock Manager (BCM)
+ * @unit: divisor used to convert bytes/sec bw value to an RPMh msg
+ * @width: multiplier used to convert bytes/sec bw value to an RPMh msg
+ * @vcd: virtual clock domain that this bcm belongs to
+ * @reserved: reserved field
+ */
+struct bcm_db {
+ __le32 unit;
+ __le16 width;
+ u8 vcd;
+ u8 reserved;
+};
+
+static u64 bcm_div(u64 num, u32 base)
+{
+ /* Ensure that small votes aren't lost. */
+ if (num && num < base)
+ return 1;
+
+ do_div(num, base);
+
+ return num;
+}
+
+static int a6xx_gmu_rpmh_bw_votes_init(const struct a6xx_info *info,
+ struct a6xx_gmu *gmu)
+{
+ const struct bcm_db *bcm_data[GMU_MAX_BCMS] = { 0 };
+ unsigned int bcm_index, bw_index;
+
+ /* Retrieve BCM data from cmd-db */
+ for (bcm_index = 0; bcm_index < GMU_MAX_BCMS; bcm_index++) {
+ size_t count;
+
+ /* Skip unconfigured BCM */
+ if (!info->bcm[bcm_index].name)
+ continue;
+
+ bcm_data[bcm_index] = cmd_db_read_aux_data(
+ info->bcm[bcm_index].name,
+ &count);
+ if (IS_ERR(bcm_data[bcm_index]))
+ return PTR_ERR(bcm_data[bcm_index]);
+
+ if (!count)
+ return -EINVAL;
+ }
+
+ /* Generate BCM votes values for each bandwidth & BCM */
+ for (bw_index = 0; bw_index < gmu->nr_gpu_bws; bw_index++) {
+ u32 *data = gmu->gpu_bw_votes[bw_index];
+ u32 bw = gmu->gpu_bw_table[bw_index];
+
+ /* Calculations loosely copied from bcm_aggregate() & tcs_cmd_gen() */
+ for (bcm_index = 0; bcm_index < GMU_MAX_BCMS; bcm_index++) {
+ bool commit = false;
+ u64 peak, vote;
+ u16 width;
+ u32 unit;
+
+ /* Skip unconfigured BCM */
+ if (!info->bcm[bcm_index].name || !bcm_data[bcm_index])
+ continue;
+
+ if (bcm_index == GMU_MAX_BCMS - 1 ||
+ (bcm_data[bcm_index + 1] &&
+ bcm_data[bcm_index]->vcd != bcm_data[bcm_index + 1]->vcd))
+ commit = true;
+
+ if (!bw) {
+ data[bcm_index] = BCM_TCS_CMD(commit, false, 0, 0);
+ continue;
+ }
+
+ if (info->bcm[bcm_index].fixed) {
+ u32 perfmode = 0;
+
+ if (bw >= info->bcm[bcm_index].perfmode_bw)
+ perfmode = info->bcm[bcm_index].perfmode;
+
+ data[bcm_index] = BCM_TCS_CMD(commit, true, 0, perfmode);
+ continue;
+ }
+
+ /* Multiply the bandwidth by the width of the connection */
+ width = le16_to_cpu(bcm_data[bcm_index]->width);
+ peak = bcm_div((u64)bw * width, info->bcm[bcm_index].buswidth);
+
+ /* Input bandwidth value is in KBps, scale the value to BCM unit */
+ unit = le32_to_cpu(bcm_data[bcm_index]->unit);
+ vote = bcm_div(peak * 1000ULL, unit);
+
+ if (vote > BCM_TCS_CMD_VOTE_MASK)
+ vote = BCM_TCS_CMD_VOTE_MASK;
+
+ data[bcm_index] = BCM_TCS_CMD(commit, true, vote, vote);
+ }
+ }
+
+ return 0;
+}
+
/* Return the 'arc-level' for the given frequency */
static unsigned int a6xx_gmu_get_arc_level(struct device *dev,
unsigned long freq)
@@ -1390,12 +1494,15 @@ static int a6xx_gmu_rpmh_arc_votes_init(struct device *dev, u32 *votes,
* The GMU votes with the RPMh for itself and on behalf of the GPU but we need
* to construct the list of votes on the CPU and send it over. Query the RPMh
* voltage levels and build the votes
+ * The GMU can also vote for DDR interconnects, use the OPP bandwidth entries
+ * and BCM parameters to build the votes.
*/
static int a6xx_gmu_rpmh_votes_init(struct a6xx_gmu *gmu)
{
struct a6xx_gpu *a6xx_gpu = container_of(gmu, struct a6xx_gpu, gmu);
struct adreno_gpu *adreno_gpu = &a6xx_gpu->base;
+ const struct a6xx_info *info = adreno_gpu->info->a6xx;
struct msm_gpu *gpu = &adreno_gpu->base;
int ret;
@@ -1407,6 +1514,10 @@ static int a6xx_gmu_rpmh_votes_init(struct a6xx_gmu *gmu)
ret |= a6xx_gmu_rpmh_arc_votes_init(gmu->dev, gmu->cx_arc_votes,
gmu->gmu_freqs, gmu->nr_gmu_freqs, "cx.lvl");
+ /* Build the interconnect votes */
+ if (adreno_gpu->info->features & ADRENO_FEAT_GMU_BW_VOTE)
+ ret |= a6xx_gmu_rpmh_bw_votes_init(info, gmu);
+
return ret;
}
@@ -1442,6 +1553,38 @@ static int a6xx_gmu_build_freq_table(struct device *dev, unsigned long *freqs,
return index;
}
+static int a6xx_gmu_build_bw_table(struct device *dev, unsigned long *bandwidths,
+ u32 size)
+{
+ int count = dev_pm_opp_get_opp_count(dev);
+ struct dev_pm_opp *opp;
+ int i, index = 0;
+ unsigned int bandwidth = 1;
+
+ /*
+ * The OPP table doesn't contain the "off" bandwidth level so we need to
+ * add 1 to the table size to account for it
+ */
+
+ if (WARN(count + 1 > size,
+ "The GMU bandwidth table is being truncated\n"))
+ count = size - 1;
+
+ /* Set the "off" bandwidth */
+ bandwidths[index++] = 0;
+
+ for (i = 0; i < count; i++) {
+ opp = dev_pm_opp_find_bw_ceil(dev, &bandwidth, 0);
+ if (IS_ERR(opp))
+ break;
+
+ dev_pm_opp_put(opp);
+ bandwidths[index++] = bandwidth++;
+ }
+
+ return index;
+}
+
static int a6xx_gmu_pwrlevels_probe(struct a6xx_gmu *gmu)
{
struct a6xx_gpu *a6xx_gpu = container_of(gmu, struct a6xx_gpu, gmu);
@@ -1472,6 +1615,16 @@ static int a6xx_gmu_pwrlevels_probe(struct a6xx_gmu *gmu)
gmu->current_perf_index = gmu->nr_gpu_freqs - 1;
+ /*
+ * The GMU also handles GPU Interconnect Votes so build a list
+ * of DDR bandwidths from the GPU OPP table
+ */
+ if (adreno_gpu->info->features & ADRENO_FEAT_GMU_BW_VOTE)
+ gmu->nr_gpu_bws = a6xx_gmu_build_bw_table(&gpu->pdev->dev,
+ gmu->gpu_bw_table, ARRAY_SIZE(gmu->gpu_bw_table));
+
+ gmu->current_perf_index = gmu->nr_gpu_freqs - 1;
+
/* Build the list of RPMh votes that we'll send to the GMU */
return a6xx_gmu_rpmh_votes_init(gmu);
}
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.h b/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
index b4a79f88ccf45cfe651c86d2a9da39541c5772b3..03603eadc0f9ed866899c95e99f333a511ebc3c1 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
@@ -19,6 +19,16 @@ struct a6xx_gmu_bo {
u64 iova;
};
+#define GMU_MAX_BCMS 3
+
+struct a6xx_bcm {
+ char *name;
+ unsigned int buswidth;
+ bool fixed;
+ unsigned int perfmode;
+ unsigned int perfmode_bw;
+};
+
/*
* These define the different GMU wake up options - these define how both the
* CPU and the GMU bring up the hardware
@@ -82,6 +92,10 @@ struct a6xx_gmu {
unsigned long gpu_freqs[16];
u32 gx_arc_votes[16];
+ int nr_gpu_bws;
+ unsigned long gpu_bw_table[16];
+ u32 gpu_bw_votes[16][GMU_MAX_BCMS];
+
int nr_gmu_freqs;
unsigned long gmu_freqs[4];
u32 cx_arc_votes[4];
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.h b/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
index 4aceffb6aae89c781facc2a6e4a82b20b341b6cb..5b80919e595fa1ba0a3afcca55feb89e60870cb1 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
@@ -44,6 +44,7 @@ struct a6xx_info {
u32 gmu_chipid;
u32 gmu_cgc_mode;
u32 prim_fifo_threshold;
+ const struct a6xx_bcm bcm[GMU_MAX_BCMS];
};
struct a6xx_gpu {
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* Re: [PATCH v2 05/11] drm/msm: adreno: add plumbing to generate bandwidth vote table for GMU
2024-11-19 17:56 ` [PATCH v2 05/11] drm/msm: adreno: add plumbing to generate bandwidth vote table for GMU Neil Armstrong
@ 2024-11-20 11:42 ` Dmitry Baryshkov
0 siblings, 0 replies; 19+ messages in thread
From: Dmitry Baryshkov @ 2024-11-20 11:42 UTC (permalink / raw)
To: Neil Armstrong
Cc: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Marijn Suijten, David Airlie, Simona Vetter,
Bjorn Andersson, Rob Herring, Krzysztof Kozlowski, Conor Dooley,
Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree
On Tue, Nov 19, 2024 at 06:56:40PM +0100, Neil Armstrong wrote:
> The Adreno GMU Management Unit (GMU) can also scale DDR Bandwidth along
> the Frequency and Power Domain level, but by default we leave the
> OPP core scale the interconnect ddr path.
>
> In order to calculate vote values used by the GPU Management
> Unit (GMU), we need to parse all the possible OPP Bandwidths and
> create a vote value to be sent to the appropriate Bus Control
> Modules (BCMs) declared in the GPU info struct.
>
> The vote array will then be used to dynamically generate the GMU
> bw_table sent during the GMU power-up.
>
> Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
LGTM, two minor nits below.
> ---
> drivers/gpu/drm/msm/adreno/a6xx_gmu.c | 153 ++++++++++++++++++++++++++++++++++
> drivers/gpu/drm/msm/adreno/a6xx_gmu.h | 14 ++++
> drivers/gpu/drm/msm/adreno/a6xx_gpu.h | 1 +
> 3 files changed, 168 insertions(+)
>
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
> index 14db7376c712d19446b38152e480bd5a1e0a5198..f6814d92a4edb29ba8a34a34aabb8b2324e9c6a4 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
> @@ -9,6 +9,7 @@
> #include <linux/pm_domain.h>
> #include <linux/pm_opp.h>
> #include <soc/qcom/cmd-db.h>
> +#include <soc/qcom/tcs.h>
> #include <drm/drm_gem.h>
>
> #include "a6xx_gpu.h"
> @@ -1287,6 +1288,109 @@ static int a6xx_gmu_memory_probe(struct a6xx_gmu *gmu)
> return 0;
> }
>
> +/**
> + * struct bcm_db - Auxiliary data pertaining to each Bus Clock Manager (BCM)
> + * @unit: divisor used to convert bytes/sec bw value to an RPMh msg
> + * @width: multiplier used to convert bytes/sec bw value to an RPMh msg
> + * @vcd: virtual clock domain that this bcm belongs to
> + * @reserved: reserved field
> + */
> +struct bcm_db {
> + __le32 unit;
> + __le16 width;
> + u8 vcd;
> + u8 reserved;
> +};
> +
> +static u64 bcm_div(u64 num, u32 base)
> +{
> + /* Ensure that small votes aren't lost. */
> + if (num && num < base)
> + return 1;
> +
> + do_div(num, base);
> +
> + return num;
> +}
> +
> +static int a6xx_gmu_rpmh_bw_votes_init(const struct a6xx_info *info,
> + struct a6xx_gmu *gmu)
> +{
> + const struct bcm_db *bcm_data[GMU_MAX_BCMS] = { 0 };
> + unsigned int bcm_index, bw_index;
> +
> + /* Retrieve BCM data from cmd-db */
> + for (bcm_index = 0; bcm_index < GMU_MAX_BCMS; bcm_index++) {
> + size_t count;
> +
> + /* Skip unconfigured BCM */
> + if (!info->bcm[bcm_index].name)
> + continue;
> +
> + bcm_data[bcm_index] = cmd_db_read_aux_data(
> + info->bcm[bcm_index].name,
> + &count);
> + if (IS_ERR(bcm_data[bcm_index]))
> + return PTR_ERR(bcm_data[bcm_index]);
> +
> + if (!count)
> + return -EINVAL;
> + }
> +
> + /* Generate BCM votes values for each bandwidth & BCM */
> + for (bw_index = 0; bw_index < gmu->nr_gpu_bws; bw_index++) {
> + u32 *data = gmu->gpu_bw_votes[bw_index];
> + u32 bw = gmu->gpu_bw_table[bw_index];
> +
> + /* Calculations loosely copied from bcm_aggregate() & tcs_cmd_gen() */
> + for (bcm_index = 0; bcm_index < GMU_MAX_BCMS; bcm_index++) {
> + bool commit = false;
> + u64 peak, vote;
> + u16 width;
> + u32 unit;
> +
> + /* Skip unconfigured BCM */
> + if (!info->bcm[bcm_index].name || !bcm_data[bcm_index])
> + continue;
Nit: you don't care about the .name anymore, the first check can be
dropped.
> +
> + if (bcm_index == GMU_MAX_BCMS - 1 ||
> + (bcm_data[bcm_index + 1] &&
> + bcm_data[bcm_index]->vcd != bcm_data[bcm_index + 1]->vcd))
> + commit = true;
> +
> + if (!bw) {
> + data[bcm_index] = BCM_TCS_CMD(commit, false, 0, 0);
> + continue;
> + }
> +
> + if (info->bcm[bcm_index].fixed) {
> + u32 perfmode = 0;
> +
> + if (bw >= info->bcm[bcm_index].perfmode_bw)
> + perfmode = info->bcm[bcm_index].perfmode;
> +
> + data[bcm_index] = BCM_TCS_CMD(commit, true, 0, perfmode);
> + continue;
> + }
> +
> + /* Multiply the bandwidth by the width of the connection */
> + width = le16_to_cpu(bcm_data[bcm_index]->width);
> + peak = bcm_div((u64)bw * width, info->bcm[bcm_index].buswidth);
> +
> + /* Input bandwidth value is in KBps, scale the value to BCM unit */
> + unit = le32_to_cpu(bcm_data[bcm_index]->unit);
> + vote = bcm_div(peak * 1000ULL, unit);
> +
> + if (vote > BCM_TCS_CMD_VOTE_MASK)
> + vote = BCM_TCS_CMD_VOTE_MASK;
> +
> + data[bcm_index] = BCM_TCS_CMD(commit, true, vote, vote);
> + }
> + }
> +
> + return 0;
> +}
> +
> /* Return the 'arc-level' for the given frequency */
> static unsigned int a6xx_gmu_get_arc_level(struct device *dev,
> unsigned long freq)
> @@ -1390,12 +1494,15 @@ static int a6xx_gmu_rpmh_arc_votes_init(struct device *dev, u32 *votes,
> * The GMU votes with the RPMh for itself and on behalf of the GPU but we need
> * to construct the list of votes on the CPU and send it over. Query the RPMh
> * voltage levels and build the votes
> + * The GMU can also vote for DDR interconnects, use the OPP bandwidth entries
> + * and BCM parameters to build the votes.
> */
>
> static int a6xx_gmu_rpmh_votes_init(struct a6xx_gmu *gmu)
> {
> struct a6xx_gpu *a6xx_gpu = container_of(gmu, struct a6xx_gpu, gmu);
> struct adreno_gpu *adreno_gpu = &a6xx_gpu->base;
> + const struct a6xx_info *info = adreno_gpu->info->a6xx;
> struct msm_gpu *gpu = &adreno_gpu->base;
> int ret;
>
> @@ -1407,6 +1514,10 @@ static int a6xx_gmu_rpmh_votes_init(struct a6xx_gmu *gmu)
> ret |= a6xx_gmu_rpmh_arc_votes_init(gmu->dev, gmu->cx_arc_votes,
> gmu->gmu_freqs, gmu->nr_gmu_freqs, "cx.lvl");
>
> + /* Build the interconnect votes */
> + if (adreno_gpu->info->features & ADRENO_FEAT_GMU_BW_VOTE)
> + ret |= a6xx_gmu_rpmh_bw_votes_init(info, gmu);
> +
> return ret;
> }
>
> @@ -1442,6 +1553,38 @@ static int a6xx_gmu_build_freq_table(struct device *dev, unsigned long *freqs,
> return index;
> }
>
> +static int a6xx_gmu_build_bw_table(struct device *dev, unsigned long *bandwidths,
> + u32 size)
> +{
> + int count = dev_pm_opp_get_opp_count(dev);
> + struct dev_pm_opp *opp;
> + int i, index = 0;
> + unsigned int bandwidth = 1;
> +
> + /*
> + * The OPP table doesn't contain the "off" bandwidth level so we need to
> + * add 1 to the table size to account for it
> + */
> +
> + if (WARN(count + 1 > size,
> + "The GMU bandwidth table is being truncated\n"))
> + count = size - 1;
> +
> + /* Set the "off" bandwidth */
> + bandwidths[index++] = 0;
> +
> + for (i = 0; i < count; i++) {
> + opp = dev_pm_opp_find_bw_ceil(dev, &bandwidth, 0);
> + if (IS_ERR(opp))
> + break;
> +
> + dev_pm_opp_put(opp);
> + bandwidths[index++] = bandwidth++;
> + }
> +
> + return index;
> +}
> +
> static int a6xx_gmu_pwrlevels_probe(struct a6xx_gmu *gmu)
> {
> struct a6xx_gpu *a6xx_gpu = container_of(gmu, struct a6xx_gpu, gmu);
> @@ -1472,6 +1615,16 @@ static int a6xx_gmu_pwrlevels_probe(struct a6xx_gmu *gmu)
>
> gmu->current_perf_index = gmu->nr_gpu_freqs - 1;
>
> + /*
> + * The GMU also handles GPU Interconnect Votes so build a list
> + * of DDR bandwidths from the GPU OPP table
> + */
> + if (adreno_gpu->info->features & ADRENO_FEAT_GMU_BW_VOTE)
> + gmu->nr_gpu_bws = a6xx_gmu_build_bw_table(&gpu->pdev->dev,
> + gmu->gpu_bw_table, ARRAY_SIZE(gmu->gpu_bw_table));
> +
> + gmu->current_perf_index = gmu->nr_gpu_freqs - 1;
> +
> /* Build the list of RPMh votes that we'll send to the GMU */
> return a6xx_gmu_rpmh_votes_init(gmu);
> }
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.h b/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
> index b4a79f88ccf45cfe651c86d2a9da39541c5772b3..03603eadc0f9ed866899c95e99f333a511ebc3c1 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
> @@ -19,6 +19,16 @@ struct a6xx_gmu_bo {
> u64 iova;
> };
>
> +#define GMU_MAX_BCMS 3
> +
> +struct a6xx_bcm {
> + char *name;
> + unsigned int buswidth;
> + bool fixed;
> + unsigned int perfmode;
> + unsigned int perfmode_bw;
> +};
> +
> /*
> * These define the different GMU wake up options - these define how both the
> * CPU and the GMU bring up the hardware
> @@ -82,6 +92,10 @@ struct a6xx_gmu {
> unsigned long gpu_freqs[16];
> u32 gx_arc_votes[16];
>
> + int nr_gpu_bws;
> + unsigned long gpu_bw_table[16];
> + u32 gpu_bw_votes[16][GMU_MAX_BCMS];
We still have magic 16 here. GPU_MAX_FREQUENCIES? GPU_FREQ_TABLE_SIZE?
> +
> int nr_gmu_freqs;
> unsigned long gmu_freqs[4];
> u32 cx_arc_votes[4];
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.h b/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
> index 4aceffb6aae89c781facc2a6e4a82b20b341b6cb..5b80919e595fa1ba0a3afcca55feb89e60870cb1 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
> @@ -44,6 +44,7 @@ struct a6xx_info {
> u32 gmu_chipid;
> u32 gmu_cgc_mode;
> u32 prim_fifo_threshold;
> + const struct a6xx_bcm bcm[GMU_MAX_BCMS];
> };
>
> struct a6xx_gpu {
>
> --
> 2.34.1
>
--
With best wishes
Dmitry
^ permalink raw reply [flat|nested] 19+ messages in thread
* [PATCH v2 06/11] drm/msm: adreno: dynamically generate GMU bw table
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
` (4 preceding siblings ...)
2024-11-19 17:56 ` [PATCH v2 05/11] drm/msm: adreno: add plumbing to generate bandwidth vote table for GMU Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-19 17:56 ` [PATCH v2 07/11] drm/msm: adreno: find bandwidth index of OPP and set it along freq index Neil Armstrong
` (4 subsequent siblings)
10 siblings, 0 replies; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
The Adreno GPU Management Unit (GMU) can also scale the ddr
bandwidth along the frequency and power domain level, but for
now we statically fill the bw_table with values from the
downstream driver.
Only the first entry is used, which is a disable vote, so we
currently rely on scaling via the linux interconnect paths.
Let's dynamically generate the bw_table with the vote values
previously calculated from the OPPs.
Those entried will then be used by the GMU when passing the
appropriate bandwidth level while voting for a gpu frequency.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/gpu/drm/msm/adreno/a6xx_hfi.c | 39 ++++++++++++++++++++++++++++++++---
1 file changed, 36 insertions(+), 3 deletions(-)
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_hfi.c b/drivers/gpu/drm/msm/adreno/a6xx_hfi.c
index cb8844ed46b29c4569d05eb7a24f7b27e173190f..0c8aa9f8cabe1d9cb20445a4274b728236a99fad 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_hfi.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_hfi.c
@@ -621,6 +621,36 @@ static void a740_build_bw_table(struct a6xx_hfi_msg_bw_table *msg)
msg->cnoc_cmds_data[1][0] = 0x60000001;
}
+static void a740_generate_bw_table(struct adreno_gpu *adreno_gpu, struct a6xx_gmu *gmu,
+ struct a6xx_hfi_msg_bw_table *msg)
+{
+ const struct a6xx_info *info = adreno_gpu->info->a6xx;
+ unsigned int i, j;
+
+ msg->ddr_wait_bitmask = 0x7;
+
+ for (i = 0; i < 3; i++) {
+ if (!info->bcm[i].name)
+ break;
+ msg->ddr_cmds_addrs[i] = cmd_db_read_addr(info->bcm[i].name);
+ }
+ msg->ddr_cmds_num = i;
+
+ for (i = 0; i < gmu->nr_gpu_bws; ++i)
+ for (j = 0; j < msg->ddr_cmds_num; j++)
+ msg->ddr_cmds_data[i][j] = gmu->gpu_bw_votes[i][j];
+ msg->bw_level_num = gmu->nr_gpu_bws;
+
+ /* TODO also generate CNOC commands */
+
+ msg->cnoc_cmds_num = 1;
+ msg->cnoc_wait_bitmask = 0x1;
+
+ msg->cnoc_cmds_addrs[0] = cmd_db_read_addr("CN0");
+ msg->cnoc_cmds_data[0][0] = 0x40000000;
+ msg->cnoc_cmds_data[1][0] = 0x60000001;
+}
+
static void a6xx_build_bw_table(struct a6xx_hfi_msg_bw_table *msg)
{
/* Send a single "off" entry since the 630 GMU doesn't do bus scaling */
@@ -690,9 +720,12 @@ static int a6xx_hfi_send_bw_table(struct a6xx_gmu *gmu)
a690_build_bw_table(msg);
else if (adreno_is_a730(adreno_gpu))
a730_build_bw_table(msg);
- else if (adreno_is_a740_family(adreno_gpu))
- a740_build_bw_table(msg);
- else
+ else if (adreno_is_a740_family(adreno_gpu)) {
+ if ((adreno_gpu->info->features & ADRENO_FEAT_GMU_BW_VOTE) && gmu->nr_gpu_bws)
+ a740_generate_bw_table(adreno_gpu, gmu, msg);
+ else
+ a740_build_bw_table(msg);
+ } else
a6xx_build_bw_table(msg);
gmu->bw_table = msg;
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* [PATCH v2 07/11] drm/msm: adreno: find bandwidth index of OPP and set it along freq index
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
` (5 preceding siblings ...)
2024-11-19 17:56 ` [PATCH v2 06/11] drm/msm: adreno: dynamically generate GMU bw table Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-20 11:44 ` Dmitry Baryshkov
2024-11-19 17:56 ` [PATCH v2 08/11] drm/msm: adreno: request for maximum bus bandwidth usage Neil Armstrong
` (3 subsequent siblings)
10 siblings, 1 reply; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
The Adreno GMU Management Unit (GMU) can also scale the DDR Bandwidth
along the Frequency and Power Domain level, until now we left the OPP
core scale the OPP bandwidth via the interconnect path.
In order to enable bandwidth voting via the GPU Management
Unit (GMU), when an opp is set by devfreq we also look for
the corresponding bandwidth index in the previously generated
bw_table and pass this value along the frequency index to the GMU.
Since we now vote for all resources via the GMU, setting the OPP
is no more needed, so we can completely skip calling
dev_pm_opp_set_opp() in this situation.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/gpu/drm/msm/adreno/a6xx_gmu.c | 17 +++++++++++++++--
drivers/gpu/drm/msm/adreno/a6xx_gmu.h | 2 +-
drivers/gpu/drm/msm/adreno/a6xx_hfi.c | 6 +++---
3 files changed, 19 insertions(+), 6 deletions(-)
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
index f6814d92a4edb29ba8a34a34aabb8b2324e9c6a4..dc2d0035544e7848e5c4ea27f1ea9a191f9c4991 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
@@ -113,6 +113,7 @@ void a6xx_gmu_set_freq(struct msm_gpu *gpu, struct dev_pm_opp *opp,
struct a6xx_gpu *a6xx_gpu = to_a6xx_gpu(adreno_gpu);
struct a6xx_gmu *gmu = &a6xx_gpu->gmu;
u32 perf_index;
+ u32 bw_index = 0;
unsigned long gpu_freq;
int ret = 0;
@@ -125,6 +126,16 @@ void a6xx_gmu_set_freq(struct msm_gpu *gpu, struct dev_pm_opp *opp,
if (gpu_freq == gmu->gpu_freqs[perf_index])
break;
+ /* If enabled, find the corresponding DDR bandwidth index */
+ if ((adreno_gpu->info->features & ADRENO_FEAT_GMU_BW_VOTE) && gmu->nr_gpu_bws) {
+ unsigned int bw = dev_pm_opp_get_bw(opp, true, 0);
+
+ for (bw_index = 0; bw_index < gmu->nr_gpu_bws - 1; bw_index++) {
+ if (bw == gmu->gpu_bw_table[bw_index])
+ break;
+ }
+ }
+
gmu->current_perf_index = perf_index;
gmu->freq = gmu->gpu_freqs[perf_index];
@@ -140,8 +151,10 @@ void a6xx_gmu_set_freq(struct msm_gpu *gpu, struct dev_pm_opp *opp,
return;
if (!gmu->legacy) {
- a6xx_hfi_set_freq(gmu, perf_index);
- dev_pm_opp_set_opp(&gpu->pdev->dev, opp);
+ a6xx_hfi_set_freq(gmu, perf_index, bw_index);
+ /* With Bandwidth voting, we now vote for all resources, so skip OPP set */
+ if (!bw_index)
+ dev_pm_opp_set_opp(&gpu->pdev->dev, opp);
return;
}
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.h b/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
index 03603eadc0f9ed866899c95e99f333a511ebc3c1..a42cdd0261872a08033a368af1434492c075b3fd 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.h
@@ -207,7 +207,7 @@ void a6xx_hfi_init(struct a6xx_gmu *gmu);
int a6xx_hfi_start(struct a6xx_gmu *gmu, int boot_state);
void a6xx_hfi_stop(struct a6xx_gmu *gmu);
int a6xx_hfi_send_prep_slumber(struct a6xx_gmu *gmu);
-int a6xx_hfi_set_freq(struct a6xx_gmu *gmu, int index);
+int a6xx_hfi_set_freq(struct a6xx_gmu *gmu, int perf_index, int bw_index);
bool a6xx_gmu_gx_is_on(struct a6xx_gmu *gmu);
bool a6xx_gmu_sptprac_is_on(struct a6xx_gmu *gmu);
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_hfi.c b/drivers/gpu/drm/msm/adreno/a6xx_hfi.c
index 0c8aa9f8cabe1d9cb20445a4274b728236a99fad..2d8e144e3453b30e7d3a22c9252e516763c39b83 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_hfi.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_hfi.c
@@ -759,13 +759,13 @@ static int a6xx_hfi_send_core_fw_start(struct a6xx_gmu *gmu)
sizeof(msg), NULL, 0);
}
-int a6xx_hfi_set_freq(struct a6xx_gmu *gmu, int index)
+int a6xx_hfi_set_freq(struct a6xx_gmu *gmu, int freq_index, int bw_index)
{
struct a6xx_hfi_gx_bw_perf_vote_cmd msg = { 0 };
msg.ack_type = 1; /* blocking */
- msg.freq = index;
- msg.bw = 0; /* TODO: bus scaling */
+ msg.freq = freq_index;
+ msg.bw = bw_index;
return a6xx_hfi_send_msg(gmu, HFI_H2F_MSG_GX_BW_PERF_VOTE, &msg,
sizeof(msg), NULL, 0);
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* Re: [PATCH v2 07/11] drm/msm: adreno: find bandwidth index of OPP and set it along freq index
2024-11-19 17:56 ` [PATCH v2 07/11] drm/msm: adreno: find bandwidth index of OPP and set it along freq index Neil Armstrong
@ 2024-11-20 11:44 ` Dmitry Baryshkov
0 siblings, 0 replies; 19+ messages in thread
From: Dmitry Baryshkov @ 2024-11-20 11:44 UTC (permalink / raw)
To: Neil Armstrong
Cc: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Marijn Suijten, David Airlie, Simona Vetter,
Bjorn Andersson, Rob Herring, Krzysztof Kozlowski, Conor Dooley,
Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree
On Tue, Nov 19, 2024 at 06:56:42PM +0100, Neil Armstrong wrote:
> The Adreno GMU Management Unit (GMU) can also scale the DDR Bandwidth
> along the Frequency and Power Domain level, until now we left the OPP
> core scale the OPP bandwidth via the interconnect path.
>
> In order to enable bandwidth voting via the GPU Management
> Unit (GMU), when an opp is set by devfreq we also look for
> the corresponding bandwidth index in the previously generated
> bw_table and pass this value along the frequency index to the GMU.
>
> Since we now vote for all resources via the GMU, setting the OPP
> is no more needed, so we can completely skip calling
> dev_pm_opp_set_opp() in this situation.
>
> Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
> ---
> drivers/gpu/drm/msm/adreno/a6xx_gmu.c | 17 +++++++++++++++--
> drivers/gpu/drm/msm/adreno/a6xx_gmu.h | 2 +-
> drivers/gpu/drm/msm/adreno/a6xx_hfi.c | 6 +++---
> 3 files changed, 19 insertions(+), 6 deletions(-)
>
Reviewed-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
--
With best wishes
Dmitry
^ permalink raw reply [flat|nested] 19+ messages in thread
* [PATCH v2 08/11] drm/msm: adreno: request for maximum bus bandwidth usage
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
` (6 preceding siblings ...)
2024-11-19 17:56 ` [PATCH v2 07/11] drm/msm: adreno: find bandwidth index of OPP and set it along freq index Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-19 17:56 ` [PATCH v2 09/11] drm/msm: adreno: enable GMU bandwidth for A740 and A750 Neil Armstrong
` (2 subsequent siblings)
10 siblings, 0 replies; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
When requesting a DDR bandwidth level along a GPU frequency
level via the GMU, we can also specify the bus bandwidth usage in a 16bit
quantitized value.
For now simply request the maximum bus usage.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/gpu/drm/msm/adreno/a6xx_gmu.c | 11 +++++++++++
drivers/gpu/drm/msm/adreno/a6xx_hfi.h | 5 +++++
2 files changed, 16 insertions(+)
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
index dc2d0035544e7848e5c4ea27f1ea9a191f9c4991..36c0f67fd8e109aabf09a0804bacbed3593c39d7 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
@@ -134,6 +134,17 @@ void a6xx_gmu_set_freq(struct msm_gpu *gpu, struct dev_pm_opp *opp,
if (bw == gmu->gpu_bw_table[bw_index])
break;
}
+
+ if (bw_index) {
+ /*
+ * Append AB vote to the maximum bus usage.
+ * AB represents a quantitized 16bit value of the
+ * max ddr bandwidth we could use, let's simply
+ * request the maximum for now.
+ */
+ bw_index |= AB_VOTE(MAX_AB_VOTE);
+ bw_index |= AB_VOTE_ENABLE;
+ }
}
gmu->current_perf_index = perf_index;
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_hfi.h b/drivers/gpu/drm/msm/adreno/a6xx_hfi.h
index 528110169398f69f16443a29a1594d19c36fb595..52ba4a07d7b9a709289acd244a751ace9bdaab5d 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_hfi.h
+++ b/drivers/gpu/drm/msm/adreno/a6xx_hfi.h
@@ -173,6 +173,11 @@ struct a6xx_hfi_gx_bw_perf_vote_cmd {
u32 bw;
};
+#define AB_VOTE_MASK GENMASK(31, 16)
+#define MAX_AB_VOTE (FIELD_MAX(AB_VOTE_MASK) - 1)
+#define AB_VOTE(vote) FIELD_PREP(AB_VOTE_MASK, (vote))
+#define AB_VOTE_ENABLE BIT(8)
+
#define HFI_H2F_MSG_PREPARE_SLUMBER 33
struct a6xx_hfi_prep_slumber_cmd {
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* [PATCH v2 09/11] drm/msm: adreno: enable GMU bandwidth for A740 and A750
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
` (7 preceding siblings ...)
2024-11-19 17:56 ` [PATCH v2 08/11] drm/msm: adreno: request for maximum bus bandwidth usage Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-20 11:45 ` Dmitry Baryshkov
2024-11-19 17:56 ` [PATCH v2 10/11] arm64: qcom: dts: sm8550: add interconnect and opp-peak-kBps for GPU Neil Armstrong
2024-11-19 17:56 ` [PATCH v2 11/11] arm64: qcom: dts: sm8650: " Neil Armstrong
10 siblings, 1 reply; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
Now all the DDR bandwidth voting via the GPU Management Unit (GMU)
is in place, declare the Bus Control Modules (BCMs) and the
corresponding parameters in the GPU info struct and add the
GMU_BW_VOTE feature bit to enable it.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
drivers/gpu/drm/msm/adreno/a6xx_catalog.c | 26 ++++++++++++++++++++++++--
1 file changed, 24 insertions(+), 2 deletions(-)
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
index 93f0d4bf50ba773ecde93e6c29a2fcec24ebb7b3..7cb96d524f76df67c6ee4377827a38384c1b343a 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
@@ -1379,7 +1379,8 @@ static const struct adreno_info a7xx_gpus[] = {
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
.features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV |
- ADRENO_FEAT_PREEMPTION,
+ ADRENO_FEAT_PREEMPTION |
+ ADRENO_FEAT_GMU_BW_VOTE,
.init = a6xx_gpu_init,
.zapfw = "a740_zap.mdt",
.a6xx = &(const struct a6xx_info) {
@@ -1388,6 +1389,16 @@ static const struct adreno_info a7xx_gpus[] = {
.pwrup_reglist = &a7xx_pwrup_reglist,
.gmu_chipid = 0x7020100,
.gmu_cgc_mode = 0x00020202,
+ .bcm = {
+ [0] = { .name = "SH0", .buswidth = 16 },
+ [1] = { .name = "MC0", .buswidth = 4 },
+ [2] = {
+ .name = "ACV",
+ .fixed = true,
+ .perfmode = BIT(3),
+ .perfmode_bw = 16500000,
+ },
+ },
},
.address_space_size = SZ_16G,
.preempt_record_size = 4192 * SZ_1K,
@@ -1424,7 +1435,8 @@ static const struct adreno_info a7xx_gpus[] = {
.inactive_period = DRM_MSM_INACTIVE_PERIOD,
.features = ADRENO_FEAT_HAS_CACHED_COHERENT |
ADRENO_FEAT_HAS_HW_APRIV |
- ADRENO_FEAT_PREEMPTION,
+ ADRENO_FEAT_PREEMPTION |
+ ADRENO_FEAT_GMU_BW_VOTE,
.init = a6xx_gpu_init,
.zapfw = "gen70900_zap.mbn",
.a6xx = &(const struct a6xx_info) {
@@ -1432,6 +1444,16 @@ static const struct adreno_info a7xx_gpus[] = {
.pwrup_reglist = &a7xx_pwrup_reglist,
.gmu_chipid = 0x7090100,
.gmu_cgc_mode = 0x00020202,
+ .bcm = {
+ [0] = { .name = "SH0", .buswidth = 16 },
+ [1] = { .name = "MC0", .buswidth = 4 },
+ [2] = {
+ .name = "ACV",
+ .fixed = true,
+ .perfmode = BIT(2),
+ .perfmode_bw = 10687500,
+ },
+ },
},
.address_space_size = SZ_16G,
.preempt_record_size = 3572 * SZ_1K,
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* Re: [PATCH v2 09/11] drm/msm: adreno: enable GMU bandwidth for A740 and A750
2024-11-19 17:56 ` [PATCH v2 09/11] drm/msm: adreno: enable GMU bandwidth for A740 and A750 Neil Armstrong
@ 2024-11-20 11:45 ` Dmitry Baryshkov
0 siblings, 0 replies; 19+ messages in thread
From: Dmitry Baryshkov @ 2024-11-20 11:45 UTC (permalink / raw)
To: Neil Armstrong
Cc: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Marijn Suijten, David Airlie, Simona Vetter,
Bjorn Andersson, Rob Herring, Krzysztof Kozlowski, Conor Dooley,
Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree
On Tue, Nov 19, 2024 at 06:56:44PM +0100, Neil Armstrong wrote:
> Now all the DDR bandwidth voting via the GPU Management Unit (GMU)
> is in place, declare the Bus Control Modules (BCMs) and the
> corresponding parameters in the GPU info struct and add the
> GMU_BW_VOTE feature bit to enable it.
>
> Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
> ---
> drivers/gpu/drm/msm/adreno/a6xx_catalog.c | 26 ++++++++++++++++++++++++--
> 1 file changed, 24 insertions(+), 2 deletions(-)
>
Reviewed-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
--
With best wishes
Dmitry
^ permalink raw reply [flat|nested] 19+ messages in thread
* [PATCH v2 10/11] arm64: qcom: dts: sm8550: add interconnect and opp-peak-kBps for GPU
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
` (8 preceding siblings ...)
2024-11-19 17:56 ` [PATCH v2 09/11] drm/msm: adreno: enable GMU bandwidth for A740 and A750 Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-20 11:47 ` Dmitry Baryshkov
2024-11-19 17:56 ` [PATCH v2 11/11] arm64: qcom: dts: sm8650: " Neil Armstrong
10 siblings, 1 reply; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
Each GPU OPP requires a specific peak DDR bandwidth, let's add
those to each OPP and also the related interconnect path.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
arch/arm64/boot/dts/qcom/sm8550.dtsi | 11 +++++++++++
1 file changed, 11 insertions(+)
diff --git a/arch/arm64/boot/dts/qcom/sm8550.dtsi b/arch/arm64/boot/dts/qcom/sm8550.dtsi
index 9dc0ee3eb98f8711e01934e47331b99e3bb73682..808dce3a624197d38222f53fffa280e63088c1c1 100644
--- a/arch/arm64/boot/dts/qcom/sm8550.dtsi
+++ b/arch/arm64/boot/dts/qcom/sm8550.dtsi
@@ -2113,6 +2113,9 @@ gpu: gpu@3d00000 {
qcom,gmu = <&gmu>;
#cooling-cells = <2>;
+ interconnects = <&gem_noc MASTER_GFX3D 0 &mc_virt SLAVE_EBI1 0>;
+ interconnect-names = "gfx-mem";
+
status = "disabled";
zap-shader {
@@ -2126,41 +2129,49 @@ gpu_opp_table: opp-table {
opp-680000000 {
opp-hz = /bits/ 64 <680000000>;
opp-level = <RPMH_REGULATOR_LEVEL_SVS_L1>;
+ opp-peak-kBps = <16500000>;
};
opp-615000000 {
opp-hz = /bits/ 64 <615000000>;
opp-level = <RPMH_REGULATOR_LEVEL_SVS_L0>;
+ opp-peak-kBps = <16500000>;
};
opp-550000000 {
opp-hz = /bits/ 64 <550000000>;
opp-level = <RPMH_REGULATOR_LEVEL_SVS>;
+ opp-peak-kBps = <12449218>;
};
opp-475000000 {
opp-hz = /bits/ 64 <475000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_L1>;
+ opp-peak-kBps = <8171875>;
};
opp-401000000 {
opp-hz = /bits/ 64 <401000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS>;
+ opp-peak-kBps = <6671875>;
};
opp-348000000 {
opp-hz = /bits/ 64 <348000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D0>;
+ opp-peak-kBps = <6074218>;
};
opp-295000000 {
opp-hz = /bits/ 64 <295000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D1>;
+ opp-peak-kBps = <6074218>;
};
opp-220000000 {
opp-hz = /bits/ 64 <220000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D2>;
+ opp-peak-kBps = <6074218>;
};
};
};
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* Re: [PATCH v2 10/11] arm64: qcom: dts: sm8550: add interconnect and opp-peak-kBps for GPU
2024-11-19 17:56 ` [PATCH v2 10/11] arm64: qcom: dts: sm8550: add interconnect and opp-peak-kBps for GPU Neil Armstrong
@ 2024-11-20 11:47 ` Dmitry Baryshkov
0 siblings, 0 replies; 19+ messages in thread
From: Dmitry Baryshkov @ 2024-11-20 11:47 UTC (permalink / raw)
To: Neil Armstrong
Cc: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Marijn Suijten, David Airlie, Simona Vetter,
Bjorn Andersson, Rob Herring, Krzysztof Kozlowski, Conor Dooley,
Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree
On Tue, Nov 19, 2024 at 06:56:45PM +0100, Neil Armstrong wrote:
> Each GPU OPP requires a specific peak DDR bandwidth, let's add
> those to each OPP and also the related interconnect path.
>
> Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
> ---
> arch/arm64/boot/dts/qcom/sm8550.dtsi | 11 +++++++++++
> 1 file changed, 11 insertions(+)
>
> diff --git a/arch/arm64/boot/dts/qcom/sm8550.dtsi b/arch/arm64/boot/dts/qcom/sm8550.dtsi
> index 9dc0ee3eb98f8711e01934e47331b99e3bb73682..808dce3a624197d38222f53fffa280e63088c1c1 100644
> --- a/arch/arm64/boot/dts/qcom/sm8550.dtsi
> +++ b/arch/arm64/boot/dts/qcom/sm8550.dtsi
> @@ -2113,6 +2113,9 @@ gpu: gpu@3d00000 {
> qcom,gmu = <&gmu>;
> #cooling-cells = <2>;
>
> + interconnects = <&gem_noc MASTER_GFX3D 0 &mc_virt SLAVE_EBI1 0>;
QCOM_ICC_TAG_ALWAYS
> + interconnect-names = "gfx-mem";
> +
> status = "disabled";
>
> zap-shader {
> @@ -2126,41 +2129,49 @@ gpu_opp_table: opp-table {
> opp-680000000 {
> opp-hz = /bits/ 64 <680000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_SVS_L1>;
> + opp-peak-kBps = <16500000>;
> };
>
> opp-615000000 {
> opp-hz = /bits/ 64 <615000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_SVS_L0>;
> + opp-peak-kBps = <16500000>;
> };
>
> opp-550000000 {
> opp-hz = /bits/ 64 <550000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_SVS>;
> + opp-peak-kBps = <12449218>;
> };
>
> opp-475000000 {
> opp-hz = /bits/ 64 <475000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_L1>;
> + opp-peak-kBps = <8171875>;
> };
>
> opp-401000000 {
> opp-hz = /bits/ 64 <401000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS>;
> + opp-peak-kBps = <6671875>;
> };
>
> opp-348000000 {
> opp-hz = /bits/ 64 <348000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D0>;
> + opp-peak-kBps = <6074218>;
> };
>
> opp-295000000 {
> opp-hz = /bits/ 64 <295000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D1>;
> + opp-peak-kBps = <6074218>;
> };
>
> opp-220000000 {
> opp-hz = /bits/ 64 <220000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D2>;
> + opp-peak-kBps = <6074218>;
> };
> };
> };
>
> --
> 2.34.1
>
--
With best wishes
Dmitry
^ permalink raw reply [flat|nested] 19+ messages in thread
* [PATCH v2 11/11] arm64: qcom: dts: sm8650: add interconnect and opp-peak-kBps for GPU
2024-11-19 17:56 [PATCH v2 00/11] drm/msm: adreno: add support for DDR bandwidth scaling via GMU Neil Armstrong
` (9 preceding siblings ...)
2024-11-19 17:56 ` [PATCH v2 10/11] arm64: qcom: dts: sm8550: add interconnect and opp-peak-kBps for GPU Neil Armstrong
@ 2024-11-19 17:56 ` Neil Armstrong
2024-11-20 11:47 ` Dmitry Baryshkov
10 siblings, 1 reply; 19+ messages in thread
From: Neil Armstrong @ 2024-11-19 17:56 UTC (permalink / raw)
To: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Dmitry Baryshkov, Marijn Suijten, David Airlie,
Simona Vetter, Bjorn Andersson, Rob Herring, Krzysztof Kozlowski,
Conor Dooley
Cc: Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree, Neil Armstrong
Each GPU OPP requires a specific peak DDR bandwidth, let's add
those to each OPP and also the related interconnect path.
Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
---
arch/arm64/boot/dts/qcom/sm8650.dtsi | 14 ++++++++++++++
1 file changed, 14 insertions(+)
diff --git a/arch/arm64/boot/dts/qcom/sm8650.dtsi b/arch/arm64/boot/dts/qcom/sm8650.dtsi
index 01ac3769ffa62ffb83c5c51878e2823e1982eb67..331c5140c16bf013190d6da136c0920009d2646b 100644
--- a/arch/arm64/boot/dts/qcom/sm8650.dtsi
+++ b/arch/arm64/boot/dts/qcom/sm8650.dtsi
@@ -2636,6 +2636,9 @@ gpu: gpu@3d00000 {
qcom,gmu = <&gmu>;
#cooling-cells = <2>;
+ interconnects = <&gem_noc MASTER_GFX3D 0 &mc_virt SLAVE_EBI1 0>;
+ interconnect-names = "gfx-mem";
+
status = "disabled";
zap-shader {
@@ -2649,56 +2652,67 @@ gpu_opp_table: opp-table {
opp-231000000 {
opp-hz = /bits/ 64 <231000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D2>;
+ opp-peak-kBps = <2136718>;
};
opp-310000000 {
opp-hz = /bits/ 64 <310000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D1>;
+ opp-peak-kBps = <6074218>;
};
opp-366000000 {
opp-hz = /bits/ 64 <366000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D0>;
+ opp-peak-kBps = <6074218>;
};
opp-422000000 {
opp-hz = /bits/ 64 <422000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS>;
+ opp-peak-kBps = <8171875>;
};
opp-500000000 {
opp-hz = /bits/ 64 <500000000>;
opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_L1>;
+ opp-peak-kBps = <8171875>;
};
opp-578000000 {
opp-hz = /bits/ 64 <578000000>;
opp-level = <RPMH_REGULATOR_LEVEL_SVS>;
+ opp-peak-kBps = <12449218>;
};
opp-629000000 {
opp-hz = /bits/ 64 <629000000>;
opp-level = <RPMH_REGULATOR_LEVEL_SVS_L0>;
+ opp-peak-kBps = <12449218>;
};
opp-680000000 {
opp-hz = /bits/ 64 <680000000>;
opp-level = <RPMH_REGULATOR_LEVEL_SVS_L1>;
+ opp-peak-kBps = <16500000>;
};
opp-720000000 {
opp-hz = /bits/ 64 <720000000>;
opp-level = <RPMH_REGULATOR_LEVEL_SVS_L2>;
+ opp-peak-kBps = <16500000>;
};
opp-770000000 {
opp-hz = /bits/ 64 <770000000>;
opp-level = <RPMH_REGULATOR_LEVEL_NOM>;
+ opp-peak-kBps = <16500000>;
};
opp-834000000 {
opp-hz = /bits/ 64 <834000000>;
opp-level = <RPMH_REGULATOR_LEVEL_NOM_L1>;
+ opp-peak-kBps = <16500000>;
};
};
};
--
2.34.1
^ permalink raw reply related [flat|nested] 19+ messages in thread* Re: [PATCH v2 11/11] arm64: qcom: dts: sm8650: add interconnect and opp-peak-kBps for GPU
2024-11-19 17:56 ` [PATCH v2 11/11] arm64: qcom: dts: sm8650: " Neil Armstrong
@ 2024-11-20 11:47 ` Dmitry Baryshkov
0 siblings, 0 replies; 19+ messages in thread
From: Dmitry Baryshkov @ 2024-11-20 11:47 UTC (permalink / raw)
To: Neil Armstrong
Cc: Akhil P Oommen, Viresh Kumar, Nishanth Menon, Stephen Boyd,
Rafael J. Wysocki, Rob Clark, Sean Paul, Konrad Dybcio,
Abhinav Kumar, Marijn Suijten, David Airlie, Simona Vetter,
Bjorn Andersson, Rob Herring, Krzysztof Kozlowski, Conor Dooley,
Connor Abbott, linux-pm, linux-kernel, linux-arm-msm, dri-devel,
freedreno, devicetree
On Tue, Nov 19, 2024 at 06:56:46PM +0100, Neil Armstrong wrote:
> Each GPU OPP requires a specific peak DDR bandwidth, let's add
> those to each OPP and also the related interconnect path.
>
> Signed-off-by: Neil Armstrong <neil.armstrong@linaro.org>
> ---
> arch/arm64/boot/dts/qcom/sm8650.dtsi | 14 ++++++++++++++
> 1 file changed, 14 insertions(+)
>
> diff --git a/arch/arm64/boot/dts/qcom/sm8650.dtsi b/arch/arm64/boot/dts/qcom/sm8650.dtsi
> index 01ac3769ffa62ffb83c5c51878e2823e1982eb67..331c5140c16bf013190d6da136c0920009d2646b 100644
> --- a/arch/arm64/boot/dts/qcom/sm8650.dtsi
> +++ b/arch/arm64/boot/dts/qcom/sm8650.dtsi
> @@ -2636,6 +2636,9 @@ gpu: gpu@3d00000 {
> qcom,gmu = <&gmu>;
> #cooling-cells = <2>;
>
> + interconnects = <&gem_noc MASTER_GFX3D 0 &mc_virt SLAVE_EBI1 0>;
QCOM_ICC_TAG_ALWAYS
> + interconnect-names = "gfx-mem";
> +
> status = "disabled";
>
> zap-shader {
> @@ -2649,56 +2652,67 @@ gpu_opp_table: opp-table {
> opp-231000000 {
> opp-hz = /bits/ 64 <231000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D2>;
> + opp-peak-kBps = <2136718>;
> };
>
> opp-310000000 {
> opp-hz = /bits/ 64 <310000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D1>;
> + opp-peak-kBps = <6074218>;
> };
>
> opp-366000000 {
> opp-hz = /bits/ 64 <366000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_D0>;
> + opp-peak-kBps = <6074218>;
> };
>
> opp-422000000 {
> opp-hz = /bits/ 64 <422000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS>;
> + opp-peak-kBps = <8171875>;
> };
>
> opp-500000000 {
> opp-hz = /bits/ 64 <500000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_LOW_SVS_L1>;
> + opp-peak-kBps = <8171875>;
> };
>
> opp-578000000 {
> opp-hz = /bits/ 64 <578000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_SVS>;
> + opp-peak-kBps = <12449218>;
> };
>
> opp-629000000 {
> opp-hz = /bits/ 64 <629000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_SVS_L0>;
> + opp-peak-kBps = <12449218>;
> };
>
> opp-680000000 {
> opp-hz = /bits/ 64 <680000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_SVS_L1>;
> + opp-peak-kBps = <16500000>;
> };
>
> opp-720000000 {
> opp-hz = /bits/ 64 <720000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_SVS_L2>;
> + opp-peak-kBps = <16500000>;
> };
>
> opp-770000000 {
> opp-hz = /bits/ 64 <770000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_NOM>;
> + opp-peak-kBps = <16500000>;
> };
>
> opp-834000000 {
> opp-hz = /bits/ 64 <834000000>;
> opp-level = <RPMH_REGULATOR_LEVEL_NOM_L1>;
> + opp-peak-kBps = <16500000>;
> };
> };
> };
>
> --
> 2.34.1
>
--
With best wishes
Dmitry
^ permalink raw reply [flat|nested] 19+ messages in thread