From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 1379EC7619A for ; Wed, 12 Apr 2023 08:11:48 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229978AbjDLILq (ORCPT ); Wed, 12 Apr 2023 04:11:46 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:52102 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229882AbjDLILn (ORCPT ); Wed, 12 Apr 2023 04:11:43 -0400 Received: from mail-wr1-x436.google.com (mail-wr1-x436.google.com [IPv6:2a00:1450:4864:20::436]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 44F2761B4 for ; Wed, 12 Apr 2023 01:11:36 -0700 (PDT) Received: by mail-wr1-x436.google.com with SMTP id ffacd0b85a97d-2f27a10f72bso165786f8f.0 for ; Wed, 12 Apr 2023 01:11:36 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ffwll.ch; s=google; t=1681287095; x=1683879095; h=in-reply-to:content-transfer-encoding:content-disposition :mime-version:references:mail-followup-to:message-id:subject:cc:to :from:date:from:to:cc:subject:date:message-id:reply-to; bh=ApQh3OO8N/EkSlcg/GoWQjRUynfLNJg2zkmAQ0bDPGw=; b=gfyWMUpShdnQKWyAjpTEdc2LlGQvtN0yklmxrUAwTj29EHlkO5IhoMlUjmWqHs/QD8 Mb7XfysniPM4irakrgTheWv1Pk91OvVsHCAzzqQyCbxRl8G5wB9UshcnPPA6YQ4eGZmy QsJIeA5rz0iifCC/VUKfFFyWjgyIK9rVjnbmk= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; t=1681287095; x=1683879095; h=in-reply-to:content-transfer-encoding:content-disposition :mime-version:references:mail-followup-to:message-id:subject:cc:to :from:date:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=ApQh3OO8N/EkSlcg/GoWQjRUynfLNJg2zkmAQ0bDPGw=; b=zeQia75TT8A8ulUoBlayNPvHjAqfqJZ/A4H7jT3Kdvbxp7WUloibKz9kl0cIyhY4y+ W+IHpV0ITwCi/rB02XQQ/MjpfCDda1Tunc/Zet2ouC+ZOURjQPI5487ryRRWpBWc0DP5 ouCHFBu+ncvcLkNPcmaNv+lHgq0kw8vBNPqeKksLaSMlK0A92Bn36X1BjWoQuA7JFx9e yoV2g1kJdBxobgRcvWiNi6unpW5UIUqX3eFAVixG3u78ebB1dwNkMHaCZEJ6wEwyfDn+ OZ229lgz63GT81cD0W+Jdgv/7YJkChzU2n94NbrV3e9TsDAS2U1AAt44upfzCNZ8T9Jg vctg== X-Gm-Message-State: AAQBX9cVJaTAR3LChyeh5bxOZJ9TXf9HQY4aWEv8i4/DI314FaKEPlKs /K0sqw5f8nU1g9QKb6CMm54FDA== X-Google-Smtp-Source: AKy350bnBOA1yKbUJKGFfK7Ajlx+yHLgcIBWXcno+RB9JgfyhFepkx/MEZdWOedrsJ7jrgeu2MHhLQ== X-Received: by 2002:a5d:6588:0:b0:2e4:aa61:a50 with SMTP id q8-20020a5d6588000000b002e4aa610a50mr9004356wru.1.1681287094654; Wed, 12 Apr 2023 01:11:34 -0700 (PDT) Received: from phenom.ffwll.local (212-51-149-33.fiber7.init7.net. [212.51.149.33]) by smtp.gmail.com with ESMTPSA id m3-20020adfdc43000000b002c5691f13eesm16556148wrj.50.2023.04.12.01.11.33 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 12 Apr 2023 01:11:34 -0700 (PDT) Date: Wed, 12 Apr 2023 10:11:32 +0200 From: Daniel Vetter To: Dmitry Baryshkov Cc: Rob Clark , dri-devel@lists.freedesktop.org, Rob Clark , Tvrtko Ursulin , "open list:DOCUMENTATION" , linux-arm-msm@vger.kernel.org, Emil Velikov , Christopher Healy , open list , Sean Paul , Boris Brezillon , freedreno@lists.freedesktop.org Subject: Re: [Freedreno] [PATCH v2 0/2] drm: fdinfo memory stats Message-ID: Mail-Followup-To: Dmitry Baryshkov , Rob Clark , dri-devel@lists.freedesktop.org, Rob Clark , Tvrtko Ursulin , "open list:DOCUMENTATION" , linux-arm-msm@vger.kernel.org, Emil Velikov , Christopher Healy , open list , Sean Paul , Boris Brezillon , freedreno@lists.freedesktop.org References: <20230410210608.1873968-1-robdclark@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: X-Operating-System: Linux phenom 6.1.0-7-amd64 Precedence: bulk List-ID: X-Mailing-List: linux-doc@vger.kernel.org On Wed, Apr 12, 2023 at 01:36:52AM +0300, Dmitry Baryshkov wrote: > On 11/04/2023 21:28, Rob Clark wrote: > > On Tue, Apr 11, 2023 at 10:36 AM Dmitry Baryshkov > > wrote: > > > > > > On Tue, 11 Apr 2023 at 20:13, Rob Clark wrote: > > > > > > > > On Tue, Apr 11, 2023 at 9:53 AM Daniel Vetter wrote: > > > > > > > > > > On Tue, Apr 11, 2023 at 09:47:32AM -0700, Rob Clark wrote: > > > > > > On Mon, Apr 10, 2023 at 2:06 PM Rob Clark wrote: > > > > > > > > > > > > > > From: Rob Clark > > > > > > > > > > > > > > Similar motivation to other similar recent attempt[1]. But with an > > > > > > > attempt to have some shared code for this. As well as documentation. > > > > > > > > > > > > > > It is probably a bit UMA-centric, I guess devices with VRAM might want > > > > > > > some placement stats as well. But this seems like a reasonable start. > > > > > > > > > > > > > > Basic gputop support: https://patchwork.freedesktop.org/series/116236/ > > > > > > > And already nvtop support: https://github.com/Syllo/nvtop/pull/204 > > > > > > > > > > > > On a related topic, I'm wondering if it would make sense to report > > > > > > some more global things (temp, freq, etc) via fdinfo? Some of this, > > > > > > tools like nvtop could get by trawling sysfs or other driver specific > > > > > > ways. But maybe it makes sense to have these sort of things reported > > > > > > in a standardized way (even though they aren't really per-drm_file) > > > > > > > > > > I think that's a bit much layering violation, we'd essentially have to > > > > > reinvent the hwmon sysfs uapi in fdinfo. Not really a business I want to > > > > > be in :-) > > > > > > > > I guess this is true for temp (where there are thermal zones with > > > > potentially multiple temp sensors.. but I'm still digging my way thru > > > > the thermal_cooling_device stuff) > > > > > > It is slightly ugly. All thermal zones and cooling devices are virtual > > > devices (so, even no connection to the particular tsens device). One > > > can either enumerate them by checking > > > /sys/class/thermal/thermal_zoneN/type or enumerate them through > > > /sys/class/hwmon. For cooling devices again the only enumeration is > > > through /sys/class/thermal/cooling_deviceN/type. > > > > > > Probably it should be possible to push cooling devices and thermal > > > zones under corresponding providers. However I do not know if there is > > > a good way to correlate cooling device (ideally a part of GPU) to the > > > thermal_zone (which in our case is provided by tsens / temp_alarm > > > rather than GPU itself). > > > > > > > > > > > But what about freq? I think, esp for cases where some "fw thing" is > > > > controlling the freq we end up needing to use gpu counters to measure > > > > the freq. > > > > > > For the freq it is slightly easier: /sys/class/devfreq/*, devices are > > > registered under proper parent (IOW, GPU). So one can read > > > /sys/class/devfreq/3d00000.gpu/cur_freq or > > > /sys/bus/platform/devices/3d00000.gpu/devfreq/3d00000.gpu/cur_freq. > > > > > > However because of the components usage, there is no link from > > > /sys/class/drm/card0 > > > (/sys/devices/platform/soc@0/ae00000.display-subsystem/ae01000.display-controller/drm/card0) > > > to /sys/devices/platform/soc@0/3d00000.gpu, the GPU unit. > > > > > > Getting all these items together in a platform-independent way would > > > be definitely an important but complex topic. > > > > But I don't believe any of the pci gpu's use devfreq ;-) > > > > And also, you can't expect the CPU to actually know the freq when fw > > is the one controlling freq. We can, currently, have a reasonable > > approximation from devfreq but that stops if IFPC is implemented. And > > other GPUs have even less direct control. So freq is a thing that I > > don't think we should try to get from "common frameworks" > > I think it might be useful to add another passive devfreq governor type for > external frequencies. This way we can use the same interface to export > non-CPU-controlled frequencies. Yeah this sounds like a decent idea to me too. It might also solve the fun of various pci devices having very non-standard freq controls in sysfs (looking at least at i915 here ...) I guess it would minimally be a good idea if we could document this, or maybe have a reference implementation in nvtop or whatever the cool thing is rn. -Daniel > > > > > BR, > > -R > > > > > > > > > > > What might be needed is better glue to go from the fd or fdinfo to the > > > > > right hw device and then crawl around the hwmon in sysfs automatically. I > > > > > would not be surprised at all if we really suck on this, probably more > > > > > likely on SoC than pci gpus where at least everything should be under the > > > > > main pci sysfs device. > > > > > > > > yeah, I *think* userspace would have to look at /proc/device-tree to > > > > find the cooling device(s) associated with the gpu.. at least I don't > > > > see a straightforward way to figure it out just for sysfs > > > > > > > > BR, > > > > -R > > > > > > > > > -Daniel > > > > > > > > > > > > > > > > > BR, > > > > > > -R > > > > > > > > > > > > > > > > > > > [1] https://patchwork.freedesktop.org/series/112397/ > > > > > > > > > > > > > > Rob Clark (2): > > > > > > > drm: Add fdinfo memory stats > > > > > > > drm/msm: Add memory stats to fdinfo > > > > > > > > > > > > > > Documentation/gpu/drm-usage-stats.rst | 21 +++++++ > > > > > > > drivers/gpu/drm/drm_file.c | 79 +++++++++++++++++++++++++++ > > > > > > > drivers/gpu/drm/msm/msm_drv.c | 25 ++++++++- > > > > > > > drivers/gpu/drm/msm/msm_gpu.c | 2 - > > > > > > > include/drm/drm_file.h | 10 ++++ > > > > > > > 5 files changed, 134 insertions(+), 3 deletions(-) > > > > > > > > > > > > > > -- > > > > > > > 2.39.2 > > > > > > > > > > > > > > > > > -- > > > > > Daniel Vetter > > > > > Software Engineer, Intel Corporation > > > > > http://blog.ffwll.ch > > > > > > > > > > > > -- > > > With best wishes > > > Dmitry > > -- > With best wishes > Dmitry > -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch