From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from gabe.freedesktop.org (gabe.freedesktop.org [131.252.210.177]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.lore.kernel.org (Postfix) with ESMTPS id 35A51C25B4E for ; Tue, 24 Jan 2023 07:58:57 +0000 (UTC) Received: from gabe.freedesktop.org (localhost [127.0.0.1]) by gabe.freedesktop.org (Postfix) with ESMTP id AE50C10E61A; Tue, 24 Jan 2023 07:58:56 +0000 (UTC) Received: from mail-wm1-x334.google.com (mail-wm1-x334.google.com [IPv6:2a00:1450:4864:20::334]) by gabe.freedesktop.org (Postfix) with ESMTPS id 334D710E627 for ; Tue, 24 Jan 2023 07:58:54 +0000 (UTC) Received: by mail-wm1-x334.google.com with SMTP id c4-20020a1c3504000000b003d9e2f72093so12203385wma.1 for ; Mon, 23 Jan 2023 23:58:54 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=in-reply-to:from:references:cc:to:content-language:subject :user-agent:mime-version:date:message-id:from:to:cc:subject:date :message-id:reply-to; bh=Q45PPlDRTdSxHRBWlvDnZEsuycTmkbTXDYXqoQchTbs=; b=ZtEZHwqppDr5kJ8XXZuyt4iCexl+cjiuTectMTPWAUl5aN3t+zlXoUlS6Qo+R/WyJy NgqZb59vkdu30Ce+nRVDp64EwpehmfXTv4cVerd8ZSUw7pXAbVfzmJMoUmD65Aw+fEw2 ZKZASpSOM9N6x0TxZfO4hzCT5Vi5sPD8UZ2+MlXRVuK8AZ0deJHY0aXG8uo4HIv8Q/zo LWzOuMPgKr8faBpwm8lUQqkKSA8fCaQSzS5zf5Eydr4Uj4OkZ3gqCjek6MZrcbxyVyi/ 3ZAzTFOhO/V0QSnuGVkspxhrpxS5ASZMDrYHU0QqxJN6mifQ56HPJ2nXffVyQBupf1YQ XvBQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=in-reply-to:from:references:cc:to:content-language:subject :user-agent:mime-version:date:message-id:x-gm-message-state:from:to :cc:subject:date:message-id:reply-to; bh=Q45PPlDRTdSxHRBWlvDnZEsuycTmkbTXDYXqoQchTbs=; b=E7OJrVY2sDqsoJ/gdA13Prvyk71XK4V/l+sKOyH6zyzw3RV6K+LCHHEd00JKn6LSEE 4TokW74FhiOFN+hYPq4cpM9fEW9vQz3ZltKjC5vKZiII7AhzkBlXzT0EVUYckexonmuF aWmbO9m5pn7HFcVikI1c8l5w+89VDTbUcqcPsE6BqrJxC6Po5WSiFdDolpob0vKl5M+Y Ifp4S2HI5WJShNwj8qiflmKcPxW1UQnmkic6RSCu2YaLM6ZhFmDpP1ijfzk8MrDZu6zN WufygmY480Jc3cIAAv7RdgB9Ve0gB0YIcd2qUSNMeBSCRcJW4Bel7QjR0/bwGdPWqC4q A5gQ== X-Gm-Message-State: AFqh2kowZDfuWwffOavQ00WWEvgUb79ysGuLJEn42LDAq2P/CccZZGP9 TjhDUfydX8Jk+g/dcsAuDYA= X-Google-Smtp-Source: AMrXdXs45wB5/P5BbppUKH060vGwQ+aJQpjQx+8hhYPlzS8BFJN4sYE08MuNhLcMpfa5lRuX7KgdvA== X-Received: by 2002:a05:600c:a11:b0:3db:f1f:bc31 with SMTP id z17-20020a05600c0a1100b003db0f1fbc31mr23748822wmp.16.1674547132548; Mon, 23 Jan 2023 23:58:52 -0800 (PST) Received: from [192.168.178.21] (p5b0ea2e7.dip0.t-ipconnect.de. [91.14.162.231]) by smtp.gmail.com with ESMTPSA id f20-20020a05600c43d400b003db07420d14sm1202454wmn.39.2023.01.23.23.58.51 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Mon, 23 Jan 2023 23:58:51 -0800 (PST) Content-Type: multipart/alternative; boundary="------------01ql92H3kEw1yz4OsY66O6XR" Message-ID: Date: Tue, 24 Jan 2023 08:58:50 +0100 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Thunderbird/102.4.2 Subject: Re: [PATCH 2/2] drm/amdgpu: add AMDGPU_INFO_VM_STAT to return GPU VM Content-Language: en-US To: =?UTF-8?B?TWFyZWsgT2zFocOhaw==?= References: <355bab4f-2ca8-3dd2-4cbc-264931fe9d7a@gmail.com> <1d0ddc01-b315-a5ff-f4cc-30b8aedfeb57@gmail.com> <4992933e-ad45-5f7a-b7af-39c6d0948321@gmail.com> From: =?UTF-8?Q?Christian_K=c3=b6nig?= In-Reply-To: X-BeenThere: amd-gfx@lists.freedesktop.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: Discussion list for AMD gfx List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: amd-gfx mailing list Errors-To: amd-gfx-bounces@lists.freedesktop.org Sender: "amd-gfx" This is a multi-part message in MIME format. --------------01ql92H3kEw1yz4OsY66O6XR Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit How are the counters which the HUD consumes declared? See what I want to avoid is a) to nail down the interface with the kernel on specific values and b) make it possible to easily expose new values. In other words what we could do with fdinfo is to have something like this: GALLIUM_FDINFO_HUD=drm-memory-vram,amd-evicted-vram,amd-mclk glxgears And the HUD just displays the values the kernel provides without the need to re-compile mesa when we want to add some more values nor have the values as part of the UAPI. Christian. Am 24.01.23 um 08:37 schrieb Marek Olšák: > The Gallium HUD doesn't consume strings. It only consumes values that > are exposed as counters from the driver. In this case, we need the > driver to expose evicted stats as counters. Each counter can set > whether the value is absolute (e.g. memory usage) or monotonic (e.g. > perf counter). Parsing fdinfo to get the values is undesirable. > > Marek > > On Mon, Jan 23, 2023 at 4:31 AM Christian König > wrote: > > Let's do this as valid in fdinfo. > > This way we can easily extend whatever the kernel wants to display > as statistics in the userspace HUD. > > Regards, > Christian. > > Am 21.01.23 um 01:45 schrieb Marek Olšák: >> We badly need a way to query evicted memory usage. It's essential >> for investigating performance problems and it uncovered the buddy >> allocator disaster. Please either suggest an alternative, suggest >> changes, or review. We need it ASAP. >> >> Thanks, >> Marek >> >> On Tue, Jan 10, 2023 at 11:55 AM Marek Olšák >> wrote: >> >> On Tue, Jan 10, 2023 at 11:23 AM Christian König >> wrote: >> >> Am 10.01.23 um 16:28 schrieb Marek Olšák: >>> On Wed, Jan 4, 2023 at 9:51 AM Christian König >>> wrote: >>> >>> Am 04.01.23 um 00:08 schrieb Marek Olšák: >>>> I see about the access now, but did you even look >>>> at the patch? >>> >>> I did look at the patch, but I haven't fully >>> understood yet what you are trying to do here. >>> >>> >>> First and foremost, it returns the evicted size of VRAM >>> and visible VRAM, and returns visible VRAM usage. It >>> should be obvious which stat includes the size of another. >>> >>> >>>> Because what the patch does isn't even exposed to >>>> common drm code, such as the preferred domain and >>>> visible VRAM placement, so it can't be in fdinfo >>>> right now. >>>> >>>> Or do you even know what fdinfo contains? Because >>>> it contains nothing useful. It only has VRAM and >>>> GTT usage, which we already have in the INFO ioctl, >>>> so it has nothing that we need. We mainly need the >>>> eviction information and visible VRAM information >>>> now. Everything else is a bonus. >>> >>> Well the main question is what are you trying to get >>> from that information? The eviction list for example >>> is completely meaningless to userspace, that stuff >>> is only temporary and will be cleared on the next CS >>> again. >>> >>> >>> I don't know what you mean. The returned eviction stats >>> look correct and are stable (they don't change much). >>> You can suggest changes if you think some numbers are >>> not reported correctly. >>> >>> >>> What we could expose is the VRAM over-commit value, >>> e.g. how much BOs which where supposed to be in VRAM >>> are in GTT now. I think that's what you are looking >>> for here, right? >>> >>> >>> The VRAM overcommit value is "evicted_vram". >>> >>> >>>> Also, it's undesirable to open and parse a text >>>> file if we can just call an ioctl. >>> >>> Well I see the reasoning for that, but I also see >>> why other drivers do a lot of the stuff we have as >>> IOCTL as separate files in sysfs, fdinfo or debugfs. >>> >>> Especially repeating all the static information >>> which were already available under sysfs in the INFO >>> IOCTL was a design mistake as far as I can see. Just >>> compare what AMDGPU and the KFD code is doing to >>> what for example i915 is doing. >>> >>> Same for things like debug information about a >>> process. The fdinfo stuff can be queried from >>> external tools (gdb, gputop, umr etc...) as well >>> which makes that interface more preferred. >>> >>> >>> Nothing uses fdinfo in Mesa. No driver uses sysfs in >>> Mesa except drm shims, noop drivers, and Intel for perf >>> metrics. sysfs itself is an unusable mess for the PCIe >>> query and is missing information. >>> >>> I'm not against exposing more stuff through sysfs and >>> fdinfo for tools, but I don't see any reason why drivers >>> should use it (other than for slowing down queries and >>> initialization). >> >> That's what I'm asking: Is this for some tool or to make >> some driver decision based on it? >> >> If you just want the numbers for over displaying then I >> think it would be better to put this into fdinfo together >> with the other existing stuff there. >> >> >> If you want to make allocation decisions based on this >> then we should have that as IOCTL or even better as >> mmap() page between kernel and userspace. But in this >> case I would also calculation the numbers completely >> different as well. >> >> See we have at least the following things in the kernel: >> 1. The eviction list in the VM. >>     Those are the BOs which are currently evicted and >> tried to moved back in on the next CS. >> >> 2. The VRAM over commit value. >>     In other words how much more VRAM than available has >> the application tried to allocate? >> >> 3. The visible VRAM usage by this application. >> >> The end goal is that the eviction list will go away, e.g. >> we will always have stable allocations based on >> allocations of other applications and not constantly swap >> things in and out. >> >> When you now expose the eviction list to userspace we >> will be stuck with this interface forever. >> >> >> It's for the GALLIUM HUD. >> >> The only missing thing is the size of all evicted VRAM >> allocations, and the size of all evicted visible VRAM >> allocations. >> >> 1. No list is exposed. Only sums of buffer sizes are exposed. >> Also, the eviction list has no meaning here. All lists are >> treated equally, and mem_type is compared with >> preferred_domains to determine where buffers are and where >> they should be. >> >> 2. I'm not interested in the overcommit value. I'm only >> interested in knowing the number of bytes of evicted VRAM >> right now. It can be as variable as the CPU load, but in >> practice it shouldn't be because PCIe doesn't have the >> bandwidth to move things quickly. >> >> 3. Yes, that's true. >> >> Marek >> > --------------01ql92H3kEw1yz4OsY66O6XR Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: 8bit How are the counters which the HUD consumes declared?

See what I want to avoid is a) to nail down the interface with the kernel on specific values and b) make it possible to easily expose new values.

In other words what we could do with fdinfo is to have something like this:

GALLIUM_FDINFO_HUD=drm-memory-vram,amd-evicted-vram,amd-mclk glxgears

And the HUD just displays the values the kernel provides without the need to re-compile mesa when we want to add some more values nor have the values as part of the UAPI.

Christian.

Am 24.01.23 um 08:37 schrieb Marek Olšák:
The Gallium HUD doesn't consume strings. It only consumes values that are exposed as counters from the driver. In this case, we need the driver to expose evicted stats as counters. Each counter can set whether the value is absolute (e.g. memory usage) or monotonic (e.g. perf counter). Parsing fdinfo to get the values is undesirable.

Marek

On Mon, Jan 23, 2023 at 4:31 AM Christian König <ckoenig.leichtzumerken@gmail.com> wrote:
Let's do this as valid in fdinfo.

This way we can easily extend whatever the kernel wants to display as statistics in the userspace HUD.

Regards,
Christian.

Am 21.01.23 um 01:45 schrieb Marek Olšák:
We badly need a way to query evicted memory usage. It's essential for investigating performance problems and it uncovered the buddy allocator disaster. Please either suggest an alternative, suggest changes, or review. We need it ASAP.

Thanks,
Marek

On Tue, Jan 10, 2023 at 11:55 AM Marek Olšák <maraeo@gmail.com> wrote:
On Tue, Jan 10, 2023 at 11:23 AM Christian König <ckoenig.leichtzumerken@gmail.com> wrote:
Am 10.01.23 um 16:28 schrieb Marek Olšák:
On Wed, Jan 4, 2023 at 9:51 AM Christian König <ckoenig.leichtzumerken@gmail.com> wrote:
Am 04.01.23 um 00:08 schrieb Marek Olšák:
I see about the access now, but did you even look at the patch?

I did look at the patch, but I haven't fully understood yet what you are trying to do here.

First and foremost, it returns the evicted size of VRAM and visible VRAM, and returns visible VRAM usage. It should be obvious which stat includes the size of another.


Because what the patch does isn't even exposed to common drm code, such as the preferred domain and visible VRAM placement, so it can't be in fdinfo right now.

Or do you even know what fdinfo contains? Because it contains nothing useful. It only has VRAM and GTT usage, which we already have in the INFO ioctl, so it has nothing that we need. We mainly need the eviction information and visible VRAM information now. Everything else is a bonus.

Well the main question is what are you trying to get from that information? The eviction list for example is completely meaningless to userspace, that stuff is only temporary and will be cleared on the next CS again.

I don't know what you mean. The returned eviction stats look correct and are stable (they don't change much). You can suggest changes if you think some numbers are not reported correctly.
 

What we could expose is the VRAM over-commit value, e.g. how much BOs which where supposed to be in VRAM are in GTT now. I think that's what you are looking for here, right?

The VRAM overcommit value is "evicted_vram".
 

Also, it's undesirable to open and parse a text file if we can just call an ioctl.

Well I see the reasoning for that, but I also see why other drivers do a lot of the stuff we have as IOCTL as separate files in sysfs, fdinfo or debugfs.

Especially repeating all the static information which were already available under sysfs in the INFO IOCTL was a design mistake as far as I can see. Just compare what AMDGPU and the KFD code is doing to what for example i915 is doing.

Same for things like debug information about a process. The fdinfo stuff can be queried from external tools (gdb, gputop, umr etc...) as well which makes that interface more preferred.

Nothing uses fdinfo in Mesa. No driver uses sysfs in Mesa except drm shims, noop drivers, and Intel for perf metrics. sysfs itself is an unusable mess for the PCIe query and is missing information.

I'm not against exposing more stuff through sysfs and fdinfo for tools, but I don't see any reason why drivers should use it (other than for slowing down queries and initialization).

That's what I'm asking: Is this for some tool or to make some driver decision based on it?

If you just want the numbers for over displaying then I think it would be better to put this into fdinfo together with the other existing stuff there.

If you want to make allocation decisions based on this then we should have that as IOCTL or even better as mmap() page between kernel and userspace. But in this case I would also calculation the numbers completely different as well.

See we have at least the following things in the kernel:
1. The eviction list in the VM.
    Those are the BOs which are currently evicted and tried to moved back in on the next CS.

2. The VRAM over commit value.
    In other words how much more VRAM than available has the application tried to allocate?

3. The visible VRAM usage by this application.

The end goal is that the eviction list will go away, e.g. we will always have stable allocations based on allocations of other applications and not constantly swap things in and out.

When you now expose the eviction list to userspace we will be stuck with this interface forever.

It's for the GALLIUM HUD.

The only missing thing is the size of all evicted VRAM allocations, and the size of all evicted visible VRAM allocations.

1. No list is exposed. Only sums of buffer sizes are exposed. Also, the eviction list has no meaning here. All lists are treated equally, and mem_type is compared with preferred_domains to determine where buffers are and where they should be.

2. I'm not interested in the overcommit value. I'm only interested in knowing the number of bytes of evicted VRAM right now. It can be as variable as the CPU load, but in practice it shouldn't be because PCIe doesn't have the bandwidth to move things quickly.

3. Yes, that's true.

Marek



--------------01ql92H3kEw1yz4OsY66O6XR--