Intel-GFX Archive on lore.kernel.org
 help / color / mirror / Atom feed
From: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
To: Jani Nikula <jani.nikula@intel.com>, Intel-gfx@lists.freedesktop.org
Cc: Lucas De Marchi <lucas.demarchi@intel.com>,
	dri-devel@lists.freedesktop.org
Subject: Re: [Intel-gfx] [RFC] drm/i915: Split out intel_vtd_active and run_as_guest to own header
Date: Thu, 24 Mar 2022 13:29:02 +0000	[thread overview]
Message-ID: <2a91ffe1-71a2-98a0-daa3-23aee0b1c29d@linux.intel.com> (raw)
In-Reply-To: <87r16rh8b2.fsf@intel.com>


On 24/03/2022 11:57, Jani Nikula wrote:
> On Thu, 24 Mar 2022, Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com> wrote:
>> On 24/03/2022 09:31, Jani Nikula wrote:
>>> On Tue, 22 Mar 2022, Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com> wrote:
>>>> From: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
>>>>
>>>> ...
>>>>
>>>> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
>>>> Cc: Jani Nikula <jani.nikula@intel.com>
>>>> Cc: Lucas De Marchi <lucas.demarchi@intel.com>
>>>> ---
>>>> Typed up how I see it - bash away.
>>>
>>> So is intel_vtd_active() so performance critical that it needs to be
>>> inline?
>>>
>>> We're passing struct drm_i915_private * everywhere we can, and it just
>>> feels silly to use struct drm_device * to avoid the include.
>>>
>>> Static inlines considered harmful. :p
>>
>> Same as it is ;), and gee, who was it that he said he was just trying to
>> declutter i915_drv.h.. ;p
> 
> Not at the cost of clarity elsewhere!

To be clear now you oppose intel_vtd_active taking struct device? I 
thought you expressed general agreement when I presented the idea in the 
previous thread.

I don't mind hugely to go either way, but I also don't see how taking 
struct device makes anything unclear. (I only think 
intel_vtd_run_as_guest is really wrong in this story but that's old news.)

And if I make it take i915 then I would want to name it i915_vtd_active 
as well. But then you wouldn't like that.

Should we just stuff all this into i915_utils for now, as I think Lucas 
suggested? Static inline or not, I don't care.

Regards,

Tvrtko

  reply	other threads:[~2022-03-24 13:29 UTC|newest]

Thread overview: 15+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-03-22 16:44 [Intel-gfx] [RFC] drm/i915: Split out intel_vtd_active and run_as_guest to own header Tvrtko Ursulin
2022-03-22 17:05 ` [Intel-gfx] ✗ Fi.CI.CHECKPATCH: warning for " Patchwork
2022-03-22 17:06 ` [Intel-gfx] ✗ Fi.CI.SPARSE: " Patchwork
2022-03-22 17:10 ` [Intel-gfx] ✗ Fi.CI.DOCS: " Patchwork
2022-03-22 17:40 ` [Intel-gfx] ✓ Fi.CI.BAT: success " Patchwork
2022-03-23  1:03 ` [Intel-gfx] ✓ Fi.CI.IGT: " Patchwork
2022-03-24  9:31 ` [Intel-gfx] [RFC] " Jani Nikula
2022-03-24  9:32   ` Jani Nikula
2022-03-24 11:48   ` Tvrtko Ursulin
2022-03-24 11:57     ` Jani Nikula
2022-03-24 13:29       ` Tvrtko Ursulin [this message]
2022-03-24 18:57         ` Jani Nikula
2022-03-25  8:47           ` Tvrtko Ursulin
2022-03-25 12:09             ` Jani Nikula
2022-03-28 20:48               ` Lucas De Marchi

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=2a91ffe1-71a2-98a0-daa3-23aee0b1c29d@linux.intel.com \
    --to=tvrtko.ursulin@linux.intel.com \
    --cc=Intel-gfx@lists.freedesktop.org \
    --cc=dri-devel@lists.freedesktop.org \
    --cc=jani.nikula@intel.com \
    --cc=lucas.demarchi@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox