public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Anton Blanchard <anton@samba.org>
To: William Lee Irwin III <wli@holomorphy.com>, linux-kernel@vger.kernel.org
Subject: Re: 48GB NUMA-Q boots, with major IO-APIC hassles
Date: Wed, 15 Jan 2003 23:32:09 +1100	[thread overview]
Message-ID: <20030115123209.GA6588@krispykreme> (raw)
In-Reply-To: <20030115115525.GI919@holomorphy.com>


> PCI config cycles must be qualified by segment here just to get the
> right address, so there's definitely a requirement for stuff. One
> "advantage" of NUMA-Q (if it can be called that) is that firmware/BIOS
> and hardware pushes a bus number mangling scheme that is more or less
> mandatory, so things can at largely implicitly taken care of with the
> bus number and the firmware's mapping of the mangled bus numbers to the
> cross-quad portio area. But this scheme does not have cooperation from
> PCI-PCI bridges, where NUMA-Q mangling scheme -noncompliant physical
> bus ID's are kept in the bus number registers.

Can you store the segment in pci_bus->sysdata and then use that in
your config read/write macros? What do you mean by the mangling?
Does each host bridge have a separate segment identifier? If so why
would the PCI-PCI bridges below it need to have incorrect IDs?

> I have no idea what to do about these; I just sort of hope and pray the
> backward-compatibility constraints won't hurt me. At the level of things
> exported to userspace my main concern is largely that things like disk
> arrays will actually be accessible as raw devices or mountable as fs's,
> cooperation with userspace drivers and accurate reporting is kind of
> secondary to me aside from satisfying backward-compatibility concerns
> from Pee Cee -centric sides of things.

Its possible customers will run X on their server, thats when getting
things like /proc/bus/pci/* right becomes important :) In fact on
sparc64 and ppc64 X should work with domains fairly easily because
they are already use /proc/bus/pci/* to mmap PCI BARs.

I think X should always use this method, reading PCI BARs and mmap'ing
/dev/mem is just awful.

Anton

  reply	other threads:[~2003-01-15 12:27 UTC|newest]

Thread overview: 17+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2003-01-15 10:58 48GB NUMA-Q boots, with major IO-APIC hassles William Lee Irwin III
2003-01-15 11:24 ` Anton Blanchard
2003-01-15 11:55   ` William Lee Irwin III
2003-01-15 12:32     ` Anton Blanchard [this message]
2003-01-15 13:10       ` William Lee Irwin III
2003-01-15 15:24 ` Martin J. Bligh
2003-01-15 15:34   ` William Lee Irwin III
2003-01-19  1:43 ` William Lee Irwin III
2003-01-19  1:50   ` William Lee Irwin III
2003-01-19  2:13     ` Zwane Mwaikambo
2003-01-19  2:27       ` William Lee Irwin III
2003-01-19  2:32     ` Zwane Mwaikambo
2003-01-19  2:55       ` William Lee Irwin III
2003-01-19  3:08         ` William Lee Irwin III
2003-03-28  5:08 ` William Lee Irwin III
  -- strict thread matches above, loose matches on Subject: below --
2003-01-15 17:32 Protasevich, Natalie
2003-01-15 22:01 ` Martin J. Bligh

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20030115123209.GA6588@krispykreme \
    --to=anton@samba.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=wli@holomorphy.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox