From: Borislav Petkov <bp@amd64.org>
To: Mauro Carvalho Chehab <mchehab@redhat.com>
Cc: Borislav Petkov <bp@amd64.org>,
Greg KH <gregkh@linuxfoundation.org>,
Linux Edac Mailing List <linux-edac@vger.kernel.org>,
Linux Kernel Mailing List <linux-kernel@vger.kernel.org>
Subject: Re: [PATCH 0/6] Add a per-dimm structure
Date: Thu, 15 Mar 2012 22:38:44 +0100 [thread overview]
Message-ID: <20120315213844.GA3781@aftab> (raw)
In-Reply-To: <4F61E35B.9000906@redhat.com>
On Thu, Mar 15, 2012 at 09:40:59AM -0300, Mauro Carvalho Chehab wrote:
> > What are you talking about? Those per-rank counters should be the same
> > as the per-csrow ch0 and ch1 counters...
>
> Yes, but with your proposal, the per-csrow counters will not be added
> (the equivalent of):
> /sys/devices/system/edac/mc/mc0/csrow0/ue_count
> /sys/devices/system/edac/mc/mc0/csrow0/ce_count
What the hell? Those are already there:
/sys/devices/system/edac/mc/mc0/csrow0/
|-- ce_count
|-- ch0_ce_count
|-- ch0_dimm_label
|-- ch1_ce_count
|-- ch1_dimm_label
|-- dev_type
|-- edac_mode
|-- mem_type
|-- size_mb
`-- ue_count
and since userspace uses them, they cannot be removed.
> > It depends - if the 128 bit word comes from a single DIMM (unganged
> > mode) then you have a per-rank UE.
>
> True, and there are other types of ECC logic that would allow to identify
> what DIMM/rank produced the error.
>
> Yet, the typical case is to use two DIMMs for a 128-bits cacheline
> on separate channels, due to performance improvements, and ECC chipkill
> using the 128+16 bits, as it improves the probability of error correction.
... and in this typical case, on smart hardware you can get the rank
too. If one cannot discern between the two DIMMs, then there should be
one counter and the other one should be a symlink to that counter, or
something to that effect.
> >> Of course, the EDAC logic could increment multiple UE error counters
> >> in such case, (meaning that an error happened on either one of the
> >> affected DIMMs/Ranks) but this is a different behavior than the
> >> current API.
> >
> > Well, the API should be changed to accomodate such configurations.
>
> True, but changing the propagation logic to propagate the error down
> to the several DIMMs from where the error might have occurred is:
>
> - the opposite of the current propagation logic;
>
> - the opposite on how ITU-T TMN architecture and all EMS/NMS
> implementations I'm aware with work.
>
> So, using such propagation logic doesn't sound right to me. What I'm
> saying is that, if all the driver can be sure is that the error happened
> at the csrow level, it should not propagate the errors to the channel
> level.
>
> So, I think that csrow-level counter is needed (and the equivalent
> "group" counters for non-rank-based memory controllers).
See above, we already have 'ce_count' and 'ue_count' and those are
csrow-level counters.
>
> Regards,
> Mauro.
>
--
Regards/Gruss,
Boris.
Advanced Micro Devices GmbH
Einsteinring 24, 85609 Dornach
GM: Alberto Bozzo
Reg: Dornach, Landkreis Muenchen
HRB Nr. 43632 WEEE Registernr: 129 19551
next prev parent reply other threads:[~2012-03-15 21:39 UTC|newest]
Thread overview: 41+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-03-07 11:40 [PATCH 0/6] Add a per-dimm structure Mauro Carvalho Chehab
2012-03-07 11:40 ` [PATCH 1/6] edac: Create a dimm struct and move the labels into it Mauro Carvalho Chehab
2012-03-07 11:40 ` [PATCH 2/6] edac: Add per dimm's sysfs nodes Mauro Carvalho Chehab
2012-03-07 11:40 ` [PATCH 3/6] edac: move dimm properties to struct memset_info Mauro Carvalho Chehab
2012-03-07 11:40 ` [PATCH 4/6] edac: Don't initialize csrow's first_page & friends when not needed Mauro Carvalho Chehab
2012-03-07 11:40 ` [PATCH 5/6] edac: move nr_pages to dimm struct Mauro Carvalho Chehab
2012-03-07 11:40 ` [PATCH 6/6] edac: Add per-dimm sysfs show nodes Mauro Carvalho Chehab
2012-03-08 21:57 ` [PATCH 0/6] Add a per-dimm structure Borislav Petkov
2012-03-09 10:32 ` Mauro Carvalho Chehab
2012-03-09 14:38 ` Borislav Petkov
2012-03-09 16:40 ` Mauro Carvalho Chehab
2012-03-09 18:47 ` Borislav Petkov
2012-03-09 19:46 ` Mauro Carvalho Chehab
2012-03-11 11:34 ` Borislav Petkov
2012-03-11 12:32 ` Mauro Carvalho Chehab
2012-03-12 16:39 ` Borislav Petkov
2012-03-12 17:03 ` Luck, Tony
2012-03-12 18:10 ` Borislav Petkov
2012-03-13 23:32 ` Greg KH
2012-03-14 19:35 ` Mauro Carvalho Chehab
2012-03-14 20:43 ` Greg KH
2012-03-14 22:20 ` Mauro Carvalho Chehab
2012-03-14 23:32 ` Greg KH
2012-03-15 2:22 ` Mauro Carvalho Chehab
2012-03-15 15:00 ` Greg KH
2012-03-14 22:31 ` Borislav Petkov
2012-03-14 22:40 ` Greg KH
2012-03-15 1:37 ` Mauro Carvalho Chehab
2012-03-15 1:44 ` Mauro Carvalho Chehab
2012-03-15 11:31 ` Borislav Petkov
2012-03-15 12:40 ` Mauro Carvalho Chehab
2012-03-15 21:38 ` Borislav Petkov [this message]
2012-03-16 8:47 ` Mauro Carvalho Chehab
2012-03-16 11:15 ` Borislav Petkov
2012-03-16 12:07 ` Mauro Carvalho Chehab
2012-03-16 14:07 ` Mauro Carvalho Chehab
2012-03-16 15:31 ` Greg KH
2012-03-16 16:54 ` Borislav Petkov
2012-03-16 15:30 ` Greg KH
2012-03-16 15:44 ` Mauro Carvalho Chehab
2012-03-16 16:01 ` Greg KH
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20120315213844.GA3781@aftab \
--to=bp@amd64.org \
--cc=gregkh@linuxfoundation.org \
--cc=linux-edac@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=mchehab@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox