From: Dave Jiang <dave.jiang@intel.com>
To: Shiyang Ruan <ruansy.fnst@fujitsu.com>,
Dan Williams <dan.j.williams@intel.com>,
qemu-devel@nongnu.org, linux-cxl@vger.kernel.org
Cc: jonathan.cameron@huawei.com, dave@stgolabs.net,
ira.weiny@intel.com, alison.schofield@intel.com,
vishal.l.verma@intel.com
Subject: Re: [RFC PATCH] cxl: avoid duplicating report from MCE & device
Date: Fri, 19 Jul 2024 09:04:50 -0700 [thread overview]
Message-ID: <ec0f2e8d-e783-4211-8d41-2b9a6df67049@intel.com> (raw)
In-Reply-To: <1fab087b-eb6c-485e-a2ed-f86e8dfacc5d@fujitsu.com>
On 7/1/24 7:12 PM, Shiyang Ruan wrote:
>
>
> 在 2024/6/25 21:56, Shiyang Ruan 写道:
>>
>>
>> 在 2024/6/22 1:51, Dan Williams 写道:
>>> Shiyang Ruan wrote:
>>>> Background:
>>>> Since CXL device is a memory device, while CPU consumes a poison page of
>>>> CXL device, it always triggers a MCE by interrupt (INT18), no matter
>>>> which-First path is configured. This is the first report. Then
>>>> currently, in FW-First path, the poison event is transferred according
>>>> to the following process: CXL device -> firmware -> OS:ACPI->APEI->GHES
>>>> -> CPER -> trace report. This is the second one. These two reports
>>>> are indicating the same poisoning page, which is the so-called "duplicate
>>>> report"[1]. And the memory_failure() handling I'm trying to add in
>>>> OS-First path could also be another duplicate report.
>>>>
>>>> Hope the flow below could make it easier to understand:
>>>> CPU accesses bad memory on CXL device, then
>>>> -> MCE (INT18), *always* report (1)
>>>> -> * FW-First (implemented now)
>>>> -> CXL device -> FW
>>>> -> OS:ACPI->APEI->GHES->CPER -> trace report (2.a)
>>>> * OS-First (not implemented yet, I'm working on it)
>>>> -> CXL device -> MSI
>>>> -> OS:CXL driver -> memory_failure() (2.b)
>>>> so, the (1) and (2.a/b) are duplicated.
>>>>
>>>> (I didn't get response in my reply for [1] while I have to make patch to
>>>> solve this problem, so please correct me if my understanding is wrong.)
>>>
>>> The CPU MCE may not be in the loop. Consider the case of patrol scrub,
>>> or device-DMA accessing poison. In that case the device will signal a
>>> component event and the CPU may never issue the MCE.
>>
>> My other patch: "cxl/core: add poison creation event handler", adds calling memory_failure() when an event is received form device, which checks the poison record (insert if not exists) to make sure poison would be reported/handled, but not twice, no matter CPU issues the MCE later or earlier. And the lock of poison record makes sure that it's fine even in race condition.
>>
>>>
>>> What is missing for me from this description is *why* does the duplicate
>>> report matter in practice? If all that happens is that the kernel
>>> repeats the lookup to offline the page and set the HWPoison bit, is that
>>> duplicated work worth adding more tracking?
>>
>> Besides setting the HWPoison bit for the poison page, memory_failure() also finds and notifies those processes who are accessing the poison page, and tries to recovery the page. And there seems no lock to prevent the 2nd memory_failure() and clearing poison operation from being called in race, then the HWPoison bit could be unsure. I think that's the problem.
>>
>> > So, I think all CXL poison notification events should trigger an
>> > action optional memory_failure(). I expect this needs to make sure
>> > that duplicates re not a problem. I.e. in the case of CPU consumption
>> > of CXL poison, that causes a synchronous MF_ACTION_REQUIRED event via
>> > the MCE path *and* it may trigger the device to send an error record
>> > for the same page. As far as I can see, duplicate reports (MCE + CXL
>> > device) are unavoidable.
>>
>> As you mentioned in my other patch, this problem should be solved at first. My patch adds the poison record (tracking, which might not become very large) to prevent duplicating report.
>
> Ping~
>
> And I had some problems when using xarray, please see below.
>
>>
>>>
>>>> This patch adds a new notifier_block and MCE_PRIO_CXL, for CXL memdev
>>>> to check whether the current poison page has been reported (if yes,
>>>> stop the notifier chain, won't call the following memory_failure()
>>>> to report), into `x86_mce_decoder_chain`. In this way, if the poison
>>>> page already handled(recorded and reported) in (1) or (2), the other one
>>>> won't duplicate the report. The record could be clear when
>>>> cxl_clear_poison() is called.
>>>>
>>>> [1] https://lore.kernel.org/linux-cxl/664d948fb86f0_e8be294f8@dwillia2-mobl3.amr.corp.intel.com.notmuch/
>>>>
>>>> Signed-off-by: Shiyang Ruan <ruansy.fnst@fujitsu.com>
>>>> ---
>>>> arch/x86/include/asm/mce.h | 1 +
>>>> drivers/cxl/core/mbox.c | 130 +++++++++++++++++++++++++++++++++++++
>>>> drivers/cxl/core/memdev.c | 6 +-
>>>> drivers/cxl/cxlmem.h | 3 +
>>>> 4 files changed, 139 insertions(+), 1 deletion(-)
>>>>
>>>> diff --git a/arch/x86/include/asm/mce.h b/arch/x86/include/asm/mce.h
>>>> index dfd2e9699bd7..d8109c48e7d9 100644
>>>> --- a/arch/x86/include/asm/mce.h
>>>> +++ b/arch/x86/include/asm/mce.h
>>>> @@ -182,6 +182,7 @@ enum mce_notifier_prios {
>>>> MCE_PRIO_NFIT,
>>>> MCE_PRIO_EXTLOG,
>>>> MCE_PRIO_UC,
>>>> + MCE_PRIO_CXL,
>>>> MCE_PRIO_EARLY,
>>>> MCE_PRIO_CEC,
>>>> MCE_PRIO_HIGHEST = MCE_PRIO_CEC
>>>> diff --git a/drivers/cxl/core/mbox.c b/drivers/cxl/core/mbox.c
>>>> index 2626f3fff201..0eb3c5401e81 100644
>>>> --- a/drivers/cxl/core/mbox.c
>>>> +++ b/drivers/cxl/core/mbox.c
>>>> @@ -4,6 +4,8 @@
>>>> #include <linux/debugfs.h>
>>>> #include <linux/ktime.h>
>>>> #include <linux/mutex.h>
>>>> +#include <linux/notifier.h>
>>>> +#include <asm/mce.h>
>>>> #include <asm/unaligned.h>
>>>> #include <cxlpci.h>
>>>> #include <cxlmem.h>
>>>> @@ -880,6 +882,9 @@ void cxl_event_trace_record(const struct cxl_memdev *cxlmd,
>>>> if (cxlr)
>>>> hpa = cxl_trace_hpa(cxlr, cxlmd, dpa);
>>>> + if (hpa != ULLONG_MAX && cxl_mce_recorded(hpa))
>>>> + return;
>>>> +
>>>> if (event_type == CXL_CPER_EVENT_GEN_MEDIA)
>>>> trace_cxl_general_media(cxlmd, type, cxlr, hpa,
>>>> &evt->gen_media);
>>>> @@ -1408,6 +1413,127 @@ int cxl_poison_state_init(struct cxl_memdev_state *mds)
>>>> }
>>>> EXPORT_SYMBOL_NS_GPL(cxl_poison_state_init, CXL);
>>>> +struct cxl_mce_record {
>>>> + struct list_head node;
>>>> + u64 hpa;
>>>> +};
>>>> +LIST_HEAD(cxl_mce_records);
>>>> +DEFINE_MUTEX(cxl_mce_mutex);
>>>
>>> I would recommend an xarray for this use case as that already has its
>>> own internal locking and efficient memory allocation for new nodes.
>>>
>>> However, the "why" question needs to be answered first.
>>
>> xarray does look better. I didn't think of it befre. Thanks for suggestion.
>
> I'm trying using xarray but but I'm running into two problems.
>
> The first one is: xarray stores an entry with a specified index, but here we only need to store the PFN. I'm not sure how to specific an index for a PFN. So I think xarray might not suitable for my case.
> The second one: because we only need to store/search PFN, the list node only contains a list_head and a PFN. But xarray seems to require more memory for each node, which causes more overhead.
Use PFN as the index of the xarray since it's unique. And that makes finding the entry easy with xarray because you just pass in the PFN as the key. You can store an xarray entry with no data. At this point you are just detecting whether the entry has been stored (valid) or not. Hope that helps.
>
> Maybe I'm overthinking this?
>
>
> --
> Thanks
> Ruan.
>
>>
>> --
>> Thanks
>> Ruan.
next prev parent reply other threads:[~2024-07-19 16:04 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-06-18 16:53 [RFC PATCH] cxl: avoid duplicating report from MCE & device Shiyang Ruan
2024-06-18 23:35 ` Dave Jiang
2024-06-19 9:24 ` Shiyang Ruan
2024-06-20 15:51 ` Dave Jiang
2024-06-21 10:18 ` Shiyang Ruan
2024-06-20 17:02 ` Jonathan Cameron
2024-06-21 10:16 ` Shiyang Ruan
2024-06-21 17:21 ` Jonathan Cameron
2024-06-21 17:59 ` Dan Williams
2024-06-21 18:45 ` Jonathan Cameron
2024-06-21 20:44 ` Luck, Tony
2024-06-26 6:03 ` Shiyang Ruan
2024-06-26 15:56 ` Luck, Tony
2024-06-21 17:51 ` Dan Williams
2024-06-25 13:56 ` Shiyang Ruan
2024-07-02 2:12 ` Shiyang Ruan
2024-07-19 16:04 ` Dave Jiang [this message]
2024-07-22 7:01 ` Shiyang Ruan
2024-07-25 2:51 ` Yasunori Gotou (Fujitsu)
2024-07-19 6:24 ` Shiyang Ruan
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=ec0f2e8d-e783-4211-8d41-2b9a6df67049@intel.com \
--to=dave.jiang@intel.com \
--cc=alison.schofield@intel.com \
--cc=dan.j.williams@intel.com \
--cc=dave@stgolabs.net \
--cc=ira.weiny@intel.com \
--cc=jonathan.cameron@huawei.com \
--cc=linux-cxl@vger.kernel.org \
--cc=qemu-devel@nongnu.org \
--cc=ruansy.fnst@fujitsu.com \
--cc=vishal.l.verma@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox