From: Feng Tang <feng.tang@intel.com>
To: Yu Liao <liaoyu15@huawei.com>
Cc: Xiongfeng Wang <wangxiongfeng2@huawei.com>,
Zhang Rui <rui.zhang@intel.com>,
Thomas Gleixner <tglx@linutronix.de>,
Bjorn Helgaas <helgaas@kernel.org>,
Ingo Molnar <mingo@redhat.com>, Borislav Petkov <bp@alien8.de>,
<x86@kernel.org>, <linux-kernel@vger.kernel.org>,
"Bjorn Helgaas" <bhelgaas@google.com>,
Kai-Heng Feng <kai.heng.feng@canonical.com>,
<len.brown@intel.com>, Xie XiuQi <xiexiuqi@huawei.com>,
Kefeng Wang <wangkefeng.wang@huawei.com>
Subject: Re: [PATCH] x86/PCI: Convert force_disable_hpet() to standard quirk
Date: Fri, 30 Sep 2022 18:13:14 +0800 [thread overview]
Message-ID: <YzbBOuX37XpUiP4y@feng-clx> (raw)
In-Reply-To: <119b669e-aafb-4d73-e94e-ef119f909cfa@huawei.com>
On Fri, Sep 30, 2022 at 05:45:29PM +0800, Yu Liao wrote:
[...]
> >>>>
> >>>> Hi, Zhang Rui, we have met the same problem as you mentioned above. I have
> >>>> tested the following modification. It can solve the problem. Do you have plan
> >>>> to push it to upstream ?
> >>>
> >>> Hi Liao Yu,
> >>>
> >>> Could you provoide more details? Like, what ARCH is the platform (x86
> >>> or others), client or sever, if sever, how many sockets (2S/4S/8S)?
> >>>
> >>> The error kernel log will also be helpful.
> >>
> >> Hi, Feng Tang,
> >>
> >> It's a X86 Sever. lscpu print the following information:
> >>
> >> Architecture: x86_64
> >> CPU op-mode(s): 32-bit, 64-bit
> >> Byte Order: Little Endian
> >> Address sizes: 46 bits physical, 48 bits virtual
> >> CPU(s): 224
> >> On-line CPU(s) list: 0-223
> >> Thread(s) per core: 2
> >> Core(s) per socket: 28
> >> Socket(s): 4
> >> NUMA node(s): 4
> >> Vendor ID: GenuineIntel
> >> CPU family: 6
> >> Model: 85
> >> Model name: Intel(R) Xeon(R) Platinum 8180 CPU @ 2.50GHz
> >> Stepping: 4
> >> CPU MHz: 3199.379
> >> CPU max MHz: 3800.0000
> >> CPU min MHz: 1000.0000
> >> BogoMIPS: 5000.00
> >> Virtualization: VT-x
> >> L1d cache: 3.5 MiB
> >> L1i cache: 3.5 MiB
> >> L2 cache: 112 MiB
> >> L3 cache: 154 MiB
> >> NUMA node0 CPU(s): 0-27,112-139
> >> NUMA node1 CPU(s): 28-55,140-167
> >> NUMA node2 CPU(s): 56-83,168-195
> >> NUMA node3 CPU(s): 84-111,196-223
> >>
> >> Part of the kernel log is as follows.
> >>
> >> [ 1.144402] smp: Brought up 4 nodes, 224 CPUs
> >> [ 1.144402] smpboot: Max logical packages: 4
> >> [ 1.144402] smpboot: Total of 224 processors activated (1121097.93 BogoMIPS)
> >> [ 1.520003] clocksource: timekeeping watchdog on CPU2: Marking clocksource
> >> 'tsc-early' as unstable because the skew is too large:
> >> [ 1.520010] clocksource: 'refined-jiffies' wd_now:
> >> fffb7210 wd_last: fffb7018 mask: ffffffff
> >> [ 1.520013] clocksource: 'tsc-early' cs_now:
> >> 6606717afddd0 cs_last: 66065eff88ad4 mask: ffffffffffffffff
> >> [ 1.520015] tsc: Marking TSC unstable due to clocksource watchdog
> >> [ 5.164635] node 0 initialised, 98233092 pages in 4013ms
> >> [ 5.209294] node 3 initialised, 98923232 pages in 4057ms
> >> [ 5.220001] node 2 initialised, 99054870 pages in 4068ms
> >> [ 5.222282] node 1 initialised, 99054870 pages in 4070ms
> >
> > Thanks Xiaofeng for the info.
> >
> > Could you try the below patch? It is kinda extension of
> >
> > b50db7095fe0 ("x86/tsc: Disable clocksource watchdog for TSC on qualified platorms")
> >
> > which I have run limited test on some 4 sockets Haswell and Cascadelake
> > AP x86 servers.
> >
> >
> > Thanks,
> > Feng
> > ---
> > diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c
> > index cafacb2e58cc..b4ea79cb1d1a 100644
> > --- a/arch/x86/kernel/tsc.c
> > +++ b/arch/x86/kernel/tsc.c
> > @@ -1217,7 +1217,7 @@ static void __init check_system_tsc_reliable(void)
> > if (boot_cpu_has(X86_FEATURE_CONSTANT_TSC) &&
> > boot_cpu_has(X86_FEATURE_NONSTOP_TSC) &&
> > boot_cpu_has(X86_FEATURE_TSC_ADJUST) &&
> > - nr_online_nodes <= 2)
> > + nr_online_nodes <= 8)
> > tsc_disable_clocksource_watchdog();
> > }
> >
> >
> Hi Feng,
>
> I tested this patch on a previous server and it fixes the issue.
Thanks for the testing, please do let us know if there is any TSC
problem after long time or stress running.
Plan to send the patch for merging.
Thanks,
Feng
> Thanks,
> Yu
>
>
next prev parent reply other threads:[~2022-09-30 10:13 UTC|newest]
Thread overview: 19+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-11-19 18:19 [PATCH] x86/PCI: Convert force_disable_hpet() to standard quirk Bjorn Helgaas
2020-11-24 23:27 ` Bjorn Helgaas
2020-11-25 12:46 ` Thomas Gleixner
2020-11-25 19:13 ` Bjorn Helgaas
2020-11-26 0:50 ` Thomas Gleixner
2020-11-26 1:24 ` Feng Tang
2020-11-26 23:27 ` Thomas Gleixner
2020-11-27 6:11 ` Feng Tang
2020-11-30 19:21 ` Thomas Gleixner
2020-12-01 8:34 ` Feng Tang
2020-12-02 7:28 ` Zhang Rui
2022-09-29 15:52 ` Yu Liao
2022-09-30 0:38 ` Feng Tang
2022-09-30 1:05 ` Xiongfeng Wang
2022-09-30 1:15 ` Feng Tang
2022-09-30 9:45 ` Yu Liao
2022-09-30 10:13 ` Feng Tang [this message]
2022-10-01 5:18 ` Zhang Rui
2022-10-01 12:00 ` Feng Tang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=YzbBOuX37XpUiP4y@feng-clx \
--to=feng.tang@intel.com \
--cc=bhelgaas@google.com \
--cc=bp@alien8.de \
--cc=helgaas@kernel.org \
--cc=kai.heng.feng@canonical.com \
--cc=len.brown@intel.com \
--cc=liaoyu15@huawei.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@redhat.com \
--cc=rui.zhang@intel.com \
--cc=tglx@linutronix.de \
--cc=wangkefeng.wang@huawei.com \
--cc=wangxiongfeng2@huawei.com \
--cc=x86@kernel.org \
--cc=xiexiuqi@huawei.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox