public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: tip-bot for Fenghua Yu <tipbot@zytor.com>
To: linux-tip-commits@vger.kernel.org
Cc: linux-kernel@vger.kernel.org, vikas.shivappa@linux.intel.com,
	fenghua.yu@intel.com, tglx@linutronix.de, mingo@kernel.org,
	hpa@zytor.com
Subject: [tip:x86/cache] x86/intel_rdt: Intel haswell Cache Allocation enumeration
Date: Fri, 18 Dec 2015 13:36:30 -0800	[thread overview]
Message-ID: <tip-8741b655628d89380bfbe0ded7a83c0bc2293a72@git.kernel.org> (raw)
In-Reply-To: <1450392376-6397-10-git-send-email-fenghua.yu@intel.com>

Commit-ID:  8741b655628d89380bfbe0ded7a83c0bc2293a72
Gitweb:     http://git.kernel.org/tip/8741b655628d89380bfbe0ded7a83c0bc2293a72
Author:     Fenghua Yu <fenghua.yu@intel.com>
AuthorDate: Thu, 17 Dec 2015 14:46:14 -0800
Committer:  H. Peter Anvin <hpa@linux.intel.com>
CommitDate: Fri, 18 Dec 2015 13:17:56 -0800

x86/intel_rdt: Intel haswell Cache Allocation enumeration

From: Vikas Shivappa <vikas.shivappa@linux.intel.com>

This patch is specific to Intel haswell (hsw) server SKUs. Cache
Allocation on hsw server needs to be enumerated separately as HSW does
not have support for CPUID enumeration for Cache Allocation. This patch
does a probe by writing a CLOSid (Class of service id) into high 32 bits
of IA32_PQR_MSR and see if the bits stick. The probe is only done after
confirming that the CPU is HSW server. Other hardcoded values are:

 - L3 cache bit mask must be at least two bits.
 - Maximum CLOSids supported is always 4.
 - Maximum bits support in cache bit mask is always 20.

Signed-off-by: Vikas Shivappa <vikas.shivappa@linux.intel.com>
Link: http://lkml.kernel.org/r/1450392376-6397-10-git-send-email-fenghua.yu@intel.com
Signed-off-by: Fenghua Yu <fenghua.yu@intel.com>
---
 arch/x86/kernel/cpu/intel_rdt.c | 59 +++++++++++++++++++++++++++++++++++++++--
 1 file changed, 57 insertions(+), 2 deletions(-)

diff --git a/arch/x86/kernel/cpu/intel_rdt.c b/arch/x86/kernel/cpu/intel_rdt.c
index 31f8588..ecaf8e6 100644
--- a/arch/x86/kernel/cpu/intel_rdt.c
+++ b/arch/x86/kernel/cpu/intel_rdt.c
@@ -38,6 +38,10 @@ static struct clos_cbm_table *cctable;
  */
 unsigned long *closmap;
 /*
+ * Minimum bits required in Cache bitmask.
+ */
+static unsigned int min_bitmask_len = 1;
+/*
  * Mask of CPUs for writing CBM values. We only need one CPU per-socket.
  */
 static cpumask_t rdt_cpumask;
@@ -54,6 +58,57 @@ struct rdt_remote_data {
 	u64 val;
 };
 
+/*
+ * cache_alloc_hsw_probe() - Have to probe for Intel haswell server CPUs
+ * as it does not have CPUID enumeration support for Cache allocation.
+ *
+ * Probes by writing to the high 32 bits(CLOSid) of the IA32_PQR_MSR and
+ * testing if the bits stick. Max CLOSids is always 4 and max cbm length
+ * is always 20 on hsw server parts. The minimum cache bitmask length
+ * allowed for HSW server is always 2 bits. Hardcode all of them.
+ */
+static inline bool cache_alloc_hsw_probe(void)
+{
+	u32 l, h_old, h_new, h_tmp;
+
+	if (rdmsr_safe(MSR_IA32_PQR_ASSOC, &l, &h_old))
+		return false;
+
+	/*
+	 * Default value is always 0 if feature is present.
+	 */
+	h_tmp = h_old ^ 0x1U;
+	if (wrmsr_safe(MSR_IA32_PQR_ASSOC, l, h_tmp) ||
+	    rdmsr_safe(MSR_IA32_PQR_ASSOC, &l, &h_new))
+		return false;
+
+	if (h_tmp != h_new)
+		return false;
+
+	wrmsr_safe(MSR_IA32_PQR_ASSOC, l, h_old);
+
+	boot_cpu_data.x86_cache_max_closid = 4;
+	boot_cpu_data.x86_cache_max_cbm_len = 20;
+	min_bitmask_len = 2;
+
+	return true;
+}
+
+static inline bool cache_alloc_supported(struct cpuinfo_x86 *c)
+{
+	if (cpu_has(c, X86_FEATURE_CAT_L3))
+		return true;
+
+	/*
+	 * Probe for Haswell server CPUs.
+	 */
+	if (c->x86 == 0x6 && c->x86_model == 0x3f)
+		return cache_alloc_hsw_probe();
+
+	return false;
+}
+
+
 void __intel_rdt_sched_in(void *dummy)
 {
 	struct intel_pqr_state *state = this_cpu_ptr(&pqr_state);
@@ -126,7 +181,7 @@ static bool cbm_validate(unsigned long var)
 	unsigned long first_bit, zero_bit;
 	u64 max_cbm;
 
-	if (bitmap_weight(&var, max_cbm_len) < 1)
+	if (bitmap_weight(&var, max_cbm_len) < min_bitmask_len)
 		return false;
 
 	max_cbm = (1ULL << max_cbm_len) - 1;
@@ -310,7 +365,7 @@ static int __init intel_rdt_late_init(void)
 	u32 maxid, max_cbm_len;
 	int err = 0, size, i;
 
-	if (!cpu_has(c, X86_FEATURE_CAT_L3))
+	if (!cache_alloc_supported(c))
 		return -ENODEV;
 
 	maxid = c->x86_cache_max_closid;

  reply	other threads:[~2015-12-18 21:36 UTC|newest]

Thread overview: 36+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2015-12-17 22:46 [PATCH V16 00/11] x86: Intel Cache Allocation Technology Support Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 01/11] x86/intel_cqm: Modify hot cpu notification handling Fenghua Yu
2015-12-18 21:33   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 02/11] x86/intel_rapl: " Fenghua Yu
2015-12-18 21:34   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 03/11] x86/intel_rdt: Cache Allocation documentation Fenghua Yu
2015-12-18 21:34   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 04/11] x86/intel_rdt: Add support for Cache Allocation detection Fenghua Yu
2015-12-18 21:34   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 05/11] x86/intel_rdt: Add Class of service management Fenghua Yu
2015-12-18 21:35   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 06/11] x86/intel_rdt: Add L3 cache capacity bitmask management Fenghua Yu
2015-12-18 21:35   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 07/11] x86/intel_rdt: Implement scheduling support for Intel RDT Fenghua Yu
2015-12-18 21:35   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 08/11] x86/intel_rdt: Hot cpu support for Cache Allocation Fenghua Yu
2015-12-18 21:36   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 09/11] x86/intel_rdt: Intel haswell Cache Allocation enumeration Fenghua Yu
2015-12-18 21:36   ` tip-bot for Fenghua Yu [this message]
2015-12-17 22:46 ` [PATCH V16 10/11] x86,cgroup/intel_rdt : Add intel_rdt cgroup documentation Fenghua Yu
2015-12-18 21:36   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-17 22:46 ` [PATCH V16 11/11] x86,cgroup/intel_rdt : Add a cgroup interface to manage Intel cache allocation Fenghua Yu
2015-12-18 21:37   ` [tip:x86/cache] " tip-bot for Fenghua Yu
2015-12-19 10:42   ` [PATCH V16 11/11] " Thomas Gleixner
2015-12-20  0:57     ` Marcelo Tosatti
2015-12-21 13:44       ` Thomas Gleixner
2015-12-21 15:48       ` Luiz Capitulino
2015-12-21 17:05         ` Marcelo Tosatti
2016-01-02 22:53           ` Richard Weinberger
2016-01-04 21:44             ` Yu, Fenghua
2016-01-04 21:47               ` Richard Weinberger
2015-12-18 17:45 ` [PATCH V16 00/11] x86: Intel Cache Allocation Technology Support Christoph Lameter
2015-12-18 20:49   ` Marcelo Tosatti
2015-12-21 12:53     ` Christoph Lameter
2015-12-21 15:55 ` Luiz Capitulino
2015-12-23 15:50 ` Tejun Heo

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=tip-8741b655628d89380bfbe0ded7a83c0bc2293a72@git.kernel.org \
    --to=tipbot@zytor.com \
    --cc=fenghua.yu@intel.com \
    --cc=hpa@zytor.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-tip-commits@vger.kernel.org \
    --cc=mingo@kernel.org \
    --cc=tglx@linutronix.de \
    --cc=vikas.shivappa@linux.intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox