From: Jonathan Cameron <Jonathan.Cameron@Huawei.com>
To: Dan Williams <dan.j.williams@intel.com>,
Vikram Sethi <vsethi@nvidia.com>
Cc: <linux-cxl@vger.kernel.org>
Subject: Re: [PATCH 1/3] cxl/region: Move cache invalidation before region teardown, and before setup
Date: Thu, 22 Jun 2023 10:00:19 +0100 [thread overview]
Message-ID: <20230622100019.00001f6c@Huawei.com> (raw)
In-Reply-To: <168696506886.3590522.4597053660991916591.stgit@dwillia2-xfh.jf.intel.com>
On Fri, 16 Jun 2023 18:24:28 -0700
Dan Williams <dan.j.williams@intel.com> wrote:
> Vikram raised a concern with the theoretical case of a CPU sending
> MemClnEvict to a device that is not prepared to receive. MemClnEvict is
> a message that is sent after a CPU has taken ownership of a cacheline
> from accelerator memory (HDM-DB). In the case of hotplug or HDM decoder
> reconfiguration it is possible that the CPU is holding old contents for
> a new device that has taken over the physical address range being cached
> by the CPU.
>
> To avoid this scenario, invalidate caches prior to tearing down an HDM
> decoder configuration.
>
> Now, this poses another problem that it is possible for something to
> speculate into that space while the decode configuration is still up, so
> to close that gap also invalidate prior to establish new contents behind
> a given physical address range.
>
> With this change the cache invalidation is now explicit and need not be
> checked in cxl_region_probe(), and that obviates the need for
> CXL_REGION_F_INCOHERENT.
>
> Cc: Jonathan Cameron <Jonathan.Cameron@Huawei.com>
> Fixes: d18bc74aced6 ("cxl/region: Manage CPU caches relative to DPA invalidation events")
> Reported-by: Vikram Sethi <vsethi@nvidia.com>
> Closes: http://lore.kernel.org/r/BYAPR12MB33364B5EB908BF7239BB996BBD53A@BYAPR12MB3336.namprd12.prod.outlook.com
> Signed-off-by: Dan Williams <dan.j.williams@intel.com>
A far as I can tell (I feel my brain slowly frying when considering the speculation
opportunities) this should do the job.
Reviewed-by: Jonathan Cameron <Jonathan.Cameron@huawei.com>
p.s. it sucks - architecture folk, we need less destructive ways of doing this...
> ---
> drivers/cxl/core/region.c | 66 +++++++++++++++++++++++++--------------------
> drivers/cxl/cxl.h | 8 +----
> 2 files changed, 38 insertions(+), 36 deletions(-)
>
> diff --git a/drivers/cxl/core/region.c b/drivers/cxl/core/region.c
> index 013f3656e661..d48c5916f2e9 100644
> --- a/drivers/cxl/core/region.c
> +++ b/drivers/cxl/core/region.c
> @@ -125,10 +125,38 @@ static struct cxl_region_ref *cxl_rr_load(struct cxl_port *port,
> return xa_load(&port->regions, (unsigned long)cxlr);
> }
>
> +static int cxl_region_invalidate_memregion(struct cxl_region *cxlr)
> +{
> + if (!cpu_cache_has_invalidate_memregion()) {
> + if (IS_ENABLED(CONFIG_CXL_REGION_INVALIDATION_TEST)) {
> + dev_warn_once(
> + &cxlr->dev,
> + "Bypassing cpu_cache_invalidate_memregion() for testing!\n");
> + return 0;
> + } else {
> + dev_err(&cxlr->dev,
> + "Failed to synchronize CPU cache state\n");
> + return -ENXIO;
> + }
> + }
> +
> + cpu_cache_invalidate_memregion(IORES_DESC_CXL);
> + return 0;
> +}
> +
> static int cxl_region_decode_reset(struct cxl_region *cxlr, int count)
> {
> struct cxl_region_params *p = &cxlr->params;
> - int i;
> + int i, rc = 0;
> +
> + /*
> + * Before region teardown attempt to flush, and if the flush
> + * fails cancel the region teardown for data consistency
> + * concerns
> + */
> + rc = cxl_region_invalidate_memregion(cxlr);
> + if (rc)
> + return rc;
>
> for (i = count - 1; i >= 0; i--) {
> struct cxl_endpoint_decoder *cxled = p->targets[i];
> @@ -136,7 +164,6 @@ static int cxl_region_decode_reset(struct cxl_region *cxlr, int count)
> struct cxl_port *iter = cxled_to_port(cxled);
> struct cxl_dev_state *cxlds = cxlmd->cxlds;
> struct cxl_ep *ep;
> - int rc = 0;
>
> if (cxlds->rcd)
> goto endpoint_reset;
> @@ -256,6 +283,14 @@ static ssize_t commit_store(struct device *dev, struct device_attribute *attr,
> goto out;
> }
>
> + /*
> + * Invalidate caches before region setup to drop any speculative
> + * consumption of this address space
> + */
> + rc = cxl_region_invalidate_memregion(cxlr);
> + if (rc)
> + return rc;
> +
> if (commit)
> rc = cxl_region_decode_commit(cxlr);
> else {
> @@ -1686,7 +1721,6 @@ static int cxl_region_attach(struct cxl_region *cxlr,
> if (rc)
> goto err_decrement;
> p->state = CXL_CONFIG_ACTIVE;
> - set_bit(CXL_REGION_F_INCOHERENT, &cxlr->flags);
> }
>
> cxled->cxld.interleave_ways = p->interleave_ways;
> @@ -2815,30 +2849,6 @@ int cxl_add_to_region(struct cxl_port *root, struct cxl_endpoint_decoder *cxled)
> }
> EXPORT_SYMBOL_NS_GPL(cxl_add_to_region, CXL);
>
> -static int cxl_region_invalidate_memregion(struct cxl_region *cxlr)
> -{
> - if (!test_bit(CXL_REGION_F_INCOHERENT, &cxlr->flags))
> - return 0;
> -
> - if (!cpu_cache_has_invalidate_memregion()) {
> - if (IS_ENABLED(CONFIG_CXL_REGION_INVALIDATION_TEST)) {
> - dev_warn_once(
> - &cxlr->dev,
> - "Bypassing cpu_cache_invalidate_memregion() for testing!\n");
> - clear_bit(CXL_REGION_F_INCOHERENT, &cxlr->flags);
> - return 0;
> - } else {
> - dev_err(&cxlr->dev,
> - "Failed to synchronize CPU cache state\n");
> - return -ENXIO;
> - }
> - }
> -
> - cpu_cache_invalidate_memregion(IORES_DESC_CXL);
> - clear_bit(CXL_REGION_F_INCOHERENT, &cxlr->flags);
> - return 0;
> -}
> -
> static int is_system_ram(struct resource *res, void *arg)
> {
> struct cxl_region *cxlr = arg;
> @@ -2866,8 +2876,6 @@ static int cxl_region_probe(struct device *dev)
> goto out;
> }
>
> - rc = cxl_region_invalidate_memregion(cxlr);
> -
> /*
> * From this point on any path that changes the region's state away from
> * CXL_CONFIG_COMMIT is also responsible for releasing the driver.
> diff --git a/drivers/cxl/cxl.h b/drivers/cxl/cxl.h
> index f0c428cb9a71..187abd8ba673 100644
> --- a/drivers/cxl/cxl.h
> +++ b/drivers/cxl/cxl.h
> @@ -462,18 +462,12 @@ struct cxl_region_params {
> int nr_targets;
> };
>
> -/*
> - * Flag whether this region needs to have its HPA span synchronized with
> - * CPU cache state at region activation time.
> - */
> -#define CXL_REGION_F_INCOHERENT 0
> -
> /*
> * Indicate whether this region has been assembled by autodetection or
> * userspace assembly. Prevent endpoint decoders outside of automatic
> * detection from being added to the region.
> */
> -#define CXL_REGION_F_AUTO 1
> +#define CXL_REGION_F_AUTO 0
>
> /**
> * struct cxl_region - CXL region
>
next prev parent reply other threads:[~2023-06-22 9:09 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-06-17 1:24 [PATCH 0/3] cxl/region: Cache management and region decode reset fixes Dan Williams
2023-06-17 1:24 ` [PATCH 1/3] cxl/region: Move cache invalidation before region teardown, and before setup Dan Williams
2023-06-21 0:00 ` Dave Jiang
2023-06-22 9:00 ` Jonathan Cameron [this message]
2023-06-17 1:24 ` [PATCH 2/3] cxl/region: Flag partially torn down regions as unusable Dan Williams
2023-06-21 0:04 ` Dave Jiang
2023-06-22 9:14 ` Jonathan Cameron
2023-06-17 1:24 ` [PATCH 3/3] cxl/region: Fix state transitions after reset failure Dan Williams
2023-06-21 0:06 ` Dave Jiang
2023-06-22 9:18 ` Jonathan Cameron
2023-06-25 20:42 ` Dan Williams
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20230622100019.00001f6c@Huawei.com \
--to=jonathan.cameron@huawei.com \
--cc=dan.j.williams@intel.com \
--cc=linux-cxl@vger.kernel.org \
--cc=vsethi@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox