From: Jonathan Cameron <Jonathan.Cameron@huawei.com>
To: Jean-Philippe Brucker <jean-philippe.brucker@arm.com>
Cc: <linux-arm-kernel@lists.infradead.org>,
<linux-pci@vger.kernel.org>, <linux-acpi@vger.kernel.org>,
<devicetree@vger.kernel.org>, <iommu@lists.linux-foundation.org>,
<kvm@vger.kernel.org>, <joro@8bytes.org>, <robh+dt@kernel.org>,
<mark.rutland@arm.com>, <catalin.marinas@arm.com>,
<will.deacon@arm.com>, <lorenzo.pieralisi@arm.com>,
<hanjun.guo@linaro.org>, <sudeep.holla@arm.com>,
<rjw@rjwysocki.net>, <lenb@kernel.org>, <robin.murphy@arm.com>,
<bhelgaas@google.com>, <alex.williamson@redhat.com>,
<tn@semihalf.com>, <liubo95@huawei.com>,
<thunder.leizhen@huawei.com>, <xieyisheng1@huawei.com>,
<xuzaibo@huawei.com>, <ilias.apalodimas@linaro.org>,
<shunyong.yang@hxt-semitech.com>, <nwatters@codeaurora.org>,
<okaya@codeaurora.org>, <jcrouse@codeaurora.org>,
<rfranz@cavium.com>, <dwmw2@infradead.org>,
<jacob.jun.pan@linux.intel.com>, <yi.l.liu@intel.com>,
<ashok.raj@intel.com>, <robdclark@gmail.com>,
<christian.koenig@amd.com>, <bharatku@xilinx.com>
Subject: Re: [PATCH 07/37] iommu: Add a page fault handler
Date: Thu, 8 Mar 2018 15:40:35 +0000 [thread overview]
Message-ID: <20180308164035.000065c2@huawei.com> (raw)
In-Reply-To: <20180212183352.22730-8-jean-philippe.brucker@arm.com>
On Mon, 12 Feb 2018 18:33:22 +0000
Jean-Philippe Brucker <jean-philippe.brucker@arm.com> wrote:
> Some systems allow devices to handle IOMMU translation faults in the core
> mm. For example systems supporting the PCI PRI extension or Arm SMMU stall
> model. Infrastructure for reporting such recoverable page faults was
> recently added to the IOMMU core, for SVA virtualization. Extend
> iommu_report_device_fault() to handle host page faults as well.
>
> * IOMMU drivers instantiate a fault workqueue, using
> iommu_fault_queue_init() and iommu_fault_queue_destroy().
>
> * When it receives a fault event, supposedly in an IRQ handler, the IOMMU
> driver reports the fault using iommu_report_device_fault()
>
> * If the device driver registered a handler (e.g. VFIO), pass down the
> fault event. Otherwise submit it to the fault queue, to be handled in a
> thread.
>
> * When the fault corresponds to an io_mm, call the mm fault handler on it
> (in next patch).
>
> * Once the fault is handled, the mm wrapper or the device driver reports
> success of failure with iommu_page_response(). The translation is either
> retried or aborted, depending on the response code.
>
> Signed-off-by: Jean-Philippe Brucker <jean-philippe.brucker@arm.com>
A few really minor points inline... Basically looks good to me.
> ---
> drivers/iommu/Kconfig | 10 ++
> drivers/iommu/Makefile | 1 +
> drivers/iommu/io-pgfault.c | 282 +++++++++++++++++++++++++++++++++++++++++++++
> drivers/iommu/iommu-sva.c | 3 -
> drivers/iommu/iommu.c | 31 ++---
> include/linux/iommu.h | 34 +++++-
> 6 files changed, 339 insertions(+), 22 deletions(-)
> create mode 100644 drivers/iommu/io-pgfault.c
>
> diff --git a/drivers/iommu/Kconfig b/drivers/iommu/Kconfig
> index 146eebe9a4bb..e751bb9958ba 100644
> --- a/drivers/iommu/Kconfig
> +++ b/drivers/iommu/Kconfig
> @@ -85,6 +85,15 @@ config IOMMU_SVA
>
> If unsure, say N here.
>
> +config IOMMU_FAULT
> + bool "Fault handler for the IOMMU API"
> + select IOMMU_API
> + help
> + Enable the generic fault handler for the IOMMU API, that handles
> + recoverable page faults or inject them into guests.
> +
> + If unsure, say N here.
> +
> config FSL_PAMU
> bool "Freescale IOMMU support"
> depends on PCI
> @@ -156,6 +165,7 @@ config INTEL_IOMMU
> select IOMMU_API
> select IOMMU_IOVA
> select DMAR_TABLE
> + select IOMMU_FAULT
> help
> DMA remapping (DMAR) devices support enables independent address
> translations for Direct Memory Access (DMA) from devices.
> diff --git a/drivers/iommu/Makefile b/drivers/iommu/Makefile
> index 1dbcc89ebe4c..f4324e29035e 100644
> --- a/drivers/iommu/Makefile
> +++ b/drivers/iommu/Makefile
> @@ -4,6 +4,7 @@ obj-$(CONFIG_IOMMU_API) += iommu-traces.o
> obj-$(CONFIG_IOMMU_API) += iommu-sysfs.o
> obj-$(CONFIG_IOMMU_DMA) += dma-iommu.o
> obj-$(CONFIG_IOMMU_SVA) += iommu-sva.o
> +obj-$(CONFIG_IOMMU_FAULT) += io-pgfault.o
> obj-$(CONFIG_IOMMU_IO_PGTABLE) += io-pgtable.o
> obj-$(CONFIG_IOMMU_IO_PGTABLE_ARMV7S) += io-pgtable-arm-v7s.o
> obj-$(CONFIG_IOMMU_IO_PGTABLE_LPAE) += io-pgtable-arm.o
> diff --git a/drivers/iommu/io-pgfault.c b/drivers/iommu/io-pgfault.c
> new file mode 100644
> index 000000000000..33309ed316d2
> --- /dev/null
> +++ b/drivers/iommu/io-pgfault.c
> @@ -0,0 +1,282 @@
> +/*
> + * Handle device page faults
> + *
> + * Copyright (C) 2018 ARM Ltd.
> + * Author: Jean-Philippe Brucker <jean-philippe.brucker@arm.com>
> + *
> + * SPDX-License-Identifier: GPL-2.0
> + */
> +
> +#include <linux/iommu.h>
> +#include <linux/list.h>
> +#include <linux/slab.h>
> +#include <linux/workqueue.h>
> +
> +static struct workqueue_struct *iommu_fault_queue;
> +static DECLARE_RWSEM(iommu_fault_queue_sem);
> +static refcount_t iommu_fault_queue_refs = REFCOUNT_INIT(0);
> +static BLOCKING_NOTIFIER_HEAD(iommu_fault_queue_flush_notifiers);
> +
> +/* Used to store incomplete fault groups */
> +static LIST_HEAD(iommu_partial_faults);
> +static DEFINE_SPINLOCK(iommu_partial_faults_lock);
> +
> +struct iommu_fault_context {
> + struct device *dev;
> + struct iommu_fault_event evt;
> + struct list_head head;
> +};
> +
> +struct iommu_fault_group {
> + struct iommu_domain *domain;
> + struct iommu_fault_context last_fault;
> + struct list_head faults;
> + struct work_struct work;
> +};
> +
> +/*
> + * iommu_fault_complete() - Finish handling a fault
> + *
> + * Send a response if necessary and pass on the sanitized status code
> + */
> +static int iommu_fault_complete(struct iommu_domain *domain, struct device *dev,
> + struct iommu_fault_event *evt, int status)
> +{
> + struct page_response_msg resp = {
> + .addr = evt->addr,
> + .pasid = evt->pasid,
> + .pasid_present = evt->pasid_valid,
> + .page_req_group_id = evt->page_req_group_id,
Really trivial, but if you want to align the equals signs, the all need indenting
one more tab.
> + .type = IOMMU_PAGE_GROUP_RESP,
> + .private_data = evt->iommu_private,
> + };
> +
> + /*
> + * There is no "handling" an unrecoverable fault, so the only valid
> + * return values are 0 or an error.
> + */
> + if (evt->type == IOMMU_FAULT_DMA_UNRECOV)
> + return status > 0 ? 0 : status;
> +
> + /* Someone took ownership of the fault and will complete it later */
> + if (status == IOMMU_PAGE_RESP_HANDLED)
> + return 0;
> +
> + /*
> + * There was an internal error with handling the recoverable fault. Try
> + * to complete the fault if possible.
> + */
> + if (status < 0)
> + status = IOMMU_PAGE_RESP_INVALID;
> +
> + if (WARN_ON(!domain->ops->page_response))
> + /*
> + * The IOMMU driver shouldn't have submitted recoverable faults
> + * if it cannot receive a response.
> + */
> + return -EINVAL;
> +
> + resp.resp_code = status;
> + return domain->ops->page_response(domain, dev, &resp);
> +}
> +
> +static int iommu_fault_handle_single(struct iommu_fault_context *fault)
> +{
> + /* TODO */
> + return -ENODEV;
> +}
> +
> +static void iommu_fault_handle_group(struct work_struct *work)
> +{
> + struct iommu_fault_group *group;
> + struct iommu_fault_context *fault, *next;
> + int status = IOMMU_PAGE_RESP_SUCCESS;
> +
> + group = container_of(work, struct iommu_fault_group, work);
> +
> + list_for_each_entry_safe(fault, next, &group->faults, head) {
> + struct iommu_fault_event *evt = &fault->evt;
> + /*
> + * Errors are sticky: don't handle subsequent faults in the
> + * group if there is an error.
> + */
> + if (status == IOMMU_PAGE_RESP_SUCCESS)
> + status = iommu_fault_handle_single(fault);
> +
> + if (!evt->last_req)
> + kfree(fault);
> + }
> +
> + iommu_fault_complete(group->domain, group->last_fault.dev,
> + &group->last_fault.evt, status);
> + kfree(group);
> +}
> +
> +static int iommu_queue_fault(struct iommu_domain *domain, struct device *dev,
> + struct iommu_fault_event *evt)
> +{
> + struct iommu_fault_group *group;
> + struct iommu_fault_context *fault, *next;
> +
> + if (!iommu_fault_queue)
> + return -ENOSYS;
> +
> + if (!evt->last_req) {
> + fault = kzalloc(sizeof(*fault), GFP_KERNEL);
> + if (!fault)
> + return -ENOMEM;
> +
> + fault->evt = *evt;
> + fault->dev = dev;
> +
> + /* Non-last request of a group. Postpone until the last one */
> + spin_lock(&iommu_partial_faults_lock);
> + list_add_tail(&fault->head, &iommu_partial_faults);
> + spin_unlock(&iommu_partial_faults_lock);
> +
> + return IOMMU_PAGE_RESP_HANDLED;
> + }
> +
> + group = kzalloc(sizeof(*group), GFP_KERNEL);
> + if (!group)
> + return -ENOMEM;
> +
> + group->last_fault.evt = *evt;
> + group->last_fault.dev = dev;
> + group->domain = domain;
> + INIT_LIST_HEAD(&group->faults);
> + list_add(&group->last_fault.head, &group->faults);
> + INIT_WORK(&group->work, iommu_fault_handle_group);
> +
> + /* See if we have pending faults for this group */
> + spin_lock(&iommu_partial_faults_lock);
> + list_for_each_entry_safe(fault, next, &iommu_partial_faults, head) {
> + if (fault->evt.page_req_group_id == evt->page_req_group_id &&
> + fault->dev == dev) {
> + list_del(&fault->head);
> + /* Insert *before* the last fault */
> + list_add(&fault->head, &group->faults);
> + }
> + }
> + spin_unlock(&iommu_partial_faults_lock);
> +
> + queue_work(iommu_fault_queue, &group->work);
> +
> + /* Postpone the fault completion */
> + return IOMMU_PAGE_RESP_HANDLED;
> +}
> +
> +/**
> + * iommu_report_device_fault() - Handle fault in device driver or mm
> + *
> + * If the device driver expressed interest in handling fault, report it through
> + * the callback. If the fault is recoverable, try to page in the address.
> + */
> +int iommu_report_device_fault(struct device *dev, struct iommu_fault_event *evt)
> +{
> + int ret = -ENOSYS;
> + struct iommu_domain *domain = iommu_get_domain_for_dev(dev);
> +
> + if (!domain)
> + return -ENODEV;
> +
> + /*
> + * if upper layers showed interest and installed a fault handler,
> + * invoke it.
> + */
> + if (iommu_has_device_fault_handler(dev)) {
> + struct iommu_fault_param *param = dev->iommu_param->fault_param;
> +
> + return param->handler(evt, param->data);
> + }
> +
> + /* If the handler is blocking, handle fault in the workqueue */
> + if (evt->type == IOMMU_FAULT_PAGE_REQ)
> + ret = iommu_queue_fault(domain, dev, evt);
> +
> + return iommu_fault_complete(domain, dev, evt, ret);
> +}
> +EXPORT_SYMBOL_GPL(iommu_report_device_fault);
> +
> +/**
> + * iommu_fault_queue_register() - register an IOMMU driver to the fault queue
> + * @flush_notifier: a notifier block that is called before the fault queue is
> + * flushed. The IOMMU driver should commit all faults that are pending in its
> + * low-level queues at the time of the call, into the fault queue. The notifier
> + * takes a device pointer as argument, hinting what endpoint is causing the
> + * flush. When the device is NULL, all faults should be committed.
> + */
> +int iommu_fault_queue_register(struct notifier_block *flush_notifier)
> +{
> + /*
> + * The WQ is unordered because the low-level handler enqueues faults by
> + * group. PRI requests within a group have to be ordered, but once
> + * that's dealt with, the high-level function can handle groups out of
> + * order.
> + */
> + down_write(&iommu_fault_queue_sem);
> + if (!iommu_fault_queue) {
> + iommu_fault_queue = alloc_workqueue("iommu_fault_queue",
> + WQ_UNBOUND, 0);
> + if (iommu_fault_queue)
> + refcount_set(&iommu_fault_queue_refs, 1);
> + } else {
> + refcount_inc(&iommu_fault_queue_refs);
> + }
> + up_write(&iommu_fault_queue_sem);
> +
> + if (!iommu_fault_queue)
> + return -ENOMEM;
> +
> + if (flush_notifier)
> + blocking_notifier_chain_register(&iommu_fault_queue_flush_notifiers,
> + flush_notifier);
> +
> + return 0;
> +}
> +EXPORT_SYMBOL_GPL(iommu_fault_queue_register);
> +
> +/**
> + * iommu_fault_queue_flush() - Ensure that all queued faults have been
> + * processed.
> + * @dev: the endpoint whose faults need to be flushed. If NULL, flush all
> + * pending faults.
> + *
> + * Users must call this function when releasing a PASID, to ensure that all
> + * pending faults affecting this PASID have been handled, and won't affect the
> + * address space of a subsequent process that reuses this PASID.
> + */
> +void iommu_fault_queue_flush(struct device *dev)
> +{
> + blocking_notifier_call_chain(&iommu_fault_queue_flush_notifiers, 0, dev);
> +
> + down_read(&iommu_fault_queue_sem);
> + /*
> + * Don't flush the partial faults list. All PRGs with the PASID are
> + * complete and have been submitted to the queue.
> + */
> + if (iommu_fault_queue)
> + flush_workqueue(iommu_fault_queue);
> + up_read(&iommu_fault_queue_sem);
> +}
> +EXPORT_SYMBOL_GPL(iommu_fault_queue_flush);
> +
> +/**
> + * iommu_fault_queue_unregister() - Unregister an IOMMU driver from the fault
> + * queue.
> + * @flush_notifier: same parameter as iommu_fault_queue_register
> + */
> +void iommu_fault_queue_unregister(struct notifier_block *flush_notifier)
> +{
> + down_write(&iommu_fault_queue_sem);
> + if (refcount_dec_and_test(&iommu_fault_queue_refs)) {
> + destroy_workqueue(iommu_fault_queue);
> + iommu_fault_queue = NULL;
> + }
> + up_write(&iommu_fault_queue_sem);
> +
> + if (flush_notifier)
> + blocking_notifier_chain_unregister(&iommu_fault_queue_flush_notifiers,
> + flush_notifier);
I would expect the ordering in queue_unregister to be the reverse of queue
register (to make it obvious there are no races).
That would put this last block at the start before potentially destroying
the work queue. If I'm missing something then perhaps a comment to
explain why the ordering is not the obvious one?
> +}
> +EXPORT_SYMBOL_GPL(iommu_fault_queue_unregister);
> diff --git a/drivers/iommu/iommu-sva.c b/drivers/iommu/iommu-sva.c
> index 4bc2a8c12465..d7b231cd7355 100644
> --- a/drivers/iommu/iommu-sva.c
> +++ b/drivers/iommu/iommu-sva.c
> @@ -102,9 +102,6 @@
> * the device table and PASID 0 would be available to the allocator.
> */
>
> -/* TODO: stub for the fault queue. Remove later. */
> -#define iommu_fault_queue_flush(...)
> -
> struct iommu_bond {
> struct io_mm *io_mm;
> struct device *dev;
> diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c
> index 1d60b32a6744..c475893ec7dc 100644
> --- a/drivers/iommu/iommu.c
> +++ b/drivers/iommu/iommu.c
> @@ -798,6 +798,17 @@ int iommu_group_unregister_notifier(struct iommu_group *group,
> }
> EXPORT_SYMBOL_GPL(iommu_group_unregister_notifier);
>
> +/**
> + * iommu_register_device_fault_handler() - Register a device fault handler
> + * @dev: the device
> + * @handler: the fault handler
> + * @data: private data passed as argument to the callback
> + *
> + * When an IOMMU fault event is received, call this handler with the fault event
> + * and data as argument.
> + *
> + * Return 0 if the fault handler was installed successfully, or an error.
> + */
> int iommu_register_device_fault_handler(struct device *dev,
> iommu_dev_fault_handler_t handler,
> void *data)
> @@ -825,6 +836,13 @@ int iommu_register_device_fault_handler(struct device *dev,
> }
> EXPORT_SYMBOL_GPL(iommu_register_device_fault_handler);
>
> +/**
> + * iommu_unregister_device_fault_handler() - Unregister the device fault handler
> + * @dev: the device
> + *
> + * Remove the device fault handler installed with
> + * iommu_register_device_fault_handler().
> + */
> int iommu_unregister_device_fault_handler(struct device *dev)
> {
> struct iommu_param *idata = dev->iommu_param;
> @@ -840,19 +858,6 @@ int iommu_unregister_device_fault_handler(struct device *dev)
> }
> EXPORT_SYMBOL_GPL(iommu_unregister_device_fault_handler);
>
> -
> -int iommu_report_device_fault(struct device *dev, struct iommu_fault_event *evt)
> -{
> - /* we only report device fault if there is a handler registered */
> - if (!dev->iommu_param || !dev->iommu_param->fault_param ||
> - !dev->iommu_param->fault_param->handler)
> - return -ENOSYS;
> -
> - return dev->iommu_param->fault_param->handler(evt,
> - dev->iommu_param->fault_param->data);
> -}
> -EXPORT_SYMBOL_GPL(iommu_report_device_fault);
> -
> /**
> * iommu_group_id - Return ID for a group
> * @group: the group to ID
> diff --git a/include/linux/iommu.h b/include/linux/iommu.h
> index 226ab4f3ae0e..65e56f28e0ce 100644
> --- a/include/linux/iommu.h
> +++ b/include/linux/iommu.h
> @@ -205,6 +205,7 @@ struct page_response_msg {
> u32 resp_code:4;
> #define IOMMU_PAGE_RESP_SUCCESS 0
> #define IOMMU_PAGE_RESP_INVALID 1
> +#define IOMMU_PAGE_RESP_HANDLED 2
> #define IOMMU_PAGE_RESP_FAILURE 0xF
>
> u32 pasid_present:1;
> @@ -534,7 +535,6 @@ extern int iommu_register_device_fault_handler(struct device *dev,
>
> extern int iommu_unregister_device_fault_handler(struct device *dev);
>
> -extern int iommu_report_device_fault(struct device *dev, struct iommu_fault_event *evt);
> extern int iommu_page_response(struct iommu_domain *domain, struct device *dev,
> struct page_response_msg *msg);
>
> @@ -836,11 +836,6 @@ static inline bool iommu_has_device_fault_handler(struct device *dev)
> return false;
> }
>
> -static inline int iommu_report_device_fault(struct device *dev, struct iommu_fault_event *evt)
> -{
> - return 0;
> -}
> -
> static inline int iommu_page_response(struct iommu_domain *domain, struct device *dev,
> struct page_response_msg *msg)
> {
> @@ -1005,4 +1000,31 @@ static inline struct mm_struct *iommu_sva_find(int pasid)
> }
> #endif /* CONFIG_IOMMU_SVA */
>
> +#ifdef CONFIG_IOMMU_FAULT
> +extern int iommu_fault_queue_register(struct notifier_block *flush_notifier);
> +extern void iommu_fault_queue_flush(struct device *dev);
> +extern void iommu_fault_queue_unregister(struct notifier_block *flush_notifier);
> +extern int iommu_report_device_fault(struct device *dev,
> + struct iommu_fault_event *evt);
> +#else /* CONFIG_IOMMU_FAULT */
> +static inline int iommu_fault_queue_register(struct notifier_block *flush_notifier)
> +{
> + return -ENODEV;
> +}
> +
> +static inline void iommu_fault_queue_flush(struct device *dev)
> +{
> +}
> +
> +static inline void iommu_fault_queue_unregister(struct notifier_block *flush_notifier)
> +{
> +}
> +
> +static inline int iommu_report_device_fault(struct device *dev,
> + struct iommu_fault_event *evt)
> +{
> + return 0;
> +}
> +#endif /* CONFIG_IOMMU_FAULT */
> +
> #endif /* __LINUX_IOMMU_H */
next prev parent reply other threads:[~2018-03-08 15:40 UTC|newest]
Thread overview: 105+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-02-12 18:33 [PATCH 00/37] Shared Virtual Addressing for the IOMMU Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 01/37] iommu: Introduce Shared Virtual Addressing API Jean-Philippe Brucker
2018-02-13 7:31 ` Tian, Kevin
2018-02-13 12:40 ` Jean-Philippe Brucker
2018-02-13 23:43 ` Tian, Kevin
2018-02-15 12:42 ` Jean-Philippe Brucker
2018-02-27 6:21 ` Tian, Kevin
2018-02-28 16:20 ` Jean-Philippe Brucker
2018-02-15 9:59 ` Joerg Roedel
2018-02-15 12:43 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 02/37] iommu/sva: Bind process address spaces to devices Jean-Philippe Brucker
2018-02-13 7:54 ` Tian, Kevin
2018-02-13 12:57 ` Jean-Philippe Brucker
2018-02-13 23:34 ` Tian, Kevin
2018-02-15 12:40 ` Jean-Philippe Brucker
2018-03-01 3:03 ` Liu, Yi L
2018-03-02 16:03 ` Jean-Philippe Brucker
2018-02-15 10:21 ` joro
2018-02-15 12:29 ` Christian König
2018-02-15 12:46 ` Jean-Philippe Brucker
2018-02-28 20:34 ` Sinan Kaya
2018-03-02 12:32 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 03/37] iommu/sva: Manage process address spaces Jean-Philippe Brucker
2018-03-01 6:52 ` Lu Baolu
2018-03-01 8:04 ` Christian König
2018-03-02 16:42 ` Jean-Philippe Brucker
2018-03-02 16:19 ` Jean-Philippe Brucker
2018-03-05 15:28 ` Sinan Kaya
2018-03-06 10:37 ` Jean-Philippe Brucker
2018-04-10 18:53 ` Sinan Kaya
2018-04-13 10:59 ` Jean-Philippe Brucker
2018-04-24 1:32 ` Sinan Kaya
2018-04-24 9:33 ` Jean-Philippe Brucker
2018-04-24 17:17 ` Sinan Kaya
2018-02-12 18:33 ` [PATCH 04/37] iommu/sva: Add a mm_exit callback for device drivers Jean-Philippe Brucker
2018-02-13 8:11 ` Tian, Kevin
2018-02-13 12:57 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 05/37] iommu/sva: Track mm changes with an MMU notifier Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 06/37] iommu/sva: Search mm by PASID Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 07/37] iommu: Add a page fault handler Jean-Philippe Brucker
2018-02-14 7:18 ` Jacob Pan
2018-02-15 13:49 ` Jean-Philippe Brucker
2018-03-05 21:44 ` Sinan Kaya
2018-03-06 10:24 ` Jean-Philippe Brucker
2018-03-05 21:53 ` Sinan Kaya
2018-03-06 10:46 ` Jean-Philippe Brucker
2018-03-06 12:52 ` okaya
2018-03-08 15:40 ` Jonathan Cameron [this message]
2018-03-14 13:08 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 08/37] iommu/fault: Handle mm faults Jean-Philippe Brucker
2018-02-14 18:46 ` Jacob Pan
2018-02-15 13:51 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 09/37] iommu/fault: Let handler return a fault response Jean-Philippe Brucker
2018-02-20 23:19 ` Jacob Pan
2018-02-21 10:28 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 10/37] iommu/fault: Allow blocking fault handlers Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 11/37] dt-bindings: document stall and PASID properties for IOMMU masters Jean-Philippe Brucker
2018-02-19 2:51 ` Rob Herring
2018-02-20 11:28 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 12/37] iommu/of: Add stall and pasid properties to iommu_fwspec Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 13/37] arm64: mm: Pin down ASIDs for sharing mm with devices Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 14/37] iommu/arm-smmu-v3: Link domains and devices Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 15/37] iommu/io-pgtable-arm: Factor out ARM LPAE register defines Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 16/37] iommu: Add generic PASID table library Jean-Philippe Brucker
2018-02-27 18:51 ` Jacob Pan
2018-02-28 16:22 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 17/37] iommu/arm-smmu-v3: Move context descriptor code Jean-Philippe Brucker
2018-03-09 11:44 ` Jonathan Cameron
2018-03-14 13:08 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 18/37] iommu/arm-smmu-v3: Add support for Substream IDs Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 19/37] iommu/arm-smmu-v3: Add second level of context descriptor table Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 20/37] iommu/arm-smmu-v3: Share process page tables Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 21/37] iommu/arm-smmu-v3: Seize private ASID Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 22/37] iommu/arm-smmu-v3: Add support for VHE Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 23/37] iommu/arm-smmu-v3: Enable broadcast TLB maintenance Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 24/37] iommu/arm-smmu-v3: Add SVA feature checking Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 25/37] iommu/arm-smmu-v3: Implement mm operations Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 26/37] iommu/arm-smmu-v3: Add support for Hardware Translation Table Update Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 27/37] iommu/arm-smmu-v3: Register fault workqueue Jean-Philippe Brucker
2018-03-08 17:44 ` Jonathan Cameron
2018-03-14 13:08 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 28/37] iommu/arm-smmu-v3: Maintain a SID->device structure Jean-Philippe Brucker
2018-03-08 17:34 ` Jonathan Cameron
2018-03-14 13:09 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 29/37] iommu/arm-smmu-v3: Add stall support for platform devices Jean-Philippe Brucker
2018-02-13 1:46 ` Xu Zaibo
2018-02-13 12:58 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 30/37] ACPI/IORT: Check ATS capability in root complex nodes Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 31/37] iommu/arm-smmu-v3: Add support for PCI ATS Jean-Philippe Brucker
2018-03-08 16:17 ` Jonathan Cameron
2018-03-14 13:09 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 32/37] iommu/arm-smmu-v3: Hook up ATC invalidation to mm ops Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 33/37] iommu/arm-smmu-v3: Disable tagged pointers Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 34/37] PCI: Make "PRG Response PASID Required" handling common Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 35/37] iommu/arm-smmu-v3: Add support for PRI Jean-Philippe Brucker
2018-03-05 12:29 ` Dongdong Liu
2018-03-05 13:09 ` Jean-Philippe Brucker
2018-03-08 16:24 ` Jonathan Cameron
2018-03-14 13:10 ` Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 36/37] iommu/arm-smmu-v3: Add support for PCI PASID Jean-Philippe Brucker
2018-02-12 18:33 ` [PATCH 37/37] vfio: Add support for Shared Virtual Addressing Jean-Philippe Brucker
2018-02-16 19:33 ` Alex Williamson
2018-02-20 11:26 ` Jean-Philippe Brucker
2018-02-28 1:26 ` Sinan Kaya
2018-02-28 16:25 ` Jean-Philippe Brucker
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20180308164035.000065c2@huawei.com \
--to=jonathan.cameron@huawei.com \
--cc=alex.williamson@redhat.com \
--cc=ashok.raj@intel.com \
--cc=bharatku@xilinx.com \
--cc=bhelgaas@google.com \
--cc=catalin.marinas@arm.com \
--cc=christian.koenig@amd.com \
--cc=devicetree@vger.kernel.org \
--cc=dwmw2@infradead.org \
--cc=hanjun.guo@linaro.org \
--cc=ilias.apalodimas@linaro.org \
--cc=iommu@lists.linux-foundation.org \
--cc=jacob.jun.pan@linux.intel.com \
--cc=jcrouse@codeaurora.org \
--cc=jean-philippe.brucker@arm.com \
--cc=joro@8bytes.org \
--cc=kvm@vger.kernel.org \
--cc=lenb@kernel.org \
--cc=linux-acpi@vger.kernel.org \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-pci@vger.kernel.org \
--cc=liubo95@huawei.com \
--cc=lorenzo.pieralisi@arm.com \
--cc=mark.rutland@arm.com \
--cc=nwatters@codeaurora.org \
--cc=okaya@codeaurora.org \
--cc=rfranz@cavium.com \
--cc=rjw@rjwysocki.net \
--cc=robdclark@gmail.com \
--cc=robh+dt@kernel.org \
--cc=robin.murphy@arm.com \
--cc=shunyong.yang@hxt-semitech.com \
--cc=sudeep.holla@arm.com \
--cc=thunder.leizhen@huawei.com \
--cc=tn@semihalf.com \
--cc=will.deacon@arm.com \
--cc=xieyisheng1@huawei.com \
--cc=xuzaibo@huawei.com \
--cc=yi.l.liu@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).