From mboxrd@z Thu Jan 1 00:00:00 1970 From: Baoquan He Subject: [PATCH v5 8/8] iommu/amd: Update domain into to dte entry during device driver init Date: Thu, 15 Sep 2016 23:03:26 +0800 Message-ID: <1473951806-25511-9-git-send-email-bhe@redhat.com> References: <1473951806-25511-1-git-send-email-bhe@redhat.com> Mime-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Return-path: In-Reply-To: <1473951806-25511-1-git-send-email-bhe-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org> List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Sender: iommu-bounces-cunTk1MwBs9QetFLy7KEm3xJsTq8ys+cHZ5vskTnxNA@public.gmane.org Errors-To: iommu-bounces-cunTk1MwBs9QetFLy7KEm3xJsTq8ys+cHZ5vskTnxNA@public.gmane.org To: joro-zLv9SwRftAIdnm+yROfE0A@public.gmane.org Cc: kexec-IAPFreCvJWM7uuMidbF8XUB+6BGkLq7r@public.gmane.org, linux-kernel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org, Vincent.Wan-5C7GfCeVMHo@public.gmane.org, iommu-cunTk1MwBs9QetFLy7KEm3xJsTq8ys+cHZ5vskTnxNA@public.gmane.org, dyoung-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org List-Id: iommu@lists.linux-foundation.org All devices are supposed to reset themselves at device driver initialization stage. At this time if in kdump kernel those on-flight DMA will be stopped because of device reset. It's best time to update the protection domain info, especially pte_root, to dte entry which the device relates to. Signed-off-by: Baoquan He --- drivers/iommu/amd_iommu.c | 21 +++++++++++++++++++++ 1 file changed, 21 insertions(+) diff --git a/drivers/iommu/amd_iommu.c b/drivers/iommu/amd_iommu.c index 6c37300..00b64ee 100644 --- a/drivers/iommu/amd_iommu.c +++ b/drivers/iommu/amd_iommu.c @@ -2310,6 +2310,10 @@ static dma_addr_t __map_single(struct device *dev, unsigned int pages; int prot = 0; int i; + struct iommu_dev_data *dev_data = get_dev_data(dev); + struct protection_domain *domain = get_domain(dev); + u16 alias = amd_iommu_alias_table[dev_data->devid]; + struct amd_iommu *iommu = amd_iommu_rlookup_table[dev_data->devid]; pages = iommu_num_pages(paddr, size, PAGE_SIZE); paddr &= PAGE_MASK; @@ -2319,6 +2323,13 @@ static dma_addr_t __map_single(struct device *dev, goto out; prot = dir2prot(direction); + if (translation_pre_enabled(iommu) && !dev_data->domain_updated) { + dev_data->domain_updated = true; + set_dte_entry(dev_data->devid, domain, dev_data->ats.enabled); + if (alias != dev_data->devid) + set_dte_entry(alias, domain, dev_data->ats.enabled); + device_flush_dte(dev_data); + } start = address; for (i = 0; i < pages; ++i) { @@ -2470,6 +2481,9 @@ static int map_sg(struct device *dev, struct scatterlist *sglist, struct scatterlist *s; unsigned long address; u64 dma_mask; + struct iommu_dev_data *dev_data = get_dev_data(dev); + u16 alias = amd_iommu_alias_table[dev_data->devid]; + struct amd_iommu *iommu = amd_iommu_rlookup_table[dev_data->devid]; domain = get_domain(dev); if (IS_ERR(domain)) @@ -2485,6 +2499,13 @@ static int map_sg(struct device *dev, struct scatterlist *sglist, goto out_err; prot = dir2prot(direction); + if (translation_pre_enabled(iommu) && !dev_data->domain_updated) { + dev_data->domain_updated = true; + set_dte_entry(dev_data->devid, domain, dev_data->ats.enabled); + if (alias != dev_data->devid) + set_dte_entry(alias, domain, dev_data->ats.enabled); + device_flush_dte(dev_data); + } /* Map all sg entries */ for_each_sg(sglist, s, nelems, i) { -- 2.5.5