public inbox for kvm@vger.kernel.org
 help / color / mirror / Atom feed
* [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
@ 2008-11-27 13:49 Han, Weidong
  2008-11-28 13:27 ` Joerg Roedel
  0 siblings, 1 reply; 7+ messages in thread
From: Han, Weidong @ 2008-11-27 13:49 UTC (permalink / raw)
  To: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity'
  Cc: Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
	'iommu@lists.linux-foundation.org'

[-- Attachment #1: Type: text/plain, Size: 47927 bytes --]

In order to support multiple device assignment for KVM, this patch does following main changes:
   - extend dmar_domain to own multiple devices from different iommus, use a bitmap of iommus to replace iommu pointer in dmar_domain.
   - implement independent low level functions for kvm, then won't impact native VT-d.
   - "SAGAW" capability may be different across iommus, that's to say the VT-d page table levels may be different among iommus. This patch uses a defaut agaw, and skip top levels of page tables for iommus which have smaller agaw than default.
   - rename the APIs for kvm VT-d, make it more readable.


Signed-off-by: Weidong Han <weidong.han@intel.com>
---
 drivers/pci/dmar.c            |   15 +
 drivers/pci/intel-iommu.c     |  698 ++++++++++++++++++++++++++++++++++------
 include/linux/dma_remapping.h |   21 +-
 include/linux/intel-iommu.h   |   21 +-
 4 files changed, 637 insertions(+), 118 deletions(-)

diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
index 691b3ad..d6bdced 100644
--- a/drivers/pci/dmar.c
+++ b/drivers/pci/dmar.c
@@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
        dmar_tbl = NULL;
 }

+extern int width_to_agaw(int width);

 int alloc_iommu(struct dmar_drhd_unit *drhd)
 {
@@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
        int map_size;
        u32 ver;
        static int iommu_allocated = 0;
+       unsigned long sagaw;
+       int agaw;

        iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);
        if (!iommu)
@@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
        iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
        iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);

+       /* set agaw, "SAGAW" may be different across iommus */
+       sagaw = cap_sagaw(iommu->cap);
+       for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
+            agaw >= 0; agaw--)
+               if (test_bit(agaw, &sagaw))
+                       break;
+       if (agaw < 0) {
+               printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n", sagaw);
+               goto error;
+       }
+       iommu->agaw = agaw;
+
        /* the registers might be more than one page */
        map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
                cap_max_fault_reg_offset(iommu->cap));
diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
index 5c8baa4..55b96c4 100644
--- a/drivers/pci/intel-iommu.c
+++ b/drivers/pci/intel-iommu.c
@@ -50,8 +50,6 @@
 #define IOAPIC_RANGE_END       (0xfeefffff)
 #define IOVA_START_ADDR                (0x1000)

-#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
-
 #define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)


@@ -64,6 +62,7 @@ struct deferred_flush_tables {
        int next;
        struct iova *iova[HIGH_WATER_MARK];
        struct dmar_domain *domain[HIGH_WATER_MARK];
+       struct intel_iommu *iommu;
 };

 static struct deferred_flush_tables *deferred_flush;
@@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
        kmem_cache_free(iommu_iova_cache, iova);
 }

+/* in native case, each domain is related to only one iommu */
+static struct intel_iommu *domain_get_only_iommu(struct dmar_domain *domain)
+{
+       struct dmar_drhd_unit *drhd;
+
+       for_each_drhd_unit(drhd) {
+               if (drhd->ignored)
+                       continue;
+               if (test_bit(drhd->iommu->seq_id, &domain->iommu_bmp))
+                       return drhd->iommu;
+       }
+
+       return NULL;
+}
+
+static void domain_flush_cache(struct dmar_domain *domain,
+                              void *addr, int size)
+{
+       struct intel_iommu *iommu;
+
+       if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
+               struct dmar_drhd_unit *drhd;
+
+               for_each_drhd_unit(drhd) {
+                       if (drhd->ignored)
+                               continue;
+                       iommu = drhd->iommu;
+
+                       if (!test_bit(iommu->seq_id, &domain->iommu_bmp))
+                               continue;
+
+                       if (!ecap_coherent(iommu->ecap))
+                               clflush_cache_range(addr, size);
+               }
+       }
+       else {
+               iommu = domain_get_only_iommu(domain);
+               if (iommu && !ecap_coherent(iommu->ecap))
+                       clflush_cache_range(addr, size);
+       }
+}
+
+static struct intel_iommu *device_find_matched_iommu(u8 bus, u8 devfn)
+{
+       struct dmar_drhd_unit *drhd = NULL;
+       int i;
+
+       for_each_drhd_unit(drhd) {
+               if (drhd->ignored)
+                       continue;
+
+               for (i = 0; i < drhd->devices_cnt; i++)
+                       if (drhd->devices[i]->bus->number == bus &&
+                           drhd->devices[i]->devfn == devfn)
+                               return drhd->iommu;
+
+               if (drhd->include_all)
+                       return drhd->iommu;
+       }
+
+       return NULL;
+}
+
 /* Gets context entry for a given bus and devfn */
 static struct context_entry * device_to_context_entry(struct intel_iommu *iommu,
                u8 bus, u8 devfn)
@@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)

 }

-static inline int width_to_agaw(int width)
+int width_to_agaw(int width)
 {
        return (width - 30) / LEVEL_STRIDE;
 }
@@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
                                        flags);
                                return NULL;
                        }
-                       __iommu_flush_cache(domain->iommu, tmp_page,
-                                       PAGE_SIZE);
+                       domain_flush_cache(domain, tmp_page, PAGE_SIZE);
                        dma_set_pte_addr(*pte, virt_to_phys(tmp_page));
                        /*
                         * high level table always sets r/w, last level page
@@ -356,7 +417,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
                         */
                        dma_set_pte_readable(*pte);
                        dma_set_pte_writable(*pte);
-                       __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+                       domain_flush_cache(domain, pte, sizeof(*pte));
                }
                parent = phys_to_virt(dma_pte_addr(*pte));
                level--;
@@ -399,7 +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain, u64 addr)

        if (pte) {
                dma_clear_pte(*pte);
-               __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+               domain_flush_cache(domain, pte, sizeof(*pte));
        }
 }

@@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct dmar_domain *domain,
                                free_pgtable_page(
                                        phys_to_virt(dma_pte_addr(*pte)));
                                dma_clear_pte(*pte);
-                               __iommu_flush_cache(domain->iommu,
-                                               pte, sizeof(*pte));
+                               domain_flush_cache(domain, pte, sizeof(*pte));
                        }
                        tmp += level_size(level);
                }
@@ -948,8 +1008,8 @@ static int iommu_init_domains(struct intel_iommu *iommu)
        return 0;
 }

-
 static void domain_exit(struct dmar_domain *domain);
+static void vm_domain_exit(struct dmar_domain *domain);

 void free_dmar_iommu(struct intel_iommu *iommu)
 {
@@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
        for (; i < cap_ndoms(iommu->cap); ) {
                domain = iommu->domains[i];
                clear_bit(i, iommu->domain_ids);
-               domain_exit(domain);
+
+               if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
+                       if (--domain->iommu_count == 0)
+                               vm_domain_exit(domain);
+               }
+               else
+                       domain_exit(domain);
+
                i = find_next_bit(iommu->domain_ids,
                        cap_ndoms(iommu->cap), i+1);
        }
@@ -1006,8 +1073,11 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)

        set_bit(num, iommu->domain_ids);
        domain->id = num;
-       domain->iommu = iommu;
+       memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
+       set_bit(iommu->seq_id, &domain->iommu_bmp);
        iommu->domains[num] = domain;
+       domain->iommu_count = 1;
+       domain->flags = 0;
        spin_unlock_irqrestore(&iommu->lock, flags);

        return domain;
@@ -1016,10 +1086,12 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
 static void iommu_free_domain(struct dmar_domain *domain)
 {
        unsigned long flags;
+       struct intel_iommu *iommu;

-       spin_lock_irqsave(&domain->iommu->lock, flags);
-       clear_bit(domain->id, domain->iommu->domain_ids);
-       spin_unlock_irqrestore(&domain->iommu->lock, flags);
+       iommu = domain_get_only_iommu(domain);
+       spin_lock_irqsave(&iommu->lock, flags);
+       clear_bit(domain->id, iommu->domain_ids);
+       spin_unlock_irqrestore(&iommu->lock, flags);
 }

 static struct iova_domain reserved_iova_list;
@@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
        domain_reserve_special_ranges(domain);

        /* calculate AGAW */
-       iommu = domain->iommu;
+       iommu = domain_get_only_iommu(domain);
        if (guest_width > cap_mgaw(iommu->cap))
                guest_width = cap_mgaw(iommu->cap);
        domain->gaw = guest_width;
@@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
        sagaw = cap_sagaw(iommu->cap);
        if (!test_bit(agaw, &sagaw)) {
                /* hardware doesn't support it, choose a bigger one */
-               pr_debug("IOMMU: hardware doesn't support agaw %d\n", agaw);
+               pr_debug("IOMMU: hardware doesn't support agaw %d\n",
+                 agaw);
                agaw = find_next_bit(&sagaw, 5, agaw);
                if (agaw >= 5)
                        return -ENODEV;
        }
        domain->agaw = agaw;
+
        INIT_LIST_HEAD(&domain->devices);

        /* always allocate the top pgd */
        domain->pgd = (struct dma_pte *)alloc_pgtable_page();
        if (!domain->pgd)
                return -ENOMEM;
-       __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
+       domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
        return 0;
 }

@@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain *domain)
 }

 static int domain_context_mapping_one(struct dmar_domain *domain,
-               u8 bus, u8 devfn)
+               struct intel_iommu *iommu, u8 bus, u8 devfn)
 {
        struct context_entry *context;
-       struct intel_iommu *iommu = domain->iommu;
        unsigned long flags;

        pr_debug("Set context mapping for %02x:%02x.%d\n",
@@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
 {
        int ret;
        struct pci_dev *tmp, *parent;
+       struct intel_iommu *iommu;
+
+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return  -ENODEV;

-       ret = domain_context_mapping_one(domain, pdev->bus->number,
-               pdev->devfn);
+       ret = domain_context_mapping_one(domain, iommu,
+               pdev->bus->number, pdev->devfn);
        if (ret)
                return ret;

@@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
        /* Secondary interface's bus number and devfn 0 */
        parent = pdev->bus->self;
        while (parent != tmp) {
-               ret = domain_context_mapping_one(domain, parent->bus->number,
-                       parent->devfn);
+               ret = domain_context_mapping_one(domain, iommu,
+                       parent->bus->number, parent->devfn);
                if (ret)
                        return ret;
                parent = parent->bus->self;
        }
        if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
-               return domain_context_mapping_one(domain,
+               return domain_context_mapping_one(domain, iommu,
                        tmp->subordinate->number, 0);
        else /* this is a legacy PCI bridge */
-               return domain_context_mapping_one(domain,
+               return domain_context_mapping_one(domain, iommu,
                        tmp->bus->number, tmp->devfn);
 }

-static int domain_context_mapped(struct dmar_domain *domain,
-       struct pci_dev *pdev)
+static int domain_context_mapped(struct pci_dev *pdev)
 {
        int ret;
        struct pci_dev *tmp, *parent;
+       struct intel_iommu *iommu;

-       ret = device_context_mapped(domain->iommu,
+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return 0;
+
+       ret = device_context_mapped(iommu,
                pdev->bus->number, pdev->devfn);
        if (!ret)
                return ret;
@@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct dmar_domain *domain,
        /* Secondary interface's bus number and devfn 0 */
        parent = pdev->bus->self;
        while (parent != tmp) {
-               ret = device_context_mapped(domain->iommu, parent->bus->number,
+               ret = device_context_mapped(iommu, parent->bus->number,
                        parent->devfn);
                if (!ret)
                        return ret;
                parent = parent->bus->self;
        }
        if (tmp->is_pcie)
-               return device_context_mapped(domain->iommu,
+               return device_context_mapped(iommu,
                        tmp->subordinate->number, 0);
        else
-               return device_context_mapped(domain->iommu,
+               return device_context_mapped(iommu,
                        tmp->bus->number, tmp->devfn);
 }

@@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
                BUG_ON(dma_pte_addr(*pte));
                dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
                dma_set_pte_prot(*pte, prot);
-               __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+               domain_flush_cache(domain, pte, sizeof(*pte));
                start_pfn++;
                index++;
        }
        return 0;
 }

-static void detach_domain_for_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
+static void detach_domain_for_dev(struct dmar_domain *domain,
+                                 u8 bus, u8 devfn)
 {
-       clear_context_table(domain->iommu, bus, devfn);
-       domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
-                                          DMA_CCMD_GLOBAL_INVL, 0);
-       domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
-                                        DMA_TLB_GLOBAL_FLUSH, 0);
+       struct intel_iommu *iommu;
+
+       iommu = device_find_matched_iommu(bus, devfn);
+       if (!iommu)
+               return;
+
+       clear_context_table(iommu, bus, devfn);
+       iommu->flush.flush_context(iommu, 0, 0, 0,
+                                  DMA_CCMD_GLOBAL_INVL, 0);
+       iommu->flush.flush_iotlb(iommu, 0, 0, 0,
+                                DMA_TLB_GLOBAL_FLUSH, 0);
 }

 static void domain_remove_dev_info(struct dmar_domain *domain)
@@ -1336,7 +1425,6 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
 {
        struct dmar_domain *domain, *found = NULL;
        struct intel_iommu *iommu;
-       struct dmar_drhd_unit *drhd;
        struct device_domain_info *info, *tmp;
        struct pci_dev *dev_tmp;
        unsigned long flags;
@@ -1371,13 +1459,9 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
        }

        /* Allocate new domain for the device */
-       drhd = dmar_find_matched_drhd_unit(pdev);
-       if (!drhd) {
-               printk(KERN_ERR "IOMMU: can't find DMAR for device %s\n",
-                       pci_name(pdev));
-               return NULL;
-       }
-       iommu = drhd->iommu;
+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return  NULL;

        domain = iommu_alloc_domain(iommu);
        if (!domain)
@@ -1400,7 +1484,7 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
                info->dev = NULL;
                info->domain = domain;
                /* This domain is shared by devices under p2p bridge */
-               domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
+               domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;

                /* pcie-to-pci bridge already has a domain, uses it */
                found = NULL;
@@ -1805,7 +1889,7 @@ get_valid_domain_for_dev(struct pci_dev *pdev)
        }

        /* make sure context mapping is ok */
-       if (unlikely(!domain_context_mapped(domain, pdev))) {
+       if (unlikely(!domain_context_mapped(pdev))) {
                ret = domain_context_mapping(domain, pdev);
                if (ret) {
                        printk(KERN_ERR
@@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
 {
        struct pci_dev *pdev = to_pci_dev(hwdev);
        struct dmar_domain *domain;
+       struct intel_iommu *iommu;
        phys_addr_t start_paddr;
        struct iova *iova;
        int prot = 0;
@@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
        if (!domain)
                return 0;

+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return 0;
+
        size = aligned_size((u64)paddr, size);

        iova = __intel_alloc_iova(hwdev, domain, size, pdev->dma_mask);
@@ -1849,7 +1938,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
         * mappings..
         */
        if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
-                       !cap_zlr(domain->iommu->cap))
+                       !cap_zlr(iommu->cap))
                prot |= DMA_PTE_READ;
        if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
                prot |= DMA_PTE_WRITE;
@@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
                goto error;

        /* it's a non-present to present mapping */
-       ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
+       ret = iommu_flush_iotlb_psi(iommu, domain->id,
                        start_paddr, size >> VTD_PAGE_SHIFT, 1);
        if (ret)
-               iommu_flush_write_buffer(domain->iommu);
+               iommu_flush_write_buffer(iommu);

        return start_paddr + ((u64)paddr & (~PAGE_MASK));

@@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
        /* just flush them all */
        for (i = 0; i < g_num_of_iommus; i++) {
                if (deferred_flush[i].next) {
-                       struct intel_iommu *iommu =
-                               deferred_flush[i].domain[0]->iommu;
+                       struct intel_iommu *iommu = deferred_flush[i].iommu;

                        iommu->flush.flush_iotlb(iommu, 0, 0, 0,
                                                 DMA_TLB_GLOBAL_FLUSH, 0);
@@ -1921,7 +2009,8 @@ static void flush_unmaps_timeout(unsigned long data)
        spin_unlock_irqrestore(&async_umap_flush_lock, flags);
 }

-static void add_unmap(struct dmar_domain *dom, struct iova *iova)
+static void add_unmap(struct dmar_domain *dom,
+                     struct intel_iommu *iommu, struct iova *iova)
 {
        unsigned long flags;
        int next, iommu_id;
@@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova)
        if (list_size == HIGH_WATER_MARK)
                flush_unmaps();

-       iommu_id = dom->iommu->seq_id;
+       iommu_id = iommu->seq_id;

        next = deferred_flush[iommu_id].next;
        deferred_flush[iommu_id].domain[next] = dom;
        deferred_flush[iommu_id].iova[next] = iova;
+       deferred_flush[iommu_id].iommu = iommu;
        deferred_flush[iommu_id].next++;

        if (!timer_on) {
@@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
        struct dmar_domain *domain;
        unsigned long start_addr;
        struct iova *iova;
+       struct intel_iommu *iommu;

        if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
                return;
        domain = find_domain(pdev);
        BUG_ON(!domain);

+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return;
+
        iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
        if (!iova)
                return;
@@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
        /* free page tables */
        dma_pte_free_pagetable(domain, start_addr, start_addr + size);
        if (intel_iommu_strict) {
-               if (iommu_flush_iotlb_psi(domain->iommu,
+               if (iommu_flush_iotlb_psi(iommu,
                        domain->id, start_addr, size >> VTD_PAGE_SHIFT, 0))
-                       iommu_flush_write_buffer(domain->iommu);
+                       iommu_flush_write_buffer(iommu);
                /* free iova */
                __free_iova(&domain->iovad, iova);
        } else {
-               add_unmap(domain, iova);
+               add_unmap(domain, iommu, iova);
                /*
                 * queue up the release of the unmap to save the 1/6th of the
                 * cpu used up by the iotlb flush operation...
@@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
        size_t size = 0;
        void *addr;
        struct scatterlist *sg;
+       struct intel_iommu *iommu;

        if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
                return;

        domain = find_domain(pdev);

+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return;
+
        iova = find_iova(&domain->iovad, IOVA_PFN(sglist[0].dma_address));
        if (!iova)
                return;
@@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
        /* free page tables */
        dma_pte_free_pagetable(domain, start_addr, start_addr + size);

-       if (iommu_flush_iotlb_psi(domain->iommu, domain->id, start_addr,
+       if (iommu_flush_iotlb_psi(iommu, domain->id, start_addr,
                        size >> VTD_PAGE_SHIFT, 0))
-               iommu_flush_write_buffer(domain->iommu);
+               iommu_flush_write_buffer(iommu);

        /* free iova */
        __free_iova(&domain->iovad, iova);
@@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
        int ret;
        struct scatterlist *sg;
        unsigned long start_addr;
+       struct intel_iommu *iommu;

        BUG_ON(dir == DMA_NONE);
        if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
@@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
        if (!domain)
                return 0;

+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return 0;
+
        for_each_sg(sglist, sg, nelems, i) {
                addr = SG_ENT_VIRT_ADDRESS(sg);
                addr = (void *)virt_to_phys(addr);
@@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
         * mappings..
         */
        if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
-                       !cap_zlr(domain->iommu->cap))
+                       !cap_zlr(iommu->cap))
                prot |= DMA_PTE_READ;
        if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
                prot |= DMA_PTE_WRITE;
@@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
        }

        /* it's a non-present to present mapping */
-       if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
+       if (iommu_flush_iotlb_psi(iommu, domain->id,
                        start_addr, offset >> VTD_PAGE_SHIFT, 1))
-               iommu_flush_write_buffer(domain->iommu);
+               iommu_flush_write_buffer(iommu);
        return nelems;
 }

@@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
        return 0;
 }

-void intel_iommu_domain_exit(struct dmar_domain *domain)
+/* domain id for virtual machine, it won't be set in context */
+static unsigned long vm_domid;
+
+static int vm_domain_min_agaw(struct dmar_domain *domain)
+{
+       struct dmar_drhd_unit *drhd;
+       struct intel_iommu *iommu;
+       int min_agaw = domain->agaw;
+
+       for_each_drhd_unit(drhd) {
+               if (drhd->ignored)
+                       continue;
+               iommu = drhd->iommu;
+
+               if (test_bit(iommu->seq_id, &domain->iommu_bmp))
+                       if (min_agaw > iommu->agaw)
+                               min_agaw = iommu->agaw;
+       }
+
+       return min_agaw;
+}
+
+static int vm_domain_add_dev_info(struct dmar_domain *domain,
+                                 struct pci_dev *pdev)
+{
+       struct device_domain_info *info;
+       unsigned long flags;
+
+       info = alloc_devinfo_mem();
+       if (!info)
+               return -ENOMEM;
+
+       info->bus = pdev->bus->number;
+       info->devfn = pdev->devfn;
+       info->dev = pdev;
+       info->domain = domain;
+
+       spin_lock_irqsave(&device_domain_lock, flags);
+       list_add(&info->link, &domain->devices);
+       list_add(&info->global, &device_domain_list);
+       pdev->dev.archdata.iommu = info;
+       spin_unlock_irqrestore(&device_domain_lock, flags);
+
+       return 0;
+}
+
+static void vm_domain_remove_one_dev_info(struct dmar_domain *domain,
+                                         struct pci_dev *pdev)
+{
+       struct device_domain_info *info;
+       struct intel_iommu *iommu;
+       unsigned long flags;
+       int found = 0;
+
+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+
+       spin_lock_irqsave(&device_domain_lock, flags);
+       while (!list_empty(&domain->devices)) {
+               info = list_entry(domain->devices.next,
+                       struct device_domain_info, link);
+               if (info->bus == pdev->bus->number &&
+                   info->devfn == pdev->devfn) {
+                       list_del(&info->link);
+                       list_del(&info->global);
+                       if (info->dev)
+                               info->dev->dev.archdata.iommu = NULL;
+                       spin_unlock_irqrestore(&device_domain_lock, flags);
+
+                       detach_domain_for_dev(info->domain,
+                                             info->bus, info->devfn);
+                       free_devinfo_mem(info);
+
+                       spin_lock_irqsave(&device_domain_lock, flags);
+
+                       if (found)
+                               break;
+                       else
+                               continue;
+               }
+
+               /* if there is no other devices under the same iommu
+                * owned by this domain, clear this iommu in iommu_bmp
+                */
+               if (device_find_matched_iommu(info->bus, info->devfn) == iommu)
+                       found = 1;
+       }
+
+       if (found == 0) {
+               spin_lock_irqsave(&iommu->lock, flags);
+               clear_bit(iommu->seq_id, &domain->iommu_bmp);
+               domain->iommu_count--;
+               spin_unlock_irqrestore(&iommu->lock, flags);
+       }
+
+       spin_unlock_irqrestore(&device_domain_lock, flags);
+}
+
+static void vm_domain_remove_all_dev_info(struct dmar_domain *domain)
+{
+       struct device_domain_info *info;
+       struct intel_iommu *iommu;
+       unsigned long flags;
+
+       spin_lock_irqsave(&device_domain_lock, flags);
+       while (!list_empty(&domain->devices)) {
+               info = list_entry(domain->devices.next,
+                       struct device_domain_info, link);
+               list_del(&info->link);
+               list_del(&info->global);
+               if (info->dev)
+                       info->dev->dev.archdata.iommu = NULL;
+
+               spin_unlock_irqrestore(&device_domain_lock, flags);
+               detach_domain_for_dev(info->domain,
+                                     info->bus, info->devfn);
+
+               /* clear this iommu in iommu_bmp */
+               iommu = device_find_matched_iommu(info->bus, info->devfn);
+               spin_lock_irqsave(&iommu->lock, flags);
+               if (test_and_clear_bit(iommu->seq_id,
+                                      &domain->iommu_bmp))
+                       domain->iommu_count--;
+               spin_unlock_irqrestore(&iommu->lock, flags);
+
+               free_devinfo_mem(info);
+               spin_lock_irqsave(&device_domain_lock, flags);
+       }
+       spin_unlock_irqrestore(&device_domain_lock, flags);
+}
+
+static int vm_domain_context_mapping_one(struct dmar_domain *domain,
+               struct intel_iommu *iommu, u8 bus, u8 devfn)
+{
+       struct context_entry *context;
+       unsigned long flags;
+       struct dma_pte *pgd;
+       unsigned long num;
+       unsigned long ndomains;
+       int id;
+       int agaw;
+       int found = 0;
+
+       pr_debug("Set context mapping for %02x:%02x.%d\n",
+               bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
+       BUG_ON(!domain->pgd);
+       context = device_to_context_entry(iommu, bus, devfn);
+       if (!context)
+               return -ENOMEM;
+       spin_lock_irqsave(&iommu->lock, flags);
+       if (context_present(*context)) {
+               spin_unlock_irqrestore(&iommu->lock, flags);
+               return 0;
+       }
+
+       id = domain->id;
+
+       /* find an available domain id for this device in iommu */
+       ndomains = cap_ndoms(iommu->cap);
+       num = find_first_bit(iommu->domain_ids, ndomains);
+       for (; num < ndomains; ) {
+               if (iommu->domains[num] == domain) {
+                       id = num;
+                       found = 1;
+                       break;
+               }
+               num = find_next_bit(iommu->domain_ids,
+                 cap_ndoms(iommu->cap), num+1);
+       }
+
+       if (found == 0) {
+               num = find_first_zero_bit(iommu->domain_ids, ndomains);
+               if (num >= ndomains) {
+                       spin_unlock_irqrestore(&iommu->lock, flags);
+                       printk(KERN_ERR "IOMMU: no free domain ids\n");
+                       return -EFAULT;
+               }
+
+               set_bit(num, iommu->domain_ids);
+               iommu->domains[num] = domain;
+               id = num;
+       }
+
+       pgd = domain->pgd;
+
+       /* Skip top levels of page tables for
+        * iommu which has less agaw than default.
+        */
+       for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
+               pgd = phys_to_virt(dma_pte_addr(*pgd));
+               if (!dma_pte_present(*pgd)) {
+                       spin_unlock_irqrestore(&iommu->lock, flags);
+                       return -ENOMEM;
+               }
+       }
+
+       context_set_domain_id(*context, id);
+       context_set_address_width(*context, iommu->agaw);
+       context_set_address_root(*context, virt_to_phys(pgd));
+       context_set_translation_type(*context, CONTEXT_TT_MULTI_LEVEL);
+       context_set_fault_enable(*context);
+       context_set_present(*context);
+       __iommu_flush_cache(iommu, context, sizeof(*context));
+
+       /* it's a non-present to present mapping */
+       if (iommu->flush.flush_context(iommu, id,
+               (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
+               DMA_CCMD_DEVICE_INVL, 1))
+               iommu_flush_write_buffer(iommu);
+       else
+               iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_DSI_FLUSH, 0);
+
+       if (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp))
+               domain->iommu_count++;
+       spin_unlock_irqrestore(&iommu->lock, flags);
+       return 0;
+}
+
+static int
+vm_domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
+{
+       int ret;
+       struct pci_dev *tmp, *parent;
+       struct intel_iommu *iommu;
+
+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return  -ENODEV;
+
+       ret = vm_domain_context_mapping_one(domain, iommu,
+               pdev->bus->number, pdev->devfn);
+       if (ret)
+               return ret;
+
+       /* dependent device mapping */
+       tmp = pci_find_upstream_pcie_bridge(pdev);
+       if (!tmp)
+               return 0;
+       /* Secondary interface's bus number and devfn 0 */
+       parent = pdev->bus->self;
+       while (parent != tmp) {
+               ret = vm_domain_context_mapping_one(domain, iommu,
+                       parent->bus->number, parent->devfn);
+               if (ret)
+                       return ret;
+               parent = parent->bus->self;
+       }
+       if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
+               return vm_domain_context_mapping_one(domain, iommu,
+                       tmp->subordinate->number, 0);
+       else /* this is a legacy PCI bridge */
+               return vm_domain_context_mapping_one(domain, iommu,
+                       tmp->bus->number, tmp->devfn);
+}
+
+
+static int vm_domain_init(struct dmar_domain *domain, int guest_width)
+{
+       int adjust_width;
+
+       init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
+       spin_lock_init(&domain->mapping_lock);
+
+       domain_reserve_special_ranges(domain);
+
+       /* calculate AGAW */
+       domain->gaw = guest_width;
+       adjust_width = guestwidth_to_adjustwidth(guest_width);
+       domain->agaw = width_to_agaw(adjust_width);
+
+       INIT_LIST_HEAD(&domain->devices);
+
+       /* always allocate the top pgd */
+       domain->pgd = (struct dma_pte *)alloc_pgtable_page();
+       if (!domain->pgd)
+               return -ENOMEM;
+       domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
+       return 0;
+}
+
+static void iommu_free_vm_domain(struct dmar_domain *domain)
+{
+       unsigned long flags;
+       struct dmar_drhd_unit *drhd;
+       struct intel_iommu *iommu;
+       unsigned long i;
+       unsigned long ndomains;
+
+       for_each_drhd_unit(drhd) {
+               if (drhd->ignored)
+                       continue;
+               iommu = drhd->iommu;
+
+               ndomains = cap_ndoms(iommu->cap);
+               i = find_first_bit(iommu->domain_ids, ndomains);
+               for (; i < ndomains; ) {
+                       if (iommu->domains[i] == domain) {
+                               spin_lock_irqsave(&iommu->lock, flags);
+                               clear_bit(i, iommu->domain_ids);
+                               iommu->domains[i] = NULL;
+                               spin_unlock_irqrestore(&iommu->lock, flags);
+                               break;
+                       }
+                       i = find_next_bit(iommu->domain_ids, ndomains, i+1);
+               }
+       }
+}
+
+static void vm_domain_exit(struct dmar_domain *domain)
 {
        u64 end;

@@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
        if (!domain)
                return;

+       vm_domain_remove_all_dev_info(domain);
+       /* destroy iovas */
+       put_iova_domain(&domain->iovad);
        end = DOMAIN_MAX_ADDR(domain->gaw);
-       end = end & (~VTD_PAGE_MASK);
+       end &= VTD_PAGE_MASK;

        /* clear ptes */
        dma_pte_clear_range(domain, 0, end);
@@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
        /* free page tables */
        dma_pte_free_pagetable(domain, 0, end);

-       iommu_free_domain(domain);
+       iommu_free_vm_domain(domain);
        free_domain_mem(domain);
 }
-EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);

-struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
+static struct dmar_domain *iommu_alloc_vm_domain(void)
 {
-       struct dmar_drhd_unit *drhd;
        struct dmar_domain *domain;
-       struct intel_iommu *iommu;

-       drhd = dmar_find_matched_drhd_unit(pdev);
-       if (!drhd) {
-               printk(KERN_ERR "intel_iommu_domain_alloc: drhd == NULL\n");
+       domain = alloc_domain_mem();
+       if (!domain)
                return NULL;
-       }

-       iommu = drhd->iommu;
-       if (!iommu) {
-               printk(KERN_ERR
-                       "intel_iommu_domain_alloc: iommu == NULL\n");
-               return NULL;
-       }
-       domain = iommu_alloc_domain(iommu);
+       domain->id = vm_domid++;
+       domain->iommu_count = 0;
+       domain->max_addr = 0;
+       memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
+       domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE;
+
+       return domain;
+}
+
+struct dmar_domain *intel_iommu_alloc_domain(void)
+{
+       struct dmar_domain *domain;
+
+       domain = iommu_alloc_vm_domain();
        if (!domain) {
                printk(KERN_ERR
                        "intel_iommu_domain_alloc: domain == NULL\n");
                return NULL;
        }
-       if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
+       if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
                printk(KERN_ERR
                        "intel_iommu_domain_alloc: domain_init() failed\n");
-               intel_iommu_domain_exit(domain);
+               vm_domain_exit(domain);
                return NULL;
        }
+
        return domain;
 }
-EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
+EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
+
+void intel_iommu_free_domain(struct dmar_domain *domain)
+{
+       vm_domain_exit(domain);
+}
+EXPORT_SYMBOL_GPL(intel_iommu_free_domain);

-int intel_iommu_context_mapping(
-       struct dmar_domain *domain, struct pci_dev *pdev)
+int intel_iommu_assign_device(struct dmar_domain *domain,
+                             struct pci_dev *pdev)
 {
-       int rc;
-       rc = domain_context_mapping(domain, pdev);
-       return rc;
+       struct intel_iommu *iommu;
+       int addr_width;
+       u64 end;
+       int ret;
+
+       /* normally pdev is not mapped */
+       if (unlikely(domain_context_mapped(pdev))) {
+               struct dmar_domain *old_domain;
+
+               old_domain = find_domain(pdev);
+               if (old_domain) {
+                       if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE)
+                               vm_domain_remove_one_dev_info(old_domain, pdev);
+                       else
+                               domain_remove_dev_info(old_domain);
+               }
+       }
+
+       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+       if (!iommu)
+               return  -ENODEV;
+
+       /* check if this iommu agaw is sufficient for max mapped address */
+       addr_width = agaw_to_width(iommu->agaw);
+       end = DOMAIN_MAX_ADDR(addr_width);
+       end = end & VTD_PAGE_MASK;
+       if (end < domain->max_addr) {
+               printk(KERN_ERR "%s: iommu agaw (%d) is not "
+                      "sufficient for the mapped address (%llx)\n",
+                      __func__, iommu->agaw, domain->max_addr);
+               return -EFAULT;
+       }
+
+       ret = vm_domain_context_mapping(domain, pdev);
+       if (ret)
+               return ret;
+
+       ret = vm_domain_add_dev_info(domain, pdev);
+       return ret;
 }
-EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
+EXPORT_SYMBOL_GPL(intel_iommu_assign_device);

-int intel_iommu_page_mapping(
-       struct dmar_domain *domain, dma_addr_t iova,
-       u64 hpa, size_t size, int prot)
+
+void intel_iommu_deassign_device(struct dmar_domain *domain,
+                                struct pci_dev *pdev)
 {
-       int rc;
-       rc = domain_page_mapping(domain, iova, hpa, size, prot);
-       return rc;
+       vm_domain_remove_one_dev_info(domain, pdev);
 }
-EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
+EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);

-void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
+int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
+                         u64 hpa, size_t size, int prot)
 {
-       detach_domain_for_dev(domain, bus, devfn);
+       u64 max_addr;
+       int addr_width;
+       int ret;
+
+       max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
+       if (domain->max_addr < max_addr) {
+               int min_agaw;
+               u64 end;
+
+               /* check if minimum agaw is sufficient for mapped address */
+               min_agaw = vm_domain_min_agaw(domain);
+               addr_width = agaw_to_width(min_agaw);
+               end = DOMAIN_MAX_ADDR(addr_width);
+               end = end & VTD_PAGE_MASK;
+               if (end < max_addr) {
+                       printk(KERN_ERR "%s: iommu agaw (%d) is not "
+                              "sufficient for the mapped address (%llx)\n",
+                              __func__, min_agaw, max_addr);
+                       return -EFAULT;
+               }
+               domain->max_addr = max_addr;
+       }
+
+       ret = domain_page_mapping(domain, iova, hpa, size, prot);
+       return ret;
 }
-EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
+EXPORT_SYMBOL_GPL(intel_iommu_map_pages);

-struct dmar_domain *
-intel_iommu_find_domain(struct pci_dev *pdev)
+void intel_iommu_unmap_pages(struct dmar_domain *domain,
+                            dma_addr_t iova, size_t size)
 {
-       return find_domain(pdev);
+       dma_addr_t base;
+
+       /* The address might not be aligned */
+       base = iova & PAGE_MASK;
+       size = PAGE_ALIGN(size);
+       dma_pte_clear_range(domain, base, base + size);
 }
-EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
+EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);

 int intel_iommu_found(void)
 {
diff --git a/include/linux/dma_remapping.h b/include/linux/dma_remapping.h
index 952df39..3568ae6 100644
--- a/include/linux/dma_remapping.h
+++ b/include/linux/dma_remapping.h
@@ -111,11 +111,21 @@ struct dma_pte {
                (p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
 #define dma_pte_present(p) (((p).val & 3) != 0)

+/* domain flags, one domain owns one device by default */
+
+/* devices under the same p2p bridge are owned in one domain */
+#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES       (1 << 0)
+
+/* domain represents a virtual machine, more than one devices
+ * across iommus may be owned in one domain, e.g. kvm guest.
+ */
+#define DOMAIN_FLAG_VIRTUAL_MACHINE            (1 << 1)
+
 struct intel_iommu;

 struct dmar_domain {
        int     id;                     /* domain id */
-       struct intel_iommu *iommu;      /* back pointer to owning iommu */
+       unsigned long iommu_bmp;        /* bitmap of iommus this domain uses*/

        struct list_head devices;       /* all devices' list */
        struct iova_domain iovad;       /* iova's that belong to this domain */
@@ -123,12 +133,13 @@ struct dmar_domain {
        struct dma_pte  *pgd;           /* virtual address */
        spinlock_t      mapping_lock;   /* page table lock */
        int             gaw;            /* max guest address width */
+       int             agaw;           /* adjusted guest address width */

-       /* adjusted guest address width, 0 is level 2 30-bit */
-       int             agaw;
+       int             flags;          /* domain flag */

-#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
-       int             flags;
+       /* following fields are used in virtual machine case */
+       int             iommu_count;    /* reference count of iommu */
+       u64             max_addr;       /* maximum mapped address */
 };

 /* PCI domain-device relationship */
diff --git a/include/linux/intel-iommu.h b/include/linux/intel-iommu.h
index 3d017cf..c2f37b8 100644
--- a/include/linux/intel-iommu.h
+++ b/include/linux/intel-iommu.h
@@ -219,6 +219,8 @@ do {                                                                        \
        }                                                               \
 } while (0)

+#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
+
 #define QI_LENGTH      256     /* queue length */

 enum {
@@ -299,6 +301,7 @@ struct intel_iommu {
        struct dmar_domain **domains; /* ptr to domains */
        spinlock_t      lock; /* protect context, domain ids */
        struct root_entry *root_entry; /* virtual address */
+       int             agaw;

        unsigned int irq;
        unsigned char name[7];    /* Device Name */
@@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu *iommu, u16 did, u64 addr,

 extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu *iommu);

-void intel_iommu_domain_exit(struct dmar_domain *domain);
-struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
-int intel_iommu_context_mapping(struct dmar_domain *domain,
-                               struct pci_dev *pdev);
-int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
-                            u64 hpa, size_t size, int prot);
-void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn);
-struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
+struct dmar_domain *intel_iommu_alloc_domain(void);
+void intel_iommu_free_domain(struct dmar_domain *domain);
+int intel_iommu_assign_device(struct dmar_domain *domain,
+                             struct pci_dev *pdev);
+void intel_iommu_deassign_device(struct dmar_domain *domain,
+                                struct pci_dev *pdev);
+int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
+                         u64 hpa, size_t size, int prot);
+void intel_iommu_unmap_pages(struct dmar_domain *domain,
+                            dma_addr_t iova, size_t size);
 u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);

 #ifdef CONFIG_DMAR
--
1.5.1

[-- Attachment #2: 0001-vtd-multiple-device-assignment-for-kvm-v2.patch --]
[-- Type: application/octet-stream, Size: 37293 bytes --]

From 260c3191401888fd4de7028991653960a05ca6ac Mon Sep 17 00:00:00 2001
From: Weidong Han <weidong.han@intel.com>
Date: Thu, 27 Nov 2008 21:26:47 +0800
Subject: [PATCH] vtd multiple device assignment for kvm 


Signed-off-by: Weidong Han <weidong.han@intel.com>
---
 drivers/pci/dmar.c            |   15 +
 drivers/pci/intel-iommu.c     |  698 ++++++++++++++++++++++++++++++++++------
 include/linux/dma_remapping.h |   21 +-
 include/linux/intel-iommu.h   |   21 +-
 4 files changed, 637 insertions(+), 118 deletions(-)

diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
index 691b3ad..d6bdced 100644
--- a/drivers/pci/dmar.c
+++ b/drivers/pci/dmar.c
@@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
 	dmar_tbl = NULL;
 }
 
+extern int width_to_agaw(int width);
 
 int alloc_iommu(struct dmar_drhd_unit *drhd)
 {
@@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
 	int map_size;
 	u32 ver;
 	static int iommu_allocated = 0;
+	unsigned long sagaw;
+	int agaw;
 
 	iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);
 	if (!iommu)
@@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
 	iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
 	iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
 
+	/* set agaw, "SAGAW" may be different across iommus */
+	sagaw = cap_sagaw(iommu->cap);
+	for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
+	     agaw >= 0; agaw--)
+		if (test_bit(agaw, &sagaw))
+			break;
+	if (agaw < 0) {
+		printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n", sagaw);
+		goto error;
+	}
+	iommu->agaw = agaw;
+
 	/* the registers might be more than one page */
 	map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
 		cap_max_fault_reg_offset(iommu->cap));
diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
index 5c8baa4..55b96c4 100644
--- a/drivers/pci/intel-iommu.c
+++ b/drivers/pci/intel-iommu.c
@@ -50,8 +50,6 @@
 #define IOAPIC_RANGE_END	(0xfeefffff)
 #define IOVA_START_ADDR		(0x1000)
 
-#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
-
 #define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
 
 
@@ -64,6 +62,7 @@ struct deferred_flush_tables {
 	int next;
 	struct iova *iova[HIGH_WATER_MARK];
 	struct dmar_domain *domain[HIGH_WATER_MARK];
+	struct intel_iommu *iommu;
 };
 
 static struct deferred_flush_tables *deferred_flush;
@@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
 	kmem_cache_free(iommu_iova_cache, iova);
 }
 
+/* in native case, each domain is related to only one iommu */
+static struct intel_iommu *domain_get_only_iommu(struct dmar_domain *domain)
+{
+	struct dmar_drhd_unit *drhd;
+
+	for_each_drhd_unit(drhd) {
+		if (drhd->ignored)
+			continue;
+		if (test_bit(drhd->iommu->seq_id, &domain->iommu_bmp))
+			return drhd->iommu;
+	}
+
+	return NULL;
+}
+
+static void domain_flush_cache(struct dmar_domain *domain,
+			       void *addr, int size)
+{
+	struct intel_iommu *iommu;
+
+	if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
+		struct dmar_drhd_unit *drhd;
+
+		for_each_drhd_unit(drhd) {
+			if (drhd->ignored)
+				continue;
+			iommu = drhd->iommu;
+
+			if (!test_bit(iommu->seq_id, &domain->iommu_bmp))
+				continue;
+
+			if (!ecap_coherent(iommu->ecap))
+                		clflush_cache_range(addr, size);
+		}
+	}
+	else {
+		iommu = domain_get_only_iommu(domain);
+		if (iommu && !ecap_coherent(iommu->ecap))
+                	clflush_cache_range(addr, size);
+	}
+}
+
+static struct intel_iommu *device_find_matched_iommu(u8 bus, u8 devfn)
+{
+	struct dmar_drhd_unit *drhd = NULL;
+	int i;
+
+	for_each_drhd_unit(drhd) {
+		if (drhd->ignored)
+			continue;
+
+		for (i = 0; i < drhd->devices_cnt; i++)
+			if (drhd->devices[i]->bus->number == bus &&
+			    drhd->devices[i]->devfn == devfn)
+				return drhd->iommu;
+
+		if (drhd->include_all)
+			return drhd->iommu;
+	}
+
+	return NULL;
+}
+
 /* Gets context entry for a given bus and devfn */
 static struct context_entry * device_to_context_entry(struct intel_iommu *iommu,
 		u8 bus, u8 devfn)
@@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
 
 }
 
-static inline int width_to_agaw(int width)
+int width_to_agaw(int width)
 {
 	return (width - 30) / LEVEL_STRIDE;
 }
@@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
 					flags);
 				return NULL;
 			}
-			__iommu_flush_cache(domain->iommu, tmp_page,
-					PAGE_SIZE);
+			domain_flush_cache(domain, tmp_page, PAGE_SIZE);
 			dma_set_pte_addr(*pte, virt_to_phys(tmp_page));
 			/*
 			 * high level table always sets r/w, last level page
@@ -356,7 +417,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
 			 */
 			dma_set_pte_readable(*pte);
 			dma_set_pte_writable(*pte);
-			__iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+			domain_flush_cache(domain, pte, sizeof(*pte));
 		}
 		parent = phys_to_virt(dma_pte_addr(*pte));
 		level--;
@@ -399,7 +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain, u64 addr)
 
 	if (pte) {
 		dma_clear_pte(*pte);
-		__iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+		domain_flush_cache(domain, pte, sizeof(*pte));
 	}
 }
 
@@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct dmar_domain *domain,
 				free_pgtable_page(
 					phys_to_virt(dma_pte_addr(*pte)));
 				dma_clear_pte(*pte);
-				__iommu_flush_cache(domain->iommu,
-						pte, sizeof(*pte));
+				domain_flush_cache(domain, pte, sizeof(*pte));
 			}
 			tmp += level_size(level);
 		}
@@ -948,8 +1008,8 @@ static int iommu_init_domains(struct intel_iommu *iommu)
 	return 0;
 }
 
-
 static void domain_exit(struct dmar_domain *domain);
+static void vm_domain_exit(struct dmar_domain *domain);
 
 void free_dmar_iommu(struct intel_iommu *iommu)
 {
@@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
 	for (; i < cap_ndoms(iommu->cap); ) {
 		domain = iommu->domains[i];
 		clear_bit(i, iommu->domain_ids);
-		domain_exit(domain);
+
+		if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
+			if (--domain->iommu_count == 0)
+				vm_domain_exit(domain);
+		}
+		else
+			domain_exit(domain);
+
 		i = find_next_bit(iommu->domain_ids,
 			cap_ndoms(iommu->cap), i+1);
 	}
@@ -1006,8 +1073,11 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
 
 	set_bit(num, iommu->domain_ids);
 	domain->id = num;
-	domain->iommu = iommu;
+	memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
+	set_bit(iommu->seq_id, &domain->iommu_bmp);
 	iommu->domains[num] = domain;
+	domain->iommu_count = 1;
+	domain->flags = 0;
 	spin_unlock_irqrestore(&iommu->lock, flags);
 
 	return domain;
@@ -1016,10 +1086,12 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
 static void iommu_free_domain(struct dmar_domain *domain)
 {
 	unsigned long flags;
+	struct intel_iommu *iommu;
 
-	spin_lock_irqsave(&domain->iommu->lock, flags);
-	clear_bit(domain->id, domain->iommu->domain_ids);
-	spin_unlock_irqrestore(&domain->iommu->lock, flags);
+	iommu = domain_get_only_iommu(domain);
+	spin_lock_irqsave(&iommu->lock, flags);
+	clear_bit(domain->id, iommu->domain_ids);
+	spin_unlock_irqrestore(&iommu->lock, flags);
 }
 
 static struct iova_domain reserved_iova_list;
@@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
 	domain_reserve_special_ranges(domain);
 
 	/* calculate AGAW */
-	iommu = domain->iommu;
+	iommu = domain_get_only_iommu(domain);
 	if (guest_width > cap_mgaw(iommu->cap))
 		guest_width = cap_mgaw(iommu->cap);
 	domain->gaw = guest_width;
@@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
 	sagaw = cap_sagaw(iommu->cap);
 	if (!test_bit(agaw, &sagaw)) {
 		/* hardware doesn't support it, choose a bigger one */
-		pr_debug("IOMMU: hardware doesn't support agaw %d\n", agaw);
+		pr_debug("IOMMU: hardware doesn't support agaw %d\n",
+		  agaw);
 		agaw = find_next_bit(&sagaw, 5, agaw);
 		if (agaw >= 5)
 			return -ENODEV;
 	}
 	domain->agaw = agaw;
+
 	INIT_LIST_HEAD(&domain->devices);
 
 	/* always allocate the top pgd */
 	domain->pgd = (struct dma_pte *)alloc_pgtable_page();
 	if (!domain->pgd)
 		return -ENOMEM;
-	__iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
+	domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
 	return 0;
 }
 
@@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain *domain)
 }
 
 static int domain_context_mapping_one(struct dmar_domain *domain,
-		u8 bus, u8 devfn)
+		struct intel_iommu *iommu, u8 bus, u8 devfn)
 {
 	struct context_entry *context;
-	struct intel_iommu *iommu = domain->iommu;
 	unsigned long flags;
 
 	pr_debug("Set context mapping for %02x:%02x.%d\n",
@@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
 {
 	int ret;
 	struct pci_dev *tmp, *parent;
+	struct intel_iommu *iommu;
+
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return 	-ENODEV;
 
-	ret = domain_context_mapping_one(domain, pdev->bus->number,
-		pdev->devfn);
+	ret = domain_context_mapping_one(domain, iommu,
+		pdev->bus->number, pdev->devfn);
 	if (ret)
 		return ret;
 
@@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
 	/* Secondary interface's bus number and devfn 0 */
 	parent = pdev->bus->self;
 	while (parent != tmp) {
-		ret = domain_context_mapping_one(domain, parent->bus->number,
-			parent->devfn);
+		ret = domain_context_mapping_one(domain, iommu,
+			parent->bus->number, parent->devfn);
 		if (ret)
 			return ret;
 		parent = parent->bus->self;
 	}
 	if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
-		return domain_context_mapping_one(domain,
+		return domain_context_mapping_one(domain, iommu,
 			tmp->subordinate->number, 0);
 	else /* this is a legacy PCI bridge */
-		return domain_context_mapping_one(domain,
+		return domain_context_mapping_one(domain, iommu,
 			tmp->bus->number, tmp->devfn);
 }
 
-static int domain_context_mapped(struct dmar_domain *domain,
-	struct pci_dev *pdev)
+static int domain_context_mapped(struct pci_dev *pdev)
 {
 	int ret;
 	struct pci_dev *tmp, *parent;
+	struct intel_iommu *iommu;
 
-	ret = device_context_mapped(domain->iommu,
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return 0;
+
+	ret = device_context_mapped(iommu,
 		pdev->bus->number, pdev->devfn);
 	if (!ret)
 		return ret;
@@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct dmar_domain *domain,
 	/* Secondary interface's bus number and devfn 0 */
 	parent = pdev->bus->self;
 	while (parent != tmp) {
-		ret = device_context_mapped(domain->iommu, parent->bus->number,
+		ret = device_context_mapped(iommu, parent->bus->number,
 			parent->devfn);
 		if (!ret)
 			return ret;
 		parent = parent->bus->self;
 	}
 	if (tmp->is_pcie)
-		return device_context_mapped(domain->iommu,
+		return device_context_mapped(iommu,
 			tmp->subordinate->number, 0);
 	else
-		return device_context_mapped(domain->iommu,
+		return device_context_mapped(iommu,
 			tmp->bus->number, tmp->devfn);
 }
 
@@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
 		BUG_ON(dma_pte_addr(*pte));
 		dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
 		dma_set_pte_prot(*pte, prot);
-		__iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+		domain_flush_cache(domain, pte, sizeof(*pte));
 		start_pfn++;
 		index++;
 	}
 	return 0;
 }
 
-static void detach_domain_for_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
+static void detach_domain_for_dev(struct dmar_domain *domain,
+				  u8 bus, u8 devfn)
 {
-	clear_context_table(domain->iommu, bus, devfn);
-	domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
-					   DMA_CCMD_GLOBAL_INVL, 0);
-	domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
-					 DMA_TLB_GLOBAL_FLUSH, 0);
+	struct intel_iommu *iommu;
+
+	iommu = device_find_matched_iommu(bus, devfn);
+	if (!iommu)
+		return;
+
+	clear_context_table(iommu, bus, devfn);
+	iommu->flush.flush_context(iommu, 0, 0, 0,
+				   DMA_CCMD_GLOBAL_INVL, 0);
+	iommu->flush.flush_iotlb(iommu, 0, 0, 0,
+				 DMA_TLB_GLOBAL_FLUSH, 0);
 }
 
 static void domain_remove_dev_info(struct dmar_domain *domain)
@@ -1336,7 +1425,6 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
 {
 	struct dmar_domain *domain, *found = NULL;
 	struct intel_iommu *iommu;
-	struct dmar_drhd_unit *drhd;
 	struct device_domain_info *info, *tmp;
 	struct pci_dev *dev_tmp;
 	unsigned long flags;
@@ -1371,13 +1459,9 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
 	}
 
 	/* Allocate new domain for the device */
-	drhd = dmar_find_matched_drhd_unit(pdev);
-	if (!drhd) {
-		printk(KERN_ERR "IOMMU: can't find DMAR for device %s\n",
-			pci_name(pdev));
-		return NULL;
-	}
-	iommu = drhd->iommu;
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return 	NULL;
 
 	domain = iommu_alloc_domain(iommu);
 	if (!domain)
@@ -1400,7 +1484,7 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
 		info->dev = NULL;
 		info->domain = domain;
 		/* This domain is shared by devices under p2p bridge */
-		domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
+		domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
 
 		/* pcie-to-pci bridge already has a domain, uses it */
 		found = NULL;
@@ -1805,7 +1889,7 @@ get_valid_domain_for_dev(struct pci_dev *pdev)
 	}
 
 	/* make sure context mapping is ok */
-	if (unlikely(!domain_context_mapped(domain, pdev))) {
+	if (unlikely(!domain_context_mapped(pdev))) {
 		ret = domain_context_mapping(domain, pdev);
 		if (ret) {
 			printk(KERN_ERR
@@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
 {
 	struct pci_dev *pdev = to_pci_dev(hwdev);
 	struct dmar_domain *domain;
+	struct intel_iommu *iommu;
 	phys_addr_t start_paddr;
 	struct iova *iova;
 	int prot = 0;
@@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
 	if (!domain)
 		return 0;
 
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return 0;
+
 	size = aligned_size((u64)paddr, size);
 
 	iova = __intel_alloc_iova(hwdev, domain, size, pdev->dma_mask);
@@ -1849,7 +1938,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
 	 * mappings..
 	 */
 	if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
-			!cap_zlr(domain->iommu->cap))
+			!cap_zlr(iommu->cap))
 		prot |= DMA_PTE_READ;
 	if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
 		prot |= DMA_PTE_WRITE;
@@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
 		goto error;
 
 	/* it's a non-present to present mapping */
-	ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
+	ret = iommu_flush_iotlb_psi(iommu, domain->id,
 			start_paddr, size >> VTD_PAGE_SHIFT, 1);
 	if (ret)
-		iommu_flush_write_buffer(domain->iommu);
+		iommu_flush_write_buffer(iommu);
 
 	return start_paddr + ((u64)paddr & (~PAGE_MASK));
 
@@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
 	/* just flush them all */
 	for (i = 0; i < g_num_of_iommus; i++) {
 		if (deferred_flush[i].next) {
-			struct intel_iommu *iommu =
-				deferred_flush[i].domain[0]->iommu;
+			struct intel_iommu *iommu = deferred_flush[i].iommu;
 
 			iommu->flush.flush_iotlb(iommu, 0, 0, 0,
 						 DMA_TLB_GLOBAL_FLUSH, 0);
@@ -1921,7 +2009,8 @@ static void flush_unmaps_timeout(unsigned long data)
 	spin_unlock_irqrestore(&async_umap_flush_lock, flags);
 }
 
-static void add_unmap(struct dmar_domain *dom, struct iova *iova)
+static void add_unmap(struct dmar_domain *dom,
+		      struct intel_iommu *iommu, struct iova *iova)
 {
 	unsigned long flags;
 	int next, iommu_id;
@@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova)
 	if (list_size == HIGH_WATER_MARK)
 		flush_unmaps();
 
-	iommu_id = dom->iommu->seq_id;
+	iommu_id = iommu->seq_id;
 
 	next = deferred_flush[iommu_id].next;
 	deferred_flush[iommu_id].domain[next] = dom;
 	deferred_flush[iommu_id].iova[next] = iova;
+	deferred_flush[iommu_id].iommu = iommu;
 	deferred_flush[iommu_id].next++;
 
 	if (!timer_on) {
@@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
 	struct dmar_domain *domain;
 	unsigned long start_addr;
 	struct iova *iova;
+	struct intel_iommu *iommu;
 
 	if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
 		return;
 	domain = find_domain(pdev);
 	BUG_ON(!domain);
 
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return;
+
 	iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
 	if (!iova)
 		return;
@@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
 	/* free page tables */
 	dma_pte_free_pagetable(domain, start_addr, start_addr + size);
 	if (intel_iommu_strict) {
-		if (iommu_flush_iotlb_psi(domain->iommu,
+		if (iommu_flush_iotlb_psi(iommu,
 			domain->id, start_addr, size >> VTD_PAGE_SHIFT, 0))
-			iommu_flush_write_buffer(domain->iommu);
+			iommu_flush_write_buffer(iommu);
 		/* free iova */
 		__free_iova(&domain->iovad, iova);
 	} else {
-		add_unmap(domain, iova);
+		add_unmap(domain, iommu, iova);
 		/*
 		 * queue up the release of the unmap to save the 1/6th of the
 		 * cpu used up by the iotlb flush operation...
@@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
 	size_t size = 0;
 	void *addr;
 	struct scatterlist *sg;
+	struct intel_iommu *iommu;
 
 	if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
 		return;
 
 	domain = find_domain(pdev);
 
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return;
+
 	iova = find_iova(&domain->iovad, IOVA_PFN(sglist[0].dma_address));
 	if (!iova)
 		return;
@@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
 	/* free page tables */
 	dma_pte_free_pagetable(domain, start_addr, start_addr + size);
 
-	if (iommu_flush_iotlb_psi(domain->iommu, domain->id, start_addr,
+	if (iommu_flush_iotlb_psi(iommu, domain->id, start_addr,
 			size >> VTD_PAGE_SHIFT, 0))
-		iommu_flush_write_buffer(domain->iommu);
+		iommu_flush_write_buffer(iommu);
 
 	/* free iova */
 	__free_iova(&domain->iovad, iova);
@@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
 	int ret;
 	struct scatterlist *sg;
 	unsigned long start_addr;
+	struct intel_iommu *iommu;
 
 	BUG_ON(dir == DMA_NONE);
 	if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
@@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
 	if (!domain)
 		return 0;
 
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return 0;
+
 	for_each_sg(sglist, sg, nelems, i) {
 		addr = SG_ENT_VIRT_ADDRESS(sg);
 		addr = (void *)virt_to_phys(addr);
@@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
 	 * mappings..
 	 */
 	if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
-			!cap_zlr(domain->iommu->cap))
+			!cap_zlr(iommu->cap))
 		prot |= DMA_PTE_READ;
 	if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
 		prot |= DMA_PTE_WRITE;
@@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
 	}
 
 	/* it's a non-present to present mapping */
-	if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
+	if (iommu_flush_iotlb_psi(iommu, domain->id,
 			start_addr, offset >> VTD_PAGE_SHIFT, 1))
-		iommu_flush_write_buffer(domain->iommu);
+		iommu_flush_write_buffer(iommu);
 	return nelems;
 }
 
@@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
 	return 0;
 }
 
-void intel_iommu_domain_exit(struct dmar_domain *domain)
+/* domain id for virtual machine, it won't be set in context */
+static unsigned long vm_domid;
+
+static int vm_domain_min_agaw(struct dmar_domain *domain)
+{
+	struct dmar_drhd_unit *drhd;
+	struct intel_iommu *iommu;
+	int min_agaw = domain->agaw;
+
+	for_each_drhd_unit(drhd) {
+		if (drhd->ignored)
+			continue;
+		iommu = drhd->iommu;
+
+		if (test_bit(iommu->seq_id, &domain->iommu_bmp))
+			if (min_agaw > iommu->agaw)
+				min_agaw = iommu->agaw;
+	}
+
+	return min_agaw;
+}
+
+static int vm_domain_add_dev_info(struct dmar_domain *domain,
+				  struct pci_dev *pdev)
+{
+	struct device_domain_info *info;
+	unsigned long flags;
+
+	info = alloc_devinfo_mem();
+	if (!info)
+		return -ENOMEM;
+
+	info->bus = pdev->bus->number;
+	info->devfn = pdev->devfn;
+	info->dev = pdev;
+	info->domain = domain;
+
+	spin_lock_irqsave(&device_domain_lock, flags);
+	list_add(&info->link, &domain->devices);
+	list_add(&info->global, &device_domain_list);
+	pdev->dev.archdata.iommu = info;
+	spin_unlock_irqrestore(&device_domain_lock, flags);
+	
+	return 0;
+}
+
+static void vm_domain_remove_one_dev_info(struct dmar_domain *domain,
+					  struct pci_dev *pdev)
+{
+	struct device_domain_info *info;
+	struct intel_iommu *iommu;
+	unsigned long flags;
+	int found = 0;
+
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+
+	spin_lock_irqsave(&device_domain_lock, flags);
+	while (!list_empty(&domain->devices)) {
+		info = list_entry(domain->devices.next,
+			struct device_domain_info, link);
+		if (info->bus == pdev->bus->number &&
+		    info->devfn == pdev->devfn) {
+			list_del(&info->link);
+			list_del(&info->global);
+			if (info->dev)
+				info->dev->dev.archdata.iommu = NULL;
+			spin_unlock_irqrestore(&device_domain_lock, flags);
+
+			detach_domain_for_dev(info->domain,
+					      info->bus, info->devfn);
+			free_devinfo_mem(info);
+
+			spin_lock_irqsave(&device_domain_lock, flags);
+
+			if (found)
+				break;
+			else
+				continue;
+		}
+
+		/* if there is no other devices under the same iommu
+		 * owned by this domain, clear this iommu in iommu_bmp 
+		 */
+		if (device_find_matched_iommu(info->bus, info->devfn) == iommu)
+			found = 1;
+	}
+
+	if (found == 0) {
+		spin_lock_irqsave(&iommu->lock, flags);
+		clear_bit(iommu->seq_id, &domain->iommu_bmp);
+		domain->iommu_count--;
+		spin_unlock_irqrestore(&iommu->lock, flags);
+	}
+
+	spin_unlock_irqrestore(&device_domain_lock, flags);
+}
+
+static void vm_domain_remove_all_dev_info(struct dmar_domain *domain)
+{
+	struct device_domain_info *info;
+	struct intel_iommu *iommu;
+	unsigned long flags;
+
+	spin_lock_irqsave(&device_domain_lock, flags);
+	while (!list_empty(&domain->devices)) {
+		info = list_entry(domain->devices.next,
+			struct device_domain_info, link);
+		list_del(&info->link);
+		list_del(&info->global);
+		if (info->dev)
+			info->dev->dev.archdata.iommu = NULL;
+
+		spin_unlock_irqrestore(&device_domain_lock, flags);
+		detach_domain_for_dev(info->domain,
+				      info->bus, info->devfn);
+
+		/* clear this iommu in iommu_bmp */
+		iommu = device_find_matched_iommu(info->bus, info->devfn);
+		spin_lock_irqsave(&iommu->lock, flags);
+		if (test_and_clear_bit(iommu->seq_id,
+				       &domain->iommu_bmp))
+			domain->iommu_count--;
+		spin_unlock_irqrestore(&iommu->lock, flags);
+
+		free_devinfo_mem(info);
+		spin_lock_irqsave(&device_domain_lock, flags);
+	}
+	spin_unlock_irqrestore(&device_domain_lock, flags);
+}
+
+static int vm_domain_context_mapping_one(struct dmar_domain *domain,
+		struct intel_iommu *iommu, u8 bus, u8 devfn)
+{
+	struct context_entry *context;
+	unsigned long flags;
+	struct dma_pte *pgd;
+	unsigned long num;
+	unsigned long ndomains;
+	int id;
+	int agaw;
+	int found = 0;
+
+	pr_debug("Set context mapping for %02x:%02x.%d\n",
+		bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
+	BUG_ON(!domain->pgd);
+	context = device_to_context_entry(iommu, bus, devfn);
+	if (!context)
+		return -ENOMEM;
+	spin_lock_irqsave(&iommu->lock, flags);
+	if (context_present(*context)) {
+		spin_unlock_irqrestore(&iommu->lock, flags);
+		return 0;
+	}
+
+	id = domain->id;
+
+	/* find an available domain id for this device in iommu */
+	ndomains = cap_ndoms(iommu->cap);
+	num = find_first_bit(iommu->domain_ids, ndomains);
+	for (; num < ndomains; ) {
+		if (iommu->domains[num] == domain) {
+			id = num;
+			found = 1;
+			break;
+		}
+		num = find_next_bit(iommu->domain_ids,
+		  cap_ndoms(iommu->cap), num+1);
+	}
+
+	if (found == 0) {
+		num = find_first_zero_bit(iommu->domain_ids, ndomains);
+		if (num >= ndomains) {
+			spin_unlock_irqrestore(&iommu->lock, flags);
+			printk(KERN_ERR "IOMMU: no free domain ids\n");
+			return -EFAULT;
+		}
+
+		set_bit(num, iommu->domain_ids);
+		iommu->domains[num] = domain;
+		id = num;
+	}
+
+	pgd = domain->pgd;
+
+	/* Skip top levels of page tables for
+	 * iommu which has less agaw than default.
+	 */
+	for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
+		pgd = phys_to_virt(dma_pte_addr(*pgd));
+		if (!dma_pte_present(*pgd)) {
+			spin_unlock_irqrestore(&iommu->lock, flags);
+			return -ENOMEM;
+		}
+	}
+
+	context_set_domain_id(*context, id);
+	context_set_address_width(*context, iommu->agaw);
+	context_set_address_root(*context, virt_to_phys(pgd));
+	context_set_translation_type(*context, CONTEXT_TT_MULTI_LEVEL);
+	context_set_fault_enable(*context);
+	context_set_present(*context);
+	__iommu_flush_cache(iommu, context, sizeof(*context));
+
+	/* it's a non-present to present mapping */
+	if (iommu->flush.flush_context(iommu, id,
+		(((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
+		DMA_CCMD_DEVICE_INVL, 1))
+		iommu_flush_write_buffer(iommu);
+	else
+		iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_DSI_FLUSH, 0);
+
+	if (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp))
+		domain->iommu_count++;
+	spin_unlock_irqrestore(&iommu->lock, flags);
+	return 0;
+}
+
+static int
+vm_domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
+{
+	int ret;
+	struct pci_dev *tmp, *parent;
+	struct intel_iommu *iommu;
+
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return 	-ENODEV;
+
+	ret = vm_domain_context_mapping_one(domain, iommu,
+		pdev->bus->number, pdev->devfn);
+	if (ret)
+		return ret;
+
+	/* dependent device mapping */
+	tmp = pci_find_upstream_pcie_bridge(pdev);
+	if (!tmp)
+		return 0;
+	/* Secondary interface's bus number and devfn 0 */
+	parent = pdev->bus->self;
+	while (parent != tmp) {
+		ret = vm_domain_context_mapping_one(domain, iommu,
+			parent->bus->number, parent->devfn);
+		if (ret)
+			return ret;
+		parent = parent->bus->self;
+	}
+	if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
+		return vm_domain_context_mapping_one(domain, iommu,
+			tmp->subordinate->number, 0);
+	else /* this is a legacy PCI bridge */
+		return vm_domain_context_mapping_one(domain, iommu,
+			tmp->bus->number, tmp->devfn);
+}
+
+
+static int vm_domain_init(struct dmar_domain *domain, int guest_width)
+{
+	int adjust_width;
+
+	init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
+	spin_lock_init(&domain->mapping_lock);
+
+	domain_reserve_special_ranges(domain);
+
+	/* calculate AGAW */
+	domain->gaw = guest_width;
+	adjust_width = guestwidth_to_adjustwidth(guest_width);
+	domain->agaw = width_to_agaw(adjust_width);
+
+	INIT_LIST_HEAD(&domain->devices);
+
+	/* always allocate the top pgd */
+	domain->pgd = (struct dma_pte *)alloc_pgtable_page();
+	if (!domain->pgd)
+		return -ENOMEM;
+	domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
+	return 0;
+}
+
+static void iommu_free_vm_domain(struct dmar_domain *domain)
+{
+	unsigned long flags;
+	struct dmar_drhd_unit *drhd;
+	struct intel_iommu *iommu;
+	unsigned long i;
+	unsigned long ndomains;
+
+	for_each_drhd_unit(drhd) {
+		if (drhd->ignored)
+			continue;
+		iommu = drhd->iommu;
+
+		ndomains = cap_ndoms(iommu->cap);
+		i = find_first_bit(iommu->domain_ids, ndomains);
+		for (; i < ndomains; ) {
+			if (iommu->domains[i] == domain) {
+				spin_lock_irqsave(&iommu->lock, flags);
+				clear_bit(i, iommu->domain_ids);
+				iommu->domains[i] = NULL;
+				spin_unlock_irqrestore(&iommu->lock, flags);
+				break;
+			}
+			i = find_next_bit(iommu->domain_ids, ndomains, i+1);
+		}
+	}
+}
+
+static void vm_domain_exit(struct dmar_domain *domain)
 {
 	u64 end;
 
@@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
 	if (!domain)
 		return;
 
+	vm_domain_remove_all_dev_info(domain);
+	/* destroy iovas */
+	put_iova_domain(&domain->iovad);
 	end = DOMAIN_MAX_ADDR(domain->gaw);
-	end = end & (~VTD_PAGE_MASK);
+	end &= VTD_PAGE_MASK;
 
 	/* clear ptes */
 	dma_pte_clear_range(domain, 0, end);
@@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
 	/* free page tables */
 	dma_pte_free_pagetable(domain, 0, end);
 
-	iommu_free_domain(domain);
+	iommu_free_vm_domain(domain);
 	free_domain_mem(domain);
 }
-EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
 
-struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
+static struct dmar_domain *iommu_alloc_vm_domain(void)
 {
-	struct dmar_drhd_unit *drhd;
 	struct dmar_domain *domain;
-	struct intel_iommu *iommu;
 
-	drhd = dmar_find_matched_drhd_unit(pdev);
-	if (!drhd) {
-		printk(KERN_ERR "intel_iommu_domain_alloc: drhd == NULL\n");
+	domain = alloc_domain_mem();
+	if (!domain)
 		return NULL;
-	}
 
-	iommu = drhd->iommu;
-	if (!iommu) {
-		printk(KERN_ERR
-			"intel_iommu_domain_alloc: iommu == NULL\n");
-		return NULL;
-	}
-	domain = iommu_alloc_domain(iommu);
+	domain->id = vm_domid++;
+	domain->iommu_count = 0;
+	domain->max_addr = 0;
+	memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
+	domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE;
+
+	return domain;
+}
+
+struct dmar_domain *intel_iommu_alloc_domain(void)
+{
+	struct dmar_domain *domain;
+
+	domain = iommu_alloc_vm_domain();
 	if (!domain) {
 		printk(KERN_ERR
 			"intel_iommu_domain_alloc: domain == NULL\n");
 		return NULL;
 	}
-	if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
+	if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
 		printk(KERN_ERR
 			"intel_iommu_domain_alloc: domain_init() failed\n");
-		intel_iommu_domain_exit(domain);
+		vm_domain_exit(domain);
 		return NULL;
 	}
+
 	return domain;
 }
-EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
+EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
+
+void intel_iommu_free_domain(struct dmar_domain *domain)
+{
+	vm_domain_exit(domain);
+}
+EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
 
-int intel_iommu_context_mapping(
-	struct dmar_domain *domain, struct pci_dev *pdev)
+int intel_iommu_assign_device(struct dmar_domain *domain,
+			      struct pci_dev *pdev)
 {
-	int rc;
-	rc = domain_context_mapping(domain, pdev);
-	return rc;
+	struct intel_iommu *iommu;
+	int addr_width;
+	u64 end;
+	int ret;
+
+	/* normally pdev is not mapped */
+	if (unlikely(domain_context_mapped(pdev))) {
+		struct dmar_domain *old_domain;
+
+		old_domain = find_domain(pdev);
+		if (old_domain) {
+			if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE)
+				vm_domain_remove_one_dev_info(old_domain, pdev);
+			else
+				domain_remove_dev_info(old_domain);
+		}
+	}
+
+	iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+	if (!iommu)
+		return 	-ENODEV;
+
+	/* check if this iommu agaw is sufficient for max mapped address */
+	addr_width = agaw_to_width(iommu->agaw);
+	end = DOMAIN_MAX_ADDR(addr_width);
+	end = end & VTD_PAGE_MASK;
+	if (end < domain->max_addr) {
+		printk(KERN_ERR "%s: iommu agaw (%d) is not "
+		       "sufficient for the mapped address (%llx)\n",
+		       __func__, iommu->agaw, domain->max_addr);
+		return -EFAULT;
+	}
+	
+	ret = vm_domain_context_mapping(domain, pdev);
+	if (ret)
+		return ret;
+
+	ret = vm_domain_add_dev_info(domain, pdev);
+	return ret;
 }
-EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
+EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
 
-int intel_iommu_page_mapping(
-	struct dmar_domain *domain, dma_addr_t iova,
-	u64 hpa, size_t size, int prot)
+
+void intel_iommu_deassign_device(struct dmar_domain *domain,
+				 struct pci_dev *pdev)
 {
-	int rc;
-	rc = domain_page_mapping(domain, iova, hpa, size, prot);
-	return rc;
+	vm_domain_remove_one_dev_info(domain, pdev);
 }
-EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
+EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
 
-void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
+int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
+			  u64 hpa, size_t size, int prot)
 {
-	detach_domain_for_dev(domain, bus, devfn);
+	u64 max_addr;
+	int addr_width;
+	int ret;
+
+	max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
+	if (domain->max_addr < max_addr) {
+		int min_agaw;
+		u64 end;
+
+		/* check if minimum agaw is sufficient for mapped address */
+		min_agaw = vm_domain_min_agaw(domain);
+		addr_width = agaw_to_width(min_agaw);
+		end = DOMAIN_MAX_ADDR(addr_width);
+		end = end & VTD_PAGE_MASK;
+		if (end < max_addr) {
+			printk(KERN_ERR "%s: iommu agaw (%d) is not "
+			       "sufficient for the mapped address (%llx)\n",
+			       __func__, min_agaw, max_addr);
+			return -EFAULT;
+		}
+		domain->max_addr = max_addr;
+	}
+	
+	ret = domain_page_mapping(domain, iova, hpa, size, prot);
+	return ret;
 }
-EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
+EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
 
-struct dmar_domain *
-intel_iommu_find_domain(struct pci_dev *pdev)
+void intel_iommu_unmap_pages(struct dmar_domain *domain,
+			     dma_addr_t iova, size_t size)
 {
-	return find_domain(pdev);
+	dma_addr_t base;
+	
+	/* The address might not be aligned */
+	base = iova & PAGE_MASK;
+	size = PAGE_ALIGN(size);
+	dma_pte_clear_range(domain, base, base + size);
 }
-EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
+EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
 
 int intel_iommu_found(void)
 {
diff --git a/include/linux/dma_remapping.h b/include/linux/dma_remapping.h
index 952df39..3568ae6 100644
--- a/include/linux/dma_remapping.h
+++ b/include/linux/dma_remapping.h
@@ -111,11 +111,21 @@ struct dma_pte {
 		(p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
 #define dma_pte_present(p) (((p).val & 3) != 0)
 
+/* domain flags, one domain owns one device by default */
+
+/* devices under the same p2p bridge are owned in one domain */ 
+#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES	(1 << 0)
+
+/* domain represents a virtual machine, more than one devices 
+ * across iommus may be owned in one domain, e.g. kvm guest.
+ */
+#define DOMAIN_FLAG_VIRTUAL_MACHINE		(1 << 1)
+
 struct intel_iommu;
 
 struct dmar_domain {
 	int	id;			/* domain id */
-	struct intel_iommu *iommu;	/* back pointer to owning iommu */
+	unsigned long iommu_bmp;	/* bitmap of iommus this domain uses*/
 
 	struct list_head devices; 	/* all devices' list */
 	struct iova_domain iovad;	/* iova's that belong to this domain */
@@ -123,12 +133,13 @@ struct dmar_domain {
 	struct dma_pte	*pgd;		/* virtual address */
 	spinlock_t	mapping_lock;	/* page table lock */
 	int		gaw;		/* max guest address width */
+	int		agaw;		/* adjusted guest address width */
 
-	/* adjusted guest address width, 0 is level 2 30-bit */
-	int		agaw;
+	int		flags;		/* domain flag */
 
-#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
-	int		flags;
+	/* following fields are used in virtual machine case */
+	int		iommu_count;	/* reference count of iommu */
+	u64		max_addr;	/* maximum mapped address */
 };
 
 /* PCI domain-device relationship */
diff --git a/include/linux/intel-iommu.h b/include/linux/intel-iommu.h
index 3d017cf..c2f37b8 100644
--- a/include/linux/intel-iommu.h
+++ b/include/linux/intel-iommu.h
@@ -219,6 +219,8 @@ do {									\
 	}								\
 } while (0)
 
+#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
+
 #define QI_LENGTH	256	/* queue length */
 
 enum {
@@ -299,6 +301,7 @@ struct intel_iommu {
 	struct dmar_domain **domains; /* ptr to domains */
 	spinlock_t	lock; /* protect context, domain ids */
 	struct root_entry *root_entry; /* virtual address */
+	int		agaw;
 
 	unsigned int irq;
 	unsigned char name[7];    /* Device Name */
@@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu *iommu, u16 did, u64 addr,
 
 extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu *iommu);
 
-void intel_iommu_domain_exit(struct dmar_domain *domain);
-struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
-int intel_iommu_context_mapping(struct dmar_domain *domain,
-				struct pci_dev *pdev);
-int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
-			     u64 hpa, size_t size, int prot);
-void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn);
-struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
+struct dmar_domain *intel_iommu_alloc_domain(void);
+void intel_iommu_free_domain(struct dmar_domain *domain);
+int intel_iommu_assign_device(struct dmar_domain *domain,
+			      struct pci_dev *pdev);
+void intel_iommu_deassign_device(struct dmar_domain *domain,
+				 struct pci_dev *pdev);
+int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
+			  u64 hpa, size_t size, int prot);
+void intel_iommu_unmap_pages(struct dmar_domain *domain,
+			     dma_addr_t iova, size_t size);
 u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
 
 #ifdef CONFIG_DMAR
-- 
1.5.1


^ permalink raw reply related	[flat|nested] 7+ messages in thread

* Re: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
  2008-11-27 13:49 [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM Han, Weidong
@ 2008-11-28 13:27 ` Joerg Roedel
  2008-12-01  6:17   ` Han, Weidong
  0 siblings, 1 reply; 7+ messages in thread
From: Joerg Roedel @ 2008-11-28 13:27 UTC (permalink / raw)
  To: Han, Weidong
  Cc: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity',
	Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
	'iommu@lists.linux-foundation.org'

Hmm, I tried to apply this patch against avi/master and linus/master but
get merge conflicts. Where do these patches apply cleanly?

Joerg

On Thu, Nov 27, 2008 at 09:49:04PM +0800, Han, Weidong wrote:
> In order to support multiple device assignment for KVM, this patch does following main changes:
>    - extend dmar_domain to own multiple devices from different iommus, use a bitmap of iommus to replace iommu pointer in dmar_domain.
>    - implement independent low level functions for kvm, then won't impact native VT-d.
>    - "SAGAW" capability may be different across iommus, that's to say the VT-d page table levels may be different among iommus. This patch uses a defaut agaw, and skip top levels of page tables for iommus which have smaller agaw than default.
>    - rename the APIs for kvm VT-d, make it more readable.
> 
> 
> Signed-off-by: Weidong Han <weidong.han@intel.com>
> ---
>  drivers/pci/dmar.c            |   15 +
>  drivers/pci/intel-iommu.c     |  698 ++++++++++++++++++++++++++++++++++------
>  include/linux/dma_remapping.h |   21 +-
>  include/linux/intel-iommu.h   |   21 +-
>  4 files changed, 637 insertions(+), 118 deletions(-)
> 
> diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
> index 691b3ad..d6bdced 100644
> --- a/drivers/pci/dmar.c
> +++ b/drivers/pci/dmar.c
> @@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
>         dmar_tbl = NULL;
>  }
> 
> +extern int width_to_agaw(int width);
> 
>  int alloc_iommu(struct dmar_drhd_unit *drhd)
>  {
> @@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
>         int map_size;
>         u32 ver;
>         static int iommu_allocated = 0;
> +       unsigned long sagaw;
> +       int agaw;
> 
>         iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);
>         if (!iommu)
> @@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
>         iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
>         iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
> 
> +       /* set agaw, "SAGAW" may be different across iommus */
> +       sagaw = cap_sagaw(iommu->cap);
> +       for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
> +            agaw >= 0; agaw--)
> +               if (test_bit(agaw, &sagaw))
> +                       break;
> +       if (agaw < 0) {
> +               printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n", sagaw);
> +               goto error;
> +       }
> +       iommu->agaw = agaw;
> +
>         /* the registers might be more than one page */
>         map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
>                 cap_max_fault_reg_offset(iommu->cap));
> diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
> index 5c8baa4..55b96c4 100644
> --- a/drivers/pci/intel-iommu.c
> +++ b/drivers/pci/intel-iommu.c
> @@ -50,8 +50,6 @@
>  #define IOAPIC_RANGE_END       (0xfeefffff)
>  #define IOVA_START_ADDR                (0x1000)
> 
> -#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
> -
>  #define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
> 
> 
> @@ -64,6 +62,7 @@ struct deferred_flush_tables {
>         int next;
>         struct iova *iova[HIGH_WATER_MARK];
>         struct dmar_domain *domain[HIGH_WATER_MARK];
> +       struct intel_iommu *iommu;
>  };
> 
>  static struct deferred_flush_tables *deferred_flush;
> @@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
>         kmem_cache_free(iommu_iova_cache, iova);
>  }
> 
> +/* in native case, each domain is related to only one iommu */
> +static struct intel_iommu *domain_get_only_iommu(struct dmar_domain *domain)
> +{
> +       struct dmar_drhd_unit *drhd;
> +
> +       for_each_drhd_unit(drhd) {
> +               if (drhd->ignored)
> +                       continue;
> +               if (test_bit(drhd->iommu->seq_id, &domain->iommu_bmp))
> +                       return drhd->iommu;
> +       }
> +
> +       return NULL;
> +}
> +
> +static void domain_flush_cache(struct dmar_domain *domain,
> +                              void *addr, int size)
> +{
> +       struct intel_iommu *iommu;
> +
> +       if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
> +               struct dmar_drhd_unit *drhd;
> +
> +               for_each_drhd_unit(drhd) {
> +                       if (drhd->ignored)
> +                               continue;
> +                       iommu = drhd->iommu;
> +
> +                       if (!test_bit(iommu->seq_id, &domain->iommu_bmp))
> +                               continue;
> +
> +                       if (!ecap_coherent(iommu->ecap))
> +                               clflush_cache_range(addr, size);
> +               }
> +       }
> +       else {
> +               iommu = domain_get_only_iommu(domain);
> +               if (iommu && !ecap_coherent(iommu->ecap))
> +                       clflush_cache_range(addr, size);
> +       }
> +}
> +
> +static struct intel_iommu *device_find_matched_iommu(u8 bus, u8 devfn)
> +{
> +       struct dmar_drhd_unit *drhd = NULL;
> +       int i;
> +
> +       for_each_drhd_unit(drhd) {
> +               if (drhd->ignored)
> +                       continue;
> +
> +               for (i = 0; i < drhd->devices_cnt; i++)
> +                       if (drhd->devices[i]->bus->number == bus &&
> +                           drhd->devices[i]->devfn == devfn)
> +                               return drhd->iommu;
> +
> +               if (drhd->include_all)
> +                       return drhd->iommu;
> +       }
> +
> +       return NULL;
> +}
> +
>  /* Gets context entry for a given bus and devfn */
>  static struct context_entry * device_to_context_entry(struct intel_iommu *iommu,
>                 u8 bus, u8 devfn)
> @@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
> 
>  }
> 
> -static inline int width_to_agaw(int width)
> +int width_to_agaw(int width)
>  {
>         return (width - 30) / LEVEL_STRIDE;
>  }
> @@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
>                                         flags);
>                                 return NULL;
>                         }
> -                       __iommu_flush_cache(domain->iommu, tmp_page,
> -                                       PAGE_SIZE);
> +                       domain_flush_cache(domain, tmp_page, PAGE_SIZE);
>                         dma_set_pte_addr(*pte, virt_to_phys(tmp_page));
>                         /*
>                          * high level table always sets r/w, last level page
> @@ -356,7 +417,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
>                          */
>                         dma_set_pte_readable(*pte);
>                         dma_set_pte_writable(*pte);
> -                       __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
> +                       domain_flush_cache(domain, pte, sizeof(*pte));
>                 }
>                 parent = phys_to_virt(dma_pte_addr(*pte));
>                 level--;
> @@ -399,7 +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain, u64 addr)
> 
>         if (pte) {
>                 dma_clear_pte(*pte);
> -               __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
> +               domain_flush_cache(domain, pte, sizeof(*pte));
>         }
>  }
> 
> @@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct dmar_domain *domain,
>                                 free_pgtable_page(
>                                         phys_to_virt(dma_pte_addr(*pte)));
>                                 dma_clear_pte(*pte);
> -                               __iommu_flush_cache(domain->iommu,
> -                                               pte, sizeof(*pte));
> +                               domain_flush_cache(domain, pte, sizeof(*pte));
>                         }
>                         tmp += level_size(level);
>                 }
> @@ -948,8 +1008,8 @@ static int iommu_init_domains(struct intel_iommu *iommu)
>         return 0;
>  }
> 
> -
>  static void domain_exit(struct dmar_domain *domain);
> +static void vm_domain_exit(struct dmar_domain *domain);
> 
>  void free_dmar_iommu(struct intel_iommu *iommu)
>  {
> @@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
>         for (; i < cap_ndoms(iommu->cap); ) {
>                 domain = iommu->domains[i];
>                 clear_bit(i, iommu->domain_ids);
> -               domain_exit(domain);
> +
> +               if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
> +                       if (--domain->iommu_count == 0)
> +                               vm_domain_exit(domain);
> +               }
> +               else
> +                       domain_exit(domain);
> +
>                 i = find_next_bit(iommu->domain_ids,
>                         cap_ndoms(iommu->cap), i+1);
>         }
> @@ -1006,8 +1073,11 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
> 
>         set_bit(num, iommu->domain_ids);
>         domain->id = num;
> -       domain->iommu = iommu;
> +       memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
> +       set_bit(iommu->seq_id, &domain->iommu_bmp);
>         iommu->domains[num] = domain;
> +       domain->iommu_count = 1;
> +       domain->flags = 0;
>         spin_unlock_irqrestore(&iommu->lock, flags);
> 
>         return domain;
> @@ -1016,10 +1086,12 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
>  static void iommu_free_domain(struct dmar_domain *domain)
>  {
>         unsigned long flags;
> +       struct intel_iommu *iommu;
> 
> -       spin_lock_irqsave(&domain->iommu->lock, flags);
> -       clear_bit(domain->id, domain->iommu->domain_ids);
> -       spin_unlock_irqrestore(&domain->iommu->lock, flags);
> +       iommu = domain_get_only_iommu(domain);
> +       spin_lock_irqsave(&iommu->lock, flags);
> +       clear_bit(domain->id, iommu->domain_ids);
> +       spin_unlock_irqrestore(&iommu->lock, flags);
>  }
> 
>  static struct iova_domain reserved_iova_list;
> @@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
>         domain_reserve_special_ranges(domain);
> 
>         /* calculate AGAW */
> -       iommu = domain->iommu;
> +       iommu = domain_get_only_iommu(domain);
>         if (guest_width > cap_mgaw(iommu->cap))
>                 guest_width = cap_mgaw(iommu->cap);
>         domain->gaw = guest_width;
> @@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
>         sagaw = cap_sagaw(iommu->cap);
>         if (!test_bit(agaw, &sagaw)) {
>                 /* hardware doesn't support it, choose a bigger one */
> -               pr_debug("IOMMU: hardware doesn't support agaw %d\n", agaw);
> +               pr_debug("IOMMU: hardware doesn't support agaw %d\n",
> +                 agaw);
>                 agaw = find_next_bit(&sagaw, 5, agaw);
>                 if (agaw >= 5)
>                         return -ENODEV;
>         }
>         domain->agaw = agaw;
> +
>         INIT_LIST_HEAD(&domain->devices);
> 
>         /* always allocate the top pgd */
>         domain->pgd = (struct dma_pte *)alloc_pgtable_page();
>         if (!domain->pgd)
>                 return -ENOMEM;
> -       __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
> +       domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
>         return 0;
>  }
> 
> @@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain *domain)
>  }
> 
>  static int domain_context_mapping_one(struct dmar_domain *domain,
> -               u8 bus, u8 devfn)
> +               struct intel_iommu *iommu, u8 bus, u8 devfn)
>  {
>         struct context_entry *context;
> -       struct intel_iommu *iommu = domain->iommu;
>         unsigned long flags;
> 
>         pr_debug("Set context mapping for %02x:%02x.%d\n",
> @@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
>  {
>         int ret;
>         struct pci_dev *tmp, *parent;
> +       struct intel_iommu *iommu;
> +
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return  -ENODEV;
> 
> -       ret = domain_context_mapping_one(domain, pdev->bus->number,
> -               pdev->devfn);
> +       ret = domain_context_mapping_one(domain, iommu,
> +               pdev->bus->number, pdev->devfn);
>         if (ret)
>                 return ret;
> 
> @@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
>         /* Secondary interface's bus number and devfn 0 */
>         parent = pdev->bus->self;
>         while (parent != tmp) {
> -               ret = domain_context_mapping_one(domain, parent->bus->number,
> -                       parent->devfn);
> +               ret = domain_context_mapping_one(domain, iommu,
> +                       parent->bus->number, parent->devfn);
>                 if (ret)
>                         return ret;
>                 parent = parent->bus->self;
>         }
>         if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
> -               return domain_context_mapping_one(domain,
> +               return domain_context_mapping_one(domain, iommu,
>                         tmp->subordinate->number, 0);
>         else /* this is a legacy PCI bridge */
> -               return domain_context_mapping_one(domain,
> +               return domain_context_mapping_one(domain, iommu,
>                         tmp->bus->number, tmp->devfn);
>  }
> 
> -static int domain_context_mapped(struct dmar_domain *domain,
> -       struct pci_dev *pdev)
> +static int domain_context_mapped(struct pci_dev *pdev)
>  {
>         int ret;
>         struct pci_dev *tmp, *parent;
> +       struct intel_iommu *iommu;
> 
> -       ret = device_context_mapped(domain->iommu,
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return 0;
> +
> +       ret = device_context_mapped(iommu,
>                 pdev->bus->number, pdev->devfn);
>         if (!ret)
>                 return ret;
> @@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct dmar_domain *domain,
>         /* Secondary interface's bus number and devfn 0 */
>         parent = pdev->bus->self;
>         while (parent != tmp) {
> -               ret = device_context_mapped(domain->iommu, parent->bus->number,
> +               ret = device_context_mapped(iommu, parent->bus->number,
>                         parent->devfn);
>                 if (!ret)
>                         return ret;
>                 parent = parent->bus->self;
>         }
>         if (tmp->is_pcie)
> -               return device_context_mapped(domain->iommu,
> +               return device_context_mapped(iommu,
>                         tmp->subordinate->number, 0);
>         else
> -               return device_context_mapped(domain->iommu,
> +               return device_context_mapped(iommu,
>                         tmp->bus->number, tmp->devfn);
>  }
> 
> @@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
>                 BUG_ON(dma_pte_addr(*pte));
>                 dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
>                 dma_set_pte_prot(*pte, prot);
> -               __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
> +               domain_flush_cache(domain, pte, sizeof(*pte));
>                 start_pfn++;
>                 index++;
>         }
>         return 0;
>  }
> 
> -static void detach_domain_for_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
> +static void detach_domain_for_dev(struct dmar_domain *domain,
> +                                 u8 bus, u8 devfn)
>  {
> -       clear_context_table(domain->iommu, bus, devfn);
> -       domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
> -                                          DMA_CCMD_GLOBAL_INVL, 0);
> -       domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
> -                                        DMA_TLB_GLOBAL_FLUSH, 0);
> +       struct intel_iommu *iommu;
> +
> +       iommu = device_find_matched_iommu(bus, devfn);
> +       if (!iommu)
> +               return;
> +
> +       clear_context_table(iommu, bus, devfn);
> +       iommu->flush.flush_context(iommu, 0, 0, 0,
> +                                  DMA_CCMD_GLOBAL_INVL, 0);
> +       iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> +                                DMA_TLB_GLOBAL_FLUSH, 0);
>  }
> 
>  static void domain_remove_dev_info(struct dmar_domain *domain)
> @@ -1336,7 +1425,6 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
>  {
>         struct dmar_domain *domain, *found = NULL;
>         struct intel_iommu *iommu;
> -       struct dmar_drhd_unit *drhd;
>         struct device_domain_info *info, *tmp;
>         struct pci_dev *dev_tmp;
>         unsigned long flags;
> @@ -1371,13 +1459,9 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
>         }
> 
>         /* Allocate new domain for the device */
> -       drhd = dmar_find_matched_drhd_unit(pdev);
> -       if (!drhd) {
> -               printk(KERN_ERR "IOMMU: can't find DMAR for device %s\n",
> -                       pci_name(pdev));
> -               return NULL;
> -       }
> -       iommu = drhd->iommu;
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return  NULL;
> 
>         domain = iommu_alloc_domain(iommu);
>         if (!domain)
> @@ -1400,7 +1484,7 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
>                 info->dev = NULL;
>                 info->domain = domain;
>                 /* This domain is shared by devices under p2p bridge */
> -               domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
> +               domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
> 
>                 /* pcie-to-pci bridge already has a domain, uses it */
>                 found = NULL;
> @@ -1805,7 +1889,7 @@ get_valid_domain_for_dev(struct pci_dev *pdev)
>         }
> 
>         /* make sure context mapping is ok */
> -       if (unlikely(!domain_context_mapped(domain, pdev))) {
> +       if (unlikely(!domain_context_mapped(pdev))) {
>                 ret = domain_context_mapping(domain, pdev);
>                 if (ret) {
>                         printk(KERN_ERR
> @@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
>  {
>         struct pci_dev *pdev = to_pci_dev(hwdev);
>         struct dmar_domain *domain;
> +       struct intel_iommu *iommu;
>         phys_addr_t start_paddr;
>         struct iova *iova;
>         int prot = 0;
> @@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
>         if (!domain)
>                 return 0;
> 
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return 0;
> +
>         size = aligned_size((u64)paddr, size);
> 
>         iova = __intel_alloc_iova(hwdev, domain, size, pdev->dma_mask);
> @@ -1849,7 +1938,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
>          * mappings..
>          */
>         if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
> -                       !cap_zlr(domain->iommu->cap))
> +                       !cap_zlr(iommu->cap))
>                 prot |= DMA_PTE_READ;
>         if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
>                 prot |= DMA_PTE_WRITE;
> @@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
>                 goto error;
> 
>         /* it's a non-present to present mapping */
> -       ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
> +       ret = iommu_flush_iotlb_psi(iommu, domain->id,
>                         start_paddr, size >> VTD_PAGE_SHIFT, 1);
>         if (ret)
> -               iommu_flush_write_buffer(domain->iommu);
> +               iommu_flush_write_buffer(iommu);
> 
>         return start_paddr + ((u64)paddr & (~PAGE_MASK));
> 
> @@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
>         /* just flush them all */
>         for (i = 0; i < g_num_of_iommus; i++) {
>                 if (deferred_flush[i].next) {
> -                       struct intel_iommu *iommu =
> -                               deferred_flush[i].domain[0]->iommu;
> +                       struct intel_iommu *iommu = deferred_flush[i].iommu;
> 
>                         iommu->flush.flush_iotlb(iommu, 0, 0, 0,
>                                                  DMA_TLB_GLOBAL_FLUSH, 0);
> @@ -1921,7 +2009,8 @@ static void flush_unmaps_timeout(unsigned long data)
>         spin_unlock_irqrestore(&async_umap_flush_lock, flags);
>  }
> 
> -static void add_unmap(struct dmar_domain *dom, struct iova *iova)
> +static void add_unmap(struct dmar_domain *dom,
> +                     struct intel_iommu *iommu, struct iova *iova)
>  {
>         unsigned long flags;
>         int next, iommu_id;
> @@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova)
>         if (list_size == HIGH_WATER_MARK)
>                 flush_unmaps();
> 
> -       iommu_id = dom->iommu->seq_id;
> +       iommu_id = iommu->seq_id;
> 
>         next = deferred_flush[iommu_id].next;
>         deferred_flush[iommu_id].domain[next] = dom;
>         deferred_flush[iommu_id].iova[next] = iova;
> +       deferred_flush[iommu_id].iommu = iommu;
>         deferred_flush[iommu_id].next++;
> 
>         if (!timer_on) {
> @@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
>         struct dmar_domain *domain;
>         unsigned long start_addr;
>         struct iova *iova;
> +       struct intel_iommu *iommu;
> 
>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
>                 return;
>         domain = find_domain(pdev);
>         BUG_ON(!domain);
> 
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return;
> +
>         iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
>         if (!iova)
>                 return;
> @@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
>         /* free page tables */
>         dma_pte_free_pagetable(domain, start_addr, start_addr + size);
>         if (intel_iommu_strict) {
> -               if (iommu_flush_iotlb_psi(domain->iommu,
> +               if (iommu_flush_iotlb_psi(iommu,
>                         domain->id, start_addr, size >> VTD_PAGE_SHIFT, 0))
> -                       iommu_flush_write_buffer(domain->iommu);
> +                       iommu_flush_write_buffer(iommu);
>                 /* free iova */
>                 __free_iova(&domain->iovad, iova);
>         } else {
> -               add_unmap(domain, iova);
> +               add_unmap(domain, iommu, iova);
>                 /*
>                  * queue up the release of the unmap to save the 1/6th of the
>                  * cpu used up by the iotlb flush operation...
> @@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
>         size_t size = 0;
>         void *addr;
>         struct scatterlist *sg;
> +       struct intel_iommu *iommu;
> 
>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
>                 return;
> 
>         domain = find_domain(pdev);
> 
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return;
> +
>         iova = find_iova(&domain->iovad, IOVA_PFN(sglist[0].dma_address));
>         if (!iova)
>                 return;
> @@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
>         /* free page tables */
>         dma_pte_free_pagetable(domain, start_addr, start_addr + size);
> 
> -       if (iommu_flush_iotlb_psi(domain->iommu, domain->id, start_addr,
> +       if (iommu_flush_iotlb_psi(iommu, domain->id, start_addr,
>                         size >> VTD_PAGE_SHIFT, 0))
> -               iommu_flush_write_buffer(domain->iommu);
> +               iommu_flush_write_buffer(iommu);
> 
>         /* free iova */
>         __free_iova(&domain->iovad, iova);
> @@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
>         int ret;
>         struct scatterlist *sg;
>         unsigned long start_addr;
> +       struct intel_iommu *iommu;
> 
>         BUG_ON(dir == DMA_NONE);
>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> @@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
>         if (!domain)
>                 return 0;
> 
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return 0;
> +
>         for_each_sg(sglist, sg, nelems, i) {
>                 addr = SG_ENT_VIRT_ADDRESS(sg);
>                 addr = (void *)virt_to_phys(addr);
> @@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
>          * mappings..
>          */
>         if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
> -                       !cap_zlr(domain->iommu->cap))
> +                       !cap_zlr(iommu->cap))
>                 prot |= DMA_PTE_READ;
>         if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
>                 prot |= DMA_PTE_WRITE;
> @@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
>         }
> 
>         /* it's a non-present to present mapping */
> -       if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
> +       if (iommu_flush_iotlb_psi(iommu, domain->id,
>                         start_addr, offset >> VTD_PAGE_SHIFT, 1))
> -               iommu_flush_write_buffer(domain->iommu);
> +               iommu_flush_write_buffer(iommu);
>         return nelems;
>  }
> 
> @@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
>         return 0;
>  }
> 
> -void intel_iommu_domain_exit(struct dmar_domain *domain)
> +/* domain id for virtual machine, it won't be set in context */
> +static unsigned long vm_domid;
> +
> +static int vm_domain_min_agaw(struct dmar_domain *domain)
> +{
> +       struct dmar_drhd_unit *drhd;
> +       struct intel_iommu *iommu;
> +       int min_agaw = domain->agaw;
> +
> +       for_each_drhd_unit(drhd) {
> +               if (drhd->ignored)
> +                       continue;
> +               iommu = drhd->iommu;
> +
> +               if (test_bit(iommu->seq_id, &domain->iommu_bmp))
> +                       if (min_agaw > iommu->agaw)
> +                               min_agaw = iommu->agaw;
> +       }
> +
> +       return min_agaw;
> +}
> +
> +static int vm_domain_add_dev_info(struct dmar_domain *domain,
> +                                 struct pci_dev *pdev)
> +{
> +       struct device_domain_info *info;
> +       unsigned long flags;
> +
> +       info = alloc_devinfo_mem();
> +       if (!info)
> +               return -ENOMEM;
> +
> +       info->bus = pdev->bus->number;
> +       info->devfn = pdev->devfn;
> +       info->dev = pdev;
> +       info->domain = domain;
> +
> +       spin_lock_irqsave(&device_domain_lock, flags);
> +       list_add(&info->link, &domain->devices);
> +       list_add(&info->global, &device_domain_list);
> +       pdev->dev.archdata.iommu = info;
> +       spin_unlock_irqrestore(&device_domain_lock, flags);
> +
> +       return 0;
> +}
> +
> +static void vm_domain_remove_one_dev_info(struct dmar_domain *domain,
> +                                         struct pci_dev *pdev)
> +{
> +       struct device_domain_info *info;
> +       struct intel_iommu *iommu;
> +       unsigned long flags;
> +       int found = 0;
> +
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +
> +       spin_lock_irqsave(&device_domain_lock, flags);
> +       while (!list_empty(&domain->devices)) {
> +               info = list_entry(domain->devices.next,
> +                       struct device_domain_info, link);
> +               if (info->bus == pdev->bus->number &&
> +                   info->devfn == pdev->devfn) {
> +                       list_del(&info->link);
> +                       list_del(&info->global);
> +                       if (info->dev)
> +                               info->dev->dev.archdata.iommu = NULL;
> +                       spin_unlock_irqrestore(&device_domain_lock, flags);
> +
> +                       detach_domain_for_dev(info->domain,
> +                                             info->bus, info->devfn);
> +                       free_devinfo_mem(info);
> +
> +                       spin_lock_irqsave(&device_domain_lock, flags);
> +
> +                       if (found)
> +                               break;
> +                       else
> +                               continue;
> +               }
> +
> +               /* if there is no other devices under the same iommu
> +                * owned by this domain, clear this iommu in iommu_bmp
> +                */
> +               if (device_find_matched_iommu(info->bus, info->devfn) == iommu)
> +                       found = 1;
> +       }
> +
> +       if (found == 0) {
> +               spin_lock_irqsave(&iommu->lock, flags);
> +               clear_bit(iommu->seq_id, &domain->iommu_bmp);
> +               domain->iommu_count--;
> +               spin_unlock_irqrestore(&iommu->lock, flags);
> +       }
> +
> +       spin_unlock_irqrestore(&device_domain_lock, flags);
> +}
> +
> +static void vm_domain_remove_all_dev_info(struct dmar_domain *domain)
> +{
> +       struct device_domain_info *info;
> +       struct intel_iommu *iommu;
> +       unsigned long flags;
> +
> +       spin_lock_irqsave(&device_domain_lock, flags);
> +       while (!list_empty(&domain->devices)) {
> +               info = list_entry(domain->devices.next,
> +                       struct device_domain_info, link);
> +               list_del(&info->link);
> +               list_del(&info->global);
> +               if (info->dev)
> +                       info->dev->dev.archdata.iommu = NULL;
> +
> +               spin_unlock_irqrestore(&device_domain_lock, flags);
> +               detach_domain_for_dev(info->domain,
> +                                     info->bus, info->devfn);
> +
> +               /* clear this iommu in iommu_bmp */
> +               iommu = device_find_matched_iommu(info->bus, info->devfn);
> +               spin_lock_irqsave(&iommu->lock, flags);
> +               if (test_and_clear_bit(iommu->seq_id,
> +                                      &domain->iommu_bmp))
> +                       domain->iommu_count--;
> +               spin_unlock_irqrestore(&iommu->lock, flags);
> +
> +               free_devinfo_mem(info);
> +               spin_lock_irqsave(&device_domain_lock, flags);
> +       }
> +       spin_unlock_irqrestore(&device_domain_lock, flags);
> +}
> +
> +static int vm_domain_context_mapping_one(struct dmar_domain *domain,
> +               struct intel_iommu *iommu, u8 bus, u8 devfn)
> +{
> +       struct context_entry *context;
> +       unsigned long flags;
> +       struct dma_pte *pgd;
> +       unsigned long num;
> +       unsigned long ndomains;
> +       int id;
> +       int agaw;
> +       int found = 0;
> +
> +       pr_debug("Set context mapping for %02x:%02x.%d\n",
> +               bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
> +       BUG_ON(!domain->pgd);
> +       context = device_to_context_entry(iommu, bus, devfn);
> +       if (!context)
> +               return -ENOMEM;
> +       spin_lock_irqsave(&iommu->lock, flags);
> +       if (context_present(*context)) {
> +               spin_unlock_irqrestore(&iommu->lock, flags);
> +               return 0;
> +       }
> +
> +       id = domain->id;
> +
> +       /* find an available domain id for this device in iommu */
> +       ndomains = cap_ndoms(iommu->cap);
> +       num = find_first_bit(iommu->domain_ids, ndomains);
> +       for (; num < ndomains; ) {
> +               if (iommu->domains[num] == domain) {
> +                       id = num;
> +                       found = 1;
> +                       break;
> +               }
> +               num = find_next_bit(iommu->domain_ids,
> +                 cap_ndoms(iommu->cap), num+1);
> +       }
> +
> +       if (found == 0) {
> +               num = find_first_zero_bit(iommu->domain_ids, ndomains);
> +               if (num >= ndomains) {
> +                       spin_unlock_irqrestore(&iommu->lock, flags);
> +                       printk(KERN_ERR "IOMMU: no free domain ids\n");
> +                       return -EFAULT;
> +               }
> +
> +               set_bit(num, iommu->domain_ids);
> +               iommu->domains[num] = domain;
> +               id = num;
> +       }
> +
> +       pgd = domain->pgd;
> +
> +       /* Skip top levels of page tables for
> +        * iommu which has less agaw than default.
> +        */
> +       for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
> +               pgd = phys_to_virt(dma_pte_addr(*pgd));
> +               if (!dma_pte_present(*pgd)) {
> +                       spin_unlock_irqrestore(&iommu->lock, flags);
> +                       return -ENOMEM;
> +               }
> +       }
> +
> +       context_set_domain_id(*context, id);
> +       context_set_address_width(*context, iommu->agaw);
> +       context_set_address_root(*context, virt_to_phys(pgd));
> +       context_set_translation_type(*context, CONTEXT_TT_MULTI_LEVEL);
> +       context_set_fault_enable(*context);
> +       context_set_present(*context);
> +       __iommu_flush_cache(iommu, context, sizeof(*context));
> +
> +       /* it's a non-present to present mapping */
> +       if (iommu->flush.flush_context(iommu, id,
> +               (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
> +               DMA_CCMD_DEVICE_INVL, 1))
> +               iommu_flush_write_buffer(iommu);
> +       else
> +               iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_DSI_FLUSH, 0);
> +
> +       if (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp))
> +               domain->iommu_count++;
> +       spin_unlock_irqrestore(&iommu->lock, flags);
> +       return 0;
> +}
> +
> +static int
> +vm_domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
> +{
> +       int ret;
> +       struct pci_dev *tmp, *parent;
> +       struct intel_iommu *iommu;
> +
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return  -ENODEV;
> +
> +       ret = vm_domain_context_mapping_one(domain, iommu,
> +               pdev->bus->number, pdev->devfn);
> +       if (ret)
> +               return ret;
> +
> +       /* dependent device mapping */
> +       tmp = pci_find_upstream_pcie_bridge(pdev);
> +       if (!tmp)
> +               return 0;
> +       /* Secondary interface's bus number and devfn 0 */
> +       parent = pdev->bus->self;
> +       while (parent != tmp) {
> +               ret = vm_domain_context_mapping_one(domain, iommu,
> +                       parent->bus->number, parent->devfn);
> +               if (ret)
> +                       return ret;
> +               parent = parent->bus->self;
> +       }
> +       if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
> +               return vm_domain_context_mapping_one(domain, iommu,
> +                       tmp->subordinate->number, 0);
> +       else /* this is a legacy PCI bridge */
> +               return vm_domain_context_mapping_one(domain, iommu,
> +                       tmp->bus->number, tmp->devfn);
> +}
> +
> +
> +static int vm_domain_init(struct dmar_domain *domain, int guest_width)
> +{
> +       int adjust_width;
> +
> +       init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
> +       spin_lock_init(&domain->mapping_lock);
> +
> +       domain_reserve_special_ranges(domain);
> +
> +       /* calculate AGAW */
> +       domain->gaw = guest_width;
> +       adjust_width = guestwidth_to_adjustwidth(guest_width);
> +       domain->agaw = width_to_agaw(adjust_width);
> +
> +       INIT_LIST_HEAD(&domain->devices);
> +
> +       /* always allocate the top pgd */
> +       domain->pgd = (struct dma_pte *)alloc_pgtable_page();
> +       if (!domain->pgd)
> +               return -ENOMEM;
> +       domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
> +       return 0;
> +}
> +
> +static void iommu_free_vm_domain(struct dmar_domain *domain)
> +{
> +       unsigned long flags;
> +       struct dmar_drhd_unit *drhd;
> +       struct intel_iommu *iommu;
> +       unsigned long i;
> +       unsigned long ndomains;
> +
> +       for_each_drhd_unit(drhd) {
> +               if (drhd->ignored)
> +                       continue;
> +               iommu = drhd->iommu;
> +
> +               ndomains = cap_ndoms(iommu->cap);
> +               i = find_first_bit(iommu->domain_ids, ndomains);
> +               for (; i < ndomains; ) {
> +                       if (iommu->domains[i] == domain) {
> +                               spin_lock_irqsave(&iommu->lock, flags);
> +                               clear_bit(i, iommu->domain_ids);
> +                               iommu->domains[i] = NULL;
> +                               spin_unlock_irqrestore(&iommu->lock, flags);
> +                               break;
> +                       }
> +                       i = find_next_bit(iommu->domain_ids, ndomains, i+1);
> +               }
> +       }
> +}
> +
> +static void vm_domain_exit(struct dmar_domain *domain)
>  {
>         u64 end;
> 
> @@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
>         if (!domain)
>                 return;
> 
> +       vm_domain_remove_all_dev_info(domain);
> +       /* destroy iovas */
> +       put_iova_domain(&domain->iovad);
>         end = DOMAIN_MAX_ADDR(domain->gaw);
> -       end = end & (~VTD_PAGE_MASK);
> +       end &= VTD_PAGE_MASK;
> 
>         /* clear ptes */
>         dma_pte_clear_range(domain, 0, end);
> @@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
>         /* free page tables */
>         dma_pte_free_pagetable(domain, 0, end);
> 
> -       iommu_free_domain(domain);
> +       iommu_free_vm_domain(domain);
>         free_domain_mem(domain);
>  }
> -EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
> 
> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
> +static struct dmar_domain *iommu_alloc_vm_domain(void)
>  {
> -       struct dmar_drhd_unit *drhd;
>         struct dmar_domain *domain;
> -       struct intel_iommu *iommu;
> 
> -       drhd = dmar_find_matched_drhd_unit(pdev);
> -       if (!drhd) {
> -               printk(KERN_ERR "intel_iommu_domain_alloc: drhd == NULL\n");
> +       domain = alloc_domain_mem();
> +       if (!domain)
>                 return NULL;
> -       }
> 
> -       iommu = drhd->iommu;
> -       if (!iommu) {
> -               printk(KERN_ERR
> -                       "intel_iommu_domain_alloc: iommu == NULL\n");
> -               return NULL;
> -       }
> -       domain = iommu_alloc_domain(iommu);
> +       domain->id = vm_domid++;
> +       domain->iommu_count = 0;
> +       domain->max_addr = 0;
> +       memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
> +       domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE;
> +
> +       return domain;
> +}
> +
> +struct dmar_domain *intel_iommu_alloc_domain(void)
> +{
> +       struct dmar_domain *domain;
> +
> +       domain = iommu_alloc_vm_domain();
>         if (!domain) {
>                 printk(KERN_ERR
>                         "intel_iommu_domain_alloc: domain == NULL\n");
>                 return NULL;
>         }
> -       if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
> +       if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
>                 printk(KERN_ERR
>                         "intel_iommu_domain_alloc: domain_init() failed\n");
> -               intel_iommu_domain_exit(domain);
> +               vm_domain_exit(domain);
>                 return NULL;
>         }
> +
>         return domain;
>  }
> -EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
> +EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
> +
> +void intel_iommu_free_domain(struct dmar_domain *domain)
> +{
> +       vm_domain_exit(domain);
> +}
> +EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
> 
> -int intel_iommu_context_mapping(
> -       struct dmar_domain *domain, struct pci_dev *pdev)
> +int intel_iommu_assign_device(struct dmar_domain *domain,
> +                             struct pci_dev *pdev)
>  {
> -       int rc;
> -       rc = domain_context_mapping(domain, pdev);
> -       return rc;
> +       struct intel_iommu *iommu;
> +       int addr_width;
> +       u64 end;
> +       int ret;
> +
> +       /* normally pdev is not mapped */
> +       if (unlikely(domain_context_mapped(pdev))) {
> +               struct dmar_domain *old_domain;
> +
> +               old_domain = find_domain(pdev);
> +               if (old_domain) {
> +                       if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE)
> +                               vm_domain_remove_one_dev_info(old_domain, pdev);
> +                       else
> +                               domain_remove_dev_info(old_domain);
> +               }
> +       }
> +
> +       iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +       if (!iommu)
> +               return  -ENODEV;
> +
> +       /* check if this iommu agaw is sufficient for max mapped address */
> +       addr_width = agaw_to_width(iommu->agaw);
> +       end = DOMAIN_MAX_ADDR(addr_width);
> +       end = end & VTD_PAGE_MASK;
> +       if (end < domain->max_addr) {
> +               printk(KERN_ERR "%s: iommu agaw (%d) is not "
> +                      "sufficient for the mapped address (%llx)\n",
> +                      __func__, iommu->agaw, domain->max_addr);
> +               return -EFAULT;
> +       }
> +
> +       ret = vm_domain_context_mapping(domain, pdev);
> +       if (ret)
> +               return ret;
> +
> +       ret = vm_domain_add_dev_info(domain, pdev);
> +       return ret;
>  }
> -EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
> +EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
> 
> -int intel_iommu_page_mapping(
> -       struct dmar_domain *domain, dma_addr_t iova,
> -       u64 hpa, size_t size, int prot)
> +
> +void intel_iommu_deassign_device(struct dmar_domain *domain,
> +                                struct pci_dev *pdev)
>  {
> -       int rc;
> -       rc = domain_page_mapping(domain, iova, hpa, size, prot);
> -       return rc;
> +       vm_domain_remove_one_dev_info(domain, pdev);
>  }
> -EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
> +EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
> 
> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
> +int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
> +                         u64 hpa, size_t size, int prot)
>  {
> -       detach_domain_for_dev(domain, bus, devfn);
> +       u64 max_addr;
> +       int addr_width;
> +       int ret;
> +
> +       max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
> +       if (domain->max_addr < max_addr) {
> +               int min_agaw;
> +               u64 end;
> +
> +               /* check if minimum agaw is sufficient for mapped address */
> +               min_agaw = vm_domain_min_agaw(domain);
> +               addr_width = agaw_to_width(min_agaw);
> +               end = DOMAIN_MAX_ADDR(addr_width);
> +               end = end & VTD_PAGE_MASK;
> +               if (end < max_addr) {
> +                       printk(KERN_ERR "%s: iommu agaw (%d) is not "
> +                              "sufficient for the mapped address (%llx)\n",
> +                              __func__, min_agaw, max_addr);
> +                       return -EFAULT;
> +               }
> +               domain->max_addr = max_addr;
> +       }
> +
> +       ret = domain_page_mapping(domain, iova, hpa, size, prot);
> +       return ret;
>  }
> -EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
> +EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
> 
> -struct dmar_domain *
> -intel_iommu_find_domain(struct pci_dev *pdev)
> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
> +                            dma_addr_t iova, size_t size)
>  {
> -       return find_domain(pdev);
> +       dma_addr_t base;
> +
> +       /* The address might not be aligned */
> +       base = iova & PAGE_MASK;
> +       size = PAGE_ALIGN(size);
> +       dma_pte_clear_range(domain, base, base + size);
>  }
> -EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
> +EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
> 
>  int intel_iommu_found(void)
>  {
> diff --git a/include/linux/dma_remapping.h b/include/linux/dma_remapping.h
> index 952df39..3568ae6 100644
> --- a/include/linux/dma_remapping.h
> +++ b/include/linux/dma_remapping.h
> @@ -111,11 +111,21 @@ struct dma_pte {
>                 (p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
>  #define dma_pte_present(p) (((p).val & 3) != 0)
> 
> +/* domain flags, one domain owns one device by default */
> +
> +/* devices under the same p2p bridge are owned in one domain */
> +#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES       (1 << 0)
> +
> +/* domain represents a virtual machine, more than one devices
> + * across iommus may be owned in one domain, e.g. kvm guest.
> + */
> +#define DOMAIN_FLAG_VIRTUAL_MACHINE            (1 << 1)
> +
>  struct intel_iommu;
> 
>  struct dmar_domain {
>         int     id;                     /* domain id */
> -       struct intel_iommu *iommu;      /* back pointer to owning iommu */
> +       unsigned long iommu_bmp;        /* bitmap of iommus this domain uses*/
> 
>         struct list_head devices;       /* all devices' list */
>         struct iova_domain iovad;       /* iova's that belong to this domain */
> @@ -123,12 +133,13 @@ struct dmar_domain {
>         struct dma_pte  *pgd;           /* virtual address */
>         spinlock_t      mapping_lock;   /* page table lock */
>         int             gaw;            /* max guest address width */
> +       int             agaw;           /* adjusted guest address width */
> 
> -       /* adjusted guest address width, 0 is level 2 30-bit */
> -       int             agaw;
> +       int             flags;          /* domain flag */
> 
> -#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
> -       int             flags;
> +       /* following fields are used in virtual machine case */
> +       int             iommu_count;    /* reference count of iommu */
> +       u64             max_addr;       /* maximum mapped address */
>  };
> 
>  /* PCI domain-device relationship */
> diff --git a/include/linux/intel-iommu.h b/include/linux/intel-iommu.h
> index 3d017cf..c2f37b8 100644
> --- a/include/linux/intel-iommu.h
> +++ b/include/linux/intel-iommu.h
> @@ -219,6 +219,8 @@ do {                                                                        \
>         }                                                               \
>  } while (0)
> 
> +#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
> +
>  #define QI_LENGTH      256     /* queue length */
> 
>  enum {
> @@ -299,6 +301,7 @@ struct intel_iommu {
>         struct dmar_domain **domains; /* ptr to domains */
>         spinlock_t      lock; /* protect context, domain ids */
>         struct root_entry *root_entry; /* virtual address */
> +       int             agaw;
> 
>         unsigned int irq;
>         unsigned char name[7];    /* Device Name */
> @@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu *iommu, u16 did, u64 addr,
> 
>  extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu *iommu);
> 
> -void intel_iommu_domain_exit(struct dmar_domain *domain);
> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
> -int intel_iommu_context_mapping(struct dmar_domain *domain,
> -                               struct pci_dev *pdev);
> -int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
> -                            u64 hpa, size_t size, int prot);
> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn);
> -struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
> +struct dmar_domain *intel_iommu_alloc_domain(void);
> +void intel_iommu_free_domain(struct dmar_domain *domain);
> +int intel_iommu_assign_device(struct dmar_domain *domain,
> +                             struct pci_dev *pdev);
> +void intel_iommu_deassign_device(struct dmar_domain *domain,
> +                                struct pci_dev *pdev);
> +int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
> +                         u64 hpa, size_t size, int prot);
> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
> +                            dma_addr_t iova, size_t size);
>  u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
> 
>  #ifdef CONFIG_DMAR
> --
> 1.5.1



-- 
           |           AMD Saxony Limited Liability Company & Co. KG
 Operating |         Wilschdorfer Landstr. 101, 01109 Dresden, Germany
 System    |                  Register Court Dresden: HRA 4896
 Research  |              General Partner authorized to represent:
 Center    |             AMD Saxony LLC (Wilmington, Delaware, US)
           | General Manager of AMD Saxony LLC: Dr. Hans-R. Deppe, Thomas McCoy


^ permalink raw reply	[flat|nested] 7+ messages in thread

* RE: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
  2008-11-28 13:27 ` Joerg Roedel
@ 2008-12-01  6:17   ` Han, Weidong
  2008-12-01 11:55     ` 'Joerg Roedel'
  0 siblings, 1 reply; 7+ messages in thread
From: Han, Weidong @ 2008-12-01  6:17 UTC (permalink / raw)
  To: 'Joerg Roedel'
  Cc: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity',
	Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
	'iommu@lists.linux-foundation.org'

It's developed based on commit 0f7d3ee6 on avi/master, but it still can be applied on latest avi/master (commit 90755652).

Regards,
Weidong

Joerg Roedel wrote:
> Hmm, I tried to apply this patch against avi/master and linus/master
> but get merge conflicts. Where do these patches apply cleanly?
>
> Joerg
>
> On Thu, Nov 27, 2008 at 09:49:04PM +0800, Han, Weidong wrote:
>> In order to support multiple device assignment for KVM, this patch
>> does following main changes:
>>    - extend dmar_domain to own multiple devices from different
>> iommus, use a bitmap of iommus to replace iommu pointer in
>> dmar_domain.
>>    - implement independent low level functions for kvm, then won't
>> impact native VT-d.
>>    - "SAGAW" capability may be different across iommus, that's to
>> say the VT-d page table levels may be different among iommus. This
>> patch uses a defaut agaw, and skip top levels of page tables for
>> iommus which have smaller agaw than default.
>>    - rename the APIs for kvm VT-d, make it more readable.
>>
>>
>> Signed-off-by: Weidong Han <weidong.han@intel.com>
>> ---
>>  drivers/pci/dmar.c            |   15 +
>>  drivers/pci/intel-iommu.c     |  698
>>  ++++++++++++++++++++++++++++++++++------
>>  include/linux/dma_remapping.h |   21 +- include/linux/intel-iommu.h
>>  |   21 +- 4 files changed, 637 insertions(+), 118 deletions(-)
>>
>> diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
>> index 691b3ad..d6bdced 100644
>> --- a/drivers/pci/dmar.c
>> +++ b/drivers/pci/dmar.c
>> @@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
>>  dmar_tbl = NULL; }
>>
>> +extern int width_to_agaw(int width);
>>
>>  int alloc_iommu(struct dmar_drhd_unit *drhd)
>>  {
>> @@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
>>         int map_size; u32 ver;
>>         static int iommu_allocated = 0;
>> +       unsigned long sagaw;
>> +       int agaw;
>>
>>         iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);         if
>> (!iommu) @@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit
>>         *drhd) iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
>>         iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
>>
>> +       /* set agaw, "SAGAW" may be different across iommus */
>> +       sagaw = cap_sagaw(iommu->cap);
>> +       for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
>> +            agaw >= 0; agaw--)
>> +               if (test_bit(agaw, &sagaw))
>> +                       break;
>> +       if (agaw < 0) {
>> +               printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n",
>> sagaw); +               goto error; +       }
>> +       iommu->agaw = agaw;
>> +
>>         /* the registers might be more than one page */
>>         map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
>>                 cap_max_fault_reg_offset(iommu->cap));
>> diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
>> index 5c8baa4..55b96c4 100644
>> --- a/drivers/pci/intel-iommu.c
>> +++ b/drivers/pci/intel-iommu.c
>> @@ -50,8 +50,6 @@
>>  #define IOAPIC_RANGE_END       (0xfeefffff)
>>  #define IOVA_START_ADDR                (0x1000)
>>
>> -#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
>> -
>>  #define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
>>
>>
>> @@ -64,6 +62,7 @@ struct deferred_flush_tables {
>>         int next;
>>         struct iova *iova[HIGH_WATER_MARK];
>>         struct dmar_domain *domain[HIGH_WATER_MARK];
>> +       struct intel_iommu *iommu;
>>  };
>>
>>  static struct deferred_flush_tables *deferred_flush;
>> @@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
>>         kmem_cache_free(iommu_iova_cache, iova);
>>  }
>>
>> +/* in native case, each domain is related to only one iommu */
>> +static struct intel_iommu *domain_get_only_iommu(struct dmar_domain
>> *domain) +{ +       struct dmar_drhd_unit *drhd;
>> +
>> +       for_each_drhd_unit(drhd) {
>> +               if (drhd->ignored)
>> +                       continue;
>> +               if (test_bit(drhd->iommu->seq_id,
>> &domain->iommu_bmp)) +                       return drhd->iommu;
>> +       }
>> +
>> +       return NULL;
>> +}
>> +
>> +static void domain_flush_cache(struct dmar_domain *domain,
>> +                              void *addr, int size) +{
>> +       struct intel_iommu *iommu;
>> +
>> +       if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
>> +               struct dmar_drhd_unit *drhd;
>> +
>> +               for_each_drhd_unit(drhd) {
>> +                       if (drhd->ignored)
>> +                               continue;
>> +                       iommu = drhd->iommu;
>> +
>> +                       if (!test_bit(iommu->seq_id,
>> &domain->iommu_bmp)) +                               continue;
>> +
>> +                       if (!ecap_coherent(iommu->ecap))
>> +                               clflush_cache_range(addr, size); +
>> } +       }
>> +       else {
>> +               iommu = domain_get_only_iommu(domain);
>> +               if (iommu && !ecap_coherent(iommu->ecap))
>> +                       clflush_cache_range(addr, size); +       }
>> +}
>> +
>> +static struct intel_iommu *device_find_matched_iommu(u8 bus, u8
>> devfn) +{ +       struct dmar_drhd_unit *drhd = NULL;
>> +       int i;
>> +
>> +       for_each_drhd_unit(drhd) {
>> +               if (drhd->ignored)
>> +                       continue;
>> +
>> +               for (i = 0; i < drhd->devices_cnt; i++)
>> +                       if (drhd->devices[i]->bus->number == bus &&
>> +                           drhd->devices[i]->devfn == devfn)
>> +                               return drhd->iommu; +
>> +               if (drhd->include_all)
>> +                       return drhd->iommu;
>> +       }
>> +
>> +       return NULL;
>> +}
>> +
>>  /* Gets context entry for a given bus and devfn */
>>  static struct context_entry * device_to_context_entry(struct
>>                 intel_iommu *iommu, u8 bus, u8 devfn)
>> @@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
>>
>>  }
>>
>> -static inline int width_to_agaw(int width)
>> +int width_to_agaw(int width)
>>  {
>>         return (width - 30) / LEVEL_STRIDE;
>>  }
>> @@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct
>>                                         dmar_domain *domain, u64
>>                                 addr) flags); return NULL;
>>                         }
>> -                       __iommu_flush_cache(domain->iommu, tmp_page,
>> -                                       PAGE_SIZE);
>> +                       domain_flush_cache(domain, tmp_page,
>>                         PAGE_SIZE); dma_set_pte_addr(*pte,
>>                          virt_to_phys(tmp_page));
>> /* * high level table always sets r/w, last level page @@ -356,7
>>                         +417,7 @@ static struct dma_pte *
>>                         addr_to_dma_pte(struct dmar_domain *domain,
>> u64 addr)                          */ dma_set_pte_readable(*pte);
>> dma_set_pte_writable(*pte); -
>>                 __iommu_flush_cache(domain->iommu, pte,
>> sizeof(*pte)); +                       domain_flush_cache(domain,
>> pte, sizeof(*pte));                 } parent =
>> phys_to_virt(dma_pte_addr(*pte));                 level--; @@ -399,7
>> +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain,
>> u64 addr)
>>
>>         if (pte) {
>>                 dma_clear_pte(*pte);
>> -               __iommu_flush_cache(domain->iommu, pte,
>> sizeof(*pte)); +               domain_flush_cache(domain, pte,
>>  sizeof(*pte));         } }
>>
>> @@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct
>>                                 dmar_domain *domain,
>>                                         free_pgtable_page(
>>                                 phys_to_virt(dma_pte_addr(*pte)));
>> dma_clear_pte(*pte);
>> -                               __iommu_flush_cache(domain->iommu,
>> -                                               pte, sizeof(*pte));
>> +                               domain_flush_cache(domain, pte,
>>                         sizeof(*pte));                         } tmp
>>                 += level_size(level); }
>> @@ -948,8 +1008,8 @@ static int iommu_init_domains(struct
>>  intel_iommu *iommu)         return 0; }
>>
>> -
>>  static void domain_exit(struct dmar_domain *domain);
>> +static void vm_domain_exit(struct dmar_domain *domain);
>>
>>  void free_dmar_iommu(struct intel_iommu *iommu)
>>  {
>> @@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
>>         for (; i < cap_ndoms(iommu->cap); ) {
>>                 domain = iommu->domains[i];
>>                 clear_bit(i, iommu->domain_ids);
>> -               domain_exit(domain);
>> +
>> +               if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
>> +                       if (--domain->iommu_count == 0)
>> +                               vm_domain_exit(domain); +
>> } +               else
>> +                       domain_exit(domain);
>> +
>>                 i = find_next_bit(iommu->domain_ids,
>>                         cap_ndoms(iommu->cap), i+1);         }
>> @@ -1006,8 +1073,11 @@ static struct dmar_domain *
>> iommu_alloc_domain(struct intel_iommu *iommu)
>>
>>         set_bit(num, iommu->domain_ids);
>>         domain->id = num;
>> -       domain->iommu = iommu;
>> +       memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
>> +       set_bit(iommu->seq_id, &domain->iommu_bmp);
>>         iommu->domains[num] = domain;
>> +       domain->iommu_count = 1;
>> +       domain->flags = 0;
>>         spin_unlock_irqrestore(&iommu->lock, flags);
>>
>>         return domain;
>> @@ -1016,10 +1086,12 @@ static struct dmar_domain *
>>  iommu_alloc_domain(struct intel_iommu *iommu) static void
>>         iommu_free_domain(struct dmar_domain *domain)  { unsigned
>> long flags; +       struct intel_iommu *iommu;
>>
>> -       spin_lock_irqsave(&domain->iommu->lock, flags);
>> -       clear_bit(domain->id, domain->iommu->domain_ids);
>> -       spin_unlock_irqrestore(&domain->iommu->lock, flags);
>> +       iommu = domain_get_only_iommu(domain);
>> +       spin_lock_irqsave(&iommu->lock, flags);
>> +       clear_bit(domain->id, iommu->domain_ids);
>> +       spin_unlock_irqrestore(&iommu->lock, flags);  }
>>
>>  static struct iova_domain reserved_iova_list;
>> @@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain
>>         *domain, int guest_width)
>> domain_reserve_special_ranges(domain);
>>
>>         /* calculate AGAW */
>> -       iommu = domain->iommu;
>> +       iommu = domain_get_only_iommu(domain);
>>         if (guest_width > cap_mgaw(iommu->cap))
>>                 guest_width = cap_mgaw(iommu->cap);
>>         domain->gaw = guest_width;
>> @@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain
>>         *domain, int guest_width) sagaw = cap_sagaw(iommu->cap);
>>         if (!test_bit(agaw, &sagaw)) {
>>                 /* hardware doesn't support it, choose a bigger one
>> */ -               pr_debug("IOMMU: hardware doesn't support agaw
>> %d\n", agaw); +               pr_debug("IOMMU: hardware doesn't
>>                 support agaw %d\n", +                 agaw); agaw =
>>                 find_next_bit(&sagaw, 5, agaw); if (agaw >= 5)
>>                         return -ENODEV;
>>         }
>>         domain->agaw = agaw;
>> +
>>         INIT_LIST_HEAD(&domain->devices);
>>
>>         /* always allocate the top pgd */
>>         domain->pgd = (struct dma_pte *)alloc_pgtable_page();
>>                 if (!domain->pgd) return -ENOMEM;
>> -       __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
>> +       domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
>>  return 0; }
>>
>> @@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain
>> *domain)  }
>>
>>  static int domain_context_mapping_one(struct dmar_domain *domain,
>> -               u8 bus, u8 devfn)
>> +               struct intel_iommu *iommu, u8 bus, u8 devfn)  {
>>         struct context_entry *context;
>> -       struct intel_iommu *iommu = domain->iommu;
>>         unsigned long flags;
>>
>>         pr_debug("Set context mapping for %02x:%02x.%d\n",
>> @@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain
>>         *domain, struct pci_dev *pdev)  { int ret;
>>         struct pci_dev *tmp, *parent;
>> +       struct intel_iommu *iommu;
>> +
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return  -ENODEV;
>>
>> -       ret = domain_context_mapping_one(domain, pdev->bus->number,
>> -               pdev->devfn);
>> +       ret = domain_context_mapping_one(domain, iommu,
>> +               pdev->bus->number, pdev->devfn);
>>         if (ret)
>>                 return ret;
>>
>> @@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain
>>         *domain, struct pci_dev *pdev) /* Secondary interface's bus
>>         number and devfn 0 */ parent = pdev->bus->self;
>>         while (parent != tmp) {
>> -               ret = domain_context_mapping_one(domain,
>> parent->bus->number,
>> -                       parent->devfn);
>> +               ret = domain_context_mapping_one(domain, iommu,
>> +                       parent->bus->number, parent->devfn);
>>                         if (ret) return ret;
>>                 parent = parent->bus->self;
>>         }
>>         if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
>> -               return domain_context_mapping_one(domain,
>> +               return domain_context_mapping_one(domain, iommu,
>>                         tmp->subordinate->number, 0);
>>         else /* this is a legacy PCI bridge */
>> -               return domain_context_mapping_one(domain,
>> +               return domain_context_mapping_one(domain, iommu,
>>                         tmp->bus->number, tmp->devfn);  }
>>
>> -static int domain_context_mapped(struct dmar_domain *domain,
>> -       struct pci_dev *pdev)
>> +static int domain_context_mapped(struct pci_dev *pdev)  {
>>         int ret;
>>         struct pci_dev *tmp, *parent;
>> +       struct intel_iommu *iommu;
>>
>> -       ret = device_context_mapped(domain->iommu,
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return 0;
>> +
>> +       ret = device_context_mapped(iommu,
>>                 pdev->bus->number, pdev->devfn);
>>         if (!ret)
>>                 return ret;
>> @@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct
>>         dmar_domain *domain, /* Secondary interface's bus number and
>>         devfn 0 */ parent = pdev->bus->self;
>>         while (parent != tmp) {
>> -               ret = device_context_mapped(domain->iommu,
>> parent->bus->number, +               ret =
>>                         device_context_mapped(iommu,
>>                 parent->bus->number, parent->devfn); if (!ret)
>>                         return ret;
>>                 parent = parent->bus->self;
>>         }
>>         if (tmp->is_pcie)
>> -               return device_context_mapped(domain->iommu,
>> +               return device_context_mapped(iommu,
>>                         tmp->subordinate->number, 0);         else
>> -               return device_context_mapped(domain->iommu,
>> +               return device_context_mapped(iommu,
>>                         tmp->bus->number, tmp->devfn);  }
>>
>> @@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain
>>                 *domain, dma_addr_t iova, BUG_ON(dma_pte_addr(*pte));
>>                 dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
>>                 dma_set_pte_prot(*pte, prot);
>> -               __iommu_flush_cache(domain->iommu, pte,
>> sizeof(*pte)); +               domain_flush_cache(domain, pte,
>>                 sizeof(*pte));                 start_pfn++; index++;
>>         }
>>         return 0;
>>  }
>>
>> -static void detach_domain_for_dev(struct dmar_domain *domain, u8
>> bus, u8 devfn) +static void detach_domain_for_dev(struct dmar_domain
>> *domain, +                                 u8 bus, u8 devfn)  {
>> -       clear_context_table(domain->iommu, bus, devfn);
>> -       domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
>> -                                          DMA_CCMD_GLOBAL_INVL, 0);
>> -       domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
>> -                                        DMA_TLB_GLOBAL_FLUSH, 0);
>> +       struct intel_iommu *iommu;
>> +
>> +       iommu = device_find_matched_iommu(bus, devfn); +       if
>> (!iommu) +               return;
>> +
>> +       clear_context_table(iommu, bus, devfn);
>> +       iommu->flush.flush_context(iommu, 0, 0, 0,
>> +                                  DMA_CCMD_GLOBAL_INVL, 0);
>> +       iommu->flush.flush_iotlb(iommu, 0, 0, 0,
>> +                                DMA_TLB_GLOBAL_FLUSH, 0);  }
>>
>>  static void domain_remove_dev_info(struct dmar_domain *domain)
>> @@ -1336,7 +1425,6 @@ static struct dmar_domain
>>         *get_domain_for_dev(struct pci_dev *pdev, int gaw)  { struct
>>         dmar_domain *domain, *found = NULL; struct intel_iommu
>> *iommu; -       struct dmar_drhd_unit *drhd;
>>         struct device_domain_info *info, *tmp;
>>         struct pci_dev *dev_tmp;
>>         unsigned long flags;
>> @@ -1371,13 +1459,9 @@ static struct dmar_domain
>> *get_domain_for_dev(struct pci_dev *pdev, int gaw)         }
>>
>>         /* Allocate new domain for the device */
>> -       drhd = dmar_find_matched_drhd_unit(pdev);
>> -       if (!drhd) {
>> -               printk(KERN_ERR "IOMMU: can't find DMAR for device
>> %s\n",
>> -                       pci_name(pdev));
>> -               return NULL;
>> -       }
>> -       iommu = drhd->iommu;
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return  NULL;
>>
>>         domain = iommu_alloc_domain(iommu);
>>         if (!domain)
>> @@ -1400,7 +1484,7 @@ static struct dmar_domain
>>                 *get_domain_for_dev(struct pci_dev *pdev, int gaw)
>>                 info->dev = NULL; info->domain = domain;
>>                 /* This domain is shared by devices under p2p bridge
>> */ -               domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
>> +               domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
>>
>>                 /* pcie-to-pci bridge already has a domain, uses it
>> */                 found = NULL; @@ -1805,7 +1889,7 @@
>> get_valid_domain_for_dev(struct pci_dev *pdev)         }
>>
>>         /* make sure context mapping is ok */
>> -       if (unlikely(!domain_context_mapped(domain, pdev))) {
>> +       if (unlikely(!domain_context_mapped(pdev))) {
>>                 ret = domain_context_mapping(domain, pdev);
>>                         if (ret) { printk(KERN_ERR
>> @@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct
>>         device *hwdev, phys_addr_t paddr,  { struct pci_dev *pdev =
>>         to_pci_dev(hwdev); struct dmar_domain *domain;
>> +       struct intel_iommu *iommu;
>>         phys_addr_t start_paddr;
>>         struct iova *iova;
>>         int prot = 0;
>> @@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct
>>                 device *hwdev, phys_addr_t paddr,         if
>> (!domain) return 0;
>>
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return 0;
>> +
>>         size = aligned_size((u64)paddr, size);
>>
>>         iova = __intel_alloc_iova(hwdev, domain, size,
>> pdev->dma_mask); @@ -1849,7 +1938,7 @@ static dma_addr_t
>> __intel_map_single(struct device *hwdev, phys_addr_t paddr,
>>          * mappings..
>>          */
>>         if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
>> -                       !cap_zlr(domain->iommu->cap))
>> +                       !cap_zlr(iommu->cap))
>>                 prot |= DMA_PTE_READ;
>>         if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
>>                 prot |= DMA_PTE_WRITE;
>> @@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct
>> device *hwdev, phys_addr_t paddr,                 goto error;
>>
>>         /* it's a non-present to present mapping */
>> -       ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
>> +       ret = iommu_flush_iotlb_psi(iommu, domain->id,
>>                         start_paddr, size >> VTD_PAGE_SHIFT, 1);
>> if (ret) -               iommu_flush_write_buffer(domain->iommu);
>> +               iommu_flush_write_buffer(iommu);
>>
>>         return start_paddr + ((u64)paddr & (~PAGE_MASK));
>>
>> @@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
>>         /* just flush them all */
>>         for (i = 0; i < g_num_of_iommus; i++) {
>>                 if (deferred_flush[i].next) {
>> -                       struct intel_iommu *iommu =
>> -                               deferred_flush[i].domain[0]->iommu;
>> +                       struct intel_iommu *iommu =
>> deferred_flush[i].iommu;
>>
>>                         iommu->flush.flush_iotlb(iommu, 0, 0, 0,
>>
>> DMA_TLB_GLOBAL_FLUSH, 0); @@ -1921,7 +2009,8 @@ static void
>>         flush_unmaps_timeout(unsigned long data)
>> spin_unlock_irqrestore(&async_umap_flush_lock, flags);  }
>>
>> -static void add_unmap(struct dmar_domain *dom, struct iova *iova)
>> +static void add_unmap(struct dmar_domain *dom,
>> +                     struct intel_iommu *iommu, struct iova *iova)
>>         { unsigned long flags;
>>         int next, iommu_id;
>> @@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain
>>         *dom, struct iova *iova) if (list_size == HIGH_WATER_MARK)
>>                 flush_unmaps();
>>
>> -       iommu_id = dom->iommu->seq_id;
>> +       iommu_id = iommu->seq_id;
>>
>>         next = deferred_flush[iommu_id].next;
>>         deferred_flush[iommu_id].domain[next] = dom;
>>         deferred_flush[iommu_id].iova[next] = iova;
>> +       deferred_flush[iommu_id].iommu = iommu;
>>         deferred_flush[iommu_id].next++;
>>
>>         if (!timer_on) {
>> @@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev,
>>         dma_addr_t dev_addr, size_t size, struct dmar_domain *domain;
>>         unsigned long start_addr;
>>         struct iova *iova;
>> +       struct intel_iommu *iommu;
>>
>>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
>>         return; domain = find_domain(pdev);
>>         BUG_ON(!domain);
>>
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return;
>> +
>>         iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
>>                 if (!iova) return;
>> @@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev,
>>         dma_addr_t dev_addr, size_t size, /* free page tables */
>>         dma_pte_free_pagetable(domain, start_addr, start_addr +
>>         size); if (intel_iommu_strict) {
>> -               if (iommu_flush_iotlb_psi(domain->iommu,
>> +               if (iommu_flush_iotlb_psi(iommu,
>>                         domain->id, start_addr, size >>
>> VTD_PAGE_SHIFT, 0)) -
>> iommu_flush_write_buffer(domain->iommu); +
>>                 iommu_flush_write_buffer(iommu); /* free iova */
>>                 __free_iova(&domain->iovad, iova);
>>         } else {
>> -               add_unmap(domain, iova);
>> +               add_unmap(domain, iommu, iova);
>>                 /*
>>                  * queue up the release of the unmap to save the
>> 1/6th of the
>>                  * cpu used up by the iotlb flush operation...
>> @@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev,
>>         struct scatterlist *sglist,         size_t size = 0; void
>>         *addr; struct scatterlist *sg;
>> +       struct intel_iommu *iommu;
>>
>>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
>> return;
>>
>>         domain = find_domain(pdev);
>>
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return;
>> +
>>         iova = find_iova(&domain->iovad,
>>                 IOVA_PFN(sglist[0].dma_address));         if (!iova)
>> return; @@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device
>>         *hwdev, struct scatterlist *sglist, /* free page tables */
>>         dma_pte_free_pagetable(domain, start_addr, start_addr +
>> size);
>>
>> -       if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
>> start_addr, +       if (iommu_flush_iotlb_psi(iommu, domain->id,
>>                         start_addr, size >> VTD_PAGE_SHIFT, 0))
>> -               iommu_flush_write_buffer(domain->iommu);
>> +               iommu_flush_write_buffer(iommu);
>>
>>         /* free iova */
>>         __free_iova(&domain->iovad, iova);
>> @@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct
>>         scatterlist *sglist, int nelems,         int ret; struct
>>         scatterlist *sg; unsigned long start_addr;
>> +       struct intel_iommu *iommu;
>>
>>         BUG_ON(dir == DMA_NONE);
>>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
>> @@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct
>>                 scatterlist *sglist, int nelems,         if
>> (!domain) return 0;
>>
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return 0;
>> +
>>         for_each_sg(sglist, sg, nelems, i) {
>>                 addr = SG_ENT_VIRT_ADDRESS(sg);
>>                 addr = (void *)virt_to_phys(addr);
>> @@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct
>> scatterlist *sglist, int nelems,
>>          * mappings..
>>          */
>>         if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
>> -                       !cap_zlr(domain->iommu->cap))
>> +                       !cap_zlr(iommu->cap))
>>                 prot |= DMA_PTE_READ;
>>         if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
>>                 prot |= DMA_PTE_WRITE;
>> @@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct
>> scatterlist *sglist, int nelems,         }
>>
>>         /* it's a non-present to present mapping */
>> -       if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
>> +       if (iommu_flush_iotlb_psi(iommu, domain->id,
>>                         start_addr, offset >> VTD_PAGE_SHIFT, 1))
>> -               iommu_flush_write_buffer(domain->iommu);
>> +               iommu_flush_write_buffer(iommu);
>>         return nelems;
>>  }
>>
>> @@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
>>  return 0; }
>>
>> -void intel_iommu_domain_exit(struct dmar_domain *domain)
>> +/* domain id for virtual machine, it won't be set in context */
>> +static unsigned long vm_domid;
>> +
>> +static int vm_domain_min_agaw(struct dmar_domain *domain) +{
>> +       struct dmar_drhd_unit *drhd;
>> +       struct intel_iommu *iommu;
>> +       int min_agaw = domain->agaw;
>> +
>> +       for_each_drhd_unit(drhd) {
>> +               if (drhd->ignored)
>> +                       continue;
>> +               iommu = drhd->iommu;
>> +
>> +               if (test_bit(iommu->seq_id, &domain->iommu_bmp))
>> +                       if (min_agaw > iommu->agaw)
>> +                               min_agaw = iommu->agaw; +       }
>> +
>> +       return min_agaw;
>> +}
>> +
>> +static int vm_domain_add_dev_info(struct dmar_domain *domain,
>> +                                 struct pci_dev *pdev) +{
>> +       struct device_domain_info *info;
>> +       unsigned long flags;
>> +
>> +       info = alloc_devinfo_mem();
>> +       if (!info)
>> +               return -ENOMEM;
>> +
>> +       info->bus = pdev->bus->number;
>> +       info->devfn = pdev->devfn;
>> +       info->dev = pdev;
>> +       info->domain = domain;
>> +
>> +       spin_lock_irqsave(&device_domain_lock, flags);
>> +       list_add(&info->link, &domain->devices);
>> +       list_add(&info->global, &device_domain_list);
>> +       pdev->dev.archdata.iommu = info;
>> +       spin_unlock_irqrestore(&device_domain_lock, flags); +
>> +       return 0;
>> +}
>> +
>> +static void vm_domain_remove_one_dev_info(struct dmar_domain
>> *domain, +                                         struct pci_dev
>> *pdev) +{ +       struct device_domain_info *info;
>> +       struct intel_iommu *iommu;
>> +       unsigned long flags;
>> +       int found = 0;
>> +
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + +       spin_lock_irqsave(&device_domain_lock,
>> flags); +       while (!list_empty(&domain->devices)) {
>> +               info = list_entry(domain->devices.next,
>> +                       struct device_domain_info, link);
>> +               if (info->bus == pdev->bus->number &&
>> +                   info->devfn == pdev->devfn) {
>> +                       list_del(&info->link);
>> +                       list_del(&info->global);
>> +                       if (info->dev)
>> +                               info->dev->dev.archdata.iommu = NULL;
>> +                       spin_unlock_irqrestore(&device_domain_lock,
>> flags); + +                       detach_domain_for_dev(info->domain,
>> +                                             info->bus,
>> info->devfn); +                       free_devinfo_mem(info);
>> +
>> +                       spin_lock_irqsave(&device_domain_lock,
>> flags); + +                       if (found)
>> +                               break;
>> +                       else
>> +                               continue;
>> +               }
>> +
>> +               /* if there is no other devices under the same iommu
>> +                * owned by this domain, clear this iommu in
>> iommu_bmp +                */ +               if
>> (device_find_matched_iommu(info->bus, info->devfn) == iommu) +
>> found = 1; +       }
>> +
>> +       if (found == 0) {
>> +               spin_lock_irqsave(&iommu->lock, flags);
>> +               clear_bit(iommu->seq_id, &domain->iommu_bmp);
>> +               domain->iommu_count--;
>> +               spin_unlock_irqrestore(&iommu->lock, flags); +
>> } +
>> +       spin_unlock_irqrestore(&device_domain_lock, flags); +}
>> +
>> +static void vm_domain_remove_all_dev_info(struct dmar_domain
>> *domain) +{ +       struct device_domain_info *info;
>> +       struct intel_iommu *iommu;
>> +       unsigned long flags;
>> +
>> +       spin_lock_irqsave(&device_domain_lock, flags);
>> +       while (!list_empty(&domain->devices)) {
>> +               info = list_entry(domain->devices.next,
>> +                       struct device_domain_info, link);
>> +               list_del(&info->link);
>> +               list_del(&info->global);
>> +               if (info->dev)
>> +                       info->dev->dev.archdata.iommu = NULL; +
>> +               spin_unlock_irqrestore(&device_domain_lock, flags);
>> +               detach_domain_for_dev(info->domain,
>> +                                     info->bus, info->devfn); +
>> +               /* clear this iommu in iommu_bmp */
>> +               iommu = device_find_matched_iommu(info->bus,
>> info->devfn); +               spin_lock_irqsave(&iommu->lock, flags);
>> +               if (test_and_clear_bit(iommu->seq_id,
>> +                                      &domain->iommu_bmp))
>> +                       domain->iommu_count--;
>> +               spin_unlock_irqrestore(&iommu->lock, flags); +
>> +               free_devinfo_mem(info);
>> +               spin_lock_irqsave(&device_domain_lock, flags); +
>> } +       spin_unlock_irqrestore(&device_domain_lock, flags); +}
>> +
>> +static int vm_domain_context_mapping_one(struct dmar_domain *domain,
>> +               struct intel_iommu *iommu, u8 bus, u8 devfn) +{
>> +       struct context_entry *context;
>> +       unsigned long flags;
>> +       struct dma_pte *pgd;
>> +       unsigned long num;
>> +       unsigned long ndomains;
>> +       int id;
>> +       int agaw;
>> +       int found = 0;
>> +
>> +       pr_debug("Set context mapping for %02x:%02x.%d\n",
>> +               bus, PCI_SLOT(devfn), PCI_FUNC(devfn)); +
>> BUG_ON(!domain->pgd); +       context =
>> device_to_context_entry(iommu, bus, devfn); +       if (!context) +
>> return -ENOMEM; +       spin_lock_irqsave(&iommu->lock, flags);
>> +       if (context_present(*context)) {
>> +               spin_unlock_irqrestore(&iommu->lock, flags); +
>> return 0; +       }
>> +
>> +       id = domain->id;
>> +
>> +       /* find an available domain id for this device in iommu */
>> +       ndomains = cap_ndoms(iommu->cap);
>> +       num = find_first_bit(iommu->domain_ids, ndomains);
>> +       for (; num < ndomains; ) {
>> +               if (iommu->domains[num] == domain) {
>> +                       id = num;
>> +                       found = 1;
>> +                       break;
>> +               }
>> +               num = find_next_bit(iommu->domain_ids,
>> +                 cap_ndoms(iommu->cap), num+1);
>> +       }
>> +
>> +       if (found == 0) {
>> +               num = find_first_zero_bit(iommu->domain_ids,
>> ndomains); +               if (num >= ndomains) {
>> +                       spin_unlock_irqrestore(&iommu->lock, flags);
>> +                       printk(KERN_ERR "IOMMU: no free domain
>> ids\n"); +                       return -EFAULT;
>> +               }
>> +
>> +               set_bit(num, iommu->domain_ids);
>> +               iommu->domains[num] = domain;
>> +               id = num;
>> +       }
>> +
>> +       pgd = domain->pgd;
>> +
>> +       /* Skip top levels of page tables for
>> +        * iommu which has less agaw than default.
>> +        */
>> +       for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
>> +               pgd = phys_to_virt(dma_pte_addr(*pgd));
>> +               if (!dma_pte_present(*pgd)) {
>> +                       spin_unlock_irqrestore(&iommu->lock, flags);
>> +                       return -ENOMEM;
>> +               }
>> +       }
>> +
>> +       context_set_domain_id(*context, id);
>> +       context_set_address_width(*context, iommu->agaw);
>> +       context_set_address_root(*context, virt_to_phys(pgd));
>> +       context_set_translation_type(*context,
>> CONTEXT_TT_MULTI_LEVEL); +       context_set_fault_enable(*context);
>> +       context_set_present(*context);
>> +       __iommu_flush_cache(iommu, context, sizeof(*context)); +
>> +       /* it's a non-present to present mapping */
>> +       if (iommu->flush.flush_context(iommu, id,
>> +               (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
>> +               DMA_CCMD_DEVICE_INVL, 1))
>> +               iommu_flush_write_buffer(iommu);
>> +       else
>> +               iommu->flush.flush_iotlb(iommu, 0, 0, 0,
>> DMA_TLB_DSI_FLUSH, 0); + +       if
>> (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp)) +
>> domain->iommu_count++; +       spin_unlock_irqrestore(&iommu->lock,
>> flags); +       return 0; +}
>> +
>> +static int
>> +vm_domain_context_mapping(struct dmar_domain *domain, struct
>> pci_dev *pdev) +{ +       int ret;
>> +       struct pci_dev *tmp, *parent;
>> +       struct intel_iommu *iommu;
>> +
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return  -ENODEV;
>> +
>> +       ret = vm_domain_context_mapping_one(domain, iommu,
>> +               pdev->bus->number, pdev->devfn);
>> +       if (ret)
>> +               return ret;
>> +
>> +       /* dependent device mapping */
>> +       tmp = pci_find_upstream_pcie_bridge(pdev);
>> +       if (!tmp)
>> +               return 0;
>> +       /* Secondary interface's bus number and devfn 0 */
>> +       parent = pdev->bus->self;
>> +       while (parent != tmp) {
>> +               ret = vm_domain_context_mapping_one(domain, iommu,
>> +                       parent->bus->number, parent->devfn); +
>> if (ret) +                       return ret;
>> +               parent = parent->bus->self;
>> +       }
>> +       if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
>> +               return vm_domain_context_mapping_one(domain, iommu,
>> +                       tmp->subordinate->number, 0);
>> +       else /* this is a legacy PCI bridge */
>> +               return vm_domain_context_mapping_one(domain, iommu,
>> +                       tmp->bus->number, tmp->devfn); +}
>> +
>> +
>> +static int vm_domain_init(struct dmar_domain *domain, int
>> guest_width) +{ +       int adjust_width;
>> +
>> +       init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
>> +       spin_lock_init(&domain->mapping_lock);
>> +
>> +       domain_reserve_special_ranges(domain);
>> +
>> +       /* calculate AGAW */
>> +       domain->gaw = guest_width;
>> +       adjust_width = guestwidth_to_adjustwidth(guest_width);
>> +       domain->agaw = width_to_agaw(adjust_width); +
>> +       INIT_LIST_HEAD(&domain->devices);
>> +
>> +       /* always allocate the top pgd */
>> +       domain->pgd = (struct dma_pte *)alloc_pgtable_page(); +
>> if (!domain->pgd) +               return -ENOMEM;
>> +       domain_flush_cache(domain, domain->pgd, PAGE_SIZE); +
>> return 0; +}
>> +
>> +static void iommu_free_vm_domain(struct dmar_domain *domain) +{
>> +       unsigned long flags;
>> +       struct dmar_drhd_unit *drhd;
>> +       struct intel_iommu *iommu;
>> +       unsigned long i;
>> +       unsigned long ndomains;
>> +
>> +       for_each_drhd_unit(drhd) {
>> +               if (drhd->ignored)
>> +                       continue;
>> +               iommu = drhd->iommu;
>> +
>> +               ndomains = cap_ndoms(iommu->cap);
>> +               i = find_first_bit(iommu->domain_ids, ndomains);
>> +               for (; i < ndomains; ) {
>> +                       if (iommu->domains[i] == domain) {
>> +                               spin_lock_irqsave(&iommu->lock,
>> flags); +                               clear_bit(i,
>> iommu->domain_ids); +
>> iommu->domains[i] = NULL; +
>> spin_unlock_irqrestore(&iommu->lock, flags); +
>> break; +                       }
>> +                       i = find_next_bit(iommu->domain_ids,
>> ndomains, i+1); +               } +       }
>> +}
>> +
>> +static void vm_domain_exit(struct dmar_domain *domain)  {
>>         u64 end;
>>
>> @@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct
>>                 dmar_domain *domain)         if (!domain) return;
>>
>> +       vm_domain_remove_all_dev_info(domain);
>> +       /* destroy iovas */
>> +       put_iova_domain(&domain->iovad);
>>         end = DOMAIN_MAX_ADDR(domain->gaw);
>> -       end = end & (~VTD_PAGE_MASK);
>> +       end &= VTD_PAGE_MASK;
>>
>>         /* clear ptes */
>>         dma_pte_clear_range(domain, 0, end);
>> @@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct
>>         dmar_domain *domain) /* free page tables */
>>         dma_pte_free_pagetable(domain, 0, end);
>>
>> -       iommu_free_domain(domain);
>> +       iommu_free_vm_domain(domain);
>>         free_domain_mem(domain);
>>  }
>> -EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
>>
>> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
>> +static struct dmar_domain *iommu_alloc_vm_domain(void)  {
>> -       struct dmar_drhd_unit *drhd;
>>         struct dmar_domain *domain;
>> -       struct intel_iommu *iommu;
>>
>> -       drhd = dmar_find_matched_drhd_unit(pdev);
>> -       if (!drhd) {
>> -               printk(KERN_ERR "intel_iommu_domain_alloc: drhd ==
>> NULL\n"); +       domain = alloc_domain_mem();
>> +       if (!domain)
>>                 return NULL;
>> -       }
>>
>> -       iommu = drhd->iommu;
>> -       if (!iommu) {
>> -               printk(KERN_ERR
>> -                       "intel_iommu_domain_alloc: iommu == NULL\n");
>> -               return NULL;
>> -       }
>> -       domain = iommu_alloc_domain(iommu);
>> +       domain->id = vm_domid++;
>> +       domain->iommu_count = 0;
>> +       domain->max_addr = 0;
>> +       memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
>> +       domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE; +
>> +       return domain;
>> +}
>> +
>> +struct dmar_domain *intel_iommu_alloc_domain(void) +{
>> +       struct dmar_domain *domain;
>> +
>> +       domain = iommu_alloc_vm_domain();
>>         if (!domain) {
>>                 printk(KERN_ERR
>>                         "intel_iommu_domain_alloc: domain ==
>>         NULL\n");                 return NULL; }
>> -       if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
>> +       if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
>>                 printk(KERN_ERR
>>                         "intel_iommu_domain_alloc: domain_init()
>> failed\n"); -               intel_iommu_domain_exit(domain);
>> +               vm_domain_exit(domain);
>>                 return NULL;
>>         }
>> +
>>         return domain;
>>  }
>> -EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
>> +EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
>> +
>> +void intel_iommu_free_domain(struct dmar_domain *domain) +{
>> +       vm_domain_exit(domain);
>> +}
>> +EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
>>
>> -int intel_iommu_context_mapping(
>> -       struct dmar_domain *domain, struct pci_dev *pdev)
>> +int intel_iommu_assign_device(struct dmar_domain *domain,
>> +                             struct pci_dev *pdev)  {
>> -       int rc;
>> -       rc = domain_context_mapping(domain, pdev);
>> -       return rc;
>> +       struct intel_iommu *iommu;
>> +       int addr_width;
>> +       u64 end;
>> +       int ret;
>> +
>> +       /* normally pdev is not mapped */
>> +       if (unlikely(domain_context_mapped(pdev))) {
>> +               struct dmar_domain *old_domain;
>> +
>> +               old_domain = find_domain(pdev);
>> +               if (old_domain) {
>> +                       if (domain->flags &
>> DOMAIN_FLAG_VIRTUAL_MACHINE) +
>> vm_domain_remove_one_dev_info(old_domain, pdev); +
>> else +
>> domain_remove_dev_info(old_domain); +               } +       } +
>> +       iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); +       if (!iommu) +               return  -ENODEV;
>> +
>> +       /* check if this iommu agaw is sufficient for max mapped
>> address */ +       addr_width = agaw_to_width(iommu->agaw);
>> +       end = DOMAIN_MAX_ADDR(addr_width);
>> +       end = end & VTD_PAGE_MASK;
>> +       if (end < domain->max_addr) {
>> +               printk(KERN_ERR "%s: iommu agaw (%d) is not "
>> +                      "sufficient for the mapped address (%llx)\n",
>> +                      __func__, iommu->agaw, domain->max_addr);
>> +               return -EFAULT;
>> +       }
>> +
>> +       ret = vm_domain_context_mapping(domain, pdev); +       if
>> (ret) +               return ret;
>> +
>> +       ret = vm_domain_add_dev_info(domain, pdev); +       return
>>  ret; }
>> -EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
>> +EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
>>
>> -int intel_iommu_page_mapping(
>> -       struct dmar_domain *domain, dma_addr_t iova,
>> -       u64 hpa, size_t size, int prot)
>> +
>> +void intel_iommu_deassign_device(struct dmar_domain *domain,
>> +                                struct pci_dev *pdev)  {
>> -       int rc;
>> -       rc = domain_page_mapping(domain, iova, hpa, size, prot);
>> -       return rc;
>> +       vm_domain_remove_one_dev_info(domain, pdev);  }
>> -EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
>> +EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
>>
>> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8
>> devfn) +int intel_iommu_map_pages(struct dmar_domain *domain,
>> dma_addr_t iova, +                         u64 hpa, size_t size, int
>> prot)  { -       detach_domain_for_dev(domain, bus, devfn);
>> +       u64 max_addr;
>> +       int addr_width;
>> +       int ret;
>> +
>> +       max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
>> +       if (domain->max_addr < max_addr) {
>> +               int min_agaw;
>> +               u64 end;
>> +
>> +               /* check if minimum agaw is sufficient for mapped
>> address */ +               min_agaw = vm_domain_min_agaw(domain);
>> +               addr_width = agaw_to_width(min_agaw);
>> +               end = DOMAIN_MAX_ADDR(addr_width);
>> +               end = end & VTD_PAGE_MASK;
>> +               if (end < max_addr) {
>> +                       printk(KERN_ERR "%s: iommu agaw (%d) is not "
>> +                              "sufficient for the mapped address
>> (%llx)\n", +                              __func__, min_agaw,
>> max_addr); +                       return -EFAULT;
>> +               }
>> +               domain->max_addr = max_addr;
>> +       }
>> +
>> +       ret = domain_page_mapping(domain, iova, hpa, size, prot); +
>>  return ret; }
>> -EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
>> +EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
>>
>> -struct dmar_domain *
>> -intel_iommu_find_domain(struct pci_dev *pdev)
>> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
>> +                            dma_addr_t iova, size_t size)  {
>> -       return find_domain(pdev);
>> +       dma_addr_t base;
>> +
>> +       /* The address might not be aligned */
>> +       base = iova & PAGE_MASK;
>> +       size = PAGE_ALIGN(size);
>> +       dma_pte_clear_range(domain, base, base + size);  }
>> -EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
>> +EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
>>
>>  int intel_iommu_found(void)
>>  {
>> diff --git a/include/linux/dma_remapping.h
>> b/include/linux/dma_remapping.h
>> index 952df39..3568ae6 100644
>> --- a/include/linux/dma_remapping.h
>> +++ b/include/linux/dma_remapping.h
>> @@ -111,11 +111,21 @@ struct dma_pte {
>>                 (p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
>>  #define dma_pte_present(p) (((p).val & 3) != 0)
>>
>> +/* domain flags, one domain owns one device by default */ +
>> +/* devices under the same p2p bridge are owned in one domain */
>> +#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES       (1 << 0) +
>> +/* domain represents a virtual machine, more than one devices
>> + * across iommus may be owned in one domain, e.g. kvm guest. + */
>> +#define DOMAIN_FLAG_VIRTUAL_MACHINE            (1 << 1) +
>>  struct intel_iommu;
>>
>>  struct dmar_domain {
>>         int     id;                     /* domain id */
>> -       struct intel_iommu *iommu;      /* back pointer to owning
>> iommu */ +       unsigned long iommu_bmp;        /* bitmap of iommus
>> this domain uses*/
>>
>>         struct list_head devices;       /* all devices' list */
>>         struct iova_domain iovad;       /* iova's that belong to
>> this domain */ @@ -123,12 +133,13 @@ struct dmar_domain {
>>         struct dma_pte  *pgd;           /* virtual address */
>>         spinlock_t      mapping_lock;   /* page table lock */
>>         int             gaw;            /* max guest address width */
>> +       int             agaw;           /* adjusted guest address
>> width */
>>
>> -       /* adjusted guest address width, 0 is level 2 30-bit */
>> -       int             agaw;
>> +       int             flags;          /* domain flag */
>>
>> -#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
>> -       int             flags;
>> +       /* following fields are used in virtual machine case */
>> +       int             iommu_count;    /* reference count of iommu
>> */ +       u64             max_addr;       /* maximum mapped address
>> */  };
>>
>>  /* PCI domain-device relationship */
>> diff --git a/include/linux/intel-iommu.h
>> b/include/linux/intel-iommu.h
>> index 3d017cf..c2f37b8 100644
>> --- a/include/linux/intel-iommu.h
>> +++ b/include/linux/intel-iommu.h
>> @@ -219,6 +219,8 @@ do {
>>         \ }
>> \  } while (0)
>>
>> +#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
>> +
>>  #define QI_LENGTH      256     /* queue length */
>>
>>  enum {
>> @@ -299,6 +301,7 @@ struct intel_iommu {
>>         struct dmar_domain **domains; /* ptr to domains */
>>         spinlock_t      lock; /* protect context, domain ids */
>>         struct root_entry *root_entry; /* virtual address */ +
>> int             agaw;
>>
>>         unsigned int irq;
>>         unsigned char name[7];    /* Device Name */
>> @@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu
>> *iommu, u16 did, u64 addr,
>>
>>  extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu
>> *iommu);
>>
>> -void intel_iommu_domain_exit(struct dmar_domain *domain);
>> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
>> -int intel_iommu_context_mapping(struct dmar_domain *domain,
>> -                               struct pci_dev *pdev);
>> -int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t
>> iova,
>> -                            u64 hpa, size_t size, int prot);
>> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8
>> devfn);
>> -struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
>> +struct dmar_domain *intel_iommu_alloc_domain(void);
>> +void intel_iommu_free_domain(struct dmar_domain *domain);
>> +int intel_iommu_assign_device(struct dmar_domain *domain,
>> +                             struct pci_dev *pdev);
>> +void intel_iommu_deassign_device(struct dmar_domain *domain,
>> +                                struct pci_dev *pdev);
>> +int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t
>> iova, +                         u64 hpa, size_t size, int prot);
>> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
>> +                            dma_addr_t iova, size_t size);
>>  u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
>>
>>  #ifdef CONFIG_DMAR
>> --
>> 1.5.1


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
  2008-12-01  6:17   ` Han, Weidong
@ 2008-12-01 11:55     ` 'Joerg Roedel'
  2008-12-01 13:22       ` Han, Weidong
  0 siblings, 1 reply; 7+ messages in thread
From: 'Joerg Roedel' @ 2008-12-01 11:55 UTC (permalink / raw)
  To: Han, Weidong
  Cc: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity',
	Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
	'iommu@lists.linux-foundation.org'

Hmm, I get these errors using git-am:

Applying VT-d: Support multiple device assignment for KVM
.dotest/patch:1344: space before tab in indent.
                                clflush_cache_range(addr, size);
.dotest/patch:1350: space before tab in indent.
                        clflush_cache_range(addr, size);
.dotest/patch:1907: trailing whitespace.

.dotest/patch:1946: trailing whitespace.
                 * owned by this domain, clear this iommu in iommu_bmp 
.dotest/patch:2300: trailing whitespace.

error: patch failed: drivers/pci/dmar.c:484
error: drivers/pci/dmar.c: patch does not apply
error: patch failed: drivers/pci/intel-iommu.c:50
error: drivers/pci/intel-iommu.c: patch does not apply
error: patch failed: include/linux/dma_remapping.h:111
error: include/linux/dma_remapping.h: patch does not apply
error: patch failed: include/linux/intel-iommu.h:219
error: include/linux/intel-iommu.h: patch does not apply
Patch failed at 0001.

Joerg

On Mon, Dec 01, 2008 at 02:17:38PM +0800, Han, Weidong wrote:
> It's developed based on commit 0f7d3ee6 on avi/master, but it still can be applied on latest avi/master (commit 90755652).
> 
> Regards,
> Weidong
> 
> Joerg Roedel wrote:
> > Hmm, I tried to apply this patch against avi/master and linus/master
> > but get merge conflicts. Where do these patches apply cleanly?
> >
> > Joerg
> >
> > On Thu, Nov 27, 2008 at 09:49:04PM +0800, Han, Weidong wrote:
> >> In order to support multiple device assignment for KVM, this patch
> >> does following main changes:
> >>    - extend dmar_domain to own multiple devices from different
> >> iommus, use a bitmap of iommus to replace iommu pointer in
> >> dmar_domain.
> >>    - implement independent low level functions for kvm, then won't
> >> impact native VT-d.
> >>    - "SAGAW" capability may be different across iommus, that's to
> >> say the VT-d page table levels may be different among iommus. This
> >> patch uses a defaut agaw, and skip top levels of page tables for
> >> iommus which have smaller agaw than default.
> >>    - rename the APIs for kvm VT-d, make it more readable.
> >>
> >>
> >> Signed-off-by: Weidong Han <weidong.han@intel.com>
> >> ---
> >>  drivers/pci/dmar.c            |   15 +
> >>  drivers/pci/intel-iommu.c     |  698
> >>  ++++++++++++++++++++++++++++++++++------
> >>  include/linux/dma_remapping.h |   21 +- include/linux/intel-iommu.h
> >>  |   21 +- 4 files changed, 637 insertions(+), 118 deletions(-)
> >>
> >> diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
> >> index 691b3ad..d6bdced 100644
> >> --- a/drivers/pci/dmar.c
> >> +++ b/drivers/pci/dmar.c
> >> @@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
> >>  dmar_tbl = NULL; }
> >>
> >> +extern int width_to_agaw(int width);
> >>
> >>  int alloc_iommu(struct dmar_drhd_unit *drhd)
> >>  {
> >> @@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
> >>         int map_size; u32 ver;
> >>         static int iommu_allocated = 0;
> >> +       unsigned long sagaw;
> >> +       int agaw;
> >>
> >>         iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);         if
> >> (!iommu) @@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit
> >>         *drhd) iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
> >>         iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
> >>
> >> +       /* set agaw, "SAGAW" may be different across iommus */
> >> +       sagaw = cap_sagaw(iommu->cap);
> >> +       for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
> >> +            agaw >= 0; agaw--)
> >> +               if (test_bit(agaw, &sagaw))
> >> +                       break;
> >> +       if (agaw < 0) {
> >> +               printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n",
> >> sagaw); +               goto error; +       }
> >> +       iommu->agaw = agaw;
> >> +
> >>         /* the registers might be more than one page */
> >>         map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
> >>                 cap_max_fault_reg_offset(iommu->cap));
> >> diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
> >> index 5c8baa4..55b96c4 100644
> >> --- a/drivers/pci/intel-iommu.c
> >> +++ b/drivers/pci/intel-iommu.c
> >> @@ -50,8 +50,6 @@
> >>  #define IOAPIC_RANGE_END       (0xfeefffff)
> >>  #define IOVA_START_ADDR                (0x1000)
> >>
> >> -#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
> >> -
> >>  #define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
> >>
> >>
> >> @@ -64,6 +62,7 @@ struct deferred_flush_tables {
> >>         int next;
> >>         struct iova *iova[HIGH_WATER_MARK];
> >>         struct dmar_domain *domain[HIGH_WATER_MARK];
> >> +       struct intel_iommu *iommu;
> >>  };
> >>
> >>  static struct deferred_flush_tables *deferred_flush;
> >> @@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
> >>         kmem_cache_free(iommu_iova_cache, iova);
> >>  }
> >>
> >> +/* in native case, each domain is related to only one iommu */
> >> +static struct intel_iommu *domain_get_only_iommu(struct dmar_domain
> >> *domain) +{ +       struct dmar_drhd_unit *drhd;
> >> +
> >> +       for_each_drhd_unit(drhd) {
> >> +               if (drhd->ignored)
> >> +                       continue;
> >> +               if (test_bit(drhd->iommu->seq_id,
> >> &domain->iommu_bmp)) +                       return drhd->iommu;
> >> +       }
> >> +
> >> +       return NULL;
> >> +}
> >> +
> >> +static void domain_flush_cache(struct dmar_domain *domain,
> >> +                              void *addr, int size) +{
> >> +       struct intel_iommu *iommu;
> >> +
> >> +       if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
> >> +               struct dmar_drhd_unit *drhd;
> >> +
> >> +               for_each_drhd_unit(drhd) {
> >> +                       if (drhd->ignored)
> >> +                               continue;
> >> +                       iommu = drhd->iommu;
> >> +
> >> +                       if (!test_bit(iommu->seq_id,
> >> &domain->iommu_bmp)) +                               continue;
> >> +
> >> +                       if (!ecap_coherent(iommu->ecap))
> >> +                               clflush_cache_range(addr, size); +
> >> } +       }
> >> +       else {
> >> +               iommu = domain_get_only_iommu(domain);
> >> +               if (iommu && !ecap_coherent(iommu->ecap))
> >> +                       clflush_cache_range(addr, size); +       }
> >> +}
> >> +
> >> +static struct intel_iommu *device_find_matched_iommu(u8 bus, u8
> >> devfn) +{ +       struct dmar_drhd_unit *drhd = NULL;
> >> +       int i;
> >> +
> >> +       for_each_drhd_unit(drhd) {
> >> +               if (drhd->ignored)
> >> +                       continue;
> >> +
> >> +               for (i = 0; i < drhd->devices_cnt; i++)
> >> +                       if (drhd->devices[i]->bus->number == bus &&
> >> +                           drhd->devices[i]->devfn == devfn)
> >> +                               return drhd->iommu; +
> >> +               if (drhd->include_all)
> >> +                       return drhd->iommu;
> >> +       }
> >> +
> >> +       return NULL;
> >> +}
> >> +
> >>  /* Gets context entry for a given bus and devfn */
> >>  static struct context_entry * device_to_context_entry(struct
> >>                 intel_iommu *iommu, u8 bus, u8 devfn)
> >> @@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
> >>
> >>  }
> >>
> >> -static inline int width_to_agaw(int width)
> >> +int width_to_agaw(int width)
> >>  {
> >>         return (width - 30) / LEVEL_STRIDE;
> >>  }
> >> @@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct
> >>                                         dmar_domain *domain, u64
> >>                                 addr) flags); return NULL;
> >>                         }
> >> -                       __iommu_flush_cache(domain->iommu, tmp_page,
> >> -                                       PAGE_SIZE);
> >> +                       domain_flush_cache(domain, tmp_page,
> >>                         PAGE_SIZE); dma_set_pte_addr(*pte,
> >>                          virt_to_phys(tmp_page));
> >> /* * high level table always sets r/w, last level page @@ -356,7
> >>                         +417,7 @@ static struct dma_pte *
> >>                         addr_to_dma_pte(struct dmar_domain *domain,
> >> u64 addr)                          */ dma_set_pte_readable(*pte);
> >> dma_set_pte_writable(*pte); -
> >>                 __iommu_flush_cache(domain->iommu, pte,
> >> sizeof(*pte)); +                       domain_flush_cache(domain,
> >> pte, sizeof(*pte));                 } parent =
> >> phys_to_virt(dma_pte_addr(*pte));                 level--; @@ -399,7
> >> +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain,
> >> u64 addr)
> >>
> >>         if (pte) {
> >>                 dma_clear_pte(*pte);
> >> -               __iommu_flush_cache(domain->iommu, pte,
> >> sizeof(*pte)); +               domain_flush_cache(domain, pte,
> >>  sizeof(*pte));         } }
> >>
> >> @@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct
> >>                                 dmar_domain *domain,
> >>                                         free_pgtable_page(
> >>                                 phys_to_virt(dma_pte_addr(*pte)));
> >> dma_clear_pte(*pte);
> >> -                               __iommu_flush_cache(domain->iommu,
> >> -                                               pte, sizeof(*pte));
> >> +                               domain_flush_cache(domain, pte,
> >>                         sizeof(*pte));                         } tmp
> >>                 += level_size(level); }
> >> @@ -948,8 +1008,8 @@ static int iommu_init_domains(struct
> >>  intel_iommu *iommu)         return 0; }
> >>
> >> -
> >>  static void domain_exit(struct dmar_domain *domain);
> >> +static void vm_domain_exit(struct dmar_domain *domain);
> >>
> >>  void free_dmar_iommu(struct intel_iommu *iommu)
> >>  {
> >> @@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
> >>         for (; i < cap_ndoms(iommu->cap); ) {
> >>                 domain = iommu->domains[i];
> >>                 clear_bit(i, iommu->domain_ids);
> >> -               domain_exit(domain);
> >> +
> >> +               if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
> >> +                       if (--domain->iommu_count == 0)
> >> +                               vm_domain_exit(domain); +
> >> } +               else
> >> +                       domain_exit(domain);
> >> +
> >>                 i = find_next_bit(iommu->domain_ids,
> >>                         cap_ndoms(iommu->cap), i+1);         }
> >> @@ -1006,8 +1073,11 @@ static struct dmar_domain *
> >> iommu_alloc_domain(struct intel_iommu *iommu)
> >>
> >>         set_bit(num, iommu->domain_ids);
> >>         domain->id = num;
> >> -       domain->iommu = iommu;
> >> +       memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
> >> +       set_bit(iommu->seq_id, &domain->iommu_bmp);
> >>         iommu->domains[num] = domain;
> >> +       domain->iommu_count = 1;
> >> +       domain->flags = 0;
> >>         spin_unlock_irqrestore(&iommu->lock, flags);
> >>
> >>         return domain;
> >> @@ -1016,10 +1086,12 @@ static struct dmar_domain *
> >>  iommu_alloc_domain(struct intel_iommu *iommu) static void
> >>         iommu_free_domain(struct dmar_domain *domain)  { unsigned
> >> long flags; +       struct intel_iommu *iommu;
> >>
> >> -       spin_lock_irqsave(&domain->iommu->lock, flags);
> >> -       clear_bit(domain->id, domain->iommu->domain_ids);
> >> -       spin_unlock_irqrestore(&domain->iommu->lock, flags);
> >> +       iommu = domain_get_only_iommu(domain);
> >> +       spin_lock_irqsave(&iommu->lock, flags);
> >> +       clear_bit(domain->id, iommu->domain_ids);
> >> +       spin_unlock_irqrestore(&iommu->lock, flags);  }
> >>
> >>  static struct iova_domain reserved_iova_list;
> >> @@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain
> >>         *domain, int guest_width)
> >> domain_reserve_special_ranges(domain);
> >>
> >>         /* calculate AGAW */
> >> -       iommu = domain->iommu;
> >> +       iommu = domain_get_only_iommu(domain);
> >>         if (guest_width > cap_mgaw(iommu->cap))
> >>                 guest_width = cap_mgaw(iommu->cap);
> >>         domain->gaw = guest_width;
> >> @@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain
> >>         *domain, int guest_width) sagaw = cap_sagaw(iommu->cap);
> >>         if (!test_bit(agaw, &sagaw)) {
> >>                 /* hardware doesn't support it, choose a bigger one
> >> */ -               pr_debug("IOMMU: hardware doesn't support agaw
> >> %d\n", agaw); +               pr_debug("IOMMU: hardware doesn't
> >>                 support agaw %d\n", +                 agaw); agaw =
> >>                 find_next_bit(&sagaw, 5, agaw); if (agaw >= 5)
> >>                         return -ENODEV;
> >>         }
> >>         domain->agaw = agaw;
> >> +
> >>         INIT_LIST_HEAD(&domain->devices);
> >>
> >>         /* always allocate the top pgd */
> >>         domain->pgd = (struct dma_pte *)alloc_pgtable_page();
> >>                 if (!domain->pgd) return -ENOMEM;
> >> -       __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
> >> +       domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
> >>  return 0; }
> >>
> >> @@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain
> >> *domain)  }
> >>
> >>  static int domain_context_mapping_one(struct dmar_domain *domain,
> >> -               u8 bus, u8 devfn)
> >> +               struct intel_iommu *iommu, u8 bus, u8 devfn)  {
> >>         struct context_entry *context;
> >> -       struct intel_iommu *iommu = domain->iommu;
> >>         unsigned long flags;
> >>
> >>         pr_debug("Set context mapping for %02x:%02x.%d\n",
> >> @@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain
> >>         *domain, struct pci_dev *pdev)  { int ret;
> >>         struct pci_dev *tmp, *parent;
> >> +       struct intel_iommu *iommu;
> >> +
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return  -ENODEV;
> >>
> >> -       ret = domain_context_mapping_one(domain, pdev->bus->number,
> >> -               pdev->devfn);
> >> +       ret = domain_context_mapping_one(domain, iommu,
> >> +               pdev->bus->number, pdev->devfn);
> >>         if (ret)
> >>                 return ret;
> >>
> >> @@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain
> >>         *domain, struct pci_dev *pdev) /* Secondary interface's bus
> >>         number and devfn 0 */ parent = pdev->bus->self;
> >>         while (parent != tmp) {
> >> -               ret = domain_context_mapping_one(domain,
> >> parent->bus->number,
> >> -                       parent->devfn);
> >> +               ret = domain_context_mapping_one(domain, iommu,
> >> +                       parent->bus->number, parent->devfn);
> >>                         if (ret) return ret;
> >>                 parent = parent->bus->self;
> >>         }
> >>         if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
> >> -               return domain_context_mapping_one(domain,
> >> +               return domain_context_mapping_one(domain, iommu,
> >>                         tmp->subordinate->number, 0);
> >>         else /* this is a legacy PCI bridge */
> >> -               return domain_context_mapping_one(domain,
> >> +               return domain_context_mapping_one(domain, iommu,
> >>                         tmp->bus->number, tmp->devfn);  }
> >>
> >> -static int domain_context_mapped(struct dmar_domain *domain,
> >> -       struct pci_dev *pdev)
> >> +static int domain_context_mapped(struct pci_dev *pdev)  {
> >>         int ret;
> >>         struct pci_dev *tmp, *parent;
> >> +       struct intel_iommu *iommu;
> >>
> >> -       ret = device_context_mapped(domain->iommu,
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return 0;
> >> +
> >> +       ret = device_context_mapped(iommu,
> >>                 pdev->bus->number, pdev->devfn);
> >>         if (!ret)
> >>                 return ret;
> >> @@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct
> >>         dmar_domain *domain, /* Secondary interface's bus number and
> >>         devfn 0 */ parent = pdev->bus->self;
> >>         while (parent != tmp) {
> >> -               ret = device_context_mapped(domain->iommu,
> >> parent->bus->number, +               ret =
> >>                         device_context_mapped(iommu,
> >>                 parent->bus->number, parent->devfn); if (!ret)
> >>                         return ret;
> >>                 parent = parent->bus->self;
> >>         }
> >>         if (tmp->is_pcie)
> >> -               return device_context_mapped(domain->iommu,
> >> +               return device_context_mapped(iommu,
> >>                         tmp->subordinate->number, 0);         else
> >> -               return device_context_mapped(domain->iommu,
> >> +               return device_context_mapped(iommu,
> >>                         tmp->bus->number, tmp->devfn);  }
> >>
> >> @@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain
> >>                 *domain, dma_addr_t iova, BUG_ON(dma_pte_addr(*pte));
> >>                 dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
> >>                 dma_set_pte_prot(*pte, prot);
> >> -               __iommu_flush_cache(domain->iommu, pte,
> >> sizeof(*pte)); +               domain_flush_cache(domain, pte,
> >>                 sizeof(*pte));                 start_pfn++; index++;
> >>         }
> >>         return 0;
> >>  }
> >>
> >> -static void detach_domain_for_dev(struct dmar_domain *domain, u8
> >> bus, u8 devfn) +static void detach_domain_for_dev(struct dmar_domain
> >> *domain, +                                 u8 bus, u8 devfn)  {
> >> -       clear_context_table(domain->iommu, bus, devfn);
> >> -       domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
> >> -                                          DMA_CCMD_GLOBAL_INVL, 0);
> >> -       domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
> >> -                                        DMA_TLB_GLOBAL_FLUSH, 0);
> >> +       struct intel_iommu *iommu;
> >> +
> >> +       iommu = device_find_matched_iommu(bus, devfn); +       if
> >> (!iommu) +               return;
> >> +
> >> +       clear_context_table(iommu, bus, devfn);
> >> +       iommu->flush.flush_context(iommu, 0, 0, 0,
> >> +                                  DMA_CCMD_GLOBAL_INVL, 0);
> >> +       iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> >> +                                DMA_TLB_GLOBAL_FLUSH, 0);  }
> >>
> >>  static void domain_remove_dev_info(struct dmar_domain *domain)
> >> @@ -1336,7 +1425,6 @@ static struct dmar_domain
> >>         *get_domain_for_dev(struct pci_dev *pdev, int gaw)  { struct
> >>         dmar_domain *domain, *found = NULL; struct intel_iommu
> >> *iommu; -       struct dmar_drhd_unit *drhd;
> >>         struct device_domain_info *info, *tmp;
> >>         struct pci_dev *dev_tmp;
> >>         unsigned long flags;
> >> @@ -1371,13 +1459,9 @@ static struct dmar_domain
> >> *get_domain_for_dev(struct pci_dev *pdev, int gaw)         }
> >>
> >>         /* Allocate new domain for the device */
> >> -       drhd = dmar_find_matched_drhd_unit(pdev);
> >> -       if (!drhd) {
> >> -               printk(KERN_ERR "IOMMU: can't find DMAR for device
> >> %s\n",
> >> -                       pci_name(pdev));
> >> -               return NULL;
> >> -       }
> >> -       iommu = drhd->iommu;
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return  NULL;
> >>
> >>         domain = iommu_alloc_domain(iommu);
> >>         if (!domain)
> >> @@ -1400,7 +1484,7 @@ static struct dmar_domain
> >>                 *get_domain_for_dev(struct pci_dev *pdev, int gaw)
> >>                 info->dev = NULL; info->domain = domain;
> >>                 /* This domain is shared by devices under p2p bridge
> >> */ -               domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
> >> +               domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
> >>
> >>                 /* pcie-to-pci bridge already has a domain, uses it
> >> */                 found = NULL; @@ -1805,7 +1889,7 @@
> >> get_valid_domain_for_dev(struct pci_dev *pdev)         }
> >>
> >>         /* make sure context mapping is ok */
> >> -       if (unlikely(!domain_context_mapped(domain, pdev))) {
> >> +       if (unlikely(!domain_context_mapped(pdev))) {
> >>                 ret = domain_context_mapping(domain, pdev);
> >>                         if (ret) { printk(KERN_ERR
> >> @@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct
> >>         device *hwdev, phys_addr_t paddr,  { struct pci_dev *pdev =
> >>         to_pci_dev(hwdev); struct dmar_domain *domain;
> >> +       struct intel_iommu *iommu;
> >>         phys_addr_t start_paddr;
> >>         struct iova *iova;
> >>         int prot = 0;
> >> @@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct
> >>                 device *hwdev, phys_addr_t paddr,         if
> >> (!domain) return 0;
> >>
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return 0;
> >> +
> >>         size = aligned_size((u64)paddr, size);
> >>
> >>         iova = __intel_alloc_iova(hwdev, domain, size,
> >> pdev->dma_mask); @@ -1849,7 +1938,7 @@ static dma_addr_t
> >> __intel_map_single(struct device *hwdev, phys_addr_t paddr,
> >>          * mappings..
> >>          */
> >>         if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
> >> -                       !cap_zlr(domain->iommu->cap))
> >> +                       !cap_zlr(iommu->cap))
> >>                 prot |= DMA_PTE_READ;
> >>         if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
> >>                 prot |= DMA_PTE_WRITE;
> >> @@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct
> >> device *hwdev, phys_addr_t paddr,                 goto error;
> >>
> >>         /* it's a non-present to present mapping */
> >> -       ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
> >> +       ret = iommu_flush_iotlb_psi(iommu, domain->id,
> >>                         start_paddr, size >> VTD_PAGE_SHIFT, 1);
> >> if (ret) -               iommu_flush_write_buffer(domain->iommu);
> >> +               iommu_flush_write_buffer(iommu);
> >>
> >>         return start_paddr + ((u64)paddr & (~PAGE_MASK));
> >>
> >> @@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
> >>         /* just flush them all */
> >>         for (i = 0; i < g_num_of_iommus; i++) {
> >>                 if (deferred_flush[i].next) {
> >> -                       struct intel_iommu *iommu =
> >> -                               deferred_flush[i].domain[0]->iommu;
> >> +                       struct intel_iommu *iommu =
> >> deferred_flush[i].iommu;
> >>
> >>                         iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> >>
> >> DMA_TLB_GLOBAL_FLUSH, 0); @@ -1921,7 +2009,8 @@ static void
> >>         flush_unmaps_timeout(unsigned long data)
> >> spin_unlock_irqrestore(&async_umap_flush_lock, flags);  }
> >>
> >> -static void add_unmap(struct dmar_domain *dom, struct iova *iova)
> >> +static void add_unmap(struct dmar_domain *dom,
> >> +                     struct intel_iommu *iommu, struct iova *iova)
> >>         { unsigned long flags;
> >>         int next, iommu_id;
> >> @@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain
> >>         *dom, struct iova *iova) if (list_size == HIGH_WATER_MARK)
> >>                 flush_unmaps();
> >>
> >> -       iommu_id = dom->iommu->seq_id;
> >> +       iommu_id = iommu->seq_id;
> >>
> >>         next = deferred_flush[iommu_id].next;
> >>         deferred_flush[iommu_id].domain[next] = dom;
> >>         deferred_flush[iommu_id].iova[next] = iova;
> >> +       deferred_flush[iommu_id].iommu = iommu;
> >>         deferred_flush[iommu_id].next++;
> >>
> >>         if (!timer_on) {
> >> @@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev,
> >>         dma_addr_t dev_addr, size_t size, struct dmar_domain *domain;
> >>         unsigned long start_addr;
> >>         struct iova *iova;
> >> +       struct intel_iommu *iommu;
> >>
> >>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> >>         return; domain = find_domain(pdev);
> >>         BUG_ON(!domain);
> >>
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return;
> >> +
> >>         iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
> >>                 if (!iova) return;
> >> @@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev,
> >>         dma_addr_t dev_addr, size_t size, /* free page tables */
> >>         dma_pte_free_pagetable(domain, start_addr, start_addr +
> >>         size); if (intel_iommu_strict) {
> >> -               if (iommu_flush_iotlb_psi(domain->iommu,
> >> +               if (iommu_flush_iotlb_psi(iommu,
> >>                         domain->id, start_addr, size >>
> >> VTD_PAGE_SHIFT, 0)) -
> >> iommu_flush_write_buffer(domain->iommu); +
> >>                 iommu_flush_write_buffer(iommu); /* free iova */
> >>                 __free_iova(&domain->iovad, iova);
> >>         } else {
> >> -               add_unmap(domain, iova);
> >> +               add_unmap(domain, iommu, iova);
> >>                 /*
> >>                  * queue up the release of the unmap to save the
> >> 1/6th of the
> >>                  * cpu used up by the iotlb flush operation...
> >> @@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev,
> >>         struct scatterlist *sglist,         size_t size = 0; void
> >>         *addr; struct scatterlist *sg;
> >> +       struct intel_iommu *iommu;
> >>
> >>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> >> return;
> >>
> >>         domain = find_domain(pdev);
> >>
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return;
> >> +
> >>         iova = find_iova(&domain->iovad,
> >>                 IOVA_PFN(sglist[0].dma_address));         if (!iova)
> >> return; @@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device
> >>         *hwdev, struct scatterlist *sglist, /* free page tables */
> >>         dma_pte_free_pagetable(domain, start_addr, start_addr +
> >> size);
> >>
> >> -       if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
> >> start_addr, +       if (iommu_flush_iotlb_psi(iommu, domain->id,
> >>                         start_addr, size >> VTD_PAGE_SHIFT, 0))
> >> -               iommu_flush_write_buffer(domain->iommu);
> >> +               iommu_flush_write_buffer(iommu);
> >>
> >>         /* free iova */
> >>         __free_iova(&domain->iovad, iova);
> >> @@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct
> >>         scatterlist *sglist, int nelems,         int ret; struct
> >>         scatterlist *sg; unsigned long start_addr;
> >> +       struct intel_iommu *iommu;
> >>
> >>         BUG_ON(dir == DMA_NONE);
> >>         if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> >> @@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct
> >>                 scatterlist *sglist, int nelems,         if
> >> (!domain) return 0;
> >>
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return 0;
> >> +
> >>         for_each_sg(sglist, sg, nelems, i) {
> >>                 addr = SG_ENT_VIRT_ADDRESS(sg);
> >>                 addr = (void *)virt_to_phys(addr);
> >> @@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct
> >> scatterlist *sglist, int nelems,
> >>          * mappings..
> >>          */
> >>         if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
> >> -                       !cap_zlr(domain->iommu->cap))
> >> +                       !cap_zlr(iommu->cap))
> >>                 prot |= DMA_PTE_READ;
> >>         if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
> >>                 prot |= DMA_PTE_WRITE;
> >> @@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct
> >> scatterlist *sglist, int nelems,         }
> >>
> >>         /* it's a non-present to present mapping */
> >> -       if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
> >> +       if (iommu_flush_iotlb_psi(iommu, domain->id,
> >>                         start_addr, offset >> VTD_PAGE_SHIFT, 1))
> >> -               iommu_flush_write_buffer(domain->iommu);
> >> +               iommu_flush_write_buffer(iommu);
> >>         return nelems;
> >>  }
> >>
> >> @@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
> >>  return 0; }
> >>
> >> -void intel_iommu_domain_exit(struct dmar_domain *domain)
> >> +/* domain id for virtual machine, it won't be set in context */
> >> +static unsigned long vm_domid;
> >> +
> >> +static int vm_domain_min_agaw(struct dmar_domain *domain) +{
> >> +       struct dmar_drhd_unit *drhd;
> >> +       struct intel_iommu *iommu;
> >> +       int min_agaw = domain->agaw;
> >> +
> >> +       for_each_drhd_unit(drhd) {
> >> +               if (drhd->ignored)
> >> +                       continue;
> >> +               iommu = drhd->iommu;
> >> +
> >> +               if (test_bit(iommu->seq_id, &domain->iommu_bmp))
> >> +                       if (min_agaw > iommu->agaw)
> >> +                               min_agaw = iommu->agaw; +       }
> >> +
> >> +       return min_agaw;
> >> +}
> >> +
> >> +static int vm_domain_add_dev_info(struct dmar_domain *domain,
> >> +                                 struct pci_dev *pdev) +{
> >> +       struct device_domain_info *info;
> >> +       unsigned long flags;
> >> +
> >> +       info = alloc_devinfo_mem();
> >> +       if (!info)
> >> +               return -ENOMEM;
> >> +
> >> +       info->bus = pdev->bus->number;
> >> +       info->devfn = pdev->devfn;
> >> +       info->dev = pdev;
> >> +       info->domain = domain;
> >> +
> >> +       spin_lock_irqsave(&device_domain_lock, flags);
> >> +       list_add(&info->link, &domain->devices);
> >> +       list_add(&info->global, &device_domain_list);
> >> +       pdev->dev.archdata.iommu = info;
> >> +       spin_unlock_irqrestore(&device_domain_lock, flags); +
> >> +       return 0;
> >> +}
> >> +
> >> +static void vm_domain_remove_one_dev_info(struct dmar_domain
> >> *domain, +                                         struct pci_dev
> >> *pdev) +{ +       struct device_domain_info *info;
> >> +       struct intel_iommu *iommu;
> >> +       unsigned long flags;
> >> +       int found = 0;
> >> +
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + +       spin_lock_irqsave(&device_domain_lock,
> >> flags); +       while (!list_empty(&domain->devices)) {
> >> +               info = list_entry(domain->devices.next,
> >> +                       struct device_domain_info, link);
> >> +               if (info->bus == pdev->bus->number &&
> >> +                   info->devfn == pdev->devfn) {
> >> +                       list_del(&info->link);
> >> +                       list_del(&info->global);
> >> +                       if (info->dev)
> >> +                               info->dev->dev.archdata.iommu = NULL;
> >> +                       spin_unlock_irqrestore(&device_domain_lock,
> >> flags); + +                       detach_domain_for_dev(info->domain,
> >> +                                             info->bus,
> >> info->devfn); +                       free_devinfo_mem(info);
> >> +
> >> +                       spin_lock_irqsave(&device_domain_lock,
> >> flags); + +                       if (found)
> >> +                               break;
> >> +                       else
> >> +                               continue;
> >> +               }
> >> +
> >> +               /* if there is no other devices under the same iommu
> >> +                * owned by this domain, clear this iommu in
> >> iommu_bmp +                */ +               if
> >> (device_find_matched_iommu(info->bus, info->devfn) == iommu) +
> >> found = 1; +       }
> >> +
> >> +       if (found == 0) {
> >> +               spin_lock_irqsave(&iommu->lock, flags);
> >> +               clear_bit(iommu->seq_id, &domain->iommu_bmp);
> >> +               domain->iommu_count--;
> >> +               spin_unlock_irqrestore(&iommu->lock, flags); +
> >> } +
> >> +       spin_unlock_irqrestore(&device_domain_lock, flags); +}
> >> +
> >> +static void vm_domain_remove_all_dev_info(struct dmar_domain
> >> *domain) +{ +       struct device_domain_info *info;
> >> +       struct intel_iommu *iommu;
> >> +       unsigned long flags;
> >> +
> >> +       spin_lock_irqsave(&device_domain_lock, flags);
> >> +       while (!list_empty(&domain->devices)) {
> >> +               info = list_entry(domain->devices.next,
> >> +                       struct device_domain_info, link);
> >> +               list_del(&info->link);
> >> +               list_del(&info->global);
> >> +               if (info->dev)
> >> +                       info->dev->dev.archdata.iommu = NULL; +
> >> +               spin_unlock_irqrestore(&device_domain_lock, flags);
> >> +               detach_domain_for_dev(info->domain,
> >> +                                     info->bus, info->devfn); +
> >> +               /* clear this iommu in iommu_bmp */
> >> +               iommu = device_find_matched_iommu(info->bus,
> >> info->devfn); +               spin_lock_irqsave(&iommu->lock, flags);
> >> +               if (test_and_clear_bit(iommu->seq_id,
> >> +                                      &domain->iommu_bmp))
> >> +                       domain->iommu_count--;
> >> +               spin_unlock_irqrestore(&iommu->lock, flags); +
> >> +               free_devinfo_mem(info);
> >> +               spin_lock_irqsave(&device_domain_lock, flags); +
> >> } +       spin_unlock_irqrestore(&device_domain_lock, flags); +}
> >> +
> >> +static int vm_domain_context_mapping_one(struct dmar_domain *domain,
> >> +               struct intel_iommu *iommu, u8 bus, u8 devfn) +{
> >> +       struct context_entry *context;
> >> +       unsigned long flags;
> >> +       struct dma_pte *pgd;
> >> +       unsigned long num;
> >> +       unsigned long ndomains;
> >> +       int id;
> >> +       int agaw;
> >> +       int found = 0;
> >> +
> >> +       pr_debug("Set context mapping for %02x:%02x.%d\n",
> >> +               bus, PCI_SLOT(devfn), PCI_FUNC(devfn)); +
> >> BUG_ON(!domain->pgd); +       context =
> >> device_to_context_entry(iommu, bus, devfn); +       if (!context) +
> >> return -ENOMEM; +       spin_lock_irqsave(&iommu->lock, flags);
> >> +       if (context_present(*context)) {
> >> +               spin_unlock_irqrestore(&iommu->lock, flags); +
> >> return 0; +       }
> >> +
> >> +       id = domain->id;
> >> +
> >> +       /* find an available domain id for this device in iommu */
> >> +       ndomains = cap_ndoms(iommu->cap);
> >> +       num = find_first_bit(iommu->domain_ids, ndomains);
> >> +       for (; num < ndomains; ) {
> >> +               if (iommu->domains[num] == domain) {
> >> +                       id = num;
> >> +                       found = 1;
> >> +                       break;
> >> +               }
> >> +               num = find_next_bit(iommu->domain_ids,
> >> +                 cap_ndoms(iommu->cap), num+1);
> >> +       }
> >> +
> >> +       if (found == 0) {
> >> +               num = find_first_zero_bit(iommu->domain_ids,
> >> ndomains); +               if (num >= ndomains) {
> >> +                       spin_unlock_irqrestore(&iommu->lock, flags);
> >> +                       printk(KERN_ERR "IOMMU: no free domain
> >> ids\n"); +                       return -EFAULT;
> >> +               }
> >> +
> >> +               set_bit(num, iommu->domain_ids);
> >> +               iommu->domains[num] = domain;
> >> +               id = num;
> >> +       }
> >> +
> >> +       pgd = domain->pgd;
> >> +
> >> +       /* Skip top levels of page tables for
> >> +        * iommu which has less agaw than default.
> >> +        */
> >> +       for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
> >> +               pgd = phys_to_virt(dma_pte_addr(*pgd));
> >> +               if (!dma_pte_present(*pgd)) {
> >> +                       spin_unlock_irqrestore(&iommu->lock, flags);
> >> +                       return -ENOMEM;
> >> +               }
> >> +       }
> >> +
> >> +       context_set_domain_id(*context, id);
> >> +       context_set_address_width(*context, iommu->agaw);
> >> +       context_set_address_root(*context, virt_to_phys(pgd));
> >> +       context_set_translation_type(*context,
> >> CONTEXT_TT_MULTI_LEVEL); +       context_set_fault_enable(*context);
> >> +       context_set_present(*context);
> >> +       __iommu_flush_cache(iommu, context, sizeof(*context)); +
> >> +       /* it's a non-present to present mapping */
> >> +       if (iommu->flush.flush_context(iommu, id,
> >> +               (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
> >> +               DMA_CCMD_DEVICE_INVL, 1))
> >> +               iommu_flush_write_buffer(iommu);
> >> +       else
> >> +               iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> >> DMA_TLB_DSI_FLUSH, 0); + +       if
> >> (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp)) +
> >> domain->iommu_count++; +       spin_unlock_irqrestore(&iommu->lock,
> >> flags); +       return 0; +}
> >> +
> >> +static int
> >> +vm_domain_context_mapping(struct dmar_domain *domain, struct
> >> pci_dev *pdev) +{ +       int ret;
> >> +       struct pci_dev *tmp, *parent;
> >> +       struct intel_iommu *iommu;
> >> +
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return  -ENODEV;
> >> +
> >> +       ret = vm_domain_context_mapping_one(domain, iommu,
> >> +               pdev->bus->number, pdev->devfn);
> >> +       if (ret)
> >> +               return ret;
> >> +
> >> +       /* dependent device mapping */
> >> +       tmp = pci_find_upstream_pcie_bridge(pdev);
> >> +       if (!tmp)
> >> +               return 0;
> >> +       /* Secondary interface's bus number and devfn 0 */
> >> +       parent = pdev->bus->self;
> >> +       while (parent != tmp) {
> >> +               ret = vm_domain_context_mapping_one(domain, iommu,
> >> +                       parent->bus->number, parent->devfn); +
> >> if (ret) +                       return ret;
> >> +               parent = parent->bus->self;
> >> +       }
> >> +       if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
> >> +               return vm_domain_context_mapping_one(domain, iommu,
> >> +                       tmp->subordinate->number, 0);
> >> +       else /* this is a legacy PCI bridge */
> >> +               return vm_domain_context_mapping_one(domain, iommu,
> >> +                       tmp->bus->number, tmp->devfn); +}
> >> +
> >> +
> >> +static int vm_domain_init(struct dmar_domain *domain, int
> >> guest_width) +{ +       int adjust_width;
> >> +
> >> +       init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
> >> +       spin_lock_init(&domain->mapping_lock);
> >> +
> >> +       domain_reserve_special_ranges(domain);
> >> +
> >> +       /* calculate AGAW */
> >> +       domain->gaw = guest_width;
> >> +       adjust_width = guestwidth_to_adjustwidth(guest_width);
> >> +       domain->agaw = width_to_agaw(adjust_width); +
> >> +       INIT_LIST_HEAD(&domain->devices);
> >> +
> >> +       /* always allocate the top pgd */
> >> +       domain->pgd = (struct dma_pte *)alloc_pgtable_page(); +
> >> if (!domain->pgd) +               return -ENOMEM;
> >> +       domain_flush_cache(domain, domain->pgd, PAGE_SIZE); +
> >> return 0; +}
> >> +
> >> +static void iommu_free_vm_domain(struct dmar_domain *domain) +{
> >> +       unsigned long flags;
> >> +       struct dmar_drhd_unit *drhd;
> >> +       struct intel_iommu *iommu;
> >> +       unsigned long i;
> >> +       unsigned long ndomains;
> >> +
> >> +       for_each_drhd_unit(drhd) {
> >> +               if (drhd->ignored)
> >> +                       continue;
> >> +               iommu = drhd->iommu;
> >> +
> >> +               ndomains = cap_ndoms(iommu->cap);
> >> +               i = find_first_bit(iommu->domain_ids, ndomains);
> >> +               for (; i < ndomains; ) {
> >> +                       if (iommu->domains[i] == domain) {
> >> +                               spin_lock_irqsave(&iommu->lock,
> >> flags); +                               clear_bit(i,
> >> iommu->domain_ids); +
> >> iommu->domains[i] = NULL; +
> >> spin_unlock_irqrestore(&iommu->lock, flags); +
> >> break; +                       }
> >> +                       i = find_next_bit(iommu->domain_ids,
> >> ndomains, i+1); +               } +       }
> >> +}
> >> +
> >> +static void vm_domain_exit(struct dmar_domain *domain)  {
> >>         u64 end;
> >>
> >> @@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct
> >>                 dmar_domain *domain)         if (!domain) return;
> >>
> >> +       vm_domain_remove_all_dev_info(domain);
> >> +       /* destroy iovas */
> >> +       put_iova_domain(&domain->iovad);
> >>         end = DOMAIN_MAX_ADDR(domain->gaw);
> >> -       end = end & (~VTD_PAGE_MASK);
> >> +       end &= VTD_PAGE_MASK;
> >>
> >>         /* clear ptes */
> >>         dma_pte_clear_range(domain, 0, end);
> >> @@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct
> >>         dmar_domain *domain) /* free page tables */
> >>         dma_pte_free_pagetable(domain, 0, end);
> >>
> >> -       iommu_free_domain(domain);
> >> +       iommu_free_vm_domain(domain);
> >>         free_domain_mem(domain);
> >>  }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
> >>
> >> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
> >> +static struct dmar_domain *iommu_alloc_vm_domain(void)  {
> >> -       struct dmar_drhd_unit *drhd;
> >>         struct dmar_domain *domain;
> >> -       struct intel_iommu *iommu;
> >>
> >> -       drhd = dmar_find_matched_drhd_unit(pdev);
> >> -       if (!drhd) {
> >> -               printk(KERN_ERR "intel_iommu_domain_alloc: drhd ==
> >> NULL\n"); +       domain = alloc_domain_mem();
> >> +       if (!domain)
> >>                 return NULL;
> >> -       }
> >>
> >> -       iommu = drhd->iommu;
> >> -       if (!iommu) {
> >> -               printk(KERN_ERR
> >> -                       "intel_iommu_domain_alloc: iommu == NULL\n");
> >> -               return NULL;
> >> -       }
> >> -       domain = iommu_alloc_domain(iommu);
> >> +       domain->id = vm_domid++;
> >> +       domain->iommu_count = 0;
> >> +       domain->max_addr = 0;
> >> +       memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
> >> +       domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE; +
> >> +       return domain;
> >> +}
> >> +
> >> +struct dmar_domain *intel_iommu_alloc_domain(void) +{
> >> +       struct dmar_domain *domain;
> >> +
> >> +       domain = iommu_alloc_vm_domain();
> >>         if (!domain) {
> >>                 printk(KERN_ERR
> >>                         "intel_iommu_domain_alloc: domain ==
> >>         NULL\n");                 return NULL; }
> >> -       if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
> >> +       if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
> >>                 printk(KERN_ERR
> >>                         "intel_iommu_domain_alloc: domain_init()
> >> failed\n"); -               intel_iommu_domain_exit(domain);
> >> +               vm_domain_exit(domain);
> >>                 return NULL;
> >>         }
> >> +
> >>         return domain;
> >>  }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
> >> +
> >> +void intel_iommu_free_domain(struct dmar_domain *domain) +{
> >> +       vm_domain_exit(domain);
> >> +}
> >> +EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
> >>
> >> -int intel_iommu_context_mapping(
> >> -       struct dmar_domain *domain, struct pci_dev *pdev)
> >> +int intel_iommu_assign_device(struct dmar_domain *domain,
> >> +                             struct pci_dev *pdev)  {
> >> -       int rc;
> >> -       rc = domain_context_mapping(domain, pdev);
> >> -       return rc;
> >> +       struct intel_iommu *iommu;
> >> +       int addr_width;
> >> +       u64 end;
> >> +       int ret;
> >> +
> >> +       /* normally pdev is not mapped */
> >> +       if (unlikely(domain_context_mapped(pdev))) {
> >> +               struct dmar_domain *old_domain;
> >> +
> >> +               old_domain = find_domain(pdev);
> >> +               if (old_domain) {
> >> +                       if (domain->flags &
> >> DOMAIN_FLAG_VIRTUAL_MACHINE) +
> >> vm_domain_remove_one_dev_info(old_domain, pdev); +
> >> else +
> >> domain_remove_dev_info(old_domain); +               } +       } +
> >> +       iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); +       if (!iommu) +               return  -ENODEV;
> >> +
> >> +       /* check if this iommu agaw is sufficient for max mapped
> >> address */ +       addr_width = agaw_to_width(iommu->agaw);
> >> +       end = DOMAIN_MAX_ADDR(addr_width);
> >> +       end = end & VTD_PAGE_MASK;
> >> +       if (end < domain->max_addr) {
> >> +               printk(KERN_ERR "%s: iommu agaw (%d) is not "
> >> +                      "sufficient for the mapped address (%llx)\n",
> >> +                      __func__, iommu->agaw, domain->max_addr);
> >> +               return -EFAULT;
> >> +       }
> >> +
> >> +       ret = vm_domain_context_mapping(domain, pdev); +       if
> >> (ret) +               return ret;
> >> +
> >> +       ret = vm_domain_add_dev_info(domain, pdev); +       return
> >>  ret; }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
> >>
> >> -int intel_iommu_page_mapping(
> >> -       struct dmar_domain *domain, dma_addr_t iova,
> >> -       u64 hpa, size_t size, int prot)
> >> +
> >> +void intel_iommu_deassign_device(struct dmar_domain *domain,
> >> +                                struct pci_dev *pdev)  {
> >> -       int rc;
> >> -       rc = domain_page_mapping(domain, iova, hpa, size, prot);
> >> -       return rc;
> >> +       vm_domain_remove_one_dev_info(domain, pdev);  }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
> >>
> >> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8
> >> devfn) +int intel_iommu_map_pages(struct dmar_domain *domain,
> >> dma_addr_t iova, +                         u64 hpa, size_t size, int
> >> prot)  { -       detach_domain_for_dev(domain, bus, devfn);
> >> +       u64 max_addr;
> >> +       int addr_width;
> >> +       int ret;
> >> +
> >> +       max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
> >> +       if (domain->max_addr < max_addr) {
> >> +               int min_agaw;
> >> +               u64 end;
> >> +
> >> +               /* check if minimum agaw is sufficient for mapped
> >> address */ +               min_agaw = vm_domain_min_agaw(domain);
> >> +               addr_width = agaw_to_width(min_agaw);
> >> +               end = DOMAIN_MAX_ADDR(addr_width);
> >> +               end = end & VTD_PAGE_MASK;
> >> +               if (end < max_addr) {
> >> +                       printk(KERN_ERR "%s: iommu agaw (%d) is not "
> >> +                              "sufficient for the mapped address
> >> (%llx)\n", +                              __func__, min_agaw,
> >> max_addr); +                       return -EFAULT;
> >> +               }
> >> +               domain->max_addr = max_addr;
> >> +       }
> >> +
> >> +       ret = domain_page_mapping(domain, iova, hpa, size, prot); +
> >>  return ret; }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
> >>
> >> -struct dmar_domain *
> >> -intel_iommu_find_domain(struct pci_dev *pdev)
> >> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
> >> +                            dma_addr_t iova, size_t size)  {
> >> -       return find_domain(pdev);
> >> +       dma_addr_t base;
> >> +
> >> +       /* The address might not be aligned */
> >> +       base = iova & PAGE_MASK;
> >> +       size = PAGE_ALIGN(size);
> >> +       dma_pte_clear_range(domain, base, base + size);  }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
> >>
> >>  int intel_iommu_found(void)
> >>  {
> >> diff --git a/include/linux/dma_remapping.h
> >> b/include/linux/dma_remapping.h
> >> index 952df39..3568ae6 100644
> >> --- a/include/linux/dma_remapping.h
> >> +++ b/include/linux/dma_remapping.h
> >> @@ -111,11 +111,21 @@ struct dma_pte {
> >>                 (p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
> >>  #define dma_pte_present(p) (((p).val & 3) != 0)
> >>
> >> +/* domain flags, one domain owns one device by default */ +
> >> +/* devices under the same p2p bridge are owned in one domain */
> >> +#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES       (1 << 0) +
> >> +/* domain represents a virtual machine, more than one devices
> >> + * across iommus may be owned in one domain, e.g. kvm guest. + */
> >> +#define DOMAIN_FLAG_VIRTUAL_MACHINE            (1 << 1) +
> >>  struct intel_iommu;
> >>
> >>  struct dmar_domain {
> >>         int     id;                     /* domain id */
> >> -       struct intel_iommu *iommu;      /* back pointer to owning
> >> iommu */ +       unsigned long iommu_bmp;        /* bitmap of iommus
> >> this domain uses*/
> >>
> >>         struct list_head devices;       /* all devices' list */
> >>         struct iova_domain iovad;       /* iova's that belong to
> >> this domain */ @@ -123,12 +133,13 @@ struct dmar_domain {
> >>         struct dma_pte  *pgd;           /* virtual address */
> >>         spinlock_t      mapping_lock;   /* page table lock */
> >>         int             gaw;            /* max guest address width */
> >> +       int             agaw;           /* adjusted guest address
> >> width */
> >>
> >> -       /* adjusted guest address width, 0 is level 2 30-bit */
> >> -       int             agaw;
> >> +       int             flags;          /* domain flag */
> >>
> >> -#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
> >> -       int             flags;
> >> +       /* following fields are used in virtual machine case */
> >> +       int             iommu_count;    /* reference count of iommu
> >> */ +       u64             max_addr;       /* maximum mapped address
> >> */  };
> >>
> >>  /* PCI domain-device relationship */
> >> diff --git a/include/linux/intel-iommu.h
> >> b/include/linux/intel-iommu.h
> >> index 3d017cf..c2f37b8 100644
> >> --- a/include/linux/intel-iommu.h
> >> +++ b/include/linux/intel-iommu.h
> >> @@ -219,6 +219,8 @@ do {
> >>         \ }
> >> \  } while (0)
> >>
> >> +#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
> >> +
> >>  #define QI_LENGTH      256     /* queue length */
> >>
> >>  enum {
> >> @@ -299,6 +301,7 @@ struct intel_iommu {
> >>         struct dmar_domain **domains; /* ptr to domains */
> >>         spinlock_t      lock; /* protect context, domain ids */
> >>         struct root_entry *root_entry; /* virtual address */ +
> >> int             agaw;
> >>
> >>         unsigned int irq;
> >>         unsigned char name[7];    /* Device Name */
> >> @@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu
> >> *iommu, u16 did, u64 addr,
> >>
> >>  extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu
> >> *iommu);
> >>
> >> -void intel_iommu_domain_exit(struct dmar_domain *domain);
> >> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
> >> -int intel_iommu_context_mapping(struct dmar_domain *domain,
> >> -                               struct pci_dev *pdev);
> >> -int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t
> >> iova,
> >> -                            u64 hpa, size_t size, int prot);
> >> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8
> >> devfn);
> >> -struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
> >> +struct dmar_domain *intel_iommu_alloc_domain(void);
> >> +void intel_iommu_free_domain(struct dmar_domain *domain);
> >> +int intel_iommu_assign_device(struct dmar_domain *domain,
> >> +                             struct pci_dev *pdev);
> >> +void intel_iommu_deassign_device(struct dmar_domain *domain,
> >> +                                struct pci_dev *pdev);
> >> +int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t
> >> iova, +                         u64 hpa, size_t size, int prot);
> >> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
> >> +                            dma_addr_t iova, size_t size);
> >>  u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
> >>
> >>  #ifdef CONFIG_DMAR
> >> --
> >> 1.5.1
> 
> 

-- 
           |           AMD Saxony Limited Liability Company & Co. KG
 Operating |         Wilschdorfer Landstr. 101, 01109 Dresden, Germany
 System    |                  Register Court Dresden: HRA 4896
 Research  |              General Partner authorized to represent:
 Center    |             AMD Saxony LLC (Wilmington, Delaware, US)
           | General Manager of AMD Saxony LLC: Dr. Hans-R. Deppe, Thomas McCoy


^ permalink raw reply	[flat|nested] 7+ messages in thread

* RE: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
  2008-12-01 11:55     ` 'Joerg Roedel'
@ 2008-12-01 13:22       ` Han, Weidong
  2008-12-01 14:04         ` Joerg Roedel
  0 siblings, 1 reply; 7+ messages in thread
From: Han, Weidong @ 2008-12-01 13:22 UTC (permalink / raw)
  To: 'Joerg Roedel'
  Cc: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity',
	Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
	'iommu@lists.linux-foundation.org'

Sorry, this patch has style problem. I will update it and also split it to smaller patches for easy reviewing.

Regards,
Weidong

'Joerg Roedel' wrote:
> Hmm, I get these errors using git-am:
> 
> Applying VT-d: Support multiple device assignment for KVM
> .dotest/patch:1344: space before tab in indent.
>                                 clflush_cache_range(addr, size);
> .dotest/patch:1350: space before tab in indent.
>                         clflush_cache_range(addr, size);
> .dotest/patch:1907: trailing whitespace.
> 
> .dotest/patch:1946: trailing whitespace.
>                  * owned by this domain, clear this iommu in iommu_bmp
> .dotest/patch:2300: trailing whitespace.
> 
> error: patch failed: drivers/pci/dmar.c:484
> error: drivers/pci/dmar.c: patch does not apply
> error: patch failed: drivers/pci/intel-iommu.c:50
> error: drivers/pci/intel-iommu.c: patch does not apply
> error: patch failed: include/linux/dma_remapping.h:111
> error: include/linux/dma_remapping.h: patch does not apply
> error: patch failed: include/linux/intel-iommu.h:219
> error: include/linux/intel-iommu.h: patch does not apply
> Patch failed at 0001.
> 
> Joerg
> 

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
  2008-12-01 13:22       ` Han, Weidong
@ 2008-12-01 14:04         ` Joerg Roedel
  2008-12-02  0:31           ` Han, Weidong
  0 siblings, 1 reply; 7+ messages in thread
From: Joerg Roedel @ 2008-12-01 14:04 UTC (permalink / raw)
  To: Han, Weidong
  Cc: 'Joerg Roedel', Woodhouse, David, 'Jesse Barnes',
	'Avi Kivity', Kay, Allen M, Yu, Fenghua,
	'kvm@vger.kernel.org',
	'iommu@lists.linux-foundation.org'

Ok, I got them to apply. I also did the checkpatch cleanups. To speed
things up a bit I would suggest that I rebase my patchset on your
patches and send it out in a single series. Any problems with this
approach?

Joerg

On Mon, Dec 01, 2008 at 09:22:42PM +0800, Han, Weidong wrote:
> Sorry, this patch has style problem. I will update it and also split it to smaller patches for easy reviewing.
> 
> Regards,
> Weidong
> 
> 'Joerg Roedel' wrote:
> > Hmm, I get these errors using git-am:
> > 
> > Applying VT-d: Support multiple device assignment for KVM
> > .dotest/patch:1344: space before tab in indent.
> >                                 clflush_cache_range(addr, size);
> > .dotest/patch:1350: space before tab in indent.
> >                         clflush_cache_range(addr, size);
> > .dotest/patch:1907: trailing whitespace.
> > 
> > .dotest/patch:1946: trailing whitespace.
> >                  * owned by this domain, clear this iommu in iommu_bmp
> > .dotest/patch:2300: trailing whitespace.
> > 
> > error: patch failed: drivers/pci/dmar.c:484
> > error: drivers/pci/dmar.c: patch does not apply
> > error: patch failed: drivers/pci/intel-iommu.c:50
> > error: drivers/pci/intel-iommu.c: patch does not apply
> > error: patch failed: include/linux/dma_remapping.h:111
> > error: include/linux/dma_remapping.h: patch does not apply
> > error: patch failed: include/linux/intel-iommu.h:219
> > error: include/linux/intel-iommu.h: patch does not apply
> > Patch failed at 0001.
> > 
> > Joerg
> > 
> --
> To unsubscribe from this list: send the line "unsubscribe kvm" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 7+ messages in thread

* RE: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
  2008-12-01 14:04         ` Joerg Roedel
@ 2008-12-02  0:31           ` Han, Weidong
  0 siblings, 0 replies; 7+ messages in thread
From: Han, Weidong @ 2008-12-02  0:31 UTC (permalink / raw)
  To: 'Joerg Roedel'
  Cc: 'Joerg Roedel', Woodhouse, David, 'Jesse Barnes',
	'Avi Kivity', Kay, Allen M, Yu, Fenghua,
	'kvm@vger.kernel.org',
	'iommu@lists.linux-foundation.org'

It's fine. You only needs to change the APIs to generic APIs. I will update it soon.

Regards,
Weidong

Joerg Roedel wrote:
> Ok, I got them to apply. I also did the checkpatch cleanups. To speed
> things up a bit I would suggest that I rebase my patchset on your
> patches and send it out in a single series. Any problems with this
> approach?
> 
> Joerg
> 
> On Mon, Dec 01, 2008 at 09:22:42PM +0800, Han, Weidong wrote:
>> Sorry, this patch has style problem. I will update it and also split
>> it to smaller patches for easy reviewing. 
>> 
>> Regards,
>> Weidong
>> 
>> 'Joerg Roedel' wrote:
>>> Hmm, I get these errors using git-am:
>>> 
>>> Applying VT-d: Support multiple device assignment for KVM
>>> .dotest/patch:1344: space before tab in indent.
>>>                                 clflush_cache_range(addr, size);
>>> .dotest/patch:1350: space before tab in indent.
>>>                         clflush_cache_range(addr, size);
>>> .dotest/patch:1907: trailing whitespace.
>>> 
>>> .dotest/patch:1946: trailing whitespace.
>>>                  * owned by this domain, clear this iommu in
>>> iommu_bmp .dotest/patch:2300: trailing whitespace.
>>> 
>>> error: patch failed: drivers/pci/dmar.c:484
>>> error: drivers/pci/dmar.c: patch does not apply
>>> error: patch failed: drivers/pci/intel-iommu.c:50
>>> error: drivers/pci/intel-iommu.c: patch does not apply
>>> error: patch failed: include/linux/dma_remapping.h:111
>>> error: include/linux/dma_remapping.h: patch does not apply
>>> error: patch failed: include/linux/intel-iommu.h:219
>>> error: include/linux/intel-iommu.h: patch does not apply
>>> Patch failed at 0001.
>>> 
>>> Joerg
>>> 
>> --
>> To unsubscribe from this list: send the line "unsubscribe kvm" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2008-12-02  0:31 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2008-11-27 13:49 [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM Han, Weidong
2008-11-28 13:27 ` Joerg Roedel
2008-12-01  6:17   ` Han, Weidong
2008-12-01 11:55     ` 'Joerg Roedel'
2008-12-01 13:22       ` Han, Weidong
2008-12-01 14:04         ` Joerg Roedel
2008-12-02  0:31           ` Han, Weidong

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox