* [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
@ 2008-11-27 13:49 Han, Weidong
2008-11-28 13:27 ` Joerg Roedel
0 siblings, 1 reply; 7+ messages in thread
From: Han, Weidong @ 2008-11-27 13:49 UTC (permalink / raw)
To: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity'
Cc: Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
'iommu@lists.linux-foundation.org'
[-- Attachment #1: Type: text/plain, Size: 47927 bytes --]
In order to support multiple device assignment for KVM, this patch does following main changes:
- extend dmar_domain to own multiple devices from different iommus, use a bitmap of iommus to replace iommu pointer in dmar_domain.
- implement independent low level functions for kvm, then won't impact native VT-d.
- "SAGAW" capability may be different across iommus, that's to say the VT-d page table levels may be different among iommus. This patch uses a defaut agaw, and skip top levels of page tables for iommus which have smaller agaw than default.
- rename the APIs for kvm VT-d, make it more readable.
Signed-off-by: Weidong Han <weidong.han@intel.com>
---
drivers/pci/dmar.c | 15 +
drivers/pci/intel-iommu.c | 698 ++++++++++++++++++++++++++++++++++------
include/linux/dma_remapping.h | 21 +-
include/linux/intel-iommu.h | 21 +-
4 files changed, 637 insertions(+), 118 deletions(-)
diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
index 691b3ad..d6bdced 100644
--- a/drivers/pci/dmar.c
+++ b/drivers/pci/dmar.c
@@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
dmar_tbl = NULL;
}
+extern int width_to_agaw(int width);
int alloc_iommu(struct dmar_drhd_unit *drhd)
{
@@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
int map_size;
u32 ver;
static int iommu_allocated = 0;
+ unsigned long sagaw;
+ int agaw;
iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);
if (!iommu)
@@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
+ /* set agaw, "SAGAW" may be different across iommus */
+ sagaw = cap_sagaw(iommu->cap);
+ for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
+ agaw >= 0; agaw--)
+ if (test_bit(agaw, &sagaw))
+ break;
+ if (agaw < 0) {
+ printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n", sagaw);
+ goto error;
+ }
+ iommu->agaw = agaw;
+
/* the registers might be more than one page */
map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
cap_max_fault_reg_offset(iommu->cap));
diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
index 5c8baa4..55b96c4 100644
--- a/drivers/pci/intel-iommu.c
+++ b/drivers/pci/intel-iommu.c
@@ -50,8 +50,6 @@
#define IOAPIC_RANGE_END (0xfeefffff)
#define IOVA_START_ADDR (0x1000)
-#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
-
#define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
@@ -64,6 +62,7 @@ struct deferred_flush_tables {
int next;
struct iova *iova[HIGH_WATER_MARK];
struct dmar_domain *domain[HIGH_WATER_MARK];
+ struct intel_iommu *iommu;
};
static struct deferred_flush_tables *deferred_flush;
@@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
kmem_cache_free(iommu_iova_cache, iova);
}
+/* in native case, each domain is related to only one iommu */
+static struct intel_iommu *domain_get_only_iommu(struct dmar_domain *domain)
+{
+ struct dmar_drhd_unit *drhd;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+ if (test_bit(drhd->iommu->seq_id, &domain->iommu_bmp))
+ return drhd->iommu;
+ }
+
+ return NULL;
+}
+
+static void domain_flush_cache(struct dmar_domain *domain,
+ void *addr, int size)
+{
+ struct intel_iommu *iommu;
+
+ if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
+ struct dmar_drhd_unit *drhd;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+ iommu = drhd->iommu;
+
+ if (!test_bit(iommu->seq_id, &domain->iommu_bmp))
+ continue;
+
+ if (!ecap_coherent(iommu->ecap))
+ clflush_cache_range(addr, size);
+ }
+ }
+ else {
+ iommu = domain_get_only_iommu(domain);
+ if (iommu && !ecap_coherent(iommu->ecap))
+ clflush_cache_range(addr, size);
+ }
+}
+
+static struct intel_iommu *device_find_matched_iommu(u8 bus, u8 devfn)
+{
+ struct dmar_drhd_unit *drhd = NULL;
+ int i;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+
+ for (i = 0; i < drhd->devices_cnt; i++)
+ if (drhd->devices[i]->bus->number == bus &&
+ drhd->devices[i]->devfn == devfn)
+ return drhd->iommu;
+
+ if (drhd->include_all)
+ return drhd->iommu;
+ }
+
+ return NULL;
+}
+
/* Gets context entry for a given bus and devfn */
static struct context_entry * device_to_context_entry(struct intel_iommu *iommu,
u8 bus, u8 devfn)
@@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
}
-static inline int width_to_agaw(int width)
+int width_to_agaw(int width)
{
return (width - 30) / LEVEL_STRIDE;
}
@@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
flags);
return NULL;
}
- __iommu_flush_cache(domain->iommu, tmp_page,
- PAGE_SIZE);
+ domain_flush_cache(domain, tmp_page, PAGE_SIZE);
dma_set_pte_addr(*pte, virt_to_phys(tmp_page));
/*
* high level table always sets r/w, last level page
@@ -356,7 +417,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
*/
dma_set_pte_readable(*pte);
dma_set_pte_writable(*pte);
- __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+ domain_flush_cache(domain, pte, sizeof(*pte));
}
parent = phys_to_virt(dma_pte_addr(*pte));
level--;
@@ -399,7 +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain, u64 addr)
if (pte) {
dma_clear_pte(*pte);
- __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+ domain_flush_cache(domain, pte, sizeof(*pte));
}
}
@@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct dmar_domain *domain,
free_pgtable_page(
phys_to_virt(dma_pte_addr(*pte)));
dma_clear_pte(*pte);
- __iommu_flush_cache(domain->iommu,
- pte, sizeof(*pte));
+ domain_flush_cache(domain, pte, sizeof(*pte));
}
tmp += level_size(level);
}
@@ -948,8 +1008,8 @@ static int iommu_init_domains(struct intel_iommu *iommu)
return 0;
}
-
static void domain_exit(struct dmar_domain *domain);
+static void vm_domain_exit(struct dmar_domain *domain);
void free_dmar_iommu(struct intel_iommu *iommu)
{
@@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
for (; i < cap_ndoms(iommu->cap); ) {
domain = iommu->domains[i];
clear_bit(i, iommu->domain_ids);
- domain_exit(domain);
+
+ if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
+ if (--domain->iommu_count == 0)
+ vm_domain_exit(domain);
+ }
+ else
+ domain_exit(domain);
+
i = find_next_bit(iommu->domain_ids,
cap_ndoms(iommu->cap), i+1);
}
@@ -1006,8 +1073,11 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
set_bit(num, iommu->domain_ids);
domain->id = num;
- domain->iommu = iommu;
+ memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
+ set_bit(iommu->seq_id, &domain->iommu_bmp);
iommu->domains[num] = domain;
+ domain->iommu_count = 1;
+ domain->flags = 0;
spin_unlock_irqrestore(&iommu->lock, flags);
return domain;
@@ -1016,10 +1086,12 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
static void iommu_free_domain(struct dmar_domain *domain)
{
unsigned long flags;
+ struct intel_iommu *iommu;
- spin_lock_irqsave(&domain->iommu->lock, flags);
- clear_bit(domain->id, domain->iommu->domain_ids);
- spin_unlock_irqrestore(&domain->iommu->lock, flags);
+ iommu = domain_get_only_iommu(domain);
+ spin_lock_irqsave(&iommu->lock, flags);
+ clear_bit(domain->id, iommu->domain_ids);
+ spin_unlock_irqrestore(&iommu->lock, flags);
}
static struct iova_domain reserved_iova_list;
@@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
domain_reserve_special_ranges(domain);
/* calculate AGAW */
- iommu = domain->iommu;
+ iommu = domain_get_only_iommu(domain);
if (guest_width > cap_mgaw(iommu->cap))
guest_width = cap_mgaw(iommu->cap);
domain->gaw = guest_width;
@@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
sagaw = cap_sagaw(iommu->cap);
if (!test_bit(agaw, &sagaw)) {
/* hardware doesn't support it, choose a bigger one */
- pr_debug("IOMMU: hardware doesn't support agaw %d\n", agaw);
+ pr_debug("IOMMU: hardware doesn't support agaw %d\n",
+ agaw);
agaw = find_next_bit(&sagaw, 5, agaw);
if (agaw >= 5)
return -ENODEV;
}
domain->agaw = agaw;
+
INIT_LIST_HEAD(&domain->devices);
/* always allocate the top pgd */
domain->pgd = (struct dma_pte *)alloc_pgtable_page();
if (!domain->pgd)
return -ENOMEM;
- __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
+ domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
return 0;
}
@@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain *domain)
}
static int domain_context_mapping_one(struct dmar_domain *domain,
- u8 bus, u8 devfn)
+ struct intel_iommu *iommu, u8 bus, u8 devfn)
{
struct context_entry *context;
- struct intel_iommu *iommu = domain->iommu;
unsigned long flags;
pr_debug("Set context mapping for %02x:%02x.%d\n",
@@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
{
int ret;
struct pci_dev *tmp, *parent;
+ struct intel_iommu *iommu;
+
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return -ENODEV;
- ret = domain_context_mapping_one(domain, pdev->bus->number,
- pdev->devfn);
+ ret = domain_context_mapping_one(domain, iommu,
+ pdev->bus->number, pdev->devfn);
if (ret)
return ret;
@@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
/* Secondary interface's bus number and devfn 0 */
parent = pdev->bus->self;
while (parent != tmp) {
- ret = domain_context_mapping_one(domain, parent->bus->number,
- parent->devfn);
+ ret = domain_context_mapping_one(domain, iommu,
+ parent->bus->number, parent->devfn);
if (ret)
return ret;
parent = parent->bus->self;
}
if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
- return domain_context_mapping_one(domain,
+ return domain_context_mapping_one(domain, iommu,
tmp->subordinate->number, 0);
else /* this is a legacy PCI bridge */
- return domain_context_mapping_one(domain,
+ return domain_context_mapping_one(domain, iommu,
tmp->bus->number, tmp->devfn);
}
-static int domain_context_mapped(struct dmar_domain *domain,
- struct pci_dev *pdev)
+static int domain_context_mapped(struct pci_dev *pdev)
{
int ret;
struct pci_dev *tmp, *parent;
+ struct intel_iommu *iommu;
- ret = device_context_mapped(domain->iommu,
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return 0;
+
+ ret = device_context_mapped(iommu,
pdev->bus->number, pdev->devfn);
if (!ret)
return ret;
@@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct dmar_domain *domain,
/* Secondary interface's bus number and devfn 0 */
parent = pdev->bus->self;
while (parent != tmp) {
- ret = device_context_mapped(domain->iommu, parent->bus->number,
+ ret = device_context_mapped(iommu, parent->bus->number,
parent->devfn);
if (!ret)
return ret;
parent = parent->bus->self;
}
if (tmp->is_pcie)
- return device_context_mapped(domain->iommu,
+ return device_context_mapped(iommu,
tmp->subordinate->number, 0);
else
- return device_context_mapped(domain->iommu,
+ return device_context_mapped(iommu,
tmp->bus->number, tmp->devfn);
}
@@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
BUG_ON(dma_pte_addr(*pte));
dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
dma_set_pte_prot(*pte, prot);
- __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+ domain_flush_cache(domain, pte, sizeof(*pte));
start_pfn++;
index++;
}
return 0;
}
-static void detach_domain_for_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
+static void detach_domain_for_dev(struct dmar_domain *domain,
+ u8 bus, u8 devfn)
{
- clear_context_table(domain->iommu, bus, devfn);
- domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
- DMA_CCMD_GLOBAL_INVL, 0);
- domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
- DMA_TLB_GLOBAL_FLUSH, 0);
+ struct intel_iommu *iommu;
+
+ iommu = device_find_matched_iommu(bus, devfn);
+ if (!iommu)
+ return;
+
+ clear_context_table(iommu, bus, devfn);
+ iommu->flush.flush_context(iommu, 0, 0, 0,
+ DMA_CCMD_GLOBAL_INVL, 0);
+ iommu->flush.flush_iotlb(iommu, 0, 0, 0,
+ DMA_TLB_GLOBAL_FLUSH, 0);
}
static void domain_remove_dev_info(struct dmar_domain *domain)
@@ -1336,7 +1425,6 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
{
struct dmar_domain *domain, *found = NULL;
struct intel_iommu *iommu;
- struct dmar_drhd_unit *drhd;
struct device_domain_info *info, *tmp;
struct pci_dev *dev_tmp;
unsigned long flags;
@@ -1371,13 +1459,9 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
}
/* Allocate new domain for the device */
- drhd = dmar_find_matched_drhd_unit(pdev);
- if (!drhd) {
- printk(KERN_ERR "IOMMU: can't find DMAR for device %s\n",
- pci_name(pdev));
- return NULL;
- }
- iommu = drhd->iommu;
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return NULL;
domain = iommu_alloc_domain(iommu);
if (!domain)
@@ -1400,7 +1484,7 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
info->dev = NULL;
info->domain = domain;
/* This domain is shared by devices under p2p bridge */
- domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
+ domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
/* pcie-to-pci bridge already has a domain, uses it */
found = NULL;
@@ -1805,7 +1889,7 @@ get_valid_domain_for_dev(struct pci_dev *pdev)
}
/* make sure context mapping is ok */
- if (unlikely(!domain_context_mapped(domain, pdev))) {
+ if (unlikely(!domain_context_mapped(pdev))) {
ret = domain_context_mapping(domain, pdev);
if (ret) {
printk(KERN_ERR
@@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
{
struct pci_dev *pdev = to_pci_dev(hwdev);
struct dmar_domain *domain;
+ struct intel_iommu *iommu;
phys_addr_t start_paddr;
struct iova *iova;
int prot = 0;
@@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
if (!domain)
return 0;
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return 0;
+
size = aligned_size((u64)paddr, size);
iova = __intel_alloc_iova(hwdev, domain, size, pdev->dma_mask);
@@ -1849,7 +1938,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
* mappings..
*/
if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
- !cap_zlr(domain->iommu->cap))
+ !cap_zlr(iommu->cap))
prot |= DMA_PTE_READ;
if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
prot |= DMA_PTE_WRITE;
@@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
goto error;
/* it's a non-present to present mapping */
- ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
+ ret = iommu_flush_iotlb_psi(iommu, domain->id,
start_paddr, size >> VTD_PAGE_SHIFT, 1);
if (ret)
- iommu_flush_write_buffer(domain->iommu);
+ iommu_flush_write_buffer(iommu);
return start_paddr + ((u64)paddr & (~PAGE_MASK));
@@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
/* just flush them all */
for (i = 0; i < g_num_of_iommus; i++) {
if (deferred_flush[i].next) {
- struct intel_iommu *iommu =
- deferred_flush[i].domain[0]->iommu;
+ struct intel_iommu *iommu = deferred_flush[i].iommu;
iommu->flush.flush_iotlb(iommu, 0, 0, 0,
DMA_TLB_GLOBAL_FLUSH, 0);
@@ -1921,7 +2009,8 @@ static void flush_unmaps_timeout(unsigned long data)
spin_unlock_irqrestore(&async_umap_flush_lock, flags);
}
-static void add_unmap(struct dmar_domain *dom, struct iova *iova)
+static void add_unmap(struct dmar_domain *dom,
+ struct intel_iommu *iommu, struct iova *iova)
{
unsigned long flags;
int next, iommu_id;
@@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova)
if (list_size == HIGH_WATER_MARK)
flush_unmaps();
- iommu_id = dom->iommu->seq_id;
+ iommu_id = iommu->seq_id;
next = deferred_flush[iommu_id].next;
deferred_flush[iommu_id].domain[next] = dom;
deferred_flush[iommu_id].iova[next] = iova;
+ deferred_flush[iommu_id].iommu = iommu;
deferred_flush[iommu_id].next++;
if (!timer_on) {
@@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
struct dmar_domain *domain;
unsigned long start_addr;
struct iova *iova;
+ struct intel_iommu *iommu;
if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
return;
domain = find_domain(pdev);
BUG_ON(!domain);
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return;
+
iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
if (!iova)
return;
@@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
/* free page tables */
dma_pte_free_pagetable(domain, start_addr, start_addr + size);
if (intel_iommu_strict) {
- if (iommu_flush_iotlb_psi(domain->iommu,
+ if (iommu_flush_iotlb_psi(iommu,
domain->id, start_addr, size >> VTD_PAGE_SHIFT, 0))
- iommu_flush_write_buffer(domain->iommu);
+ iommu_flush_write_buffer(iommu);
/* free iova */
__free_iova(&domain->iovad, iova);
} else {
- add_unmap(domain, iova);
+ add_unmap(domain, iommu, iova);
/*
* queue up the release of the unmap to save the 1/6th of the
* cpu used up by the iotlb flush operation...
@@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
size_t size = 0;
void *addr;
struct scatterlist *sg;
+ struct intel_iommu *iommu;
if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
return;
domain = find_domain(pdev);
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return;
+
iova = find_iova(&domain->iovad, IOVA_PFN(sglist[0].dma_address));
if (!iova)
return;
@@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
/* free page tables */
dma_pte_free_pagetable(domain, start_addr, start_addr + size);
- if (iommu_flush_iotlb_psi(domain->iommu, domain->id, start_addr,
+ if (iommu_flush_iotlb_psi(iommu, domain->id, start_addr,
size >> VTD_PAGE_SHIFT, 0))
- iommu_flush_write_buffer(domain->iommu);
+ iommu_flush_write_buffer(iommu);
/* free iova */
__free_iova(&domain->iovad, iova);
@@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
int ret;
struct scatterlist *sg;
unsigned long start_addr;
+ struct intel_iommu *iommu;
BUG_ON(dir == DMA_NONE);
if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
@@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
if (!domain)
return 0;
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return 0;
+
for_each_sg(sglist, sg, nelems, i) {
addr = SG_ENT_VIRT_ADDRESS(sg);
addr = (void *)virt_to_phys(addr);
@@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
* mappings..
*/
if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
- !cap_zlr(domain->iommu->cap))
+ !cap_zlr(iommu->cap))
prot |= DMA_PTE_READ;
if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
prot |= DMA_PTE_WRITE;
@@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
}
/* it's a non-present to present mapping */
- if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
+ if (iommu_flush_iotlb_psi(iommu, domain->id,
start_addr, offset >> VTD_PAGE_SHIFT, 1))
- iommu_flush_write_buffer(domain->iommu);
+ iommu_flush_write_buffer(iommu);
return nelems;
}
@@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
return 0;
}
-void intel_iommu_domain_exit(struct dmar_domain *domain)
+/* domain id for virtual machine, it won't be set in context */
+static unsigned long vm_domid;
+
+static int vm_domain_min_agaw(struct dmar_domain *domain)
+{
+ struct dmar_drhd_unit *drhd;
+ struct intel_iommu *iommu;
+ int min_agaw = domain->agaw;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+ iommu = drhd->iommu;
+
+ if (test_bit(iommu->seq_id, &domain->iommu_bmp))
+ if (min_agaw > iommu->agaw)
+ min_agaw = iommu->agaw;
+ }
+
+ return min_agaw;
+}
+
+static int vm_domain_add_dev_info(struct dmar_domain *domain,
+ struct pci_dev *pdev)
+{
+ struct device_domain_info *info;
+ unsigned long flags;
+
+ info = alloc_devinfo_mem();
+ if (!info)
+ return -ENOMEM;
+
+ info->bus = pdev->bus->number;
+ info->devfn = pdev->devfn;
+ info->dev = pdev;
+ info->domain = domain;
+
+ spin_lock_irqsave(&device_domain_lock, flags);
+ list_add(&info->link, &domain->devices);
+ list_add(&info->global, &device_domain_list);
+ pdev->dev.archdata.iommu = info;
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+
+ return 0;
+}
+
+static void vm_domain_remove_one_dev_info(struct dmar_domain *domain,
+ struct pci_dev *pdev)
+{
+ struct device_domain_info *info;
+ struct intel_iommu *iommu;
+ unsigned long flags;
+ int found = 0;
+
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+
+ spin_lock_irqsave(&device_domain_lock, flags);
+ while (!list_empty(&domain->devices)) {
+ info = list_entry(domain->devices.next,
+ struct device_domain_info, link);
+ if (info->bus == pdev->bus->number &&
+ info->devfn == pdev->devfn) {
+ list_del(&info->link);
+ list_del(&info->global);
+ if (info->dev)
+ info->dev->dev.archdata.iommu = NULL;
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+
+ detach_domain_for_dev(info->domain,
+ info->bus, info->devfn);
+ free_devinfo_mem(info);
+
+ spin_lock_irqsave(&device_domain_lock, flags);
+
+ if (found)
+ break;
+ else
+ continue;
+ }
+
+ /* if there is no other devices under the same iommu
+ * owned by this domain, clear this iommu in iommu_bmp
+ */
+ if (device_find_matched_iommu(info->bus, info->devfn) == iommu)
+ found = 1;
+ }
+
+ if (found == 0) {
+ spin_lock_irqsave(&iommu->lock, flags);
+ clear_bit(iommu->seq_id, &domain->iommu_bmp);
+ domain->iommu_count--;
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ }
+
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+}
+
+static void vm_domain_remove_all_dev_info(struct dmar_domain *domain)
+{
+ struct device_domain_info *info;
+ struct intel_iommu *iommu;
+ unsigned long flags;
+
+ spin_lock_irqsave(&device_domain_lock, flags);
+ while (!list_empty(&domain->devices)) {
+ info = list_entry(domain->devices.next,
+ struct device_domain_info, link);
+ list_del(&info->link);
+ list_del(&info->global);
+ if (info->dev)
+ info->dev->dev.archdata.iommu = NULL;
+
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+ detach_domain_for_dev(info->domain,
+ info->bus, info->devfn);
+
+ /* clear this iommu in iommu_bmp */
+ iommu = device_find_matched_iommu(info->bus, info->devfn);
+ spin_lock_irqsave(&iommu->lock, flags);
+ if (test_and_clear_bit(iommu->seq_id,
+ &domain->iommu_bmp))
+ domain->iommu_count--;
+ spin_unlock_irqrestore(&iommu->lock, flags);
+
+ free_devinfo_mem(info);
+ spin_lock_irqsave(&device_domain_lock, flags);
+ }
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+}
+
+static int vm_domain_context_mapping_one(struct dmar_domain *domain,
+ struct intel_iommu *iommu, u8 bus, u8 devfn)
+{
+ struct context_entry *context;
+ unsigned long flags;
+ struct dma_pte *pgd;
+ unsigned long num;
+ unsigned long ndomains;
+ int id;
+ int agaw;
+ int found = 0;
+
+ pr_debug("Set context mapping for %02x:%02x.%d\n",
+ bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
+ BUG_ON(!domain->pgd);
+ context = device_to_context_entry(iommu, bus, devfn);
+ if (!context)
+ return -ENOMEM;
+ spin_lock_irqsave(&iommu->lock, flags);
+ if (context_present(*context)) {
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ return 0;
+ }
+
+ id = domain->id;
+
+ /* find an available domain id for this device in iommu */
+ ndomains = cap_ndoms(iommu->cap);
+ num = find_first_bit(iommu->domain_ids, ndomains);
+ for (; num < ndomains; ) {
+ if (iommu->domains[num] == domain) {
+ id = num;
+ found = 1;
+ break;
+ }
+ num = find_next_bit(iommu->domain_ids,
+ cap_ndoms(iommu->cap), num+1);
+ }
+
+ if (found == 0) {
+ num = find_first_zero_bit(iommu->domain_ids, ndomains);
+ if (num >= ndomains) {
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ printk(KERN_ERR "IOMMU: no free domain ids\n");
+ return -EFAULT;
+ }
+
+ set_bit(num, iommu->domain_ids);
+ iommu->domains[num] = domain;
+ id = num;
+ }
+
+ pgd = domain->pgd;
+
+ /* Skip top levels of page tables for
+ * iommu which has less agaw than default.
+ */
+ for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
+ pgd = phys_to_virt(dma_pte_addr(*pgd));
+ if (!dma_pte_present(*pgd)) {
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ return -ENOMEM;
+ }
+ }
+
+ context_set_domain_id(*context, id);
+ context_set_address_width(*context, iommu->agaw);
+ context_set_address_root(*context, virt_to_phys(pgd));
+ context_set_translation_type(*context, CONTEXT_TT_MULTI_LEVEL);
+ context_set_fault_enable(*context);
+ context_set_present(*context);
+ __iommu_flush_cache(iommu, context, sizeof(*context));
+
+ /* it's a non-present to present mapping */
+ if (iommu->flush.flush_context(iommu, id,
+ (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
+ DMA_CCMD_DEVICE_INVL, 1))
+ iommu_flush_write_buffer(iommu);
+ else
+ iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_DSI_FLUSH, 0);
+
+ if (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp))
+ domain->iommu_count++;
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ return 0;
+}
+
+static int
+vm_domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
+{
+ int ret;
+ struct pci_dev *tmp, *parent;
+ struct intel_iommu *iommu;
+
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return -ENODEV;
+
+ ret = vm_domain_context_mapping_one(domain, iommu,
+ pdev->bus->number, pdev->devfn);
+ if (ret)
+ return ret;
+
+ /* dependent device mapping */
+ tmp = pci_find_upstream_pcie_bridge(pdev);
+ if (!tmp)
+ return 0;
+ /* Secondary interface's bus number and devfn 0 */
+ parent = pdev->bus->self;
+ while (parent != tmp) {
+ ret = vm_domain_context_mapping_one(domain, iommu,
+ parent->bus->number, parent->devfn);
+ if (ret)
+ return ret;
+ parent = parent->bus->self;
+ }
+ if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
+ return vm_domain_context_mapping_one(domain, iommu,
+ tmp->subordinate->number, 0);
+ else /* this is a legacy PCI bridge */
+ return vm_domain_context_mapping_one(domain, iommu,
+ tmp->bus->number, tmp->devfn);
+}
+
+
+static int vm_domain_init(struct dmar_domain *domain, int guest_width)
+{
+ int adjust_width;
+
+ init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
+ spin_lock_init(&domain->mapping_lock);
+
+ domain_reserve_special_ranges(domain);
+
+ /* calculate AGAW */
+ domain->gaw = guest_width;
+ adjust_width = guestwidth_to_adjustwidth(guest_width);
+ domain->agaw = width_to_agaw(adjust_width);
+
+ INIT_LIST_HEAD(&domain->devices);
+
+ /* always allocate the top pgd */
+ domain->pgd = (struct dma_pte *)alloc_pgtable_page();
+ if (!domain->pgd)
+ return -ENOMEM;
+ domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
+ return 0;
+}
+
+static void iommu_free_vm_domain(struct dmar_domain *domain)
+{
+ unsigned long flags;
+ struct dmar_drhd_unit *drhd;
+ struct intel_iommu *iommu;
+ unsigned long i;
+ unsigned long ndomains;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+ iommu = drhd->iommu;
+
+ ndomains = cap_ndoms(iommu->cap);
+ i = find_first_bit(iommu->domain_ids, ndomains);
+ for (; i < ndomains; ) {
+ if (iommu->domains[i] == domain) {
+ spin_lock_irqsave(&iommu->lock, flags);
+ clear_bit(i, iommu->domain_ids);
+ iommu->domains[i] = NULL;
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ break;
+ }
+ i = find_next_bit(iommu->domain_ids, ndomains, i+1);
+ }
+ }
+}
+
+static void vm_domain_exit(struct dmar_domain *domain)
{
u64 end;
@@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
if (!domain)
return;
+ vm_domain_remove_all_dev_info(domain);
+ /* destroy iovas */
+ put_iova_domain(&domain->iovad);
end = DOMAIN_MAX_ADDR(domain->gaw);
- end = end & (~VTD_PAGE_MASK);
+ end &= VTD_PAGE_MASK;
/* clear ptes */
dma_pte_clear_range(domain, 0, end);
@@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
/* free page tables */
dma_pte_free_pagetable(domain, 0, end);
- iommu_free_domain(domain);
+ iommu_free_vm_domain(domain);
free_domain_mem(domain);
}
-EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
-struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
+static struct dmar_domain *iommu_alloc_vm_domain(void)
{
- struct dmar_drhd_unit *drhd;
struct dmar_domain *domain;
- struct intel_iommu *iommu;
- drhd = dmar_find_matched_drhd_unit(pdev);
- if (!drhd) {
- printk(KERN_ERR "intel_iommu_domain_alloc: drhd == NULL\n");
+ domain = alloc_domain_mem();
+ if (!domain)
return NULL;
- }
- iommu = drhd->iommu;
- if (!iommu) {
- printk(KERN_ERR
- "intel_iommu_domain_alloc: iommu == NULL\n");
- return NULL;
- }
- domain = iommu_alloc_domain(iommu);
+ domain->id = vm_domid++;
+ domain->iommu_count = 0;
+ domain->max_addr = 0;
+ memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
+ domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE;
+
+ return domain;
+}
+
+struct dmar_domain *intel_iommu_alloc_domain(void)
+{
+ struct dmar_domain *domain;
+
+ domain = iommu_alloc_vm_domain();
if (!domain) {
printk(KERN_ERR
"intel_iommu_domain_alloc: domain == NULL\n");
return NULL;
}
- if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
+ if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
printk(KERN_ERR
"intel_iommu_domain_alloc: domain_init() failed\n");
- intel_iommu_domain_exit(domain);
+ vm_domain_exit(domain);
return NULL;
}
+
return domain;
}
-EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
+EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
+
+void intel_iommu_free_domain(struct dmar_domain *domain)
+{
+ vm_domain_exit(domain);
+}
+EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
-int intel_iommu_context_mapping(
- struct dmar_domain *domain, struct pci_dev *pdev)
+int intel_iommu_assign_device(struct dmar_domain *domain,
+ struct pci_dev *pdev)
{
- int rc;
- rc = domain_context_mapping(domain, pdev);
- return rc;
+ struct intel_iommu *iommu;
+ int addr_width;
+ u64 end;
+ int ret;
+
+ /* normally pdev is not mapped */
+ if (unlikely(domain_context_mapped(pdev))) {
+ struct dmar_domain *old_domain;
+
+ old_domain = find_domain(pdev);
+ if (old_domain) {
+ if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE)
+ vm_domain_remove_one_dev_info(old_domain, pdev);
+ else
+ domain_remove_dev_info(old_domain);
+ }
+ }
+
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return -ENODEV;
+
+ /* check if this iommu agaw is sufficient for max mapped address */
+ addr_width = agaw_to_width(iommu->agaw);
+ end = DOMAIN_MAX_ADDR(addr_width);
+ end = end & VTD_PAGE_MASK;
+ if (end < domain->max_addr) {
+ printk(KERN_ERR "%s: iommu agaw (%d) is not "
+ "sufficient for the mapped address (%llx)\n",
+ __func__, iommu->agaw, domain->max_addr);
+ return -EFAULT;
+ }
+
+ ret = vm_domain_context_mapping(domain, pdev);
+ if (ret)
+ return ret;
+
+ ret = vm_domain_add_dev_info(domain, pdev);
+ return ret;
}
-EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
+EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
-int intel_iommu_page_mapping(
- struct dmar_domain *domain, dma_addr_t iova,
- u64 hpa, size_t size, int prot)
+
+void intel_iommu_deassign_device(struct dmar_domain *domain,
+ struct pci_dev *pdev)
{
- int rc;
- rc = domain_page_mapping(domain, iova, hpa, size, prot);
- return rc;
+ vm_domain_remove_one_dev_info(domain, pdev);
}
-EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
+EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
-void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
+int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
+ u64 hpa, size_t size, int prot)
{
- detach_domain_for_dev(domain, bus, devfn);
+ u64 max_addr;
+ int addr_width;
+ int ret;
+
+ max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
+ if (domain->max_addr < max_addr) {
+ int min_agaw;
+ u64 end;
+
+ /* check if minimum agaw is sufficient for mapped address */
+ min_agaw = vm_domain_min_agaw(domain);
+ addr_width = agaw_to_width(min_agaw);
+ end = DOMAIN_MAX_ADDR(addr_width);
+ end = end & VTD_PAGE_MASK;
+ if (end < max_addr) {
+ printk(KERN_ERR "%s: iommu agaw (%d) is not "
+ "sufficient for the mapped address (%llx)\n",
+ __func__, min_agaw, max_addr);
+ return -EFAULT;
+ }
+ domain->max_addr = max_addr;
+ }
+
+ ret = domain_page_mapping(domain, iova, hpa, size, prot);
+ return ret;
}
-EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
+EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
-struct dmar_domain *
-intel_iommu_find_domain(struct pci_dev *pdev)
+void intel_iommu_unmap_pages(struct dmar_domain *domain,
+ dma_addr_t iova, size_t size)
{
- return find_domain(pdev);
+ dma_addr_t base;
+
+ /* The address might not be aligned */
+ base = iova & PAGE_MASK;
+ size = PAGE_ALIGN(size);
+ dma_pte_clear_range(domain, base, base + size);
}
-EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
+EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
int intel_iommu_found(void)
{
diff --git a/include/linux/dma_remapping.h b/include/linux/dma_remapping.h
index 952df39..3568ae6 100644
--- a/include/linux/dma_remapping.h
+++ b/include/linux/dma_remapping.h
@@ -111,11 +111,21 @@ struct dma_pte {
(p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
#define dma_pte_present(p) (((p).val & 3) != 0)
+/* domain flags, one domain owns one device by default */
+
+/* devices under the same p2p bridge are owned in one domain */
+#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES (1 << 0)
+
+/* domain represents a virtual machine, more than one devices
+ * across iommus may be owned in one domain, e.g. kvm guest.
+ */
+#define DOMAIN_FLAG_VIRTUAL_MACHINE (1 << 1)
+
struct intel_iommu;
struct dmar_domain {
int id; /* domain id */
- struct intel_iommu *iommu; /* back pointer to owning iommu */
+ unsigned long iommu_bmp; /* bitmap of iommus this domain uses*/
struct list_head devices; /* all devices' list */
struct iova_domain iovad; /* iova's that belong to this domain */
@@ -123,12 +133,13 @@ struct dmar_domain {
struct dma_pte *pgd; /* virtual address */
spinlock_t mapping_lock; /* page table lock */
int gaw; /* max guest address width */
+ int agaw; /* adjusted guest address width */
- /* adjusted guest address width, 0 is level 2 30-bit */
- int agaw;
+ int flags; /* domain flag */
-#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
- int flags;
+ /* following fields are used in virtual machine case */
+ int iommu_count; /* reference count of iommu */
+ u64 max_addr; /* maximum mapped address */
};
/* PCI domain-device relationship */
diff --git a/include/linux/intel-iommu.h b/include/linux/intel-iommu.h
index 3d017cf..c2f37b8 100644
--- a/include/linux/intel-iommu.h
+++ b/include/linux/intel-iommu.h
@@ -219,6 +219,8 @@ do { \
} \
} while (0)
+#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
+
#define QI_LENGTH 256 /* queue length */
enum {
@@ -299,6 +301,7 @@ struct intel_iommu {
struct dmar_domain **domains; /* ptr to domains */
spinlock_t lock; /* protect context, domain ids */
struct root_entry *root_entry; /* virtual address */
+ int agaw;
unsigned int irq;
unsigned char name[7]; /* Device Name */
@@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu *iommu, u16 did, u64 addr,
extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu *iommu);
-void intel_iommu_domain_exit(struct dmar_domain *domain);
-struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
-int intel_iommu_context_mapping(struct dmar_domain *domain,
- struct pci_dev *pdev);
-int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
- u64 hpa, size_t size, int prot);
-void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn);
-struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
+struct dmar_domain *intel_iommu_alloc_domain(void);
+void intel_iommu_free_domain(struct dmar_domain *domain);
+int intel_iommu_assign_device(struct dmar_domain *domain,
+ struct pci_dev *pdev);
+void intel_iommu_deassign_device(struct dmar_domain *domain,
+ struct pci_dev *pdev);
+int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
+ u64 hpa, size_t size, int prot);
+void intel_iommu_unmap_pages(struct dmar_domain *domain,
+ dma_addr_t iova, size_t size);
u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
#ifdef CONFIG_DMAR
--
1.5.1
[-- Attachment #2: 0001-vtd-multiple-device-assignment-for-kvm-v2.patch --]
[-- Type: application/octet-stream, Size: 37293 bytes --]
From 260c3191401888fd4de7028991653960a05ca6ac Mon Sep 17 00:00:00 2001
From: Weidong Han <weidong.han@intel.com>
Date: Thu, 27 Nov 2008 21:26:47 +0800
Subject: [PATCH] vtd multiple device assignment for kvm
Signed-off-by: Weidong Han <weidong.han@intel.com>
---
drivers/pci/dmar.c | 15 +
drivers/pci/intel-iommu.c | 698 ++++++++++++++++++++++++++++++++++------
include/linux/dma_remapping.h | 21 +-
include/linux/intel-iommu.h | 21 +-
4 files changed, 637 insertions(+), 118 deletions(-)
diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
index 691b3ad..d6bdced 100644
--- a/drivers/pci/dmar.c
+++ b/drivers/pci/dmar.c
@@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
dmar_tbl = NULL;
}
+extern int width_to_agaw(int width);
int alloc_iommu(struct dmar_drhd_unit *drhd)
{
@@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
int map_size;
u32 ver;
static int iommu_allocated = 0;
+ unsigned long sagaw;
+ int agaw;
iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);
if (!iommu)
@@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
+ /* set agaw, "SAGAW" may be different across iommus */
+ sagaw = cap_sagaw(iommu->cap);
+ for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
+ agaw >= 0; agaw--)
+ if (test_bit(agaw, &sagaw))
+ break;
+ if (agaw < 0) {
+ printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n", sagaw);
+ goto error;
+ }
+ iommu->agaw = agaw;
+
/* the registers might be more than one page */
map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
cap_max_fault_reg_offset(iommu->cap));
diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
index 5c8baa4..55b96c4 100644
--- a/drivers/pci/intel-iommu.c
+++ b/drivers/pci/intel-iommu.c
@@ -50,8 +50,6 @@
#define IOAPIC_RANGE_END (0xfeefffff)
#define IOVA_START_ADDR (0x1000)
-#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
-
#define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
@@ -64,6 +62,7 @@ struct deferred_flush_tables {
int next;
struct iova *iova[HIGH_WATER_MARK];
struct dmar_domain *domain[HIGH_WATER_MARK];
+ struct intel_iommu *iommu;
};
static struct deferred_flush_tables *deferred_flush;
@@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
kmem_cache_free(iommu_iova_cache, iova);
}
+/* in native case, each domain is related to only one iommu */
+static struct intel_iommu *domain_get_only_iommu(struct dmar_domain *domain)
+{
+ struct dmar_drhd_unit *drhd;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+ if (test_bit(drhd->iommu->seq_id, &domain->iommu_bmp))
+ return drhd->iommu;
+ }
+
+ return NULL;
+}
+
+static void domain_flush_cache(struct dmar_domain *domain,
+ void *addr, int size)
+{
+ struct intel_iommu *iommu;
+
+ if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
+ struct dmar_drhd_unit *drhd;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+ iommu = drhd->iommu;
+
+ if (!test_bit(iommu->seq_id, &domain->iommu_bmp))
+ continue;
+
+ if (!ecap_coherent(iommu->ecap))
+ clflush_cache_range(addr, size);
+ }
+ }
+ else {
+ iommu = domain_get_only_iommu(domain);
+ if (iommu && !ecap_coherent(iommu->ecap))
+ clflush_cache_range(addr, size);
+ }
+}
+
+static struct intel_iommu *device_find_matched_iommu(u8 bus, u8 devfn)
+{
+ struct dmar_drhd_unit *drhd = NULL;
+ int i;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+
+ for (i = 0; i < drhd->devices_cnt; i++)
+ if (drhd->devices[i]->bus->number == bus &&
+ drhd->devices[i]->devfn == devfn)
+ return drhd->iommu;
+
+ if (drhd->include_all)
+ return drhd->iommu;
+ }
+
+ return NULL;
+}
+
/* Gets context entry for a given bus and devfn */
static struct context_entry * device_to_context_entry(struct intel_iommu *iommu,
u8 bus, u8 devfn)
@@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
}
-static inline int width_to_agaw(int width)
+int width_to_agaw(int width)
{
return (width - 30) / LEVEL_STRIDE;
}
@@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
flags);
return NULL;
}
- __iommu_flush_cache(domain->iommu, tmp_page,
- PAGE_SIZE);
+ domain_flush_cache(domain, tmp_page, PAGE_SIZE);
dma_set_pte_addr(*pte, virt_to_phys(tmp_page));
/*
* high level table always sets r/w, last level page
@@ -356,7 +417,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
*/
dma_set_pte_readable(*pte);
dma_set_pte_writable(*pte);
- __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+ domain_flush_cache(domain, pte, sizeof(*pte));
}
parent = phys_to_virt(dma_pte_addr(*pte));
level--;
@@ -399,7 +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain, u64 addr)
if (pte) {
dma_clear_pte(*pte);
- __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+ domain_flush_cache(domain, pte, sizeof(*pte));
}
}
@@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct dmar_domain *domain,
free_pgtable_page(
phys_to_virt(dma_pte_addr(*pte)));
dma_clear_pte(*pte);
- __iommu_flush_cache(domain->iommu,
- pte, sizeof(*pte));
+ domain_flush_cache(domain, pte, sizeof(*pte));
}
tmp += level_size(level);
}
@@ -948,8 +1008,8 @@ static int iommu_init_domains(struct intel_iommu *iommu)
return 0;
}
-
static void domain_exit(struct dmar_domain *domain);
+static void vm_domain_exit(struct dmar_domain *domain);
void free_dmar_iommu(struct intel_iommu *iommu)
{
@@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
for (; i < cap_ndoms(iommu->cap); ) {
domain = iommu->domains[i];
clear_bit(i, iommu->domain_ids);
- domain_exit(domain);
+
+ if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
+ if (--domain->iommu_count == 0)
+ vm_domain_exit(domain);
+ }
+ else
+ domain_exit(domain);
+
i = find_next_bit(iommu->domain_ids,
cap_ndoms(iommu->cap), i+1);
}
@@ -1006,8 +1073,11 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
set_bit(num, iommu->domain_ids);
domain->id = num;
- domain->iommu = iommu;
+ memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
+ set_bit(iommu->seq_id, &domain->iommu_bmp);
iommu->domains[num] = domain;
+ domain->iommu_count = 1;
+ domain->flags = 0;
spin_unlock_irqrestore(&iommu->lock, flags);
return domain;
@@ -1016,10 +1086,12 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
static void iommu_free_domain(struct dmar_domain *domain)
{
unsigned long flags;
+ struct intel_iommu *iommu;
- spin_lock_irqsave(&domain->iommu->lock, flags);
- clear_bit(domain->id, domain->iommu->domain_ids);
- spin_unlock_irqrestore(&domain->iommu->lock, flags);
+ iommu = domain_get_only_iommu(domain);
+ spin_lock_irqsave(&iommu->lock, flags);
+ clear_bit(domain->id, iommu->domain_ids);
+ spin_unlock_irqrestore(&iommu->lock, flags);
}
static struct iova_domain reserved_iova_list;
@@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
domain_reserve_special_ranges(domain);
/* calculate AGAW */
- iommu = domain->iommu;
+ iommu = domain_get_only_iommu(domain);
if (guest_width > cap_mgaw(iommu->cap))
guest_width = cap_mgaw(iommu->cap);
domain->gaw = guest_width;
@@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
sagaw = cap_sagaw(iommu->cap);
if (!test_bit(agaw, &sagaw)) {
/* hardware doesn't support it, choose a bigger one */
- pr_debug("IOMMU: hardware doesn't support agaw %d\n", agaw);
+ pr_debug("IOMMU: hardware doesn't support agaw %d\n",
+ agaw);
agaw = find_next_bit(&sagaw, 5, agaw);
if (agaw >= 5)
return -ENODEV;
}
domain->agaw = agaw;
+
INIT_LIST_HEAD(&domain->devices);
/* always allocate the top pgd */
domain->pgd = (struct dma_pte *)alloc_pgtable_page();
if (!domain->pgd)
return -ENOMEM;
- __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
+ domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
return 0;
}
@@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain *domain)
}
static int domain_context_mapping_one(struct dmar_domain *domain,
- u8 bus, u8 devfn)
+ struct intel_iommu *iommu, u8 bus, u8 devfn)
{
struct context_entry *context;
- struct intel_iommu *iommu = domain->iommu;
unsigned long flags;
pr_debug("Set context mapping for %02x:%02x.%d\n",
@@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
{
int ret;
struct pci_dev *tmp, *parent;
+ struct intel_iommu *iommu;
+
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return -ENODEV;
- ret = domain_context_mapping_one(domain, pdev->bus->number,
- pdev->devfn);
+ ret = domain_context_mapping_one(domain, iommu,
+ pdev->bus->number, pdev->devfn);
if (ret)
return ret;
@@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
/* Secondary interface's bus number and devfn 0 */
parent = pdev->bus->self;
while (parent != tmp) {
- ret = domain_context_mapping_one(domain, parent->bus->number,
- parent->devfn);
+ ret = domain_context_mapping_one(domain, iommu,
+ parent->bus->number, parent->devfn);
if (ret)
return ret;
parent = parent->bus->self;
}
if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
- return domain_context_mapping_one(domain,
+ return domain_context_mapping_one(domain, iommu,
tmp->subordinate->number, 0);
else /* this is a legacy PCI bridge */
- return domain_context_mapping_one(domain,
+ return domain_context_mapping_one(domain, iommu,
tmp->bus->number, tmp->devfn);
}
-static int domain_context_mapped(struct dmar_domain *domain,
- struct pci_dev *pdev)
+static int domain_context_mapped(struct pci_dev *pdev)
{
int ret;
struct pci_dev *tmp, *parent;
+ struct intel_iommu *iommu;
- ret = device_context_mapped(domain->iommu,
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return 0;
+
+ ret = device_context_mapped(iommu,
pdev->bus->number, pdev->devfn);
if (!ret)
return ret;
@@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct dmar_domain *domain,
/* Secondary interface's bus number and devfn 0 */
parent = pdev->bus->self;
while (parent != tmp) {
- ret = device_context_mapped(domain->iommu, parent->bus->number,
+ ret = device_context_mapped(iommu, parent->bus->number,
parent->devfn);
if (!ret)
return ret;
parent = parent->bus->self;
}
if (tmp->is_pcie)
- return device_context_mapped(domain->iommu,
+ return device_context_mapped(iommu,
tmp->subordinate->number, 0);
else
- return device_context_mapped(domain->iommu,
+ return device_context_mapped(iommu,
tmp->bus->number, tmp->devfn);
}
@@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
BUG_ON(dma_pte_addr(*pte));
dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
dma_set_pte_prot(*pte, prot);
- __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
+ domain_flush_cache(domain, pte, sizeof(*pte));
start_pfn++;
index++;
}
return 0;
}
-static void detach_domain_for_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
+static void detach_domain_for_dev(struct dmar_domain *domain,
+ u8 bus, u8 devfn)
{
- clear_context_table(domain->iommu, bus, devfn);
- domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
- DMA_CCMD_GLOBAL_INVL, 0);
- domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
- DMA_TLB_GLOBAL_FLUSH, 0);
+ struct intel_iommu *iommu;
+
+ iommu = device_find_matched_iommu(bus, devfn);
+ if (!iommu)
+ return;
+
+ clear_context_table(iommu, bus, devfn);
+ iommu->flush.flush_context(iommu, 0, 0, 0,
+ DMA_CCMD_GLOBAL_INVL, 0);
+ iommu->flush.flush_iotlb(iommu, 0, 0, 0,
+ DMA_TLB_GLOBAL_FLUSH, 0);
}
static void domain_remove_dev_info(struct dmar_domain *domain)
@@ -1336,7 +1425,6 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
{
struct dmar_domain *domain, *found = NULL;
struct intel_iommu *iommu;
- struct dmar_drhd_unit *drhd;
struct device_domain_info *info, *tmp;
struct pci_dev *dev_tmp;
unsigned long flags;
@@ -1371,13 +1459,9 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
}
/* Allocate new domain for the device */
- drhd = dmar_find_matched_drhd_unit(pdev);
- if (!drhd) {
- printk(KERN_ERR "IOMMU: can't find DMAR for device %s\n",
- pci_name(pdev));
- return NULL;
- }
- iommu = drhd->iommu;
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return NULL;
domain = iommu_alloc_domain(iommu);
if (!domain)
@@ -1400,7 +1484,7 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
info->dev = NULL;
info->domain = domain;
/* This domain is shared by devices under p2p bridge */
- domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
+ domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
/* pcie-to-pci bridge already has a domain, uses it */
found = NULL;
@@ -1805,7 +1889,7 @@ get_valid_domain_for_dev(struct pci_dev *pdev)
}
/* make sure context mapping is ok */
- if (unlikely(!domain_context_mapped(domain, pdev))) {
+ if (unlikely(!domain_context_mapped(pdev))) {
ret = domain_context_mapping(domain, pdev);
if (ret) {
printk(KERN_ERR
@@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
{
struct pci_dev *pdev = to_pci_dev(hwdev);
struct dmar_domain *domain;
+ struct intel_iommu *iommu;
phys_addr_t start_paddr;
struct iova *iova;
int prot = 0;
@@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
if (!domain)
return 0;
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return 0;
+
size = aligned_size((u64)paddr, size);
iova = __intel_alloc_iova(hwdev, domain, size, pdev->dma_mask);
@@ -1849,7 +1938,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
* mappings..
*/
if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
- !cap_zlr(domain->iommu->cap))
+ !cap_zlr(iommu->cap))
prot |= DMA_PTE_READ;
if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
prot |= DMA_PTE_WRITE;
@@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
goto error;
/* it's a non-present to present mapping */
- ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
+ ret = iommu_flush_iotlb_psi(iommu, domain->id,
start_paddr, size >> VTD_PAGE_SHIFT, 1);
if (ret)
- iommu_flush_write_buffer(domain->iommu);
+ iommu_flush_write_buffer(iommu);
return start_paddr + ((u64)paddr & (~PAGE_MASK));
@@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
/* just flush them all */
for (i = 0; i < g_num_of_iommus; i++) {
if (deferred_flush[i].next) {
- struct intel_iommu *iommu =
- deferred_flush[i].domain[0]->iommu;
+ struct intel_iommu *iommu = deferred_flush[i].iommu;
iommu->flush.flush_iotlb(iommu, 0, 0, 0,
DMA_TLB_GLOBAL_FLUSH, 0);
@@ -1921,7 +2009,8 @@ static void flush_unmaps_timeout(unsigned long data)
spin_unlock_irqrestore(&async_umap_flush_lock, flags);
}
-static void add_unmap(struct dmar_domain *dom, struct iova *iova)
+static void add_unmap(struct dmar_domain *dom,
+ struct intel_iommu *iommu, struct iova *iova)
{
unsigned long flags;
int next, iommu_id;
@@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova)
if (list_size == HIGH_WATER_MARK)
flush_unmaps();
- iommu_id = dom->iommu->seq_id;
+ iommu_id = iommu->seq_id;
next = deferred_flush[iommu_id].next;
deferred_flush[iommu_id].domain[next] = dom;
deferred_flush[iommu_id].iova[next] = iova;
+ deferred_flush[iommu_id].iommu = iommu;
deferred_flush[iommu_id].next++;
if (!timer_on) {
@@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
struct dmar_domain *domain;
unsigned long start_addr;
struct iova *iova;
+ struct intel_iommu *iommu;
if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
return;
domain = find_domain(pdev);
BUG_ON(!domain);
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return;
+
iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
if (!iova)
return;
@@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
/* free page tables */
dma_pte_free_pagetable(domain, start_addr, start_addr + size);
if (intel_iommu_strict) {
- if (iommu_flush_iotlb_psi(domain->iommu,
+ if (iommu_flush_iotlb_psi(iommu,
domain->id, start_addr, size >> VTD_PAGE_SHIFT, 0))
- iommu_flush_write_buffer(domain->iommu);
+ iommu_flush_write_buffer(iommu);
/* free iova */
__free_iova(&domain->iovad, iova);
} else {
- add_unmap(domain, iova);
+ add_unmap(domain, iommu, iova);
/*
* queue up the release of the unmap to save the 1/6th of the
* cpu used up by the iotlb flush operation...
@@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
size_t size = 0;
void *addr;
struct scatterlist *sg;
+ struct intel_iommu *iommu;
if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
return;
domain = find_domain(pdev);
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return;
+
iova = find_iova(&domain->iovad, IOVA_PFN(sglist[0].dma_address));
if (!iova)
return;
@@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
/* free page tables */
dma_pte_free_pagetable(domain, start_addr, start_addr + size);
- if (iommu_flush_iotlb_psi(domain->iommu, domain->id, start_addr,
+ if (iommu_flush_iotlb_psi(iommu, domain->id, start_addr,
size >> VTD_PAGE_SHIFT, 0))
- iommu_flush_write_buffer(domain->iommu);
+ iommu_flush_write_buffer(iommu);
/* free iova */
__free_iova(&domain->iovad, iova);
@@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
int ret;
struct scatterlist *sg;
unsigned long start_addr;
+ struct intel_iommu *iommu;
BUG_ON(dir == DMA_NONE);
if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
@@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
if (!domain)
return 0;
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return 0;
+
for_each_sg(sglist, sg, nelems, i) {
addr = SG_ENT_VIRT_ADDRESS(sg);
addr = (void *)virt_to_phys(addr);
@@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
* mappings..
*/
if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
- !cap_zlr(domain->iommu->cap))
+ !cap_zlr(iommu->cap))
prot |= DMA_PTE_READ;
if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
prot |= DMA_PTE_WRITE;
@@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
}
/* it's a non-present to present mapping */
- if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
+ if (iommu_flush_iotlb_psi(iommu, domain->id,
start_addr, offset >> VTD_PAGE_SHIFT, 1))
- iommu_flush_write_buffer(domain->iommu);
+ iommu_flush_write_buffer(iommu);
return nelems;
}
@@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
return 0;
}
-void intel_iommu_domain_exit(struct dmar_domain *domain)
+/* domain id for virtual machine, it won't be set in context */
+static unsigned long vm_domid;
+
+static int vm_domain_min_agaw(struct dmar_domain *domain)
+{
+ struct dmar_drhd_unit *drhd;
+ struct intel_iommu *iommu;
+ int min_agaw = domain->agaw;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+ iommu = drhd->iommu;
+
+ if (test_bit(iommu->seq_id, &domain->iommu_bmp))
+ if (min_agaw > iommu->agaw)
+ min_agaw = iommu->agaw;
+ }
+
+ return min_agaw;
+}
+
+static int vm_domain_add_dev_info(struct dmar_domain *domain,
+ struct pci_dev *pdev)
+{
+ struct device_domain_info *info;
+ unsigned long flags;
+
+ info = alloc_devinfo_mem();
+ if (!info)
+ return -ENOMEM;
+
+ info->bus = pdev->bus->number;
+ info->devfn = pdev->devfn;
+ info->dev = pdev;
+ info->domain = domain;
+
+ spin_lock_irqsave(&device_domain_lock, flags);
+ list_add(&info->link, &domain->devices);
+ list_add(&info->global, &device_domain_list);
+ pdev->dev.archdata.iommu = info;
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+
+ return 0;
+}
+
+static void vm_domain_remove_one_dev_info(struct dmar_domain *domain,
+ struct pci_dev *pdev)
+{
+ struct device_domain_info *info;
+ struct intel_iommu *iommu;
+ unsigned long flags;
+ int found = 0;
+
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+
+ spin_lock_irqsave(&device_domain_lock, flags);
+ while (!list_empty(&domain->devices)) {
+ info = list_entry(domain->devices.next,
+ struct device_domain_info, link);
+ if (info->bus == pdev->bus->number &&
+ info->devfn == pdev->devfn) {
+ list_del(&info->link);
+ list_del(&info->global);
+ if (info->dev)
+ info->dev->dev.archdata.iommu = NULL;
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+
+ detach_domain_for_dev(info->domain,
+ info->bus, info->devfn);
+ free_devinfo_mem(info);
+
+ spin_lock_irqsave(&device_domain_lock, flags);
+
+ if (found)
+ break;
+ else
+ continue;
+ }
+
+ /* if there is no other devices under the same iommu
+ * owned by this domain, clear this iommu in iommu_bmp
+ */
+ if (device_find_matched_iommu(info->bus, info->devfn) == iommu)
+ found = 1;
+ }
+
+ if (found == 0) {
+ spin_lock_irqsave(&iommu->lock, flags);
+ clear_bit(iommu->seq_id, &domain->iommu_bmp);
+ domain->iommu_count--;
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ }
+
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+}
+
+static void vm_domain_remove_all_dev_info(struct dmar_domain *domain)
+{
+ struct device_domain_info *info;
+ struct intel_iommu *iommu;
+ unsigned long flags;
+
+ spin_lock_irqsave(&device_domain_lock, flags);
+ while (!list_empty(&domain->devices)) {
+ info = list_entry(domain->devices.next,
+ struct device_domain_info, link);
+ list_del(&info->link);
+ list_del(&info->global);
+ if (info->dev)
+ info->dev->dev.archdata.iommu = NULL;
+
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+ detach_domain_for_dev(info->domain,
+ info->bus, info->devfn);
+
+ /* clear this iommu in iommu_bmp */
+ iommu = device_find_matched_iommu(info->bus, info->devfn);
+ spin_lock_irqsave(&iommu->lock, flags);
+ if (test_and_clear_bit(iommu->seq_id,
+ &domain->iommu_bmp))
+ domain->iommu_count--;
+ spin_unlock_irqrestore(&iommu->lock, flags);
+
+ free_devinfo_mem(info);
+ spin_lock_irqsave(&device_domain_lock, flags);
+ }
+ spin_unlock_irqrestore(&device_domain_lock, flags);
+}
+
+static int vm_domain_context_mapping_one(struct dmar_domain *domain,
+ struct intel_iommu *iommu, u8 bus, u8 devfn)
+{
+ struct context_entry *context;
+ unsigned long flags;
+ struct dma_pte *pgd;
+ unsigned long num;
+ unsigned long ndomains;
+ int id;
+ int agaw;
+ int found = 0;
+
+ pr_debug("Set context mapping for %02x:%02x.%d\n",
+ bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
+ BUG_ON(!domain->pgd);
+ context = device_to_context_entry(iommu, bus, devfn);
+ if (!context)
+ return -ENOMEM;
+ spin_lock_irqsave(&iommu->lock, flags);
+ if (context_present(*context)) {
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ return 0;
+ }
+
+ id = domain->id;
+
+ /* find an available domain id for this device in iommu */
+ ndomains = cap_ndoms(iommu->cap);
+ num = find_first_bit(iommu->domain_ids, ndomains);
+ for (; num < ndomains; ) {
+ if (iommu->domains[num] == domain) {
+ id = num;
+ found = 1;
+ break;
+ }
+ num = find_next_bit(iommu->domain_ids,
+ cap_ndoms(iommu->cap), num+1);
+ }
+
+ if (found == 0) {
+ num = find_first_zero_bit(iommu->domain_ids, ndomains);
+ if (num >= ndomains) {
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ printk(KERN_ERR "IOMMU: no free domain ids\n");
+ return -EFAULT;
+ }
+
+ set_bit(num, iommu->domain_ids);
+ iommu->domains[num] = domain;
+ id = num;
+ }
+
+ pgd = domain->pgd;
+
+ /* Skip top levels of page tables for
+ * iommu which has less agaw than default.
+ */
+ for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
+ pgd = phys_to_virt(dma_pte_addr(*pgd));
+ if (!dma_pte_present(*pgd)) {
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ return -ENOMEM;
+ }
+ }
+
+ context_set_domain_id(*context, id);
+ context_set_address_width(*context, iommu->agaw);
+ context_set_address_root(*context, virt_to_phys(pgd));
+ context_set_translation_type(*context, CONTEXT_TT_MULTI_LEVEL);
+ context_set_fault_enable(*context);
+ context_set_present(*context);
+ __iommu_flush_cache(iommu, context, sizeof(*context));
+
+ /* it's a non-present to present mapping */
+ if (iommu->flush.flush_context(iommu, id,
+ (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
+ DMA_CCMD_DEVICE_INVL, 1))
+ iommu_flush_write_buffer(iommu);
+ else
+ iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_DSI_FLUSH, 0);
+
+ if (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp))
+ domain->iommu_count++;
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ return 0;
+}
+
+static int
+vm_domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
+{
+ int ret;
+ struct pci_dev *tmp, *parent;
+ struct intel_iommu *iommu;
+
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return -ENODEV;
+
+ ret = vm_domain_context_mapping_one(domain, iommu,
+ pdev->bus->number, pdev->devfn);
+ if (ret)
+ return ret;
+
+ /* dependent device mapping */
+ tmp = pci_find_upstream_pcie_bridge(pdev);
+ if (!tmp)
+ return 0;
+ /* Secondary interface's bus number and devfn 0 */
+ parent = pdev->bus->self;
+ while (parent != tmp) {
+ ret = vm_domain_context_mapping_one(domain, iommu,
+ parent->bus->number, parent->devfn);
+ if (ret)
+ return ret;
+ parent = parent->bus->self;
+ }
+ if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
+ return vm_domain_context_mapping_one(domain, iommu,
+ tmp->subordinate->number, 0);
+ else /* this is a legacy PCI bridge */
+ return vm_domain_context_mapping_one(domain, iommu,
+ tmp->bus->number, tmp->devfn);
+}
+
+
+static int vm_domain_init(struct dmar_domain *domain, int guest_width)
+{
+ int adjust_width;
+
+ init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
+ spin_lock_init(&domain->mapping_lock);
+
+ domain_reserve_special_ranges(domain);
+
+ /* calculate AGAW */
+ domain->gaw = guest_width;
+ adjust_width = guestwidth_to_adjustwidth(guest_width);
+ domain->agaw = width_to_agaw(adjust_width);
+
+ INIT_LIST_HEAD(&domain->devices);
+
+ /* always allocate the top pgd */
+ domain->pgd = (struct dma_pte *)alloc_pgtable_page();
+ if (!domain->pgd)
+ return -ENOMEM;
+ domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
+ return 0;
+}
+
+static void iommu_free_vm_domain(struct dmar_domain *domain)
+{
+ unsigned long flags;
+ struct dmar_drhd_unit *drhd;
+ struct intel_iommu *iommu;
+ unsigned long i;
+ unsigned long ndomains;
+
+ for_each_drhd_unit(drhd) {
+ if (drhd->ignored)
+ continue;
+ iommu = drhd->iommu;
+
+ ndomains = cap_ndoms(iommu->cap);
+ i = find_first_bit(iommu->domain_ids, ndomains);
+ for (; i < ndomains; ) {
+ if (iommu->domains[i] == domain) {
+ spin_lock_irqsave(&iommu->lock, flags);
+ clear_bit(i, iommu->domain_ids);
+ iommu->domains[i] = NULL;
+ spin_unlock_irqrestore(&iommu->lock, flags);
+ break;
+ }
+ i = find_next_bit(iommu->domain_ids, ndomains, i+1);
+ }
+ }
+}
+
+static void vm_domain_exit(struct dmar_domain *domain)
{
u64 end;
@@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
if (!domain)
return;
+ vm_domain_remove_all_dev_info(domain);
+ /* destroy iovas */
+ put_iova_domain(&domain->iovad);
end = DOMAIN_MAX_ADDR(domain->gaw);
- end = end & (~VTD_PAGE_MASK);
+ end &= VTD_PAGE_MASK;
/* clear ptes */
dma_pte_clear_range(domain, 0, end);
@@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
/* free page tables */
dma_pte_free_pagetable(domain, 0, end);
- iommu_free_domain(domain);
+ iommu_free_vm_domain(domain);
free_domain_mem(domain);
}
-EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
-struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
+static struct dmar_domain *iommu_alloc_vm_domain(void)
{
- struct dmar_drhd_unit *drhd;
struct dmar_domain *domain;
- struct intel_iommu *iommu;
- drhd = dmar_find_matched_drhd_unit(pdev);
- if (!drhd) {
- printk(KERN_ERR "intel_iommu_domain_alloc: drhd == NULL\n");
+ domain = alloc_domain_mem();
+ if (!domain)
return NULL;
- }
- iommu = drhd->iommu;
- if (!iommu) {
- printk(KERN_ERR
- "intel_iommu_domain_alloc: iommu == NULL\n");
- return NULL;
- }
- domain = iommu_alloc_domain(iommu);
+ domain->id = vm_domid++;
+ domain->iommu_count = 0;
+ domain->max_addr = 0;
+ memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
+ domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE;
+
+ return domain;
+}
+
+struct dmar_domain *intel_iommu_alloc_domain(void)
+{
+ struct dmar_domain *domain;
+
+ domain = iommu_alloc_vm_domain();
if (!domain) {
printk(KERN_ERR
"intel_iommu_domain_alloc: domain == NULL\n");
return NULL;
}
- if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
+ if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
printk(KERN_ERR
"intel_iommu_domain_alloc: domain_init() failed\n");
- intel_iommu_domain_exit(domain);
+ vm_domain_exit(domain);
return NULL;
}
+
return domain;
}
-EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
+EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
+
+void intel_iommu_free_domain(struct dmar_domain *domain)
+{
+ vm_domain_exit(domain);
+}
+EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
-int intel_iommu_context_mapping(
- struct dmar_domain *domain, struct pci_dev *pdev)
+int intel_iommu_assign_device(struct dmar_domain *domain,
+ struct pci_dev *pdev)
{
- int rc;
- rc = domain_context_mapping(domain, pdev);
- return rc;
+ struct intel_iommu *iommu;
+ int addr_width;
+ u64 end;
+ int ret;
+
+ /* normally pdev is not mapped */
+ if (unlikely(domain_context_mapped(pdev))) {
+ struct dmar_domain *old_domain;
+
+ old_domain = find_domain(pdev);
+ if (old_domain) {
+ if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE)
+ vm_domain_remove_one_dev_info(old_domain, pdev);
+ else
+ domain_remove_dev_info(old_domain);
+ }
+ }
+
+ iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
+ if (!iommu)
+ return -ENODEV;
+
+ /* check if this iommu agaw is sufficient for max mapped address */
+ addr_width = agaw_to_width(iommu->agaw);
+ end = DOMAIN_MAX_ADDR(addr_width);
+ end = end & VTD_PAGE_MASK;
+ if (end < domain->max_addr) {
+ printk(KERN_ERR "%s: iommu agaw (%d) is not "
+ "sufficient for the mapped address (%llx)\n",
+ __func__, iommu->agaw, domain->max_addr);
+ return -EFAULT;
+ }
+
+ ret = vm_domain_context_mapping(domain, pdev);
+ if (ret)
+ return ret;
+
+ ret = vm_domain_add_dev_info(domain, pdev);
+ return ret;
}
-EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
+EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
-int intel_iommu_page_mapping(
- struct dmar_domain *domain, dma_addr_t iova,
- u64 hpa, size_t size, int prot)
+
+void intel_iommu_deassign_device(struct dmar_domain *domain,
+ struct pci_dev *pdev)
{
- int rc;
- rc = domain_page_mapping(domain, iova, hpa, size, prot);
- return rc;
+ vm_domain_remove_one_dev_info(domain, pdev);
}
-EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
+EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
-void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
+int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
+ u64 hpa, size_t size, int prot)
{
- detach_domain_for_dev(domain, bus, devfn);
+ u64 max_addr;
+ int addr_width;
+ int ret;
+
+ max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
+ if (domain->max_addr < max_addr) {
+ int min_agaw;
+ u64 end;
+
+ /* check if minimum agaw is sufficient for mapped address */
+ min_agaw = vm_domain_min_agaw(domain);
+ addr_width = agaw_to_width(min_agaw);
+ end = DOMAIN_MAX_ADDR(addr_width);
+ end = end & VTD_PAGE_MASK;
+ if (end < max_addr) {
+ printk(KERN_ERR "%s: iommu agaw (%d) is not "
+ "sufficient for the mapped address (%llx)\n",
+ __func__, min_agaw, max_addr);
+ return -EFAULT;
+ }
+ domain->max_addr = max_addr;
+ }
+
+ ret = domain_page_mapping(domain, iova, hpa, size, prot);
+ return ret;
}
-EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
+EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
-struct dmar_domain *
-intel_iommu_find_domain(struct pci_dev *pdev)
+void intel_iommu_unmap_pages(struct dmar_domain *domain,
+ dma_addr_t iova, size_t size)
{
- return find_domain(pdev);
+ dma_addr_t base;
+
+ /* The address might not be aligned */
+ base = iova & PAGE_MASK;
+ size = PAGE_ALIGN(size);
+ dma_pte_clear_range(domain, base, base + size);
}
-EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
+EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
int intel_iommu_found(void)
{
diff --git a/include/linux/dma_remapping.h b/include/linux/dma_remapping.h
index 952df39..3568ae6 100644
--- a/include/linux/dma_remapping.h
+++ b/include/linux/dma_remapping.h
@@ -111,11 +111,21 @@ struct dma_pte {
(p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
#define dma_pte_present(p) (((p).val & 3) != 0)
+/* domain flags, one domain owns one device by default */
+
+/* devices under the same p2p bridge are owned in one domain */
+#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES (1 << 0)
+
+/* domain represents a virtual machine, more than one devices
+ * across iommus may be owned in one domain, e.g. kvm guest.
+ */
+#define DOMAIN_FLAG_VIRTUAL_MACHINE (1 << 1)
+
struct intel_iommu;
struct dmar_domain {
int id; /* domain id */
- struct intel_iommu *iommu; /* back pointer to owning iommu */
+ unsigned long iommu_bmp; /* bitmap of iommus this domain uses*/
struct list_head devices; /* all devices' list */
struct iova_domain iovad; /* iova's that belong to this domain */
@@ -123,12 +133,13 @@ struct dmar_domain {
struct dma_pte *pgd; /* virtual address */
spinlock_t mapping_lock; /* page table lock */
int gaw; /* max guest address width */
+ int agaw; /* adjusted guest address width */
- /* adjusted guest address width, 0 is level 2 30-bit */
- int agaw;
+ int flags; /* domain flag */
-#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
- int flags;
+ /* following fields are used in virtual machine case */
+ int iommu_count; /* reference count of iommu */
+ u64 max_addr; /* maximum mapped address */
};
/* PCI domain-device relationship */
diff --git a/include/linux/intel-iommu.h b/include/linux/intel-iommu.h
index 3d017cf..c2f37b8 100644
--- a/include/linux/intel-iommu.h
+++ b/include/linux/intel-iommu.h
@@ -219,6 +219,8 @@ do { \
} \
} while (0)
+#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
+
#define QI_LENGTH 256 /* queue length */
enum {
@@ -299,6 +301,7 @@ struct intel_iommu {
struct dmar_domain **domains; /* ptr to domains */
spinlock_t lock; /* protect context, domain ids */
struct root_entry *root_entry; /* virtual address */
+ int agaw;
unsigned int irq;
unsigned char name[7]; /* Device Name */
@@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu *iommu, u16 did, u64 addr,
extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu *iommu);
-void intel_iommu_domain_exit(struct dmar_domain *domain);
-struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
-int intel_iommu_context_mapping(struct dmar_domain *domain,
- struct pci_dev *pdev);
-int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
- u64 hpa, size_t size, int prot);
-void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn);
-struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
+struct dmar_domain *intel_iommu_alloc_domain(void);
+void intel_iommu_free_domain(struct dmar_domain *domain);
+int intel_iommu_assign_device(struct dmar_domain *domain,
+ struct pci_dev *pdev);
+void intel_iommu_deassign_device(struct dmar_domain *domain,
+ struct pci_dev *pdev);
+int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
+ u64 hpa, size_t size, int prot);
+void intel_iommu_unmap_pages(struct dmar_domain *domain,
+ dma_addr_t iova, size_t size);
u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
#ifdef CONFIG_DMAR
--
1.5.1
^ permalink raw reply related [flat|nested] 7+ messages in thread
* Re: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
2008-11-27 13:49 [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM Han, Weidong
@ 2008-11-28 13:27 ` Joerg Roedel
2008-12-01 6:17 ` Han, Weidong
0 siblings, 1 reply; 7+ messages in thread
From: Joerg Roedel @ 2008-11-28 13:27 UTC (permalink / raw)
To: Han, Weidong
Cc: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity',
Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
'iommu@lists.linux-foundation.org'
Hmm, I tried to apply this patch against avi/master and linus/master but
get merge conflicts. Where do these patches apply cleanly?
Joerg
On Thu, Nov 27, 2008 at 09:49:04PM +0800, Han, Weidong wrote:
> In order to support multiple device assignment for KVM, this patch does following main changes:
> - extend dmar_domain to own multiple devices from different iommus, use a bitmap of iommus to replace iommu pointer in dmar_domain.
> - implement independent low level functions for kvm, then won't impact native VT-d.
> - "SAGAW" capability may be different across iommus, that's to say the VT-d page table levels may be different among iommus. This patch uses a defaut agaw, and skip top levels of page tables for iommus which have smaller agaw than default.
> - rename the APIs for kvm VT-d, make it more readable.
>
>
> Signed-off-by: Weidong Han <weidong.han@intel.com>
> ---
> drivers/pci/dmar.c | 15 +
> drivers/pci/intel-iommu.c | 698 ++++++++++++++++++++++++++++++++++------
> include/linux/dma_remapping.h | 21 +-
> include/linux/intel-iommu.h | 21 +-
> 4 files changed, 637 insertions(+), 118 deletions(-)
>
> diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
> index 691b3ad..d6bdced 100644
> --- a/drivers/pci/dmar.c
> +++ b/drivers/pci/dmar.c
> @@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
> dmar_tbl = NULL;
> }
>
> +extern int width_to_agaw(int width);
>
> int alloc_iommu(struct dmar_drhd_unit *drhd)
> {
> @@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
> int map_size;
> u32 ver;
> static int iommu_allocated = 0;
> + unsigned long sagaw;
> + int agaw;
>
> iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);
> if (!iommu)
> @@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
> iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
> iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
>
> + /* set agaw, "SAGAW" may be different across iommus */
> + sagaw = cap_sagaw(iommu->cap);
> + for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
> + agaw >= 0; agaw--)
> + if (test_bit(agaw, &sagaw))
> + break;
> + if (agaw < 0) {
> + printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n", sagaw);
> + goto error;
> + }
> + iommu->agaw = agaw;
> +
> /* the registers might be more than one page */
> map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
> cap_max_fault_reg_offset(iommu->cap));
> diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
> index 5c8baa4..55b96c4 100644
> --- a/drivers/pci/intel-iommu.c
> +++ b/drivers/pci/intel-iommu.c
> @@ -50,8 +50,6 @@
> #define IOAPIC_RANGE_END (0xfeefffff)
> #define IOVA_START_ADDR (0x1000)
>
> -#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
> -
> #define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
>
>
> @@ -64,6 +62,7 @@ struct deferred_flush_tables {
> int next;
> struct iova *iova[HIGH_WATER_MARK];
> struct dmar_domain *domain[HIGH_WATER_MARK];
> + struct intel_iommu *iommu;
> };
>
> static struct deferred_flush_tables *deferred_flush;
> @@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
> kmem_cache_free(iommu_iova_cache, iova);
> }
>
> +/* in native case, each domain is related to only one iommu */
> +static struct intel_iommu *domain_get_only_iommu(struct dmar_domain *domain)
> +{
> + struct dmar_drhd_unit *drhd;
> +
> + for_each_drhd_unit(drhd) {
> + if (drhd->ignored)
> + continue;
> + if (test_bit(drhd->iommu->seq_id, &domain->iommu_bmp))
> + return drhd->iommu;
> + }
> +
> + return NULL;
> +}
> +
> +static void domain_flush_cache(struct dmar_domain *domain,
> + void *addr, int size)
> +{
> + struct intel_iommu *iommu;
> +
> + if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
> + struct dmar_drhd_unit *drhd;
> +
> + for_each_drhd_unit(drhd) {
> + if (drhd->ignored)
> + continue;
> + iommu = drhd->iommu;
> +
> + if (!test_bit(iommu->seq_id, &domain->iommu_bmp))
> + continue;
> +
> + if (!ecap_coherent(iommu->ecap))
> + clflush_cache_range(addr, size);
> + }
> + }
> + else {
> + iommu = domain_get_only_iommu(domain);
> + if (iommu && !ecap_coherent(iommu->ecap))
> + clflush_cache_range(addr, size);
> + }
> +}
> +
> +static struct intel_iommu *device_find_matched_iommu(u8 bus, u8 devfn)
> +{
> + struct dmar_drhd_unit *drhd = NULL;
> + int i;
> +
> + for_each_drhd_unit(drhd) {
> + if (drhd->ignored)
> + continue;
> +
> + for (i = 0; i < drhd->devices_cnt; i++)
> + if (drhd->devices[i]->bus->number == bus &&
> + drhd->devices[i]->devfn == devfn)
> + return drhd->iommu;
> +
> + if (drhd->include_all)
> + return drhd->iommu;
> + }
> +
> + return NULL;
> +}
> +
> /* Gets context entry for a given bus and devfn */
> static struct context_entry * device_to_context_entry(struct intel_iommu *iommu,
> u8 bus, u8 devfn)
> @@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
>
> }
>
> -static inline int width_to_agaw(int width)
> +int width_to_agaw(int width)
> {
> return (width - 30) / LEVEL_STRIDE;
> }
> @@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
> flags);
> return NULL;
> }
> - __iommu_flush_cache(domain->iommu, tmp_page,
> - PAGE_SIZE);
> + domain_flush_cache(domain, tmp_page, PAGE_SIZE);
> dma_set_pte_addr(*pte, virt_to_phys(tmp_page));
> /*
> * high level table always sets r/w, last level page
> @@ -356,7 +417,7 @@ static struct dma_pte * addr_to_dma_pte(struct dmar_domain *domain, u64 addr)
> */
> dma_set_pte_readable(*pte);
> dma_set_pte_writable(*pte);
> - __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
> + domain_flush_cache(domain, pte, sizeof(*pte));
> }
> parent = phys_to_virt(dma_pte_addr(*pte));
> level--;
> @@ -399,7 +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain, u64 addr)
>
> if (pte) {
> dma_clear_pte(*pte);
> - __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
> + domain_flush_cache(domain, pte, sizeof(*pte));
> }
> }
>
> @@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct dmar_domain *domain,
> free_pgtable_page(
> phys_to_virt(dma_pte_addr(*pte)));
> dma_clear_pte(*pte);
> - __iommu_flush_cache(domain->iommu,
> - pte, sizeof(*pte));
> + domain_flush_cache(domain, pte, sizeof(*pte));
> }
> tmp += level_size(level);
> }
> @@ -948,8 +1008,8 @@ static int iommu_init_domains(struct intel_iommu *iommu)
> return 0;
> }
>
> -
> static void domain_exit(struct dmar_domain *domain);
> +static void vm_domain_exit(struct dmar_domain *domain);
>
> void free_dmar_iommu(struct intel_iommu *iommu)
> {
> @@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
> for (; i < cap_ndoms(iommu->cap); ) {
> domain = iommu->domains[i];
> clear_bit(i, iommu->domain_ids);
> - domain_exit(domain);
> +
> + if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
> + if (--domain->iommu_count == 0)
> + vm_domain_exit(domain);
> + }
> + else
> + domain_exit(domain);
> +
> i = find_next_bit(iommu->domain_ids,
> cap_ndoms(iommu->cap), i+1);
> }
> @@ -1006,8 +1073,11 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
>
> set_bit(num, iommu->domain_ids);
> domain->id = num;
> - domain->iommu = iommu;
> + memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
> + set_bit(iommu->seq_id, &domain->iommu_bmp);
> iommu->domains[num] = domain;
> + domain->iommu_count = 1;
> + domain->flags = 0;
> spin_unlock_irqrestore(&iommu->lock, flags);
>
> return domain;
> @@ -1016,10 +1086,12 @@ static struct dmar_domain * iommu_alloc_domain(struct intel_iommu *iommu)
> static void iommu_free_domain(struct dmar_domain *domain)
> {
> unsigned long flags;
> + struct intel_iommu *iommu;
>
> - spin_lock_irqsave(&domain->iommu->lock, flags);
> - clear_bit(domain->id, domain->iommu->domain_ids);
> - spin_unlock_irqrestore(&domain->iommu->lock, flags);
> + iommu = domain_get_only_iommu(domain);
> + spin_lock_irqsave(&iommu->lock, flags);
> + clear_bit(domain->id, iommu->domain_ids);
> + spin_unlock_irqrestore(&iommu->lock, flags);
> }
>
> static struct iova_domain reserved_iova_list;
> @@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
> domain_reserve_special_ranges(domain);
>
> /* calculate AGAW */
> - iommu = domain->iommu;
> + iommu = domain_get_only_iommu(domain);
> if (guest_width > cap_mgaw(iommu->cap))
> guest_width = cap_mgaw(iommu->cap);
> domain->gaw = guest_width;
> @@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
> sagaw = cap_sagaw(iommu->cap);
> if (!test_bit(agaw, &sagaw)) {
> /* hardware doesn't support it, choose a bigger one */
> - pr_debug("IOMMU: hardware doesn't support agaw %d\n", agaw);
> + pr_debug("IOMMU: hardware doesn't support agaw %d\n",
> + agaw);
> agaw = find_next_bit(&sagaw, 5, agaw);
> if (agaw >= 5)
> return -ENODEV;
> }
> domain->agaw = agaw;
> +
> INIT_LIST_HEAD(&domain->devices);
>
> /* always allocate the top pgd */
> domain->pgd = (struct dma_pte *)alloc_pgtable_page();
> if (!domain->pgd)
> return -ENOMEM;
> - __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
> + domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
> return 0;
> }
>
> @@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain *domain)
> }
>
> static int domain_context_mapping_one(struct dmar_domain *domain,
> - u8 bus, u8 devfn)
> + struct intel_iommu *iommu, u8 bus, u8 devfn)
> {
> struct context_entry *context;
> - struct intel_iommu *iommu = domain->iommu;
> unsigned long flags;
>
> pr_debug("Set context mapping for %02x:%02x.%d\n",
> @@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
> {
> int ret;
> struct pci_dev *tmp, *parent;
> + struct intel_iommu *iommu;
> +
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return -ENODEV;
>
> - ret = domain_context_mapping_one(domain, pdev->bus->number,
> - pdev->devfn);
> + ret = domain_context_mapping_one(domain, iommu,
> + pdev->bus->number, pdev->devfn);
> if (ret)
> return ret;
>
> @@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
> /* Secondary interface's bus number and devfn 0 */
> parent = pdev->bus->self;
> while (parent != tmp) {
> - ret = domain_context_mapping_one(domain, parent->bus->number,
> - parent->devfn);
> + ret = domain_context_mapping_one(domain, iommu,
> + parent->bus->number, parent->devfn);
> if (ret)
> return ret;
> parent = parent->bus->self;
> }
> if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
> - return domain_context_mapping_one(domain,
> + return domain_context_mapping_one(domain, iommu,
> tmp->subordinate->number, 0);
> else /* this is a legacy PCI bridge */
> - return domain_context_mapping_one(domain,
> + return domain_context_mapping_one(domain, iommu,
> tmp->bus->number, tmp->devfn);
> }
>
> -static int domain_context_mapped(struct dmar_domain *domain,
> - struct pci_dev *pdev)
> +static int domain_context_mapped(struct pci_dev *pdev)
> {
> int ret;
> struct pci_dev *tmp, *parent;
> + struct intel_iommu *iommu;
>
> - ret = device_context_mapped(domain->iommu,
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return 0;
> +
> + ret = device_context_mapped(iommu,
> pdev->bus->number, pdev->devfn);
> if (!ret)
> return ret;
> @@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct dmar_domain *domain,
> /* Secondary interface's bus number and devfn 0 */
> parent = pdev->bus->self;
> while (parent != tmp) {
> - ret = device_context_mapped(domain->iommu, parent->bus->number,
> + ret = device_context_mapped(iommu, parent->bus->number,
> parent->devfn);
> if (!ret)
> return ret;
> parent = parent->bus->self;
> }
> if (tmp->is_pcie)
> - return device_context_mapped(domain->iommu,
> + return device_context_mapped(iommu,
> tmp->subordinate->number, 0);
> else
> - return device_context_mapped(domain->iommu,
> + return device_context_mapped(iommu,
> tmp->bus->number, tmp->devfn);
> }
>
> @@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
> BUG_ON(dma_pte_addr(*pte));
> dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
> dma_set_pte_prot(*pte, prot);
> - __iommu_flush_cache(domain->iommu, pte, sizeof(*pte));
> + domain_flush_cache(domain, pte, sizeof(*pte));
> start_pfn++;
> index++;
> }
> return 0;
> }
>
> -static void detach_domain_for_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
> +static void detach_domain_for_dev(struct dmar_domain *domain,
> + u8 bus, u8 devfn)
> {
> - clear_context_table(domain->iommu, bus, devfn);
> - domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
> - DMA_CCMD_GLOBAL_INVL, 0);
> - domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
> - DMA_TLB_GLOBAL_FLUSH, 0);
> + struct intel_iommu *iommu;
> +
> + iommu = device_find_matched_iommu(bus, devfn);
> + if (!iommu)
> + return;
> +
> + clear_context_table(iommu, bus, devfn);
> + iommu->flush.flush_context(iommu, 0, 0, 0,
> + DMA_CCMD_GLOBAL_INVL, 0);
> + iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> + DMA_TLB_GLOBAL_FLUSH, 0);
> }
>
> static void domain_remove_dev_info(struct dmar_domain *domain)
> @@ -1336,7 +1425,6 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
> {
> struct dmar_domain *domain, *found = NULL;
> struct intel_iommu *iommu;
> - struct dmar_drhd_unit *drhd;
> struct device_domain_info *info, *tmp;
> struct pci_dev *dev_tmp;
> unsigned long flags;
> @@ -1371,13 +1459,9 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
> }
>
> /* Allocate new domain for the device */
> - drhd = dmar_find_matched_drhd_unit(pdev);
> - if (!drhd) {
> - printk(KERN_ERR "IOMMU: can't find DMAR for device %s\n",
> - pci_name(pdev));
> - return NULL;
> - }
> - iommu = drhd->iommu;
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return NULL;
>
> domain = iommu_alloc_domain(iommu);
> if (!domain)
> @@ -1400,7 +1484,7 @@ static struct dmar_domain *get_domain_for_dev(struct pci_dev *pdev, int gaw)
> info->dev = NULL;
> info->domain = domain;
> /* This domain is shared by devices under p2p bridge */
> - domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
> + domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
>
> /* pcie-to-pci bridge already has a domain, uses it */
> found = NULL;
> @@ -1805,7 +1889,7 @@ get_valid_domain_for_dev(struct pci_dev *pdev)
> }
>
> /* make sure context mapping is ok */
> - if (unlikely(!domain_context_mapped(domain, pdev))) {
> + if (unlikely(!domain_context_mapped(pdev))) {
> ret = domain_context_mapping(domain, pdev);
> if (ret) {
> printk(KERN_ERR
> @@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
> {
> struct pci_dev *pdev = to_pci_dev(hwdev);
> struct dmar_domain *domain;
> + struct intel_iommu *iommu;
> phys_addr_t start_paddr;
> struct iova *iova;
> int prot = 0;
> @@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
> if (!domain)
> return 0;
>
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return 0;
> +
> size = aligned_size((u64)paddr, size);
>
> iova = __intel_alloc_iova(hwdev, domain, size, pdev->dma_mask);
> @@ -1849,7 +1938,7 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
> * mappings..
> */
> if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
> - !cap_zlr(domain->iommu->cap))
> + !cap_zlr(iommu->cap))
> prot |= DMA_PTE_READ;
> if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
> prot |= DMA_PTE_WRITE;
> @@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct device *hwdev, phys_addr_t paddr,
> goto error;
>
> /* it's a non-present to present mapping */
> - ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
> + ret = iommu_flush_iotlb_psi(iommu, domain->id,
> start_paddr, size >> VTD_PAGE_SHIFT, 1);
> if (ret)
> - iommu_flush_write_buffer(domain->iommu);
> + iommu_flush_write_buffer(iommu);
>
> return start_paddr + ((u64)paddr & (~PAGE_MASK));
>
> @@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
> /* just flush them all */
> for (i = 0; i < g_num_of_iommus; i++) {
> if (deferred_flush[i].next) {
> - struct intel_iommu *iommu =
> - deferred_flush[i].domain[0]->iommu;
> + struct intel_iommu *iommu = deferred_flush[i].iommu;
>
> iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> DMA_TLB_GLOBAL_FLUSH, 0);
> @@ -1921,7 +2009,8 @@ static void flush_unmaps_timeout(unsigned long data)
> spin_unlock_irqrestore(&async_umap_flush_lock, flags);
> }
>
> -static void add_unmap(struct dmar_domain *dom, struct iova *iova)
> +static void add_unmap(struct dmar_domain *dom,
> + struct intel_iommu *iommu, struct iova *iova)
> {
> unsigned long flags;
> int next, iommu_id;
> @@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova)
> if (list_size == HIGH_WATER_MARK)
> flush_unmaps();
>
> - iommu_id = dom->iommu->seq_id;
> + iommu_id = iommu->seq_id;
>
> next = deferred_flush[iommu_id].next;
> deferred_flush[iommu_id].domain[next] = dom;
> deferred_flush[iommu_id].iova[next] = iova;
> + deferred_flush[iommu_id].iommu = iommu;
> deferred_flush[iommu_id].next++;
>
> if (!timer_on) {
> @@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
> struct dmar_domain *domain;
> unsigned long start_addr;
> struct iova *iova;
> + struct intel_iommu *iommu;
>
> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> return;
> domain = find_domain(pdev);
> BUG_ON(!domain);
>
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return;
> +
> iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
> if (!iova)
> return;
> @@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
> /* free page tables */
> dma_pte_free_pagetable(domain, start_addr, start_addr + size);
> if (intel_iommu_strict) {
> - if (iommu_flush_iotlb_psi(domain->iommu,
> + if (iommu_flush_iotlb_psi(iommu,
> domain->id, start_addr, size >> VTD_PAGE_SHIFT, 0))
> - iommu_flush_write_buffer(domain->iommu);
> + iommu_flush_write_buffer(iommu);
> /* free iova */
> __free_iova(&domain->iovad, iova);
> } else {
> - add_unmap(domain, iova);
> + add_unmap(domain, iommu, iova);
> /*
> * queue up the release of the unmap to save the 1/6th of the
> * cpu used up by the iotlb flush operation...
> @@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
> size_t size = 0;
> void *addr;
> struct scatterlist *sg;
> + struct intel_iommu *iommu;
>
> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> return;
>
> domain = find_domain(pdev);
>
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return;
> +
> iova = find_iova(&domain->iovad, IOVA_PFN(sglist[0].dma_address));
> if (!iova)
> return;
> @@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist,
> /* free page tables */
> dma_pte_free_pagetable(domain, start_addr, start_addr + size);
>
> - if (iommu_flush_iotlb_psi(domain->iommu, domain->id, start_addr,
> + if (iommu_flush_iotlb_psi(iommu, domain->id, start_addr,
> size >> VTD_PAGE_SHIFT, 0))
> - iommu_flush_write_buffer(domain->iommu);
> + iommu_flush_write_buffer(iommu);
>
> /* free iova */
> __free_iova(&domain->iovad, iova);
> @@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
> int ret;
> struct scatterlist *sg;
> unsigned long start_addr;
> + struct intel_iommu *iommu;
>
> BUG_ON(dir == DMA_NONE);
> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> @@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
> if (!domain)
> return 0;
>
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return 0;
> +
> for_each_sg(sglist, sg, nelems, i) {
> addr = SG_ENT_VIRT_ADDRESS(sg);
> addr = (void *)virt_to_phys(addr);
> @@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
> * mappings..
> */
> if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
> - !cap_zlr(domain->iommu->cap))
> + !cap_zlr(iommu->cap))
> prot |= DMA_PTE_READ;
> if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
> prot |= DMA_PTE_WRITE;
> @@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
> }
>
> /* it's a non-present to present mapping */
> - if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
> + if (iommu_flush_iotlb_psi(iommu, domain->id,
> start_addr, offset >> VTD_PAGE_SHIFT, 1))
> - iommu_flush_write_buffer(domain->iommu);
> + iommu_flush_write_buffer(iommu);
> return nelems;
> }
>
> @@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
> return 0;
> }
>
> -void intel_iommu_domain_exit(struct dmar_domain *domain)
> +/* domain id for virtual machine, it won't be set in context */
> +static unsigned long vm_domid;
> +
> +static int vm_domain_min_agaw(struct dmar_domain *domain)
> +{
> + struct dmar_drhd_unit *drhd;
> + struct intel_iommu *iommu;
> + int min_agaw = domain->agaw;
> +
> + for_each_drhd_unit(drhd) {
> + if (drhd->ignored)
> + continue;
> + iommu = drhd->iommu;
> +
> + if (test_bit(iommu->seq_id, &domain->iommu_bmp))
> + if (min_agaw > iommu->agaw)
> + min_agaw = iommu->agaw;
> + }
> +
> + return min_agaw;
> +}
> +
> +static int vm_domain_add_dev_info(struct dmar_domain *domain,
> + struct pci_dev *pdev)
> +{
> + struct device_domain_info *info;
> + unsigned long flags;
> +
> + info = alloc_devinfo_mem();
> + if (!info)
> + return -ENOMEM;
> +
> + info->bus = pdev->bus->number;
> + info->devfn = pdev->devfn;
> + info->dev = pdev;
> + info->domain = domain;
> +
> + spin_lock_irqsave(&device_domain_lock, flags);
> + list_add(&info->link, &domain->devices);
> + list_add(&info->global, &device_domain_list);
> + pdev->dev.archdata.iommu = info;
> + spin_unlock_irqrestore(&device_domain_lock, flags);
> +
> + return 0;
> +}
> +
> +static void vm_domain_remove_one_dev_info(struct dmar_domain *domain,
> + struct pci_dev *pdev)
> +{
> + struct device_domain_info *info;
> + struct intel_iommu *iommu;
> + unsigned long flags;
> + int found = 0;
> +
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> +
> + spin_lock_irqsave(&device_domain_lock, flags);
> + while (!list_empty(&domain->devices)) {
> + info = list_entry(domain->devices.next,
> + struct device_domain_info, link);
> + if (info->bus == pdev->bus->number &&
> + info->devfn == pdev->devfn) {
> + list_del(&info->link);
> + list_del(&info->global);
> + if (info->dev)
> + info->dev->dev.archdata.iommu = NULL;
> + spin_unlock_irqrestore(&device_domain_lock, flags);
> +
> + detach_domain_for_dev(info->domain,
> + info->bus, info->devfn);
> + free_devinfo_mem(info);
> +
> + spin_lock_irqsave(&device_domain_lock, flags);
> +
> + if (found)
> + break;
> + else
> + continue;
> + }
> +
> + /* if there is no other devices under the same iommu
> + * owned by this domain, clear this iommu in iommu_bmp
> + */
> + if (device_find_matched_iommu(info->bus, info->devfn) == iommu)
> + found = 1;
> + }
> +
> + if (found == 0) {
> + spin_lock_irqsave(&iommu->lock, flags);
> + clear_bit(iommu->seq_id, &domain->iommu_bmp);
> + domain->iommu_count--;
> + spin_unlock_irqrestore(&iommu->lock, flags);
> + }
> +
> + spin_unlock_irqrestore(&device_domain_lock, flags);
> +}
> +
> +static void vm_domain_remove_all_dev_info(struct dmar_domain *domain)
> +{
> + struct device_domain_info *info;
> + struct intel_iommu *iommu;
> + unsigned long flags;
> +
> + spin_lock_irqsave(&device_domain_lock, flags);
> + while (!list_empty(&domain->devices)) {
> + info = list_entry(domain->devices.next,
> + struct device_domain_info, link);
> + list_del(&info->link);
> + list_del(&info->global);
> + if (info->dev)
> + info->dev->dev.archdata.iommu = NULL;
> +
> + spin_unlock_irqrestore(&device_domain_lock, flags);
> + detach_domain_for_dev(info->domain,
> + info->bus, info->devfn);
> +
> + /* clear this iommu in iommu_bmp */
> + iommu = device_find_matched_iommu(info->bus, info->devfn);
> + spin_lock_irqsave(&iommu->lock, flags);
> + if (test_and_clear_bit(iommu->seq_id,
> + &domain->iommu_bmp))
> + domain->iommu_count--;
> + spin_unlock_irqrestore(&iommu->lock, flags);
> +
> + free_devinfo_mem(info);
> + spin_lock_irqsave(&device_domain_lock, flags);
> + }
> + spin_unlock_irqrestore(&device_domain_lock, flags);
> +}
> +
> +static int vm_domain_context_mapping_one(struct dmar_domain *domain,
> + struct intel_iommu *iommu, u8 bus, u8 devfn)
> +{
> + struct context_entry *context;
> + unsigned long flags;
> + struct dma_pte *pgd;
> + unsigned long num;
> + unsigned long ndomains;
> + int id;
> + int agaw;
> + int found = 0;
> +
> + pr_debug("Set context mapping for %02x:%02x.%d\n",
> + bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
> + BUG_ON(!domain->pgd);
> + context = device_to_context_entry(iommu, bus, devfn);
> + if (!context)
> + return -ENOMEM;
> + spin_lock_irqsave(&iommu->lock, flags);
> + if (context_present(*context)) {
> + spin_unlock_irqrestore(&iommu->lock, flags);
> + return 0;
> + }
> +
> + id = domain->id;
> +
> + /* find an available domain id for this device in iommu */
> + ndomains = cap_ndoms(iommu->cap);
> + num = find_first_bit(iommu->domain_ids, ndomains);
> + for (; num < ndomains; ) {
> + if (iommu->domains[num] == domain) {
> + id = num;
> + found = 1;
> + break;
> + }
> + num = find_next_bit(iommu->domain_ids,
> + cap_ndoms(iommu->cap), num+1);
> + }
> +
> + if (found == 0) {
> + num = find_first_zero_bit(iommu->domain_ids, ndomains);
> + if (num >= ndomains) {
> + spin_unlock_irqrestore(&iommu->lock, flags);
> + printk(KERN_ERR "IOMMU: no free domain ids\n");
> + return -EFAULT;
> + }
> +
> + set_bit(num, iommu->domain_ids);
> + iommu->domains[num] = domain;
> + id = num;
> + }
> +
> + pgd = domain->pgd;
> +
> + /* Skip top levels of page tables for
> + * iommu which has less agaw than default.
> + */
> + for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
> + pgd = phys_to_virt(dma_pte_addr(*pgd));
> + if (!dma_pte_present(*pgd)) {
> + spin_unlock_irqrestore(&iommu->lock, flags);
> + return -ENOMEM;
> + }
> + }
> +
> + context_set_domain_id(*context, id);
> + context_set_address_width(*context, iommu->agaw);
> + context_set_address_root(*context, virt_to_phys(pgd));
> + context_set_translation_type(*context, CONTEXT_TT_MULTI_LEVEL);
> + context_set_fault_enable(*context);
> + context_set_present(*context);
> + __iommu_flush_cache(iommu, context, sizeof(*context));
> +
> + /* it's a non-present to present mapping */
> + if (iommu->flush.flush_context(iommu, id,
> + (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
> + DMA_CCMD_DEVICE_INVL, 1))
> + iommu_flush_write_buffer(iommu);
> + else
> + iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_DSI_FLUSH, 0);
> +
> + if (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp))
> + domain->iommu_count++;
> + spin_unlock_irqrestore(&iommu->lock, flags);
> + return 0;
> +}
> +
> +static int
> +vm_domain_context_mapping(struct dmar_domain *domain, struct pci_dev *pdev)
> +{
> + int ret;
> + struct pci_dev *tmp, *parent;
> + struct intel_iommu *iommu;
> +
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return -ENODEV;
> +
> + ret = vm_domain_context_mapping_one(domain, iommu,
> + pdev->bus->number, pdev->devfn);
> + if (ret)
> + return ret;
> +
> + /* dependent device mapping */
> + tmp = pci_find_upstream_pcie_bridge(pdev);
> + if (!tmp)
> + return 0;
> + /* Secondary interface's bus number and devfn 0 */
> + parent = pdev->bus->self;
> + while (parent != tmp) {
> + ret = vm_domain_context_mapping_one(domain, iommu,
> + parent->bus->number, parent->devfn);
> + if (ret)
> + return ret;
> + parent = parent->bus->self;
> + }
> + if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
> + return vm_domain_context_mapping_one(domain, iommu,
> + tmp->subordinate->number, 0);
> + else /* this is a legacy PCI bridge */
> + return vm_domain_context_mapping_one(domain, iommu,
> + tmp->bus->number, tmp->devfn);
> +}
> +
> +
> +static int vm_domain_init(struct dmar_domain *domain, int guest_width)
> +{
> + int adjust_width;
> +
> + init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
> + spin_lock_init(&domain->mapping_lock);
> +
> + domain_reserve_special_ranges(domain);
> +
> + /* calculate AGAW */
> + domain->gaw = guest_width;
> + adjust_width = guestwidth_to_adjustwidth(guest_width);
> + domain->agaw = width_to_agaw(adjust_width);
> +
> + INIT_LIST_HEAD(&domain->devices);
> +
> + /* always allocate the top pgd */
> + domain->pgd = (struct dma_pte *)alloc_pgtable_page();
> + if (!domain->pgd)
> + return -ENOMEM;
> + domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
> + return 0;
> +}
> +
> +static void iommu_free_vm_domain(struct dmar_domain *domain)
> +{
> + unsigned long flags;
> + struct dmar_drhd_unit *drhd;
> + struct intel_iommu *iommu;
> + unsigned long i;
> + unsigned long ndomains;
> +
> + for_each_drhd_unit(drhd) {
> + if (drhd->ignored)
> + continue;
> + iommu = drhd->iommu;
> +
> + ndomains = cap_ndoms(iommu->cap);
> + i = find_first_bit(iommu->domain_ids, ndomains);
> + for (; i < ndomains; ) {
> + if (iommu->domains[i] == domain) {
> + spin_lock_irqsave(&iommu->lock, flags);
> + clear_bit(i, iommu->domain_ids);
> + iommu->domains[i] = NULL;
> + spin_unlock_irqrestore(&iommu->lock, flags);
> + break;
> + }
> + i = find_next_bit(iommu->domain_ids, ndomains, i+1);
> + }
> + }
> +}
> +
> +static void vm_domain_exit(struct dmar_domain *domain)
> {
> u64 end;
>
> @@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
> if (!domain)
> return;
>
> + vm_domain_remove_all_dev_info(domain);
> + /* destroy iovas */
> + put_iova_domain(&domain->iovad);
> end = DOMAIN_MAX_ADDR(domain->gaw);
> - end = end & (~VTD_PAGE_MASK);
> + end &= VTD_PAGE_MASK;
>
> /* clear ptes */
> dma_pte_clear_range(domain, 0, end);
> @@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct dmar_domain *domain)
> /* free page tables */
> dma_pte_free_pagetable(domain, 0, end);
>
> - iommu_free_domain(domain);
> + iommu_free_vm_domain(domain);
> free_domain_mem(domain);
> }
> -EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
>
> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
> +static struct dmar_domain *iommu_alloc_vm_domain(void)
> {
> - struct dmar_drhd_unit *drhd;
> struct dmar_domain *domain;
> - struct intel_iommu *iommu;
>
> - drhd = dmar_find_matched_drhd_unit(pdev);
> - if (!drhd) {
> - printk(KERN_ERR "intel_iommu_domain_alloc: drhd == NULL\n");
> + domain = alloc_domain_mem();
> + if (!domain)
> return NULL;
> - }
>
> - iommu = drhd->iommu;
> - if (!iommu) {
> - printk(KERN_ERR
> - "intel_iommu_domain_alloc: iommu == NULL\n");
> - return NULL;
> - }
> - domain = iommu_alloc_domain(iommu);
> + domain->id = vm_domid++;
> + domain->iommu_count = 0;
> + domain->max_addr = 0;
> + memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
> + domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE;
> +
> + return domain;
> +}
> +
> +struct dmar_domain *intel_iommu_alloc_domain(void)
> +{
> + struct dmar_domain *domain;
> +
> + domain = iommu_alloc_vm_domain();
> if (!domain) {
> printk(KERN_ERR
> "intel_iommu_domain_alloc: domain == NULL\n");
> return NULL;
> }
> - if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
> + if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
> printk(KERN_ERR
> "intel_iommu_domain_alloc: domain_init() failed\n");
> - intel_iommu_domain_exit(domain);
> + vm_domain_exit(domain);
> return NULL;
> }
> +
> return domain;
> }
> -EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
> +EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
> +
> +void intel_iommu_free_domain(struct dmar_domain *domain)
> +{
> + vm_domain_exit(domain);
> +}
> +EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
>
> -int intel_iommu_context_mapping(
> - struct dmar_domain *domain, struct pci_dev *pdev)
> +int intel_iommu_assign_device(struct dmar_domain *domain,
> + struct pci_dev *pdev)
> {
> - int rc;
> - rc = domain_context_mapping(domain, pdev);
> - return rc;
> + struct intel_iommu *iommu;
> + int addr_width;
> + u64 end;
> + int ret;
> +
> + /* normally pdev is not mapped */
> + if (unlikely(domain_context_mapped(pdev))) {
> + struct dmar_domain *old_domain;
> +
> + old_domain = find_domain(pdev);
> + if (old_domain) {
> + if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE)
> + vm_domain_remove_one_dev_info(old_domain, pdev);
> + else
> + domain_remove_dev_info(old_domain);
> + }
> + }
> +
> + iommu = device_find_matched_iommu(pdev->bus->number, pdev->devfn);
> + if (!iommu)
> + return -ENODEV;
> +
> + /* check if this iommu agaw is sufficient for max mapped address */
> + addr_width = agaw_to_width(iommu->agaw);
> + end = DOMAIN_MAX_ADDR(addr_width);
> + end = end & VTD_PAGE_MASK;
> + if (end < domain->max_addr) {
> + printk(KERN_ERR "%s: iommu agaw (%d) is not "
> + "sufficient for the mapped address (%llx)\n",
> + __func__, iommu->agaw, domain->max_addr);
> + return -EFAULT;
> + }
> +
> + ret = vm_domain_context_mapping(domain, pdev);
> + if (ret)
> + return ret;
> +
> + ret = vm_domain_add_dev_info(domain, pdev);
> + return ret;
> }
> -EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
> +EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
>
> -int intel_iommu_page_mapping(
> - struct dmar_domain *domain, dma_addr_t iova,
> - u64 hpa, size_t size, int prot)
> +
> +void intel_iommu_deassign_device(struct dmar_domain *domain,
> + struct pci_dev *pdev)
> {
> - int rc;
> - rc = domain_page_mapping(domain, iova, hpa, size, prot);
> - return rc;
> + vm_domain_remove_one_dev_info(domain, pdev);
> }
> -EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
> +EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
>
> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn)
> +int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
> + u64 hpa, size_t size, int prot)
> {
> - detach_domain_for_dev(domain, bus, devfn);
> + u64 max_addr;
> + int addr_width;
> + int ret;
> +
> + max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
> + if (domain->max_addr < max_addr) {
> + int min_agaw;
> + u64 end;
> +
> + /* check if minimum agaw is sufficient for mapped address */
> + min_agaw = vm_domain_min_agaw(domain);
> + addr_width = agaw_to_width(min_agaw);
> + end = DOMAIN_MAX_ADDR(addr_width);
> + end = end & VTD_PAGE_MASK;
> + if (end < max_addr) {
> + printk(KERN_ERR "%s: iommu agaw (%d) is not "
> + "sufficient for the mapped address (%llx)\n",
> + __func__, min_agaw, max_addr);
> + return -EFAULT;
> + }
> + domain->max_addr = max_addr;
> + }
> +
> + ret = domain_page_mapping(domain, iova, hpa, size, prot);
> + return ret;
> }
> -EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
> +EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
>
> -struct dmar_domain *
> -intel_iommu_find_domain(struct pci_dev *pdev)
> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
> + dma_addr_t iova, size_t size)
> {
> - return find_domain(pdev);
> + dma_addr_t base;
> +
> + /* The address might not be aligned */
> + base = iova & PAGE_MASK;
> + size = PAGE_ALIGN(size);
> + dma_pte_clear_range(domain, base, base + size);
> }
> -EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
> +EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
>
> int intel_iommu_found(void)
> {
> diff --git a/include/linux/dma_remapping.h b/include/linux/dma_remapping.h
> index 952df39..3568ae6 100644
> --- a/include/linux/dma_remapping.h
> +++ b/include/linux/dma_remapping.h
> @@ -111,11 +111,21 @@ struct dma_pte {
> (p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
> #define dma_pte_present(p) (((p).val & 3) != 0)
>
> +/* domain flags, one domain owns one device by default */
> +
> +/* devices under the same p2p bridge are owned in one domain */
> +#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES (1 << 0)
> +
> +/* domain represents a virtual machine, more than one devices
> + * across iommus may be owned in one domain, e.g. kvm guest.
> + */
> +#define DOMAIN_FLAG_VIRTUAL_MACHINE (1 << 1)
> +
> struct intel_iommu;
>
> struct dmar_domain {
> int id; /* domain id */
> - struct intel_iommu *iommu; /* back pointer to owning iommu */
> + unsigned long iommu_bmp; /* bitmap of iommus this domain uses*/
>
> struct list_head devices; /* all devices' list */
> struct iova_domain iovad; /* iova's that belong to this domain */
> @@ -123,12 +133,13 @@ struct dmar_domain {
> struct dma_pte *pgd; /* virtual address */
> spinlock_t mapping_lock; /* page table lock */
> int gaw; /* max guest address width */
> + int agaw; /* adjusted guest address width */
>
> - /* adjusted guest address width, 0 is level 2 30-bit */
> - int agaw;
> + int flags; /* domain flag */
>
> -#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
> - int flags;
> + /* following fields are used in virtual machine case */
> + int iommu_count; /* reference count of iommu */
> + u64 max_addr; /* maximum mapped address */
> };
>
> /* PCI domain-device relationship */
> diff --git a/include/linux/intel-iommu.h b/include/linux/intel-iommu.h
> index 3d017cf..c2f37b8 100644
> --- a/include/linux/intel-iommu.h
> +++ b/include/linux/intel-iommu.h
> @@ -219,6 +219,8 @@ do { \
> } \
> } while (0)
>
> +#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
> +
> #define QI_LENGTH 256 /* queue length */
>
> enum {
> @@ -299,6 +301,7 @@ struct intel_iommu {
> struct dmar_domain **domains; /* ptr to domains */
> spinlock_t lock; /* protect context, domain ids */
> struct root_entry *root_entry; /* virtual address */
> + int agaw;
>
> unsigned int irq;
> unsigned char name[7]; /* Device Name */
> @@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu *iommu, u16 did, u64 addr,
>
> extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu *iommu);
>
> -void intel_iommu_domain_exit(struct dmar_domain *domain);
> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
> -int intel_iommu_context_mapping(struct dmar_domain *domain,
> - struct pci_dev *pdev);
> -int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t iova,
> - u64 hpa, size_t size, int prot);
> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8 devfn);
> -struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
> +struct dmar_domain *intel_iommu_alloc_domain(void);
> +void intel_iommu_free_domain(struct dmar_domain *domain);
> +int intel_iommu_assign_device(struct dmar_domain *domain,
> + struct pci_dev *pdev);
> +void intel_iommu_deassign_device(struct dmar_domain *domain,
> + struct pci_dev *pdev);
> +int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t iova,
> + u64 hpa, size_t size, int prot);
> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
> + dma_addr_t iova, size_t size);
> u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
>
> #ifdef CONFIG_DMAR
> --
> 1.5.1
--
| AMD Saxony Limited Liability Company & Co. KG
Operating | Wilschdorfer Landstr. 101, 01109 Dresden, Germany
System | Register Court Dresden: HRA 4896
Research | General Partner authorized to represent:
Center | AMD Saxony LLC (Wilmington, Delaware, US)
| General Manager of AMD Saxony LLC: Dr. Hans-R. Deppe, Thomas McCoy
^ permalink raw reply [flat|nested] 7+ messages in thread
* RE: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
2008-11-28 13:27 ` Joerg Roedel
@ 2008-12-01 6:17 ` Han, Weidong
2008-12-01 11:55 ` 'Joerg Roedel'
0 siblings, 1 reply; 7+ messages in thread
From: Han, Weidong @ 2008-12-01 6:17 UTC (permalink / raw)
To: 'Joerg Roedel'
Cc: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity',
Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
'iommu@lists.linux-foundation.org'
It's developed based on commit 0f7d3ee6 on avi/master, but it still can be applied on latest avi/master (commit 90755652).
Regards,
Weidong
Joerg Roedel wrote:
> Hmm, I tried to apply this patch against avi/master and linus/master
> but get merge conflicts. Where do these patches apply cleanly?
>
> Joerg
>
> On Thu, Nov 27, 2008 at 09:49:04PM +0800, Han, Weidong wrote:
>> In order to support multiple device assignment for KVM, this patch
>> does following main changes:
>> - extend dmar_domain to own multiple devices from different
>> iommus, use a bitmap of iommus to replace iommu pointer in
>> dmar_domain.
>> - implement independent low level functions for kvm, then won't
>> impact native VT-d.
>> - "SAGAW" capability may be different across iommus, that's to
>> say the VT-d page table levels may be different among iommus. This
>> patch uses a defaut agaw, and skip top levels of page tables for
>> iommus which have smaller agaw than default.
>> - rename the APIs for kvm VT-d, make it more readable.
>>
>>
>> Signed-off-by: Weidong Han <weidong.han@intel.com>
>> ---
>> drivers/pci/dmar.c | 15 +
>> drivers/pci/intel-iommu.c | 698
>> ++++++++++++++++++++++++++++++++++------
>> include/linux/dma_remapping.h | 21 +- include/linux/intel-iommu.h
>> | 21 +- 4 files changed, 637 insertions(+), 118 deletions(-)
>>
>> diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
>> index 691b3ad..d6bdced 100644
>> --- a/drivers/pci/dmar.c
>> +++ b/drivers/pci/dmar.c
>> @@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
>> dmar_tbl = NULL; }
>>
>> +extern int width_to_agaw(int width);
>>
>> int alloc_iommu(struct dmar_drhd_unit *drhd)
>> {
>> @@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
>> int map_size; u32 ver;
>> static int iommu_allocated = 0;
>> + unsigned long sagaw;
>> + int agaw;
>>
>> iommu = kzalloc(sizeof(*iommu), GFP_KERNEL); if
>> (!iommu) @@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit
>> *drhd) iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
>> iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
>>
>> + /* set agaw, "SAGAW" may be different across iommus */
>> + sagaw = cap_sagaw(iommu->cap);
>> + for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
>> + agaw >= 0; agaw--)
>> + if (test_bit(agaw, &sagaw))
>> + break;
>> + if (agaw < 0) {
>> + printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n",
>> sagaw); + goto error; + }
>> + iommu->agaw = agaw;
>> +
>> /* the registers might be more than one page */
>> map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
>> cap_max_fault_reg_offset(iommu->cap));
>> diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
>> index 5c8baa4..55b96c4 100644
>> --- a/drivers/pci/intel-iommu.c
>> +++ b/drivers/pci/intel-iommu.c
>> @@ -50,8 +50,6 @@
>> #define IOAPIC_RANGE_END (0xfeefffff)
>> #define IOVA_START_ADDR (0x1000)
>>
>> -#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
>> -
>> #define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
>>
>>
>> @@ -64,6 +62,7 @@ struct deferred_flush_tables {
>> int next;
>> struct iova *iova[HIGH_WATER_MARK];
>> struct dmar_domain *domain[HIGH_WATER_MARK];
>> + struct intel_iommu *iommu;
>> };
>>
>> static struct deferred_flush_tables *deferred_flush;
>> @@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
>> kmem_cache_free(iommu_iova_cache, iova);
>> }
>>
>> +/* in native case, each domain is related to only one iommu */
>> +static struct intel_iommu *domain_get_only_iommu(struct dmar_domain
>> *domain) +{ + struct dmar_drhd_unit *drhd;
>> +
>> + for_each_drhd_unit(drhd) {
>> + if (drhd->ignored)
>> + continue;
>> + if (test_bit(drhd->iommu->seq_id,
>> &domain->iommu_bmp)) + return drhd->iommu;
>> + }
>> +
>> + return NULL;
>> +}
>> +
>> +static void domain_flush_cache(struct dmar_domain *domain,
>> + void *addr, int size) +{
>> + struct intel_iommu *iommu;
>> +
>> + if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
>> + struct dmar_drhd_unit *drhd;
>> +
>> + for_each_drhd_unit(drhd) {
>> + if (drhd->ignored)
>> + continue;
>> + iommu = drhd->iommu;
>> +
>> + if (!test_bit(iommu->seq_id,
>> &domain->iommu_bmp)) + continue;
>> +
>> + if (!ecap_coherent(iommu->ecap))
>> + clflush_cache_range(addr, size); +
>> } + }
>> + else {
>> + iommu = domain_get_only_iommu(domain);
>> + if (iommu && !ecap_coherent(iommu->ecap))
>> + clflush_cache_range(addr, size); + }
>> +}
>> +
>> +static struct intel_iommu *device_find_matched_iommu(u8 bus, u8
>> devfn) +{ + struct dmar_drhd_unit *drhd = NULL;
>> + int i;
>> +
>> + for_each_drhd_unit(drhd) {
>> + if (drhd->ignored)
>> + continue;
>> +
>> + for (i = 0; i < drhd->devices_cnt; i++)
>> + if (drhd->devices[i]->bus->number == bus &&
>> + drhd->devices[i]->devfn == devfn)
>> + return drhd->iommu; +
>> + if (drhd->include_all)
>> + return drhd->iommu;
>> + }
>> +
>> + return NULL;
>> +}
>> +
>> /* Gets context entry for a given bus and devfn */
>> static struct context_entry * device_to_context_entry(struct
>> intel_iommu *iommu, u8 bus, u8 devfn)
>> @@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
>>
>> }
>>
>> -static inline int width_to_agaw(int width)
>> +int width_to_agaw(int width)
>> {
>> return (width - 30) / LEVEL_STRIDE;
>> }
>> @@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct
>> dmar_domain *domain, u64
>> addr) flags); return NULL;
>> }
>> - __iommu_flush_cache(domain->iommu, tmp_page,
>> - PAGE_SIZE);
>> + domain_flush_cache(domain, tmp_page,
>> PAGE_SIZE); dma_set_pte_addr(*pte,
>> virt_to_phys(tmp_page));
>> /* * high level table always sets r/w, last level page @@ -356,7
>> +417,7 @@ static struct dma_pte *
>> addr_to_dma_pte(struct dmar_domain *domain,
>> u64 addr) */ dma_set_pte_readable(*pte);
>> dma_set_pte_writable(*pte); -
>> __iommu_flush_cache(domain->iommu, pte,
>> sizeof(*pte)); + domain_flush_cache(domain,
>> pte, sizeof(*pte)); } parent =
>> phys_to_virt(dma_pte_addr(*pte)); level--; @@ -399,7
>> +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain,
>> u64 addr)
>>
>> if (pte) {
>> dma_clear_pte(*pte);
>> - __iommu_flush_cache(domain->iommu, pte,
>> sizeof(*pte)); + domain_flush_cache(domain, pte,
>> sizeof(*pte)); } }
>>
>> @@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct
>> dmar_domain *domain,
>> free_pgtable_page(
>> phys_to_virt(dma_pte_addr(*pte)));
>> dma_clear_pte(*pte);
>> - __iommu_flush_cache(domain->iommu,
>> - pte, sizeof(*pte));
>> + domain_flush_cache(domain, pte,
>> sizeof(*pte)); } tmp
>> += level_size(level); }
>> @@ -948,8 +1008,8 @@ static int iommu_init_domains(struct
>> intel_iommu *iommu) return 0; }
>>
>> -
>> static void domain_exit(struct dmar_domain *domain);
>> +static void vm_domain_exit(struct dmar_domain *domain);
>>
>> void free_dmar_iommu(struct intel_iommu *iommu)
>> {
>> @@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
>> for (; i < cap_ndoms(iommu->cap); ) {
>> domain = iommu->domains[i];
>> clear_bit(i, iommu->domain_ids);
>> - domain_exit(domain);
>> +
>> + if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
>> + if (--domain->iommu_count == 0)
>> + vm_domain_exit(domain); +
>> } + else
>> + domain_exit(domain);
>> +
>> i = find_next_bit(iommu->domain_ids,
>> cap_ndoms(iommu->cap), i+1); }
>> @@ -1006,8 +1073,11 @@ static struct dmar_domain *
>> iommu_alloc_domain(struct intel_iommu *iommu)
>>
>> set_bit(num, iommu->domain_ids);
>> domain->id = num;
>> - domain->iommu = iommu;
>> + memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
>> + set_bit(iommu->seq_id, &domain->iommu_bmp);
>> iommu->domains[num] = domain;
>> + domain->iommu_count = 1;
>> + domain->flags = 0;
>> spin_unlock_irqrestore(&iommu->lock, flags);
>>
>> return domain;
>> @@ -1016,10 +1086,12 @@ static struct dmar_domain *
>> iommu_alloc_domain(struct intel_iommu *iommu) static void
>> iommu_free_domain(struct dmar_domain *domain) { unsigned
>> long flags; + struct intel_iommu *iommu;
>>
>> - spin_lock_irqsave(&domain->iommu->lock, flags);
>> - clear_bit(domain->id, domain->iommu->domain_ids);
>> - spin_unlock_irqrestore(&domain->iommu->lock, flags);
>> + iommu = domain_get_only_iommu(domain);
>> + spin_lock_irqsave(&iommu->lock, flags);
>> + clear_bit(domain->id, iommu->domain_ids);
>> + spin_unlock_irqrestore(&iommu->lock, flags); }
>>
>> static struct iova_domain reserved_iova_list;
>> @@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain
>> *domain, int guest_width)
>> domain_reserve_special_ranges(domain);
>>
>> /* calculate AGAW */
>> - iommu = domain->iommu;
>> + iommu = domain_get_only_iommu(domain);
>> if (guest_width > cap_mgaw(iommu->cap))
>> guest_width = cap_mgaw(iommu->cap);
>> domain->gaw = guest_width;
>> @@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain
>> *domain, int guest_width) sagaw = cap_sagaw(iommu->cap);
>> if (!test_bit(agaw, &sagaw)) {
>> /* hardware doesn't support it, choose a bigger one
>> */ - pr_debug("IOMMU: hardware doesn't support agaw
>> %d\n", agaw); + pr_debug("IOMMU: hardware doesn't
>> support agaw %d\n", + agaw); agaw =
>> find_next_bit(&sagaw, 5, agaw); if (agaw >= 5)
>> return -ENODEV;
>> }
>> domain->agaw = agaw;
>> +
>> INIT_LIST_HEAD(&domain->devices);
>>
>> /* always allocate the top pgd */
>> domain->pgd = (struct dma_pte *)alloc_pgtable_page();
>> if (!domain->pgd) return -ENOMEM;
>> - __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
>> + domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
>> return 0; }
>>
>> @@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain
>> *domain) }
>>
>> static int domain_context_mapping_one(struct dmar_domain *domain,
>> - u8 bus, u8 devfn)
>> + struct intel_iommu *iommu, u8 bus, u8 devfn) {
>> struct context_entry *context;
>> - struct intel_iommu *iommu = domain->iommu;
>> unsigned long flags;
>>
>> pr_debug("Set context mapping for %02x:%02x.%d\n",
>> @@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain
>> *domain, struct pci_dev *pdev) { int ret;
>> struct pci_dev *tmp, *parent;
>> + struct intel_iommu *iommu;
>> +
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return -ENODEV;
>>
>> - ret = domain_context_mapping_one(domain, pdev->bus->number,
>> - pdev->devfn);
>> + ret = domain_context_mapping_one(domain, iommu,
>> + pdev->bus->number, pdev->devfn);
>> if (ret)
>> return ret;
>>
>> @@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain
>> *domain, struct pci_dev *pdev) /* Secondary interface's bus
>> number and devfn 0 */ parent = pdev->bus->self;
>> while (parent != tmp) {
>> - ret = domain_context_mapping_one(domain,
>> parent->bus->number,
>> - parent->devfn);
>> + ret = domain_context_mapping_one(domain, iommu,
>> + parent->bus->number, parent->devfn);
>> if (ret) return ret;
>> parent = parent->bus->self;
>> }
>> if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
>> - return domain_context_mapping_one(domain,
>> + return domain_context_mapping_one(domain, iommu,
>> tmp->subordinate->number, 0);
>> else /* this is a legacy PCI bridge */
>> - return domain_context_mapping_one(domain,
>> + return domain_context_mapping_one(domain, iommu,
>> tmp->bus->number, tmp->devfn); }
>>
>> -static int domain_context_mapped(struct dmar_domain *domain,
>> - struct pci_dev *pdev)
>> +static int domain_context_mapped(struct pci_dev *pdev) {
>> int ret;
>> struct pci_dev *tmp, *parent;
>> + struct intel_iommu *iommu;
>>
>> - ret = device_context_mapped(domain->iommu,
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return 0;
>> +
>> + ret = device_context_mapped(iommu,
>> pdev->bus->number, pdev->devfn);
>> if (!ret)
>> return ret;
>> @@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct
>> dmar_domain *domain, /* Secondary interface's bus number and
>> devfn 0 */ parent = pdev->bus->self;
>> while (parent != tmp) {
>> - ret = device_context_mapped(domain->iommu,
>> parent->bus->number, + ret =
>> device_context_mapped(iommu,
>> parent->bus->number, parent->devfn); if (!ret)
>> return ret;
>> parent = parent->bus->self;
>> }
>> if (tmp->is_pcie)
>> - return device_context_mapped(domain->iommu,
>> + return device_context_mapped(iommu,
>> tmp->subordinate->number, 0); else
>> - return device_context_mapped(domain->iommu,
>> + return device_context_mapped(iommu,
>> tmp->bus->number, tmp->devfn); }
>>
>> @@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain
>> *domain, dma_addr_t iova, BUG_ON(dma_pte_addr(*pte));
>> dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
>> dma_set_pte_prot(*pte, prot);
>> - __iommu_flush_cache(domain->iommu, pte,
>> sizeof(*pte)); + domain_flush_cache(domain, pte,
>> sizeof(*pte)); start_pfn++; index++;
>> }
>> return 0;
>> }
>>
>> -static void detach_domain_for_dev(struct dmar_domain *domain, u8
>> bus, u8 devfn) +static void detach_domain_for_dev(struct dmar_domain
>> *domain, + u8 bus, u8 devfn) {
>> - clear_context_table(domain->iommu, bus, devfn);
>> - domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
>> - DMA_CCMD_GLOBAL_INVL, 0);
>> - domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
>> - DMA_TLB_GLOBAL_FLUSH, 0);
>> + struct intel_iommu *iommu;
>> +
>> + iommu = device_find_matched_iommu(bus, devfn); + if
>> (!iommu) + return;
>> +
>> + clear_context_table(iommu, bus, devfn);
>> + iommu->flush.flush_context(iommu, 0, 0, 0,
>> + DMA_CCMD_GLOBAL_INVL, 0);
>> + iommu->flush.flush_iotlb(iommu, 0, 0, 0,
>> + DMA_TLB_GLOBAL_FLUSH, 0); }
>>
>> static void domain_remove_dev_info(struct dmar_domain *domain)
>> @@ -1336,7 +1425,6 @@ static struct dmar_domain
>> *get_domain_for_dev(struct pci_dev *pdev, int gaw) { struct
>> dmar_domain *domain, *found = NULL; struct intel_iommu
>> *iommu; - struct dmar_drhd_unit *drhd;
>> struct device_domain_info *info, *tmp;
>> struct pci_dev *dev_tmp;
>> unsigned long flags;
>> @@ -1371,13 +1459,9 @@ static struct dmar_domain
>> *get_domain_for_dev(struct pci_dev *pdev, int gaw) }
>>
>> /* Allocate new domain for the device */
>> - drhd = dmar_find_matched_drhd_unit(pdev);
>> - if (!drhd) {
>> - printk(KERN_ERR "IOMMU: can't find DMAR for device
>> %s\n",
>> - pci_name(pdev));
>> - return NULL;
>> - }
>> - iommu = drhd->iommu;
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return NULL;
>>
>> domain = iommu_alloc_domain(iommu);
>> if (!domain)
>> @@ -1400,7 +1484,7 @@ static struct dmar_domain
>> *get_domain_for_dev(struct pci_dev *pdev, int gaw)
>> info->dev = NULL; info->domain = domain;
>> /* This domain is shared by devices under p2p bridge
>> */ - domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
>> + domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
>>
>> /* pcie-to-pci bridge already has a domain, uses it
>> */ found = NULL; @@ -1805,7 +1889,7 @@
>> get_valid_domain_for_dev(struct pci_dev *pdev) }
>>
>> /* make sure context mapping is ok */
>> - if (unlikely(!domain_context_mapped(domain, pdev))) {
>> + if (unlikely(!domain_context_mapped(pdev))) {
>> ret = domain_context_mapping(domain, pdev);
>> if (ret) { printk(KERN_ERR
>> @@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct
>> device *hwdev, phys_addr_t paddr, { struct pci_dev *pdev =
>> to_pci_dev(hwdev); struct dmar_domain *domain;
>> + struct intel_iommu *iommu;
>> phys_addr_t start_paddr;
>> struct iova *iova;
>> int prot = 0;
>> @@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct
>> device *hwdev, phys_addr_t paddr, if
>> (!domain) return 0;
>>
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return 0;
>> +
>> size = aligned_size((u64)paddr, size);
>>
>> iova = __intel_alloc_iova(hwdev, domain, size,
>> pdev->dma_mask); @@ -1849,7 +1938,7 @@ static dma_addr_t
>> __intel_map_single(struct device *hwdev, phys_addr_t paddr,
>> * mappings..
>> */
>> if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
>> - !cap_zlr(domain->iommu->cap))
>> + !cap_zlr(iommu->cap))
>> prot |= DMA_PTE_READ;
>> if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
>> prot |= DMA_PTE_WRITE;
>> @@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct
>> device *hwdev, phys_addr_t paddr, goto error;
>>
>> /* it's a non-present to present mapping */
>> - ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
>> + ret = iommu_flush_iotlb_psi(iommu, domain->id,
>> start_paddr, size >> VTD_PAGE_SHIFT, 1);
>> if (ret) - iommu_flush_write_buffer(domain->iommu);
>> + iommu_flush_write_buffer(iommu);
>>
>> return start_paddr + ((u64)paddr & (~PAGE_MASK));
>>
>> @@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
>> /* just flush them all */
>> for (i = 0; i < g_num_of_iommus; i++) {
>> if (deferred_flush[i].next) {
>> - struct intel_iommu *iommu =
>> - deferred_flush[i].domain[0]->iommu;
>> + struct intel_iommu *iommu =
>> deferred_flush[i].iommu;
>>
>> iommu->flush.flush_iotlb(iommu, 0, 0, 0,
>>
>> DMA_TLB_GLOBAL_FLUSH, 0); @@ -1921,7 +2009,8 @@ static void
>> flush_unmaps_timeout(unsigned long data)
>> spin_unlock_irqrestore(&async_umap_flush_lock, flags); }
>>
>> -static void add_unmap(struct dmar_domain *dom, struct iova *iova)
>> +static void add_unmap(struct dmar_domain *dom,
>> + struct intel_iommu *iommu, struct iova *iova)
>> { unsigned long flags;
>> int next, iommu_id;
>> @@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain
>> *dom, struct iova *iova) if (list_size == HIGH_WATER_MARK)
>> flush_unmaps();
>>
>> - iommu_id = dom->iommu->seq_id;
>> + iommu_id = iommu->seq_id;
>>
>> next = deferred_flush[iommu_id].next;
>> deferred_flush[iommu_id].domain[next] = dom;
>> deferred_flush[iommu_id].iova[next] = iova;
>> + deferred_flush[iommu_id].iommu = iommu;
>> deferred_flush[iommu_id].next++;
>>
>> if (!timer_on) {
>> @@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev,
>> dma_addr_t dev_addr, size_t size, struct dmar_domain *domain;
>> unsigned long start_addr;
>> struct iova *iova;
>> + struct intel_iommu *iommu;
>>
>> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
>> return; domain = find_domain(pdev);
>> BUG_ON(!domain);
>>
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return;
>> +
>> iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
>> if (!iova) return;
>> @@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev,
>> dma_addr_t dev_addr, size_t size, /* free page tables */
>> dma_pte_free_pagetable(domain, start_addr, start_addr +
>> size); if (intel_iommu_strict) {
>> - if (iommu_flush_iotlb_psi(domain->iommu,
>> + if (iommu_flush_iotlb_psi(iommu,
>> domain->id, start_addr, size >>
>> VTD_PAGE_SHIFT, 0)) -
>> iommu_flush_write_buffer(domain->iommu); +
>> iommu_flush_write_buffer(iommu); /* free iova */
>> __free_iova(&domain->iovad, iova);
>> } else {
>> - add_unmap(domain, iova);
>> + add_unmap(domain, iommu, iova);
>> /*
>> * queue up the release of the unmap to save the
>> 1/6th of the
>> * cpu used up by the iotlb flush operation...
>> @@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev,
>> struct scatterlist *sglist, size_t size = 0; void
>> *addr; struct scatterlist *sg;
>> + struct intel_iommu *iommu;
>>
>> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
>> return;
>>
>> domain = find_domain(pdev);
>>
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return;
>> +
>> iova = find_iova(&domain->iovad,
>> IOVA_PFN(sglist[0].dma_address)); if (!iova)
>> return; @@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device
>> *hwdev, struct scatterlist *sglist, /* free page tables */
>> dma_pte_free_pagetable(domain, start_addr, start_addr +
>> size);
>>
>> - if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
>> start_addr, + if (iommu_flush_iotlb_psi(iommu, domain->id,
>> start_addr, size >> VTD_PAGE_SHIFT, 0))
>> - iommu_flush_write_buffer(domain->iommu);
>> + iommu_flush_write_buffer(iommu);
>>
>> /* free iova */
>> __free_iova(&domain->iovad, iova);
>> @@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct
>> scatterlist *sglist, int nelems, int ret; struct
>> scatterlist *sg; unsigned long start_addr;
>> + struct intel_iommu *iommu;
>>
>> BUG_ON(dir == DMA_NONE);
>> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
>> @@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct
>> scatterlist *sglist, int nelems, if
>> (!domain) return 0;
>>
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return 0;
>> +
>> for_each_sg(sglist, sg, nelems, i) {
>> addr = SG_ENT_VIRT_ADDRESS(sg);
>> addr = (void *)virt_to_phys(addr);
>> @@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct
>> scatterlist *sglist, int nelems,
>> * mappings..
>> */
>> if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
>> - !cap_zlr(domain->iommu->cap))
>> + !cap_zlr(iommu->cap))
>> prot |= DMA_PTE_READ;
>> if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
>> prot |= DMA_PTE_WRITE;
>> @@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct
>> scatterlist *sglist, int nelems, }
>>
>> /* it's a non-present to present mapping */
>> - if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
>> + if (iommu_flush_iotlb_psi(iommu, domain->id,
>> start_addr, offset >> VTD_PAGE_SHIFT, 1))
>> - iommu_flush_write_buffer(domain->iommu);
>> + iommu_flush_write_buffer(iommu);
>> return nelems;
>> }
>>
>> @@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
>> return 0; }
>>
>> -void intel_iommu_domain_exit(struct dmar_domain *domain)
>> +/* domain id for virtual machine, it won't be set in context */
>> +static unsigned long vm_domid;
>> +
>> +static int vm_domain_min_agaw(struct dmar_domain *domain) +{
>> + struct dmar_drhd_unit *drhd;
>> + struct intel_iommu *iommu;
>> + int min_agaw = domain->agaw;
>> +
>> + for_each_drhd_unit(drhd) {
>> + if (drhd->ignored)
>> + continue;
>> + iommu = drhd->iommu;
>> +
>> + if (test_bit(iommu->seq_id, &domain->iommu_bmp))
>> + if (min_agaw > iommu->agaw)
>> + min_agaw = iommu->agaw; + }
>> +
>> + return min_agaw;
>> +}
>> +
>> +static int vm_domain_add_dev_info(struct dmar_domain *domain,
>> + struct pci_dev *pdev) +{
>> + struct device_domain_info *info;
>> + unsigned long flags;
>> +
>> + info = alloc_devinfo_mem();
>> + if (!info)
>> + return -ENOMEM;
>> +
>> + info->bus = pdev->bus->number;
>> + info->devfn = pdev->devfn;
>> + info->dev = pdev;
>> + info->domain = domain;
>> +
>> + spin_lock_irqsave(&device_domain_lock, flags);
>> + list_add(&info->link, &domain->devices);
>> + list_add(&info->global, &device_domain_list);
>> + pdev->dev.archdata.iommu = info;
>> + spin_unlock_irqrestore(&device_domain_lock, flags); +
>> + return 0;
>> +}
>> +
>> +static void vm_domain_remove_one_dev_info(struct dmar_domain
>> *domain, + struct pci_dev
>> *pdev) +{ + struct device_domain_info *info;
>> + struct intel_iommu *iommu;
>> + unsigned long flags;
>> + int found = 0;
>> +
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + + spin_lock_irqsave(&device_domain_lock,
>> flags); + while (!list_empty(&domain->devices)) {
>> + info = list_entry(domain->devices.next,
>> + struct device_domain_info, link);
>> + if (info->bus == pdev->bus->number &&
>> + info->devfn == pdev->devfn) {
>> + list_del(&info->link);
>> + list_del(&info->global);
>> + if (info->dev)
>> + info->dev->dev.archdata.iommu = NULL;
>> + spin_unlock_irqrestore(&device_domain_lock,
>> flags); + + detach_domain_for_dev(info->domain,
>> + info->bus,
>> info->devfn); + free_devinfo_mem(info);
>> +
>> + spin_lock_irqsave(&device_domain_lock,
>> flags); + + if (found)
>> + break;
>> + else
>> + continue;
>> + }
>> +
>> + /* if there is no other devices under the same iommu
>> + * owned by this domain, clear this iommu in
>> iommu_bmp + */ + if
>> (device_find_matched_iommu(info->bus, info->devfn) == iommu) +
>> found = 1; + }
>> +
>> + if (found == 0) {
>> + spin_lock_irqsave(&iommu->lock, flags);
>> + clear_bit(iommu->seq_id, &domain->iommu_bmp);
>> + domain->iommu_count--;
>> + spin_unlock_irqrestore(&iommu->lock, flags); +
>> } +
>> + spin_unlock_irqrestore(&device_domain_lock, flags); +}
>> +
>> +static void vm_domain_remove_all_dev_info(struct dmar_domain
>> *domain) +{ + struct device_domain_info *info;
>> + struct intel_iommu *iommu;
>> + unsigned long flags;
>> +
>> + spin_lock_irqsave(&device_domain_lock, flags);
>> + while (!list_empty(&domain->devices)) {
>> + info = list_entry(domain->devices.next,
>> + struct device_domain_info, link);
>> + list_del(&info->link);
>> + list_del(&info->global);
>> + if (info->dev)
>> + info->dev->dev.archdata.iommu = NULL; +
>> + spin_unlock_irqrestore(&device_domain_lock, flags);
>> + detach_domain_for_dev(info->domain,
>> + info->bus, info->devfn); +
>> + /* clear this iommu in iommu_bmp */
>> + iommu = device_find_matched_iommu(info->bus,
>> info->devfn); + spin_lock_irqsave(&iommu->lock, flags);
>> + if (test_and_clear_bit(iommu->seq_id,
>> + &domain->iommu_bmp))
>> + domain->iommu_count--;
>> + spin_unlock_irqrestore(&iommu->lock, flags); +
>> + free_devinfo_mem(info);
>> + spin_lock_irqsave(&device_domain_lock, flags); +
>> } + spin_unlock_irqrestore(&device_domain_lock, flags); +}
>> +
>> +static int vm_domain_context_mapping_one(struct dmar_domain *domain,
>> + struct intel_iommu *iommu, u8 bus, u8 devfn) +{
>> + struct context_entry *context;
>> + unsigned long flags;
>> + struct dma_pte *pgd;
>> + unsigned long num;
>> + unsigned long ndomains;
>> + int id;
>> + int agaw;
>> + int found = 0;
>> +
>> + pr_debug("Set context mapping for %02x:%02x.%d\n",
>> + bus, PCI_SLOT(devfn), PCI_FUNC(devfn)); +
>> BUG_ON(!domain->pgd); + context =
>> device_to_context_entry(iommu, bus, devfn); + if (!context) +
>> return -ENOMEM; + spin_lock_irqsave(&iommu->lock, flags);
>> + if (context_present(*context)) {
>> + spin_unlock_irqrestore(&iommu->lock, flags); +
>> return 0; + }
>> +
>> + id = domain->id;
>> +
>> + /* find an available domain id for this device in iommu */
>> + ndomains = cap_ndoms(iommu->cap);
>> + num = find_first_bit(iommu->domain_ids, ndomains);
>> + for (; num < ndomains; ) {
>> + if (iommu->domains[num] == domain) {
>> + id = num;
>> + found = 1;
>> + break;
>> + }
>> + num = find_next_bit(iommu->domain_ids,
>> + cap_ndoms(iommu->cap), num+1);
>> + }
>> +
>> + if (found == 0) {
>> + num = find_first_zero_bit(iommu->domain_ids,
>> ndomains); + if (num >= ndomains) {
>> + spin_unlock_irqrestore(&iommu->lock, flags);
>> + printk(KERN_ERR "IOMMU: no free domain
>> ids\n"); + return -EFAULT;
>> + }
>> +
>> + set_bit(num, iommu->domain_ids);
>> + iommu->domains[num] = domain;
>> + id = num;
>> + }
>> +
>> + pgd = domain->pgd;
>> +
>> + /* Skip top levels of page tables for
>> + * iommu which has less agaw than default.
>> + */
>> + for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
>> + pgd = phys_to_virt(dma_pte_addr(*pgd));
>> + if (!dma_pte_present(*pgd)) {
>> + spin_unlock_irqrestore(&iommu->lock, flags);
>> + return -ENOMEM;
>> + }
>> + }
>> +
>> + context_set_domain_id(*context, id);
>> + context_set_address_width(*context, iommu->agaw);
>> + context_set_address_root(*context, virt_to_phys(pgd));
>> + context_set_translation_type(*context,
>> CONTEXT_TT_MULTI_LEVEL); + context_set_fault_enable(*context);
>> + context_set_present(*context);
>> + __iommu_flush_cache(iommu, context, sizeof(*context)); +
>> + /* it's a non-present to present mapping */
>> + if (iommu->flush.flush_context(iommu, id,
>> + (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
>> + DMA_CCMD_DEVICE_INVL, 1))
>> + iommu_flush_write_buffer(iommu);
>> + else
>> + iommu->flush.flush_iotlb(iommu, 0, 0, 0,
>> DMA_TLB_DSI_FLUSH, 0); + + if
>> (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp)) +
>> domain->iommu_count++; + spin_unlock_irqrestore(&iommu->lock,
>> flags); + return 0; +}
>> +
>> +static int
>> +vm_domain_context_mapping(struct dmar_domain *domain, struct
>> pci_dev *pdev) +{ + int ret;
>> + struct pci_dev *tmp, *parent;
>> + struct intel_iommu *iommu;
>> +
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return -ENODEV;
>> +
>> + ret = vm_domain_context_mapping_one(domain, iommu,
>> + pdev->bus->number, pdev->devfn);
>> + if (ret)
>> + return ret;
>> +
>> + /* dependent device mapping */
>> + tmp = pci_find_upstream_pcie_bridge(pdev);
>> + if (!tmp)
>> + return 0;
>> + /* Secondary interface's bus number and devfn 0 */
>> + parent = pdev->bus->self;
>> + while (parent != tmp) {
>> + ret = vm_domain_context_mapping_one(domain, iommu,
>> + parent->bus->number, parent->devfn); +
>> if (ret) + return ret;
>> + parent = parent->bus->self;
>> + }
>> + if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
>> + return vm_domain_context_mapping_one(domain, iommu,
>> + tmp->subordinate->number, 0);
>> + else /* this is a legacy PCI bridge */
>> + return vm_domain_context_mapping_one(domain, iommu,
>> + tmp->bus->number, tmp->devfn); +}
>> +
>> +
>> +static int vm_domain_init(struct dmar_domain *domain, int
>> guest_width) +{ + int adjust_width;
>> +
>> + init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
>> + spin_lock_init(&domain->mapping_lock);
>> +
>> + domain_reserve_special_ranges(domain);
>> +
>> + /* calculate AGAW */
>> + domain->gaw = guest_width;
>> + adjust_width = guestwidth_to_adjustwidth(guest_width);
>> + domain->agaw = width_to_agaw(adjust_width); +
>> + INIT_LIST_HEAD(&domain->devices);
>> +
>> + /* always allocate the top pgd */
>> + domain->pgd = (struct dma_pte *)alloc_pgtable_page(); +
>> if (!domain->pgd) + return -ENOMEM;
>> + domain_flush_cache(domain, domain->pgd, PAGE_SIZE); +
>> return 0; +}
>> +
>> +static void iommu_free_vm_domain(struct dmar_domain *domain) +{
>> + unsigned long flags;
>> + struct dmar_drhd_unit *drhd;
>> + struct intel_iommu *iommu;
>> + unsigned long i;
>> + unsigned long ndomains;
>> +
>> + for_each_drhd_unit(drhd) {
>> + if (drhd->ignored)
>> + continue;
>> + iommu = drhd->iommu;
>> +
>> + ndomains = cap_ndoms(iommu->cap);
>> + i = find_first_bit(iommu->domain_ids, ndomains);
>> + for (; i < ndomains; ) {
>> + if (iommu->domains[i] == domain) {
>> + spin_lock_irqsave(&iommu->lock,
>> flags); + clear_bit(i,
>> iommu->domain_ids); +
>> iommu->domains[i] = NULL; +
>> spin_unlock_irqrestore(&iommu->lock, flags); +
>> break; + }
>> + i = find_next_bit(iommu->domain_ids,
>> ndomains, i+1); + } + }
>> +}
>> +
>> +static void vm_domain_exit(struct dmar_domain *domain) {
>> u64 end;
>>
>> @@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct
>> dmar_domain *domain) if (!domain) return;
>>
>> + vm_domain_remove_all_dev_info(domain);
>> + /* destroy iovas */
>> + put_iova_domain(&domain->iovad);
>> end = DOMAIN_MAX_ADDR(domain->gaw);
>> - end = end & (~VTD_PAGE_MASK);
>> + end &= VTD_PAGE_MASK;
>>
>> /* clear ptes */
>> dma_pte_clear_range(domain, 0, end);
>> @@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct
>> dmar_domain *domain) /* free page tables */
>> dma_pte_free_pagetable(domain, 0, end);
>>
>> - iommu_free_domain(domain);
>> + iommu_free_vm_domain(domain);
>> free_domain_mem(domain);
>> }
>> -EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
>>
>> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
>> +static struct dmar_domain *iommu_alloc_vm_domain(void) {
>> - struct dmar_drhd_unit *drhd;
>> struct dmar_domain *domain;
>> - struct intel_iommu *iommu;
>>
>> - drhd = dmar_find_matched_drhd_unit(pdev);
>> - if (!drhd) {
>> - printk(KERN_ERR "intel_iommu_domain_alloc: drhd ==
>> NULL\n"); + domain = alloc_domain_mem();
>> + if (!domain)
>> return NULL;
>> - }
>>
>> - iommu = drhd->iommu;
>> - if (!iommu) {
>> - printk(KERN_ERR
>> - "intel_iommu_domain_alloc: iommu == NULL\n");
>> - return NULL;
>> - }
>> - domain = iommu_alloc_domain(iommu);
>> + domain->id = vm_domid++;
>> + domain->iommu_count = 0;
>> + domain->max_addr = 0;
>> + memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
>> + domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE; +
>> + return domain;
>> +}
>> +
>> +struct dmar_domain *intel_iommu_alloc_domain(void) +{
>> + struct dmar_domain *domain;
>> +
>> + domain = iommu_alloc_vm_domain();
>> if (!domain) {
>> printk(KERN_ERR
>> "intel_iommu_domain_alloc: domain ==
>> NULL\n"); return NULL; }
>> - if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
>> + if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
>> printk(KERN_ERR
>> "intel_iommu_domain_alloc: domain_init()
>> failed\n"); - intel_iommu_domain_exit(domain);
>> + vm_domain_exit(domain);
>> return NULL;
>> }
>> +
>> return domain;
>> }
>> -EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
>> +EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
>> +
>> +void intel_iommu_free_domain(struct dmar_domain *domain) +{
>> + vm_domain_exit(domain);
>> +}
>> +EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
>>
>> -int intel_iommu_context_mapping(
>> - struct dmar_domain *domain, struct pci_dev *pdev)
>> +int intel_iommu_assign_device(struct dmar_domain *domain,
>> + struct pci_dev *pdev) {
>> - int rc;
>> - rc = domain_context_mapping(domain, pdev);
>> - return rc;
>> + struct intel_iommu *iommu;
>> + int addr_width;
>> + u64 end;
>> + int ret;
>> +
>> + /* normally pdev is not mapped */
>> + if (unlikely(domain_context_mapped(pdev))) {
>> + struct dmar_domain *old_domain;
>> +
>> + old_domain = find_domain(pdev);
>> + if (old_domain) {
>> + if (domain->flags &
>> DOMAIN_FLAG_VIRTUAL_MACHINE) +
>> vm_domain_remove_one_dev_info(old_domain, pdev); +
>> else +
>> domain_remove_dev_info(old_domain); + } + } +
>> + iommu = device_find_matched_iommu(pdev->bus->number,
>> pdev->devfn); + if (!iommu) + return -ENODEV;
>> +
>> + /* check if this iommu agaw is sufficient for max mapped
>> address */ + addr_width = agaw_to_width(iommu->agaw);
>> + end = DOMAIN_MAX_ADDR(addr_width);
>> + end = end & VTD_PAGE_MASK;
>> + if (end < domain->max_addr) {
>> + printk(KERN_ERR "%s: iommu agaw (%d) is not "
>> + "sufficient for the mapped address (%llx)\n",
>> + __func__, iommu->agaw, domain->max_addr);
>> + return -EFAULT;
>> + }
>> +
>> + ret = vm_domain_context_mapping(domain, pdev); + if
>> (ret) + return ret;
>> +
>> + ret = vm_domain_add_dev_info(domain, pdev); + return
>> ret; }
>> -EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
>> +EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
>>
>> -int intel_iommu_page_mapping(
>> - struct dmar_domain *domain, dma_addr_t iova,
>> - u64 hpa, size_t size, int prot)
>> +
>> +void intel_iommu_deassign_device(struct dmar_domain *domain,
>> + struct pci_dev *pdev) {
>> - int rc;
>> - rc = domain_page_mapping(domain, iova, hpa, size, prot);
>> - return rc;
>> + vm_domain_remove_one_dev_info(domain, pdev); }
>> -EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
>> +EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
>>
>> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8
>> devfn) +int intel_iommu_map_pages(struct dmar_domain *domain,
>> dma_addr_t iova, + u64 hpa, size_t size, int
>> prot) { - detach_domain_for_dev(domain, bus, devfn);
>> + u64 max_addr;
>> + int addr_width;
>> + int ret;
>> +
>> + max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
>> + if (domain->max_addr < max_addr) {
>> + int min_agaw;
>> + u64 end;
>> +
>> + /* check if minimum agaw is sufficient for mapped
>> address */ + min_agaw = vm_domain_min_agaw(domain);
>> + addr_width = agaw_to_width(min_agaw);
>> + end = DOMAIN_MAX_ADDR(addr_width);
>> + end = end & VTD_PAGE_MASK;
>> + if (end < max_addr) {
>> + printk(KERN_ERR "%s: iommu agaw (%d) is not "
>> + "sufficient for the mapped address
>> (%llx)\n", + __func__, min_agaw,
>> max_addr); + return -EFAULT;
>> + }
>> + domain->max_addr = max_addr;
>> + }
>> +
>> + ret = domain_page_mapping(domain, iova, hpa, size, prot); +
>> return ret; }
>> -EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
>> +EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
>>
>> -struct dmar_domain *
>> -intel_iommu_find_domain(struct pci_dev *pdev)
>> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
>> + dma_addr_t iova, size_t size) {
>> - return find_domain(pdev);
>> + dma_addr_t base;
>> +
>> + /* The address might not be aligned */
>> + base = iova & PAGE_MASK;
>> + size = PAGE_ALIGN(size);
>> + dma_pte_clear_range(domain, base, base + size); }
>> -EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
>> +EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
>>
>> int intel_iommu_found(void)
>> {
>> diff --git a/include/linux/dma_remapping.h
>> b/include/linux/dma_remapping.h
>> index 952df39..3568ae6 100644
>> --- a/include/linux/dma_remapping.h
>> +++ b/include/linux/dma_remapping.h
>> @@ -111,11 +111,21 @@ struct dma_pte {
>> (p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
>> #define dma_pte_present(p) (((p).val & 3) != 0)
>>
>> +/* domain flags, one domain owns one device by default */ +
>> +/* devices under the same p2p bridge are owned in one domain */
>> +#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES (1 << 0) +
>> +/* domain represents a virtual machine, more than one devices
>> + * across iommus may be owned in one domain, e.g. kvm guest. + */
>> +#define DOMAIN_FLAG_VIRTUAL_MACHINE (1 << 1) +
>> struct intel_iommu;
>>
>> struct dmar_domain {
>> int id; /* domain id */
>> - struct intel_iommu *iommu; /* back pointer to owning
>> iommu */ + unsigned long iommu_bmp; /* bitmap of iommus
>> this domain uses*/
>>
>> struct list_head devices; /* all devices' list */
>> struct iova_domain iovad; /* iova's that belong to
>> this domain */ @@ -123,12 +133,13 @@ struct dmar_domain {
>> struct dma_pte *pgd; /* virtual address */
>> spinlock_t mapping_lock; /* page table lock */
>> int gaw; /* max guest address width */
>> + int agaw; /* adjusted guest address
>> width */
>>
>> - /* adjusted guest address width, 0 is level 2 30-bit */
>> - int agaw;
>> + int flags; /* domain flag */
>>
>> -#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
>> - int flags;
>> + /* following fields are used in virtual machine case */
>> + int iommu_count; /* reference count of iommu
>> */ + u64 max_addr; /* maximum mapped address
>> */ };
>>
>> /* PCI domain-device relationship */
>> diff --git a/include/linux/intel-iommu.h
>> b/include/linux/intel-iommu.h
>> index 3d017cf..c2f37b8 100644
>> --- a/include/linux/intel-iommu.h
>> +++ b/include/linux/intel-iommu.h
>> @@ -219,6 +219,8 @@ do {
>> \ }
>> \ } while (0)
>>
>> +#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
>> +
>> #define QI_LENGTH 256 /* queue length */
>>
>> enum {
>> @@ -299,6 +301,7 @@ struct intel_iommu {
>> struct dmar_domain **domains; /* ptr to domains */
>> spinlock_t lock; /* protect context, domain ids */
>> struct root_entry *root_entry; /* virtual address */ +
>> int agaw;
>>
>> unsigned int irq;
>> unsigned char name[7]; /* Device Name */
>> @@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu
>> *iommu, u16 did, u64 addr,
>>
>> extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu
>> *iommu);
>>
>> -void intel_iommu_domain_exit(struct dmar_domain *domain);
>> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
>> -int intel_iommu_context_mapping(struct dmar_domain *domain,
>> - struct pci_dev *pdev);
>> -int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t
>> iova,
>> - u64 hpa, size_t size, int prot);
>> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8
>> devfn);
>> -struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
>> +struct dmar_domain *intel_iommu_alloc_domain(void);
>> +void intel_iommu_free_domain(struct dmar_domain *domain);
>> +int intel_iommu_assign_device(struct dmar_domain *domain,
>> + struct pci_dev *pdev);
>> +void intel_iommu_deassign_device(struct dmar_domain *domain,
>> + struct pci_dev *pdev);
>> +int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t
>> iova, + u64 hpa, size_t size, int prot);
>> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
>> + dma_addr_t iova, size_t size);
>> u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
>>
>> #ifdef CONFIG_DMAR
>> --
>> 1.5.1
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
2008-12-01 6:17 ` Han, Weidong
@ 2008-12-01 11:55 ` 'Joerg Roedel'
2008-12-01 13:22 ` Han, Weidong
0 siblings, 1 reply; 7+ messages in thread
From: 'Joerg Roedel' @ 2008-12-01 11:55 UTC (permalink / raw)
To: Han, Weidong
Cc: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity',
Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
'iommu@lists.linux-foundation.org'
Hmm, I get these errors using git-am:
Applying VT-d: Support multiple device assignment for KVM
.dotest/patch:1344: space before tab in indent.
clflush_cache_range(addr, size);
.dotest/patch:1350: space before tab in indent.
clflush_cache_range(addr, size);
.dotest/patch:1907: trailing whitespace.
.dotest/patch:1946: trailing whitespace.
* owned by this domain, clear this iommu in iommu_bmp
.dotest/patch:2300: trailing whitespace.
error: patch failed: drivers/pci/dmar.c:484
error: drivers/pci/dmar.c: patch does not apply
error: patch failed: drivers/pci/intel-iommu.c:50
error: drivers/pci/intel-iommu.c: patch does not apply
error: patch failed: include/linux/dma_remapping.h:111
error: include/linux/dma_remapping.h: patch does not apply
error: patch failed: include/linux/intel-iommu.h:219
error: include/linux/intel-iommu.h: patch does not apply
Patch failed at 0001.
Joerg
On Mon, Dec 01, 2008 at 02:17:38PM +0800, Han, Weidong wrote:
> It's developed based on commit 0f7d3ee6 on avi/master, but it still can be applied on latest avi/master (commit 90755652).
>
> Regards,
> Weidong
>
> Joerg Roedel wrote:
> > Hmm, I tried to apply this patch against avi/master and linus/master
> > but get merge conflicts. Where do these patches apply cleanly?
> >
> > Joerg
> >
> > On Thu, Nov 27, 2008 at 09:49:04PM +0800, Han, Weidong wrote:
> >> In order to support multiple device assignment for KVM, this patch
> >> does following main changes:
> >> - extend dmar_domain to own multiple devices from different
> >> iommus, use a bitmap of iommus to replace iommu pointer in
> >> dmar_domain.
> >> - implement independent low level functions for kvm, then won't
> >> impact native VT-d.
> >> - "SAGAW" capability may be different across iommus, that's to
> >> say the VT-d page table levels may be different among iommus. This
> >> patch uses a defaut agaw, and skip top levels of page tables for
> >> iommus which have smaller agaw than default.
> >> - rename the APIs for kvm VT-d, make it more readable.
> >>
> >>
> >> Signed-off-by: Weidong Han <weidong.han@intel.com>
> >> ---
> >> drivers/pci/dmar.c | 15 +
> >> drivers/pci/intel-iommu.c | 698
> >> ++++++++++++++++++++++++++++++++++------
> >> include/linux/dma_remapping.h | 21 +- include/linux/intel-iommu.h
> >> | 21 +- 4 files changed, 637 insertions(+), 118 deletions(-)
> >>
> >> diff --git a/drivers/pci/dmar.c b/drivers/pci/dmar.c
> >> index 691b3ad..d6bdced 100644
> >> --- a/drivers/pci/dmar.c
> >> +++ b/drivers/pci/dmar.c
> >> @@ -484,6 +484,7 @@ void __init detect_intel_iommu(void)
> >> dmar_tbl = NULL; }
> >>
> >> +extern int width_to_agaw(int width);
> >>
> >> int alloc_iommu(struct dmar_drhd_unit *drhd)
> >> {
> >> @@ -491,6 +492,8 @@ int alloc_iommu(struct dmar_drhd_unit *drhd)
> >> int map_size; u32 ver;
> >> static int iommu_allocated = 0;
> >> + unsigned long sagaw;
> >> + int agaw;
> >>
> >> iommu = kzalloc(sizeof(*iommu), GFP_KERNEL); if
> >> (!iommu) @@ -506,6 +509,18 @@ int alloc_iommu(struct dmar_drhd_unit
> >> *drhd) iommu->cap = dmar_readq(iommu->reg + DMAR_CAP_REG);
> >> iommu->ecap = dmar_readq(iommu->reg + DMAR_ECAP_REG);
> >>
> >> + /* set agaw, "SAGAW" may be different across iommus */
> >> + sagaw = cap_sagaw(iommu->cap);
> >> + for (agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
> >> + agaw >= 0; agaw--)
> >> + if (test_bit(agaw, &sagaw))
> >> + break;
> >> + if (agaw < 0) {
> >> + printk(KERN_ERR "IOMMU: unsupported sagaw %lx\n",
> >> sagaw); + goto error; + }
> >> + iommu->agaw = agaw;
> >> +
> >> /* the registers might be more than one page */
> >> map_size = max_t(int, ecap_max_iotlb_offset(iommu->ecap),
> >> cap_max_fault_reg_offset(iommu->cap));
> >> diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
> >> index 5c8baa4..55b96c4 100644
> >> --- a/drivers/pci/intel-iommu.c
> >> +++ b/drivers/pci/intel-iommu.c
> >> @@ -50,8 +50,6 @@
> >> #define IOAPIC_RANGE_END (0xfeefffff)
> >> #define IOVA_START_ADDR (0x1000)
> >>
> >> -#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
> >> -
> >> #define DOMAIN_MAX_ADDR(gaw) ((((u64)1) << gaw) - 1)
> >>
> >>
> >> @@ -64,6 +62,7 @@ struct deferred_flush_tables {
> >> int next;
> >> struct iova *iova[HIGH_WATER_MARK];
> >> struct dmar_domain *domain[HIGH_WATER_MARK];
> >> + struct intel_iommu *iommu;
> >> };
> >>
> >> static struct deferred_flush_tables *deferred_flush;
> >> @@ -184,6 +183,69 @@ void free_iova_mem(struct iova *iova)
> >> kmem_cache_free(iommu_iova_cache, iova);
> >> }
> >>
> >> +/* in native case, each domain is related to only one iommu */
> >> +static struct intel_iommu *domain_get_only_iommu(struct dmar_domain
> >> *domain) +{ + struct dmar_drhd_unit *drhd;
> >> +
> >> + for_each_drhd_unit(drhd) {
> >> + if (drhd->ignored)
> >> + continue;
> >> + if (test_bit(drhd->iommu->seq_id,
> >> &domain->iommu_bmp)) + return drhd->iommu;
> >> + }
> >> +
> >> + return NULL;
> >> +}
> >> +
> >> +static void domain_flush_cache(struct dmar_domain *domain,
> >> + void *addr, int size) +{
> >> + struct intel_iommu *iommu;
> >> +
> >> + if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
> >> + struct dmar_drhd_unit *drhd;
> >> +
> >> + for_each_drhd_unit(drhd) {
> >> + if (drhd->ignored)
> >> + continue;
> >> + iommu = drhd->iommu;
> >> +
> >> + if (!test_bit(iommu->seq_id,
> >> &domain->iommu_bmp)) + continue;
> >> +
> >> + if (!ecap_coherent(iommu->ecap))
> >> + clflush_cache_range(addr, size); +
> >> } + }
> >> + else {
> >> + iommu = domain_get_only_iommu(domain);
> >> + if (iommu && !ecap_coherent(iommu->ecap))
> >> + clflush_cache_range(addr, size); + }
> >> +}
> >> +
> >> +static struct intel_iommu *device_find_matched_iommu(u8 bus, u8
> >> devfn) +{ + struct dmar_drhd_unit *drhd = NULL;
> >> + int i;
> >> +
> >> + for_each_drhd_unit(drhd) {
> >> + if (drhd->ignored)
> >> + continue;
> >> +
> >> + for (i = 0; i < drhd->devices_cnt; i++)
> >> + if (drhd->devices[i]->bus->number == bus &&
> >> + drhd->devices[i]->devfn == devfn)
> >> + return drhd->iommu; +
> >> + if (drhd->include_all)
> >> + return drhd->iommu;
> >> + }
> >> +
> >> + return NULL;
> >> +}
> >> +
> >> /* Gets context entry for a given bus and devfn */
> >> static struct context_entry * device_to_context_entry(struct
> >> intel_iommu *iommu, u8 bus, u8 devfn)
> >> @@ -287,7 +349,7 @@ static inline int agaw_to_width(int agaw)
> >>
> >> }
> >>
> >> -static inline int width_to_agaw(int width)
> >> +int width_to_agaw(int width)
> >> {
> >> return (width - 30) / LEVEL_STRIDE;
> >> }
> >> @@ -347,8 +409,7 @@ static struct dma_pte * addr_to_dma_pte(struct
> >> dmar_domain *domain, u64
> >> addr) flags); return NULL;
> >> }
> >> - __iommu_flush_cache(domain->iommu, tmp_page,
> >> - PAGE_SIZE);
> >> + domain_flush_cache(domain, tmp_page,
> >> PAGE_SIZE); dma_set_pte_addr(*pte,
> >> virt_to_phys(tmp_page));
> >> /* * high level table always sets r/w, last level page @@ -356,7
> >> +417,7 @@ static struct dma_pte *
> >> addr_to_dma_pte(struct dmar_domain *domain,
> >> u64 addr) */ dma_set_pte_readable(*pte);
> >> dma_set_pte_writable(*pte); -
> >> __iommu_flush_cache(domain->iommu, pte,
> >> sizeof(*pte)); + domain_flush_cache(domain,
> >> pte, sizeof(*pte)); } parent =
> >> phys_to_virt(dma_pte_addr(*pte)); level--; @@ -399,7
> >> +460,7 @@ static void dma_pte_clear_one(struct dmar_domain *domain,
> >> u64 addr)
> >>
> >> if (pte) {
> >> dma_clear_pte(*pte);
> >> - __iommu_flush_cache(domain->iommu, pte,
> >> sizeof(*pte)); + domain_flush_cache(domain, pte,
> >> sizeof(*pte)); } }
> >>
> >> @@ -447,8 +508,7 @@ static void dma_pte_free_pagetable(struct
> >> dmar_domain *domain,
> >> free_pgtable_page(
> >> phys_to_virt(dma_pte_addr(*pte)));
> >> dma_clear_pte(*pte);
> >> - __iommu_flush_cache(domain->iommu,
> >> - pte, sizeof(*pte));
> >> + domain_flush_cache(domain, pte,
> >> sizeof(*pte)); } tmp
> >> += level_size(level); }
> >> @@ -948,8 +1008,8 @@ static int iommu_init_domains(struct
> >> intel_iommu *iommu) return 0; }
> >>
> >> -
> >> static void domain_exit(struct dmar_domain *domain);
> >> +static void vm_domain_exit(struct dmar_domain *domain);
> >>
> >> void free_dmar_iommu(struct intel_iommu *iommu)
> >> {
> >> @@ -960,7 +1020,14 @@ void free_dmar_iommu(struct intel_iommu *iommu)
> >> for (; i < cap_ndoms(iommu->cap); ) {
> >> domain = iommu->domains[i];
> >> clear_bit(i, iommu->domain_ids);
> >> - domain_exit(domain);
> >> +
> >> + if (domain->flags & DOMAIN_FLAG_VIRTUAL_MACHINE) {
> >> + if (--domain->iommu_count == 0)
> >> + vm_domain_exit(domain); +
> >> } + else
> >> + domain_exit(domain);
> >> +
> >> i = find_next_bit(iommu->domain_ids,
> >> cap_ndoms(iommu->cap), i+1); }
> >> @@ -1006,8 +1073,11 @@ static struct dmar_domain *
> >> iommu_alloc_domain(struct intel_iommu *iommu)
> >>
> >> set_bit(num, iommu->domain_ids);
> >> domain->id = num;
> >> - domain->iommu = iommu;
> >> + memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
> >> + set_bit(iommu->seq_id, &domain->iommu_bmp);
> >> iommu->domains[num] = domain;
> >> + domain->iommu_count = 1;
> >> + domain->flags = 0;
> >> spin_unlock_irqrestore(&iommu->lock, flags);
> >>
> >> return domain;
> >> @@ -1016,10 +1086,12 @@ static struct dmar_domain *
> >> iommu_alloc_domain(struct intel_iommu *iommu) static void
> >> iommu_free_domain(struct dmar_domain *domain) { unsigned
> >> long flags; + struct intel_iommu *iommu;
> >>
> >> - spin_lock_irqsave(&domain->iommu->lock, flags);
> >> - clear_bit(domain->id, domain->iommu->domain_ids);
> >> - spin_unlock_irqrestore(&domain->iommu->lock, flags);
> >> + iommu = domain_get_only_iommu(domain);
> >> + spin_lock_irqsave(&iommu->lock, flags);
> >> + clear_bit(domain->id, iommu->domain_ids);
> >> + spin_unlock_irqrestore(&iommu->lock, flags); }
> >>
> >> static struct iova_domain reserved_iova_list;
> >> @@ -1098,7 +1170,7 @@ static int domain_init(struct dmar_domain
> >> *domain, int guest_width)
> >> domain_reserve_special_ranges(domain);
> >>
> >> /* calculate AGAW */
> >> - iommu = domain->iommu;
> >> + iommu = domain_get_only_iommu(domain);
> >> if (guest_width > cap_mgaw(iommu->cap))
> >> guest_width = cap_mgaw(iommu->cap);
> >> domain->gaw = guest_width;
> >> @@ -1107,19 +1179,21 @@ static int domain_init(struct dmar_domain
> >> *domain, int guest_width) sagaw = cap_sagaw(iommu->cap);
> >> if (!test_bit(agaw, &sagaw)) {
> >> /* hardware doesn't support it, choose a bigger one
> >> */ - pr_debug("IOMMU: hardware doesn't support agaw
> >> %d\n", agaw); + pr_debug("IOMMU: hardware doesn't
> >> support agaw %d\n", + agaw); agaw =
> >> find_next_bit(&sagaw, 5, agaw); if (agaw >= 5)
> >> return -ENODEV;
> >> }
> >> domain->agaw = agaw;
> >> +
> >> INIT_LIST_HEAD(&domain->devices);
> >>
> >> /* always allocate the top pgd */
> >> domain->pgd = (struct dma_pte *)alloc_pgtable_page();
> >> if (!domain->pgd) return -ENOMEM;
> >> - __iommu_flush_cache(iommu, domain->pgd, PAGE_SIZE);
> >> + domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
> >> return 0; }
> >>
> >> @@ -1148,10 +1222,9 @@ static void domain_exit(struct dmar_domain
> >> *domain) }
> >>
> >> static int domain_context_mapping_one(struct dmar_domain *domain,
> >> - u8 bus, u8 devfn)
> >> + struct intel_iommu *iommu, u8 bus, u8 devfn) {
> >> struct context_entry *context;
> >> - struct intel_iommu *iommu = domain->iommu;
> >> unsigned long flags;
> >>
> >> pr_debug("Set context mapping for %02x:%02x.%d\n",
> >> @@ -1191,9 +1264,14 @@ domain_context_mapping(struct dmar_domain
> >> *domain, struct pci_dev *pdev) { int ret;
> >> struct pci_dev *tmp, *parent;
> >> + struct intel_iommu *iommu;
> >> +
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return -ENODEV;
> >>
> >> - ret = domain_context_mapping_one(domain, pdev->bus->number,
> >> - pdev->devfn);
> >> + ret = domain_context_mapping_one(domain, iommu,
> >> + pdev->bus->number, pdev->devfn);
> >> if (ret)
> >> return ret;
> >>
> >> @@ -1204,27 +1282,31 @@ domain_context_mapping(struct dmar_domain
> >> *domain, struct pci_dev *pdev) /* Secondary interface's bus
> >> number and devfn 0 */ parent = pdev->bus->self;
> >> while (parent != tmp) {
> >> - ret = domain_context_mapping_one(domain,
> >> parent->bus->number,
> >> - parent->devfn);
> >> + ret = domain_context_mapping_one(domain, iommu,
> >> + parent->bus->number, parent->devfn);
> >> if (ret) return ret;
> >> parent = parent->bus->self;
> >> }
> >> if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
> >> - return domain_context_mapping_one(domain,
> >> + return domain_context_mapping_one(domain, iommu,
> >> tmp->subordinate->number, 0);
> >> else /* this is a legacy PCI bridge */
> >> - return domain_context_mapping_one(domain,
> >> + return domain_context_mapping_one(domain, iommu,
> >> tmp->bus->number, tmp->devfn); }
> >>
> >> -static int domain_context_mapped(struct dmar_domain *domain,
> >> - struct pci_dev *pdev)
> >> +static int domain_context_mapped(struct pci_dev *pdev) {
> >> int ret;
> >> struct pci_dev *tmp, *parent;
> >> + struct intel_iommu *iommu;
> >>
> >> - ret = device_context_mapped(domain->iommu,
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return 0;
> >> +
> >> + ret = device_context_mapped(iommu,
> >> pdev->bus->number, pdev->devfn);
> >> if (!ret)
> >> return ret;
> >> @@ -1235,17 +1317,17 @@ static int domain_context_mapped(struct
> >> dmar_domain *domain, /* Secondary interface's bus number and
> >> devfn 0 */ parent = pdev->bus->self;
> >> while (parent != tmp) {
> >> - ret = device_context_mapped(domain->iommu,
> >> parent->bus->number, + ret =
> >> device_context_mapped(iommu,
> >> parent->bus->number, parent->devfn); if (!ret)
> >> return ret;
> >> parent = parent->bus->self;
> >> }
> >> if (tmp->is_pcie)
> >> - return device_context_mapped(domain->iommu,
> >> + return device_context_mapped(iommu,
> >> tmp->subordinate->number, 0); else
> >> - return device_context_mapped(domain->iommu,
> >> + return device_context_mapped(iommu,
> >> tmp->bus->number, tmp->devfn); }
> >>
> >> @@ -1276,20 +1358,27 @@ domain_page_mapping(struct dmar_domain
> >> *domain, dma_addr_t iova, BUG_ON(dma_pte_addr(*pte));
> >> dma_set_pte_addr(*pte, start_pfn << VTD_PAGE_SHIFT);
> >> dma_set_pte_prot(*pte, prot);
> >> - __iommu_flush_cache(domain->iommu, pte,
> >> sizeof(*pte)); + domain_flush_cache(domain, pte,
> >> sizeof(*pte)); start_pfn++; index++;
> >> }
> >> return 0;
> >> }
> >>
> >> -static void detach_domain_for_dev(struct dmar_domain *domain, u8
> >> bus, u8 devfn) +static void detach_domain_for_dev(struct dmar_domain
> >> *domain, + u8 bus, u8 devfn) {
> >> - clear_context_table(domain->iommu, bus, devfn);
> >> - domain->iommu->flush.flush_context(domain->iommu, 0, 0, 0,
> >> - DMA_CCMD_GLOBAL_INVL, 0);
> >> - domain->iommu->flush.flush_iotlb(domain->iommu, 0, 0, 0,
> >> - DMA_TLB_GLOBAL_FLUSH, 0);
> >> + struct intel_iommu *iommu;
> >> +
> >> + iommu = device_find_matched_iommu(bus, devfn); + if
> >> (!iommu) + return;
> >> +
> >> + clear_context_table(iommu, bus, devfn);
> >> + iommu->flush.flush_context(iommu, 0, 0, 0,
> >> + DMA_CCMD_GLOBAL_INVL, 0);
> >> + iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> >> + DMA_TLB_GLOBAL_FLUSH, 0); }
> >>
> >> static void domain_remove_dev_info(struct dmar_domain *domain)
> >> @@ -1336,7 +1425,6 @@ static struct dmar_domain
> >> *get_domain_for_dev(struct pci_dev *pdev, int gaw) { struct
> >> dmar_domain *domain, *found = NULL; struct intel_iommu
> >> *iommu; - struct dmar_drhd_unit *drhd;
> >> struct device_domain_info *info, *tmp;
> >> struct pci_dev *dev_tmp;
> >> unsigned long flags;
> >> @@ -1371,13 +1459,9 @@ static struct dmar_domain
> >> *get_domain_for_dev(struct pci_dev *pdev, int gaw) }
> >>
> >> /* Allocate new domain for the device */
> >> - drhd = dmar_find_matched_drhd_unit(pdev);
> >> - if (!drhd) {
> >> - printk(KERN_ERR "IOMMU: can't find DMAR for device
> >> %s\n",
> >> - pci_name(pdev));
> >> - return NULL;
> >> - }
> >> - iommu = drhd->iommu;
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return NULL;
> >>
> >> domain = iommu_alloc_domain(iommu);
> >> if (!domain)
> >> @@ -1400,7 +1484,7 @@ static struct dmar_domain
> >> *get_domain_for_dev(struct pci_dev *pdev, int gaw)
> >> info->dev = NULL; info->domain = domain;
> >> /* This domain is shared by devices under p2p bridge
> >> */ - domain->flags |= DOMAIN_FLAG_MULTIPLE_DEVICES;
> >> + domain->flags |= DOMAIN_FLAG_P2P_MULTIPLE_DEVICES;
> >>
> >> /* pcie-to-pci bridge already has a domain, uses it
> >> */ found = NULL; @@ -1805,7 +1889,7 @@
> >> get_valid_domain_for_dev(struct pci_dev *pdev) }
> >>
> >> /* make sure context mapping is ok */
> >> - if (unlikely(!domain_context_mapped(domain, pdev))) {
> >> + if (unlikely(!domain_context_mapped(pdev))) {
> >> ret = domain_context_mapping(domain, pdev);
> >> if (ret) { printk(KERN_ERR
> >> @@ -1823,6 +1907,7 @@ static dma_addr_t __intel_map_single(struct
> >> device *hwdev, phys_addr_t paddr, { struct pci_dev *pdev =
> >> to_pci_dev(hwdev); struct dmar_domain *domain;
> >> + struct intel_iommu *iommu;
> >> phys_addr_t start_paddr;
> >> struct iova *iova;
> >> int prot = 0;
> >> @@ -1836,6 +1921,10 @@ static dma_addr_t __intel_map_single(struct
> >> device *hwdev, phys_addr_t paddr, if
> >> (!domain) return 0;
> >>
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return 0;
> >> +
> >> size = aligned_size((u64)paddr, size);
> >>
> >> iova = __intel_alloc_iova(hwdev, domain, size,
> >> pdev->dma_mask); @@ -1849,7 +1938,7 @@ static dma_addr_t
> >> __intel_map_single(struct device *hwdev, phys_addr_t paddr,
> >> * mappings..
> >> */
> >> if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
> >> - !cap_zlr(domain->iommu->cap))
> >> + !cap_zlr(iommu->cap))
> >> prot |= DMA_PTE_READ;
> >> if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
> >> prot |= DMA_PTE_WRITE;
> >> @@ -1865,10 +1954,10 @@ static dma_addr_t __intel_map_single(struct
> >> device *hwdev, phys_addr_t paddr, goto error;
> >>
> >> /* it's a non-present to present mapping */
> >> - ret = iommu_flush_iotlb_psi(domain->iommu, domain->id,
> >> + ret = iommu_flush_iotlb_psi(iommu, domain->id,
> >> start_paddr, size >> VTD_PAGE_SHIFT, 1);
> >> if (ret) - iommu_flush_write_buffer(domain->iommu);
> >> + iommu_flush_write_buffer(iommu);
> >>
> >> return start_paddr + ((u64)paddr & (~PAGE_MASK));
> >>
> >> @@ -1896,8 +1985,7 @@ static void flush_unmaps(void)
> >> /* just flush them all */
> >> for (i = 0; i < g_num_of_iommus; i++) {
> >> if (deferred_flush[i].next) {
> >> - struct intel_iommu *iommu =
> >> - deferred_flush[i].domain[0]->iommu;
> >> + struct intel_iommu *iommu =
> >> deferred_flush[i].iommu;
> >>
> >> iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> >>
> >> DMA_TLB_GLOBAL_FLUSH, 0); @@ -1921,7 +2009,8 @@ static void
> >> flush_unmaps_timeout(unsigned long data)
> >> spin_unlock_irqrestore(&async_umap_flush_lock, flags); }
> >>
> >> -static void add_unmap(struct dmar_domain *dom, struct iova *iova)
> >> +static void add_unmap(struct dmar_domain *dom,
> >> + struct intel_iommu *iommu, struct iova *iova)
> >> { unsigned long flags;
> >> int next, iommu_id;
> >> @@ -1930,11 +2019,12 @@ static void add_unmap(struct dmar_domain
> >> *dom, struct iova *iova) if (list_size == HIGH_WATER_MARK)
> >> flush_unmaps();
> >>
> >> - iommu_id = dom->iommu->seq_id;
> >> + iommu_id = iommu->seq_id;
> >>
> >> next = deferred_flush[iommu_id].next;
> >> deferred_flush[iommu_id].domain[next] = dom;
> >> deferred_flush[iommu_id].iova[next] = iova;
> >> + deferred_flush[iommu_id].iommu = iommu;
> >> deferred_flush[iommu_id].next++;
> >>
> >> if (!timer_on) {
> >> @@ -1952,12 +2042,17 @@ void intel_unmap_single(struct device *dev,
> >> dma_addr_t dev_addr, size_t size, struct dmar_domain *domain;
> >> unsigned long start_addr;
> >> struct iova *iova;
> >> + struct intel_iommu *iommu;
> >>
> >> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> >> return; domain = find_domain(pdev);
> >> BUG_ON(!domain);
> >>
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return;
> >> +
> >> iova = find_iova(&domain->iovad, IOVA_PFN(dev_addr));
> >> if (!iova) return;
> >> @@ -1973,13 +2068,13 @@ void intel_unmap_single(struct device *dev,
> >> dma_addr_t dev_addr, size_t size, /* free page tables */
> >> dma_pte_free_pagetable(domain, start_addr, start_addr +
> >> size); if (intel_iommu_strict) {
> >> - if (iommu_flush_iotlb_psi(domain->iommu,
> >> + if (iommu_flush_iotlb_psi(iommu,
> >> domain->id, start_addr, size >>
> >> VTD_PAGE_SHIFT, 0)) -
> >> iommu_flush_write_buffer(domain->iommu); +
> >> iommu_flush_write_buffer(iommu); /* free iova */
> >> __free_iova(&domain->iovad, iova);
> >> } else {
> >> - add_unmap(domain, iova);
> >> + add_unmap(domain, iommu, iova);
> >> /*
> >> * queue up the release of the unmap to save the
> >> 1/6th of the
> >> * cpu used up by the iotlb flush operation...
> >> @@ -2036,12 +2131,17 @@ void intel_unmap_sg(struct device *hwdev,
> >> struct scatterlist *sglist, size_t size = 0; void
> >> *addr; struct scatterlist *sg;
> >> + struct intel_iommu *iommu;
> >>
> >> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> >> return;
> >>
> >> domain = find_domain(pdev);
> >>
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return;
> >> +
> >> iova = find_iova(&domain->iovad,
> >> IOVA_PFN(sglist[0].dma_address)); if (!iova)
> >> return; @@ -2057,9 +2157,9 @@ void intel_unmap_sg(struct device
> >> *hwdev, struct scatterlist *sglist, /* free page tables */
> >> dma_pte_free_pagetable(domain, start_addr, start_addr +
> >> size);
> >>
> >> - if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
> >> start_addr, + if (iommu_flush_iotlb_psi(iommu, domain->id,
> >> start_addr, size >> VTD_PAGE_SHIFT, 0))
> >> - iommu_flush_write_buffer(domain->iommu);
> >> + iommu_flush_write_buffer(iommu);
> >>
> >> /* free iova */
> >> __free_iova(&domain->iovad, iova);
> >> @@ -2093,6 +2193,7 @@ int intel_map_sg(struct device *hwdev, struct
> >> scatterlist *sglist, int nelems, int ret; struct
> >> scatterlist *sg; unsigned long start_addr;
> >> + struct intel_iommu *iommu;
> >>
> >> BUG_ON(dir == DMA_NONE);
> >> if (pdev->dev.archdata.iommu == DUMMY_DEVICE_DOMAIN_INFO)
> >> @@ -2102,6 +2203,10 @@ int intel_map_sg(struct device *hwdev, struct
> >> scatterlist *sglist, int nelems, if
> >> (!domain) return 0;
> >>
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return 0;
> >> +
> >> for_each_sg(sglist, sg, nelems, i) {
> >> addr = SG_ENT_VIRT_ADDRESS(sg);
> >> addr = (void *)virt_to_phys(addr);
> >> @@ -2119,7 +2224,7 @@ int intel_map_sg(struct device *hwdev, struct
> >> scatterlist *sglist, int nelems,
> >> * mappings..
> >> */
> >> if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL || \
> >> - !cap_zlr(domain->iommu->cap))
> >> + !cap_zlr(iommu->cap))
> >> prot |= DMA_PTE_READ;
> >> if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL)
> >> prot |= DMA_PTE_WRITE;
> >> @@ -2151,9 +2256,9 @@ int intel_map_sg(struct device *hwdev, struct
> >> scatterlist *sglist, int nelems, }
> >>
> >> /* it's a non-present to present mapping */
> >> - if (iommu_flush_iotlb_psi(domain->iommu, domain->id,
> >> + if (iommu_flush_iotlb_psi(iommu, domain->id,
> >> start_addr, offset >> VTD_PAGE_SHIFT, 1))
> >> - iommu_flush_write_buffer(domain->iommu);
> >> + iommu_flush_write_buffer(iommu);
> >> return nelems;
> >> }
> >>
> >> @@ -2328,7 +2433,314 @@ int __init intel_iommu_init(void)
> >> return 0; }
> >>
> >> -void intel_iommu_domain_exit(struct dmar_domain *domain)
> >> +/* domain id for virtual machine, it won't be set in context */
> >> +static unsigned long vm_domid;
> >> +
> >> +static int vm_domain_min_agaw(struct dmar_domain *domain) +{
> >> + struct dmar_drhd_unit *drhd;
> >> + struct intel_iommu *iommu;
> >> + int min_agaw = domain->agaw;
> >> +
> >> + for_each_drhd_unit(drhd) {
> >> + if (drhd->ignored)
> >> + continue;
> >> + iommu = drhd->iommu;
> >> +
> >> + if (test_bit(iommu->seq_id, &domain->iommu_bmp))
> >> + if (min_agaw > iommu->agaw)
> >> + min_agaw = iommu->agaw; + }
> >> +
> >> + return min_agaw;
> >> +}
> >> +
> >> +static int vm_domain_add_dev_info(struct dmar_domain *domain,
> >> + struct pci_dev *pdev) +{
> >> + struct device_domain_info *info;
> >> + unsigned long flags;
> >> +
> >> + info = alloc_devinfo_mem();
> >> + if (!info)
> >> + return -ENOMEM;
> >> +
> >> + info->bus = pdev->bus->number;
> >> + info->devfn = pdev->devfn;
> >> + info->dev = pdev;
> >> + info->domain = domain;
> >> +
> >> + spin_lock_irqsave(&device_domain_lock, flags);
> >> + list_add(&info->link, &domain->devices);
> >> + list_add(&info->global, &device_domain_list);
> >> + pdev->dev.archdata.iommu = info;
> >> + spin_unlock_irqrestore(&device_domain_lock, flags); +
> >> + return 0;
> >> +}
> >> +
> >> +static void vm_domain_remove_one_dev_info(struct dmar_domain
> >> *domain, + struct pci_dev
> >> *pdev) +{ + struct device_domain_info *info;
> >> + struct intel_iommu *iommu;
> >> + unsigned long flags;
> >> + int found = 0;
> >> +
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + + spin_lock_irqsave(&device_domain_lock,
> >> flags); + while (!list_empty(&domain->devices)) {
> >> + info = list_entry(domain->devices.next,
> >> + struct device_domain_info, link);
> >> + if (info->bus == pdev->bus->number &&
> >> + info->devfn == pdev->devfn) {
> >> + list_del(&info->link);
> >> + list_del(&info->global);
> >> + if (info->dev)
> >> + info->dev->dev.archdata.iommu = NULL;
> >> + spin_unlock_irqrestore(&device_domain_lock,
> >> flags); + + detach_domain_for_dev(info->domain,
> >> + info->bus,
> >> info->devfn); + free_devinfo_mem(info);
> >> +
> >> + spin_lock_irqsave(&device_domain_lock,
> >> flags); + + if (found)
> >> + break;
> >> + else
> >> + continue;
> >> + }
> >> +
> >> + /* if there is no other devices under the same iommu
> >> + * owned by this domain, clear this iommu in
> >> iommu_bmp + */ + if
> >> (device_find_matched_iommu(info->bus, info->devfn) == iommu) +
> >> found = 1; + }
> >> +
> >> + if (found == 0) {
> >> + spin_lock_irqsave(&iommu->lock, flags);
> >> + clear_bit(iommu->seq_id, &domain->iommu_bmp);
> >> + domain->iommu_count--;
> >> + spin_unlock_irqrestore(&iommu->lock, flags); +
> >> } +
> >> + spin_unlock_irqrestore(&device_domain_lock, flags); +}
> >> +
> >> +static void vm_domain_remove_all_dev_info(struct dmar_domain
> >> *domain) +{ + struct device_domain_info *info;
> >> + struct intel_iommu *iommu;
> >> + unsigned long flags;
> >> +
> >> + spin_lock_irqsave(&device_domain_lock, flags);
> >> + while (!list_empty(&domain->devices)) {
> >> + info = list_entry(domain->devices.next,
> >> + struct device_domain_info, link);
> >> + list_del(&info->link);
> >> + list_del(&info->global);
> >> + if (info->dev)
> >> + info->dev->dev.archdata.iommu = NULL; +
> >> + spin_unlock_irqrestore(&device_domain_lock, flags);
> >> + detach_domain_for_dev(info->domain,
> >> + info->bus, info->devfn); +
> >> + /* clear this iommu in iommu_bmp */
> >> + iommu = device_find_matched_iommu(info->bus,
> >> info->devfn); + spin_lock_irqsave(&iommu->lock, flags);
> >> + if (test_and_clear_bit(iommu->seq_id,
> >> + &domain->iommu_bmp))
> >> + domain->iommu_count--;
> >> + spin_unlock_irqrestore(&iommu->lock, flags); +
> >> + free_devinfo_mem(info);
> >> + spin_lock_irqsave(&device_domain_lock, flags); +
> >> } + spin_unlock_irqrestore(&device_domain_lock, flags); +}
> >> +
> >> +static int vm_domain_context_mapping_one(struct dmar_domain *domain,
> >> + struct intel_iommu *iommu, u8 bus, u8 devfn) +{
> >> + struct context_entry *context;
> >> + unsigned long flags;
> >> + struct dma_pte *pgd;
> >> + unsigned long num;
> >> + unsigned long ndomains;
> >> + int id;
> >> + int agaw;
> >> + int found = 0;
> >> +
> >> + pr_debug("Set context mapping for %02x:%02x.%d\n",
> >> + bus, PCI_SLOT(devfn), PCI_FUNC(devfn)); +
> >> BUG_ON(!domain->pgd); + context =
> >> device_to_context_entry(iommu, bus, devfn); + if (!context) +
> >> return -ENOMEM; + spin_lock_irqsave(&iommu->lock, flags);
> >> + if (context_present(*context)) {
> >> + spin_unlock_irqrestore(&iommu->lock, flags); +
> >> return 0; + }
> >> +
> >> + id = domain->id;
> >> +
> >> + /* find an available domain id for this device in iommu */
> >> + ndomains = cap_ndoms(iommu->cap);
> >> + num = find_first_bit(iommu->domain_ids, ndomains);
> >> + for (; num < ndomains; ) {
> >> + if (iommu->domains[num] == domain) {
> >> + id = num;
> >> + found = 1;
> >> + break;
> >> + }
> >> + num = find_next_bit(iommu->domain_ids,
> >> + cap_ndoms(iommu->cap), num+1);
> >> + }
> >> +
> >> + if (found == 0) {
> >> + num = find_first_zero_bit(iommu->domain_ids,
> >> ndomains); + if (num >= ndomains) {
> >> + spin_unlock_irqrestore(&iommu->lock, flags);
> >> + printk(KERN_ERR "IOMMU: no free domain
> >> ids\n"); + return -EFAULT;
> >> + }
> >> +
> >> + set_bit(num, iommu->domain_ids);
> >> + iommu->domains[num] = domain;
> >> + id = num;
> >> + }
> >> +
> >> + pgd = domain->pgd;
> >> +
> >> + /* Skip top levels of page tables for
> >> + * iommu which has less agaw than default.
> >> + */
> >> + for (agaw = domain->agaw; agaw != iommu->agaw; agaw--) {
> >> + pgd = phys_to_virt(dma_pte_addr(*pgd));
> >> + if (!dma_pte_present(*pgd)) {
> >> + spin_unlock_irqrestore(&iommu->lock, flags);
> >> + return -ENOMEM;
> >> + }
> >> + }
> >> +
> >> + context_set_domain_id(*context, id);
> >> + context_set_address_width(*context, iommu->agaw);
> >> + context_set_address_root(*context, virt_to_phys(pgd));
> >> + context_set_translation_type(*context,
> >> CONTEXT_TT_MULTI_LEVEL); + context_set_fault_enable(*context);
> >> + context_set_present(*context);
> >> + __iommu_flush_cache(iommu, context, sizeof(*context)); +
> >> + /* it's a non-present to present mapping */
> >> + if (iommu->flush.flush_context(iommu, id,
> >> + (((u16)bus) << 8) | devfn, DMA_CCMD_MASK_NOBIT,
> >> + DMA_CCMD_DEVICE_INVL, 1))
> >> + iommu_flush_write_buffer(iommu);
> >> + else
> >> + iommu->flush.flush_iotlb(iommu, 0, 0, 0,
> >> DMA_TLB_DSI_FLUSH, 0); + + if
> >> (!test_and_set_bit(iommu->seq_id, &domain->iommu_bmp)) +
> >> domain->iommu_count++; + spin_unlock_irqrestore(&iommu->lock,
> >> flags); + return 0; +}
> >> +
> >> +static int
> >> +vm_domain_context_mapping(struct dmar_domain *domain, struct
> >> pci_dev *pdev) +{ + int ret;
> >> + struct pci_dev *tmp, *parent;
> >> + struct intel_iommu *iommu;
> >> +
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return -ENODEV;
> >> +
> >> + ret = vm_domain_context_mapping_one(domain, iommu,
> >> + pdev->bus->number, pdev->devfn);
> >> + if (ret)
> >> + return ret;
> >> +
> >> + /* dependent device mapping */
> >> + tmp = pci_find_upstream_pcie_bridge(pdev);
> >> + if (!tmp)
> >> + return 0;
> >> + /* Secondary interface's bus number and devfn 0 */
> >> + parent = pdev->bus->self;
> >> + while (parent != tmp) {
> >> + ret = vm_domain_context_mapping_one(domain, iommu,
> >> + parent->bus->number, parent->devfn); +
> >> if (ret) + return ret;
> >> + parent = parent->bus->self;
> >> + }
> >> + if (tmp->is_pcie) /* this is a PCIE-to-PCI bridge */
> >> + return vm_domain_context_mapping_one(domain, iommu,
> >> + tmp->subordinate->number, 0);
> >> + else /* this is a legacy PCI bridge */
> >> + return vm_domain_context_mapping_one(domain, iommu,
> >> + tmp->bus->number, tmp->devfn); +}
> >> +
> >> +
> >> +static int vm_domain_init(struct dmar_domain *domain, int
> >> guest_width) +{ + int adjust_width;
> >> +
> >> + init_iova_domain(&domain->iovad, DMA_32BIT_PFN);
> >> + spin_lock_init(&domain->mapping_lock);
> >> +
> >> + domain_reserve_special_ranges(domain);
> >> +
> >> + /* calculate AGAW */
> >> + domain->gaw = guest_width;
> >> + adjust_width = guestwidth_to_adjustwidth(guest_width);
> >> + domain->agaw = width_to_agaw(adjust_width); +
> >> + INIT_LIST_HEAD(&domain->devices);
> >> +
> >> + /* always allocate the top pgd */
> >> + domain->pgd = (struct dma_pte *)alloc_pgtable_page(); +
> >> if (!domain->pgd) + return -ENOMEM;
> >> + domain_flush_cache(domain, domain->pgd, PAGE_SIZE); +
> >> return 0; +}
> >> +
> >> +static void iommu_free_vm_domain(struct dmar_domain *domain) +{
> >> + unsigned long flags;
> >> + struct dmar_drhd_unit *drhd;
> >> + struct intel_iommu *iommu;
> >> + unsigned long i;
> >> + unsigned long ndomains;
> >> +
> >> + for_each_drhd_unit(drhd) {
> >> + if (drhd->ignored)
> >> + continue;
> >> + iommu = drhd->iommu;
> >> +
> >> + ndomains = cap_ndoms(iommu->cap);
> >> + i = find_first_bit(iommu->domain_ids, ndomains);
> >> + for (; i < ndomains; ) {
> >> + if (iommu->domains[i] == domain) {
> >> + spin_lock_irqsave(&iommu->lock,
> >> flags); + clear_bit(i,
> >> iommu->domain_ids); +
> >> iommu->domains[i] = NULL; +
> >> spin_unlock_irqrestore(&iommu->lock, flags); +
> >> break; + }
> >> + i = find_next_bit(iommu->domain_ids,
> >> ndomains, i+1); + } + }
> >> +}
> >> +
> >> +static void vm_domain_exit(struct dmar_domain *domain) {
> >> u64 end;
> >>
> >> @@ -2336,8 +2748,11 @@ void intel_iommu_domain_exit(struct
> >> dmar_domain *domain) if (!domain) return;
> >>
> >> + vm_domain_remove_all_dev_info(domain);
> >> + /* destroy iovas */
> >> + put_iova_domain(&domain->iovad);
> >> end = DOMAIN_MAX_ADDR(domain->gaw);
> >> - end = end & (~VTD_PAGE_MASK);
> >> + end &= VTD_PAGE_MASK;
> >>
> >> /* clear ptes */
> >> dma_pte_clear_range(domain, 0, end);
> >> @@ -2345,76 +2760,149 @@ void intel_iommu_domain_exit(struct
> >> dmar_domain *domain) /* free page tables */
> >> dma_pte_free_pagetable(domain, 0, end);
> >>
> >> - iommu_free_domain(domain);
> >> + iommu_free_vm_domain(domain);
> >> free_domain_mem(domain);
> >> }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_domain_exit);
> >>
> >> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev)
> >> +static struct dmar_domain *iommu_alloc_vm_domain(void) {
> >> - struct dmar_drhd_unit *drhd;
> >> struct dmar_domain *domain;
> >> - struct intel_iommu *iommu;
> >>
> >> - drhd = dmar_find_matched_drhd_unit(pdev);
> >> - if (!drhd) {
> >> - printk(KERN_ERR "intel_iommu_domain_alloc: drhd ==
> >> NULL\n"); + domain = alloc_domain_mem();
> >> + if (!domain)
> >> return NULL;
> >> - }
> >>
> >> - iommu = drhd->iommu;
> >> - if (!iommu) {
> >> - printk(KERN_ERR
> >> - "intel_iommu_domain_alloc: iommu == NULL\n");
> >> - return NULL;
> >> - }
> >> - domain = iommu_alloc_domain(iommu);
> >> + domain->id = vm_domid++;
> >> + domain->iommu_count = 0;
> >> + domain->max_addr = 0;
> >> + memset(&domain->iommu_bmp, 0, sizeof(unsigned long));
> >> + domain->flags = DOMAIN_FLAG_VIRTUAL_MACHINE; +
> >> + return domain;
> >> +}
> >> +
> >> +struct dmar_domain *intel_iommu_alloc_domain(void) +{
> >> + struct dmar_domain *domain;
> >> +
> >> + domain = iommu_alloc_vm_domain();
> >> if (!domain) {
> >> printk(KERN_ERR
> >> "intel_iommu_domain_alloc: domain ==
> >> NULL\n"); return NULL; }
> >> - if (domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
> >> + if (vm_domain_init(domain, DEFAULT_DOMAIN_ADDRESS_WIDTH)) {
> >> printk(KERN_ERR
> >> "intel_iommu_domain_alloc: domain_init()
> >> failed\n"); - intel_iommu_domain_exit(domain);
> >> + vm_domain_exit(domain);
> >> return NULL;
> >> }
> >> +
> >> return domain;
> >> }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_domain_alloc);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_alloc_domain);
> >> +
> >> +void intel_iommu_free_domain(struct dmar_domain *domain) +{
> >> + vm_domain_exit(domain);
> >> +}
> >> +EXPORT_SYMBOL_GPL(intel_iommu_free_domain);
> >>
> >> -int intel_iommu_context_mapping(
> >> - struct dmar_domain *domain, struct pci_dev *pdev)
> >> +int intel_iommu_assign_device(struct dmar_domain *domain,
> >> + struct pci_dev *pdev) {
> >> - int rc;
> >> - rc = domain_context_mapping(domain, pdev);
> >> - return rc;
> >> + struct intel_iommu *iommu;
> >> + int addr_width;
> >> + u64 end;
> >> + int ret;
> >> +
> >> + /* normally pdev is not mapped */
> >> + if (unlikely(domain_context_mapped(pdev))) {
> >> + struct dmar_domain *old_domain;
> >> +
> >> + old_domain = find_domain(pdev);
> >> + if (old_domain) {
> >> + if (domain->flags &
> >> DOMAIN_FLAG_VIRTUAL_MACHINE) +
> >> vm_domain_remove_one_dev_info(old_domain, pdev); +
> >> else +
> >> domain_remove_dev_info(old_domain); + } + } +
> >> + iommu = device_find_matched_iommu(pdev->bus->number,
> >> pdev->devfn); + if (!iommu) + return -ENODEV;
> >> +
> >> + /* check if this iommu agaw is sufficient for max mapped
> >> address */ + addr_width = agaw_to_width(iommu->agaw);
> >> + end = DOMAIN_MAX_ADDR(addr_width);
> >> + end = end & VTD_PAGE_MASK;
> >> + if (end < domain->max_addr) {
> >> + printk(KERN_ERR "%s: iommu agaw (%d) is not "
> >> + "sufficient for the mapped address (%llx)\n",
> >> + __func__, iommu->agaw, domain->max_addr);
> >> + return -EFAULT;
> >> + }
> >> +
> >> + ret = vm_domain_context_mapping(domain, pdev); + if
> >> (ret) + return ret;
> >> +
> >> + ret = vm_domain_add_dev_info(domain, pdev); + return
> >> ret; }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_context_mapping);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_assign_device);
> >>
> >> -int intel_iommu_page_mapping(
> >> - struct dmar_domain *domain, dma_addr_t iova,
> >> - u64 hpa, size_t size, int prot)
> >> +
> >> +void intel_iommu_deassign_device(struct dmar_domain *domain,
> >> + struct pci_dev *pdev) {
> >> - int rc;
> >> - rc = domain_page_mapping(domain, iova, hpa, size, prot);
> >> - return rc;
> >> + vm_domain_remove_one_dev_info(domain, pdev); }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_page_mapping);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_deassign_device);
> >>
> >> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8
> >> devfn) +int intel_iommu_map_pages(struct dmar_domain *domain,
> >> dma_addr_t iova, + u64 hpa, size_t size, int
> >> prot) { - detach_domain_for_dev(domain, bus, devfn);
> >> + u64 max_addr;
> >> + int addr_width;
> >> + int ret;
> >> +
> >> + max_addr = (iova & VTD_PAGE_MASK) + VTD_PAGE_ALIGN(size);
> >> + if (domain->max_addr < max_addr) {
> >> + int min_agaw;
> >> + u64 end;
> >> +
> >> + /* check if minimum agaw is sufficient for mapped
> >> address */ + min_agaw = vm_domain_min_agaw(domain);
> >> + addr_width = agaw_to_width(min_agaw);
> >> + end = DOMAIN_MAX_ADDR(addr_width);
> >> + end = end & VTD_PAGE_MASK;
> >> + if (end < max_addr) {
> >> + printk(KERN_ERR "%s: iommu agaw (%d) is not "
> >> + "sufficient for the mapped address
> >> (%llx)\n", + __func__, min_agaw,
> >> max_addr); + return -EFAULT;
> >> + }
> >> + domain->max_addr = max_addr;
> >> + }
> >> +
> >> + ret = domain_page_mapping(domain, iova, hpa, size, prot); +
> >> return ret; }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_detach_dev);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_map_pages);
> >>
> >> -struct dmar_domain *
> >> -intel_iommu_find_domain(struct pci_dev *pdev)
> >> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
> >> + dma_addr_t iova, size_t size) {
> >> - return find_domain(pdev);
> >> + dma_addr_t base;
> >> +
> >> + /* The address might not be aligned */
> >> + base = iova & PAGE_MASK;
> >> + size = PAGE_ALIGN(size);
> >> + dma_pte_clear_range(domain, base, base + size); }
> >> -EXPORT_SYMBOL_GPL(intel_iommu_find_domain);
> >> +EXPORT_SYMBOL_GPL(intel_iommu_unmap_pages);
> >>
> >> int intel_iommu_found(void)
> >> {
> >> diff --git a/include/linux/dma_remapping.h
> >> b/include/linux/dma_remapping.h
> >> index 952df39..3568ae6 100644
> >> --- a/include/linux/dma_remapping.h
> >> +++ b/include/linux/dma_remapping.h
> >> @@ -111,11 +111,21 @@ struct dma_pte {
> >> (p).val |= ((addr) & VTD_PAGE_MASK); } while (0)
> >> #define dma_pte_present(p) (((p).val & 3) != 0)
> >>
> >> +/* domain flags, one domain owns one device by default */ +
> >> +/* devices under the same p2p bridge are owned in one domain */
> >> +#define DOMAIN_FLAG_P2P_MULTIPLE_DEVICES (1 << 0) +
> >> +/* domain represents a virtual machine, more than one devices
> >> + * across iommus may be owned in one domain, e.g. kvm guest. + */
> >> +#define DOMAIN_FLAG_VIRTUAL_MACHINE (1 << 1) +
> >> struct intel_iommu;
> >>
> >> struct dmar_domain {
> >> int id; /* domain id */
> >> - struct intel_iommu *iommu; /* back pointer to owning
> >> iommu */ + unsigned long iommu_bmp; /* bitmap of iommus
> >> this domain uses*/
> >>
> >> struct list_head devices; /* all devices' list */
> >> struct iova_domain iovad; /* iova's that belong to
> >> this domain */ @@ -123,12 +133,13 @@ struct dmar_domain {
> >> struct dma_pte *pgd; /* virtual address */
> >> spinlock_t mapping_lock; /* page table lock */
> >> int gaw; /* max guest address width */
> >> + int agaw; /* adjusted guest address
> >> width */
> >>
> >> - /* adjusted guest address width, 0 is level 2 30-bit */
> >> - int agaw;
> >> + int flags; /* domain flag */
> >>
> >> -#define DOMAIN_FLAG_MULTIPLE_DEVICES 1
> >> - int flags;
> >> + /* following fields are used in virtual machine case */
> >> + int iommu_count; /* reference count of iommu
> >> */ + u64 max_addr; /* maximum mapped address
> >> */ };
> >>
> >> /* PCI domain-device relationship */
> >> diff --git a/include/linux/intel-iommu.h
> >> b/include/linux/intel-iommu.h
> >> index 3d017cf..c2f37b8 100644
> >> --- a/include/linux/intel-iommu.h
> >> +++ b/include/linux/intel-iommu.h
> >> @@ -219,6 +219,8 @@ do {
> >> \ }
> >> \ } while (0)
> >>
> >> +#define DEFAULT_DOMAIN_ADDRESS_WIDTH 48
> >> +
> >> #define QI_LENGTH 256 /* queue length */
> >>
> >> enum {
> >> @@ -299,6 +301,7 @@ struct intel_iommu {
> >> struct dmar_domain **domains; /* ptr to domains */
> >> spinlock_t lock; /* protect context, domain ids */
> >> struct root_entry *root_entry; /* virtual address */ +
> >> int agaw;
> >>
> >> unsigned int irq;
> >> unsigned char name[7]; /* Device Name */
> >> @@ -334,14 +337,16 @@ extern int qi_flush_iotlb(struct intel_iommu
> >> *iommu, u16 did, u64 addr,
> >>
> >> extern void qi_submit_sync(struct qi_desc *desc, struct intel_iommu
> >> *iommu);
> >>
> >> -void intel_iommu_domain_exit(struct dmar_domain *domain);
> >> -struct dmar_domain *intel_iommu_domain_alloc(struct pci_dev *pdev);
> >> -int intel_iommu_context_mapping(struct dmar_domain *domain,
> >> - struct pci_dev *pdev);
> >> -int intel_iommu_page_mapping(struct dmar_domain *domain, dma_addr_t
> >> iova,
> >> - u64 hpa, size_t size, int prot);
> >> -void intel_iommu_detach_dev(struct dmar_domain *domain, u8 bus, u8
> >> devfn);
> >> -struct dmar_domain *intel_iommu_find_domain(struct pci_dev *pdev);
> >> +struct dmar_domain *intel_iommu_alloc_domain(void);
> >> +void intel_iommu_free_domain(struct dmar_domain *domain);
> >> +int intel_iommu_assign_device(struct dmar_domain *domain,
> >> + struct pci_dev *pdev);
> >> +void intel_iommu_deassign_device(struct dmar_domain *domain,
> >> + struct pci_dev *pdev);
> >> +int intel_iommu_map_pages(struct dmar_domain *domain, dma_addr_t
> >> iova, + u64 hpa, size_t size, int prot);
> >> +void intel_iommu_unmap_pages(struct dmar_domain *domain,
> >> + dma_addr_t iova, size_t size);
> >> u64 intel_iommu_iova_to_pfn(struct dmar_domain *domain, u64 iova);
> >>
> >> #ifdef CONFIG_DMAR
> >> --
> >> 1.5.1
>
>
--
| AMD Saxony Limited Liability Company & Co. KG
Operating | Wilschdorfer Landstr. 101, 01109 Dresden, Germany
System | Register Court Dresden: HRA 4896
Research | General Partner authorized to represent:
Center | AMD Saxony LLC (Wilmington, Delaware, US)
| General Manager of AMD Saxony LLC: Dr. Hans-R. Deppe, Thomas McCoy
^ permalink raw reply [flat|nested] 7+ messages in thread
* RE: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
2008-12-01 11:55 ` 'Joerg Roedel'
@ 2008-12-01 13:22 ` Han, Weidong
2008-12-01 14:04 ` Joerg Roedel
0 siblings, 1 reply; 7+ messages in thread
From: Han, Weidong @ 2008-12-01 13:22 UTC (permalink / raw)
To: 'Joerg Roedel'
Cc: Woodhouse, David, 'Jesse Barnes', 'Avi Kivity',
Kay, Allen M, Yu, Fenghua, 'kvm@vger.kernel.org',
'iommu@lists.linux-foundation.org'
Sorry, this patch has style problem. I will update it and also split it to smaller patches for easy reviewing.
Regards,
Weidong
'Joerg Roedel' wrote:
> Hmm, I get these errors using git-am:
>
> Applying VT-d: Support multiple device assignment for KVM
> .dotest/patch:1344: space before tab in indent.
> clflush_cache_range(addr, size);
> .dotest/patch:1350: space before tab in indent.
> clflush_cache_range(addr, size);
> .dotest/patch:1907: trailing whitespace.
>
> .dotest/patch:1946: trailing whitespace.
> * owned by this domain, clear this iommu in iommu_bmp
> .dotest/patch:2300: trailing whitespace.
>
> error: patch failed: drivers/pci/dmar.c:484
> error: drivers/pci/dmar.c: patch does not apply
> error: patch failed: drivers/pci/intel-iommu.c:50
> error: drivers/pci/intel-iommu.c: patch does not apply
> error: patch failed: include/linux/dma_remapping.h:111
> error: include/linux/dma_remapping.h: patch does not apply
> error: patch failed: include/linux/intel-iommu.h:219
> error: include/linux/intel-iommu.h: patch does not apply
> Patch failed at 0001.
>
> Joerg
>
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
2008-12-01 13:22 ` Han, Weidong
@ 2008-12-01 14:04 ` Joerg Roedel
2008-12-02 0:31 ` Han, Weidong
0 siblings, 1 reply; 7+ messages in thread
From: Joerg Roedel @ 2008-12-01 14:04 UTC (permalink / raw)
To: Han, Weidong
Cc: 'Joerg Roedel', Woodhouse, David, 'Jesse Barnes',
'Avi Kivity', Kay, Allen M, Yu, Fenghua,
'kvm@vger.kernel.org',
'iommu@lists.linux-foundation.org'
Ok, I got them to apply. I also did the checkpatch cleanups. To speed
things up a bit I would suggest that I rebase my patchset on your
patches and send it out in a single series. Any problems with this
approach?
Joerg
On Mon, Dec 01, 2008 at 09:22:42PM +0800, Han, Weidong wrote:
> Sorry, this patch has style problem. I will update it and also split it to smaller patches for easy reviewing.
>
> Regards,
> Weidong
>
> 'Joerg Roedel' wrote:
> > Hmm, I get these errors using git-am:
> >
> > Applying VT-d: Support multiple device assignment for KVM
> > .dotest/patch:1344: space before tab in indent.
> > clflush_cache_range(addr, size);
> > .dotest/patch:1350: space before tab in indent.
> > clflush_cache_range(addr, size);
> > .dotest/patch:1907: trailing whitespace.
> >
> > .dotest/patch:1946: trailing whitespace.
> > * owned by this domain, clear this iommu in iommu_bmp
> > .dotest/patch:2300: trailing whitespace.
> >
> > error: patch failed: drivers/pci/dmar.c:484
> > error: drivers/pci/dmar.c: patch does not apply
> > error: patch failed: drivers/pci/intel-iommu.c:50
> > error: drivers/pci/intel-iommu.c: patch does not apply
> > error: patch failed: include/linux/dma_remapping.h:111
> > error: include/linux/dma_remapping.h: patch does not apply
> > error: patch failed: include/linux/intel-iommu.h:219
> > error: include/linux/intel-iommu.h: patch does not apply
> > Patch failed at 0001.
> >
> > Joerg
> >
> --
> To unsubscribe from this list: send the line "unsubscribe kvm" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 7+ messages in thread
* RE: [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM
2008-12-01 14:04 ` Joerg Roedel
@ 2008-12-02 0:31 ` Han, Weidong
0 siblings, 0 replies; 7+ messages in thread
From: Han, Weidong @ 2008-12-02 0:31 UTC (permalink / raw)
To: 'Joerg Roedel'
Cc: 'Joerg Roedel', Woodhouse, David, 'Jesse Barnes',
'Avi Kivity', Kay, Allen M, Yu, Fenghua,
'kvm@vger.kernel.org',
'iommu@lists.linux-foundation.org'
It's fine. You only needs to change the APIs to generic APIs. I will update it soon.
Regards,
Weidong
Joerg Roedel wrote:
> Ok, I got them to apply. I also did the checkpatch cleanups. To speed
> things up a bit I would suggest that I rebase my patchset on your
> patches and send it out in a single series. Any problems with this
> approach?
>
> Joerg
>
> On Mon, Dec 01, 2008 at 09:22:42PM +0800, Han, Weidong wrote:
>> Sorry, this patch has style problem. I will update it and also split
>> it to smaller patches for easy reviewing.
>>
>> Regards,
>> Weidong
>>
>> 'Joerg Roedel' wrote:
>>> Hmm, I get these errors using git-am:
>>>
>>> Applying VT-d: Support multiple device assignment for KVM
>>> .dotest/patch:1344: space before tab in indent.
>>> clflush_cache_range(addr, size);
>>> .dotest/patch:1350: space before tab in indent.
>>> clflush_cache_range(addr, size);
>>> .dotest/patch:1907: trailing whitespace.
>>>
>>> .dotest/patch:1946: trailing whitespace.
>>> * owned by this domain, clear this iommu in
>>> iommu_bmp .dotest/patch:2300: trailing whitespace.
>>>
>>> error: patch failed: drivers/pci/dmar.c:484
>>> error: drivers/pci/dmar.c: patch does not apply
>>> error: patch failed: drivers/pci/intel-iommu.c:50
>>> error: drivers/pci/intel-iommu.c: patch does not apply
>>> error: patch failed: include/linux/dma_remapping.h:111
>>> error: include/linux/dma_remapping.h: patch does not apply
>>> error: patch failed: include/linux/intel-iommu.h:219
>>> error: include/linux/intel-iommu.h: patch does not apply
>>> Patch failed at 0001.
>>>
>>> Joerg
>>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe kvm" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 7+ messages in thread
end of thread, other threads:[~2008-12-02 0:31 UTC | newest]
Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2008-11-27 13:49 [PATCH 1/2] [v2] VT-d: Support multiple device assignment for KVM Han, Weidong
2008-11-28 13:27 ` Joerg Roedel
2008-12-01 6:17 ` Han, Weidong
2008-12-01 11:55 ` 'Joerg Roedel'
2008-12-01 13:22 ` Han, Weidong
2008-12-01 14:04 ` Joerg Roedel
2008-12-02 0:31 ` Han, Weidong
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox