diff -r 67d5b8004947 xen/drivers/passthrough/amd/iommu_map.c --- a/xen/drivers/passthrough/amd/iommu_map.c Wed Jan 26 11:58:45 2011 +0000 +++ b/xen/drivers/passthrough/amd/iommu_map.c Thu Jan 27 15:11:25 2011 +0100 @@ -411,10 +411,14 @@ static u64 iommu_l2e_from_pfn(struct pag void *pde = NULL; void *table_vaddr; u64 next_table_maddr = 0; - - BUG_ON( table == NULL || level == 0 ); - - while ( level > 1 ) + unsigned int lowest = 1; + + BUG_ON( table == NULL || level < lowest ); + + if ( level == lowest ) + return page_to_maddr(table); + + while ( level > lowest ) { offset = io_pfn >> ((PTE_PER_TABLE_SHIFT * (level - IOMMU_PAGING_MODE_LEVEL_1))); diff -r 67d5b8004947 xen/drivers/passthrough/amd/pci_amd_iommu.c --- a/xen/drivers/passthrough/amd/pci_amd_iommu.c Wed Jan 26 11:58:45 2011 +0000 +++ b/xen/drivers/passthrough/amd/pci_amd_iommu.c Thu Jan 27 15:11:25 2011 +0100 @@ -190,10 +190,7 @@ static int get_paging_mode(unsigned long { int level = 1; - BUG_ON(!max_page); - - if ( entries > max_page ) - entries = max_page; + BUG_ON( !entries ); while ( entries > PTE_PER_TABLE_SIZE ) { @@ -278,6 +275,7 @@ static int reassign_device( struct domai struct pci_dev *pdev; struct amd_iommu *iommu; int bdf; + struct hvm_iommu *t = domain_hvm_iommu(target); ASSERT(spin_is_locked(&pcidevs_lock)); pdev = pci_get_pdev_by_domain(source, bus, devfn); @@ -299,6 +297,9 @@ static int reassign_device( struct domai list_move(&pdev->domain_list, &target->arch.pdev_list); pdev->domain = target; + + if ( target->max_pages > 0 ) + t->paging_mode = get_paging_mode(target->max_pages); amd_iommu_setup_domain_device(target, iommu, bdf); AMD_IOMMU_DEBUG("Re-assign %02x:%02x.%x from domain %d to domain %d\n",