2008-09-14 04:48:28 +04:00
/*
* Copyright ( c ) 2006 , Intel Corporation .
*
* This program is free software ; you can redistribute it and / or modify it
* under the terms and conditions of the GNU General Public License ,
* version 2 , as published by the Free Software Foundation .
*
* This program is distributed in the hope it will be useful , but WITHOUT
* ANY WARRANTY ; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE . See the GNU General Public License for
* more details .
*
* You should have received a copy of the GNU General Public License along with
* this program ; if not , write to the Free Software Foundation , Inc . , 59 Temple
* Place - Suite 330 , Boston , MA 02111 - 1307 USA .
*
* Copyright ( C ) 2006 - 2008 Intel Corporation
* Copyright IBM Corporation , 2008
2010-05-23 19:37:00 +04:00
* Copyright 2010 Red Hat , Inc . and / or its affiliates .
*
2008-09-14 04:48:28 +04:00
* Author : Allen M . Kay < allen . m . kay @ intel . com >
* Author : Weidong Han < weidong . han @ intel . com >
* Author : Ben - Ami Yassour < benami @ il . ibm . com >
*/
# include <linux/list.h>
# include <linux/kvm_host.h>
2011-07-28 05:25:05 +04:00
# include <linux/module.h>
2008-09-14 04:48:28 +04:00
# include <linux/pci.h>
2011-07-28 05:17:59 +04:00
# include <linux/stat.h>
2008-09-14 04:48:28 +04:00
# include <linux/dmar.h>
2008-12-03 16:43:34 +03:00
# include <linux/iommu.h>
2008-09-14 04:48:28 +04:00
# include <linux/intel-iommu.h>
2012-01-13 03:02:20 +04:00
static bool allow_unsafe_assigned_interrupts ;
2011-07-14 23:27:03 +04:00
module_param_named ( allow_unsafe_assigned_interrupts ,
allow_unsafe_assigned_interrupts , bool , S_IRUGO | S_IWUSR ) ;
MODULE_PARM_DESC ( allow_unsafe_assigned_interrupts ,
" Enable device assignment on platforms without interrupt remapping support. " ) ;
2008-09-14 04:48:28 +04:00
static int kvm_iommu_unmap_memslots ( struct kvm * kvm ) ;
static void kvm_iommu_put_pages ( struct kvm * kvm ,
gfn_t base_gfn , unsigned long npages ) ;
2012-07-17 17:56:16 +04:00
static pfn_t kvm_pin_pages ( struct kvm_memory_slot * slot , gfn_t gfn ,
unsigned long size )
2010-01-11 18:38:18 +03:00
{
gfn_t end_gfn ;
pfn_t pfn ;
2012-07-17 17:56:16 +04:00
pfn = gfn_to_pfn_memslot ( slot , gfn ) ;
2010-01-11 18:38:18 +03:00
end_gfn = gfn + ( size > > PAGE_SHIFT ) ;
gfn + = 1 ;
2012-10-16 16:10:59 +04:00
if ( is_error_noslot_pfn ( pfn ) )
2010-01-11 18:38:18 +03:00
return pfn ;
while ( gfn < end_gfn )
2012-07-17 17:56:16 +04:00
gfn_to_pfn_memslot ( slot , gfn + + ) ;
2010-01-11 18:38:18 +03:00
return pfn ;
}
2009-12-23 19:35:20 +03:00
int kvm_iommu_map_pages ( struct kvm * kvm , struct kvm_memory_slot * slot )
2008-09-14 04:48:28 +04:00
{
2010-01-11 18:38:18 +03:00
gfn_t gfn , end_gfn ;
2008-09-14 04:48:28 +04:00
pfn_t pfn ;
2010-01-11 18:38:18 +03:00
int r = 0 ;
2008-12-03 16:43:34 +03:00
struct iommu_domain * domain = kvm - > arch . iommu_domain ;
2009-04-27 16:35:43 +04:00
int flags ;
2008-09-14 04:48:28 +04:00
/* check if iommu exists and in use */
if ( ! domain )
return 0 ;
2010-01-11 18:38:18 +03:00
gfn = slot - > base_gfn ;
end_gfn = gfn + slot - > npages ;
2013-01-25 02:04:09 +04:00
flags = IOMMU_READ ;
if ( ! ( slot - > flags & KVM_MEM_READONLY ) )
flags | = IOMMU_WRITE ;
2013-10-30 21:02:23 +04:00
if ( ! kvm - > arch . iommu_noncoherent )
2009-04-27 16:35:43 +04:00
flags | = IOMMU_CACHE ;
2010-01-11 18:38:18 +03:00
while ( gfn < end_gfn ) {
unsigned long page_size ;
/* Check if already mapped */
if ( iommu_iova_to_phys ( domain , gfn_to_gpa ( gfn ) ) ) {
gfn + = 1 ;
continue ;
}
/* Get the page size we could use to map */
page_size = kvm_host_page_size ( kvm , gfn ) ;
/* Make sure the page_size does not exceed the memslot */
while ( ( gfn + ( page_size > > PAGE_SHIFT ) ) > end_gfn )
page_size > > = 1 ;
/* Make sure gfn is aligned to the page size we want to map */
while ( ( gfn < < PAGE_SHIFT ) & ( page_size - 1 ) )
page_size > > = 1 ;
KVM: IOMMU: hva align mapping page size
When determining the page size we could use to map with the IOMMU, the
page size should also be aligned with the hva, not just the gfn. The
gfn may not reflect the real alignment within the hugetlbfs file.
Most of the time, this works fine. However, if the hugetlbfs file is
backed by non-contiguous huge pages, a multi-huge page memslot starts at
an unaligned offset within the hugetlbfs file, and the gfn is aligned
with respect to the huge page size, kvm_host_page_size() will return the
huge page size and we will use that to map with the IOMMU.
When we later unpin that same memslot, the IOMMU returns the unmap size
as the huge page size, and we happily unpin that many pfns in
monotonically increasing order, not realizing we are spanning
non-contiguous huge pages and partially unpin the wrong huge page.
Ensure the IOMMU mapping page size is aligned with the hva corresponding
to the gfn, which does reflect the alignment within the hugetlbfs file.
Reviewed-by: Marcelo Tosatti <mtosatti@redhat.com>
Signed-off-by: Greg Edwards <gedwards@ddn.com>
Cc: stable@vger.kernel.org
Signed-off-by: Gleb Natapov <gleb@redhat.com>
2013-11-04 20:08:12 +04:00
/* Make sure hva is aligned to the page size we want to map */
while ( __gfn_to_hva_memslot ( slot , gfn ) & ( page_size - 1 ) )
page_size > > = 1 ;
2010-01-11 18:38:18 +03:00
/*
* Pin all pages we are about to map in memory . This is
* important because we unmap and unpin in 4 kb steps later .
*/
2012-07-17 17:56:16 +04:00
pfn = kvm_pin_pages ( slot , gfn , page_size ) ;
2012-10-16 16:10:59 +04:00
if ( is_error_noslot_pfn ( pfn ) ) {
2010-01-11 18:38:18 +03:00
gfn + = 1 ;
2008-09-14 04:48:28 +04:00
continue ;
2010-01-11 18:38:18 +03:00
}
2008-09-14 04:48:28 +04:00
2010-01-11 18:38:18 +03:00
/* Map into IO address space */
r = iommu_map ( domain , gfn_to_gpa ( gfn ) , pfn_to_hpa ( pfn ) ,
iommu/core: split mapping to page sizes as supported by the hardware
When mapping a memory region, split it to page sizes as supported
by the iommu hardware. Always prefer bigger pages, when possible,
in order to reduce the TLB pressure.
The logic to do that is now added to the IOMMU core, so neither the iommu
drivers themselves nor users of the IOMMU API have to duplicate it.
This allows a more lenient granularity of mappings; traditionally the
IOMMU API took 'order' (of a page) as a mapping size, and directly let
the low level iommu drivers handle the mapping, but now that the IOMMU
core can split arbitrary memory regions into pages, we can remove this
limitation, so users don't have to split those regions by themselves.
Currently the supported page sizes are advertised once and they then
remain static. That works well for OMAP and MSM but it would probably
not fly well with intel's hardware, where the page size capabilities
seem to have the potential to be different between several DMA
remapping devices.
register_iommu() currently sets a default pgsize behavior, so we can convert
the IOMMU drivers in subsequent patches. After all the drivers
are converted, the temporary default settings will be removed.
Mainline users of the IOMMU API (kvm and omap-iovmm) are adopted
to deal with bytes instead of page order.
Many thanks to Joerg Roedel <Joerg.Roedel@amd.com> for significant review!
Signed-off-by: Ohad Ben-Cohen <ohad@wizery.com>
Cc: David Brown <davidb@codeaurora.org>
Cc: David Woodhouse <dwmw2@infradead.org>
Cc: Joerg Roedel <Joerg.Roedel@amd.com>
Cc: Stepan Moskovchenko <stepanm@codeaurora.org>
Cc: KyongHo Cho <pullip.cho@samsung.com>
Cc: Hiroshi DOYU <hdoyu@nvidia.com>
Cc: Laurent Pinchart <laurent.pinchart@ideasonboard.com>
Cc: kvm@vger.kernel.org
Signed-off-by: Joerg Roedel <joerg.roedel@amd.com>
2011-11-10 13:32:26 +04:00
page_size , flags ) ;
2008-09-25 19:32:02 +04:00
if ( r ) {
2008-12-02 16:03:39 +03:00
printk ( KERN_ERR " kvm_iommu_map_address: "
2010-07-01 18:00:12 +04:00
" iommu failed to map pfn=%llx \n " , pfn ) ;
2008-09-14 04:48:28 +04:00
goto unmap_pages ;
}
2010-01-11 18:38:18 +03:00
gfn + = page_size > > PAGE_SHIFT ;
2008-09-14 04:48:28 +04:00
}
2010-01-11 18:38:18 +03:00
2008-09-14 04:48:28 +04:00
return 0 ;
unmap_pages :
2010-01-11 18:38:18 +03:00
kvm_iommu_put_pages ( kvm , slot - > base_gfn , gfn ) ;
2008-09-14 04:48:28 +04:00
return r ;
}
static int kvm_iommu_map_memslots ( struct kvm * kvm )
{
2011-11-24 13:39:18 +04:00
int idx , r = 0 ;
2009-12-23 19:35:16 +03:00
struct kvm_memslots * slots ;
2011-11-24 13:39:18 +04:00
struct kvm_memory_slot * memslot ;
2008-09-14 04:48:28 +04:00
2013-10-30 21:02:30 +04:00
if ( kvm - > arch . iommu_noncoherent )
kvm_arch_register_noncoherent_dma ( kvm ) ;
2010-07-01 11:00:50 +04:00
idx = srcu_read_lock ( & kvm - > srcu ) ;
2010-04-19 13:41:23 +04:00
slots = kvm_memslots ( kvm ) ;
2009-12-23 19:35:16 +03:00
2011-11-24 13:39:18 +04:00
kvm_for_each_memslot ( memslot , slots ) {
r = kvm_iommu_map_pages ( kvm , memslot ) ;
2008-09-14 04:48:28 +04:00
if ( r )
break ;
}
2010-07-01 11:00:50 +04:00
srcu_read_unlock ( & kvm - > srcu , idx ) ;
2009-02-05 21:23:46 +03:00
2008-09-14 04:48:28 +04:00
return r ;
}
2008-12-02 16:03:39 +03:00
int kvm_assign_device ( struct kvm * kvm ,
struct kvm_assigned_dev_kernel * assigned_dev )
2008-09-14 04:48:28 +04:00
{
struct pci_dev * pdev = NULL ;
2008-12-03 16:43:34 +03:00
struct iommu_domain * domain = kvm - > arch . iommu_domain ;
2013-10-30 21:02:23 +04:00
int r ;
bool noncoherent ;
2008-09-14 04:48:28 +04:00
2008-12-02 16:03:39 +03:00
/* check if iommu exists and in use */
if ( ! domain )
return 0 ;
pdev = assigned_dev - > dev ;
if ( pdev = = NULL )
2008-09-14 04:48:28 +04:00
return - ENODEV ;
2008-12-02 16:03:39 +03:00
2008-12-03 16:43:34 +03:00
r = iommu_attach_device ( domain , & pdev - > dev ) ;
2008-12-02 16:03:39 +03:00
if ( r ) {
2012-10-09 04:36:11 +04:00
dev_err ( & pdev - > dev , " kvm assign device failed ret %d " , r ) ;
2008-12-02 16:03:39 +03:00
return r ;
2008-09-14 04:48:28 +04:00
}
2013-10-30 21:02:23 +04:00
noncoherent = ! iommu_domain_has_cap ( kvm - > arch . iommu_domain ,
IOMMU_CAP_CACHE_COHERENCY ) ;
2009-04-27 16:35:43 +04:00
/* Check if need to update IOMMU page table for guest memory */
2013-10-30 21:02:23 +04:00
if ( noncoherent ! = kvm - > arch . iommu_noncoherent ) {
2009-04-27 16:35:43 +04:00
kvm_iommu_unmap_memslots ( kvm ) ;
2013-10-30 21:02:23 +04:00
kvm - > arch . iommu_noncoherent = noncoherent ;
2009-04-27 16:35:43 +04:00
r = kvm_iommu_map_memslots ( kvm ) ;
if ( r )
goto out_unmap ;
}
2011-07-22 09:46:07 +04:00
pdev - > dev_flags | = PCI_DEV_FLAGS_ASSIGNED ;
2013-10-02 14:23:26 +04:00
dev_info ( & pdev - > dev , " kvm assign device \n " ) ;
2008-09-14 04:48:28 +04:00
2008-12-02 16:03:39 +03:00
return 0 ;
2009-04-27 16:35:43 +04:00
out_unmap :
kvm_iommu_unmap_memslots ( kvm ) ;
return r ;
2008-12-02 16:03:39 +03:00
}
2008-09-14 04:48:28 +04:00
2008-12-02 16:24:23 +03:00
int kvm_deassign_device ( struct kvm * kvm ,
struct kvm_assigned_dev_kernel * assigned_dev )
{
2008-12-03 16:43:34 +03:00
struct iommu_domain * domain = kvm - > arch . iommu_domain ;
2008-12-02 16:24:23 +03:00
struct pci_dev * pdev = NULL ;
/* check if iommu exists and in use */
if ( ! domain )
return 0 ;
pdev = assigned_dev - > dev ;
if ( pdev = = NULL )
return - ENODEV ;
2008-12-03 16:43:34 +03:00
iommu_detach_device ( domain , & pdev - > dev ) ;
2008-12-02 16:24:23 +03:00
2011-07-22 09:46:07 +04:00
pdev - > dev_flags & = ~ PCI_DEV_FLAGS_ASSIGNED ;
2013-10-02 14:23:26 +04:00
dev_info ( & pdev - > dev , " kvm deassign device \n " ) ;
2008-12-02 16:24:23 +03:00
return 0 ;
}
2008-12-02 16:03:39 +03:00
int kvm_iommu_map_guest ( struct kvm * kvm )
{
int r ;
2011-09-06 20:46:34 +04:00
if ( ! iommu_present ( & pci_bus_type ) ) {
2008-12-03 16:43:34 +03:00
printk ( KERN_ERR " %s: iommu not found \n " , __func__ ) ;
2008-09-14 04:48:28 +04:00
return - ENODEV ;
}
2012-04-18 07:46:44 +04:00
mutex_lock ( & kvm - > slots_lock ) ;
2011-09-06 18:03:26 +04:00
kvm - > arch . iommu_domain = iommu_domain_alloc ( & pci_bus_type ) ;
2012-04-18 07:46:44 +04:00
if ( ! kvm - > arch . iommu_domain ) {
r = - ENOMEM ;
goto out_unlock ;
}
2008-09-14 04:48:28 +04:00
2011-07-14 23:27:03 +04:00
if ( ! allow_unsafe_assigned_interrupts & &
! iommu_domain_has_cap ( kvm - > arch . iommu_domain ,
IOMMU_CAP_INTR_REMAP ) ) {
printk ( KERN_WARNING " %s: No interrupt remapping support, "
" disallowing device assignment. "
" Re-enble with \" allow_unsafe_assigned_interrupts=1 \" "
" module option. \n " , __func__ ) ;
iommu_domain_free ( kvm - > arch . iommu_domain ) ;
kvm - > arch . iommu_domain = NULL ;
2012-04-18 07:46:44 +04:00
r = - EPERM ;
goto out_unlock ;
2011-07-14 23:27:03 +04:00
}
2008-09-14 04:48:28 +04:00
r = kvm_iommu_map_memslots ( kvm ) ;
if ( r )
2012-04-18 07:46:44 +04:00
kvm_iommu_unmap_memslots ( kvm ) ;
2008-09-14 04:48:28 +04:00
2012-04-18 07:46:44 +04:00
out_unlock :
mutex_unlock ( & kvm - > slots_lock ) ;
2008-09-14 04:48:28 +04:00
return r ;
}
2010-01-11 18:38:18 +03:00
static void kvm_unpin_pages ( struct kvm * kvm , pfn_t pfn , unsigned long npages )
{
unsigned long i ;
for ( i = 0 ; i < npages ; + + i )
kvm_release_pfn_clean ( pfn + i ) ;
}
2008-09-14 04:48:28 +04:00
static void kvm_iommu_put_pages ( struct kvm * kvm ,
2008-12-02 16:03:39 +03:00
gfn_t base_gfn , unsigned long npages )
2008-09-14 04:48:28 +04:00
{
2010-01-11 18:38:18 +03:00
struct iommu_domain * domain ;
gfn_t end_gfn , gfn ;
2008-09-14 04:48:28 +04:00
pfn_t pfn ;
2008-12-02 16:03:39 +03:00
u64 phys ;
2010-01-11 18:38:18 +03:00
domain = kvm - > arch . iommu_domain ;
end_gfn = base_gfn + npages ;
gfn = base_gfn ;
2008-12-02 16:03:39 +03:00
/* check if iommu exists and in use */
if ( ! domain )
return ;
2008-09-14 04:48:28 +04:00
2010-01-11 18:38:18 +03:00
while ( gfn < end_gfn ) {
unsigned long unmap_pages ;
iommu/core: split mapping to page sizes as supported by the hardware
When mapping a memory region, split it to page sizes as supported
by the iommu hardware. Always prefer bigger pages, when possible,
in order to reduce the TLB pressure.
The logic to do that is now added to the IOMMU core, so neither the iommu
drivers themselves nor users of the IOMMU API have to duplicate it.
This allows a more lenient granularity of mappings; traditionally the
IOMMU API took 'order' (of a page) as a mapping size, and directly let
the low level iommu drivers handle the mapping, but now that the IOMMU
core can split arbitrary memory regions into pages, we can remove this
limitation, so users don't have to split those regions by themselves.
Currently the supported page sizes are advertised once and they then
remain static. That works well for OMAP and MSM but it would probably
not fly well with intel's hardware, where the page size capabilities
seem to have the potential to be different between several DMA
remapping devices.
register_iommu() currently sets a default pgsize behavior, so we can convert
the IOMMU drivers in subsequent patches. After all the drivers
are converted, the temporary default settings will be removed.
Mainline users of the IOMMU API (kvm and omap-iovmm) are adopted
to deal with bytes instead of page order.
Many thanks to Joerg Roedel <Joerg.Roedel@amd.com> for significant review!
Signed-off-by: Ohad Ben-Cohen <ohad@wizery.com>
Cc: David Brown <davidb@codeaurora.org>
Cc: David Woodhouse <dwmw2@infradead.org>
Cc: Joerg Roedel <Joerg.Roedel@amd.com>
Cc: Stepan Moskovchenko <stepanm@codeaurora.org>
Cc: KyongHo Cho <pullip.cho@samsung.com>
Cc: Hiroshi DOYU <hdoyu@nvidia.com>
Cc: Laurent Pinchart <laurent.pinchart@ideasonboard.com>
Cc: kvm@vger.kernel.org
Signed-off-by: Joerg Roedel <joerg.roedel@amd.com>
2011-11-10 13:32:26 +04:00
size_t size ;
2010-01-11 18:38:18 +03:00
/* Get physical address */
2008-12-03 16:43:34 +03:00
phys = iommu_iova_to_phys ( domain , gfn_to_gpa ( gfn ) ) ;
2012-08-03 11:36:52 +04:00
if ( ! phys ) {
gfn + + ;
continue ;
}
2010-01-11 18:38:18 +03:00
pfn = phys > > PAGE_SHIFT ;
/* Unmap address from IO address space */
iommu/core: split mapping to page sizes as supported by the hardware
When mapping a memory region, split it to page sizes as supported
by the iommu hardware. Always prefer bigger pages, when possible,
in order to reduce the TLB pressure.
The logic to do that is now added to the IOMMU core, so neither the iommu
drivers themselves nor users of the IOMMU API have to duplicate it.
This allows a more lenient granularity of mappings; traditionally the
IOMMU API took 'order' (of a page) as a mapping size, and directly let
the low level iommu drivers handle the mapping, but now that the IOMMU
core can split arbitrary memory regions into pages, we can remove this
limitation, so users don't have to split those regions by themselves.
Currently the supported page sizes are advertised once and they then
remain static. That works well for OMAP and MSM but it would probably
not fly well with intel's hardware, where the page size capabilities
seem to have the potential to be different between several DMA
remapping devices.
register_iommu() currently sets a default pgsize behavior, so we can convert
the IOMMU drivers in subsequent patches. After all the drivers
are converted, the temporary default settings will be removed.
Mainline users of the IOMMU API (kvm and omap-iovmm) are adopted
to deal with bytes instead of page order.
Many thanks to Joerg Roedel <Joerg.Roedel@amd.com> for significant review!
Signed-off-by: Ohad Ben-Cohen <ohad@wizery.com>
Cc: David Brown <davidb@codeaurora.org>
Cc: David Woodhouse <dwmw2@infradead.org>
Cc: Joerg Roedel <Joerg.Roedel@amd.com>
Cc: Stepan Moskovchenko <stepanm@codeaurora.org>
Cc: KyongHo Cho <pullip.cho@samsung.com>
Cc: Hiroshi DOYU <hdoyu@nvidia.com>
Cc: Laurent Pinchart <laurent.pinchart@ideasonboard.com>
Cc: kvm@vger.kernel.org
Signed-off-by: Joerg Roedel <joerg.roedel@amd.com>
2011-11-10 13:32:26 +04:00
size = iommu_unmap ( domain , gfn_to_gpa ( gfn ) , PAGE_SIZE ) ;
unmap_pages = 1ULL < < get_order ( size ) ;
2008-12-02 16:03:39 +03:00
2010-01-11 18:38:18 +03:00
/* Unpin all pages we just unmapped to not leak any memory */
kvm_unpin_pages ( kvm , pfn , unmap_pages ) ;
gfn + = unmap_pages ;
}
2008-09-14 04:48:28 +04:00
}
2012-04-11 19:51:49 +04:00
void kvm_iommu_unmap_pages ( struct kvm * kvm , struct kvm_memory_slot * slot )
{
kvm_iommu_put_pages ( kvm , slot - > base_gfn , slot - > npages ) ;
}
2008-09-14 04:48:28 +04:00
static int kvm_iommu_unmap_memslots ( struct kvm * kvm )
{
2011-11-24 13:39:18 +04:00
int idx ;
2009-12-23 19:35:16 +03:00
struct kvm_memslots * slots ;
2011-11-24 13:39:18 +04:00
struct kvm_memory_slot * memslot ;
2009-12-23 19:35:16 +03:00
2010-07-01 11:00:50 +04:00
idx = srcu_read_lock ( & kvm - > srcu ) ;
2010-04-19 13:41:23 +04:00
slots = kvm_memslots ( kvm ) ;
2009-02-05 21:23:46 +03:00
2011-11-24 13:39:18 +04:00
kvm_for_each_memslot ( memslot , slots )
2012-04-11 19:51:49 +04:00
kvm_iommu_unmap_pages ( kvm , memslot ) ;
2011-11-24 13:39:18 +04:00
2010-07-01 11:00:50 +04:00
srcu_read_unlock ( & kvm - > srcu , idx ) ;
2008-09-14 04:48:28 +04:00
2013-10-30 21:02:30 +04:00
if ( kvm - > arch . iommu_noncoherent )
kvm_arch_unregister_noncoherent_dma ( kvm ) ;
2008-09-14 04:48:28 +04:00
return 0 ;
}
int kvm_iommu_unmap_guest ( struct kvm * kvm )
{
2008-12-03 16:43:34 +03:00
struct iommu_domain * domain = kvm - > arch . iommu_domain ;
2008-09-14 04:48:28 +04:00
/* check if iommu exists and in use */
if ( ! domain )
return 0 ;
2012-04-18 07:46:44 +04:00
mutex_lock ( & kvm - > slots_lock ) ;
2008-09-14 04:48:28 +04:00
kvm_iommu_unmap_memslots ( kvm ) ;
2012-04-18 07:46:44 +04:00
kvm - > arch . iommu_domain = NULL ;
2013-10-30 21:02:23 +04:00
kvm - > arch . iommu_noncoherent = false ;
2012-04-18 07:46:44 +04:00
mutex_unlock ( & kvm - > slots_lock ) ;
2008-12-03 16:43:34 +03:00
iommu_domain_free ( domain ) ;
2008-09-14 04:48:28 +04:00
return 0 ;
}